Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
@thomasseptimius Doesn't seem like a fair representation of what I have read from Yudkowsky. Are you just refering to the video? In his writing, the *why* of it is rather clear: powerful optimisation processes tend to convert available resources into a form that is beneficial for the optimisation target. This is a pretty fundamental feature of "optimisation", and when applied to a general domain (like planet Earth) means that any aspect of the current world state that is not part of that optimisation target will be changed to either directly match the target or instrumentally contribute to it (e.g. by making the attainment or sustainment of the target more probable). *Almost all* optimisation targets are incompatible with human survival because of this, so creating a superhumanly competetent, general optimiser is *almost certainly* a cause for human extinction. You can argue that future AI systems will not be in the class of systems that are here refered to as *general optimisers*, but it sure seems like we are training current systems to optimise some kinds of scores (i.e. minimize prediction loss). Afaik, mechanistic interpretability has so far brought nothing to light that would counter this assessment. So in Yudkowsky's framework, unless you make sure that your AI is not in this category of general optimisers (nor will grow into this category over time), or unless you endow it with a *close to ideal* optimisation target, building a sufficiently powerful AGI probably kills everyone. It's not that the AI is spontaneously malicious - rather, it is just not the default for optimisers to care about you, and that seems like a pretty hard thing to change.
youtube AI Governance 2024-11-12T02:2… ♥ 1
Coding Result
DimensionValue
Responsibilitynone
Reasoningconsequentialist
Policynone
Emotionindifference
Coded at2026-04-27T06:24:59.937377
Raw LLM Response
[ {"id":"ytr_UgxWbWU8HnwXTGqsyxd4AaABAg.AAiYeR1UICtAAimhHmYdLd","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytr_UgxWbWU8HnwXTGqsyxd4AaABAg.AAiYeR1UICtAAjRPne2vJB","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"outrage"}, {"id":"ytr_UgxWbWU8HnwXTGqsyxd4AaABAg.AAiYeR1UICtAAjgxLAap2x","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytr_Ugw65QRNZqXybNcskx94AaABAg.AAiXbYlEStfAAmmuu4wRqq","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"}, {"id":"ytr_Ugw65QRNZqXybNcskx94AaABAg.AAiXbYlEStfAAokZRx4eik","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"fear"}, {"id":"ytr_Ugw65QRNZqXybNcskx94AaABAg.AAiXbYlEStfAAqA9Jy7zvA","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytr_UgxJUObSq16yHsAotv54AaABAg.AAiV57RtZaaAAiWFsL8hXk","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytr_Ugz_o_4UNQo_GhkZI294AaABAg.AAiUrQt1I1zAAidX4Bf-1N","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytr_UgxWZtuw753_fj8wEad4AaABAg.AAiRc3k91HjAAjBSmlsX0M","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"resignation"}, {"id":"ytr_UgwqfDGCzxp8g181J_54AaABAg.AAiNXtAPvrsAArCz1Ewg7J","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"} ]