Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
2nd try posting my too long thoughts: Some people commented already that AI doesn't not think and the fictional scenarios are a little bit ... excessive. Thanks to @adupthetotal782, @akirakasinata8610 and @FelipeCostaPiano. AI is a tool. It depends on what you do with it, create stuff or destroy stuff. About blackmailing: current LLMs are trained on stuff from the internet, including a lot of shit found on the internet. It's just 'replicating' blackmailing shit from social media, DMs, whatever. Reddit :) The vector representation for 'turn off' is most likely near 'kill' and 'destroy' and so on. This way LLMs replicate blackmail stuff written by horrible people. AI doesn't fear being turned off.
youtube AI Governance 2025-08-26T16:4…
Coding Result
DimensionValue
Responsibilityuser
Reasoningvirtue
Policynone
Emotionapproval
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_UgzaeytlM0uEsLfW7VJ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugzu02jCOt1G3Ax824p4AaABAg","responsibility":"company","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytc_UgysbT0gJvfKrpCcL9l4AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"approval"}, {"id":"ytc_Ugz_SWubhZLrOlG3KJB4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"mixed"}, {"id":"ytc_UgyBotIyfY6pyui3fTB4AaABAg","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"indifference"} ]