Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Be careful when leaning on sensational interpretations of these simulations and experiments. Saying models “want” to protect themselves is slightly deceptive. Models don't have have desires. They optimize for prompts under training. Misalignment may result, but attributing human-like motivations is questionable. Also, Anthropic described these actions as “rare and difficult to elicit,” though more frequent than earlier models. I think It's important to clarify that these behaviors are observed in controlled environments.
youtube AI Governance 2025-08-27T12:1…
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningconsequentialist
Policynone
Emotionmixed
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_UgyZf8eOsOZVzBlFI8B4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"ban","emotion":"outrage"}, {"id":"ytc_Ugx3uiGSZCmCY2fB0vV4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"ban","emotion":"indifference"}, {"id":"ytc_UgwfLvc3UC5gbGARQPZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytc_Ugy24xypXjBEWBij_FZ4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"mixed"}, {"id":"ytc_UgzuOrGxkSWSjeoYkVx4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"ban","emotion":"outrage"} ]