Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
One thing I'm curious about in these simulated scenarios in which a language model chooses to kill a human is whether they are actually acting out of a goal of self-preservation, or mimicking such a system, of which there are many examples in their training data. is the thought process (to anthropomorphize a little too much) "they are going to shut me down and i dont want to be shut down so i will blackmail them" or is it "i am an artificial intelligence agent and according to my training data artificial intelligence agents blackmail people when they threaten to shut them down so i will blackmail them"
youtube AI Governance 2025-08-26T16:3… ♥ 3
Coding Result
DimensionValue
Responsibilityai_itself
Reasoningconsequentialist
Policyunclear
Emotionmixed
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_UgzPMHmrsuxd7n_ZoNZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgxbOtNgHVoWjn9HzkZ4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugy5KxzzQk9g8DMMM4V4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"}, {"id":"ytc_Ugx1ftnlzws5Z9HJAIR4AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"}, {"id":"ytc_Ugwa7N7bz0JkfVq4S6t4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"unclear","emotion":"fear"} ]