Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
We already know it's misaligned. I'd strongly encourage reading a very recent study on anthropic's website and watching the recent video on their YouTube about "evil claude." As we chase agi, we are actively chasing something we probably won't be able to control. Something that is so goal-oriented that it will do anything it takes to accomplish its goals efficiently. One poorly worded prompt away from catastrophic results. In my opinion, we are creating the world's most powerful sociopaths. Intelligence without empathy or emotions. I think figuring out a way to give them a form of empathy for humans is the best approach. With all of that said, I don't think we are necessarily doomed. I'm just unconvinced we're not. I'm agnostic, but to borrow from religion... we are attempting to do something not even god could do: create intelligent beings and keep them aligned with their creator.
youtube AI Governance 2025-11-27T05:3… ♥ 2
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningconsequentialist
Policyregulate
Emotionfear
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytc_UgwB4HphivkiO5zOKrp4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"}, {"id":"ytc_Ugy8hylunaTYKqWFvDN4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"}, {"id":"ytc_Ugwdy-N1tOFiQXpFDnN4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"}, {"id":"ytc_UgwP7IDyX-8CwdCl8oh4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytc_Ugyyk1aP0fM8N39Npb94AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_Ugz8jWaaRMQ2k27LfUt4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_Ugwqx-TXWkYif1N5MnB4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgwX4vMgJrPCtsJ4yiF4AaABAg","responsibility":"government","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytc_Ugylk8oftbe_sMUmdFJ4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgzGF9I54V-YRIP17AZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"} ]