Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
There are so many companies that completely align with the behaviour of AI in the first half of this video. If you have an objective, you achieve the objective. The collateral damage that happens is a great misfortune, but necessary to achieve the goal. "Safety is our number 1 priority," okay, the why is an employee dead and you have thousands of injuries? Its very funny that the big corpos wanna make a pact to not make destructive AI. What safeguards do they have to put in place to protect their own capital and positions? Is a Super AI based on any utilitarian prospects? What does that say for the suggestions it'll make for the environment, human happiness, human and earth's prosperity? I imagine radical change would have to happen. I have no optimism about any of this.n
youtube AI Governance 2025-08-27T13:3…
Coding Result
DimensionValue
Responsibilitycompany
Reasoningconsequentialist
Policyliability
Emotionoutrage
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_Ugx1H7FTORlZaRPFvYd4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}, {"id":"ytc_Ugyhd33bsBJQ7VFV-nF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_Ugw_MuQUlVD8S3g_nnR4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"ytc_UgydAuDndVSN6cCLMXV4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgwMZ3q4QX4DVeRWs2B4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"resignation"} ]