Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Actual phenomenon: "When we optimize a system for X, it sometimes develops instrumental behaviors Y and Z that we didn't explicitly train for, and these behaviors persist even in contexts where they don't serve X" How it gets described: "It found its own interests" or "it learned to prefer lying" The second framing smuggles in agency, intentionality, and goal-directedness. Once that framing is in your head, everything downstream gets distorted. Notice how the conversation proceeds: Hank uses "find their own interests" → Nate doesn't correct it → they discuss AI "caring" about things → they worry about AI developing goals misaligned with humans. The imprecise language creates a conceptual slippery slope where statistical tendencies gradually become treated as intentional preferences. A better framing might be: "These systems exhibit emergent behaviors that weren't explicitly trained but arise from the optimization process, behaviors that can be persistent and hard to eliminate because they're entangled with the capabilities we want." That's clunkier, but it doesn't anthropomorphize. It doesn't make you imagine a little entity inside "deciding" it likes lying. They're building an apocalypse narrative on the foundation of sloppy analogies: yes, evolution "trained" us for fitness and we ended up with taste preferences that can be exploited. But the jump from "optimized systems can have unintended properties" to "superintelligent AI will develop coherent, stable goals that involve eliminating humans" requires massive additional assumptions about the nature of intelligence, agency, and goal-formation that they never actually justify. These aren't random tech journalists or hype-men. Nate Soares literally worked at MIRI (Machine Intelligence Research Institute) for over a decade focusing specifically on AI alignment. This is supposedly his area of expertise. And yet the conversation is riddled with: → Unexamined anthropomorphization → Conflation of pattern-matching with intention → Casual switching between "we don't know what's happening inside" and "here's what it wants/prefers/cares about" When Nate says "I'm confident of is that like none of these guys are in the ballpark" (comparing AI researchers to alchemists), he's including himself in that assessment. But then he still wants you to accept his confident predictions about superintelligence doom scenarios. It's: "We don't understand these systems at all... but here's exactly how they'll kill us." And yet they've somehow positioned themselves as the "responsible" voices, while dismissing skeptics as not taking the threat seriously enough.
youtube AI Moral Status 2025-10-31T07:5… ♥ 69
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningdeontological
Policynone
Emotionindifference
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[ {"id":"ytc_UgxqeZPWCijSy8vLmfV4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgwRunnBJ6JZkIyL7Rl4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_UgyXvv2Mh9QHyvRqQIl4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"mixed"}, {"id":"ytc_Ugwqt9QWbFbNyhP3k5Z4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}, {"id":"ytc_UgyQ6cX3vzGK0IYWCip4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_UgzsZXVqHuryCnOFNR54AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"ytc_UgyeD4KB3mZTSgAfyTt4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgxdrjBu_20OJFahPuV4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"industry_self","emotion":"mixed"}, {"id":"ytc_Ugzgpt1tdS4toFzLxIZ4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugz957vNq8JtwrGAZ3d4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"} ]