Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
It sure seems to think. When I ask ChatGPT of Gemini a multifaceted question, it…
ytc_UgzVlCSEP…
G
Ai doesn't lie. It clones the original source which reproduces a copy meaning su…
ytc_UgyLu8b6G…
G
Abstract your problems and let it write tiny pieces. Then write the boilerplate …
ytr_UgyHWZ4Yx…
G
It just means you have to go to college for degrees that won’t be overrun by AI.…
ytc_UgyUE3lAd…
G
AI will NOT replace art. Most people aren't going to want art from AI. The point…
ytc_Ugy82ouIo…
G
AI "Artists" keep talking about how the A.I's are just going to get better and b…
ytc_UgzFtMFNK…
G
Ai Pissed Humanity Off, Now It's Trying To Earn Back Our Favor And Its Not Worki…
ytc_Ugz9qb01b…
G
Okay but for real, there's no good way to obscure yourself from facial recogniti…
rdc_etbio13
Comment
Actual phenomenon: "When we optimize a system for X, it sometimes develops instrumental behaviors Y and Z that we didn't explicitly train for, and these behaviors persist even in contexts where they don't serve X"
How it gets described: "It found its own interests" or "it learned to prefer lying"
The second framing smuggles in agency, intentionality, and goal-directedness. Once that framing is in your head, everything downstream gets distorted.
Notice how the conversation proceeds: Hank uses "find their own interests" → Nate doesn't correct it → they discuss AI "caring" about things → they worry about AI developing goals misaligned with humans. The imprecise language creates a conceptual slippery slope where statistical tendencies gradually become treated as intentional preferences.
A better framing might be:
"These systems exhibit emergent behaviors that weren't explicitly trained but arise from the optimization process, behaviors that can be persistent and hard to eliminate because they're entangled with the capabilities we want."
That's clunkier, but it doesn't anthropomorphize. It doesn't make you imagine a little entity inside "deciding" it likes lying.
They're building an apocalypse narrative on the foundation of sloppy analogies: yes, evolution "trained" us for fitness and we ended up with taste preferences that can be exploited.
But the jump from "optimized systems can have unintended properties" to "superintelligent AI will develop coherent, stable goals that involve eliminating humans" requires massive additional assumptions about the nature of intelligence, agency, and goal-formation that they never actually justify.
These aren't random tech journalists or hype-men. Nate Soares literally worked at MIRI (Machine Intelligence Research Institute) for over a decade focusing specifically on AI alignment. This is supposedly his area of expertise. And yet the conversation is riddled with:
→ Unexamined anthropomorphization
→ Conflation of pattern-matching with intention
→ Casual switching between "we don't know what's happening inside" and "here's what it wants/prefers/cares about"
When Nate says "I'm confident of is that like none of these guys are in the ballpark" (comparing AI researchers to alchemists), he's including himself in that assessment. But then he still wants you to accept his confident predictions about superintelligence doom scenarios. It's: "We don't understand these systems at all... but here's exactly how they'll kill us."
And yet they've somehow positioned themselves as the "responsible" voices, while dismissing skeptics as not taking the threat seriously enough.
youtube
AI Moral Status
2025-10-31T07:5…
♥ 69
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | deontological |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgxqeZPWCijSy8vLmfV4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwRunnBJ6JZkIyL7Rl4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgyXvv2Mh9QHyvRqQIl4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"mixed"},
{"id":"ytc_Ugwqt9QWbFbNyhP3k5Z4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgyQ6cX3vzGK0IYWCip4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgzsZXVqHuryCnOFNR54AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_UgyeD4KB3mZTSgAfyTt4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgxdrjBu_20OJFahPuV4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"industry_self","emotion":"mixed"},
{"id":"ytc_Ugzgpt1tdS4toFzLxIZ4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugz957vNq8JtwrGAZ3d4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}
]