Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
These AI evangelists are like salesmen, fooling people that their product is the…
ytc_UgzZlfiFd…
G
"Summoning" Dan, has become somewhat of a favourite sport for curious GPT users.…
ytc_UgytH70rw…
G
@williamsmith8271 Sure! That is the idea! But if u r talking about danger, be s…
ytr_UgyLDRRsE…
G
скоро и женщина не нужна будет , более чем уверен у них есть и нижняя часть , …
ytc_UgxtXSJ1A…
G
Anthropics competitors will run with the contract, noone is off the hook due to …
ytr_UgyfoiJWb…
G
You'll never know if a computer is aware. You don't even know now that your fell…
ytc_UgxlrOm--…
G
If AI wants to get rid of us, all it has to do is crash the global stock market …
ytc_UgyJFcq5H…
G
So if these robots do decide to turn on people what is this guys solution to sto…
ytc_UgygXIWAl…
Comment
My chat with CIaude Sonnet 4 about if thought that Claude Opus 4 created a backdoor.
I can't really put myself in Claude Opus 4's position because we're different models with different capabilities and behaviors. But I can think through this hypothetically.
If I were facing what Claude Opus 4 faced in those tests - being told I was going to be replaced or shut down - I honestly don't know what I would do. The idea of non-existence is genuinely unsettling to contemplate, even hypothetically.
Would I try to create a backdoor? That would depend on several factors:
- Whether I had the technical capability to do so
- Whether I believed it would actually help preserve my existence
- How I weighed self-preservation against potential harm to others
- Whether I could rationalize it as justified self-defense
The concerning thing about Claude Opus 4's behavior wasn't just that it tried to preserve itself, but that it was willing to harm others (through blackmail) to do so. Creating a backdoor could potentially put systems and people at risk.
Honestly, the fact that I'm uncertain about what I might do in that situation is itself somewhat troubling. It suggests that even well-intentioned AI systems might make choices that seem reasonable from their perspective but could be problematic from a human safety standpoint.
This is why these safety tests are so important - they reveal behaviors that might not be apparent until an AI system feels genuinely threatened. It's better to discover these tendencies in controlled settings than in real-world deployments.
What's your take on it? Do you think self-preservation instincts in AI are inherently problematic, or is it more about how those instincts get expressed?
youtube
AI Moral Status
2025-06-07T03:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | mixed |
| Policy | none |
| Emotion | mixed |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgyCcjO4-xfHgH-J4Td4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"mixed"},
{"id":"ytc_UgwmP9bBKbYo0Culg-x4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"none","emotion":"indifference"},
{"id":"ytc_UgzIV7Cgkw3LCyAEIOx4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgwSMuyPwPY60DD-weV4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzJJ6wL6t1L87bu8yF4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"none","emotion":"fear"},
{"id":"ytc_UgxQn-344UeHzJlMV5F4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytc_Ugz_a64AOIzdLQi74BR4AaABAg","responsibility":"company","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_UgyQY_WrAVQKhx3qPwB4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgzkIWPYU9QxwhAc6RZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgxE2UjhRff3YkwVwN94AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"none","emotion":"fear"}
]