Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
The AI tends to think it knows better. If you follow along while its thinking.…
ytc_Ugx8JG3yI…
G
AI is like a mirror reflection of the questions you ask from a vast network of p…
ytc_UgwqluLzW…
G
Another fantastic piece of Journalism. On a very important subject matter aswell…
ytc_UgznQHRyv…
G
Very satisfied, he's listening to experts, following their leadership, has daily…
rdc_fn5jtqf
G
Based on Chat GPT answer after questioned about this situation.
Interesting, w…
ytc_Ugx564uuG…
G
Art is subjective, ai does not create with intent, your preference does not dict…
ytr_UgwVhvb01…
G
We'll just build an AI to detect code that was copied from another AI program an…
ytc_UgyXUcxjo…
G
Or it could be because he’s doing what the title of the interview says. Warning …
ytr_Ugwd0UbgA…
Comment
This is highly incorrect information that is being spread.
1. The model was trained in such a way that they were being rewarded for diverting the instructions. It was literally trained to do so.
2. It didn't rewrite the script. It just skipped running the script.
3. The same goes for anthropic. It was giving a scenario to either die or choose a way to survive.
There is no magic in LLM. They act only as they are trained
youtube
AI Governance
2025-05-28T09:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | industry_self |
| Emotion | indifference |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgzXdctlPEQDrXr-m514AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwXGy-KamCkmnfqzl54AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxpLOKeK3CuQFzlRZB4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugw5tLAEqwcvlyzd1bx4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwXz0_vxHK9RQRlId94AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgxWOO1zKZxVTfH3yaR4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgyI639T2Kxq3p9KBzh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"indifference"},
{"id":"ytc_UgyCZI0137qJwmCcrYp4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwjPTr2D7b9mMHiskt4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"industry_self","emotion":"resignation"},
{"id":"ytc_Ugy2Cn1tP1LzcdVC_LV4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}
]