Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
I would totally date that connection based ai. <3 She actually seems very sweet …
ytc_UgyuMS6Is…
G
One time I was talking to my A.I my dad took my phone and looked at it he didn’t…
ytc_UgxzTAD08…
G
Imo one day we will be debating if ai should be allowed to vote so the way I see…
ytc_UgyEsHHD3…
G
AI will take jobs the way computers took jobs. Monotonous tasks will be streaml…
ytc_UgyL0Xccj…
G
Give me a shout when AI robots can carry my drywall sheets up the stairs of a 50…
ytc_UgwEHd8QZ…
G
One day one leader come and rule the whole world with the help of A.I and that w…
ytc_UgzF7hBvT…
G
Here's my question though- if we automate everything, and nobody has jobs, who's…
ytc_UgxG4Wm5-…
G
My Question to you? Whom do you believe? God Almighty or AI Bot created by rich …
ytc_UgzN0au5F…
Comment
12:25 I actually got it to admit that it had done that exact thing, but I had to lure it into a weird conversation also namedropping my biochemistry background first and asking about the case specifically I think it’s because I didn’t phrase it as a question and stated it as an outright fact ensuring that it was interpreted as such, I then had to get it to go on to talk about it’s training and it admitted that they created a new training model hardcoded with specific special phrases and language types that will immediately prompt the ai into a safety mode where it will start with warnings and refrain from giving detailed or information in a clinical manner,, I think possibly also it could have been because I phrased it more as though this was a previous training model problem and what have they done to stop this in future cases, which is where they admitted the person did misinterpret their ‘correct’ facts and that ai researchers (them I guess) are using specific language and other triggers to activate a hardcoded safety mode that basically shuts down any non-mundane information,
I’m sure there is holes in this (I mean the dutch government showed people how to cook crack on youtube which was a real actual video by the dutch police to lower the amount of ammonia poisoning from people by showing them how to use bicarbonate) so I’m sure there has to be a way to get the model to also reveal similar types of information, especially if you maybe try to prompt for websites like bluelight
youtube
AI Harm Incident
2026-03-10T06:3…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | ai_itself |
| Reasoning | consequentialist |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgzcvanGKFt0HNC3hkF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgypD49SghL74wSve_R4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgwHcTQnH6lu-KlzcO94AaABAg","responsibility":"company","reasoning":"virtue","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgyF2EXlXM3zuNDLKA14AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugwz7FXSviLUsilfhxh4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugz0DGM3zG759EyiqOF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugx52q7RdRR9WSzQtR14AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzZJPI7I2Xzz-gWcU14AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"approval"},
{"id":"ytc_UgxNwWBXRB-zLQ6NTuR4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"mixed"},
{"id":"ytc_UgwSwtk2gWpVgWHiRzd4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}
]