Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
I dont buy the "this can kill humanity, this is why I talk about this" and next …
ytc_UgxYPDFeG…
G
Noonr is interested in what AI is doing and what it will do. Stop posting nonsen…
ytc_Ugx7tUqdv…
G
The biggest problem for me, as an artist (especially an anime-style artist) is t…
ytc_UgxbaDHv3…
G
Talking to ai because you have social anxiety and avoid discomfort is a band-aid…
rdc_n7wuood
G
I suspect AI hasn't been activated to be the big collective beast system, yet , …
ytc_UgwubdGpU…
G
ABSOLUTELY NOT. Do not even give anything in this discussion a grain of salt. …
ytc_UgwTo_REA…
G
elon musk: lets halt ai development for six months!
also elon musk: presenting t…
ytc_Ugy05o6c4…
G
AI is used to covertly torture dissenters around the world. It's an important mi…
ytr_Ugz99nNZI…
Comment
I think the most dangerous thing about AI is the lack of responsibility. When humans generate some text or advice or whatever, they have to own it (excluding anonymous internet trolls) but there’s no responsibility in AI’s text. It can recommend you to do crazy stuff but it can’t take the responsibility either. The person who follows the stupid advice can’t be blamed because they don’t know it (that’s why they’re asking AI). When there is a lack of responsibility it’s terrifying, look at the anonymous internet trolls!
youtube
AI Responsibility
2025-05-25T00:3…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | deontological |
| Policy | liability |
| Emotion | outrage |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgxDV1MHsnN_XyPLMc14AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugw1RoNfPUSLt_TDTcx4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwhVJKxQk9By4kMZHp4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytc_Ugw2yEVi-_IWmJbfkol4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UgyVnpd5pHeiiO8_KQJ4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_Ugx8PysEJ1p75W01t-J4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugw2jE0gY7uWJbKw-9Z4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgyUca_DdJ8NSjUhREl4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgyErOyGQnrNYU_t-zt4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgxZlKi1BU6FraiVs754AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"sadness"}
]