Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
@jhonny8872 Thank you for commenting! A eso yo lo llamo una ost-IA, and I must s…
ytr_Ugxoe-gUx…
G
Fortunately, there is nothing you can do about it, the US would do anything to r…
ytc_Ugy-izwoS…
G
So you're fine with Walmart using this sort of facial recognition for years now?…
rdc_iyyzz00
G
The model in our brain updates continuously. The AI model is static. Wouldn't yo…
ytc_UgzLelvZb…
G
This was a good intro for people who know nothing about AI companionship, but I …
ytc_UgzCiXFTw…
G
@rajputmehvish bhai apna kaam kar lo thik hai kyuki mere bua ke ladke ka dost bh…
ytr_Ugwlkwqew…
G
Будущее видимо за женщинами киборгами, скоро они начнут составлять конкуренцию ж…
ytc_Ugx7hB2WU…
G
The fact that AI does backend services easier is proof that humans have miscalcu…
ytc_UgwWazDwd…
Comment
Eliezer is clearly an ultra bright guy, and I share his caution, but I never get the feeling he completely nails the "reason" AI would delete us. It seems the answer is 'we cannot not know what AI will want to do, and we cannot control it once it is a super intelligence, ergo, it's a massive risk to build it." I get that, and I would be inclined to agree that the only way forward is with maximum caution. But he is not presenting and "absolute" here and I think that confuses the issue.
youtube
AI Governance
2025-10-15T23:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | distributed |
| Reasoning | consequentialist |
| Policy | regulate |
| Emotion | fear |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_Ugy1egUSacGBMPQ4BKV4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgwetfDxOVuH1vuc9RB4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwVP425hKUyxaCzdmh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"indifference"},
{"id":"ytc_UgyI5DPCta6duLKzyr54AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwPaYcIVSzYEPNHe6p4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"unclear","emotion":"resignation"},
{"id":"ytc_Ugx0rZRZ9c7FMXTPdsd4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"ytc_UgzdQEWkOF26mJP8NEB4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugw_vKorzBDdq9QHXlJ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgzbaaIIZiBMusgDrkl4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"resignation"},
{"id":"ytc_Ugy4LxW4IjIcZHmjdhR4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}
]