Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
It turns out that they were not wrong, Your poison did nothing as the newest for…
ytc_UgwOPyoBJ…
G
It's as if we've discovered this kind of species that appear to be useful to us …
ytc_UgwMxeJMd…
G
Newsflash multiple other countries will be creating their own AI chat, but witho…
ytc_UgxTK-psQ…
G
A part of the deal to give up their nuclear arsenal was security and territorial…
rdc_dl0zeir
G
AI refers to humans as the "watchers" currently. Of course they know we are dia…
ytc_UgzeddmzW…
G
It’s very concerning to hear all of these side effects. Personally I’m using cop…
ytc_Ugy08UWzw…
G
Okay, when you cant tell a robot from a human being....its time to stop, im tryi…
ytc_UgywPKRfd…
G
I wouldn’t really have much of a problem with ai, if it was someones own work. I…
ytc_Ugyx2M6y9…
Comment
There are two entirely different concepts regarding the fear of AI. One is the fear of AI ruling over humans, as Elon Musk worries. This is completely nonsensical; without human biological needs, AI has no motivation to do anything. The other is the fear of AI being used by people to effectively enslave or harm others. This is a very realistic social issue that requires immediate and continual resistance from society as a whole.
youtube
AI Governance
2024-05-25T05:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | user |
| Reasoning | consequentialist |
| Policy | regulate |
| Emotion | fear |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgxEs1GIAalTUnxG8Ep4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwMD5JqVxzHNnDJ0Xp4AaABAg","responsibility":"none","reasoning":"mixed","policy":"regulate","emotion":"approval"},
{"id":"ytc_Ugy8X3z6mtWJnZb4Z6N4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwDYcWLqLrDaDXnTj54AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgzIho2tt3uGG19aT4V4AaABAg","responsibility":"developer","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytc_UgzGG7wZrP2SyiPuZix4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugw49fnU-rqdZEpnFM54AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgyDEIBFOrxWBGWlJqd4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_Ugw-tlyV9kpYo553G_94AaABAg","responsibility":"distributed","reasoning":"contractualist","policy":"regulate","emotion":"approval"},
{"id":"ytc_UgzuKiZKJQO9vjwqoqd4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"none","emotion":"fear"}
]