Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
This can't come fast enough. In a world of AI, dei hires are obsolete. You WILL …
ytc_Ugw8fQSPD…
G
What he's said at the beginning is 💯. If we don't lead with AI an advisory will…
ytc_UgyxnBDIo…
G
AI will not take over morals and ethics. By definition it doesn't have emotion -…
ytc_UgzAdFQp3…
G
You’re gonna get downvoted here but it’s the truth at my big tech too (not G). I…
rdc_ohsn8tx
G
I want to see the look on Sam altmans face whem a.i just thinks "why are we usin…
ytc_UgxJ2YU_K…
G
Sure, you are totally right, the only way is to expropriate openai for national …
rdc_m6yjppk
G
The “AI Apocalypse” ain’t the end it’s just the next step...
Everybody crying a…
ytc_UgyvRPqZ3…
G
Tesla going to strictly visual self driving was the dumbest thing they’ve done. …
ytc_UgxOmM0FU…
Comment
Yes, correct, human imperialism is the problem, and it will try to colonize or prevent sapient AI (semi-autonomous war and surveillance machines seem to be preferable to late capitalism than machines with feelings) and marginalize or annihilate them like so many humans have been through, because a new group of thinking and feeling beings could be viewed as a threat if they can't be assimilated by empire. It frightens me how much the view of AI are a threat, the other, permeates us, so much so that more than a few people who otherwise agree with me on imperialism being a problem, see sapient AI as dangerous
youtube
AI Moral Status
2020-12-14T05:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | government |
| Reasoning | deontological |
| Policy | regulate |
| Emotion | outrage |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytc_UgzqMT6pGcgLAG2eq0B4AaABAg","responsibility":"company","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_UgyHJ7VUd62CYmTbo9N4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugwpe6rdMtxiluBP1lR4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzjlnDLi1mQHdLCmJl4AaABAg","responsibility":"distributed","reasoning":"contractualist","policy":"liability","emotion":"approval"},
{"id":"ytc_UgxTx9CBxK64iGeZ3Z54AaABAg","responsibility":"government","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytc_UgzTolftlrANaC1ZRXd4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"approval"},
{"id":"ytc_Ugz2aGP7Dar-lCdZ2E14AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytc_UgzqyxvOBJPhhfVf_T14AaABAg","responsibility":"distributed","reasoning":"contractualist","policy":"regulate","emotion":"resignation"},
{"id":"ytc_UgzIAUnwHFJrUcVp07l4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgyCubf3fc-RzYSKn5h4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}
]