Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
"AI should be used for things like [...] content moderation"
Gotta disagree her…
ytc_UgybgzJQi…
G
Honestly all of these are BS. As a human I would come up with other options like…
ytc_Ugwnwm3zX…
G
i dont think that its possibly to make a robot that has feelings. we can program…
ytc_UgyTocf-K…
G
It's written in the Bible centuries ago. AI is a dangerous tool & the final sta…
ytc_Ugyd9aOUq…
G
The sad part is Ai could of already taken over and we wouldn’t even know…
ytc_Ugz90pgTw…
G
I've first read about this in the book "Weapons of Math Destruction". A major pr…
ytc_UgwsT0OPQ…
G
When we start seeing robots like this it's the end of the world before you know …
ytc_UgyifYuew…
G
It is all a bit strange, but just now I was feeding Gemini the script for Casabl…
rdc_mlhzpw9
Comment
Pure capitalism doesn't work in the context of super AI. The premise that makes it kind of functional right now is that anyone can increase their wealth by providing some service or commodity to their community, which is the incentive that makes it useful. If 99% of the work is going to be done by AI, the competition to provide the other 1% is going to be so harsh that human labor value is going to plummet. Long term this will make the "owners" of AI owners of every other asset. In order to make AI useful for everyone, ownership of this artificial work labour must be shared, and some additional secondary benefits need to be implemented so someone is willing to do the missing 1%.
youtube
AI Governance
2025-09-07T10:3…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | distributed |
| Reasoning | consequentialist |
| Policy | liability |
| Emotion | fear |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytr_Ugyqt942k04PhY4Yh394AaABAg.AMjejH0wvmRAMjfvyLezeP","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"ytr_Ugyv7KN1SuQO-GzWb0F4AaABAg.AMj_mqUKvAKAMjcFLpQr2x","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytr_UgxH9BG6yeeTx8ciq8J4AaABAg.AMjVxf2lAt2AMjlZ1pLiAO","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytr_UgxeqCThViTul8x-Nel4AaABAg.AMjV6VO6IIAAMlZnjXSrMV","responsibility":"distributed","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytr_UgxeqCThViTul8x-Nel4AaABAg.AMjV6VO6IIAAODKKsxnYfn","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"resignation"},
{"id":"ytr_UgzRK_HM_JR54ag7iMR4AaABAg.AMjHYFuPlx-AMjRB1UwpJT","responsibility":"ai_itself","reasoning":"mixed","policy":"industry_self","emotion":"approval"},
{"id":"ytr_UgwbS69WsHo0nKuHvqN4AaABAg.AMj68UvKsQcAMj9lLJAWXi","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytr_UgzmyMVUJxCGf4uiXnh4AaABAg.AMix5qheXr-AMjTVEJXfY5","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytr_UgyOdZBiKFhp3EoyI5x4AaABAg.AMij3JwoIK2AMj0gUP0z9H","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytr_Ugyqu6plXGJxsSR6jIh4AaABAg.AMigYc3CHASAMjGc-T2iAa","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"fear"}
]