Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
If AI succeeds and it takes everyone’s jobs the economy will collapse because no…
ytc_UgwLw6Ews…
G
I don’t get what AI safety is? What’s the harm we need safety from? If it is the…
ytc_UgxoaJNGN…
G
I think dead internet theory might actually be real just by looking at these com…
ytc_UgwJdBuSe…
G
So who will be left to buy stuff? AI specialists. Which equals : no one or maybe…
ytc_UgwCwusQX…
G
What is this person smoking? Art isn’t owned by birthright, it’s a *skill* and t…
ytc_UgykOZXpZ…
G
I surely would like more information about those robot. That field is very excit…
ytc_Ugx946i8j…
G
Love the fact that he is not afraid to say "I don't know". I've seen people talk…
ytc_UgxSBX2ZW…
G
He pushed the idea of UBI pacify everyone into thinking AI advancement would be …
ytc_UgwS9eLlg…
Comment
Kinda felt like these robots only said commands that were triggered by keywords said by the stoned out dr. Would they once evolve to comprehend visuals to make their own decisions? If so how you hardwire the android to take humour or a joke. The more advanced you make AI, they will never understand humans because there robots, with no emotion. That being said the only good idea to me that brings robots into play are jobs that are far to dangerous for humans to complete. They don’t need to speak they don’t need to free roam. Just literally do the task they were created for only.
youtube
AI Moral Status
2019-11-11T09:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytc_UgxXg339nGK4Ig2Xu_J4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxowXB8auIeAE2-EnN4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"outrage"},
{"id":"ytc_UgzzAb1sItOr5kvOw1l4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"none","emotion":"mixed"},
{"id":"ytc_UgyoZXWT0yxjbJDrtxN4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugzs4Nr4q4nAtXn-R1h4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"fear"},
{"id":"ytc_UgyDk5HwrVjeMuOwKXV4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugx5glvnB8nukdO9ERh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyQHJQPZrvuxzNDbc14AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugx13mxDUPagO7vWqOh4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgzRFph_isQl6unSkm94AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}
]