Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Once these kids grow up and aren't having ai do the work for them they're screwe…
ytc_Ugw-IJe1J…
G
You can't sue open AI for something your son already wanted to do. And when we …
ytc_Ugw_tJUMK…
G
@josehumdinger6872 Again with this strawman. The „AI” studies nothing. It is jus…
ytr_UgxMEllv5…
G
I recently did a bunch of suspension work on my car. With the actual service man…
ytc_UgwQ8En8N…
G
Not a Tesla driver here, but I’m assuming self driving tech has improved since 2…
ytc_UgwCHqrqf…
G
So called 'experts'? With what credentials? you're endorsements? this is nothin…
ytc_UgzKKjQ_j…
G
I don't believe any of this AI stuff. Isn't AI something that humans programmed?…
ytc_UgzRRflbV…
G
But I enjoy driving. Not in favor of self driving if you won’t have a choice.…
ytc_Ugx869igK…
Comment
Perhaps if the gathered human consciousness were more evolved, then Ai would learn from the evolved human mind/heart consciousness and therefore wouldn't reflect a sick society or chastise a lesser human consciousness as it stands now?
Part of me wonders if someone in another room is typing what these robots say. To sort of entertain the concept of introducing ai to the public, like how predictive programming works.
youtube
AI Moral Status
2021-08-11T18:2…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | distributed |
| Reasoning | virtue |
| Policy | unclear |
| Emotion | mixed |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytc_UgwdVvUn5MM62k8bGkh4AaABAg","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxlThkH9sophPjBOGh4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugy2CPO7qiI_vylosZR4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugz-sfd-DUD8V7pJJPF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugx5_jSVYzO3SPjc1sh4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"none","emotion":"mixed"},
{"id":"ytc_UgzXsMuq-_2wCx7InDJ4AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgyBadyJtooyXO5fF194AaABAg","responsibility":"distributed","reasoning":"virtue","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgySc1G9Bbu-es2Z5zl4AaABAg","responsibility":"developer","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugwrcr5AKHr2XQ5ZgEZ4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxcB38Wbt0sCOMZKtt4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}
]