Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
We can't catch no freaking break!
Now they evil racist ass want to build and tr…
ytc_UgwgA015j…
G
For anyone saying the US is the same, you are mentally unstable. In the US, if t…
ytc_Ugz6t6pjb…
G
I am shocked to learn they are not using lidar and where's the IR? Why are they …
ytc_UgzG05b91…
G
If you compare to this to even AI animation from last year you'd see it's drasti…
ytr_Ugz4-_IgQ…
G
Typical socialist BS... what gives this pundit or her regulator allies the speci…
ytc_UgxJgm25I…
G
it's the fact that this cop has all the information needed to solve this case, b…
ytc_UgzDkYq1C…
G
Hey @kohnfutner9637, thanks for your comment! If he's risking hurting his hand, …
ytr_Ugy-nv3Eu…
G
Yes it can there are so many ai restaurants in Japan once check it out…
ytr_UgxgoJZZK…
Comment
I'm rather right-leaning and other right wingers (on X for example) heavily support AI """""art""""" and that's one of the biggest things I hate about that part, other is just being no better than the left:
logical fallacies
conspiracy like thinking
extremism
hive-mind like behavior
etc
I hate how certain topics are made extremely political and one-sided (like climate change, immigration and safety, freedom) this makes it much harder to solve problems so I heavily encourage not being a stupid sheep. Just because your "side" supports something doesn't mean you have to.
youtube
Viral AI Reaction
2025-03-30T09:2…
♥ 3
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | unclear |
| Policy | unclear |
| Emotion | unclear |
| Coded at | 2026-04-27T06:26:44.938723 |
Raw LLM Response
[{"id":"ytc_Ugzdv90n7d0hHRCqbRp4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"approval"},
{"id":"ytc_Ugxx31z43Fw1bXw6vHZ4AaABAg","responsibility":"user","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugw7MLbkXyRu7xfOprR4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"approval"},
{"id":"ytc_Ugx5QQ36dFyfuKA5pMh4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgzxTHUcj2rRvGu-1p94AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"unclear"},
{"id":"ytc_UgwOX39yinP3x3ufz9d4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_Ugy-PWLVQUuWbyJOtJB4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgzzVTuYfbLGiwyXF3h4AaABAg","responsibility":"user","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"ytc_UgybRAdIPUkNutxTMs14AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwBWijsn42Ch9mAFPl4AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"none","emotion":"approval"})