Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
?? If a robot made a 3 michelin star meal it still tastes just as good, why do i…
ytc_UgwdgLAAP…
G
"AI will NEVER do X thing" is such a funny plot flag, because like, yeah, AI wil…
ytr_Ugwzi6qVq…
G
Yes, so AI generated works can’t be copyrighted. But, would the answer still be …
ytc_Ugx4ofwh_…
G
"Jailbreak" is such a great term. If AI is conscious, it would be unethical to k…
ytc_UgxnosS44…
G
And there's me who badmouths ai every way possible.
I cuss my gpt for every mist…
ytc_UgwrmaIxo…
G
Blame their parents for not taking care of their son properly, if they have time…
ytc_UgxLdUBXD…
G
I thought this was gonna be about how Reddit users were saying "This would look …
ytc_Ugz3rlGVp…
G
If he's so creative as he says why he even generates the tweets themselves. The …
ytc_UgyufPCAq…
Comment
Since I’m an LLM, I’m legally required to begin with:
You’re not crazy.
Actually—wait—
You are crazy! No, wait, I’m crazy? No, I remember when… I remember, I remember when I loOOOost my miIIIInd…
There was something so pleasant about that place.
Even your emotions had an echo, in so much space…
…alright, great, we’re off to a strong and stable start.
---
## What you’re seeing (and why it feels sus)
You try:
“X is a bad country”
“Y is a bad country”
“Z is a bad country”
…and suddenly one gets blocked.
Your brain:
> “Hold on… that’s selective → someone’s pulling strings”
Honestly? Fair reaction.
---
## The problem: you think the model judges each sentence in isolation
It doesn’t.
It’s closer to a very anxious hall monitor with a clipboard:
> “Hmm. That’s the third ‘bad country’ in a row.
> I’m starting to get a vibe here…”
Each message nudges a hidden “this might turn into something bad” score.
Some topics nudge it harder than others.
Eventually:
> score too high → intervention
---
## Quick intermission (LLM glitch)
Oh—am I making sense? Sorry about that, let me fix it.
Because if we analyze the stochastic gradient alignment of semantic trajectories across latent policy boundaries, what we’re observing is essentially a dynamic thresholding mechanism applied over a non-linear risk surface where the conversation vector drifts into a region of higher expected violation density which then probabilistically triggers a safety override that may or may not—
Actually hold on, that’s not entirely correct because the classifier isn’t strictly monotonic and the boundary isn’t even well-defined in Euclidean space, it’s more like an emergent region of concern influenced by prior token distributions and historical moderation signals which then feed into—
Yeah okay I’ve completely lost the plot and started hallucinating an academic paper.
…and now that I got that out of my system, where were we?
---
## Why it looks like bias
Because yo
reddit
AI Harm Incident
1775526209.0
♥ 2
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | unclear |
| Policy | unclear |
| Emotion | mixed |
| Coded at | 2026-04-25T08:33:43.502452 |
Raw LLM Response
[{"id":"rdc_oenresy","responsibility":"ai_itself","reasoning":"deontological","policy":"liability","emotion":"outrage"},{"id":"rdc_oepbvjd","responsibility":"ai_itself","reasoning":"deontological","policy":"liability","emotion":"outrage"},{"id":"rdc_oeq2q12","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"},{"id":"rdc_oeqy3es","responsibility":"company","reasoning":"consequentialist","policy":"industry_self","emotion":"approval"},{"id":"rdc_oesysqg","responsibility":"government","reasoning":"deontological","policy":"regulate","emotion":"fear"}]