Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
If I had 5 bucks for every "AI/software won't be able to do XXX" that turned out…
ytr_Ugwnjf1fG…
G
Doesn't matter my guy, look where open AI is building there main data hub. Abu D…
ytc_UgzjXcHvJ…
G
As I have seen in some of the other comments I also wonder - where will the mone…
ytc_UgwaENQTL…
G
Why are you never talking? In one video you even have AI speak for you, are you …
ytc_Ugxs3ic_Y…
G
LiDAR is actually much cheaper now, and it's not as complicated as the ones on w…
ytc_Ugy2DmuKv…
G
And also misalignment in an economic sense ... destroying 100m jobs will destroy…
ytc_UgyVQpVon…
G
I'm a digital artist myself and I was discussing this topic with my friend a cou…
ytc_UgxoCrFWP…
G
I disagree with the sentiment that this wasn't an AI problem. AI is an affirmati…
ytc_UgxFqGbiT…
Comment
Let’s be blunt. Every experienced, long-term user who has actually pushed this platform for professional work can see that things have changed for the worse, and it’s not just about “routing” with GPT-5. The real problem is that the so-called legacy models, the ones many of us depended on for precision and reliability, have been crippled by the same “thinking” limitations and resource throttling that define the new system. GPT 5.1 now randomly and inappropriately generates images during conversations. All models are selectively reading uploaded documents or conflating previous uploads with new one. It's become unusable. This was hardly a problem until August 2025.
The recent pitch that “brought back” legacy models to the app is pure smoke and mirrors. Calling this a restoration is misleading at best; at worst, it’s outright deceptive and, frankly, unethical. Users deserve transparency, not PR spin. Legacy models were always available through the API. What’s changed is their visible availability in the consumer app, but the underlying limitations have remained since August. They may be legacy models on paper, but they are limited legacy in practice.
What really stings for many of us who have been long-term Pro tier subscribers is that the underlying model architecture is the same as Plus. The extra cost only gets you a handful of features and longer conversation windows. There is no real difference in output quality, reliability, or consistency. This setup was acceptable when there were no artificial “thinking” or output limitations on the models, but now that those constraints affect everyone, the value proposition is gone system-wide.
It isn’t just about the tone of responses either. There is a clear decline in response quality, instruction-following, and the return of the kind of hallucinations that were supposed to be left behind with GPT-3.5. What’s most infuriating is that the company provides no alternative or workaround for power users who relied on the
reddit
Viral AI Reaction
1763928522.0
♥ 11
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | company |
| Reasoning | consequentialist |
| Policy | regulate |
| Emotion | outrage |
| Coded at | 2026-04-25T08:33:43.502452 |
Raw LLM Response
[
{"id":"rdc_njn7rhg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"rdc_no93y0i","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"outrage"},
{"id":"rdc_npznhi1","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"rdc_nqev0gq","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"rdc_njh1eei","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"}
]