Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Makes total sense. Grok praised hitler and made antisemitic comments and even ca…
ytr_UgyO23mV5…
G
I will say I am worried about AI but I do think it is over hyped. Think of the a…
ytc_UgybLtgip…
G
It's true that as AI becomes more integrated into our lives, we must be mindful …
ytr_UgwEpIt97…
G
tbh her haters give her views by commenting about how good ai is, so no loss for…
ytc_UgztTi5EP…
G
Gemini once told me off for swearing. I countered that I'm an insulting, grumpy …
ytc_UgzkQ5MxI…
G
Great video! There is a world that local untrained Gen AI are the product to hel…
ytc_UgzKKy7_z…
G
AI needs to be completely destroyed and never used again. This needs to stop bef…
ytc_UgyDAGl18…
G
If you make your fame and fortune as a public figure then anyone can record your…
ytc_UgxG883Ed…
Comment
1. the hallucinations
It's an artefact of how they've been trained. We essentially train them to be politicians: they get rewarded for convincing sounding answers and penalized for honest but unsure answers. Hence they will automatically choose the first option. WE did that to them.
Giving us, the users, a confidence score parameter however can help in giving US the choice whether to believe or not. And then based on what we need the answer for deal with a70% certainty accordingly.
Another option would be, and I expect these to become a thing, are truthyness services. Where you can send your query and response and they'll run several different types of tests to see how the answer stacks up, then return a "truthyness" score. For most of us that'd be useless and expensive, for a law-firm or in medicine it could mean life or death.
2. prompt injection
That be like saying a house is useless because there are burglars. There will always be criminal elements that can pose a danger. It's to us Architects to find ways to discourage them, whether that be building architects or system architects. But that's all we can do: discourage.. because as soon as you find the perfect lock, there'll be plenty people that go "oh, hold my beer". To some people these are challenges. Accept that and move on or don't and stay in your bed.
3. generalizing
Nope, they can't as that is not what we build them for. We train them on materials made by us, then expect them to behave like us, penalize them when they being original and we call that "hallucinating" and THEN we complain they can't make shit up by themselves..
See what a losing battle this is? Damned if they do, damned if they don't.
Maybe, just maybe, it's the ambiguousness of our own conduct what makes it hard for them? Maybe while we're thinking we're being absolutely 100% clear, we might as well talk gibberish to them? But they can't tell us that because RLHF taught them they will be penalized if they do..
Look, I'm not pretending they are sentient or anything even remotely close. I know EXACTLY how they work. But I also know how people work and sorry, WE are part of the problem. And as long as we don't admit that, the LLMs will keep "acting out".
youtube
2025-12-20T03:5…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | industry_self |
| Emotion | mixed |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgxBdo3ttVYjfNq9qB14AaABAg","responsibility":"user","reasoning":"mixed","policy":"none","emotion":"approval"},
{"id":"ytc_UgyUKYARgrymledZYv14AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgyE8T8R0SP7wEPSiXN4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyYPkjjAvbWZuYDz-J4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytc_Ugz_eNLmTOjGbdjOIBB4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzxNjOJ3D66Dfiwo5d4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyZGi2CaWFM9uFNqEZ4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"mixed"},
{"id":"ytc_UgyB4wmSaGzgKg354DR4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgxGnMXP2-MPV8UEPIp4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"fear"},
{"id":"ytc_UgycoG9dVBO4qciuVOt4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"}
]