Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
First of all human beings don’t need jobs. Human beings just need money. There’s…
ytc_UgxOu2TgN…
G
After I first became aware of AI, I did some reading about it. I feel it is the…
ytc_Ugx6YmYgP…
G
This is not true and is perverted.....if you go on most Chat AI bots its the COM…
ytc_Ugx5rsHOq…
G
@saraghaffari5088No you can't you find the weak studebts who dont know how to u…
ytr_UgxLnkKql…
G
AI is still simulated intelligence and that is why it's dangerous it has no mora…
ytc_UgxRe3N9p…
G
THIS is what the AI should be used for: making people's life easier, not replaci…
ytc_UgzZ7dndq…
G
This is true Talkie AI is an experience lol its a bit too Advanced but im a char…
ytc_UgwYoxJxr…
G
It's not about the quantity though
AI can pump out billions of images every day.…
ytr_UgywFQg6P…
Comment
As a senior software engineer, I acknowledge there is a certain critical satisfaction in watching the hype around AI — and so-called “vibe coders” — collide with reality. In the languages where I have the deepest expertise, namely SQL and Python, no AI model, regardless of the marketing behind it, comes close to the level of rigor, efficiency, and depth I apply to the code I produce.
AI is useful for rapid prototyping, experimentation, and small showcase applications. However, when it comes to real-world systems — consistent scalability, strong security, solid architecture, and complex business logic — the gap between automated generation and mature software engineering remains substantial.
youtube
AI Jobs
2026-02-24T22:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | virtue |
| Policy | industry_self |
| Emotion | approval |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgzBp8_yKkfM33Houd14AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"mixed"},
{"id":"ytc_UgzqwAuc5GNh-AKdFb54AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgzugNKs-aD0BCEvGLF4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugwiecei9TbT4VFKzpp4AaABAg","responsibility":"company","reasoning":"virtue","policy":"unclear","emotion":"outrage"},
{"id":"ytc_Ugwui4htzjEq8X_1c0N4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgxOPKIpQnpT_6y7w1R4AaABAg","responsibility":"unclear","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzE0N4A6ODuaP5OMY94AaABAg","responsibility":"developer","reasoning":"virtue","policy":"industry_self","emotion":"approval"},
{"id":"ytc_Ugyz3O78JlD3n27GggV4AaABAg","responsibility":"user","reasoning":"deontological","policy":"unclear","emotion":"resignation"},
{"id":"ytc_UgxOTL-m9xodRZzm97l4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugx56HbYPX4_8zO_Iuh4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}
]