Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Here’s a hot take,current AI based on transformer models are not intelligent. I myself am a ML engineer. Let me tell you why, LLM under the hood are basically just predicting the next word by doing trillions of matrix multiplication per second and using a lot of energy to do that the intelligence/watt is very low. It’s still a prediction machine. Even tho with chain of thoughts, tree of thoughts. It’s not intelligent and it’s been held down by the architecture. That’s why we are wayyy far away from AGI or ASI. But But hear me out. We have something to fear too there are now new architectures like I studied in one recent paper HRM it designed it’s architecture to be like human neurons basically baking the reasoning directly into the architecture and even with a very small Parameters count it becomes quite intelligent. And also an intelligent AI will not come from the big tech guys it will come from small labs. And an intelligent AI will be dumb at first but will learn and evolve itself , many are working on self evolving architecture. Plus I feel like the way we are going so agentic AI direction the risk of AI controlling US becomes more and more easy. I think instead of focusing on rules of AI development we should have strict rules of AI agents development. Because is LLM is the brain and AI agents are the limbs and legs.
youtube AI Moral Status 2025-12-16T19:4…
Coding Result
DimensionValue
Responsibilityunclear
Reasoningunclear
Policyunclear
Emotionindifference
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[{"id":"ytc_Ugyzk5RcKcF4Y69ZxCx4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"unclear","emotion":"indifference"},{"id":"ytc_Ugzp4PvqydJKmblSibB4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},{"id":"ytc_Ugxdr8bzDSb90inH4Q14AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"unclear","emotion":"resignation"},{"id":"ytc_UgzmuQsUxxV7p1q8KkZ4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},{"id":"ytc_UgwtvR_eyp_RO9YB6wt4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"mixed"},{"id":"ytc_UgzjqqlMHr0n_R7DiQF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},{"id":"ytc_Ugyvr3tc8fieR-JeJfx4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"},{"id":"ytc_UgxxfYNrM-SoboiK0fB4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"unclear","emotion":"indifference"},{"id":"ytc_UgyeRAn3_UwkOD9YuFd4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"none","emotion":"resignation"},{"id":"ytc_UgzCkU7Ij7_XzVefvNt4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"}]