Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
AI can read so should you stop to learn reading how to read?? Should AI read boo…
ytc_Ugz5FtFe-…
G
That's ridiculous, you just forced him to affirm you. You can use manipulations …
ytc_Ugwigj-yV…
G
honestly this guy is a dumbass, saying that drawing doesn't make you an artist l…
ytc_UgwsYpUI6…
G
it's not about "wanting to end the conversation" it's because for the AI it ga…
ytr_UgyUWwn45…
G
Well, smart people, please give your best to develop a new vaccine asap - we are…
rdc_hm7geqf
G
Human inteligence created the technology to fly, and them used this technology t…
ytc_UgzXFO_xz…
G
It's interesting how, once released, an AI becomes an object of learning for oth…
ytc_UgzU9v2GC…
G
@mariselachapa646yes I have but like I said we are far out from A.I replacing mo…
ytr_Ugz19ifqx…
Comment
You should also mention the objective bottleneck of LLMs: all of their learning is squeezed through the single objective of next-token prediction. Large language models are trained to predict the next word (or token) given the previous context. No matter how complex the task appears—reasoning, summarizing, coding, or answering philosophical questions—the underlying training objective is always the same: maximize the probability of the next token. This creates a structural bottleneck. Because everything must be optimized through this single objective, the model is not directly trained to “know,” “understand,” or “reason” in the human sense. Instead, it learns statistical patterns that help it continue text in a way that resembles high-quality human responses. Any reasoning ability, uncertainty expression (such as saying “I don’t know”), or structured problem-solving emerges indirectly from this next-token training objective rather than being explicitly optimized for truth or calibrated uncertainty.
youtube
AI Moral Status
2026-03-01T12:5…
♥ 2
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | unclear |
| Policy | unclear |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[{"id":"ytc_Ugz8K7gIffnKEMKSnNB4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgyHhli5R6UqJ0qsfTJ4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgyL797_M71m5hQW-PN4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugyillgr3oYJn_d_FnV4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_Ugz5Juih4UDG8Yij1MN4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgzUqHajhQLOQu10Pr54AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwjLJk5tZcfPpq5q7N4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugy9S4Kpf-J-OMVdrWd4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"},
{"id":"ytc_Ugy9avnzUN7G8NPX67t4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugx8zuQBCFBUGuXyjcJ4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"}]