Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
I can’t wait for Ai to get rid of all these arrogant podcasters and political an…
ytc_UgwexKqGL…
G
it's not just the facial recognition software that is at fault here.. it's at be…
ytc_UgyGbZMoL…
G
the part about guessing AI vs human was spot on, tried Olovka ai for similar tas…
ytc_UgwhR4cco…
G
I never understood the desire for AI to harm anything. What would give AI a desi…
ytc_UgyufjiCX…
G
I just wonder how can it be possible not to know what algorithms are, it was a c…
ytc_UgxD0KNBy…
G
The difference is there’s not a single job that automation can’t do. Including …
ytr_UgzMOKcii…
G
Let me ask a more ominous question. How would we know if AI has not already capt…
ytc_UgyVeilOR…
G
Heard the same thing except there was a github repo where a facial recognition s…
rdc_fyiwshq
Comment
I believe that issues like hallucinations, errors, trouble with abstract thinking, and unethical behavior may be resolved by continuing to model AI systems on human brain structure. Brains are not just one big blob of interconnected neurons. They have “compartnents” that deal separately (but interactively) with vision and other senses, reasoning, movement, spatial perception, emotion and memory. These structure can be self-correcting, for example in ethical decision making, in ways that simple LLMs cannot achieve
I also believe that neurologists and neuropsychiatrists will start to use such AI systems to model conditions like depression, PTSD, bipolar disorders, schizophrenia, etc as a way to help find more effective treatments
youtube
2025-11-27T11:2…
♥ 1
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | mixed |
| Policy | none |
| Emotion | approval |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgycAHJI6QF5fdAG-lB4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxblKZaLzAT3mu6TIR4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_Ugzc0eBBvKwNR0D0OyR4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"},
{"id":"ytc_UgzdrVJ53CPxqnfxNPh4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_Ugzs2q2y09RHe1gU1Td4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"none","emotion":"outrage"},
{"id":"ytc_UgwqAMRbRMphohcakdd4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgzhAR1gvu95cACD9_54AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwhQ2fUeLaSnGl4sIJ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgwtE3N6RYpad645Cp14AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgxINFB3kenh1VjuTql4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"}
]