Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Okay, I’m really starting to have a bone to pick with this podcast.
1. The diff…
ytc_Ugy7BTtYz…
G
I think the timeframes are way off. Any good programmer who interacts with AI wi…
ytc_UgzqvP_89…
G
What made me tune out of the AI talk was the hyper focus on the weird idea of qu…
ytc_UgxiUl3e0…
G
Geoff quit goobling his knobb at 10:30 - no, ChatGPT could absolutely do his job…
ytc_UgyJj1VbT…
G
What if they start bearing babies from humans. And produce humans.
Programmed t…
ytc_UgxhGnOWa…
G
I really don’t think that the Arbor Care companies will be gone in five years.
…
ytc_UgzUrZhUi…
G
Agreed. We still seem to talk about AI as if it's human and intelligent. This is…
ytr_UgyE1VLNZ…
G
AI art is art. I love when artists charging 100 usd for an image cry about ppl u…
ytc_UgyY-FjB8…
Comment
One question I always have and which distinguishes AI (II = Inorganic Intelligence) is in the idea of desire. The difference (speaking towards the Gorilla problem) ithat can possiblky be highlighted is that boith humnans and gorillas as organic beings have this thing called desire. Desire for food, sex, etc. To survive. These desires prompt us to act. From that angle - pointing out his chess example whereas II have beat human beings easily and now also in GO - the thing is here - when will an II challenge a human to ches or GO? In ALL of these cases it was a human's desire that led to the action of them getting beaten. At no point as far as I am aware has an inorganic intelligence had any desire whatsover to llay chess or go. But if you challenge them to a match they will probably beat you. Extending that to this conversation - the bad outcomes are likely not to come from an II's desire to commit harm but more likely to come from human's asking them to do something harmful. Or to be clear - intelligence is great but its desire that is required for that intelligence to have any use in the real world.
youtube
AI Governance
2025-12-08T15:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | mixed |
| Policy | unclear |
| Emotion | mixed |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgxdZ6obicZ679rFsZl4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugx5VZM7vqsOyGrh0YN4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"},
{"id":"ytc_Ugx5tNEuirSug106Ri14AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_Ugyb9DF8UaM5EkaJxRR4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgyMR2qraTs8HKf_nLl4AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"fear"},
{"id":"ytc_Ugzq8l3DB_gE7HBtbXh4AaABAg","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"approval"},
{"id":"ytc_UgyKMEYgyPj66nxs_eJ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgxfU8Ciu6YYPft9vMZ4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgwgOGAna6C4gApUHth4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgzGc1xt39XvvtPXYnl4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}
]