Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
A fatal error humans are making: personification of AI. Ezra, the AI is not "trying to be helpful." it can't try to be helpful. It's not human. Seems to be a pretty fundamental misunderstanding of what your guest is trying to explain to you. Same reason you can't write out a rule for AI the way you seem to think you can.
youtube AI Governance 2025-10-15T23:4…
Coding Result
DimensionValue
Responsibilityai_itself
Reasoningdeontological
Policynone
Emotionindifference
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[ {"id":"ytc_Ugy1egUSacGBMPQ4BKV4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgwetfDxOVuH1vuc9RB4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"indifference"}, {"id":"ytc_UgwVP425hKUyxaCzdmh4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"indifference"}, {"id":"ytc_UgyI5DPCta6duLKzyr54AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgwPaYcIVSzYEPNHe6p4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"unclear","emotion":"resignation"}, {"id":"ytc_Ugx0rZRZ9c7FMXTPdsd4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}, {"id":"ytc_UgzdQEWkOF26mJP8NEB4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugw_vKorzBDdq9QHXlJ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_UgzbaaIIZiBMusgDrkl4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"resignation"}, {"id":"ytc_Ugy4LxW4IjIcZHmjdhR4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"} ]