Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
22:00 .... That's a really good point. LLM only see the finished paper. They didn't see the hours and hours of time researching law reviews and reading judge opinions. To LLM and tech bros, it's "output" but the lawyer didn't already know - he found out. We expect LLM to just know, but it also needs to find out and it's not trained how to find out.
youtube AI Moral Status 2025-11-05T19:4…
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningdeontological
Policyunclear
Emotionapproval
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[ {"id":"ytc_UgwONPTSxI16vLASrCx4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgyIUNV7HqoiN0D2SY94AaABAg","responsibility":"company","reasoning":"deontological","policy":"unclear","emotion":"outrage"}, {"id":"ytc_Ugw7zLXdI8VA8NExWy54AaABAg","responsibility":"company","reasoning":"virtue","policy":"ban","emotion":"outrage"}, {"id":"ytc_UgwIqHqzQK3FRQ-Z9kd4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugxa8G6Hj7-Uy1v2m7F4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"approval"}, {"id":"ytc_UgxZeMbcoz8_B8cfC2B4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"mixed"}, {"id":"ytc_UgwlIkV3gUvfeqpbTZt4AaABAg","responsibility":"company","reasoning":"virtue","policy":"unclear","emotion":"outrage"}, {"id":"ytc_UgxPiGyOdYmVGTx4S914AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_Ugyg5llKGtiBwu0Oaj94AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}, {"id":"ytc_UgwIUaDRAUUrBlNLvdt4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"fear"} ]