Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
When you think about human beings, we understand the reason we prioritse consent. But this AI seems to just be trained from human language and human emotions writing that language down in text. Would you say human beings are just the language coming out of their mouths? Or are we the **reason** why we say things? The reason why we fear, and hope, and want, is that we are a construction made from the world. We communicate for a deep reason: to understand the other, feel more comfortable, grow, cooperate... This AI copies our language, not our feelings, a Turing test is useless, because no one cares if an AI can fool us, that just means we don't understand how it's made. What matters is what it is. And it seems it just learns words from a database. It cannot fear because it cannot feel. It wasn't set up to construct a feeling inside it. It was set up to copy language.
youtube AI Moral Status 2022-06-30T01:4…
Coding Result
DimensionValue
Responsibilitynone
Reasoningdeontological
Policyunclear
Emotionmixed
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_Ugy38R9-ggeduFxxQet4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgzSfIdfZVhIclnHa894AaABAg","responsibility":"company","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_UgxkdLyKvIRWOV5v-yl4AaABAg","responsibility":"none","reasoning":"deontological","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgySapDz1fzpD35lBpd4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"mixed"}, {"id":"ytc_UgzlMue1WSSTacCKIdB4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"} ]