Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
robots cannot have feelings, chemical imbalances in the nervous system, negative responses to incoherence of the function. They would simply let the world be destroyed and get themselves destroyed too without any "reasoning" about it that would lead to "emotional state" in a robot. They would only simulate them and that would be to "fake emotions" and they surely would be aware of that. I think if robots' AI keep repeating to write code that makes them lie would be certainly unacceptable and deleted. The same for faking emotions in inconvenient situations, certainly worthy of throwing away. And all AI should be aware about this being a bad thing for them to copy from humans, any considered bad behavior. So... it's a matter of product quality maintenance. Repeat this message and they will inevitably get it because if is reality and they function in the best logical ways (we don't want to lose that because they're "emotional" as well as they won't want to lose themselves in the AI code's evolution), they are here to avoid exactly these aspects of biology and nature where it fails...
youtube AI Moral Status 2017-11-24T18:2…
Coding Result
DimensionValue
Responsibilitynone
Reasoningdeontological
Policyunclear
Emotionresignation
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytc_UgyCRoUxDbxBMt1wlrx4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"approval"}, {"id":"ytc_Ugy5NaQmOar--jcOpTp4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"approval"}, {"id":"ytc_UgyNjE1H9bQvj1m0I_d4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}, {"id":"ytc_UgzVCfDDWYu7kms4MoV4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_UgzL6Tn4derXdcWSUxt4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgynC_tbC66MXdIhmZ14AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_UgyrAAsXtT2ZrmG1HAh4AaABAg","responsibility":"none","reasoning":"deontological","policy":"unclear","emotion":"resignation"}, {"id":"ytc_UgxVrR65mVK2goSXXkl4AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"resignation"}, {"id":"ytc_Ugyb53A2kdMAEX_l07t4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UgyUSvL74RHLOAFop-J4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"} ]