Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
PLEASE, I really think it is time to apply the The Three Laws of Robotics... A robot may not injure a human being or, through inaction, allow a human being to come to harm.A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.
youtube AI Moral Status 2016-04-24T13:2…
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningdeontological
Policyregulate
Emotionfear
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[ {"id":"ytc_Uggz59Wk1uccpHgCoAEC","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"approval"}, {"id":"ytc_UggYjSCa-viEr3gCoAEC","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"ytc_UgiR6u9UeoXzQ3gCoAEC","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"fear"}, {"id":"ytc_Ugj5PcGrXVvamXgCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"unclear","emotion":"outrage"}, {"id":"ytc_UgjbVl7XsOGSwngCoAEC","responsibility":"ai_itself","reasoning":"unclear","policy":"none","emotion":"mixed"}, {"id":"ytc_UgiZPSMsSeFdyHgCoAEC","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"fear"}, {"id":"ytc_UghjZLxbv6tJH3gCoAEC","responsibility":"developer","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_UgixzC7bRoVKfHgCoAEC","responsibility":"developer","reasoning":"virtue","policy":"ban","emotion":"fear"}, {"id":"ytc_UgiD6ik4SvpopXgCoAEC","responsibility":"distributed","reasoning":"consequentialist","policy":"unclear","emotion":"outrage"}, {"id":"ytc_Ugj6ZzVkxUKOcHgCoAEC","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"} ]