Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
From Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path? "How to choose the language for describing theories is an important question, and even the question of whether the Bayesian formalism is sufficiently agnostic to the choice of theories remains open." It is solved in theory. See NiNOR Complexity.
youtube AI Responsibility 2025-05-22T03:2…
Coding Result
DimensionValue
Responsibilitynone
Reasoningunclear
Policynone
Emotionindifference
Coded at2026-04-27T06:24:59.937377
Raw LLM Response
[ {"id":"ytc_UgwG6EVp0ebYHSYeEL14AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"ytc_UgzITkFaWgclkXhuiml4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"mixed"}, {"id":"ytc_UgyaffFFNgaInKKH4wF4AaABAg","responsibility":"distributed","reasoning":"contractualist","policy":"regulate","emotion":"approval"}, {"id":"ytc_UgxUECCHVsaRbVF6XiB4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_UgxG18gOOlravQe2SWZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"ytc_UgxrupSM3gL46TWvxxZ4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgzUjlA6D0vt-8YMD694AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"ytc_UgyL-OvW5hOZY-4itrp4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_UgyDj_c0W-_4mAiHN7V4AaABAg","responsibility":"none","reasoning":"unclear","policy":"regulate","emotion":"approval"}, {"id":"ytc_Ugxkjfmbq_aYfYUXmEN4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"resignation"} ]