Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Seems like in this case the AI is just acting like a mega-google or a bunch of research interns. A human is still making the decisions at the end of the day, the AI is just wrangling all the charts and information in one place. Makes sense, if the space is 20,000 injured soldiers, do you really want to read 20,000 charts, or have an AI read them all and give you the low down? It's probably more effective than trusting a bunch of panicked 18 year old staff officer/interns falling apart because they're under mortar fire.. AI would be used to give medical officers an RTS/XCom style view of the situation. Imagine if you could just look at all the soldiers in treatment and see healthbars with sub stats like breathing, heart rate, computer imagery driven analysis like bleeding, missing limbs, etc. Would let you make decisions way faster than walking your ass all around the hospital getting in everyone's way. Imagine if in the future making these decisions would be as simple as playing Mercy or the Medic from TF2, or selecting which soldiers get healed in XCom instead of well, the paper mountain hell it is now.
reddit AI Responsibility 1648727555.0 ♥ 4
Coding Result
DimensionValue
Responsibilityuser
Reasoningutilitarian
Policyindustry_self
Emotionapproval
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_i2sd5ts","responsibility":"none","reasoning":"consequentialist","policy":"regulate","emotion":"approval"}, {"id":"rdc_i2u4e7o","responsibility":"user","reasoning":"consequentialist","policy":"industry_self","emotion":"approval"}, {"id":"rdc_i2s40i9","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"rdc_i2t3ztm","responsibility":"developer","reasoning":"deontological","policy":"liability","emotion":"fear"}, {"id":"rdc_i2s307w","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"} ]