Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Can't you just output a confident score on each statement outputted from the LLM and based on that determine if it could be trusted? As long as confidence scores are well calibrated it adds the reliability needed for true automation
youtube AI Responsibility 2025-10-05T19:4…
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningconsequentialist
Policynone
Emotionapproval
Coded at2026-04-26T23:09:12.988011
Raw LLM Response
[ {"id":"ytc_UgzVfVMDT-jX8ydE6CR4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"indifference"}, {"id":"ytc_UgxOWYuHd6Rt0x-H3BB4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_Ugz6EAMfz6VgSiUkxet4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"indifference"}, {"id":"ytc_Ugwm767MDJH6KhiyePV4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_UgwYoeFTxfneLzYaPrF4AaABAg","responsibility":"government","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"ytc_Ugyv7bsTIpa9MjnwbTJ4AaABAg","responsibility":"company","reasoning":"deontological","policy":"none","emotion":"mixed"}, {"id":"ytc_UgwZgHxYz7Go8KvfJth4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"}, {"id":"ytc_Ugy0KG_yPEjwlnBrBth4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"}, {"id":"ytc_UgxR3_RpVggpXZs4uOB4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"indifference"}, {"id":"ytc_Ugw2WnEQp69NcxvHX0p4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"approval"} ]