Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
It’s a product of over-optimization, while they are trying to optimize the model to not output certain things or optimize it to be better at certain tasks, there can be unintended issue like this that pop up, the hard thing is finding a balance of optimization and performance. John Schulman cofounder of OpenAI just presented about this today at ICML 2023. don’t know when it’ll be up on youtube but definitely look for it in the coming weeks when it comes out if you’re interested. the talk is called “Proxy objectives in reinforcement learning from human feedback”
reddit AI Responsibility 1690513882.0 ♥ 15
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningunclear
Policyindustry_self
Emotionmixed
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_jts18tg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"rdc_jtr821v","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"mixed"}, {"id":"rdc_jtqvu13","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"}, {"id":"rdc_jtrtg4s","responsibility":"user","reasoning":"deontological","policy":"none","emotion":"outrage"}, {"id":"rdc_jtswu4d","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"mixed"} ]