Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Out of curiosity, what is your usage of different models in this approach? I often find that my tactics of using GPT3.5 turbo don’t *really* generate the right level of thought in defining approaches to complex plans, but the cap on using the GPT4 model makes heavily using it impossible. Are there any ways to optimize around this?
reddit AI Moral Status 1682288943.0 ♥ 3
Coding Result
DimensionValue
Responsibilitynone
Reasoningunclear
Policynone
Emotionindifference
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_jhhxiws","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"}, {"id":"rdc_jhg6dkk","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"rdc_jhi7txb","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"rdc_jhh24a0","responsibility":"ai_itself","reasoning":"deontological","policy":"none","emotion":"fear"}, {"id":"rdc_jhh7kjp","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"} ]