Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
> Way I see it: use it like you would use Google No, use Google like you would use Google. ChatGPT is something very different. ChatGPT is designed to *sound plausible*, which means it will totally make up stuff out of whole cloth. I've encountered this frequently, I'll ask it "how do I do X?" And it will confidently give me code with APIs that don't exist, or in one case it gave me a walkthrough of a game that was basically fanfiction. ChatGPT is very good as an aid to creativity, where making stuff up is actually the goal. For writing little programs and functions where the stuff it says can be immediately validated. For a summary explanation of something when the veracity doesn't actually matter much or can be easily checked against other sources. But as a "knowledge engine", no, it's a bad idea to use it that way. I could see this technology being used in conjunction with a knowledge engine back-end of some kind to let it sound more natural but that's something other than ChatGPT.
reddit AI Governance 1676254390.0 ♥ 146
Coding Result
DimensionValue
Responsibilityai_itself
Reasoningconsequentialist
Policynone
Emotionfear
Coded at2026-04-25T08:33:43.502452
Raw LLM Response
[ {"id":"rdc_j8bijv9","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"rdc_j8cw2wo","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"outrage"}, {"id":"rdc_j8bfhaa","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}, {"id":"rdc_j8bpbdc","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}, {"id":"rdc_j8d5qa8","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"approval"} ]