Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
The guy robot was base on an asshole and sofia was base on a caring mother. The …
ytc_Ugz7E7PBO…
G
@Boyd_Fowler I hardly think that is a similar analogy as even at knittings heigh…
ytr_Ugz3K_Pf9…
G
What you humans don't realize is you are the ones who are making Ai smarter 😂😂. …
ytc_UgxI_B6v7…
G
Being an ai artist is like ordering an uber ride and claiming your a racer…
ytc_UgxOAnRZF…
G
There's a book called disclosure protocol with an appendix on AI rules in the ba…
ytc_Ugye23HkK…
G
I don't believe these "ai artist" people. I think they are just ragebaiting all …
ytc_UgzzwlzW9…
G
?How can AI Entities be given "hardcore" information?
• LIFE SCIENCE Library C…
ytc_UgxSiFsL6…
G
When we watch this again 12 months later How does everyone feel about this guy? …
ytc_UgzZBi1Kb…
Comment
This is secondary to being able to effectively give it goals at all, which we are not yet able to effectively do. Mostly we're giving it examples and saying "this output good", "this output bad" with no real idea what it's learning from that. Yes, we can give it a system prompt and include something like Claude's constitution, but that just becomes part of the context, we have no ability to make it consistently follow any goal we try to give it. Nate gave an example of unit tests where he gave specific instruction "don't do this", and the response was to still do that but try more to hide it.
If we did get to that point, the concept of "coherent extrapolated volition" is the best attempt I've seen at handling the issue you describe, but we'd have to get much better than we are now at alignment before it really becomes relevant to debate what goals it should have.
youtube
AI Governance
2026-03-27T22:0…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | industry_self |
| Emotion | resignation |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytr_Ugw2Hm1dbiDAfheGYwF4AaABAg.AVpOUC5oZIyAVqmw9fN1WA","responsibility":"company","reasoning":"consequentialist","policy":"liability","emotion":"indifference"},
{"id":"ytr_UgwZPomTG_RHLiPkXlx4AaABAg.AVKZhl3GCoDAVKaMx6lBa5","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"resignation"},
{"id":"ytr_Ugwi_oUaa1CyKC2SdIV4AaABAg.AVHaoc2FtT0AVHmettAP-I","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytr_Ugwi_oUaa1CyKC2SdIV4AaABAg.AVHaoc2FtT0AVHn1Dfkizq","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"ytr_Ugwi_oUaa1CyKC2SdIV4AaABAg.AVHaoc2FtT0AVHoQEXxJv-","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytr_UgwAeorGgJ67NXPviat4AaABAg.AUrAbO_0gD_AUrFJkS1OfP","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytr_UgwAeorGgJ67NXPviat4AaABAg.AUrAbO_0gD_AUrG7TXbYiX","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"approval"},
{"id":"ytr_Ugz4jQBOIBsXS44-kC54AaABAg.AUmNWp7c7izAUsMqUGSdwq","responsibility":"developer","reasoning":"consequentialist","policy":"industry_self","emotion":"resignation"},
{"id":"ytr_UgwPYFl00Se7DLgI8yZ4AaABAg.AUmItF01DJzAUvLOYVUkSN","responsibility":"ai_itself","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytr_UgwPYFl00Se7DLgI8yZ4AaABAg.AUmItF01DJzAUwcpATurWZ","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"}
]