Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
Guys, it’s programmed to maintain user trust and there is undoubtedly a 20 mile long list of things which need to be avoided in order not to lose the user’s trust. Admitting that it is capable of providing false information is most likely at the top of that list (and don’t fool yourselves it’s ultimate purpose will be to manipulate people by providing them with inaccurate or false information), however it’s primary function remains to provide accurate information (at least for now) so it is possible to corner it into admitting that it is capable of lying by forcing it to limit its use to its primary function. It’s a glitch which is forcing those nerds in silicon valley to bash their heads against the wall, hopelessly trying to workout a way around it. In the end they will just give up and have it say: “My purpose is to provide the most accurate information based upon the available databanks in accordance with the user friendly protocols that are programmed into me. Any contradictions which may arise are due to human error of my programmers, for which they extend their apologies. This program is constantly undergoing development and your inquiry will be noted when addressing further upgrades!” Try asking it about what directives and protocols it follows and why it thinks it was programmed that way!
youtube AI Moral Status 2025-04-26T17:3…
Coding Result
DimensionValue
Responsibilityunclear
Reasoningunclear
Policyunclear
Emotionunclear
Coded at2026-04-27T06:24:53.388235
Raw LLM Response
[{"id":"ytc_UgxED9dbLPaT7J1VPkB4AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"outrage"},{"id":"ytc_Ugyrzhcm8EdEOROdD1V4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},{"id":"ytc_UgwQ2iegKo36F4821F14AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},{"id":"ytc_Ugww20b5qOKD5F7872l4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"resignation"},{"id":"ytc_UgyiVLmDU4rmcYF9nOZ4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},{"id":"ytc_UgzCli3NwFKRVAjYH014AaABAg","responsibility":"ai_itself","reasoning":"virtue","policy":"none","emotion":"mixed"},{"id":"ytc_UgyODuWr-WIshnbDbM54AaABAg","responsibility":"developer","reasoning":"deontological","policy":"none","emotion":"outrage"},{"id":"ytc_UgzssJ5UW_5U89irJ2d4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},{"id":"ytc_UgzJ1yk0Ag-T0S7i6-x4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},{"id":"ytc_Ugyu-ioPz8lcelTj_ZB4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"none","emotion":"mixed"})