Raw LLM Responses

Inspect the exact model output for any coded comment.

Comment
8:28 I've heard about these test and, i've fond them to be a bit leading. They'll tell it to "do this thing and this thing is the only thing you care about" or something like that. Basically prompting it into acting like a goal maximizer which would obviously be misaligned. Similar thing with AI figuring out it's being actively being trained and trying to act more aligned. A lot of times they just tell the AI it is being trained which is not necessarelly a good way of knowing how it would work in reallity. How would an AI actually figure out it's being trained without being explicitely told it is?
youtube AI Governance 2025-08-26T17:3… ♥ 1
Coding Result
DimensionValue
Responsibilitydeveloper
Reasoningconsequentialist
Policyliability
Emotionresignation
Coded at2026-04-26T19:39:26.816318
Raw LLM Response
[ {"id":"ytc_UgzcCrHZJJxknEBhj6t4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"resignation"}, {"id":"ytc_UgwvyYKaV--lqBPpVsZ4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"}, {"id":"ytc_Ugwju9H471QELXm_kph4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"}, {"id":"ytc_UgyTZPkpxjQFi7ST6h94AaABAg","responsibility":"ai_itself","reasoning":"virtue","policy":"none","emotion":"approval"}, {"id":"ytc_UgwN1M2YFT83yGDmXWd4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"} ]