Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
So... This guy knew it was ai, has specificaly on his site that these are ai, an…
ytc_Ugx9F_w_e…
G
exactly this. they'll do onlyfans or post bikini and lingerie pics on twitter an…
ytr_Ugwqpwjzw…
G
I guess I will have to come out of retirement just to fix robots 🤖 and maybe get…
ytc_UgzDp-FaP…
G
Ai can do most of the work of a CEO. Companies can save a fortune getting rid of…
ytc_Ugxva3WGu…
G
But all this assumes that only your car has self-driving capabilities. If there …
ytc_UgziBNScD…
G
I believe they can but stunt doubles are usually like almost invisible there isn…
ytr_Ugz8G41vm…
G
1. We know this since 80s. It's called Moravec's paradox. Reasoning is computati…
ytc_UgxiSulYR…
G
AI is marvelous for who is allowed to be left living. Its going to be beneficia…
ytc_UgzrBTXzT…
Comment
So, you're saying that being trained on human output doesn't make the the AI human-like "at all", because "what human has read all the text?" I don't see the logic in that. If you train a neural network on the output of another neural network (the human brain), then it will become similar to that neural network.
It's good to hear the idea that "AI safety is dangerous", becuase that's exactly what I thought when a long time ago I first read Eliezer write about the impossibility of safely aligning the utility function of a super-human AGI. I think that designing an explicit utility function and then trying to make the benevolence mathematically provable, is a dangerous approach. This approach is a self-fulfilling prophecy. I think it's much safer to just train the AI on human output, and ensure its humanness that way, without ever defining an explicit utility function.
youtube
AI Governance
2024-04-10T19:2…
♥ 1
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | unclear |
| Reasoning | mixed |
| Policy | unclear |
| Emotion | mixed |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgwHB3zmkdgBtRd6Y5h4AaABAg","responsibility":"company","reasoning":"virtue","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgxdNrV3cv1eq5RZPMt4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"ban","emotion":"fear"},
{"id":"ytc_UgzYM1xYeWovRXAzHrt4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_Ugwahq6aDz5bhamDiGd4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugyog867X5QH-LdCyzF4AaABAg","responsibility":"company","reasoning":"virtue","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgzTjacYjEIQjjTV5pR4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UgzsyJlZUKZvhx3p-qp4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"industry_self","emotion":"approval"},
{"id":"ytc_UgwEGfSsJcVXZou2CA54AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugy14FnnIIqGD6dVWat4AaABAg","responsibility":"government","reasoning":"deontological","policy":"regulate","emotion":"mixed"},
{"id":"ytc_Ugz0zx2y9HYuhxrDS1V4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}
]