Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
AI will probably wipe out the working class most definitely. It will also hurt s…
ytc_UgzywyOLm…
G
It’s not AI taking away jobs. Currently, these large organizations with these l…
rdc_npq3ypv
G
I have trouble picturing images in my mind, so using ai just to get a picture in…
ytc_UgwbZ-7r2…
G
Excellent dialogue. Let's all advocate for consent and for AI safety in any way …
ytc_Ugy5ebI8V…
G
You are delusional, the AI art piece is way better than whatever they drew combi…
ytc_UgwAJyTVg…
G
It's Madame Tussaud's vs Pirates of the Caribbean. Sure, the face looks pretty r…
ytc_Ugw7DjxBh…
G
Got 5 sucks ass compared to 4... And 4 sucked ass too compared to Claude.…
ytc_UgwW54wQ-…
G
Excuse me? Has she used Google AI lately? Google AI won’t generate white people.…
ytc_Ugz1vM2Bu…
Comment
Yudkowsky: An AI will create subgoals which are not foreseeable by the designers. Most of these subgoals will include killing humans. And for humanity to end, it is sufficient if only one of the subgoals includes killing humanity as a side effect.
Problems with this argument (in arbitrary order):
1) It is not clear that the jump from AI as smart as humanity, to AI at a level ridiculously smarter than humanity, happens so fast that humanity could not adapt ( e.g. shut it off).
2) Killing humanity is not impossible, but really hard, if humanity can have a look into your brain at any time, analyze you with the help of other AIs, and shut you down or modify you whenever they want.
3) High Intelligence alone does not automatically mean that entities are effective at reaching goals. Look at many intelligent humans ...
4) Humans follow goals and generate subgoals in a similar unforeseeable manner. But these subgoals almost never include killing all of humanity.
5) It should be possible to add high-level goals which eliminate a lot of all potential dangerous subgoals by means of contradiction.
6) Just because the subgoals are not foreseeable, does not mean that most of them will include killing humans. They are unforeseeable and not arbitrary.
7) It is not clear that subgoals are not balanced out by other subgoals.
8) There might not be one AI, but actually several of them balancing each other.
9) .. Probably there are more ..
This discussion should have addressed this kind of counter-arguments.
Instead Yudkowsky did not actually try to convince anyone, and Wolfram did not address the obvious problems with Yudkowsky's theory, but instead explored all kinds of
"1+1 may not be 4 because who knows what a 1 is, in the first place?"
youtube
AI Governance
2024-11-14T09:2…
♥ 1
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | deontological |
| Policy | regulate |
| Emotion | mixed |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgwRvWP_k7v_jN9-Te14AaABAg","responsibility":"none","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_Ugyksdh6rn-4hBjfu214AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"},
{"id":"ytc_UgxlTd1d2AkohR8lVSZ4AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_UgxF1_HmuOODIl8KiOF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytc_UgzTMs1seu-Hm2wg1tB4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugzck-R6lKxbvEb8M5Z4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgyyLzF6cJe301DdxjF4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"mixed"},
{"id":"ytc_UgyL07Rq-EVfO1ActR94AaABAg","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"mixed"},
{"id":"ytc_Ugz6Llf_yDF9Gc34V9B4AaABAg","responsibility":"ai_itself","reasoning":"virtue","policy":"industry_self","emotion":"approval"},
{"id":"ytc_UgzaIf0jFeodxvBJt2d4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"ban","emotion":"outrage"}
]