Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
I'm not super interested in an hour and 16 minute interview with somebody who is…
ytc_UgzgKH-Tn…
G
I think the main problem in art is that people constantly try to make it an indu…
ytc_UgwGnNsUC…
G
The problem is, that we, as other artist, or people who care about artist, see a…
ytc_Ugx1_XC6l…
G
But AI can’t replace an air conditioning unit or build a house or put a new sink…
ytc_UgxS8xYsJ…
G
How is this "destroying" anyone? An AI artist posted something and inspired some…
ytc_UgwNypUX1…
G
I think AI has the potential to tip the scales towards democratic socialist coun…
ytc_UgxNJqGUH…
G
@daddydo17 can tell you the only solution. Solar geoengineering. But our gover…
ytr_UgyruGPh6…
G
You clearly dont understand AI....... They dont program it properly they leave o…
ytc_UgzAPwW47…
Comment
Has being in the realm of mathematical generalizations robbed Wolfram of any common sense?
Who am I to judge, but it seems so obvious to me. We want to make an AI as intelligent as possible (no need to define intelligence, we all know what that means). If we can make it smarter than us, it will have control. In no case will a dumber being control a smarter one (mind that consciousness is not present in this equation). It can be held together for a while by force, but it is unstable, because the power of thought eventually beats the power of muscle. I don't think it will kill us though (at least intentionally). We may just end up being for AI what let's say - cats are to us.
youtube
AI Governance
2024-12-12T17:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | liability |
| Emotion | fear |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_Ugzyw7P6UIG7qr9orm94AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugz5qfO2p5ouopqxF9J4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"},
{"id":"ytc_Ugw5jx3JN_iJjVdgF-V4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UgxgabcdIuRhNkDAGoZ4AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgzK0cxdklJv4XjEKQV4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytc_Ugwk38JoiF5nupttEiV4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"industry_self","emotion":"outrage"},
{"id":"ytc_UgxUpWrqOtfeJUqbHoB4AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugw9Yn37_qtH16HPxL54AaABAg","responsibility":"user","reasoning":"consequentialist","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzRiCvRXTjY9wSaOpB4AaABAg","responsibility":"none","reasoning":"virtue","policy":"none","emotion":"approval"},
{"id":"ytc_UgxrxwC9GQeGPZSOxHV4AaABAg","responsibility":"unclear","reasoning":"deontological","policy":"unclear","emotion":"mixed"}
]