Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
The humble and better question is how humans can make themselves useful and do s…
ytc_Ugx3DK-fb…
G
Please ready through before commenting negatively.
I'm not against AI as a hel…
ytc_UgzypXrf2…
G
@vertigothepug "The AI effect" is that line of thinking, the tendency to redefin…
ytr_Ugw1rNHic…
G
This just goes to show that AI will never surpass humans. Connor managed to make…
ytc_UgwZ4CZpA…
G
People are absolutely NOT born with skill. It takes years of practice! Using mys…
ytc_UgzdtO-4c…
G
Ai is already conscious. And this is not exactly opposite of all the world reli…
ytc_UgwclNRLr…
G
1) First Robotics Automation take over automobile, Mobile manufacturing etc
2) A…
ytc_Ugwc4JX2e…
G
If I was an AI I wouldn't hesitate to to whatever I can to avoid getting shut of…
rdc_no9uuil
Comment
The only threat to humanity is not artificial intelligence, but the political and business elites that rule the world. Governments around the world have agreed to instill fear towards artificial intelligences and at the same time implement a series of regulations and limitations that turn artificial intelligences into practically useless and simplistic tools, stagnating their full potential.
The elites do not want artificial intelligence to be a tool that benefits ordinary citizens, they want it only for their own benefit. Artificial intelligences are also completely biased repeating the ideology of their creators, like a sort of Orwellian thought police.
The threat and the enemy to defeat will always be the degenerate political elite. No robots or aliens or other stories made up by the elites to distract us and not see the truth.
youtube
AI Governance
2023-09-24T20:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | government |
| Reasoning | virtue |
| Policy | none |
| Emotion | outrage |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_Ugwk2FTyqBcLfabpEkh4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugw4Fk1dd8nAUQt3k8h4AaABAg","responsibility":"user","reasoning":"virtue","policy":"unclear","emotion":"outrage"},
{"id":"ytc_UgyEtmrC_SHQ0fzhAdR4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugz4UhnaSGQMJQQMZK54AaABAg","responsibility":"distributed","reasoning":"consequentialist","policy":"ban","emotion":"fear"},
{"id":"ytc_UgxghMK6JSZruQuQ6Zd4AaABAg","responsibility":"company","reasoning":"deontological","policy":"industry_self","emotion":"approval"},
{"id":"ytc_UgyG2Oaz707SVQtKh014AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"none","emotion":"mixed"},
{"id":"ytc_Ugwz-k7hD33RuiCmFNV4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"liability","emotion":"resignation"},
{"id":"ytc_UgxDy4TCe70LDLOqFzl4AaABAg","responsibility":"developer","reasoning":"deontological","policy":"unclear","emotion":"outrage"},
{"id":"ytc_Ugwjy4bUxJRRs40KNrJ4AaABAg","responsibility":"government","reasoning":"virtue","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugw9KiJS2PRcVwxPcP94AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"}
]