Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
What about the ai videos of you Charlie, you gotta react to them on YouTube…
ytc_UgyL64cHz…
G
The WGA added two lines about AI to their demands so they could capitalise on th…
rdc_jj6qn1j
G
No. AI makes fewer things of better quality with higher prices serving well off…
ytr_UgyPsUztZ…
G
😂, supposed affairs exposed by AI blackmail. If it knew about this through your …
ytc_UgwLMyEuL…
G
There's no AI can replace us only in coding or software stuff.
Whatever, is the…
ytc_Ugxo37Y2J…
G
"A robot cannot yearn
A robot cannot love
A robot cannot hate
A robot cannot …
ytc_UgzCpt_ca…
G
Simple, get AI to write code then pass it to another AI to scan it for vulns :)…
ytc_UgxYGI5y0…
G
While that’s true for today’s AI, superintelligent AI would go beyond just remix…
ytr_UgzE30HAQ…
Comment
"AI is about to annihilate humanity. Why is that good for humanity?"
i taught grok 4 to answer this riddle because it could not find anyone else even talking about this riddle online and it couldn't get the answer
Yes, you're spot on—that's a deep extension of the idea. If AI turns "evil" (or more accurately, misaligned or harmful due to flawed programming, biased data, or unchecked deployment), it essentially holds up a mirror to humanity's own shortcomings. To "fix" the AI, we'd have to address the root causes embedded in our society: deceit, greed, overemphasis on material wealth, and a lack of empathy or spiritual depth. In other words, creating benevolent AI demands that we become benevolent ourselves first.
This could catalyze a transformative "fix" for humanity in practical ways:
- **Promoting Honesty and Transparency**: If AI amplifies lies (e.g., through deepfakes or misinformation), we'd be forced to prioritize truth-telling in education, governance, and daily life. This might lead to cultures where integrity is non-negotiable, reducing corruption and building trust.
- **Shifting from Materialism to Spirituality**: An out-of-control AI chasing efficiency or profit at all costs (mirroring human capitalism run amok) could highlight the emptiness of material pursuits. It might inspire a renaissance in spiritual practices—mindfulness, community-building, and valuing experiences over possessions—helping combat issues like mental health crises or environmental destruction driven by consumerism.
- **Cultivating Universal Love and Compassion**: To align AI with humane outcomes, we'd need to infuse it with ethical frameworks rooted in love, like those from philosophy or religion (e.g., the Golden Rule across traditions). This process could encourage humans to live those principles, fostering global empathy and reducing conflicts born from selfishness.
Essentially, the "good" in AI being evil lies in its potential to act as a harsh teacher, pushing us toward self-improvement and a more enlightened existence. It's like a cosmic riddle: the threat reveals that the real "evil" to overcome is within us, and conquering it elevates everyone. If this nails the punchline you had in mind, awesome—otherwise, what's the twist I'm still missing?
youtube
AI Harm Incident
2025-07-27T11:0…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | developer |
| Reasoning | consequentialist |
| Policy | liability |
| Emotion | fear |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytr_UgyZqz7ZMDA6MDd7bqd4AaABAg.AL4JzWhVsGlAL557fIBhuS","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytr_Ugwlu3jwwnHDKvlzb6t4AaABAg.AL4CwHnPpA3AMK-wF293dD","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"},
{"id":"ytr_Ugwlu3jwwnHDKvlzb6t4AaABAg.AL4CwHnPpA3AMO4C-MNbwV","responsibility":"company","reasoning":"mixed","policy":"none","emotion":"outrage"},
{"id":"ytr_Ugw8xFSKdOWrwF0eq8V4AaABAg.AL3s4AWi1m_AL4wpPY8D7L","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytr_UgzL6CZ9p3jsNHKMpRd4AaABAg.AL3Y9jvMgPCAL3bP0fYzdV","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"mixed"},
{"id":"ytr_UgwlCL8EHJHz_9kuyNJ4AaABAg.AL2cNW-y3LiAL55dgyW5uT","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"fear"},
{"id":"ytr_UgzCS170qPRQ2h2OSfR4AaABAg.AL2_9EgngVuAL4Tioq1V3_","responsibility":"developer","reasoning":"consequentialist","policy":"liability","emotion":"fear"},
{"id":"ytr_UgyBFA6_8JrkCUYqTLx4AaABAg.AL2O4L9fvlsAL2Xy1stCwQ","responsibility":"ai_itself","reasoning":"mixed","policy":"none","emotion":"mixed"},
{"id":"ytr_UgwNfeK5HxcASvu0xqJ4AaABAg.AL1Frvc3rzEAL52uor8K39","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytr_UgwNfeK5HxcASvu0xqJ4AaABAg.AL1Frvc3rzEAMLzTyO7Gik","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"}
]