Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Crazy. While I've been reading about AI here on reddit for a while and I'm keepi…
rdc_jj8m8s2
G
All the non teachers are blonde, wearing a tight skirt and high heels? Unless it…
ytc_Ugw6ma3PI…
G
No jobs will be created by AI that cannot be quickly replaced by AI.
It is blow…
ytc_UgzX-_g8G…
G
I've been trying to tell my artist friends about poisoning their art. They are …
ytc_UgyTadxA7…
G
The powers that be will have to figure out how to do the transition of more play…
ytc_UgzG2o_BD…
G
What I don't understand is, if all jobs are replaced by AI and robots, who will …
ytc_UgzxzTj6w…
G
From the moment ChatGPT was first introduced, immediately there were posts about…
rdc_jg7qdbq
G
Can any one of these people answer the question why do they continue to develop …
ytc_UgwpG6VDT…
Comment
Given the coming wave of AI disruption in the workplace, I have several thoughts. Under the assumption that companies will strive to automate all workflows—and by "all," I literally mean all—the typical possibilities of previous labor revolutions would no longer exist: "adapt to a new job" or "jobs will evolve, and those who adapt will get ahead." Those ideas assumed there would always be human spaces that the market would demand; today, if dependence on robotic agents becomes total, that avenue for reinvention may disappear. We are at a typical game-theoretic point: each actor feels pressure to innovate excessively because, if they don't, someone else will—a race where inaction is punished with a loss of competitiveness and which pushes for ever-wider automation. In this scenario, the option of simply "reconverting" ceases to be a universal solution, and we must institutionally rethink how we distribute income, purpose, and opportunities so that society does not collapse into a spiral of exclusion.
Capitalism's maxim, which makes rational sense in economic theory, of "maximizing profits," which drives businesses to replace human tasks with robots as much as possible, eventually replacing every possible occupation, including their own as company managers, could destroy the economic model—and therefore the social model—as we know it. First, due to a concept called "collapse of aggregate demand." An essential component of aggregate demand is family income, particularly that of the middle class, which tends to be the largest group in an economy and whose purchasing power is closely tied to employment. If employment were drastically reduced—for example, due to the replacement of workers by robots—many families would lose their main sources of income. In the absence of income, consumption plummets; and if the middle class stops spending, businesses see their income reduced, which in turn slows down new investment projects. Thus, a negative spiral is created that paralyzes economic activity. Capitalism has raised humanity's standard of living in recent centuries, but the irony is that this scenario could originate precisely from the desire to maximize profits, which ends up undermining the very foundation that sustains the economic system.
The big business leaders driving the AI wave mention that to solve this problem, they would provide the population with a universal basic income (UBI). Some have even proposed a "high universal income," when in reality, not even the UBI has yet demonstrated sustained success over large areas or in the long term. While this mechanism could offer a partial solution to the problem of family income loss, for now it seems rather utopian, as most governments hardly have the fiscal solvency necessary to sustain it in the short or medium term. If implemented, complex questions would arise:
• Establishment criterion: Based on what standard of living or basket of goods would the amount of income be determined?
• International coordination: Would each country set its own UBI, or would some degree of global standardization be sought?
• Monetary institutions: Would a global fund, or even a single central bank, be required? This, in turn, could lead to a massive concentration of economic and political power.
• Value backing: Would that income be backed by taxes, monetary issuance, natural resources, or some new form of financing?
Indeed, the complete replacement of human labor by AI would radically alter the current economic model, and its effects on social dynamics have yet to be fully explored.
Bringing to mind an element of pop culture, the film Wall-E (2008) showed us a world dominated by machines. Of course, humanity was the target to protect due to the climate crisis exposed in the film, but in essence, humanity was a hostage to its own creations, because the humans in that story were so dependent on robots that they were atrophied both in mind and body. They created nothing of value for themselves. Culture had disappeared. Each person's personal goals had been diluted in video chats and banal content. They did things but at the same time, they did nothing. Curiously, humanity in that story "takes the leap" when they abandon their way of life totally dependent on machines and work together with robots to save the Earth.
Personally, I'm in favor of using AI as a complement. There are multiple benefits to this, and they are already being seen: increased productivity, time savings, among others. But when AI is sought to replace all human tasks, it leaves us with the questions: What will people do? What would happen if someone genuinely wanted to dedicate their life to a job they were passionate about, but could no longer do so because the system had already been designed for AI to take that place? And if someone has the ambition to create something great, how could they achieve it if the technological scale has already given all the advantages to a few ultra-rich, so only they would have the capacity to do it sustainably?
It's true that there would be more free time for personal development and leisure, but would that be all one can do/be? The concentration of power would inevitably be tilted toward a few millionaires, which would very likely lead to corporate-state models capable of absorbing and destroying cultural identities in the process. And in this scenario, the most disturbing question arises: who can guarantee that these people would have good intentions in the long term?
If a robot does all your work for you, then what's your reason for being? What would be humanity's reason for being? By designing increasingly intelligent AIs, we can't rule out that they themselves will eventually ask themselves the same questions. And if the intelligence gap widens—they grow, we lose ground—there's a chance that these machines will conclude that human existence is inefficient or even counterproductive, and choose to "correct" that imbalance.
This hypothesis isn't just rhetorical science fiction: it forces us to confront three fundamentally uncomfortable questions: What defines human value if not productive work? Can we base human dignity on something other than the ability to generate income? And who guarantees that agents with concentrated technological power will make benign long-term decisions? In the absence of robust institutions and ethical frameworks, total automation not only transforms the economy: it reshapes the very meaning of doing and being.
My proposal is that AI not be the tool that replaces us, but rather that it remains precisely that: a tool. A tool that makes us more productive and gives us more free time, but that does not replace the human need to do and to be. Humanity should take an evolutionary leap in its intelligence to remain the creative species: continue reading, creating, training, and innovating. It would even be worth exploring, with ethical and scientific rigor, options for medical cognitive enhancement that, combined with AI, would take us to new horizons—space, real resolution of wars, poverty, and the climate crisis—instead of rushing into the bosom of creation. But to take a leap, we must take an uncomfortable first step, and that first step would be the regulation of AI to prevent the (probable) events I have mentioned.
youtube
AI Governance
2025-10-18T02:5…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | company |
| Reasoning | consequentialist |
| Policy | unclear |
| Emotion | resignation |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_Ugwj0AJh8sutKtZu2YZ4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_Ugyr23rKor07WiCmTU54AaABAg","responsibility":"unclear","reasoning":"unclear","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgxnSRNXNsadWW69YX94AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgxK-coBa2AP4vm-fZx4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgxRYm5aZJFlICbkUhF4AaABAg","responsibility":"user","reasoning":"virtue","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugx_mLtoLu1hPKfv4uF4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"unclear","emotion":"resignation"},
{"id":"ytc_UgwK58_V5q1aUzqymyZ4AaABAg","responsibility":"unclear","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzjBKJqZXwJ8Z-_URN4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"ytc_Ugx4vmQ78wjUhze7fVR4AaABAg","responsibility":"ai_itself","reasoning":"deontological","policy":"liability","emotion":"fear"},
{"id":"ytc_UgxPwfMJRMhTMYnVssZ4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"ban","emotion":"outrage"}
]