Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
if the new new Zealand government took the wrongs they done and stood up own tha…
ytc_Ugxo75JzW…
G
the game itself hints at problems as humans tend to avoid each other by then. to…
ytr_UgxSLnOWw…
G
the movements are stiff, the textures look blurry and undefined and the heart in…
ytc_UgwmqN1pf…
G
My brother and I tested it, and expressing gratitude weirdly enough does change …
ytc_Ugz4BG4iC…
G
When something is yours it’s yours regardless of who or what tries to take it. …
ytc_Ugzb_QuV0…
G
You should just have your speech transcribed using the whisper API and then feed…
ytc_Ugw9oPWQS…
G
As a professional animator: I really really dont care if we are being “mean” to …
ytc_UgxLrHRUP…
G
I mean, putting ethics and copyright issues aside (let's say, these are all magi…
ytc_UgwI2nnLr…
Comment
Yes and no. Because AI needs an input (and an output) prompt, the current way it works probably won't do the full set of tasks better than a human can. If the only aim was productivity, humans working with AI would probably be the most effective.
Thing is, which few people account for: Companies hate workers. Personnel costs is the majority of costs of most (if not all) companies. So even if AI will do the job worse than people, *it would still be worth it for the companies.* Cutting some production in favor of vastly lowered costs (we're talking 50% or something), would probably be a winning strategy.
Not only that, but by cutting workers off of production, you remove their power. Strikes? Useless. Boycotts? Outdated. Unions? Pointless. Without being tied to production, workers can't make a reason for companies to care about their rights or needs.
So basically, AI probably won't be as good as humans are, but it doesn't need to be. It just needs to be good enough.
youtube
AI Jobs
2025-09-10T22:3…
♥ 1
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | company |
| Reasoning | deontological |
| Policy | regulate |
| Emotion | outrage |
| Coded at | 2026-04-27T06:24:59.937377 |
Raw LLM Response
[
{"id":"ytr_Ugyj4hmaQ7m2ABDVKwZ4AaABAg.AMv64sHUSOmAMw0hCsg3_E","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"resignation"},
{"id":"ytr_UgwbFVeBPLe-u8hUcSN4AaABAg.AMucBAnSrfVAMva3jtb-ye","responsibility":"company","reasoning":"mixed","policy":"liability","emotion":"indifference"},
{"id":"ytr_Ugy0bB56BfpdtS-1_u14AaABAg.AMuEjiNzuQRAMu_i0N-yEy","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytr_Ugy0bB56BfpdtS-1_u14AaABAg.AMuEjiNzuQRAMviBII6Tgn","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytr_Ugy0bB56BfpdtS-1_u14AaABAg.AMuEjiNzuQRAMvi_UQDGul","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytr_UgxdPJEGAbpI5rB-Svh4AaABAg.AE_Cn_ZPUOjAPchAc6vuQn","responsibility":"developer","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytr_UgzHyOQBTTjDmIyBO5d4AaABAg.AEQqMR514UtARylXQgKw2o","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytr_UgxHfGnh5L7CdBhzQvF4AaABAg.AEPGOMfszOMAPch3Oa4M2V","responsibility":"distributed","reasoning":"contractualist","policy":"regulate","emotion":"resignation"},
{"id":"ytr_UgzH6AIwAo7-NJjkrmJ4AaABAg.A0oHwbHLDfyA0xSbW9xpLy","responsibility":"developer","reasoning":"virtue","policy":"liability","emotion":"outrage"},
{"id":"ytr_Ugw1i8eGpRAK2YvzMH94AaABAg.9rxxb5j63hI9sJX5ANahNN","responsibility":"user","reasoning":"deontological","policy":"none","emotion":"indifference"}
]