Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
bruh i also had the same disappointment finding out that was ai. but when you zo…
ytc_UgwbdOS--…
G
Proper AI is doing the same thing you are when you look at a human for inspirati…
ytc_UgyRXzY0Y…
G
Kuch bhi ho energy to Battery se hi lega. Aur AI ko charge bhi karna padega. Ple…
ytc_UgyX2OTIe…
G
No way not at this level. This is CGI. We didn’t have this level of video ai unt…
ytr_Ugx1MSXjH…
G
The psychosis is more widespread than even OpenAI would admit. These posts scare…
rdc_n7kfe7b
G
Man Terminator really locked in the cultural idea that a Superintelligent AI wou…
ytc_UgwTRvZUr…
G
Great reference! "Ex Machina" really dives into the complexities of AI and human…
ytr_Ugwt0s58Z…
G
There are still no AI tools. LLMs are fancy autocomplete with no concept of real…
ytc_Ugyttd0zo…
Comment
Anthropic and the scale of infringement
The core shock is not just that copyrighted material was used, but the sheer industrial scale of it. Downloading millions of pirated books turns AI training into a mass copyright breach, reframing AI development as something closer to systematic data extraction than innocent experimentation or passive learning.
The fair use versus theft distinction
The ruling draws a sharp legal line: transformation may be acceptable, theft is not. Training can qualify as fair use, but illegally sourcing material poisons that defence. This distinction matters because it forces AI companies to scrutinise data pipelines, not just model outputs or abstract learning claims.
Why Anthropic settled for $1.5 billion
Anthropic did not settle out of goodwill but survival. Statutory damages multiplied across millions of works posed an existential threat. The settlement buys certainty, not permission. It resolves past wrongdoing without legitimising future use, leaving the company financially bruised and strategically constrained.
Collapse of the ethical AI narrative
The reputational damage cuts deeper than the fine. Marketing AI safety while sourcing pirated data exposes a credibility gap that regulators and courts will not ignore. The case shows ethics statements mean little without operational discipline, and that courts are increasingly willing to interrogate corporate self-mythology.
An industry-wide legal reckoning
Anthropic’s case is a warning shot, not an anomaly. Lawsuits against OpenAI, Meta, and others show systemic exposure across the sector. Even partial wins come with judicial caveats. The message is clear: current training practices sit on legally unstable ground and will be challenged repeatedly.
Why AI learning is not human learning
The human-learning analogy breaks under scrutiny. Humans absorb imperfect impressions; machines ingest perfect, scalable copies. AI can replicate, remix, and compete at volume no human can match. That difference turns training into economic substitution, not inspiration, raising copyright concerns that human comparison cannot resolve.
Licensing and the power imbalance
Mandatory licensing favours incumbents. Large firms can absorb multimillion-pound deals; startups cannot. This risks freezing innovation behind paywalls of data ownership. While creators gain leverage, the market may consolidate further, concentrating AI power among firms rich enough to buy legality at scale.
A regulated future for training data
The case accelerates a shift towards regulated data ownership. Transparency, opt-outs, and paid access are becoming the norm. Ironically, AI still depends on human creativity to improve. The future model is clear: AI may continue to learn, but only by paying its teachers.
youtube
AI Responsibility
2026-01-12T15:2…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | company |
| Reasoning | deontological |
| Policy | liability |
| Emotion | outrage |
| Coded at | 2026-04-26T23:09:12.988011 |
Raw LLM Response
[
{"id":"ytc_UgxyE_Eitu4TJs7408F4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_Ugy3uLt_lthUgXvOBS94AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyX5Jo0GchtkidUnBt4AaABAg","responsibility":"government","reasoning":"mixed","policy":"none","emotion":"outrage"},
{"id":"ytc_Ugza0w6PDGluQnj3YUB4AaABAg","responsibility":"developer","reasoning":"consequentialist","policy":"regulate","emotion":"fear"},
{"id":"ytc_Ugw9d4Wyyi3_yP1-8zd4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"approval"},
{"id":"ytc_UgyH6hYLstLAUccNba54AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwbDT2-7wbyTWWn3Vd4AaABAg","responsibility":"company","reasoning":"deontological","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgznBi8Hcu9iaXv6uMB4AaABAg","responsibility":"government","reasoning":"mixed","policy":"none","emotion":"resignation"},
{"id":"ytc_UgziC_R1ElJclM6LmV94AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"outrage"},
{"id":"ytc_UgyMjPsvWoQhjNFFPSN4AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"}
]