Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Dude its not gonna get fixed because the more logical you can make an AI, the mo…
ytc_UgxRdhV1i…
G
Oh! Wow!
Face Recognition Technology is "Racist" TOO!!!
This is the BullSh*t tha…
ytc_UgyMyp5bR…
G
Not true prominent AI researchers have the same concerns
And I don't think llm…
ytr_UgyWwKhvz…
G
never related more to the accident one
ILL JUST BE TRYING TO FIGHT AI, BEAT THE…
ytc_Ugywdz9bJ…
G
Disabled artist here as well. Different though, I'm a writer. I have autism, ADH…
ytc_UgyJeFi8s…
G
I think sixty minutes is picking on Tesla when self driving vehicles in China se…
ytc_Ugym9Nfn9…
G
I feel bad for Siri and other chatbots because they are being called primitive a…
ytc_UgzzCXZd0…
G
I noticed this during my earlier legal research that the chatgpt had habit of ci…
ytc_UgwMHdPlR…
Comment
They are glossing over all the economic problems. They weren't quiting over the AI getting to powerful, they where quitting over bad business decision, no clear way to monetize there success, capital burn rates, and other business factors that would drive the company into bankruptcy by the end of this year. LLMs are showing diminishing returns and will not reach AGI without new breakthrough. It's not a engineering problem it's still a scientific one, and there is no guarantee of progress on it. It may 10-100 years before we see progress again.
youtube
AI Governance
2026-03-17T04:1…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | company |
| Reasoning | consequentialist |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_UgxFBh7sICuefzof2kN4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwxTBPE9D4iYHNQevF4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"resignation"},
{"id":"ytc_UgwzGmtADaJEY6k8qmd4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwOFfzGwQ6MfDj6MdJ4AaABAg","responsibility":"ai_itself","reasoning":"unclear","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgwFprXAsM6XYxs3UQl4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"regulate","emotion":"outrage"},
{"id":"ytc_Ugx9UBQQH62TXoM3hoV4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyuvbCVMFKVZiHmmA14AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgwbFeUalc0LjZuDAhF4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwNWIG2RB08sYWiEpt4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"liability","emotion":"outrage"},
{"id":"ytc_UgyIP835IEjrJ4_2SXB4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"mixed"}
]