Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
@Constellasian it shouldn't have been invented in first place if I'm being fair …
ytr_UgyknPmTh…
G
She is looking like the woman behind her in black. Atleast this robot will be a …
ytc_UgwtX_yxh…
G
Would ai do a better job of governing our countries?
In which case would ai regu…
ytc_Ugwm8V9vf…
G
I was expecting a video showing accidents. Instead it's a bunch of Luddites st…
ytc_UgzeioZji…
G
great video
i myself for a long time have found myself on a middle ground on thi…
ytc_Ugy0KWXBQ…
G
Re: AI Adoption vs. Resistance. Be careful NOT to make the assumption that older…
ytc_UgypX_Zex…
G
Why is no one talking about the fact that maybe AI will take over but wont kill …
ytc_UgwkNFPKt…
G
Why did they try this in the wild? Don't a lot of legal firms still hold mock co…
ytc_Ugz9DTGE9…
Comment
The human brain has demonstrated its ability to generate new knowledge. However, can an LLM generate new knowledge? If you train a model with all the information available up until the 19th century and then ask it to explain the Michelson–Morley experiment, would it be able to develop the Theory of Relativity or something similar? Or what if I ask it to solve the technological challenge of traveling across the stars—essentially to invent a theory and develop the technology to make it possible? Could an LLM solve it? Let me question it further.
youtube
AI Moral Status
2026-03-02T14:3…
♥ 6
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | unclear |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[
{"id":"ytc_Ugw03Pw_SS-EprGVj694AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzNHcPp28P0drDgls94AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgyzhVmAkvibOQV6CTN4AaABAg","responsibility":"unclear","reasoning":"deontological","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgzT1RJ0fZewskES2SB4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_Ugz2QaUGdEwDG1-J2DN4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"unclear","emotion":"outrage"},
{"id":"ytc_UgxpYjNX6ApiZdeNnPF4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"indifference"},
{"id":"ytc_Ugy9NK4BYfqVYkWTu5p4AaABAg","responsibility":"none","reasoning":"unclear","policy":"none","emotion":"approval"},
{"id":"ytc_UgzhujBYZrjKr_tGU5h4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgzHQquKvdxUVXn-1V54AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgygGgl4-_Wb7JsD5cd4AaABAg","responsibility":"company","reasoning":"consequentialist","policy":"unclear","emotion":"resignation"}
]