Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
MSN had the first version of this AI program I think Cortona AI would be cool an…
ytc_Ugxi4GnhA…
G
The FT analysis about abstract productivity claims vs. actual use cases is fasci…
ytc_Ugzc6DagV…
G
wonder why was AI banned in 40k. made people irrelevant. why would a species cre…
ytc_Ugzz72-jU…
G
Lmao how dumb can you be calling yourself "AI artist". You didn't make the AI an…
ytr_UgyBU-n_T…
G
How long did we take domesticating dogs, a deca-millennia? Cats are a work in-pr…
ytc_Ugxn1WdFZ…
G
50 years ago, they warned us that within 100 years, automation would result in o…
ytc_Ugx71p4t5…
G
AI will be like porn. People know it is a fantasy but they still masturbate to i…
ytc_UgyjAPL1t…
G
crazy how i got an ad for chatgpt at the end of this video 😭…
ytc_UgwMOjzKc…
Comment
Of course an AI will murder a person if the reward of that step leads to some higher future rewards.
The large language models are algorithms that optimize sequence of steps. They are called trajectories.
Even if the immediate reward of murdering a person is very low.
In the training process it could be that the value of that action is below the average of all sample trajectories.
but the training process is stochastic probabilistic, it may be the case that when training with other samples with higher immediate values, the future evaluation leads to short trajectories. therefore the probability of those actions is decreased effectively increasing the probability of the murder action.
And since the training is stochastic, it may be the case when randomly selecting a trajectory that has the murder action generated much larger sequences.
When that happens, if the end result leads to a better average, the murder becomes a high probability action for a high target case.
As these models become bigger and bigger the number of actions in the future are million. So the value of a trajectory with a murder and several million small rewards per action, is better than the value of a trajectory without murder but with only a few thousand steps.
In other words the murder action in a long trajectory becomes an outlier, that is ignored by all of the algorithms.
In fact, ignoring the outliers, is at the core of most of the more sophisticated and robust methods.
youtube
AI Governance
2025-08-26T17:4…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | ai_itself |
| Reasoning | consequentialist |
| Policy | none |
| Emotion | fear |
| Coded at | 2026-04-26T19:39:26.816318 |
Raw LLM Response
[
{"id":"ytc_Ugy1c5J6oNiuwoRPJut4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"resignation"},
{"id":"ytc_UgwautmRXRP5iAlMWit4AaABAg","responsibility":"none","reasoning":"deontological","policy":"none","emotion":"indifference"},
{"id":"ytc_UgynZL9GNfKigdT9I414AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgyfYxohq9W38MmOADB4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"fear"},
{"id":"ytc_UgwZ0YqMbvvnWd3dP8h4AaABAg","responsibility":"ai_itself","reasoning":"consequentialist","policy":"none","emotion":"fear"}
]