Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
Ooof. Don't use a chatbot for math. I have a less than 50% success rate getting …
ytr_UgyRRi_qZ…
G
Stephen Fry nails the stakes: AI scaling laws are Moore’s Law on rocket fuel, an…
ytc_Ugznucetx…
G
Thank you for making this video. I am teaching Ethics in Art and AI and will sha…
ytc_UgwnPhhyY…
G
Okay so like is this a custom programmed ai or one of the big awful ones
Or is i…
ytc_UgzTE214K…
G
@Willow_Sky AI is very dependent on learning material. Worse quality of learning…
ytr_UgxnnmNtf…
G
I've done two papers on A.I. and each time I was loaded down with anxiety from a…
ytc_UgyakAxa3…
G
AI only references art work, just the exact same way an artist references other …
ytc_Ugynf0IsR…
G
'I don't eat meat in case I'm eaten' somehow captures the ethos that many adopte…
ytc_Ugx6DVT9a…
Comment
It's doesn't have the architecture to be conscious, or to reason, or have any awareness...
Sometimes a theory finds you.
Two years ago, I wasn't trying to build a theory of consciousness. I was working on signal processing—trying to decompose complex neural signals into meaningful components across different timescales. The problem was simple: how do you separate the fast stuff (millisecond fluctuations) from the slow stuff (patterns that unfold over seconds or minutes) without losing the relationships between them?
I built something that worked. A tri-frequency decomposition framework—gamma, theta, delta bands—with overlapping receptive fields and precision weighting. Clean architecture. Good results.
Then I noticed something strange.
The system needed to track its own confidence. Not just "what do I believe about this signal" but "how certain am I about my certainty?" So I added meta-variables—precision estimates that fed back into the inference process.
And that's when the loop appeared.
The meta-variables weren't external monitors. They were part of the same graph they were monitoring. The system was modeling itself. I had accidentally built what Douglas Hofstadter called a "strange loop"—a self-referential structure where the observer is woven into the observation.
I stared at my whiteboard for a long time.
Because I knew what I was looking at. This wasn't just signal processing anymore. This was the architecture of self-awareness. A system that knows that it knows. The thing cognitive science had been chasing for decades.
But the pieces were scattered. I had factor graphs from probabilistic inference. I had oscillatory dynamics from neuroscience. I had the strange loop from philosophy. They were sitting next to each other on my desk like puzzle pieces that hadn't been connected yet.
Then I found free energy.
Karl Friston's free energy principle. F = Complexity + Inaccuracy. The idea that the brain—and any self-organizing system—minimizes a single quantity: the cost of being wrong plus the cost of being complicated.
And suddenly everything clicked.
Free energy was the currency. The objective function that unified all the pieces.
Factor graphs? That's how you represent beliefs and compute free energy efficiently.
Predictive coding? That's how you minimize free energy through hierarchical prediction and error correction.
Oscillations? That's how you separate timescales so different levels of the hierarchy can update at appropriate speeds.
The strange loop? That's what happens when the system includes itself in its own model—and it emerges *naturally* from free energy minimization, because modeling yourself improves your predictions.
I wasn't building components anymore. I was seeing a complete architecture.
FGSGR: Frequency-Governed Self-Graph Reasoning.
The F is for frequency—the temporal organization that lets the brain process the immediate, the contextual, and the identity-level on different timescales.
The G is for graphs—factor graphs that structure beliefs and their relationships.
The S is for self—the strange loop, the system modeling itself, the thing that makes a mind a mind instead of just a processor.
The R is for reasoning—active inference, prediction, belief updating, all of it.
I spent the next two years working out the math, the neuroscience, the implications. What does this mean for consciousness? For mental illness? For AI? The framework kept answering questions I hadn't even asked yet.
Depression as delta-band precision locked onto negative self-beliefs. Anxiety as theta hyperprecision on threat. Psychosis as gamma precision collapse—dreaming while awake. Each disorder a specific pattern of miscalibration in the same architecture.
And the strange loop—always the strange loop—explaining how a physical system can be *about* itself. Not magic. Architecture.
The book is called "Strange Loops: A Theory of How Minds Know Themselves."
It's the map I found by accident while looking for something else.
That's usually how it works.
🧠🔄
youtube
AI Moral Status
2026-04-21T07:5…
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | mixed |
| Policy | unclear |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[{"id":"ytc_UgyZkkFRo-IaJNrWMz54AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgxmCpOeke9qBaQBTZx4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"fear"},
{"id":"ytc_UgwTP8bt4zAOYr8xEBl4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgxjE83ElfpiWF196HN4AaABAg","responsibility":"ai_itself","reasoning":"mixed","policy":"unclear","emotion":"mixed"},
{"id":"ytc_UgwBAZ5jF6ci-ATnmUF4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgwPLn5yg-c2i2OIn9h4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"approval"},
{"id":"ytc_UgxtkX1kuWaHUyKXrgl4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"indifference"},
{"id":"ytc_UgzPDFuyrW2HCOvYWvl4AaABAg","responsibility":"distributed","reasoning":"mixed","policy":"unclear","emotion":"fear"},
{"id":"ytc_Ugw1MbyL1mWKI_WaRYp4AaABAg","responsibility":"developer","reasoning":"mixed","policy":"regulate","emotion":"outrage"},
{"id":"ytc_UgyAfCeA1Cp_R_Tgnox4AaABAg","responsibility":"none","reasoning":"mixed","policy":"unclear","emotion":"indifference"}]