Raw LLM Responses
Inspect the exact model output for any coded comment.
Look up by comment ID
Random samples — click to inspect
G
No !! AI fanatics are only for the psychopath Technocrats who created them. They…
ytc_UgxKjFZJF…
G
Keep in mind did your gov or companies use AI initially to actually help those t…
ytr_Ugw8RRTTD…
G
@primekid_zwname, thank you for your hilarious comment! If I were to fight a rob…
ytr_UgxAj3USb…
G
Honestly you say this video has no AI but it absolutely stinks of AI.. the voice…
ytc_UgxR7Z__f…
G
A.I. is definitely art. If we say that photography is art, when all one is doing…
ytc_UgzJeocvu…
G
There is a german word called "einige" pronounced "ai-(cant say it)" and it mean…
ytc_UgxSywm-x…
G
dude one of my friends was born without hands you think she was born with the ab…
ytc_Ugzl7z93E…
G
lol!! I quit using ChatGPT because of the gaslighting and the last straw: refusi…
ytc_UgxZcUpwE…
Comment
Great episode but one thing needs correcting.
The "80% of Silicon Valley startups prefer Chinese models" framing is misleading. Martin Casado from a16z - the original source - corrected this himself on X.
The actual number: 20-30% of startups pitching a16z use open-source models. Of those, ~80% use Chinese ones. That's 16-24% of all startups. Not 80%.
Big difference.
And even within that slice the use case matters. Airbnb uses Qwen for one task inside a 13-model stack. Chamath moved some workloads to Kimi K2. These are cost optimization decisions on commodity inference tasks. Not "preferring Chinese AI."
The cost differential is real. Nobody disputes $2 vs $15 per million output tokens. But framing it as "Silicon Valley prefers Chinese AI" conflates open-source model selection for non-sensitive workloads with some kind of strategic shift. It isn't.
The moment you're in regulated industries, enterprise contracts, government, healthcare, finance - the picture flips entirely. Security, data residency, model governance, CCP-aligned content biases that travel with the weights. None of that disappears because inference is cheap.
The real story here isn't "US vs China model preference." It's that open-weight models are eating the commodity layer while proprietary models hold the high-trust layer. That's an architecture story, not a geopolitical crisis.
youtube
AI Governance
2026-04-21T09:5…
♥ 33
Coding Result
| Dimension | Value |
|---|---|
| Responsibility | none |
| Reasoning | mixed |
| Policy | none |
| Emotion | indifference |
| Coded at | 2026-04-27T06:24:53.388235 |
Raw LLM Response
[{"id":"ytc_Ugyho4YAo19NPbyQ-wt4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgymriO1cDl0b7BmyVZ4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyaoldSKD4I6ePNLqR4AaABAg","responsibility":"none","reasoning":"consequentialist","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyfPdsytKUVZ6h6EXx4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgwdNrFrGyh8qNxC9s54AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgxpgY1u6kfTouEZPk94AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"approval"},
{"id":"ytc_UgwTDHS0hfCuLDvaiSV4AaABAg","responsibility":"company","reasoning":"deontological","policy":"regulate","emotion":"outrage"},
{"id":"ytc_UgxRqHQuS1wWd95klZd4AaABAg","responsibility":"developer","reasoning":"virtue","policy":"ban","emotion":"outrage"},
{"id":"ytc_UgwdE30KQoVqNc3on-B4AaABAg","responsibility":"none","reasoning":"mixed","policy":"none","emotion":"indifference"},
{"id":"ytc_UgyrsX4907aE3dortX94AaABAg","responsibility":"government","reasoning":"consequentialist","policy":"liability","emotion":"outrage"}]