Astral Codex Ten Podcast
…the bad news is that they can’t agree which one. I explained the debate more , but the short version is: twin studies find that most traits are at least 50% genetic, sometimes much more. But molecular studies - that is, attempts to find the precise genes responsible - usually only found enough genes for the traits to be ~10-20% genetic. The remaining 35% was dubbed “missing heritability”. Nurturists argued that the twin studies must be wrong; hereditarians argued that missing effect must be in hard-to-find genes. The latter seemed plausible because typical genetic studies only...
info_outlineAstral Codex Ten Podcast
If we worry too much about AI safety, will this make us “lose the race with China”? (here “AI safety” means long-term concerns about alignment and hostile superintelligence, as opposed to “AI ethics” concerns like bias or intellectual property.) Everything has tradeoffs, regulation vs. progress is a common dichotomy, and the more important you think AI will be, the more important it is that the free world get it first. If you believe in superintelligence, the technological singularity, etc, then you think AI is maximally important, and this issue ought to be high on your mind. But...
info_outlineAstral Codex Ten Podcast
Most discourse on AI is low-quality. Most discourse on consciousness is super-abysmal-double-low quality. Multiply these - or maybe raise one to the exponent of the other, or something - and you get the quality of discourse on AI consciousness. It’s not great. Out-of-the-box AIs mimic human text, and humans always describe themselves as conscious. So if you ask an AI whether it is conscious, it will often say yes. But because companies know this will happen, and don’t want to give their customers existential crises, they hard-code in a command for the AIs to answer that they aren’t...
info_outlineAstral Codex Ten Podcast
ACX has been co-running a forecasting contest with Metaculus for the past few years. Lately the “co-running” has drifted towards them doing all the work and giving me credit, but that’s how I like it! included more than 4500 forecasters predicting on 33 questions covering US politics, international events, AI, and more. They’re preparing for this year’s contest, and currently looking for interesting questions. These could be any objective outcome that might or might not happen in 2026, whose answer will be known by the end of the year. Not “Will Congress do a good job?”, but...
info_outlineAstral Codex Ten Podcast
Last year, I wrote that it would be . Commenters argued that no, it would be easy, just build more jails and mental hospitals. A year later, San Francisco feels safer. Visible homelessness is way down. But there wasn’t enough time to build many more jails or mental hospitals. So what happened? Were we all wrong? Probably not. I only did a cursory investigation, and this is all low-confidence, but it looks like: There was a big decrease in tent encampments, because a series of court cases made it easier for cities to clear them. Most of the former campers are still homeless. They just...
info_outlineAstral Codex Ten Podcast
“Life is suffering” may be a Noble Truth, but it feels like a . Yes, obviously life includes suffering. But it also includes happiness. Many people live good and happy lives, and even people with hard lives experience some pleasant moments. This is the starting point of many people’s objection to Buddhism. They continue: if nirvana is just a peaceful state beyond joy or suffering, it sounds like a letdown. An endless gray mist of bare okayness, like death or Britain. If your life was previously good, it’s a step down. Even if your life sucked, maybe you would still prefer the heroism...
info_outlineAstral Codex Ten Podcast
In Jason Pargin’s , a manic pixie dream girl cajoles a shut-in incel loser to drive her and her mysterious box cross-country. The further they drive, the more evidence starts to build that she is a terrorist and her box is a nuke. As our protagonist becomes increasingly desperate to turn around and return to his comfortable world of social media feeds and psych meds, she pleads with him to come out of his shell, learn to trust people offline, and have a sense of adventure. The book’s dramatic tension comes from our simultaneously rooting for his character development and worrying that it...
info_outlineAstral Codex Ten Podcast
American Scholar has , including and . It’s good that this is getting more attention, because in theory it seems like one of the most influential things a writer could do. In practice, it leaves me feeling mostly muddled and occasionally creeped out. “Writing for AI” means different things to different people, but seems to center around: Helping AIs learn what you know. Presenting arguments for your beliefs, in the hopes that AIs come to believe them. Helping the AIs model you in enough detail to recreate / simulate you later. Going through these in order:
info_outlineAstral Codex Ten Podcast
[I haven’t independently verified each link. On average, commenters will end up spotting evidence that around two or three of the links in each links post are wrong or misleading. I correct these as I see them, and will highlight important corrections later, but I can’t guarantee I will have caught them all by the time you read this.]
info_outlineAstral Codex Ten Podcast
Would You Like To Buy A Bahama? The Bahamas is an archipelago-nation of 400,000 people scattered across 3,000 small islands. The Bahamas’ most populous island is the one with its capital, Nassau. The second-most-populous - and fifth-largest, and most-pretentiously-named - is Grand Bahama, home of Freeport, the archipelago’s second city. Grand Bahama has a unique history. In 1955, it was barely inhabited, with only 500 people scattered across a few villages. The British colonial government turned it into a charter city, awarding the charter to , an American whose Wikipedia article describes...
info_outlineI.
Eliezer Yudkowsky’s Machine Intelligence Research Institute is the original AI safety org. But the original isn’t always the best - how is Mesopotamia doing these days? As money, brainpower, and prestige pour into the field, MIRI remains what it always was - a group of loosely-organized weird people, one of whom cannot be convinced to stop wearing a sparkly top hat in public. So when I was doing AI grantmaking last year, I asked them - why should I fund you instead of the guys with the army of bright-eyed Harvard grads, or the guys who just got Geoffrey Hinton as their celebrity spokesperson? What do you have that they don’t?
MIRI answered: moral clarity.
Most people in AI safety (including me) are uncertain and confused and looking for least-bad incremental solutions. We think AI will probably be an exciting and transformative technology, but there’s some chance, 5 or 15 or 30 percent, that it might turn against humanity in a catastrophic way. Or, if it doesn’t, that there will be something less catastrophic but still bad - maybe humanity gradually fading into the background, the same way kings and nobles faded into the background during the modern era. This is scary, but AI is coming whether we like it or not, and probably there are also potential risks from delaying too hard. We’re not sure exactly what to do, but for now we want to build a firm foundation for reacting to any future threat. That means keeping AI companies honest and transparent, helping responsible companies like Anthropic stay in the race, and investing in understanding AI goal structures and the ways that AIs interpret our commands. Then at some point in the future, we’ll be close enough to the actually-scary AI that we can understand the threat model more clearly, get more popular buy-in, and decide what to do next.
MIRI thinks this is pathetic - like trying to protect against an asteroid impact by wearing a hard hat. They’re kind of cagey about their own probability of AI wiping out humanity, but it seems to be somewhere around 95 - 99%. They think plausibly-achievable gains in company responsibility, regulation quality, and AI scholarship are orders of magnitude too weak to seriously address the problem, and they don’t expect enough of a “warning shot” that they feel comfortable kicking the can down the road until everything becomes clear and action is easy. They suggest banning all AI capabilities research immediately, to be restarted only in some distant future when the situation looks more promising.
Both sides honestly believe their position and don’t want to modulate their message for PR reasons. But both sides, coincidentally, think that their message is better PR. The incrementalists think a moderate, cautious approach keeps bridges open with academia, industry, government, and other actors that prefer normal clean-shaven interlocutors who don’t emit spittle whenever they talk. MIRI thinks that the public is sick of focus-group-tested mealy-mouthed bullshit, but might be ready to rise up against AI if someone presented the case in a clear and unambivalent way.
Now Yudkowsky and his co-author, MIRI president Nate Soares, have reached new heights of unambivalence with their new book, If Anyone Builds It, Everyone Dies (release date September 16, currently available for preorder).
https://www.astralcodexten.com/p/book-review-if-anyone-builds-it-everyone