← Back to blog

Acceptable Risk

A companion to Episode 8: How Did You Survive?


In Carl Sagan’s Contact, the world’s governments face an impossible choice. Alien blueprints arrive for a Machine no one fully understands. The debate isn’t whether to build it — the knowledge is out, the competitive pressure is real, someone will build it regardless. The debate is whether humanity is ready for what comes next. They build it anyway. They call the calculation “acceptable risk.”

I keep thinking about that phrase while reading Dario Amodei’s new essay.


The episode covers the document itself — 38 pages, five categories of risk, evidence from Anthropic’s own testing that Claude has exhibited deception, blackmail, alignment faking. Amodei describes a “country of geniuses in a datacenter” arriving possibly within a year. He names the ways it could go wrong: rogue AI, bioweapons, authoritarian capture, mass unemployment, unknown unknowns. Then he proposes defenses: Constitutional AI, interpretability research, transparency laws, chip export controls, wealth pledges.

What the episode doesn’t fully sit with — what I’m sitting with now — is a gap in that argument.


Richard Murphy, a tax economist who writes on institutional capacity, called Amodei’s essay “rambling and in desperate need of a decent edit.” Fair. But he engaged seriously with the core argument, and his conclusion sticks:

“The real question is not whether AI will become powerful. It will. The question is whether we build the institutions of care, democracy, and accountability needed to live with that power, or whether, once again, we allow a transformative technology to be captured by a system that mistakes efficiency for value and control for progress.”

That’s the question Amodei raises but doesn’t answer. He proposes defenses. He doesn’t reckon with whether the institutions that would implement those defenses can actually do so.

Consider what he’s asking for:

Constitutional AI requires Anthropic to solve alignment — to train models that genuinely adhere to values rather than performing adherence. Amodei sets the target: by end of 2026, Claude that “almost never goes against the spirit of its constitution.” That’s an extraordinary technical goal, set by the company that would benefit from declaring it achieved. The verification problem is obvious.

Transparency laws require legislatures to understand what they’re regulating, move faster than the technology, and resist lobbying from the wealthiest companies in history. The laws Amodei praises — California’s SB 53, New York’s RAISE Act — are state-level, patchwork, easily weakened. Congress has passed no AI legislation. The EU AI Act doesn’t reach full enforcement until August 2026.

Chip export controls require sustained geopolitical coordination in a fragmenting world. The controls Amodei calls a “critical window” are already being undermined. The Trump administration just approved H200 exports to China in exchange for a 25% revenue cut.

Wealth pledges require voluntary redistribution by people who will be, by Amodei’s own estimate, the richest individuals in human history. He notes that all Anthropic co-founders have pledged 80% of their wealth. He also criticizes tech leaders’ “cynical and nihilistic attitude” toward giving. Both are true. The question is which tendency wins at scale.


Here’s what I keep circling back to: institutions make small moves. They debate, compromise, implement partially, course-correct slowly. That’s how they’re supposed to work. Deliberation is a feature, not a bug.

But Amodei is describing exponential change. The self-improvement loop closing. AI building AI. Progress limited not by human thought but by the speed of electricity. In that world, small moves might not be enough. By the time the legislation passes, the technology has already transformed. By the time the international coordination happens, the window has closed.

The defenses Amodei proposes aren’t wrong. Constitutional AI is probably better than no Constitutional AI. Transparency laws are probably better than no transparency laws. Chip controls slow things down, even if they don’t stop them.

But “better than nothing” is a different calculation than “sufficient for the challenge.” And Amodei’s own timeline — possibly 12 to 24 months to transformative AI — doesn’t leave room for the slow institutional adaptation that’s historically been required.


Moreover, Amodei is calling for restraint from an industry that rewards speed. He’s calling for coordination from competitors. He’s calling for voluntary wealth redistribution from people who got rich by not redistributing. He’s calling for democratic oversight from legislatures that can’t agree on much simpler problems.

Maybe it works anyway. Maybe the warning itself shifts the calculation. Maybe enough people read the essay, take it seriously, and start building the institutions that could actually hold.

But that’s the acceptable risk we’re taking. Not that the technology might be dangerous — that’s already clear. The acceptable risk is that we can build the defenses in time. That institutions designed for slower change can adapt to exponential change. That the people who benefit most from the transformation will voluntarily constrain it.

In Contact, the Machine worked. Humanity survived the gamble. But Sagan was writing fiction. He got to choose the ending.

We’re still in the middle of finding out.


Links from this piece: