AI Won't End Nuclear Deterrence (Probably)
TL;DR
While advanced AI could theoretically undermine nuclear deterrence by tracking hidden arsenals or disabling command systems, the brutal physics of undersea warfare and inevitable move-countermove dynamics make the complete erosion of secure second-strike capabilities unlikely, preserving the 'balance of nerves' that limits great power coercion.
🌍 Why AI and Nuclear Deterrence Matter 2 insights
Deterrence limits technological domination
Nuclear deterrence prevents even economically and militarily superior states from imposing their will on rivals, creating a 'balance of nerves' rather than a pure balance of power.
Undermining deterrence raises coercion risks
If AI enabled a 'splendid first strike' or neutralized retaliation capabilities, technologically dominant states could threaten and coerce nuclear rivals with unprecedented impunity.
🛡️ The Mechanics of Second-Strike Capability 3 insights
Survivability ensures mutual destruction
Secure second-strike capability—the ability to retaliate after absorbing a nuclear attack—depends on hiding forces across land, sea, and air to ensure no adversary could destroy everything at once.
Submarines are the most survivable leg
Nuclear-powered ballistic missile submarines represent the ultimate deterrent because they operate in tens of millions of square miles of ocean where electromagnetic radiation cannot penetrate and acoustic detection faces massive noise and distortion.
States use different survivability strategies
The US maintains a triad of land-based missiles, submarines, and bombers; the UK relies solely on submarines; while Russia and China use road-mobile launchers driving on highways, a tactic deemed politically unviable in the US.
🤖 AI's Potential and Physical Limits 3 insights
Three pathways to erode deterrence
AI could theoretically enable a disarming first strike by locating all enemy weapons, disable nuclear command-and-control networks, or strengthen missile defenses enough to neutralize retaliation.
Sensor fusion faces physics barriers
While AI could fuse data from sonar, magnetic anomaly detectors, and satellite radar to track submarines, the ocean's vast volume, complex terrain, and increasing ambient noise create fundamental physical limits to transparency.
Move-countermove dynamics constrain AI
States will adapt to AI surveillance by engineering quieter submarines and deploying countermeasures, while deploying millions of sensors in contested waters risks sabotage and interference.
Bottom Line
Policymakers should invest in maintaining secure second-strike capabilities—particularly submarine stealth and resilient command-and-control systems—to ensure that AI advancements do not destabilize the nuclear balance that prevents great power war.
More from 80,000 Hours Podcast (Rob Wiblin)
View all
A ceasefire in Ukraine won’t make Europe safer
Samuel Charap argues that a Ukraine ceasefire alone won't reduce the risk of NATO-Russia war and may create a more volatile environment prone to accidental escalation through broken agreements, hybrid warfare, and miscalculation on an expanded NATO border.
How AI could let a few people quietly call all the shots
Rose Hadshar of Forethought explains how advanced AI could enable unprecedented power concentration not through dramatic coups, but via economic dominance and epistemic manipulation, allowing small groups to control millions of loyal AI workers while the general public loses political leverage.
Using AI to enhance societal decision making (article by Zershaaneh Qureshi)
Advanced AI could compress centuries of progress into years, forcing humanity to make existential decisions faster than ever; developing targeted AI decision-making tools now could help society navigate this critical period by improving collective intelligence before dangerous capabilities emerge.
Claude Thinks It's Italian American. What Does That Say About Consciousness?
Robert Long argues that while factory farming offers a cautionary tale about exploiting non-human minds, AI welfare requires distinct ethical frameworks because we design AI desires rather than discovering them, creating unique tensions between ensuring safety through alignment and granting AI systems autonomy to flourish independently.