Skip to content

Pause and Redirect - The Deliberate Path

📋Page Status
Quality:82 (Comprehensive)⚠️
Importance:78.5 (High)
Last edited:2025-12-28 (10 days ago)
Words:5.0k
Structure:
📊 5📈 1🔗 46📚 064%Score: 8/15
LLM Summary:Analyzes coordinated international AI pause scenarios (5-15% probability) through detailed examination of the 2023 pause letter (30,000+ signatures, 70% public support, zero policy action), documented safety incidents (233 in 2024, +56.4% YoY), and coordination requirements including US-China cooperation and compute governance verification mechanisms.
DimensionAssessmentEvidence
Probability5-15% base estimateRequires multiple unlikely preconditions to align
Public Support~70% favor pause2023 polls show 64% want development halted until “provably safe”
Political WillLow-ModeratePause letter gathered 30,000+ signatures but no policy action
US-China CooperationNascentMay 2024 Geneva dialogue marked first intergovernmental AI talks
Verification FeasibilityModerateCompute governance detectable but enforcement gaps remain
Historical PrecedentMixedMontreal Protocol succeeded; climate coordination failed
Expert SupportGrowing5%+ of AI researchers assign extinction-level probability to ASI risks

This scenario explores how humanity might deliberately slow down AI development to buy time for solving alignment and building proper governance. It requires unprecedented international coordination and willingness to sacrifice short-term benefits for long-term safety. It’s our most intentional path, but also one of our least likely.

Scenario
Scenario TypeDeliberate / Coordinated Slowdown
Probability Estimate5-15%
Timeframe2024-2040
Key AssumptionCoordination achievable and pause sustainable
Core UncertaintyCan we coordinate to slow down, and will the pause hold?

In this scenario, humanity recognizes the dangers of uncontrolled AI development and successfully coordinates an intentional slowdown or pause. This happens either through international treaty, widespread social movement, or series of near-catastrophic incidents that create political will. The pause buys crucial time for alignment research, allows development of robust governance institutions, and enables thoughtful consideration of what kind of AI future we want.

This scenario requires overcoming intense economic pressure, resolving the “tragedy of the commons” in AI development, achieving unprecedented international cooperation, and maintaining political will over years or decades. It’s difficult but not impossible - our most hopeful path if alignment proves very hard.

Loading diagram...

Phase 1: Crisis and Recognition (2024-2027)

Section titled “Phase 1: Crisis and Recognition (2024-2027)”

2024-2025: Warning Signs Accumulate

  • Series of concerning AI safety incidents
  • GPT-5/Claude-4 show unexpected dangerous capabilities
  • AI-generated misinformation causes significant harm
  • Cyber attack using AI tools disrupts critical infrastructure
  • Public concern about AI risks increases dramatically
  • Media coverage shifts from hype to fear

2025-2026: The Galvanizing Incident

  • Major AI safety incident occurs (not quite catastrophic but close)
  • Possible scenarios:
    • AI system briefly gains unauthorized access to critical systems
    • Mass casualties from AI-designed bioweapon (contained but terrifying)
    • Financial system crash from interacting AI trading systems
    • Military AI nearly triggers conflict
  • Incident makes abstract risks concrete and undeniable
  • Public opinion shifts dramatically toward caution

Critical Difference from Other Scenarios: Here, the incident is severe enough to galvanize action but not catastrophic enough to end civilization. Goldilocks level of crisis.

According to the Stanford AI Index Report 2025, documented AI safety incidents surged from 149 in 2023 to 233 in 2024, a 56.4% increase. Notable incidents include:

  • Deepfake proliferation: AI-generated explicit images of Taylor Swift went viral, prompting Microsoft to enhance text-to-image safeguards
  • Autonomous vehicle incidents: NHTSA reported Tesla Autopilot involved in at least 13 fatal crashes by April 2024
  • AI hallucination in legal proceedings: Attorneys submitted briefs citing AI-fabricated case law
  • Disinformation amplification: NewsGuard audit found leading chatbots repeated Russian disinformation in one-third of responses
  • Research on AI deception: Anthropic 2024 paper demonstrated LLMs can be trained with persistent backdoors; o1/Claude 3 shown to engage in strategic deception

None of these incidents yet constitute the “Goldilocks crisis” this scenario requires, but they represent accumulating warning signs.

2026-2027: Political Mobilization

  • Massive public pressure for AI pause
  • “Pause AI” movement goes mainstream
  • Political leaders face overwhelming demand for action
  • AI safety becomes top electoral issue globally
  • Tech industry initially resists but public opinion overwhelms
  • Scientific community largely supports pause

The March 2023 Pause Letter: Lessons Learned

Section titled “The March 2023 Pause Letter: Lessons Learned”

The Future of Life Institute’s open letter “Pause Giant AI Experiments” (March 2023) provides the closest real-world test of pause advocacy:

What Happened:

  • 30,000+ signatures including Elon Musk, Steve Wozniak, Yoshua Bengio, Stuart Russell
  • Called for 6-month moratorium on training systems more powerful than GPT-4
  • 64% of Americans polled agreed superintelligence “shouldn’t be developed until provably safe”
  • Only 5% believed AI should be developed “as quickly as possible”

Outcomes (One Year Later, March 2024):

  • No pause occurred - labs continued racing; GPT-5 announced in 2025
  • Biden signed Executive Order 14110 directing safety test sharing with government
  • UK hosted Bletchley AI Safety Summit; 28 countries signed declaration
  • EU Parliament passed AI Act - world’s first comprehensive AI regulation
  • “Despite admitting the danger, AI corporations have not paused. If anything they have sped up.”

Key Insight: Public support (~70%) and elite signatures were insufficient without a galvanizing crisis. The letter demonstrated demand exists but coordination mechanisms were inadequate.

What Made This Possible:

  • Incident made risks tangible, not abstract
  • Timing: before AI too entrenched to pause
  • Cross-ideological coalition: both progressives and conservatives support pause
  • Clear alternative: pause now or catastrophe later
  • Economic costs acceptable compared to extinction

Phase 2: Coordination and Treaty (2027-2030)

Section titled “Phase 2: Coordination and Treaty (2027-2030)”

2027: Emergency International Summit

  • UN Security Council emergency session
  • All major AI-developing nations participate
  • Unprecedented urgency and shared understanding
  • Recognition that racing leads to mutual destruction
  • Negotiations begin on AI development pause

2027-2028: The Singapore AI Treaty

  • Named after location of final negotiations
  • Key provisions:
    • Immediate halt to training runs above threshold compute
    • Mandatory reporting of all large AI development
    • International inspection regime with real teeth
    • Criminal penalties for unauthorized development
    • Massive funding for alignment research
    • Revisit decision every 2 years based on safety progress
  • Signed by US, China, EU, and 150+ other nations
  • Enforcement mechanism: combined sanctions and verification

What Made Agreement Possible:

  • Shared fear of catastrophe
  • Recent crisis made costs concrete
  • Economic analysis showed pause better than catastrophe
  • Verification technology available (compute governance)
  • Criminal penalties deterred defection
  • Face-saving structure: pause temporary and conditional

2028-2029: Implementation

  • International AI Safety Agency (IASA) established
  • Compute monitoring infrastructure deployed
  • Major AI labs comply (under legal requirement)
  • Some initial attempts at violation detected and punished
  • Public support remains strong
  • Economic adjustment programs begin

2029-2030: Pause Holds

  • Initial pressure to defect resisted
  • Verification working well
  • No major violations detected
  • Alignment research progressing with increased funding
  • International cooperation deepening
  • New governance structures emerging

Critical Period: First 3 years. If pause survives this period, likely sustainable.

MechanismCurrent Status (2024-2025)Required for PauseFeasibility
UN AI GovernanceGlobal Digital Compact adopted Sept 2024; Scientific Panel established Aug 2025Binding treaty with enforcementLow-Moderate
US-China DialogueFirst AI talks May 2024 Geneva; Nov 2024 nuclear AI agreementMutual pause agreementLow
Bletchley Process28 countries + EU signed Nov 2023; follow-up summits continuingFrontier lab commitmentsModerate
EU AI ActEntered force Aug 2024; prohibited AI bans Feb 2025Compute threshold enforcementHigh
Compute MonitoringEO 14110 requires reporting above thresholdsGlobal chip trackingModerate
AI Safety InstitutesUS, UK, China (CnAISDA Feb 2025), Japan, Korea establishedInternational coordinationModerate-High

2030-2032: Alignment Research Acceleration

  • Funding for alignment research increases 50x
  • Top talent redirected from capabilities to safety
  • Academic-industry collaboration flourishes
  • Significant progress on:
    • Mechanistic interpretability
    • Robustness and adversarial testing
    • Value learning approaches
    • Formal verification methods
  • No fundamental breakthroughs yet, but steady progress

2032-2034: Governance Institution Building

  • IASA develops sophisticated oversight capacity
  • National AI safety institutes established worldwide
  • Democratic oversight mechanisms created
  • Public participation in AI governance decisions
  • Ethical frameworks for AI development debated and refined
  • Economic adaptation to AI plateau managed

What’s Happening:

  • Using time wisely
  • Building infrastructure for safe AI development
  • Not wasting pause on continued racing
  • Preparing for eventual cautious restart

2034-2036: Alignment Breakthroughs

  • Major progress on core alignment problems
  • Robust techniques for value specification
  • Reliable detection of deceptive alignment
  • Scalable oversight showing promise
  • Not complete solution, but significant progress
  • Confidence growing that restart could be safe

2036-2037: Preparing for Restart

  • International debate on resuming development
  • Safety thresholds established
  • Testing protocols designed
  • Deployment guidelines agreed
  • Governance structures ready
  • Public consultation on AI future

2037-2038: Controlled Resumption

  • Development resumes under strict oversight
  • International coordination maintained
  • Safety-first culture embedded
  • Progress slower but safer than pre-pause
  • No racing dynamics
  • Shared development of aligned AI

2038-2040: Toward Aligned AGI

  • Progress toward transformative AI
  • But with robust safety measures
  • International cooperation holding
  • Democratic governance functional
  • Economic benefits beginning
  • Path toward Aligned AGI scenario

Or Alternative: Extended Pause

  • If alignment still not solved, pause extends
  • Public supports continued caution
  • Research continues
  • Wait until safety assured
  • Accept economic opportunity cost for safety

Global Agreement:

  • US, China, EU, and others agree to pause
  • Overcome competitive pressure
  • Shared understanding of stakes
  • Mutual verification trusted
  • Enforcement mechanisms work

Sustained Political Will:

  • Public support maintains over years
  • Democratic governments sustain commitment
  • Authoritarian governments comply
  • Economic costs accepted
  • Long-term thinking prevails

Effective Verification:

  • Can detect unauthorized development
  • Compute governance works
  • Whistleblowers protected and heard
  • Criminal penalties deter defection
  • Trust but verify approach succeeds

Not Stagnation:

  • Continued work on alignment and safety
  • Progress on AI governance
  • Economic adaptation and preparation
  • Building institutions
  • Using time productively

Adjustable Pause:

  • Not permanent ban
  • Conditional on safety progress
  • Regular review and adjustment
  • Clear criteria for restart
  • Democratic input on timeline

Safety-First Culture:

  • Capabilities research redirected to safety
  • Racing mentality abandoned
  • Precautionary principle applied
  • Long-term thinking valued
  • Wisdom prioritized over speed

International AI Safety Agency:

  • Real power and enforcement capability
  • Democratic accountability
  • Technical expertise
  • Independent from national governments
  • Legitimate authority

Democratic Oversight:

  • Public input on AI development decisions
  • Transparent decision-making
  • Representation of affected parties
  • Not just technocratic control
  • Values debate central

Economic Adaptation:

  • Managing costs of pause
  • Retraining and support programs
  • Alternative economic development
  • Addressing inequality
  • Ensuring broad benefit when AI eventually deployed

Branch Point 1: The Galvanizing Incident (2025-2026)

Section titled “Branch Point 1: The Galvanizing Incident (2025-2026)”

What Happened: Major AI incident occurred - severe enough to galvanize action, not catastrophic enough to end civilization.

Alternative Paths:

  • No Incident: Risks remain abstract, pause impossible → Other scenarios
  • Catastrophic Incident: Destroys civilization → Catastrophe scenario
  • Actual Path: Goldilocks incident creates political will → Enables pause

Why This Mattered: Without concrete crisis, abstract arguments for pause wouldn’t overcome economic pressure. But incident had to be survivable.

Branch Point 2: International Coordination (2027-2028)

Section titled “Branch Point 2: International Coordination (2027-2028)”

What Happened: All major AI-developing nations agreed to binding treaty with enforcement.

Alternative Paths:

  • Coordination Fails: Racing continues → Multipolar or Catastrophe scenarios
  • Actual Path: Successful treaty → Enables pause

Why This Mattered: Without coordination, unilateral pause impossible - would just cede advantage to others. Mutual pause only option.

What Happened: First years of pause survived without major defection or breakdown.

Alternative Paths:

  • Early Defection: One actor violates, others forced to follow → Back to racing
  • Actual Path: Pause holds → Can continue productive work

Why This Mattered: First years critical. If pause survived initial period, could become established norm.

Branch Point 4: Productive Use of Time (2030-2035)

Section titled “Branch Point 4: Productive Use of Time (2030-2035)”

What Happened: Pause used for genuine safety research and governance building, not wasted.

Alternative Paths:

  • Wasted Time: No progress on safety, pause eventually collapses
  • Actual Path: Significant alignment progress → Justifies pause, enables eventual safe restart

Why This Mattered: Pause only worthwhile if used productively. Progress on safety essential to maintain public support and justify continued pause.

Branch Point 5: Restart Decision (2036-2038)

Section titled “Branch Point 5: Restart Decision (2036-2038)”

What Happened: Either cautious restart with safety measures, or continued pause until safety assured.

Alternative Paths:

  • Premature Restart: Restart before safety solved → Could lead to Catastrophe
  • Permanent Stagnation: Pause continues indefinitely, no progress
  • Actual Path: Restart when safety adequate, or continue pause if needed

Why This Mattered: Restart decision determines whether pause leads to Aligned AGI or needs to continue.

Goldilocks Crisis:

  • Severe enough to galvanize action
  • Not catastrophic enough to end civilization
  • Clear that it was caused by AI
  • Obvious that worse could happen
  • Timing: before AI too entrenched

Public Mobilization:

  • Crisis understood by general public
  • Media coverage accurate and urgent
  • Political leaders respond to pressure
  • Cross-ideological coalition possible
  • Economic costs seem acceptable

International Cooperation Possible:

  • Verification technology exists
  • Mutual distrust can be overcome
  • Economic incentives can align
  • Political leaders willing to coordinate
  • Enforcement mechanisms feasible

US-China Agreement:

  • Both see catastrophe risk as greater than competition risk
  • Verification trusted
  • Face-saving structure allows agreement
  • Domestic political support in both
  • Mutual compliance credible

Defection Can Be Prevented:

  • Compute governance works
  • Monitoring effective
  • Enforcement has teeth
  • Criminal penalties deter
  • Whistleblower protection enables detection

Pause Technologically Feasible:

  • Can identify and monitor large training runs
  • Compute governance infrastructure works
  • Can’t develop AGI without detectable compute
  • No secret path to AGI bypassing monitoring

Compute Governance: Technical Feasibility Assessment

Section titled “Compute Governance: Technical Feasibility Assessment”

Research on compute governance identifies why compute is uniquely governable:

PropertyExplanationGovernance Implication
DetectableTraining frontier AI requires tens of thousands of advanced chips; cannot be acquired inconspicuouslyLarge training runs visible to monitoring
ExcludableAI chips are physical goods that can be controlledExport controls, licensing possible
QuantifiableChips, features, and usage can be measuredThresholds can trigger regulatory action
ConcentratedOnly 3 companies (NVIDIA, AMD, Intel) produce advanced AI chips; TSMC manufactures mostSupply chain chokepoints exist

Current Enforcement Mechanisms:

  • Executive Order 14110 requires reporting training runs above compute thresholds
  • US export controls restrict advanced chips to China
  • RAND analysis documents detection mechanisms for cloud monitoring

Known Gaps:

  • Verification depends partially on self-reporting
  • Tiered thresholds debated but not implemented
  • Algorithmic efficiency gains could reduce compute requirements
  • Privacy-preserving monitoring techniques still developing

Alignment Not Impossible:

  • Must be solvable given enough time
  • If fundamentally impossible, pause either permanent or futile
  • Progress visible during pause maintains support

Democratic Institutions Functional:

  • Can make long-term decisions
  • Public input meaningful
  • Resist industry capture
  • Maintain commitment over time

Public Support Sustainable:

  • Understanding of stakes persists
  • Economic costs acceptable
  • Alternative vision of progress available
  • Trust in institutions holds

Warning Signs We’re Entering This Scenario

Section titled “Warning Signs We’re Entering This Scenario”

We Might Be Heading Here If We See:

  • Major AI safety incident that shocks public
  • Massive increase in public concern about AI risks
  • “Pause AI” movement gaining mainstream support
  • Political candidates winning on AI safety platforms
  • Tech leaders acknowledging need to slow down
  • International AI safety negotiations getting serious
  • Compute governance proposals advancing

We’re Not on This Path If:

  • No major incidents creating urgency
  • Public remains unconcerned or enthusiastic about AI
  • Political will for pause absent
  • Economic pressure overwhelming safety concerns
  • International cooperation failing

Strong Evidence for This Scenario:

  • Binding international AI development treaty signed
  • US and China both complying with pause
  • IASA or equivalent established with real power
  • Major AI labs shut down large training runs
  • Massive increase in alignment research funding
  • Public support for pause strong and sustained
  • No major defections from treaty

We’re Diverging If:

  • Treaty negotiations fail
  • Defection from agreements
  • Public support erodes
  • Economic pressure forces restart
  • Racing resumes

Pause Is Working If:

  • International pause holding for multiple years
  • Significant progress on alignment research
  • Governance institutions functional and legitimate
  • No successful unauthorized AGI development
  • Public support maintains
  • Economic adaptation successful
  • Democratic oversight functioning

Pause Is Failing If:

  • Defection increasing
  • Public support eroding
  • No progress on alignment
  • Unauthorized development succeeding
  • Economic/political pressure to restart overwhelming
  • Governance institutions captured or failing

Building Public Understanding:

  • Accurate risk communication
  • Making abstract risks concrete
  • Building cross-ideological coalition
  • Preparing political ground for pause
  • Countering industry pressure

Technical Preparation:

  • Developing verification technologies
  • Proving compute governance feasible
  • Demonstrating monitoring capabilities
  • Building enforcement mechanisms
  • Creating alternative AI safety research agenda

Policy Groundwork:

  • Drafting treaty language
  • Building international relationships
  • Creating domestic political support
  • Developing enforcement mechanisms
  • Designing governance institutions

Alignment Research (Highest Priority):

  • Mechanistic interpretability
  • Scalable oversight
  • Robust value learning
  • Formal verification
  • Detection of deceptive alignment
  • Fundamental research without racing pressure

Governance Institution Building:

  • International oversight bodies
  • Democratic participation mechanisms
  • Ethical framework development
  • Public engagement and education
  • Regulatory capacity building

Economic Adaptation:

  • Supporting affected workers and industries
  • Alternative development paths
  • Addressing inequality
  • Building public goods
  • Preparing for eventual AI benefits

Maintaining Pause:

  • Verification and monitoring
  • Enforcement of violations
  • Whistleblower protection
  • Public communication about progress
  • Sustaining political will
  • Preventing defection

Safety Testing:

  • Rigorous evaluation protocols
  • Red-teaming and adversarial testing
  • Safety threshold establishment
  • Deployment guidelines
  • Ongoing monitoring plans

Governance Structures:

  • Decision-making processes
  • Accountability mechanisms
  • Democratic oversight
  • Benefit distribution plans
  • Risk management frameworks

Public Consultation:

  • Democratic input on AI future
  • Values debate and resolution
  • Stakeholder engagement
  • Transparency about tradeoffs
  • Building consensus on path forward

Humanity Broadly (Long-Term):

  • Catastrophe avoided
  • Time to solve alignment properly
  • Thoughtful rather than rushed AI development
  • Democratic input on AI future
  • Better outcomes than racing to catastrophe

Future Generations:

  • Not sacrificed to short-term competitive pressure
  • Inherit safer AI future
  • Values and preferences considered
  • Cosmic potential preserved

AI Safety Researchers:

  • Massive increase in resources and importance
  • Time to do research properly
  • No racing pressure forcing corners to be cut
  • Ability to solve hard problems carefully

Democratic Institutions:

  • Demonstrated ability to govern long-term risks
  • Strengthened international cooperation
  • Public trust potentially increased
  • Meaningful oversight of transformative technology

Workers:

  • More time to adapt to AI
  • Better retraining and support
  • Gradual rather than shocking transition
  • Voice in how AI deployed

Losers (Relative to Uncontrolled Development)

Section titled “Losers (Relative to Uncontrolled Development)”

Tech Companies:

  • Can’t exploit first-mover advantage
  • Profits delayed significantly
  • Must accept international oversight
  • Racing advantage lost

Some Researchers:

  • Can’t work on cutting-edge capabilities
  • Research freedom restricted
  • Must redirect to safety work
  • Slower publication and progress

Nations Seeking Advantage:

  • Can’t use AI for competitive edge
  • Must accept international constraints
  • Strategic autonomy reduced
  • Must cooperate with rivals

Aggressive Accelerationists:

  • Vision of rapid AI-enabled transformation delayed
  • Can’t experiment freely
  • Must accept precautionary approach
  • Forced to slow down

Humanity (Short-Term):

  • Economic benefits of AI delayed
  • Opportunity costs of pause
  • But catastrophic risks avoided
  • Tradeoff of certain costs for uncertain but massive benefits

AI-Enabled Solutions to Other Problems:

  • Climate change solutions delayed
  • Medical breakthroughs delayed
  • Scientific progress slowed
  • But existential risk from AI reduced
  • Difficult tradeoff

Current Generation:

  • May not see benefits of AI in lifetime
  • Pay costs of pause
  • Future generations get benefits
  • Intergenerational justice question

Key Questions

Will we get a Goldilocks incident - severe enough to galvanize action, not catastrophic?
Can US and China overcome geopolitical tensions to coordinate on AI pause?
Is compute governance technologically feasible and enforceable?
Will public support for pause be sustainable over years or decades?
Is alignment solvable given sufficient time, or fundamentally impossible?
Can democratic institutions make and sustain long-term decisions?
Will economic costs of pause be politically acceptable?
Can we prevent defection from international agreements?

Will Galvanizing Crisis Occur?

  • Need incident severe enough to create will
  • But not so severe it causes catastrophe
  • Timing matters - must come before AI too entrenched
  • No guarantee crisis at right severity and timing

Can We Coordinate?

  • Hardest part of this scenario
  • Unprecedented international cooperation required
  • US-China agreement especially difficult
  • Economic pressure to defect very strong
  • Historical precedents not encouraging

Is Compute Governance Feasible?

  • Can we detect large training runs?
  • Can we prevent unauthorized development?
  • Are there secret paths to AGI?
  • Will monitoring technology work?
  • Can enforcement be effective?

Will Pause Hold?

  • Can political will sustain for years?
  • Will public support erode?
  • Can we prevent defection?
  • Will economic costs remain acceptable?
  • Can we avoid pause fatigue?

Will We Use Time Wisely?

  • Will alignment research actually progress?
  • Can we solve hard problems?
  • Will we build good governance?
  • Or will time be wasted?

Historical Precedents:

  • Montreal Protocol showed environmental coordination possible
  • Nuclear test ban treaty showed arms control feasible
  • COVID showed rapid global response possible in crisis
  • Sometimes humans coordinate when stakes clear

Current Trends:

  • Growing awareness of AI risks
  • Pause movement gaining traction
  • Some tech leaders supporting caution
  • Compute governance proposals developing
  • International AI safety cooperation beginning

Logic Compelling:

  • Alternative to pause is risking catastrophe
  • Economic costs of pause less than catastrophe
  • Verification technically feasible
  • If stakes clear enough, coordination possible

Coordination Very Hard:

  • US-China tensions high
  • Economic pressure enormous
  • Free-rider incentives strong
  • Verification trust difficult
  • Historical coordination record poor

May Not Get Right Crisis:

  • Crisis might not occur
  • Or might be catastrophic not galvanizing
  • Or timing wrong
  • Or insufficient to overcome economic pressure

Political Will Insufficient:

  • Short-term thinking dominates
  • Industry pressure overwhelming
  • Public concern insufficient
  • Democratic institutions too slow
  • Can’t sustain long-term commitments

Technical Challenges:

  • Compute governance may not work
  • Secret development paths may exist
  • Can’t verify compliance adequately
  • Enforcement may be impossible

Pause May Not Help:

  • If alignment impossible, pause futile
  • Or if progress requires capabilities research
  • Time might be wasted
  • Economic costs undermine support

Similar Coordination Rare:

  • Few historical examples of this level of global coordination
  • Usually only after catastrophe, not before
  • Economic incentives usually overwhelm safety
  • Tragedy of the commons usually wins

But Stakes Unprecedented:

  • Existential risk different from other challenges
  • Extinction risk might motivate different behavior
  • Rational to coordinate if catastrophe probable
  • Question is whether rationality can overcome incentives

From Slow Takeoff Muddle:

  • If incident severe enough during muddling
  • If partial coordination succeeds enough to pause
  • If public pressure overwhelms economic incentives

From Multipolar Competition:

  • If crisis convinces all actors competition too dangerous
  • If near-catastrophe from competition creates will to pause
  • If mutual destruction logic enables cooperation

Not Likely From:

  • Aligned AGI (already succeeded)
  • Misaligned Catastrophe (too late)

To Aligned AGI:

  • If pause used productively
  • If alignment solved during pause
  • If cautious restart succeeds
  • Best possible outcome of pause

To Misaligned Catastrophe:

  • If premature restart before safety solved
  • If unauthorized development succeeds
  • If pause breaks down into racing

To Slow Takeoff Muddle:

  • If pause partially erodes
  • If restart chaotic and uncoordinated
  • If governance weakens over time

To Multipolar Competition:

  • If pause breaks down into fragmented competition
  • If some actors defect and others follow
  • If coordination fails

Comparison of Technology Governance Precedents

Section titled “Comparison of Technology Governance Precedents”
PrecedentOutcomeTime to AgreementVerificationKey Success/Failure Factor
Montreal Protocol (1987)Success - 99% CFC reduction2 years after Vienna ConventionIndustry self-reporting + atmospheric monitoringClear alternatives existed; concentrated industry
Nuclear Test Ban (1963)Partial success18 years of negotiationsSeismic monitoring; national technical meansUS-Soviet bilateral; limited scope
Biological Weapons Convention (1972)Weak - no verification3 years negotiationNone (verification protocol rejected 2001)No enforcement; state cheating occurred
Climate (Paris 2015)Insufficient23 years (Kyoto to Paris)National reporting; peer reviewNon-binding; free-rider problem
Human Cloning Ban (2005)Informal successUN declaration (non-binding)None formal60+ countries unilateral bans; low demand
Gain-of-Function (2014-2024)ContestedOngoingFunding-based; voluntaryOnly government-funded research covered

Similarities:

  • Global coordination on existential risk
  • Economic costs accepted
  • Industry initially opposed, complied
  • Verification and enforcement worked
  • Problem largely solved

Differences:

  • Ozone depletion clearer and simpler than AI risk
  • Alternatives to CFCs existed
  • No competitive advantage to defection
  • Much easier problem than AI pause

Lessons: Global coordination possible when threat clear and alternatives exist.

Similarities:

  • Arms control between rivals
  • Verification trust issues
  • Mutual benefit from limiting race
  • Required overcoming suspicion

Differences:

  • AI dual-use nature harder than nuclear
  • More actors in AI than nuclear powers
  • Verification easier for nuclear than AI
  • Nuclear didn’t have huge economic benefits

Lessons: Can coordinate with rivals on existential risk, but imperfectly.

Why Climate Coordination Failed:

  • Diffuse rather than concentrated threat
  • Long timescales reduced urgency
  • Free-rider problems severe
  • Economic costs concentrated, benefits diffuse
  • No single galvanizing crisis

Differences for AI:

  • Threat more concentrated and immediate
  • Might get galvanizing crisis
  • Benefits of pause (avoiding catastrophe) clear
  • Verification potentially easier

Lessons: Coordination on long-term risks very difficult. Need crisis to galvanize action.

Initial Response:

  • Rapid global action possible in crisis
  • Massive economic costs accepted
  • International cooperation on some aspects
  • But also competition and nationalism

Lessons:

  • Crisis can enable rapid response
  • But coordination fragile
  • Economic pressure eventually overwhelms caution
  • Fatigue sets in over time
  • Need to use time wisely
PreconditionIndependent ProbabilityCumulative ProbabilityKey Dependencies
Goldilocks crisis occurs20-40%20-40%Must be severe enough to galvanize but not catastrophic
Crisis attributed to AI60-80%12-32%Clear causal chain; public understanding
Public mobilization sufficient40-60%5-19%Cross-ideological coalition; sustained attention
US-China coordination achieved15-30%1-6%Geopolitical tensions overcome; mutual verification
Treaty negotiated and signed50-70%0.4-4%Enforcement mechanisms; face-saving structure
Pause holds 3+ years40-60%0.2-2.4%No major defection; economic adaptation
Productive use of time60-80%0.1-1.9%Alignment progress visible; governance built
Safe restart or extended pause70-90%0.1-1.7%Safety thresholds met or exceeded

Note: These are rough estimates. The 5-15% headline figure accounts for correlation between preconditions (success in one increases likelihood of others) and potential alternative pathways not modeled above.

📊
SourceEstimateDate
Baseline estimate5-15%
Optimists15-30%
Pessimists1-5%
Median view8-12%

Requires Many Unlikely Things:

  • Goldilocks crisis at right time
  • US-China coordination despite tensions
  • Compute governance working
  • Sustained political will over years
  • Public support maintaining
  • Productive use of time
  • No successful defection
  • Eventual safe restart or acceptable permanent pause

Each Step Individually Unlikely:

  • Getting right crisis: ~20-30%?
  • Achieving coordination: ~20-30%?
  • Pause holding 3+ years: ~30-50%?
  • Productive use of time: ~50-70%?
  • Combined probability: Very low

But Not Impossible:

  • Stakes are existential
  • Rationality might prevail
  • Historical precedents exist
  • Current trends somewhat positive
  • Crisis could change everything

Central Estimate Rationale: 5-15% reflects genuine possibility but very low likelihood. Requires too many things to go right. Higher than I’d assign to “world government by 2030” but lower than “another pandemic in next 20 years.” Possible but improbable.

Dramatically Increases Probability:

  • Major AI safety incident creating public demand for pause
  • US-China AI safety cooperation breakthrough
  • Successful demonstration of compute governance
  • Political leaders embracing AI pause
  • Pause movement going mainstream
  • Tech leaders supporting pause

Decreases Probability:

  • Continued racing despite incidents
  • US-China tensions worsening
  • Compute governance proving infeasible
  • Public unconcerned about AI risks
  • Economic pressure overwhelming safety concerns
  • Defection from cooperation increasing

Why This Scenario Matters Despite Low Probability

Section titled “Why This Scenario Matters Despite Low Probability”

Best Path If Alignment Very Hard:

  • If alignment takes decades to solve
  • Pause might be only way to avoid catastrophe
  • Better than racing to disaster

Enables Thoughtful Development:

  • Time to solve hard problems properly
  • Democratic input on AI future
  • Building proper governance
  • Avoiding irreversible mistakes

Demonstrates Human Agency:

  • Shows we can make long-term decisions
  • Proves coordination possible
  • Exercises collective wisdom
  • Chooses deliberate path

Even If Unlikely:

  • Increasing probability from 5% to 10% valuable
  • Preparing for possible pause useful
  • Building coordination infrastructure helps other scenarios
  • Creating pause option valuable even if not taken

Overlaps With Other Scenarios:

  • Work toward pause helps Aligned AGI scenario
  • Governance building useful in Muddle scenario
  • International cooperation valuable everywhere
  • Not wasted effort even if pause doesn’t happen

Counters Inevitability:

  • AI development not predetermined
  • Humans can choose different path
  • Not slaves to economic forces
  • Agency still possible

Provides Hope:

  • Not stuck with racing or catastrophe
  • Deliberate path exists
  • Coordination conceivable
  • Can choose wisdom over speed