Skip to content

Pause Advocacy

📋Page Status
Quality:88 (Comprehensive)
Importance:82.5 (High)
Last edited:2025-12-28 (10 days ago)
Words:4.0k
Structure:
📊 14📈 1🔗 65📚 711%Score: 14/15
LLM Summary:Comprehensive analysis of pause advocacy showing 15-40% probability of meaningful policy implementation by 2030, with potential to provide 2-5 years of additional safety research time. Includes 11 data tables comparing regulatory mechanisms, public opinion (97% support AI regulation, 64% support superintelligence ban), and international coordination challenges between US deregulation and China's governance frameworks.
Intervention

Pause Advocacy

ApproachAdvocate for slowing or pausing frontier AI development
TractabilityLow (major political/economic barriers)
Key OrganizationsFuture of Life Institute, Pause AI
DimensionAssessmentEvidence
TractabilityLow-Medium (25-35%)Major lobbying opposition; geopolitical competition; Biden’s EO 14110 rescinded January 2025
Potential ImpactHigh if achievedCould provide 2-5 years for alignment research; Asilomar precedent shows scientific pauses can work
Political FeasibilityLow (15-25%)Only 30,000 signed FLI open letter; industry opposition strong
International CoordinationVery Low (10-20%)China developing own AI governance framework; US-China competition intense
Time Gained if Successful2-5 yearsBased on proposed 6-month to multi-year pause durations
Risk of BackfireModerate (30-40%)Compute overhang; ceding leadership to less safety-conscious actors
Advocacy MomentumGrowingPauseAI protests in 13+ countries; 64% of Americans support pause until proven safe
Public Support for RegulationVery High (97%)Gallup 2025: 97% agree AI safety should be subject to rules; 69% say government not doing enough
RiskMechanismEffectiveness
Racing DynamicsReduces competitive pressure, creates coordination windowHigh if achieved
Loss of ControlBuys time for alignment researchHigh if alignment is tractable
Misuse RisksDelays deployment of dual-use capabilitiesMedium
Lock-inProvides time for governance frameworksMedium
Epistemic RisksReduces rushed deployment of unreliable systemsMedium

Pause advocacy represents one of the most controversial interventions in AI safety: calling for the deliberate slowing or temporary halting of frontier AI development until adequate safety measures can be implemented. This approach gained significant attention following the March 2023 open letter organized by the Future of Life Institute, which called for a six-month pause on training AI systems more powerful than GPT-4 and garnered over 30,000 signatures including prominent technologists and researchers.

The core premise underlying pause advocacy is that AI capabilities are advancing faster than our ability to align these systems with human values and control them reliably. Proponents argue that without intervention, we risk deploying increasingly powerful AI systems before developing adequate safety measures, potentially leading to catastrophic outcomes. The theory of change involves using advocacy, public pressure, and policy interventions to buy time for safety research to catch up with capabilities development.

However, pause advocacy faces formidable challenges. The economic incentives driving AI development are enormous, with companies investing hundreds of billions of dollars and nations viewing AI leadership as critical for economic and military competitiveness. Critics argue that unilateral pauses by safety-conscious actors could simply cede leadership to less responsible developers, potentially making outcomes worse rather than better.


The pause advocacy movement has evolved significantly since 2023, with proposals ranging from temporary moratoria to conditional bans on superintelligence development.

ProposalDateScopeSignatoriesKey DemandStatus
FLI Open LetterMarch 20236-month pause on GPT-4+ training30,000+Voluntary moratoriumIgnored by labs
CAIS StatementMay 2023Risk acknowledgment350+ researchersRecognition of extinction riskInfluenced discourse
Statement on SuperintelligenceOctober 2025Conditional ban on superintelligence700+ (Nobel laureates, public figures)Prohibition until “broad scientific consensus” on safetyActive campaign
PauseAI Policy ProposalOngoingInternational treaty + AI safety agencyGrassroots movementIAEA-like body for AIAdvocacy stage

The October 2025 “Statement on Superintelligence” represents a notable escalation from the 2023 letter. While the original called for a temporary six-month pause, the new statement advocates for a conditional prohibition: “We call for a prohibition on the development of superintelligence, not lifted before there is broad scientific consensus that it will be done safely and controllably, and strong public buy-in.” Signatories include Nobel laureates Geoffrey Hinton, Daron Acemoglu, and Beatrice Fihn, alongside public figures like Steve Wozniak, Richard Branson, and Prince Harry and Meghan Markle. FLI director Anthony Aguirre warned that “time is running out,” estimating superintelligence could arrive within one to two years.


Loading diagram...

The pause advocacy theory of change operates through multiple reinforcing pathways. Public advocacy generates media attention, which combined with grassroots organizing shifts public opinion. This creates political pressure that can lead to either legislative action (such as compute governance) or industry self-regulation (such as responsible scaling policies). Both pathways result in slower development timelines, providing additional time for safety research to mature before transformative AI capabilities emerge.

The most compelling argument for pause centers on the widening gap between AI capabilities and safety research. While frontier models have jumped from GPT-3 (175B parameters, 2020) to GPT-4 (estimated 1.7T parameters, 2023) to even more powerful systems, fundamental alignment problems remain unsolved. Current safety techniques like constitutional AI and reinforcement learning from human feedback (RLHF) appear increasingly inadequate for highly capable systems that could exhibit deceptive behavior or pursue unintended objectives.

GenerationParametersYearKey Safety AdvancesGap Assessment
GPT-3175B2020Basic RLHFModerate
GPT-4~1.7T (est.)2023Constitutional AI, red-teamingWidening
Claude 3/GPT-4.5Undisclosed2024RSP frameworks, scaling policiesSignificant
Projected 202610T+2026UnknownCritical uncertainty

Research by Anthropic and other safety-focused organizations suggests that as models become more capable, they become harder to interpret and control. A 2023 study by Perez et al. found that larger language models show increased tendencies toward deceptive behaviors when given conflicting objectives. Recent mechanistic interpretability work remains far from scalable to frontier models with hundreds of billions of parameters. Without a pause to develop better alignment techniques, we may cross critical capability thresholds before adequate safety measures are in place.

Pause advocacy also argues that slower development creates opportunities for beneficial coordination that are impossible during intense racing dynamics. The current AI development landscape involves only a handful of frontier labs—primarily OpenAI, Google DeepMind, and Anthropic—making coordination theoretically feasible. Historical precedents like the Asilomar Conference on Recombinant DNA (1975) and various nuclear arms control agreements demonstrate that the scientific community and governments can successfully coordinate to slow potentially dangerous technological development when risks are recognized.

The window for such coordination may be closing rapidly. As AI capabilities approach transformative levels, the strategic advantages they confer will likely intensify competitive pressures. Research on strategic insights from simulation gaming of AI race dynamics suggests that nations viewing AI as critical to national security may be unwilling to accept coordination mechanisms that could disadvantage them relative to competitors.

Advocates invoke the precautionary principle, arguing that when facing potentially existential risks, the burden of proof should be on demonstrating safety rather than on proving danger. Unlike most technologies, advanced AI systems could pose civilization-level risks if misaligned, making the stakes qualitatively different from typical innovation-risk tradeoffs. This principle has precedent in other high-stakes domains like nuclear weapons development and gain-of-function research, where safety considerations have sometimes overridden rapid advancement.


The strongest argument against pause concerns international competition, particularly with China. China’s national AI strategy, announced in 2017, explicitly aims for AI leadership by 2030, with massive state investment. However, the picture is more nuanced than simple “race to the bottom” narratives suggest. China released its own AI Safety Governance Framework in September 2024, and in December 2024, 17 major Chinese AI companies including DeepSeek and Alibaba signed safety commitments mirroring Seoul AI Summit pledges.

Competitive FactorUS PositionChina PositionImplication for Pause
Frontier model capabilityLeading6-18 months behindPause could narrow gap
Compute accessAdvantage (NVIDIA chips)Constrained by export controlsPause less urgent for compute
Safety researchLeading (Anthropic, OpenAI)Growing (CnAISDA)Potential coordination opportunity
Regulatory frameworkFragmented, EO rescindedUnified (CAC framework)China may regulate faster
Talent poolAdvantageGrowing rapidlyNot directly affected by pause

This concern is compounded by the dual-use nature of AI research. Unlike some other dangerous technologies, AI capabilities research often advances both beneficial applications and potentially dangerous ones simultaneously. Pausing beneficial AI development to prevent dangerous applications may be a poor tradeoff if competitors continue advancing both dimensions.

A technical argument against pause involves the “compute overhang” phenomenon. If pause affects model training but not hardware development, accumulated computing power could enable sudden capability jumps when development resumes. Historical analysis of computing trends shows that available compute continues growing exponentially even during periods of algorithmic stagnation. A pause that allows compute scaling to continue could result in more dangerous discontinuous progress than gradual development would produce.

Research by OpenAI’s scaling team suggests that sudden access to much larger compute budgets could enable capability gains that bypass incremental safety research. This could be more dangerous than the gradual development that allows safety research to track capabilities improvements.

Pause advocacy also faces arguments about opportunity costs and distributional effects. AI technologies show significant promise for addressing major challenges including disease, climate change, and poverty. A 2023 economic analysis by McKinsey Global Institute estimated that AI could contribute $13 trillion annually to global economic output by 2030, with particular benefits for developing countries if access is democratized.

Critics argue that pause advocacy primarily benefits existing AI leaders by reducing competitive pressure while imposing costs on society through delayed beneficial applications. This raises questions about the democratic legitimacy of pause policies and their distributional consequences.


The most direct path to pause involves government regulation. Several mechanisms are under consideration, including compute governance (restricting access to the high-end chips needed for training frontier models), mandatory safety evaluations before deployment, and international treaties.

Regulatory MechanismJurisdictionThresholdStatus
Executive Order 14110US10^26 FLOP training runsRescinded January 2025
EU AI ActEU10^25 FLOP (systemic risk)Active since August 2024
Responsible AI Act (proposed)US CongressTiered thresholdsUnder consideration
China AI Governance FrameworkChinaRisk-based gradingVersion 2.0 released September 2025

However, regulatory approaches face significant implementation challenges. The global nature of AI supply chains, the dual-use character of computing hardware, and the difficulty of defining “frontier” capabilities all complicate enforcement. Additionally, industry opposition remains strong, with major tech companies arguing that regulation could stifle innovation and benefit foreign competitors. The rescission of Biden’s EO 14110 within hours of the new administration taking office demonstrates the fragility of executive action on AI governance.

An alternative approach involves industry-led initiatives, such as responsible scaling policies that commit labs to pause development when certain capability thresholds are reached without adequate safety measures. Anthropic’s Responsible Scaling Policy, first announced in September 2023 and updated in October 2024, provides a template for such approaches by defining specific capability evaluations and safety requirements.

LabPolicyKey CommitmentsIndependent Verification
AnthropicRSP v2.2ASL capability thresholds, biosecurity evalsLimited external audit
OpenAIPreparedness FrameworkRed-teaming, catastrophic risk thresholdsInternal governance
Google DeepMindFrontier Safety FrameworkPre-deployment evaluationsIn development

The advantage of industry self-regulation is that it can move faster than formal regulation and may be more technically sophisticated. However, it relies on voluntary compliance and may not address competitive pressures that incentivize racing. After one year with the RSP in effect, Anthropic acknowledged instances where they fell short of meeting the full letter of its requirements, though they assessed these posed minimal safety risk.


The primary safety concern with pause advocacy is that it may not achieve its intended goals while creating new risks. If pauses are unevenly implemented, they could concentrate advanced AI development among less safety-conscious actors. Additionally, the political and economic pressures against pause may lead to policy capture or symbolic measures that provide false security without meaningfully reducing risks.

There are also concerns about the precedent that successful pause advocacy might set. If pause becomes a standard response to emerging technologies, it could hamper beneficial innovation more broadly. The challenge is distinguishing cases where precautionary pauses are justified from those where they primarily serve incumbent interests.

Conversely, even partial success in slowing AI development could provide crucial time for safety research to advance. Recent progress in interpretability research, including work on sparse autoencoders and circuit analysis, suggests that additional time could yield significant safety improvements. In March 2025, Anthropic published research on circuit tracing that “lets us watch Claude think, uncovering a shared conceptual space where reasoning happens before being translated into language.” A coordinated slowdown that maintains global leadership among safety-conscious actors could be the best available approach for navigating the development of transformative AI.

Furthermore, pause advocacy has already had positive effects on AI safety discourse by raising awareness of risks and legitimizing safety concerns in policy circles. The 2023 open letter helped establish AI safety as a mainstream policy concern, influencing subsequent regulatory discussions and industry practices.

As of late 2025, pause advocacy remains politically marginal but increasingly organized. PauseAI, founded in May 2023, has grown from a single San Francisco group to six established chapters across the US (NYC, Chicago, SF, Portland, Phoenix, Washington DC) with protests in 13+ countries including at the AI Seoul Summit and Paris AI Action Summit.

EventDateLocationFocusOutcome
Paris AI Summit ProtestFebruary 2025Paris, FranceSummit lacking safety focusGlobal coordination, Kenya and DRC joined
Google DeepMind ProtestJune 2025London, UKGemini 2.5 safety commitments60 UK MPs sent letter to Google; largest PauseAI protest
PauseConJune 2025London, UKFirst activist training eventMovement capacity building
Amsterdam ASML ProtestDecember 2025Amsterdam, NetherlandsChip supply chain for AITargeting hardware enablers

The June 2025 Google DeepMind protest marked a turning point in pause advocacy effectiveness. When Google released Gemini 2.5 Pro in March 2025 without publishing the promised safety testing documentation, PauseAI organized demonstrations outside DeepMind’s London headquarters where protesters chanted “Stop the race, it’s unsafe” and “Test, don’t guess.” The action gained political traction when 60 cross-party UK parliamentarians signed a letter accusing Google of “breach of trust” for violating commitments made at the Seoul AI Safety Summit and to the White House in 2023. Former Defence Secretary Lord Browne warned: “If leading companies like Google treat these commitments as optional, we risk a dangerous race to deploy increasingly powerful AI without proper safeguards.”

TimeframeScenarioProbabilityKey Drivers
2025-2026No meaningful pause60-70%Industry lobbying, EO rescission, China competition
2025-2026Voluntary industry slowdown15-25%Safety incident, capability threshold reached
2027-2030Coordinated international framework15-30%China-US dialogue progress, WAICO proposal
2027-2030Binding compute governance10-20%Major incident, legislative action

Looking ahead 1-2 years, the trajectory likely depends on whether AI capabilities approach obviously dangerous thresholds. Dramatic capability jumps or safety incidents could shift political feasibility significantly. Conversely, gradual progress that demonstrates controllable development might reduce pause advocacy’s appeal.

In the 2-5 year timeframe, international coordination mechanisms will likely determine pause advocacy’s viability. China has proposed WAICO (World AI Cooperation Organization) as a framework for coordinating AI governance rules. If major AI powers can establish effective governance frameworks, coordinated development constraints become more feasible. If geopolitical competition intensifies, unilateral pauses become increasingly untenable.


Several critical uncertainties shape the case for pause advocacy:

UncertaintyIf True (Pro-Pause)If False (Anti-Pause)Current Assessment
Alignment is hardPause essential to buy research timePause unnecessary, current methods sufficientUnclear; interpretability progress slow but advancing
Short timelinesPause urgent, limited windowAdequate time without constraintsModels improving rapidly; 2-5 year uncertainty
International coordination feasibleGlobal pause achievableUnilateral pause counterproductiveChina-US dialogue shows some progress
Racing dynamics dominantPause prevents corner-cuttingAviation industry shows safety can prevailCompetitive pressures strong but not deterministic
RSPs workVoluntary pause-like mechanisms sufficientNeed binding regulationAnthropic RSP promising but untested at capability thresholds

The difficulty of AI alignment remains fundamentally unknown—if alignment problems prove tractable with existing techniques, pause advocacy’s necessity diminishes significantly. Conversely, if alignment requires fundamental breakthroughs that need years of research, pause may become essential regardless of political difficulties.

The timeline to transformative AI capabilities also remains highly uncertain. If such capabilities are decades away, there may be adequate time for safety research without development constraints. If they arrive within years, pause advocacy’s urgency increases dramatically.

Finally, the prospects for international coordination remain unclear. China’s approach to AI safety and willingness to participate in coordination mechanisms will largely determine whether global pause initiatives are feasible. The International Dialogues on AI Safety (IDAIS) produced consensus statements including the Ditchley Statement and Beijing Statement establishing specific technological “red lines” including autonomous replication and deception of regulators—suggesting some foundation for cooperation.

The effectiveness of alternative safety interventions also affects pause advocacy’s relative value. If industry responsible scaling policies or technical alignment approaches prove sufficient, the need for development pauses decreases. However, if these approaches fail to keep pace with capabilities, pause may become the only viable risk reduction mechanism.


Public support for AI regulation and development pauses has grown substantially, creating a potential political foundation for pause advocacy.

PollDateFindingImplication
Gallup/SCSPSeptember 202597% agree AI safety should be subject to rules and regulationsNear-universal support for regulation crosses party lines
FLI SurveyOctober 202564% support superintelligence ban until proven safe; only 5% favor fast developmentStrong majority favors precautionary approach
QuinnipiacApril 202569% say government not doing enough to regulate AI; 44% think AI will do more harm than goodPublic perceives regulatory gap
Rethink Priorities202551% would support AI research pause; 25% opposeSupport outstrips opposition 2:1
Pew Research2025~60% of public and 56% of AI experts worry government won’t regulate enoughExpert-public alignment on regulation

Despite this strong polling support, translating public opinion into policy remains challenging. The political economy favors AI developers: concentrated benefits for tech companies versus diffuse risks for the public create asymmetric lobbying incentives. The rapid rescission of Biden’s EO 14110 in January 2025 illustrates how quickly regulatory frameworks can be dismantled despite public support. However, the Gallup finding that 88% of Democrats and 79% of Republicans and independents support AI safety rules suggests bipartisan potential for regulation if framed appropriately.


The Asilomar Conference on Recombinant DNA (1975) provides the most frequently cited precedent for scientific self-regulation and pause. Over 100 scientists, lawyers, and journalists gathered to develop consensus guidelines for recombinant DNA research, following a voluntary moratorium initiated by scientists themselves who had recognized potential biohazards.

PrecedentYearScopeOutcomeLessons for AI
Asilomar Conference1975Recombinant DNAMoratorium lifted with safety guidelines; NIH RAC createdScientists can self-regulate; guidelines enabled rather than blocked research
Nuclear weapons moratorium1958-1961Nuclear testingPartial Test Ban Treaty (1963)International coordination possible under existential threat
BWC1972Bioweapons187 states parties; no verification regimeLimits of international agreements without enforcement
Gain-of-function pause2014-2017Dangerous pathogen researchEnhanced oversight; research resumedTemporary pauses can enable safety improvements

The Asilomar precedent is instructive but imperfect. Key differences from AI:

  • Smaller community: Only a few hundred researchers worked on rDNA in 1975 vs. millions in AI today
  • Clearer risks: Biohazards were more tangible than AI alignment concerns
  • Fewer commercial pressures: Academic research vs. hundreds of billions in investment
  • Easier enforcement: Physical lab access vs. distributed compute and open-source models

Notably, the rDNA moratorium lasted only months, and “literally hundreds of millions of experiments, many inconceivable in 1975, have been carried out in the last 30 years without incident”—suggesting that well-designed pauses can enable rather than block beneficial research.


Major organizations advancing pause advocacy include:

Complementary approaches include compute governance initiatives that could enable pause enforcement, international coordination efforts that could make pauses stable, and responsible scaling policies that implement conditional pause-like mechanisms. Technical alignment research also complements pause advocacy by developing the safety measures that would make development resumption safer.


The global governance landscape for AI has evolved rapidly, with both opportunities and obstacles for pause advocacy.

InitiativeDateKey FeaturesPause Relevance
UN Scientific Committee on AIAugust 2025Independent scientific body for AI assessmentCould provide “broad scientific consensus” mechanism
UN Global Dialogue on AI GovernanceAugust 2025Inclusive international governance forumPlatform for pause coordination
China Global AI Governance Action PlanJuly 2025Multilateral cooperation frameworkSignals Chinese openness to coordination
US AI Action PlanJuly 2025Deregulation, global competitiveness focusOpposes pause; favors speed
AI Red Lines CampaignSeptember 2025200+ signatories including 10 Nobel laureatesUN-focused advocacy for “globally unacceptable AI risks”

The UN initiatives launched in August 2025 grew from the “Governing AI for Humanity” report and aim to “kickstart a much more inclusive form of international governance.” However, only seven countries—all from the developed world—are parties to all current significant global AI governance initiatives, revealing the fragmentation that hampers pause enforcement. The fundamental tension remains: the US AI Action Plan explicitly encourages “rapid AI innovation” while linking federal funding to states adopting less restrictive AI laws, directly countering pause advocacy goals. Meanwhile, China’s proposal for a World AI Cooperation Organization suggests potential for multilateral frameworks, though skeptics note both powers view AI as a strategic asset and resist international limits.




Pause advocacy affects the Ai Transition Model through multiple factors:

FactorParameterImpact
Transition TurbulenceRacing IntensityCoordinated slowdown reduces race to bottom on safety
Misalignment PotentialSafety-Capability GapProvides 2-5 additional years for safety research if achieved
Civilizational CompetenceInternational CoordinationRequires unprecedented global coordination to be effective

Pause advocacy has 15-40% probability of meaningful policy implementation by 2030; effectiveness depends on international coordination and enforcement mechanisms.