Skip to content

International Compute Regimes

📋Page Status
Quality:82 (Comprehensive)
Importance:78.5 (High)
Last edited:2025-12-28 (10 days ago)
Words:5.5k
Structure:
📊 7📈 1🔗 46📚 05%Score: 11/15
LLM Summary:Assesses international compute regimes for AI governance, finding only 10-25% chance of meaningful regimes by 2035 but potential for 30-60% reduction in racing dynamics if achieved. Documents first binding AI treaty (Council of Europe, Sept 2024) but notes 118 of 193 UN states absent from major initiatives, with most frameworks remaining non-binding declarations.
Policy

International Compute Regimes

ApproachCoordinate compute governance globally
StatusEarly discussions, no formal regime

International compute regimes represent a critical but challenging approach to governing AI development through multilateral coordination. Unlike unilateral measures such as export controls or domestic regulations, these regimes would establish binding international agreements that coordinate compute governance across nations—analogous to nuclear non-proliferation treaties or chemical weapons conventions. While mostly in proposal and early discussion stages, these frameworks could potentially address fundamental limitations of national-only approaches to AI governance.

The core premise underlying international compute regimes is that unilateral governance has inherent limitations in governing a global technology. Export controls can be circumvented through third countries or alternative supply chains. Domestic compute thresholds only apply within specific jurisdictions, creating regulatory arbitrage opportunities. Monitoring systems require international cooperation to achieve comprehensive coverage of global AI development. Racing dynamics between nations can undermine individual countries’ safety measures, as competitive pressures override precautionary approaches.

Current progress remains limited to high-level political declarations, with the November 2023 Bletchley Declaration representing the first international agreement to acknowledge catastrophic AI risks, signed by 28 countries including the United States, United Kingdom, European Union members, and China. However, these agreements remain entirely non-binding and lack enforcement mechanisms, verification systems, or specific commitments beyond general principles. The path from these declarative statements to meaningful binding regimes faces substantial political, technical, and institutional challenges.

DimensionAssessmentQuantitative EstimateSources
TractabilityLow10-25% chance of meaningful regime by 2035Expert synthesis
Potential ImpactVery HighCould reduce racing risk by 30-60% if achievedModeling estimates
Current StatusEarly implementation28 countries in Bletchley process; 1 binding treaty (Council of Europe, Sept 2024); 118 UN member states absent from governance initiativesUN HLAB 2024
Time HorizonLong-term5-10+ years for meaningful compute regimes; first binding treaty achieved 2024Council of Europe Framework Convention
Negotiation CostHigh$50-200M over 5-10 years for track-1 and track-2 diplomacyPolicy analysis estimates
Verification FeasibilityUncertainHardware monitoring 40-70% coverage possible; chip tracking analogous to nuclear verificationCSET research
RiskMechanismEffectiveness
Racing DynamicsCoordinated limits prevent racesHigh (if achieved)
ProliferationGlobal verification of compute accessHigh (if achieved)
Concentration of PowerMultilateral rather than unilateral controlMedium

The following table compares major international AI governance initiatives as of December 2024, showing the evolution from non-binding declarations to the first binding treaty:

Initiative/RegimeTypeParticipantsKey ProvisionsStatusEffectiveness Assessment
Bletchley Declaration (Nov 2023)Non-binding declaration28 countries including US, UK, EU, ChinaAcknowledges catastrophic AI risks; commits to international cooperation on safetyActive; expanded to Seoul Summit May 2024Low (declaratory only); laid groundwork for future cooperation
Seoul Declaration (May 2024)Non-binding declaration10 countries + EU (Leaders’ Session); 27 countries + EU (Ministerial)Adds “innovation” and “inclusivity” to safety agenda; creates International Network of AI Safety InstitutesActive; notably China did not signLow-Medium (created institutional network but non-binding)
Council of Europe Framework Convention (Sept 2024)Binding treaty46 CoE members + 11 observers; signed by 10 states including US, UK, EUFirst legally binding AI governance agreement; focuses on human rights and democratic valuesNewly signed; not yet ratifiedMedium (first binding treaty but limited participation from Global South and adversaries)
UN AI Advisory Body (Sept 2024)Recommendations39 experts from 33 countriesProposes 7 recommendations including International Scientific Panel, Policy Dialogue, AI Standards ExchangeRecommendation phase; GA to establish process 2024-2025Uncertain (comprehensive but requires political will; 118 of 193 UN states absent from major AI governance initiatives)
IAEA-like InstitutionProposalN/ARegistry of large training runs; inspections of AI facilities; capability evaluations; international safety standardsProposal stage; conceptual papers publishedHigh potential if achieved (5-15% probability by 2030)
Compute Allocation TreatyProposalN/ACaps on aggregate training compute; allocation mechanisms; hardware governance verificationProposal stageVery High potential if achieved (2-8% probability by 2035)

Key insight: As of late 2024, international AI governance has achieved its first binding treaty (Council of Europe), but meaningful compute governance regimes with verification mechanisms remain in proposal stages. The UN reports that only 7 of 193 member states participate in the seven most prominent AI governance initiatives, leaving 118 countries—mostly in the Global South—entirely absent from discussions.


The evolving international AI governance landscape involves multiple overlapping institutions and initiatives. This diagram illustrates the relationships between key actors and governance mechanisms:

Loading diagram...

Legend: Blue = Existing initiatives (mostly non-binding); Green = Binding agreements; Yellow = Proposals; Orange = Technical enablers. Solid lines = formal relationships; Dashed lines = potential integration pathways.


An international body modeled on the International Atomic Energy Agency represents one of the most developed proposals for AI governance, endorsed by UN Secretary-General António Guterres, UK Prime Minister Rishi Sunak, and OpenAI CEO Sam Altman. In spring 2023, OpenAI cofounders proposed an “IAEA for superintelligence efforts” to govern high-capability systems. Such an institution would combine monitoring, verification, and technical assistance functions, drawing on the IAEA’s decades of experience with dual-use nuclear technology—the IAEA operates with approximately $400 million annually and employs over 2,500 staff for global nuclear verification. The proposed AI equivalent would establish a registry of large training runs exceeding specified compute thresholds, conduct on-site inspections of major AI facilities, evaluate emerging dangerous capabilities, and coordinate international safety standards.

The institutional structure would likely require a governing board with representatives from major AI-developing nations, a technical secretariat with expertise in AI systems and verification methods, and inspection teams capable of assessing both hardware infrastructure and software development practices. Unlike the IAEA, which primarily monitors physical materials, an AI equivalent would need capabilities to verify software systems, algorithmic approaches, and distributed training processes. The UN High-Level Advisory Body’s September 2024 report proposed an International Scientific Panel on AI as a first step, which would produce annual reports on AI capabilities, risks, and trends, alongside ad hoc reports on emerging AI risks—though this falls short of full IAEA-like verification authority.

Critical challenges include the absence of existing mandate or institutional precedent specifically for AI, the requirement for agreement among major powers including the United States and China, the inherent difficulty of verifying AI development compared to nuclear materials, and the pervasive dual-use nature of AI technology that complicates any restriction regime. The rapid pace of AI development also contrasts sharply with the relatively stable nuclear technology that the IAEA was designed to monitor. Lawfare analysis notes that “nuclear and AI are not similar policy problems”—AI policy is loosely defined with disagreement over field boundaries and harms, and it is decentralized without the same physical bottlenecks in materials as nuclear technology.

Compute allocation frameworks would establish international agreements limiting total compute available for AI training, potentially including caps on aggregate training compute within specified time periods, mechanisms for allocating compute access among nations or approved entities, verification systems leveraging hardware governance approaches, and graduated response protocols for treaty violations. CSET research on verification mechanisms describes how AI chip accounts might be verified using methods analogous to nuclear arms control verification, with tracking systems that “identify people with the capability to train highly-capable and broad AI models, and thus many of the most risky models.”

These treaties would draw conceptual parallels to nuclear warhead limits in arms control agreements, but applied to computational resources rather than physical weapons. Compute is considered “a unique AI governance node due to the required physical space, energy demand, and the concentrated supply chain,” making it more governable than purely software-based approaches. The semiconductor supply chain’s high concentration—dominated by a few manufacturers like NVIDIA, TSMC, and ASML—could enable monitoring and governance in ways impossible for distributed software development. Implementation would require precise definitions of “compute for AI training” versus other computational uses, mechanisms for measuring and attributing distributed training across multiple data centers, protocols for handling technological advances that change compute efficiency, and enforcement measures that balance cooperation incentives with violation consequences.

The verification challenge is particularly acute for compute allocation treaties. Unlike nuclear warheads, which exist as discrete physical objects, AI training compute is inherently fungible and can be rapidly reallocated between different purposes. Distributed training across multiple jurisdictions further complicates attribution and monitoring, while the rapid improvement in compute efficiency means that fixed quantitative limits may become obsolete quickly. However, research suggests that decentralized compute access, while often considered desirable, would make verification more difficult—international verification regimes would benefit from compute being physically centralized both nationally and internationally, as “fewer countries with computing capacity to do prohibited activities means fewer countries necessary for successful coordination.”

A non-proliferation framework for AI would restrict transfers of frontier AI capabilities, analogous to nuclear non-proliferation regimes but adapted for information technology. Such a framework might establish a “nuclear club” equivalent for frontier AI development, restricting membership to countries meeting specific safety and verification standards, implementing technology transfer restrictions on advanced AI capabilities, establishing inspection and verification rights for international monitors, and defining consequences for violations including potential exclusion from AI technology cooperation.

This approach would need to address the fundamental differences between AI and nuclear proliferation. AI capabilities spread primarily through information rather than physical materials, making control more challenging but potentially more comprehensive. The rapid pace of AI advancement means that capability advantages may be temporary, reducing incentives for non-proliferation cooperation. Additionally, the enormous civilian benefits of AI technology create stronger pressures for widespread access compared to nuclear weapons technology.


The nuclear governance regime offers the most relevant historical precedent for international AI coordination, despite significant differences in the underlying technologies. The Nuclear Non-Proliferation Treaty, which entered into force in 1970, has succeeded in preventing proliferation to an estimated 15-25 additional countries that had nuclear weapons programs or capabilities, though it failed to prevent proliferation to India, Pakistan, North Korea, and Israel.

Key successes include the development of comprehensive inspection and safeguards regimes that verify civilian nuclear programs, international consensus on the dangers of nuclear proliferation that supports ongoing cooperation, decades of institution-building that created robust verification capabilities, and effective integration of civilian benefits (nuclear power) with non-proliferation objectives through the IAEA’s dual mandate.

However, the nuclear regime’s limitations are instructive for AI governance. Some proliferation occurred despite extensive international efforts, particularly by countries willing to accept international isolation. The regime relies heavily on voluntary state cooperation, which can break down under extreme political pressure. Dual-use technology challenges persist, as civilian nuclear capabilities can contribute to weapons programs. Enforcement ultimately depends on political will rather than automatic mechanisms.

The nuclear precedent suggests that international institutions can successfully coordinate governance of dangerous technologies, but that verification systems require sustained technical and political investment. The IAEA’s budget of approximately $400 million annually and staff of over 2,500 demonstrates the resource requirements for effective international monitoring. For AI, the verification challenges may be even greater due to the software-based nature of the technology and the much larger number of relevant actors.

The Chemical Weapons Convention provides another instructive precedent, particularly regarding industry cooperation and near-universal membership. The Organization for the Prohibition of Chemical Weapons has achieved near-universal state participation with 193 member countries, successful destruction of declared chemical weapons stockpiles totaling over 71,000 metric tons, effective industry verification through routine inspections of chemical facilities, and strong international norms against chemical weapons use that persist despite occasional violations.

The chemical weapons regime’s emphasis on industry cooperation offers important lessons for AI governance. Chemical companies voluntarily report dual-use production and submit to international inspections, demonstrating that private sector cooperation with international regimes is achievable. However, this cooperation depends on clear legal frameworks, predictable inspection procedures, and protection of legitimate commercial interests.

Limitations include ongoing non-compliance challenges, as demonstrated by chemical weapons use in Syria despite OPCW membership, verification difficulties with dual-use chemicals that have legitimate commercial applications, and limited enforcement mechanisms beyond diplomatic pressure and potential economic sanctions. For AI governance, these challenges may be magnified given the broader dual-use nature of AI technology and the much larger number of relevant commercial actors.

FactorNuclear/ChemicalAI
Civilian benefitsLimited (nuclear power)Enormous economic potential
Verification targetPhysical materialsSoftware, models, compute
Primary actorsStates and state-owned entitiesStates, companies, individuals
Pace of changeRelatively slowExtremely rapid advancement
Dual-use natureClear weapons applicationsPervasive dual-use applications
Geographic distributionConcentrated in few locationsGlobally distributed development

The AI Safety Summit process, initiated with the November 2023 Bletchley Declaration, represents the most significant current progress toward international AI coordination. Twenty-eight countries, including all major AI-developing nations, signed the declaration acknowledging that frontier AI capabilities could pose catastrophic risks and committing to international cooperation on AI safety. The declaration marked the first time that the United States, European Union, and China jointly acknowledged AI existential risks in an official diplomatic document. Following the Bletchley Declaration, “AI safety was firmly entrenched as a top AI policy concern,” with both the United States and United Kingdom announcing the creation of AI Safety Institutes and the formation of an advisory panel of international AI experts.

The May 2024 Seoul AI Safety Summit expanded the agenda beyond safety to include “innovation” and “inclusivity,” producing three major documents: the Seoul Declaration (signed by 10 countries and the EU), Frontier AI Safety Commitments from companies, and the Seoul Ministerial Statement (signed by 27 countries and the EU). Key outcomes included the creation of an International Network of AI Safety Institutes bringing together organizations from the UK, US, Japan, France, Germany, Italy, Singapore, South Korea, Australia, Canada, and the EU; the publication of the first International Scientific Report on the Safety of Advanced AI on May 17, 2024; and voluntary company commitments to safety testing, research sharing, and government cooperation. However, China notably did not sign the Seoul Declaration, potentially signaling “reluctance to sign on to AI governance mechanisms they view as promoting a Western-centric view of global AI governance.”

Future summits planned for France in 2025 and beyond aim to deepen cooperation, though Carnegie analysis warns that “looking ahead to France’s 2025 summit, the prospects of stronger regulation appear even dimmer, as the initiative has downplayed safety concerns in favor of discussions about AI’s potential for economic growth.” The process faces significant challenges in moving beyond high-level declarations toward binding commitments. Political differences between participants, competitive concerns among AI companies, and technical uncertainties about effective governance approaches all constrain progress.

The United Nations Secretary-General established a High-Level Advisory Body on Artificial Intelligence in October 2023, comprising 39 preeminent AI leaders from 33 countries across all regions and sectors. The body released its final report, “Governing AI for Humanity,” in September 2024, making seven key recommendations:

  1. International Scientific Panel on AI to provide impartial scientific knowledge through annual and ad hoc reports on AI capabilities, risks, and trends
  2. Policy Dialogue on AI Governance featuring intergovernmental and multi-stakeholder meetings to foster regulatory interoperability
  3. AI Standards Exchange to empower coordination and tackle global AI risks
  4. Global AI Capacity Development Network offering training, compute resources, and datasets
  5. Global AI Fund to address capacity gaps and empower local efforts toward Sustainable Development Goals
  6. Global AI Data Framework to standardize data-related definitions, principles, and stewardship
  7. AI Office within UN Secretariat to support implementation coordination

The report identified a critical governance gap: out of 193 UN member states, only 7 participate in the seven most prominent AI governance initiatives, while 118 countries (mostly in the Global South) are entirely absent. The General Assembly will appoint co-facilitators during its 79th session (September 2024 - September 2025) to establish terms of reference for the International Scientific Panel and Global Dialogue through an intergovernmental process.

Limitations include the advisory-only nature of the body’s mandate, the slow pace of UN institutional processes compared to rapid AI development, the challenge of achieving consensus among countries with different AI capabilities and strategic interests, and the absence of enforcement mechanisms even if recommendations are adopted. However, the report’s comprehensive framework provides a potential roadmap for future binding cooperation if political will develops.

Current bilateral discussions between major AI powers remain limited but show some promise for incremental progress. US-China AI dialogue faces significant constraints due to broader geopolitical tensions and export control policies, but some track-2 discussions continue through academic and research channels. Areas of potential cooperation include basic AI safety research, international standard-setting for AI evaluation, and information sharing about AI incidents and near-misses. Brookings emphasizes that “for AI governance agreements to be fully implemented, they will need the active participation and support of China and Russia as well as other relevant states. Just as during the Cold War, logic should dictate that potential adversaries be at the negotiating table in fashioning these agreements. Otherwise, democratic countries will end up in a situation where they are self-constrained but adversaries are not.”

US-European cooperation proceeds more smoothly through existing mechanisms including the Trade and Technology Council, which has established working groups on AI standards and safety evaluation. The EU’s AI Act and US executive orders on AI show some convergence on risk-based approaches to AI regulation, though significant differences remain in implementation approaches and scope. In September 2024, the first international legally binding agreement on AI governance was opened for signature in Vilnius, Lithuania—the Council of Europe Framework Convention, negotiated by 46 Council of Europe members and 11 observer states. Ten states including the United States, United Kingdom, and European Union have already signed, marking a significant milestone despite concerns that “states from the Global South who were not represented at the negotiating table may not be easily persuaded to come on board.”

Multilateral discussions within frameworks like the OECD, G7, and G20 have produced general principles for AI governance but have not yet achieved binding commitments or specific coordination mechanisms. In July 2024, the OECD Secretary General announced that the Global Partnership on AI (GPAI) is merging with OECD’s AI policy work, “opening a new chapter in global AI governance.” The challenge is moving from broad principle agreement to operational cooperation on technical governance questions.


Verification represents perhaps the greatest technical challenge for international compute regimes. Unlike nuclear materials, which exist as discrete physical objects with measurable properties, AI training processes involve software execution across distributed hardware infrastructure. Verification systems would need to monitor compute allocation to AI training versus other purposes, track distributed training across multiple data centers and jurisdictions, assess model capabilities and potential dangers, and detect attempts to circumvent monitoring through novel training approaches. Research on verification mechanisms describes methods “analogous to nuclear arms control verification,” with AI chips serving as “the main physical instantiation of AI development and deployment” due to their specialization and concentrated supply chain.

Current research suggests that hardware-based monitoring may offer the most promising verification approach. Specialized AI chips contain unique identifiers and performance characteristics that could enable tracking of compute allocation, with the goal of identifying “people with the capability to train highly-capable and broad AI models.” By tracking AI chips, verification regimes hope to identify actors with capacity to train risky models and verify they use chips safely. However, such systems would require cooperation from chip manufacturers, standardization of monitoring protocols, and protection against spoofing or circumvention attempts. The Semiconductor Industry Association has expressed cautious interest in governance cooperation but emphasizes the need for clear technical standards and minimal business disruption. The highly concentrated semiconductor supply chain—dominated by companies like NVIDIA for chip design, TSMC for manufacturing, and ASML for lithography equipment—creates natural chokepoints that could enable governance in ways impossible for distributed software.

Software-based verification faces greater challenges due to the difficulty of distinguishing AI training from other computational tasks, the potential for obfuscation through distributed or novel training methods, and the rapid evolution of training techniques that could outpace verification capabilities. Academic research on AI training detection suggests 60-80% accuracy under favorable conditions, but real-world performance may be significantly lower when actors attempt deliberate evasion. Overall verification feasibility estimates suggest 40-70% coverage is possible through hardware monitoring combined with software assessment, though this depends heavily on international cooperation and technical development.

The current geopolitical environment poses substantial obstacles to comprehensive international compute regimes. US-China relations, in particular, are characterized by strategic competition that includes AI development as a key dimension. US export controls on AI chips to China, implemented in October 2022 and expanded in 2023, create adversarial dynamics that complicate cooperation on AI safety despite shared interests in avoiding catastrophic outcomes.

Chinese perspectives on international AI governance emphasize concerns about technological sovereignty, equitable access to AI benefits, and avoiding discriminatory restrictions that preserve Western technological advantages. Chinese officials have expressed interest in AI safety cooperation but within frameworks that ensure equal participation and avoid restrictions that disadvantage Chinese AI development. The challenge is designing regimes that address legitimate security concerns while enabling genuine cooperation on shared safety challenges.

European approaches generally favor comprehensive international regulation but face constraints from both competitive concerns about disadvantaging European AI companies and political concerns about technological dependence on the United States or China. The EU’s regulatory approach through the AI Act demonstrates capability for detailed AI governance but may complicate international coordination if EU standards diverge significantly from those preferred by other major powers.

International compute regimes must address the reality that frontier AI development occurs primarily within private companies rather than government programs. Unlike nuclear weapons development, which remains largely state-controlled, AI development involves companies with global operations, international research collaboration, and complex supply chains that span multiple jurisdictions.

Leading AI companies have expressed varying degrees of support for international governance coordination. OpenAI has proposed international oversight of frontier AI development similar to nuclear governance. Google DeepMind has supported international AI safety research cooperation. However, companies also emphasize concerns about competitive disadvantages, regulatory compliance costs, and protection of proprietary information that could be compromised by extensive monitoring requirements.

The challenge is designing regimes that achieve meaningful oversight without imposing prohibitive costs on legitimate AI development or creating advantages for actors willing to operate outside international frameworks. Industry cooperation may require clear legal protections, predictable regulatory requirements, and mechanisms for protecting competitive information while enabling safety verification.


The most realistic pathway to comprehensive international compute regimes likely involves incremental steps that build trust and institutional capacity over time. An incremental approach might begin with voluntary information sharing about large training runs exceeding specified compute thresholds, mutual notification systems for significant AI capability developments, alignment on common safety evaluation frameworks and standards, joint research cooperation on AI safety and verification technologies, and coordinated incident response for AI-related accidents or near-misses.

This approach builds on existing diplomatic processes like the AI Safety Summit while gradually expanding scope and binding nature of commitments. Early confidence-building measures could include technical workshops on verification technologies, joint research projects on AI safety evaluation, information sharing about AI governance best practices, and coordination on standards development through existing international bodies.

The incremental approach faces the risk that gradual progress may be overtaken by rapid AI development that creates fait accompli situations before adequate governance is established. However, it offers the best prospect for achieving genuine cooperation among major powers and building institutional capacity for more comprehensive governance as AI risks become more evident.

Historical experience suggests that comprehensive international regimes often emerge in response to crises that demonstrate urgent need for coordination. For AI governance, crisis-driven regime formation might follow a major AI-caused harm event, a near-miss incident that demonstrates catastrophic AI risks, a geopolitical crisis involving AI capabilities, or rapidly accelerating AI development that creates obvious race dynamics with destabilizing potential.

Crisis-driven regime formation has the advantage of creating strong political incentives for rapid agreement and enabling more comprehensive frameworks than would be achievable through gradual negotiation. The chemical weapons regime’s expansion after the 1995 Tokyo subway attacks and enhanced nuclear cooperation after the 1962 Cuban Missile Crisis demonstrate how crises can catalyze institutional development.

However, crisis-driven regimes also risk poor design due to time pressure, overreaction that unnecessarily constrains beneficial AI development, and institutional frameworks that prove inadequate as circumstances change. The challenge is preparing institutional frameworks and technical capabilities in advance so that crisis-driven windows can be used effectively rather than squandered through inadequate preparation.

Technical advances in verification, monitoring, and governance technologies could enable international regimes that are currently infeasible. Hardware governance developments that enable reliable compute monitoring, interpretability advances that allow verification of model capabilities and safety properties, distributed ledger technologies that could support transparent compute allocation tracking, and automated safety evaluation systems that reduce verification costs.

The technology-enabled approach recognizes that many current obstacles to international compute regimes are technical rather than purely political. If verification becomes reliable and cost-effective, and if governance technologies can operate without imposing excessive burdens on legitimate AI development, political obstacles to international cooperation may diminish substantially.

Research priorities for enabling technologies include developing tamper-resistant hardware monitoring systems, creating standardized interfaces for AI safety evaluation, designing privacy-preserving verification protocols that protect competitive information, and building automated systems for detecting potentially dangerous AI capabilities. The timeline for these enabling technologies ranges from 2-5 years for basic hardware governance to 5-10 years for comprehensive capability verification systems.


CriterionAssessmentEvidence BaseConfidence Level
Historical precedent viabilityModerate successNPT prevented 15-25 additional nuclear states; CWC achieved near-universal membershipHigh
AI-specific technical feasibilityUncertainHardware monitoring 40-70% effective; software verification 60-80% accuracy in favorable conditionsMedium
Geopolitical cooperation prospectsLow-moderateLimited US-China dialogue; EU-US convergence on some approaches; 28 countries in Bletchley processMedium
Industry cooperation likelihoodModerateMajor companies expressed qualified support; concerns about competitive impacts and compliance costsMedium
Enforcement mechanism viabilityLow-moderateExport controls as partial backstop; diplomatic pressure primary mechanism; 30-50% estimated effectivenessLow-Medium

Establishing meaningful international compute regimes would require substantial sustained investment across multiple categories. Official diplomatic processes (track-1 diplomacy) would cost an estimated $10-50 million annually for 5-10+ years, including negotiation support, technical expertise, and institutional development. Unofficial relationship-building (track-2 diplomacy) requires $5-20 million annually for 2-5 years to establish communication channels and explore possible agreements before formal negotiations begin.

Technical verification research and development represents a critical enabling investment, requiring $20-100 million over 3-7 years to develop monitoring technologies, safety evaluation systems, and verification protocols. An international secretariat for treaty implementation would cost $10-50 million annually if established, while supporting academic and think tank research requires $5-15 million annually for regime design analysis and ongoing assessment.

The timeline for comprehensive regimes remains highly uncertain, with optimistic scenarios requiring 5-7 years under favorable political conditions and more realistic assessments suggesting 8-15 years given current geopolitical constraints. However, incremental progress through information sharing and technical cooperation could begin within 1-3 years building on existing AI Safety Summit processes.

Several fundamental uncertainties dominate assessments of international compute regime prospects. The feasibility of meaningful US-China cooperation represents the central political question, with estimates ranging from 5-30% probability by 2030 depending on broader geopolitical developments and potential crisis catalysts. AI development timelines to catastrophic risk, estimated at 5-30 years, determine urgency and available time for institutional development.

Technical verification capabilities face uncertain development timelines, with viable systems potentially available in 3-10 years depending on research progress and industry cooperation. The role of non-state actors in frontier AI development may grow to 10-40% of total capability, affecting regime coverage and enforcement requirements. Crisis-driven windows for regime formation are estimated to account for 20-50% of potential regime creation scenarios, but their timing and nature remain unpredictable.

The interaction between these uncertainties creates a wide range of possible outcomes, from comprehensive binding regimes within a decade to continued reliance on unilateral approaches and informal cooperation. Planning for international compute regimes must account for this uncertainty through flexible approaches that can adapt to different scenarios.

International compute regimes represent a high-value, long-term investment that should complement rather than replace nearer-term governance approaches. The highest-priority investments include supporting incremental cooperation through existing diplomatic channels, funding technical research on verification and monitoring capabilities, building academic and policy expertise on regime design questions, and maintaining diplomatic engagement even during periods of broader political tension.

The approach should be simultaneously ambitious in long-term vision and realistic about near-term constraints. Building the technical capabilities, institutional relationships, and political understanding necessary for comprehensive regimes requires sustained effort over many years. However, the potential benefits of successful international coordination—including prevention of destabilizing races, verification of safety commitments, and equitable global access to AI benefits—justify substantial investment despite uncertain prospects.

The most realistic pathway involves parallel progress on multiple fronts: technical development of verification capabilities, incremental diplomatic cooperation through existing channels, academic research on regime design and implementation challenges, and preparation for potential crisis-driven windows that could accelerate institutional development. This approach maximizes the probability of achieving meaningful international cooperation while building capabilities that support other governance approaches if comprehensive regimes prove infeasible.


The following table synthesizes the current state of international AI governance and prospects for meaningful compute regimes:

DimensionCurrent Status (Dec 2024)EvidenceProspects by 2030Sources
Binding Treaties1 treaty (Council of Europe Framework Convention)Signed by 10 states including US, UK, EU in Sept 202420-40% chance of compute-specific treatyCouncil of Europe
Global Participation7 of 193 UN states participate in major initiatives118 countries (mostly Global South) entirely absentLimited; likely remains Western-ledUN HLAB Report
US-China CooperationLimited track-2 dialogue onlyChina signed Bletchley but not Seoul Declaration5-30% chance of meaningful cooperationBrookings
Institutional DevelopmentInternational Network of AI Safety Institutes established11 countries participating as of May 2024 Seoul Summit40-60% chance of expansion and formalizationSeoul Declaration
Verification Technology40-70% coverage possible via hardware monitoringAI chip tracking analogous to nuclear verification; 60-80% software detection accuracy under favorable conditions50-70% chance of viable systemsCSET Research
IAEA-like InstitutionProposal stage; UN Scientific Panel recommendedUN HLAB proposed panel with annual reports; IAEA budget $400M, 2,500 staff as reference5-15% chance of full IAEA-equivalent by 2030IAEA Proposal Analysis
Industry CooperationVoluntary commitments onlyMajor labs (OpenAI, DeepMind, Anthropic) committed to safety testing and research sharing at Seoul30-50% chance of binding industry obligationsSeoul Summit Outcomes
Networked GovernanceMultiple overlapping initiatives (Bletchley, Seoul, UN, Council of Europe)“Regime complexes” characterize 21st century governance vs treaty-based 20th century70-85% chance this remains primary modelBrookings on Networks

Key Insight: International AI governance has made notable progress in 2024 with the first binding treaty and establishment of institutional networks, but comprehensive compute governance regimes with meaningful verification remain 5-10+ years away. The most likely scenario involves continued networked governance among like-minded democracies rather than universal treaties including adversaries. Brookings argues that “networked and distributed forms of AI governance will remain the singular form of international cooperation that can respond to the rapid pace at which AI is developing.”


  • Export Controls — Current unilateral restrictions that international regimes could supplement or replace
  • Compute Thresholds — Domestic regulatory triggers that require international coordination for effectiveness
  • Compute Monitoring — Technical foundation enabling verification for international agreements


International compute regimes improve the Ai Transition Model through multiple factors:

FactorParameterImpact
Civilizational CompetenceInternational CoordinationTreaties establish binding commitments on development pace and safety standards
Transition TurbulenceRacing IntensityCoordinated limits prevent destructive race to the bottom
Transition TurbulenceAI Control ConcentrationMultilateral governance distributes oversight across stakeholders

International regimes are high-impact but low-probability interventions; 10-25% chance of meaningful regimes by 2035 but potential 30-60% reduction in racing dynamics if achieved.