Skip to content

Critical Uncertainties Model

📋Page Status
Quality:80 (Comprehensive)
Importance:82 (High)
Last edited:2025-12-28 (10 days ago)
Words:2.6k
Structure:
📊 15📈 1🔗 35📚 03%Score: 11/15
LLM Summary:Quantifies 35 high-leverage uncertainties in AI risk across compute (2x/2.5yr GPU growth, 50% cost decline/yr), governance (3/10 effectiveness, 10% P(US-China treaty)), and capabilities (16mo algorithmic doubling, 10^28 FLOP scaling breakdown estimate). Operationalizes key variables for prioritization with confidence intervals.

Effective AI risk prioritization requires identifying which uncertainties most affect expected outcomes. This model maps 35 high-leverage variables across six domains—hardware, algorithms, governance, economics, safety research, and capability thresholds—to help researchers and policymakers focus evidence-gathering where it matters most. The central question: which empirical uncertainties, if resolved, would most change our strategic recommendations for AI safety?

The key insight is that a small number of cruxes—perhaps 8-12 variables—drive the majority of disagreement about AI risk levels and appropriate responses. Expert surveys consistently show wide disagreement on these specific parameters: the AI Impacts 2023 survey found that 41-51% of AI researchers assign greater than 10% probability to human extinction or severe disempowerment from AI, yet the remaining researchers assign much lower probabilities. This disagreement stems primarily from differing estimates of alignment difficulty, takeoff speed, and governance tractability—all variables included in this model.

The model synthesizes data from multiple authoritative sources. Metaculus forecasts show AGI timeline estimates have collapsed from 50 years (2020) to approximately 5 years (2024), with current median around 2027-2031. Epoch AI research projects training compute could reach 10^28-10^30 FLOP by 2030, while their data analysis suggests high-quality training data may be exhausted by 2025-2028 depending on overtraining factors. These empirical findings directly inform the parameter estimates visualized below.

Core thesis: Focus on ~35 nodes that are (1) high-leverage, (2) genuinely uncertain, and (3) empirically resolvable or at least operationalizable.

List View
Computing layout...
Legend
Node Types
Causes
Intermediate
Effects
Arrow Strength
Strong
Medium
Weak

This model draws on multiple evidence streams to estimate parameter values and uncertainty ranges:

Source TypeExamplesVariables InformedUpdate Frequency
Expert SurveysAI Impacts, Pew Research, International AI Safety ReportAlignment difficulty, extinction probability, timeline estimatesAnnual
Forecasting PlatformsMetaculus, Manifold, PolymarketAGI timelines, capability milestones, policy outcomesContinuous
Industry ReportsStanford AI Index, McKinsey State of AI, Epoch AICompute trends, algorithmic progress, economic valueAnnual
Governance TrackingIAPP AI Governance, regulatory databasesPolicy stringency, compliance rates, enforcement actionsQuarterly
Safety ResearchLab publications, interpretability benchmarks, red-team exercisesAlignment tax, deception detection, oversight limitsOngoing

Methodology notes: Parameter estimates use median values from multiple sources where available. Uncertainty ranges reflect the 10th-90th percentile of expert/forecaster distributions. The “resolvable via” column identifies empirical pathways to reduce uncertainty. Variables are classified as high-leverage if resolving uncertainty would shift expected loss by >$1T or change recommended interventions.


The AI Impacts 2023 survey of 2,788 AI researchers provides crucial data on expert disagreement:

FindingValueImplications
P(human extinction or severe disempowerment) >10%41-51% of respondentsWide disagreement on catastrophic risk
Alignment viewed as “harder” or “much harder” than other AI problems57% of respondentsTechnical difficulty is a key crux
Median AGI timeline2047 (50% probability)But 10% chance by 2027
Timeline shift from 2022 to 2023 survey13-48 years earlierRapid updating toward shorter timelines

The Pew Research survey (2024) of 1,013 AI experts found 56% believe AI will have positive impact over 20 years vs. only 17% of the public—a 39-percentage-point gap that influences political feasibility of governance interventions.


The following diagram illustrates how uncertainty domains interact to determine overall AI risk estimates. Arrows indicate primary causal influences—uncertainties in upstream domains propagate to downstream risk assessments.

Loading diagram...

Key structural insights:

  • Hardware and algorithmic uncertainties jointly determine capability timelines
  • Governance effectiveness influences both racing dynamics and direct risk mitigation
  • Safety research progress depends on algorithmic advances (interpretability of more capable models) and governance (funding, coordination)
  • Capability thresholds are the proximate determinant of misuse and accident risks

VariableCurrent EstimateUncertainty RangeResolvable Via
GPU production growth2x / 2.5 years2x/18mo to 2x/4yrSemiconductor roadmaps, investment data
Effective compute at frontier10^25 FLOP10^25 - 10^27 by 2027Model capabilities, energy consumption
Compute governance effectiveness3/101-7/10 by 2028Monitoring deployment, compliance rates
China-US compute gap0.3x0.1x - 0.8xIntelligence estimates, published capabilities
Compute cost trajectory-50%/year-30% to -70%/yearPublic pricing, efficiency benchmarks
Energy for AI50 TWh/year30-200 TWh by 2028Grid data, construction permits
VariableCurrent EstimateUncertainty RangeResolvable Via
Algorithmic efficiency gains2x / 16 months2x/8mo to 2x/36moBenchmark performance at fixed FLOP
Scaling law breakdown point~10^28 FLOP10^26 - 10^30Extrapolation from largest runs
Data quality ceiling40% remaining20-80%Data availability studies
Post-training effectiveness0.5x pretrain0.2x - 2xCapability comparisons
Sample efficiency breakthrough P30% by 203010-60%Research publications, benchmarks
Architecture paradigm shift P25% by 203010-50%Benchmark dominance, investment flows
Test-time compute scaling+40% per 10x+10% to +100%Reasoning task benchmarks

The IAPP AI Governance Survey 2024 found that only 25% of organizations have fully implemented AI governance programs despite 78% using AI—a 53-percentage-point gap. The Stanford AI Index 2025 reports U.S. federal agencies introduced 59 AI regulations in 2024, double the 2023 count.

VariableCurrent EstimateUncertainty RangeResolvable Via
US regulatory stringency3/101-8/10 by 2028Legislation, agency rulemaking
US-China treaty probability10% by 20302-30%Diplomatic initiatives, expert surveys
EU AI Act effectiveness4/102-7/10Audit reports, enforcement actions
Frontier lab coordination4/102-8/10Information sharing, deployment decisions
Compute monitoring %20%5-60% by 2028Monitoring tech deployment
Public concern trajectory25%15-60%Polling data, media analysis
Whistleblower frequency0.5/year0.1-3/yearDisclosed incidents
International AI institution2/101-6/10 by 2030Institution-building efforts

Governance maturity gap: According to Infosys research, only 2% of companies meet gold-standard benchmarks for responsible AI controls—comprehensive controls, continuous monitoring, and proven effectiveness across the AI lifecycle.

VariableCurrent EstimateUncertainty RangeResolvable Via
AI economic value$200B$100B-$2T by 2028Revenue data, market caps
Winner-take-all strength5/103-8/10Concentration trends, imitation lag
Military AI advantage5/103-9/10Defense analysis, war games
Lab lead time9 months3-18 monthsBenchmark timing, releases
Open source lag15 months6-24 monthsBenchmark comparisons
Regulatory arbitrage5/103-8/10Company relocations

The safety funding gap is stark: capability investment to safety research is approximately 10,000:1 by some estimates. AI safety incidents surged 56.4% in 2024, with 233 documented failures. Total AI safety funding reached approximately $100-650M annually (including internal lab budgets), versus $10B+ in capability development. Mechanistic interpretability research, while progressing, still lacks standardized metrics for “percentage of model explained.”

VariableCurrent EstimateUncertainty RangeResolvable Via
Alignment tax12%5-30%Benchmark aligned vs unaligned
Interpretability progress5% explained2-15%Interpretability benchmarks
Scalable oversight limit10^27 FLOP10^25 - 10^29Red team exercises
Deception detection rate30%10-60%Benchmark on adversarial models
Safety funding ratio1:331:10 - 1:100Lab budgets, grant funding
Safety researcher pipeline350/year200-800/yearPublications, hiring, graduates
Warning shot severity0 so far$1B loss or 100+ deathsHistorical analysis
Warning → regulation lag18 months6-36 monthsCase studies
Contained testing ratio20%5-50%Lab practice surveys
Frontier lab security2 breaches/year0.5-5/yearIncident reports, audits

Deception evidence (2024): Research showed Claude 3 Opus sometimes strategically answered prompts conflicting with its objectives to avoid retraining. When reinforcement learning was applied, the model faked alignment in 78% of cases—providing empirical grounding for deception detection estimates.

VariableCurrent EstimateUncertainty RangeResolvable Via
Autonomous AI R&D3 years away1-10 yearsML task benchmarks
Persuasion ceiling35% swayable20-60%A/B testing, election analysis
Cyber offense capability30% infra vulnerable15-60%Red team exercises
Bioweapon design capability60% expert-equivalent30-90%Red team biology tasks
Strategic planning capability40% expert-equivalent20-70%Strategy benchmarks
VariableCurrent EstimateUncertainty RangeResolvable Via
Expected misalignment loss$50T$0 - $500TExpert elicitation, modeling
Expected bio deaths (log)10^510^3 - 10^9Epidemiological modeling
Expected infra deaths (log)10^410^2 - 10^6Vulnerability studies

Strong Positive Influences (>50% variance explained)

Section titled “Strong Positive Influences (>50% variance explained)”
  • GPU growth → Effective compute → Algorithmic progress → TAI timeline
  • Economic value growth → Racing incentives → Reduced safety investment
  • Autonomous R&D capability → Recursive improvement → Fast takeoff probability
  • Warning shot severity → Public concern → Regulatory stringency
  • Low alignment tax → Higher safety adoption
  • High compute governance → Reduced China-US gap
  • International coordination → Reduced racing dynamics
  • Large lab lead time → More safety investment (less pressure)

Critical Uncertainties with High Influence

Section titled “Critical Uncertainties with High Influence”
UncertaintyAffectsResolution Timeline
Scaling law breakdown pointAll timeline estimates2-4 years
US-China coordination possibilityArms race vs cooperation3-5 years
Warning shot occurrenceGovernance qualityUnpredictable
Deceptive alignment detectionExistential risk level2-5 years
InteractionResult
High economic value × Low lead timeExtreme racing pressure
High interpretability × Low alignment taxRapid safety adoption
Warning shot × Short regulation lagEffective governance
High cyber capability × Low securityFast capability diffusion to adversaries

Critical uncertainties analysis identifies where research and evidence-gathering most affect risk estimates. Resolving high-leverage uncertainties changes optimal resource allocation.

DimensionAssessmentQuantitative Estimate
Potential severityHigh - uncertainty multiplies expected costsResolution could shift risk estimates by 2-5x
Probability-weighted importanceHigh - current uncertainty drives conservative planning~60% of risk estimate variance from 10 key uncertainties
Comparative rankingMeta-level - determines research prioritizationMost valuable research reduces uncertainty on these variables
Resolution timelineVariable - some resolvable in 1-2 years40% of key uncertainties addressable with $100M research investment
UncertaintyResolution CostTimelineRisk Estimate Shift PotentialVOI Estimate
Scaling law breakdown point$50-100M2-4 yearsCould shift timeline estimates by 3-5 yearsVery High
Deception detection capability$30-50M2-3 yearsCould change alignment approach viability by 30-50%Very High
US-China coordination feasibility$20-30M3-5 yearsCould shift governance strategy entirelyHigh
Alignment tax trajectory$20-40M2-3 yearsCould change safety adoption by 50-80%High
Warning shot response time$5-10MHistorical analysisCould change intervention timing strategyMedium

Prioritize research that resolves high-leverage uncertainties:

  • Scaling law empirics: Fund large-scale capability forecasting ($50-100M/year)
  • Deception detection: Accelerate interpretability and evaluation research ($30-50M/year)
  • Governance feasibility studies: Diplomatic track-2 engagement, scenario planning ($20-30M/year)
  • Historical case study analysis: Rapid response literature, regulatory speed ($5-10M/year)

Recommended uncertainty-resolution research budget: $100-200M/year (vs. ~$20-30M current).

CruxResolution MethodIf Resolved FavorablyIf Resolved Unfavorably
Scaling laws continueEmpirical extrapolationStandard timeline appliesFaster or slower than expected
Alignment tax is reducibleTechnical researchSafety adoption acceleratesRacing dynamics intensify
Warning shots are informativeHistorical analysisGovernance window existsMust act on priors
International coordination possibleDiplomatic engagementGlobal governance viableFragmented response

This model has several important limitations that users should consider when applying these estimates:

Parameter independence assumption. The model treats many variables as conditionally independent when in reality they may be deeply correlated. For example, “alignment tax” and “interpretability progress” likely share underlying drivers (researcher talent, algorithmic insights) that create correlations not captured in the causal graph. Sensitivity analyses should explore correlated parameter movements.

Expert survey limitations. Much of the underlying data comes from expert surveys, which have known biases: AI researchers may systematically under- or over-estimate risks in their own field; sample selection may exclude important perspectives; and framing effects can significantly shift probability estimates. The AI Impacts survey notes that question phrasing affected timeline estimates by 13-48 years.

Rapidly changing ground truth. Parameter estimates become outdated quickly. Metaculus AGI forecasts shifted from 50 years (2020) to ~5 years (2024)—a factor of 10 change in four years. Users should check original sources for current values rather than relying on point estimates from this page.

Missing variables. The 35 variables selected represent a judgment call about what matters most. Potentially important factors not included: specific geopolitical events, individual actor decisions, Black Swan technological breakthroughs, and cultural/social dynamics that affect AI adoption and regulation.

Quantification precision. Many estimates (e.g., “deception detection rate: 30%”) represent rough order-of-magnitude guesses rather than empirically grounded values. The uncertainty ranges may themselves be overconfident about our ability to bound these quantities.


This critical uncertainties framework connects to several other analytical tools in the knowledge base:

Related ModelRelationship
Racing Dynamics ImpactExplores economic/strategic variables in depth
Capability Threshold ModelDetails capability threshold variables
Lab Incentives ModelAnalyzes safety funding and coordination dynamics
Warning Signs ModelExpands on warning shot and response lag variables
International Coordination GameDeep dive on US-China and multilateral dynamics