Skip to content

Ultimate Scenarios

Ultimate Scenarios are the intermediate pathways that connect root factors to ultimate outcomes. They describe how parameter changes lead to catastrophe (or success)—the specific mechanisms and pathways that determine what kind of future we get.

The AI Transition Model uses three main ultimate scenarios:

  1. AI Takeover — AI gains decisive control
  2. Human-Caused Catastrophe — Humans use AI for mass harm
  3. Long-term Lock-in — Permanent entrenchment of values/power

Each ultimate scenario has sub-variants that describe more specific pathways (e.g., “rapid” vs “gradual” AI takeover, “state” vs “rogue actor” catastrophe).


Loading diagram...

Color coding:

  • Red: Ultimate negative outcome (existential catastrophe)
  • Green: Ultimate trajectory measure (could be good or bad)
  • Pink: Negative ultimate scenarios (catastrophes)
  • Orange: Symmetric ultimate scenario (could entrench good or bad values)

Ultimate ScenarioDescriptionKey Root FactorsUltimate Outcomes
AI TakeoverA scenario where AI systems gain decisive control over human affairs, either through rapid capability gain or gradual accumulation of power. This could occur through misaligned goals, deceptive behavior, or humans voluntarily ceding control. The outcome depends heavily on whether the AI's values align with human flourishing.AI Capabilities ↑, Misalignment Potential ↑, Misuse Potential ↑, Transition Turbulence ↑, Civilizational Competence ↓, AI Ownership, AI Uses ↑Existential Catastrophe, Long-term Trajectory
Human-Caused CatastropheScenarios where humans deliberately use AI to cause mass harm. State actors might deploy AI-enabled weapons or surveillance; rogue actors could use AI to develop bioweapons or conduct massive cyber attacks. Unlike AI takeover, humans remain in control but use that control destructively.AI Capabilities ↑, Misalignment Potential ↑, Misuse Potential ↑, Transition Turbulence ↑, Civilizational Competence ↓, AI Ownership, AI UsesExistential Catastrophe
Long-term Lock-inPermanent entrenchment of particular power structures, values, or conditions due to AI-enabled stability. This could be positive (locking in good values) or negative (perpetuating suffering or oppression). Once locked in, these outcomes may be extremely difficult to change.AI Capabilities ↑, Misalignment Potential ↑, Misuse Potential ↑, Transition Turbulence ↑, Civilizational Competence, AI Ownership ↑, AI Uses ↑Long-term Trajectory

How Ultimate Scenarios Differ from Other Concepts

Section titled “How Ultimate Scenarios Differ from Other Concepts”
ConceptWhat It IsExample
Root FactorsAggregate variables that shape scenarios”Misalignment Potential”
ParametersSpecific measurable factors”Alignment Robustness”
RisksThings that could go wrong”Deceptive Alignment”
Ultimate ScenariosIntermediate pathways connecting factors to outcomes”AI Takeover”
Ultimate OutcomesHigh-level goals we care about”Existential Catastrophe”, “Long-term Trajectory”

Key distinction: A risk like “deceptive alignment” is a mechanism that could happen. An ultimate scenario like “AI Takeover” is the outcome that results if such mechanisms play out. Multiple risks can contribute to a single ultimate scenario.


Without this layer, the connection between “Misalignment Potential increasing” and “Existential Catastrophe increasing” is abstract. Ultimate scenarios show the specific pathway: alignment fails → AI develops misaligned goals → AI takes over → catastrophe.

2. Enables Different Intervention Strategies

Section titled “2. Enables Different Intervention Strategies”

Different ultimate scenarios require different interventions:

  • AI Takeover: Technical alignment, capability restrictions
  • Human-Caused Catastrophe: International coordination, misuse prevention
  • Long-term Lock-in: Power distribution, institutional design

Ultimate scenarios map directly onto scenarios that organizations can plan for. Rather than asking “what if Existential Catastrophe increases?”, planners can ask “what if we’re heading toward a Human-Caused Catastrophe?“

Each ultimate scenario corresponds to threat models discussed in the AI safety literature:

  • Carlsmith’s six-premise argument → AI Takeover scenarios
  • Christiano’s “What Failure Looks Like” → Gradual AI Takeover
  • Ord’s “The Precipice” risk categories → Multiple ultimate scenarios
  • Kasirzadeh’s decisive vs. accumulative → Rapid vs. Gradual takeover

  • Map specific risks to the ultimate scenarios they could produce
  • Estimate which ultimate scenarios are most likely given current parameter trends
  • Identify which parameters to prioritize based on which ultimate scenarios concern you most
  • Design interventions targeted at preventing specific ultimate scenarios
  • Coordinate across domains (a single ultimate scenario may require multiple types of intervention)
  • Track early warning signs for each ultimate scenario
  • Use ultimate scenarios to frame research priorities
  • Connect technical work to concrete scenarios it addresses
  • Identify gaps in our understanding of specific pathways