Skip to content

Risks & Failure Modes

This section documents risks—harmful outcomes and the mechanisms that lead to them. These are organized into four categories: accident risks (technical failures), misuse risks (intentional harm), structural risks (societal harms), and epistemic risks (threats to collective knowledge).

Each risk is tagged with a causal level: outcomes (end-state harms), pathways (mechanisms), or amplifiers (enabling conditions).

Accident

Technical failures where AI systems behave in unintended ways

Misuse

Intentional harmful applications by malicious actors

Structural

Systemic risks from how AI reshapes society and institutions

Epistemic

End-state harms to collective knowledge and truth

Risks operate at different levels of causation. Understanding this helps identify where interventions are most effective:

LevelDescriptionExamples
OutcomeEnd-state catastrophic harms we ultimately want to avoidLock-in, Authoritarian Takeover, Epistemic Collapse, Economic Disruption
PathwayMechanisms and failure modes that lead to outcome-level harmsDeceptive Alignment, Power-Seeking, Authentication Collapse, Trust Erosion
AmplifierEnabling conditions that increase probability or severityRacing Dynamics, Multipolar Trap, Sycophancy, Automation Bias

Why this matters:

  • Addressing amplifiers can prevent multiple pathways
  • Blocking pathways prevents specific outcomes
  • Understanding outcomes clarifies what we’re ultimately trying to avoid
  • Some interventions work at multiple levels simultaneously
Filter by level:
54Total
15Catastrophic
29High
18Accident
17Epistemic
8Misuse
11Structural
54 of 54 results
Solutions
64Authentication CollapseEpistemicPathwaycriticalMedium (emerging)2025-2030Emerging
74AI Authoritarian ToolsMisuseAmplifierhighHigh (occurring)CurrentGrowing
62Automation BiasEpistemicAmplifiermediumVery-high (occurring)CurrentMature
67Autonomous WeaponsMisuseOutcomehighHigh (occurring)CurrentMature
82Bioweapons RiskMisuseOutcomecatastrophicMedium (emerging)2025-2030Growing
25Concentration of PowerStructuralOutcomehighMedium-high2025-2040Growing
62Consensus ManufacturingEpistemicPathwayhighMedium (occurring)2025-2030Emerging
85Corrigibility FailureAccidentPathwaycatastrophicHigh2035Growing
38Cyber PsychosisEpistemicOutcomemedium-highMedium (emerging)2025-2030Neglected
72Cyberweapons RiskMisuseOutcomehighHigh (emerging)CurrentGrowing
85Deceptive AlignmentAccidentPathwaycatastrophicMedium2035Growing
52DeepfakesMisusePathwaymedium-highVery-high (occurring)CurrentMature
62AI DisinformationMisusePathwayhighVery-high (occurring)CurrentMature
78Distributional ShiftAccidentAmplifiermediumVery-high (occurring)CurrentMature
25Economic DisruptionStructuralOutcomemedium-highHigh2030Growing
82Emergent CapabilitiesAccidentAmplifierhighMedium (occurring)CurrentGrowing
67EnfeeblementStructuralOutcomemedium-highMedium2030Neglected
25Epistemic CollapseEpistemicOutcomehighMedium-high2030Neglected
25Erosion of Human AgencyStructuralOutcomemedium-highHigh2030Neglected
25Expertise AtrophyEpistemicPathwayhighMedium2025-2050Neglected
68Flash DynamicsStructuralAmplifierhighMedium-highCurrentNeglected
42AI-Powered FraudMisuseOutcomehighVery-high (occurring)CurrentGrowing
82Goal MisgeneralizationAccidentPathwayhighHigh (occurring)2025-2030Growing
42AI-Enabled Historical RevisionismEpistemicPathwayhighMedium (emerging)2025-2040Neglected
75Institutional Decision CaptureEpistemicOutcomehighMedium (emerging)2025-2040Emerging
92Instrumental ConvergenceAccidentPathwayhighHigh (theoretical)2035Mature
62IrreversibilityStructuralAmplifiercriticalMedium2030Growing
62AI Knowledge MonopolyEpistemicOutcomecriticalMedium2030-2050Neglected
62Epistemic Learned HelplessnessEpistemicOutcomehighMedium (emerging)2030-2050Neglected
52Legal Evidence CrisisEpistemicOutcomehighMedium (emerging)2025-2035Neglected
82Lock-inStructuralOutcomecatastrophicMedium2030-2045Growing
78Authoritarian TakeoverAccidentOutcomecatastrophicMedium2025-2050Growing
82Mesa-OptimizationAccidentPathwaycatastrophicMedium (theoretical)2035Growing
82Multipolar TrapStructuralAmplifierhighMedium-high2030Growing
85Power-Seeking AIAccidentPathwaycatastrophicMedium (theoretical)2035Mature
25Preference ManipulationEpistemicPathwayhighMedium (occurring)2025-2035Emerging
73AI ProliferationStructuralAmplifierhighHighCurrentGrowing
82Racing DynamicsStructuralAmplifierhighHigh (occurring)CurrentGrowing
25Reality FragmentationEpistemicOutcomehighMedium (occurring)2025-2035Emerging
81Reward HackingAccidentPathwayhighVery-high (occurring)CurrentMature
78SandbaggingAccidentAmplifierhighMedium2025-2030Emerging
85SchemingAccidentPathwaycatastrophicMedium2035Emerging
62Scientific Knowledge CorruptionEpistemicOutcomehighMedium (occurring)2024-2035Emerging
85Sharp Left TurnAccidentPathwaycatastrophicMedium2035Emerging
64AI Mass SurveillanceMisuseOutcomehighVery-high (occurring)CurrentMature
64SycophancyAccidentAmplifiermediumVery-high (occurring)CurrentGrowing
62Epistemic SycophancyEpistemicAmplifiermedium-highMedium (occurring)2025-2030Emerging
82Treacherous TurnAccidentPathwaycatastrophicMedium (theoretical)2035Mature
62Trust Cascade FailureEpistemicPathwaycriticalMedium (emerging)2025-2040Neglected
25Trust DeclineEpistemicPathwaymedium-highHighCurrentGrowing
64Winner-Take-All DynamicsStructuralAmplifierhighHighCurrentGrowing
Autonomous ReplicationAccidentHighMedium (emerging)2025-2030Emerging
AI-Enabled CyberattacksAccidentHighHigh (occurring)CurrentMature
AI-Enabled Biological RisksAccidentCatastrophicMedium (emerging)2025-2032Growing
Currently ObservableEmergingTheoretical/Future
SycophancySandbaggingScheming
Reward HackingDisinformation at scaleTreacherous Turn
Specification GamingDeepfakesSharp Left Turn
Concentration of PowerEmergent CapabilitiesLock-in
Reality FragmentationInstitutional CapturePower-Seeking AI

Understanding observable failures helps us reason about future risks, though the relationship between current problems and future catastrophic risks is debated.

These categories aren’t independent:

  • Accident + Misuse: Misuse is more dangerous when AI is more capable; accident risks determine capability levels
  • Structural + Accident: Racing dynamics (an amplifier) make accidents more likely by reducing safety investment
  • Epistemic + All: If we can’t agree on what risks exist, coordinating responses is impossible
  • Structural + Misuse: Concentration of power determines who might misuse AI; proliferation determines who has access
+642