Skip to content

Agent Foundations

📋Page Status
Quality:78 (Good)
Importance:72.5 (High)
Last edited:2025-12-28 (10 days ago)
Words:2.3k
Structure:
📊 10📈 1🔗 43📚 012%Score: 11/15
LLM Summary:Agent foundations research develops mathematical frameworks for understanding aligned agency, including embedded agency, decision theory, and corrigibility. MIRI's 2024 pivot to policy work reflects pessimism about theoretical progress arriving in time. Assessment indicates low tractability and high neglectedness, with value heavily dependent on alignment difficulty and timeline assumptions.

Agent foundations research seeks to develop mathematical frameworks for understanding agency, goals, and alignment before advanced AI systems manifest the problems these frameworks aim to solve. The approach treats alignment as requiring theoretical prerequisites: formal definitions of concepts like “corrigibility,” “embedded reasoning,” and “value learning” that must be established before building reliably aligned systems.

This research program was pioneered by the Machine Intelligence Research Institute (MIRI) starting in 2013. Key contributions include embedded agency (how agents reason about themselves within their environment), logical induction (managing uncertainty about mathematical truths), functional decision theory (how agents should make decisions under uncertainty), and formal corrigibility analysis (conditions under which agents accept correction).

The field faces a critical juncture. In January 2024, MIRI announced a strategic pivot away from agent foundations research, stating that “alignment research at MIRI and in the larger field had gone too slowly” and that they “now believed this research was extremely unlikely to succeed in time to prevent an unprecedented catastrophe.” MIRI has shifted focus to policy and communications work, with their agent foundations team discontinued. This development has intensified debate about whether agent foundations represents essential groundwork or a research dead end.

DimensionAssessmentConfidenceNotes
TractabilityLowMediumMIRI’s 2024 pivot suggests internal pessimism; key problems remain open after 10+ years
Value if alignment hardHighMediumMay be necessary for robust solutions to deceptive alignment
Value if alignment easyLowHighProbably not needed if empirical methods suffice
NeglectednessVery HighHighFewer than 20 full-time researchers globally; MIRI discontinuation further reduces
Current momentumDecliningHighPrimary organization pivoted away; limited successor efforts
Theory-practice gapUnknownLowCore question: will formalisms apply to neural networks?
WorldviewAssessmentReasoning
DoomerMixedMay need foundations for robust alignment, but pessimistic about tractability
Long timelinesPositiveMore time for theoretical progress to compound
OptimisticLow priorityEmpirical methods likely sufficient
Governance-focusedLow priorityPolicy interventions more tractable in relevant timeframes

Agent foundations research spans several interconnected problem domains. The 2018 “Embedded Agency” framework by Abram Demski and Scott Garrabrant unified these into a coherent research agenda, identifying four core subproblems of embedded agency: decision theory, embedded world-models, robust delegation, and subsystem alignment.

Loading diagram...
AreaCore QuestionKey ResultsStatus
Embedded agencyHow can an agent reason about itself within the world?Demski & Garrabrant (2018) framework identifies four subproblemsConceptual clarity achieved; formal solutions lacking
Decision theoryHow should AI make decisions under uncertainty?FDT/UDT developed as alternatives to CDT/EDT; logical causality still undefinedActive research; practical relevance debated
Logical inductionHow to have calibrated beliefs about math?Garrabrant et al. (2016) algorithm achieves key desiderataMajor success; integration with decision theory incomplete
CorrigibilityCan we formally define “willing to be corrected”?Soares et al. (2015) proved utility indifference approach has fundamental limitsOpen problem; no complete solution
Value learningHow to learn values from observations?CIRL (Hadfield-Menell et al. 2016) formalizes cooperative learningTheoretical framework exists; practical applicability unclear

Traditional models of rational agents treat them as separate from their environment, able to model it completely and act on it from “outside.” Real AI systems are embedded: they exist within the world they’re modeling, are made of the same stuff as their environment, and cannot fully model themselves.

The embedded agency sequence identifies four interconnected challenges:

Decision Theory: How should an embedded agent make decisions when its decision procedure is part of the world being reasoned about? Standard decision theories assume the agent can consider counterfactual actions, but an embedded agent’s reasoning is itself a physical process that affects outcomes.

Embedded World-Models: How can an agent maintain beliefs about a world that contains the agent itself? Self-referential reasoning creates paradoxes similar to Godel’s incompleteness theorems.

Robust Delegation: How can an agent create successor agents that reliably pursue the original agent’s goals? This connects to corrigibility and value learning.

Subsystem Alignment: How can an agent ensure its internal optimization processes remain aligned with its overall goals? This prefigured the mesa-optimization concerns that later became central to alignment discourse.

Logical induction (Garrabrant et al., 2016) is widely considered agent foundations’ most significant concrete achievement. The algorithm provides a computable method for assigning probabilities to mathematical statements, analogous to how Solomonoff induction handles empirical uncertainty.

A logical inductor assigns probabilities to mathematical claims that:

  • Converge to truth: Probabilities approach 1 for true statements, 0 for false ones
  • Outpace deduction: Assigns high probability to provable claims before proofs are found
  • Resist exploitation: No computationally bounded trader can profit systematically

This addresses a fundamental gap: Bayesian reasoning assumed logical omniscience, but real reasoners (including AI systems) must be uncertain about mathematical facts. Logical induction provides the first satisfactory theory for such uncertainty.

However, integration with decision theory remains incomplete. As Wei Dai and others have noted, the framework doesn’t yet combine smoothly with updateless decision theory, limiting its practical applicability.

Corrigibility (Soares et al., 2015) formalized the challenge of building AI systems that accept human correction. A corrigible agent should:

  • Not manipulate its operators
  • Preserve oversight mechanisms
  • Follow shutdown instructions
  • Ensure these properties transfer to successors

The paper proved that naive approaches fail. “Utility indifference”—giving equal utility to shutdown and continued operation—creates perverse incentives: if shutdown utility is too high, agents may engineer their own shutdown; if too low, they resist it.

No complete solution exists. Stuart Armstrong’s work at the Future of Humanity Institute continued exploring corrigibility, but fundamental obstacles remain. The problem connects to instrumental convergence: almost any goal creates incentives to resist modification.

Decision Theory: FDT, UDT, and Open Questions

Section titled “Decision Theory: FDT, UDT, and Open Questions”

Agent foundations researchers developed novel decision theories to address problems with standard frameworks:

TheoryKey InsightLimitations
CDT (Causal)Choose action with best causal consequencesFails on Newcomb-like problems
EDT (Evidential)Choose action with best correlated outcomesCan recommend “managing the news”
FDT (Functional)Choose as if choosing the output of your decision algorithmRequires undefined “logical causation”
UDT (Updateless)Pre-commit to policies before receiving observationsDoesn’t integrate well with logical induction

Functional Decision Theory (Yudkowsky & Soares, 2017) and Updateless Decision Theory (Wei Dai, 2009) were designed to handle cases where the agent’s decision procedure is predicted by other agents or affects multiple instances of itself. These theories outperform CDT and EDT in specific scenarios but require technical machinery—particularly “logical counterfactuals”—that remains undefined.

The MIRI/Open Philanthropy exchange on decision theory reveals ongoing disagreement about whether this research direction is productive or has generated more puzzles than solutions.


The Core Debate: Agent Foundations vs. Empirical Safety

Section titled “The Core Debate: Agent Foundations vs. Empirical Safety”

The most consequential disagreement in AI safety research methodology concerns whether theoretical foundations are necessary for alignment or whether empirical investigation of neural networks is more productive.

Proponents argue that alignment requires understanding agency at a fundamental level:

Theoretical prerequisites exist: Just as designing safe bridges requires understanding materials science, designing aligned AI may require understanding what goals, agency, and alignment formally mean. Ad hoc solutions may fail unpredictably on more capable systems.

Current methods won’t scale: Techniques like RLHF work on current systems but may fail catastrophically on more capable ones. The sharp left turn hypothesis suggests capabilities may generalize while alignment doesn’t—requiring foundational understanding of why.

Deceptive alignment is fundamental: Deceptive alignment and mesa-optimization may require formal analysis rather than empirical debugging. These failure modes are precisely about systems that pass empirical tests while remaining misaligned.

Conceptual clarity enables coordination: Even if formal solutions aren’t achieved, precise problem definitions help researchers coordinate and avoid talking past each other.

Critics, including many who previously supported agent foundations, argue the approach has failed:

Slow progress on key problems: After 10+ years of work, corrigibility remains unsolved, decision theory has generated more puzzles than solutions, and logical induction doesn’t integrate with the rest of the framework. MIRI’s 2024 pivot implicitly acknowledges this.

Neural networks may not fit formalisms: Agent foundations research makes few assumptions about “gears-level properties” of AI systems. But modern AI is increasingly certain to be neural networks. Mechanistic interpretability and empirical safety research can assume neural network architectures and make faster progress.

The Science of Deep Learning argument: Researchers like Neel Nanda argue that understanding how neural networks actually work (through mechanistic interpretability, scaling laws, etc.) is more tractable than developing abstract theories. Neural network properties may generalize even if specific alignment techniques don’t.

Theory-practice gap is uncertain: It’s unknown whether theoretical results will apply to systems that emerge from gradient descent. The gap between “ideal agents” studied in agent foundations and actual neural networks may be unbridgeable.

PositionProbabilityImplications
Agent foundations essential15-25%Major research effort needed; current neglect is harmful
Agent foundations helpful but not necessary30-40%Worth some investment; complements empirical work
Agent foundations irrelevant to actual AI25-35%Resources better allocated to empirical safety
Agent foundations actively harmful (delays practical work)5-15%Should be deprioritized or defunded

Evidence for tractabilityEvidence against
Logical induction solved a decades-old problemCore problems (corrigibility, decision theory) remain open after 10+ years
Progress on conceptual clarity (embedded agency framework)MIRI’s 2024 pivot suggests insider pessimism
Problems are well-defined and mathematically precisePrecision hasn’t translated to solutions
Small field—more researchers might helpField has been trying to expand for years with limited success

Current assessment: Low tractability (60-75% confidence). The difficulty appears intrinsic rather than due to resource constraints.

Evidence for transferEvidence against
Mesa-optimization concepts from agent foundations now guide empirical researchActual neural networks may not have the structure assumed by formalisms
Conceptual frameworks help identify what to look for empiricallyGradient descent produces systems very different from idealized agents
Mathematical guarantees could provide confidenceAssumptions required for guarantees may not hold in practice

Current assessment: Uncertain (low confidence). This is the crux where evidence is weakest.

Short timelines (AGI by 2030)Long timelines (AGI 2040+)
Agent foundations unlikely to produce results in timeMore time for theoretical progress to compound
Must work with practical methods on current systemsCan develop foundations before crisis
MIRI’s stated reason for pivotingTheoretical work becomes higher-value investment

Current assessment: Conditional value. Agent foundations’ expected value is 3-5x higher under long timelines assumptions.


FactorWeightNotes
Strong math/philosophy backgroundHighResearch requires formal methods, logic, decision theory
Belief alignment is fundamentally hardHighMust expect empirical methods to be insufficient
Long timeline assumptionsMediumAffects expected value of theoretical work
Tolerance for slow, uncertain progressHighField has historically moved slowly
Interest in problems for their own sakeMediumMay need intrinsic motivation given uncertain practical payoff

Advantages: Very neglected (high marginal value if tractable); intellectually interesting problems; potential for high-impact breakthroughs if successful.

Disadvantages: Low tractability; uncertain practical relevance; declining institutional support (MIRI pivot); difficult to evaluate progress.

Alternatives to consider: Mechanistic interpretability, scalable oversight, AI control, governance research.


  • MIRI 2024 Mission and Strategy Update: MIRI blog explaining pivot away from agent foundations
  • MIRI/Open Philanthropy Exchange on Decision Theory: AI Alignment Forum
  • Science of Deep Learning vs. Agent Foundations: LessWrong post arguing for empirical approaches
  • Cooperative Inverse Reinforcement Learning (Hadfield-Menell et al., 2016): arXiv - Formalizes value learning as cooperative game
  • Inverse Reinforcement Learning: AI Alignment Forum wiki

Agent foundations research improves the Ai Transition Model through Misalignment Potential:

FactorParameterImpact
Misalignment PotentialAlignment RobustnessProvides theoretical foundations for robust alignment across capability levels
Misalignment PotentialSafety-Capability GapFormal understanding may enable alignment techniques that scale with capabilities

Agent foundations work is most valuable under long timeline assumptions where theoretical progress can compound before transformative AI development.