Conjecture
Overview
Section titled “Overview”Conjecture is an AI safety research organization founded in 2021 by Connor Leahy and a team of researchers concerned about existential risks from advanced AI. The organization pursues a distinctive technical approach centered on “Cognitive Emulation” (CoEm) - building interpretable AI systems based on human cognition principles rather than aligning existing large language models.
Based in London with a team of 30-40 researchers, Conjecture raised over $10M in Series A funding in 2023. Their research agenda emphasizes mechanistic interpretability and understanding neural network internals, representing a fundamental alternative to mainstream prosaic alignment approaches pursued by organizations like Anthropic and OpenAI.
| Aspect | Assessment | Evidence | Source |
|---|---|---|---|
| Technical Innovation | High | Novel CoEm research agenda | Conjecture Blog↗ |
| Funding Security | Strong | $30M+ Series A (2023) | TechCrunch Reports↗ |
| Research Output | Moderate | Selective publication strategy | Research Publications↗ |
| Influence | Growing | European AI policy engagement | UK AISI↗ |
Risk Assessment
Section titled “Risk Assessment”| Risk Category | Severity | Likelihood | Timeline | Trend |
|---|---|---|---|---|
| CoEm Uncompetitive | High | Moderate | 3-5 years | Uncertain |
| Commercial Pressure Compromise | Medium | High | 2-3 years | Worsening |
| Research Insularity | Low | Moderate | Ongoing | Stable |
| Funding Sustainability | Medium | Low | 5+ years | Improving |
Founding and Evolution
Section titled “Founding and Evolution”Origins (2021)
Section titled “Origins (2021)”Conjecture emerged from the EleutherAI collective, an open-source AI research group that successfully recreated GPT-3 as open-source models (GPT-J, GPT-NeoX). Key founding factors:
| Factor | Impact | Details |
|---|---|---|
| EleutherAI Experience | High | Demonstrated capability replication feasibility |
| Safety Concerns | High | Recognition of risks from capability proliferation |
| European Gap | Medium | Limited AI safety ecosystem outside Bay Area |
| Funding Availability | Medium | Growing investor interest in AI safety |
Philosophical Evolution: The transition from EleutherAI’s “democratize AI” mission to Conjecture’s safety-focused approach represents a significant shift in thinking about AI development and publication strategies.
Funding Trajectory
Section titled “Funding Trajectory”| Year | Funding Stage | Amount | Impact |
|---|---|---|---|
| 2021 | Seed | Undisclosed | Initial team of ~15 researchers |
| 2023 | Series A | $30M+ | Scaled to 30-40 researchers |
| 2024 | Operating | Ongoing | Sustained research operations |
Cognitive Emulation (CoEm) Research Agenda
Section titled “Cognitive Emulation (CoEm) Research Agenda”Core Philosophy
Section titled “Core Philosophy”Conjecture’s signature approach contrasts sharply with mainstream AI development:
| Approach | Philosophy | Methods | Evaluation |
|---|---|---|---|
| Prosaic Alignment | Train powerful LLMs, align post-hoc | RLHF, Constitutional AI | Behavioral testing |
| Cognitive Emulation | Build interpretable systems from ground up | Human cognition principles | Mechanistic understanding |
Key Research Components
Section titled “Key Research Components”Mechanistic Interpretability
- Circuit discovery in neural networks
- Feature attribution and visualization
- Scaling interpretability to larger models
- Interpretability research collaboration
Architecture Design
- Modular systems for better control
- Interpretability-first design choices
- Trading capabilities for understanding
- Novel training methodologies
Model Organisms
- Smaller, interpretable test systems
- Alignment property verification
- Deception detection research
- Goal representation analysis
Key Personnel
Section titled “Key Personnel”Connor Leahy Profile
Section titled “Connor Leahy Profile”| Aspect | Details |
|---|---|
| Background | EleutherAI collective member, GPT-J contributor |
| Evolution | From open-source advocacy to safety-focused research |
| Public Role | Active AI policy engagement, podcast appearances |
| Views | Short AI timelines, high P(doom), interpretability-necessary |
Timeline Estimates: Leahy has consistently expressed short AI timeline views, suggesting AGI within years rather than decades.
Research Focus Areas
Section titled “Research Focus Areas”Mechanistic Interpretability
Section titled “Mechanistic Interpretability”| Research Area | Status | Key Questions |
|---|---|---|
| Circuit Analysis | Active | How do transformers implement reasoning? |
| Feature Extraction | Ongoing | What representations emerge in training? |
| Scaling Methods | Development | Can interpretability scale to AGI-level systems? |
| Goal Detection | Early | How can we detect goal-directedness mechanistically? |
Comparative Advantages
Section titled “Comparative Advantages”| Organization | Primary Focus | Interpretability Approach |
|---|---|---|
| Conjecture | CoEm, ground-up interpretability | Design-time interpretability |
| Anthropic | Frontier models + interpretability | Post-hoc analysis of LLMs |
| ARC | Theoretical alignment | Evaluation and ELK research |
| Redwood | AI control | Interpretability for control |
Strategic Position
Section titled “Strategic Position”Theory of Change
Section titled “Theory of Change”Conjecture’s pathway to AI safety impact:
- Develop scalable interpretability techniques for powerful AI systems
- Demonstrate CoEm viability as competitive alternative to black-box scaling
- Influence field direction toward interpretability-first development
- Inform governance with technical feasibility insights
- Build safe systems using CoEm principles if successful
European AI Safety Hub
Section titled “European AI Safety Hub”| Role | Impact | Examples |
|---|---|---|
| Geographic Diversity | High | Alternative to Bay Area concentration |
| Policy Engagement | Growing | UK AISI consultation |
| Talent Development | Moderate | European researcher recruitment |
| Community Building | Early | Workshops and collaborations |
Challenges and Criticisms
Section titled “Challenges and Criticisms”Technical Feasibility
Section titled “Technical Feasibility”| Challenge | Severity | Status |
|---|---|---|
| CoEm Competitiveness | High | Unresolved - early stage |
| Interpretability Scaling | High | Active research question |
| Human Cognition Complexity | Medium | Ongoing investigation |
| Timeline Alignment | High | Critical if AGI timelines short |
Organizational Tensions
Section titled “Organizational Tensions”Commercial Pressure vs Safety Mission
- VC funding creates return expectations
- Potential future deployment pressure
- Comparison to Anthropic’s commercialization path
Publication Strategy Criticism
- Shift from EleutherAI’s radical openness
- Selective research sharing decisions
- Balance between transparency and safety
Current Research Outputs
Section titled “Current Research Outputs”Published Work
Section titled “Published Work”| Type | Focus | Impact |
|---|---|---|
| Technical Papers | Interpretability methods | Research community |
| Blog Posts | CoEm explanations | Public understanding |
| Policy Contributions | Technical feasibility | Governance decisions |
| Open Source Tools | Interpretability software | Research ecosystem |
Research Questions
Section titled “Research Questions”❓Key Questions
Timeline and Risk Estimates
Section titled “Timeline and Risk Estimates”Based on public statements and research direction
| Source | Estimate | Date |
|---|---|---|
| Connor Leahy | AGI: 2-10 years | 2023-2024 |
| Connor Leahy | P(doom): High without major changesAssessment | 2023 |
| Conjecture Research | Prosaic alignment: InsufficientAssessment | Ongoing |
| Organization | Interpretability: Necessary for safetyAssessment | Founding |
Connor Leahy: Consistently short timeline estimates
Connor Leahy: Expressed significant concern about default trajectory
Conjecture Research: Core motivation for CoEm approach
Organization: Fundamental research assumption
Future Scenarios
Section titled “Future Scenarios”Research Trajectory Projections
Section titled “Research Trajectory Projections”| Timeline | Optimistic | Realistic | Pessimistic |
|---|---|---|---|
| 2-3 years | CoEm demonstrations, policy influence | Continued interpretability advances | Commercial pressure compromises |
| 3-5 years | Competitive interpretable systems | Mixed results, partial success | Research agenda stagnates |
| 5+ years | Field adoption of CoEm principles | Portfolio contribution to safety | Marginalized approach |
Critical Dependencies
Section titled “Critical Dependencies”| Factor | Importance | Uncertainty |
|---|---|---|
| Technical Feasibility | Critical | High - unproven at scale |
| Funding Continuity | High | Medium - VC expectations |
| AGI Timeline | Critical | High - if very short, insufficient time |
| Field Receptivity | Medium | Medium - depends on results |
Relationships and Collaborations
Section titled “Relationships and Collaborations”Within AI Safety Ecosystem
Section titled “Within AI Safety Ecosystem”| Organization | Relationship | Collaboration Type |
|---|---|---|
| Anthropic | Friendly competition | Interpretability research sharing |
| ARC | Complementary | Different technical approaches |
| MIRI | Aligned concerns | Skepticism of prosaic alignment |
| Academic Labs | Collaborative | Interpretability technique development |
Policy and Governance
Section titled “Policy and Governance”UK Engagement
- UK AI Safety Institute consultation
- Technical feasibility assessments
- European AI Act discussions
International Influence
- Growing presence in global AI safety discussions
- Alternative perspective to US-dominated discourse
- Technical grounding for governance approaches
Sources & Resources
Section titled “Sources & Resources”Primary Sources
Section titled “Primary Sources”| Type | Source | Description |
|---|---|---|
| Official Website | Conjecture.dev↗ | Research updates, team information |
| Research Papers | Google Scholar↗ | Technical publications |
| Blog Posts | Conjecture Blog↗ | Research explanations, philosophy |
| Interviews | Connor Leahy Talks↗ | Leadership perspectives |
Secondary Analysis
Section titled “Secondary Analysis”| Type | Source | Focus |
|---|---|---|
| AI Safety Analysis | LessWrong Posts↗ | Community discussion |
| Technical Reviews | Alignment Forum↗ | Research evaluation |
| Policy Reports | GovAI Analysis↗ | Governance implications |
| Funding News | TechCrunch Coverage↗ | Business developments |
Related Resources
Section titled “Related Resources”| Topic | Internal Links | External Resources |
|---|---|---|
| Interpretability | Technical Interpretability | Anthropic Interpretability↗ |
| Alignment Approaches | Why Alignment is Hard | AI Alignment Forum↗ |
| European AI Policy | UK AISI | EU AI Office↗ |
| Related Orgs | Safety Organizations | AI Safety Community↗ |