Skip to content

Conjecture

📋Page Status
Quality:72 (Good)
Importance:38.5 (Reference)
Last edited:2025-12-24 (14 days ago)
Words:1.4k
Structure:
📊 17📈 0🔗 37📚 014%Score: 10/15
LLM Summary:Conjecture pursues Cognitive Emulation (CoEm) as an interpretability-first alternative to mainstream alignment, with $30M+ Series A funding (2023) and 30-40 researchers. Key challenge is whether CoEm can remain competitive against prosaic approaches given potentially short AGI timelines.
Research Lab

Conjecture

Importance38

Conjecture is an AI safety research organization founded in 2021 by Connor Leahy and a team of researchers concerned about existential risks from advanced AI. The organization pursues a distinctive technical approach centered on “Cognitive Emulation” (CoEm) - building interpretable AI systems based on human cognition principles rather than aligning existing large language models.

Based in London with a team of 30-40 researchers, Conjecture raised over $10M in Series A funding in 2023. Their research agenda emphasizes mechanistic interpretability and understanding neural network internals, representing a fundamental alternative to mainstream prosaic alignment approaches pursued by organizations like Anthropic and OpenAI.

AspectAssessmentEvidenceSource
Technical InnovationHighNovel CoEm research agendaConjecture Blog
Funding SecurityStrong$30M+ Series A (2023)TechCrunch Reports
Research OutputModerateSelective publication strategyResearch Publications
InfluenceGrowingEuropean AI policy engagementUK AISI
Risk CategorySeverityLikelihoodTimelineTrend
CoEm UncompetitiveHighModerate3-5 yearsUncertain
Commercial Pressure CompromiseMediumHigh2-3 yearsWorsening
Research InsularityLowModerateOngoingStable
Funding SustainabilityMediumLow5+ yearsImproving

Conjecture emerged from the EleutherAI collective, an open-source AI research group that successfully recreated GPT-3 as open-source models (GPT-J, GPT-NeoX). Key founding factors:

FactorImpactDetails
EleutherAI ExperienceHighDemonstrated capability replication feasibility
Safety ConcernsHighRecognition of risks from capability proliferation
European GapMediumLimited AI safety ecosystem outside Bay Area
Funding AvailabilityMediumGrowing investor interest in AI safety

Philosophical Evolution: The transition from EleutherAI’s “democratize AI” mission to Conjecture’s safety-focused approach represents a significant shift in thinking about AI development and publication strategies.

YearFunding StageAmountImpact
2021SeedUndisclosedInitial team of ~15 researchers
2023Series A$30M+Scaled to 30-40 researchers
2024OperatingOngoingSustained research operations

Cognitive Emulation (CoEm) Research Agenda

Section titled “Cognitive Emulation (CoEm) Research Agenda”

Conjecture’s signature approach contrasts sharply with mainstream AI development:

ApproachPhilosophyMethodsEvaluation
Prosaic AlignmentTrain powerful LLMs, align post-hocRLHF, Constitutional AIBehavioral testing
Cognitive EmulationBuild interpretable systems from ground upHuman cognition principlesMechanistic understanding

Mechanistic Interpretability

  • Circuit discovery in neural networks
  • Feature attribution and visualization
  • Scaling interpretability to larger models
  • Interpretability research collaboration

Architecture Design

  • Modular systems for better control
  • Interpretability-first design choices
  • Trading capabilities for understanding
  • Novel training methodologies

Model Organisms

  • Smaller, interpretable test systems
  • Alignment property verification
  • Deception detection research
  • Goal representation analysis
Leadership Team
CL
Connor Leahy
CEO and Co-founder
SB
Sid Black
Co-founder
GA
Gabriel Alfour
CTO
AspectDetails
BackgroundEleutherAI collective member, GPT-J contributor
EvolutionFrom open-source advocacy to safety-focused research
Public RoleActive AI policy engagement, podcast appearances
ViewsShort AI timelines, high P(doom), interpretability-necessary

Timeline Estimates: Leahy has consistently expressed short AI timeline views, suggesting AGI within years rather than decades.

Research AreaStatusKey Questions
Circuit AnalysisActiveHow do transformers implement reasoning?
Feature ExtractionOngoingWhat representations emerge in training?
Scaling MethodsDevelopmentCan interpretability scale to AGI-level systems?
Goal DetectionEarlyHow can we detect goal-directedness mechanistically?
OrganizationPrimary FocusInterpretability Approach
ConjectureCoEm, ground-up interpretabilityDesign-time interpretability
AnthropicFrontier models + interpretabilityPost-hoc analysis of LLMs
ARCTheoretical alignmentEvaluation and ELK research
RedwoodAI controlInterpretability for control

Conjecture’s pathway to AI safety impact:

  1. Develop scalable interpretability techniques for powerful AI systems
  2. Demonstrate CoEm viability as competitive alternative to black-box scaling
  3. Influence field direction toward interpretability-first development
  4. Inform governance with technical feasibility insights
  5. Build safe systems using CoEm principles if successful
RoleImpactExamples
Geographic DiversityHighAlternative to Bay Area concentration
Policy EngagementGrowingUK AISI consultation
Talent DevelopmentModerateEuropean researcher recruitment
Community BuildingEarlyWorkshops and collaborations
ChallengeSeverityStatus
CoEm CompetitivenessHighUnresolved - early stage
Interpretability ScalingHighActive research question
Human Cognition ComplexityMediumOngoing investigation
Timeline AlignmentHighCritical if AGI timelines short

Commercial Pressure vs Safety Mission

Publication Strategy Criticism

  • Shift from EleutherAI’s radical openness
  • Selective research sharing decisions
  • Balance between transparency and safety
TypeFocusImpact
Technical PapersInterpretability methodsResearch community
Blog PostsCoEm explanationsPublic understanding
Policy ContributionsTechnical feasibilityGovernance decisions
Open Source ToolsInterpretability softwareResearch ecosystem

Key Questions

Can CoEm produce AI systems competitive with scaled LLMs?
Is mechanistic interpretability sufficient for AGI safety verification?
How will commercial pressures affect Conjecture's research direction?
What role should interpretability play in AI governance frameworks?
Can cognitive emulation bridge neuroscience and AI safety research?
How does CoEm relate to other alignment approaches like Constitutional AI?
Risk and Timeline Assessments
📊Conjecture Leadership Risk Estimates

Based on public statements and research direction

SourceEstimateDate
Connor LeahyAGI: 2-10 years2023-2024
Connor LeahyP(doom): High without major changesAssessment2023
Conjecture ResearchProsaic alignment: InsufficientAssessmentOngoing
OrganizationInterpretability: Necessary for safetyAssessmentFounding

Connor Leahy: Consistently short timeline estimates

Connor Leahy: Expressed significant concern about default trajectory

Conjecture Research: Core motivation for CoEm approach

Organization: Fundamental research assumption

TimelineOptimisticRealisticPessimistic
2-3 yearsCoEm demonstrations, policy influenceContinued interpretability advancesCommercial pressure compromises
3-5 yearsCompetitive interpretable systemsMixed results, partial successResearch agenda stagnates
5+ yearsField adoption of CoEm principlesPortfolio contribution to safetyMarginalized approach
FactorImportanceUncertainty
Technical FeasibilityCriticalHigh - unproven at scale
Funding ContinuityHighMedium - VC expectations
AGI TimelineCriticalHigh - if very short, insufficient time
Field ReceptivityMediumMedium - depends on results
OrganizationRelationshipCollaboration Type
AnthropicFriendly competitionInterpretability research sharing
ARCComplementaryDifferent technical approaches
MIRIAligned concernsSkepticism of prosaic alignment
Academic LabsCollaborativeInterpretability technique development

UK Engagement

International Influence

  • Growing presence in global AI safety discussions
  • Alternative perspective to US-dominated discourse
  • Technical grounding for governance approaches
TypeSourceDescription
Official WebsiteConjecture.devResearch updates, team information
Research PapersGoogle ScholarTechnical publications
Blog PostsConjecture BlogResearch explanations, philosophy
InterviewsConnor Leahy TalksLeadership perspectives
TypeSourceFocus
AI Safety AnalysisLessWrong PostsCommunity discussion
Technical ReviewsAlignment ForumResearch evaluation
Policy ReportsGovAI AnalysisGovernance implications
Funding NewsTechCrunch CoverageBusiness developments
TopicInternal LinksExternal Resources
InterpretabilityTechnical InterpretabilityAnthropic Interpretability
Alignment ApproachesWhy Alignment is HardAI Alignment Forum
European AI PolicyUK AISIEU AI Office
Related OrgsSafety OrganizationsAI Safety Community