Skip to content

About This Wiki

Entry

Transparency & Credibility

PurposeHelp you assess whether to trust this resource
Last UpdatedDecember 2024

Before using this wiki to inform your views on AI risk, you should know who made it, what biases it has, and how to evaluate its claims.


This wiki is an experimental project exploring how to structure and communicate AI safety arguments. It was created primarily using Claude Code (Anthropic’s AI coding assistant) as an experiment in AI-assisted content creation.

What this means for you:

  • The content reflects what an AI system trained on AI safety discourse would produce
  • It may inherit biases present in that discourse
  • It has been reviewed but not comprehensively fact-checked
  • It should be treated as a starting point for exploration, not a definitive reference

351
Content Pages
315
Entities Tracked
3133
External Resources
~970K
Words (approx)
MetricValue
Content pages351
Entities tracked315
External resources/citations3133
Approximate word count~970K
Average quality score4.4 / 5.0
CreatedDecember 2024
Last data build1/7/2026
CategoryCount
Risks & Risk Factors61
Responses (Agendas, Interventions, Policies)38
Organizations & Labs19
People28
Analytical Models78
Concepts & Capabilities38
Quality LevelCount
Low (1-2) — Needs improvement6
Adequate (3) — Meets basic standards19
High (4-5) — Well-developed282

For detailed quality metrics and entity gap analysis, see the Dashboard.


What Perspective Does This Wiki Represent?

Section titled “What Perspective Does This Wiki Represent?”

This wiki is written from within the AI safety / x-risk community perspective. This is a real intellectual community with:

  • Shared assumptions (AI could be transformatively powerful, alignment is hard, existential risk is possible)
  • Shared vocabulary (TAI, P(doom), alignment, mesa-optimization)
  • Shared heroes (Bostrom, Yudkowsky, Russell, etc.)
  • Shared funding sources (Open Philanthropy, EA funders)
  • Social dynamics that can create groupthink

The wiki tries to present counterarguments, but it does so from within this perspective. The structure of the arguments, the choice of what to cover, and the framing all reflect AI safety community assumptions.

Assumptions this wiki takes for granted (which are contested elsewhere):

AssumptionContested By
”AGI” is a coherent conceptSome ML researchers (LeCun, Marcus)
Existential risk from AI is possibleAI ethics community, some ML researchers
Theoretical risk analysis is valid before empirical evidenceEmpiricists, mainstream ML
Long-term future matters morallyPerson-affecting ethicists
The AI safety research agenda is roughly correctAI ethics researchers, some ML researchers

If you reject these assumptions, this wiki may still be useful for understanding what AI safety researchers believe—but you should not treat it as a neutral source.


  • Alignment Forum and LessWrong posts
  • Academic papers on AI safety
  • Public statements from researchers and organizations
  • AI safety organization websites
  • Forecasting platforms (Metaculus, AI Impacts surveys)
  1. Initial content generated with AI assistance
  2. Structured around “cruxes” (key disagreement points)
  3. Reviewed for accuracy against primary sources
  4. Counterarguments added to major claims
  5. Limitations acknowledged explicitly
  • Systematic literature review
  • Original research or interviews
  • Comprehensive fact-checking
  • Peer review by subject matter experts
  • Input from AI ethics / critical AI studies perspectives

Well-CoveredUnder-Covered
Technical alignment researchAI ethics / fairness
X-risk argumentsMundane AI harms
Western/Anglophone perspectivesNon-Western perspectives
EA-adjacent organizationsAcademic ML safety
Arguments for concernSkeptical/dismissive views
  1. Selection of “cruxes”: The cruxes chosen reflect what the AI safety community considers important, not necessarily what’s actually decision-relevant
  2. Probability estimates: Numbers are drawn primarily from AI safety-adjacent sources
  3. Organization coverage: More detail on orgs the AI safety community considers important
  4. Framing of debates: The “Is AI x-risk real?” debate is framed as a debate, not as one side potentially being correct
  • Created using Anthropic’s Claude, which has interests in AI safety being taken seriously
  • Content reflects perspectives likely to increase AI safety funding/attention
  • No financial conflicts, but intellectual/tribal ones exist

  1. Check the sources provided
  2. Look for the original source, not just citations
  3. Check if claims are contested by credible critics
  4. Note the date—AI safety discourse moves fast
  1. Treat as rough intuitions, not rigorous calculations
  2. Note whose estimates are cited (selection bias)
  3. Remember these are uncertain by definition
  4. Your informed estimate is as valid as anyone’s
  1. Check if counterarguments are addressed
  2. Look for the strongest version of opposing views
  3. Ask: “What would falsify this?”
  4. Consider whether the argument structure privileges certain conclusions
  1. Consider who benefits from the recommendation
  2. Check if alternatives are fairly presented
  3. Ask if the recommendation follows from the arguments
  4. Consider your own situation and values

A healthy knowledge resource should update based on evidence. Here’s our track record:

  • Added explicit “Limitations & Perspective” section
  • Added criticism sections to major argument pages
  • Changed fake-precise probabilities to ranges
  • Added sources for alternative viewpoints
  • Acknowledged non-neutral framing
  • Evidence of deceptive behavior in AI systems → would increase concern emphasis
  • Continued absence of concerning behaviors → would moderate concern emphasis
  • Major alignment breakthroughs → would increase optimism
  • Alignment techniques failing at scale → would increase pessimism
  • Feedback from AI ethics researchers → would improve coverage of that perspective

We don’t have systematic tracking of predictions made in this wiki vs. outcomes. This is a significant limitation for assessing calibration.


  • Open an issue on GitHub explaining the gap
  • Suggest specific sources that should be included
  • Write a critique we can link to
  • Factual errors: Open an issue with correction and source
  • Broken links: Open an issue or PR
  • Outdated information: Suggest update with new source
  • AI safety researchers: Help ensure technical accuracy
  • AI ethics researchers: Help improve coverage of that perspective
  • ML practitioners: Help ground theoretical claims in practice
  • Forecasters: Help improve probability estimates

The Meta-Question: Can You Trust AI-Generated Content About AI Risk?

Section titled “The Meta-Question: Can You Trust AI-Generated Content About AI Risk?”

This wiki was largely created with AI assistance. This creates an interesting epistemological situation:

Reasons for caution:

  • AI systems might have biases toward certain conclusions about AI
  • The AI safety community’s perspectives may be overrepresented in training data
  • AI assistance may create false confidence through fluent, confident-sounding prose

Reasons this might be okay:

  • AI assistance is transparent (you know it was used)
  • The content can be checked against primary sources
  • Human review and editing occurred
  • Explicit acknowledgment of limitations

Our recommendation: Treat this wiki as a well-organized summary of AI safety community views, not as an authoritative source. Use it to navigate the landscape and find primary sources—then form your own conclusions.


DimensionAssessment
Factual accuracyModerate—based on reputable sources but not comprehensively fact-checked
Balanced coverageLow—systematically biased toward AI safety perspective
Source qualityModerate—mix of academic papers, forum posts, and public statements
TransparencyHigh—explicit about limitations and perspective
UpdatabilityHigh—can incorporate feedback and new evidence

Bottom line: Use this wiki to understand what the AI safety community believes and why. Don’t use it as your only source on AI risk. Seek out critics and alternative perspectives.