Skip to content

Model Style Guide

This guide covers core purpose, summary requirements, format requirements, diagram types, methodological principles, and rating criteria for analytical models in this knowledge base.


A good model summary follows this pattern:

“This model [methodology/approach]. It [key finding - trajectory, critical variables, or uncertainty assessment].”

Summaries should emphasize where things are going and what matters most:

Finding TypeExample
Trajectory/Projection”…projects uplift increasing from 1.5x to 3-5x by 2030”
Critical Variables”…identifies X and Y as the key variables determining outcomes”
Risk Magnitude”…estimates this represents 5-15% of total AI risk”
Uncertainty Assessment”…finds high variance across scenarios; results depend heavily on [assumption]“
Negative Finding”…finds no significant effect under current conditions, but this changes if [X]“

Good summaries:

TopicSummary
Bioweapons uplift”This model estimates AI’s contribution to bioweapons risk over time. It projects uplift increasing from 1.5x to 3-5x by 2030, with biosecurity evasion posing the greatest concern.”
Racing dynamics”This model analyzes competitive pressures among frontier labs. It finds the key variable is whether any single lab can maintain >6 month lead; if not, racing dynamics dominate.”
Lock-in probability”This model assesses paths to irreversible outcomes. Results are highly uncertain (10-60% range) depending on governance assumptions.”
Sycophancy”This model maps feedback loops in AI training. It finds no clear evidence of runaway dynamics under current training regimes, but identifies reward hacking as the critical variable to monitor.”

Bad summaries:

SummaryProblem
”Analysis of AI bioweapons risk”No methodology, no conclusion
”This model examines how racing dynamics affect safety”No finding at all
”Current LLMs provide 1.3x uplift”Current state only, no trajectory or implications
---
title: "Racing Dynamics Impact Model"
description: "This model analyzes competitive pressures among frontier labs. It estimates a 60-80% probability that racing dynamics reduce safety investment by 30-50% compared to non-competitive scenarios."
quality: 3
lastEdited: "2025-12-26"
ratings:
novelty: 4
rigor: 3
actionability: 4
completeness: 3
---

The description field:

  • Must state what the model does (methodology/approach)
  • Must include key conclusions with quantified estimates where possible
  • Should be 1-3 sentences (max ~250 characters for good preview display)
  • Is shown in entity cards, backlinks, and search results

Add a ## Summary section at the top of every model page that references the frontmatter:

{/* Replace with actual entity ID from entities.yaml */}
<DataInfoBox entityId="bioweapons-ai-uplift" ratings={frontmatter.ratings} />
## Summary
{frontmatter.description}
## Overview
[Detailed context and background...]

This way the summary text lives only in frontmatter and is rendered on the page via {frontmatter.description}.


The knowledge base serves people making strategic decisions about AI safety:

  • Researchers deciding what to work on
  • Funders deciding where to allocate resources
  • Policymakers deciding what to regulate
  • Organizations deciding their focus areas

Models should help them decide what matters most and what to do about it.

Every model must include:

ElementQuestion AnsweredExample
Magnitude AssessmentHow big is this problem?”This affects 10-30% of total AI risk”
Comparative ImportanceHow does this rank vs. other risks?”Less important than misalignment, more than job displacement”
Resource ImplicationsWhat does this mean for prioritization?”Warrants 5-10% of safety resources”
Key CruxesWhat beliefs would change the conclusion?”If X is true, this becomes top priority”
ActionabilityWhat should actors actually do?”Labs should implement Y, funders should fund Z”

Common Mistake: Mechanism Without Magnitude

Section titled “Common Mistake: Mechanism Without Magnitude”

Bad example (from a hypothetical sycophancy model):

“The feedback loop operates through 4 phases over 10 years, with differential equations governing each variable…”

(300 lines on mechanism, 0 lines on strategic importance)

What’s missing:

  • Is sycophancy a top-5 AI risk or a minor concern?
  • Should safety orgs prioritize this over alignment research?
  • How does this compare to racing dynamics or concentration risks?
  • What beliefs would change whether this matters?

Better approach:

“Sycophancy represents approximately 5-15% of near-term AI risk, ranking below core alignment but above most misuse risks. For most safety organizations, this is a secondary priority unless they have specific comparative advantage. The key crux is whether market competition makes sycophancy inevitable—if so, regulatory intervention becomes critical.”

Include a section like this in every model:

## Strategic Importance
### Magnitude
- **Share of total AI risk:** [X-Y%]
- **Affected population:** [scope]
- **Timeline:** [when effects materialize]
### Comparative Ranking
| Risk Category | Relative Importance | Reasoning |
|---------------|--------------------:|-----------|
| Core alignment | Higher | [why] |
| This risk | Baseline | - |
| [Other risk] | Lower | [why] |
### Resource Implications
- **Who should work on this:** [actor types]
- **Suggested allocation:** [% of resources]
- **Comparative advantage:** [who is best positioned]
### Key Cruxes
1. If [X], this becomes more important because [Y]
2. If [A], this becomes less important because [B]

Write 2-3 paragraphs of flowing prose (no bullet points). The overview should:

  • Explain the model’s central insight in the first paragraph
  • Describe why understanding this matters in the second paragraph
  • Preview key findings or framework structure in the third paragraph
SectionPurposeFormat
OverviewCentral insight and importance2-3 paragraphs of prose
Conceptual FrameworkVisual structure of the modelMermaid diagram + explanation
Quantitative AnalysisNumbers, estimates, projectionsTables with uncertainty ranges
Scenario AnalysisProbability-weighted futures3-5 scenarios with probabilities
LimitationsWhat the model cannot doFlowing prose, specific caveats
Related ModelsConnections to other modelsLinked list

Include tables with 3+ columns and 4+ rows. Tables should add structured information beyond what prose conveys.

Good table characteristics:

  • Uncertainty ranges (low/central/high estimates)
  • Comparison across multiple dimensions
  • Clear headers that explain what each column means
  • Sources or confidence levels where relevant

Models should include at least one diagram. Choose the type that best represents your model’s structure.

Use for: Showing how things lead to other things, causal chains, decision processes.

Loading diagram...
flowchart TD
A[Initial Condition] --> B[Intermediate State]
B --> C[Outcome 1]
B --> D[Outcome 2]

Use for: Showing how multiple factors influence each other, feedback loops, complex interdependencies.

Loading diagram...

Use for: Showing how systems move between discrete states, regime changes, phase transitions.

Loading diagram...

For showing how different actors or systems interact in sequence, attack chains, or response protocols, use a table rather than a sequence diagram (which has rendering issues):

StepActorActionTarget
1Actor AInitial actionSystem
2SystemAlert triggeredDefender
3DefenderResponse deployedSystem
4SystemAction blockedActor A

For simple flows, a basic flowchart works well:

Loading diagram...

Use for: Showing structural relationships between concepts, taxonomies, classification systems.

Loading diagram...

Use for: Classifying items along two dimensions, prioritization frameworks, strategic positioning.

Loading diagram...

Use for: Showing progression over time, milestone projections, historical development.

Loading diagram...

Use for: Showing categories of related items, system boundaries, domain separation.

Loading diagram...

For models describing structural relationships, use schema-style representations.

  • Defining taxonomy of related concepts
  • Showing how entities relate to each other
  • Describing data structures or classification systems
  • Mapping stakeholder relationships
Loading diagram...
Loading diagram...

ConceptStock (Level)Flow (Rate)
TrustCurrent trust level (0-100%)Trust erosion rate (%/year)
CapabilityCurrent capability scoreCapability growth rate
Safety marginCurrent margin sizeMargin compression rate

Many risks involve feedback loops where effects become causes. Make these explicit.

Loading diagram...

Positive feedback loops (amplifying): The effect reinforces the cause. Negative feedback loops (stabilizing): The effect counteracts the cause.

Before modeling specific mechanisms, consider: what’s the base rate for this type of event?

Event TypeHistorical Base RateAI-Specific Adjustment
Major infrastructure failure~0.5/year globallyUnknown multiplier
Technology-driven job displacement~2-5%/decadePotentially 10x faster
Great power conflict~0.5%/yearUnknown effect

When factors co-occur, be explicit about the causal structure:

RelationshipDescriptionImplication
A causes BIntervening on A changes BTarget A to affect B
B causes AIntervening on A doesn’t change BTarget B instead
C causes bothA and B are correlated but independentTarget C to affect both
A and B cause each otherFeedback loopConsider system dynamics

Better approach:

  • Use gradient language: “largely past,” “degrading,” “limited risk”
  • Acknowledge that most systems degrade continuously
  • If using threshold framing, add explicit caveats

Better approach:

  1. Acknowledge correlations explicitly in a table
  2. Use influence diagrams instead of formulas
  3. If using formulas, add caveats about correlation assumptions

Good models should address: “Compared to what?”

ComparisonWhat it reveals
vs. no AI developmentTotal effect of AI
vs. slower developmentEffect of racing
vs. different governanceEffect of policy choices
vs. different actorsEffect of who controls AI

Each model is rated on four dimensions (1-5 scale). These ratings appear in the model’s info box.

How much does this model add beyond existing frameworks?

ScoreDescriptionExample
1Restates common knowledge”AI could be dangerous”
2Minor variation on existing modelAdding one factor to known framework
3Useful synthesis or new framingCombining two existing models in novel way
4Significant new insightNew mechanism or relationship not previously articulated
5Paradigm-shifting frameworkFundamentally new way of understanding the problem

How well-supported and internally consistent is the model?

ScoreDescriptionCharacteristics
1SpeculationNo sources, hand-wavy reasoning
2PlausibleSome logical basis, few sources
3Well-reasonedClear logic, some empirical grounding
4Strong evidence baseMultiple sources, quantified where possible
5Rigorous analysisComprehensive evidence, sensitivity analysis, peer review

How useful is this model for decision-making?

ScoreDescriptionExample outputs
1Abstract only”Things are complex”
2General direction”We should be careful”
3Specific considerations”These 3 factors matter most”
4Concrete recommendations”Prioritize X intervention over Y because Z”
5Decision-readyClear decision criteria, thresholds, action triggers

How thoroughly does the model cover its domain?

ScoreDescriptionMissing elements
1SketchMost components missing
2PartialKey components missing
3AdequateCore model complete, some gaps
4ComprehensiveThorough coverage, minor gaps
5ExhaustiveAll relevant factors, edge cases, interactions

When reviewing a model, ask:

  1. Novelty: “Have I seen this idea before? Does it change how I think?”
  2. Rigor: “Would I bet money on these claims? What’s the evidence quality?”
  3. Actionability: “Could I make a decision based on this? What would I do differently?”
  4. Completeness: “What’s missing? Would adding more change the conclusions?”

Show how conclusions change with different assumptions:

ParameterLow EstimateCentralHigh EstimateConclusion Changes?
Capability growth rate10%/yr30%/yr50%/yrYes - timeline shifts 3-5 years
Alignment difficultyEasyMediumHardYes - risk estimate changes 2-3x
Coordination probability10%30%50%No - conclusion robust

Compare interventions, scenarios, or approaches:

InterventionEffectivenessCostFeasibilityTime to ImpactOverall
Speed limitsHighLowMediumImmediate⭐⭐⭐⭐
International treatyVery HighMediumLow3-5 years⭐⭐⭐
Research fundingMediumMediumHigh5-10 years⭐⭐⭐

Show how a change affects multiple dimensions:

DimensionBefore InterventionAfter InterventionChange
Risk levelHigh (0.7)Medium (0.4)-43%
Detection time2 weeks2 days-86%
Recovery cost$10B$2B-80%

For models with sequential choices:

Loading diagram...

  • Overview is 2-3 paragraphs of flowing prose (no bullets)
  • At least one Mermaid diagram with caption
  • Quantitative tables with 3+ columns and uncertainty ranges
  • Scenario analysis with probability weights
  • Limitations section in prose format
  • Related models linked
  • Diagram type matches content (flowchart for causation, network for relationships, etc.)
  • Diagram has explanatory caption
  • Complex diagrams use subgraphs for grouping
  • Color coding is meaningful and explained
  • No false binary thresholds (or explicitly caveated)
  • Multiplicative formulas acknowledge correlations
  • Feedback loops identified where relevant
  • Stocks vs. flows distinguished
  • Base rates considered
  • Counterfactual comparisons made
  • All four ratings assigned (novelty, rigor, actionability, completeness)
  • Ratings are justified by content quality
  • Ratings are consistent with similar models
IssueFix
Binary threshold languageUse gradient language (“degrading,” “largely past”)
Multiplicative formula without caveatAdd correlation acknowledgment
Missing uncertainty rangesAdd low/central/high estimates
Flowchart for structural relationshipsUse entity-relationship or class diagram
No feedback loops shownAdd arrows showing circular dependencies
Ratings don’t match qualityAdjust ratings or improve content