Skip to content

Meta & Structural Indicators

📋Page Status
Quality:72 (Good)⚠️
Importance:62.5 (Useful)
Words:3.1k
Structure:
📊 1📈 0🔗 56📚 069%Score: 6/15
LLM Summary:Comprehensive tracking of meta-level indicators (information quality, institutional capacity, elite-public gaps, policy response times) across multiple jurisdictions with quantified baselines: 72% US public wants AI development slowed, 29-month EU policy lag from GPT-4 to enforcement, 16-point trust gap between high/low income on AI. Provides essential context for understanding governance capacity but focuses on measurement rather than intervention.

These metrics assess the structural and meta-level conditions that determine society’s ability to navigate AI development safely. Unlike direct capability or safety metrics, these measure the quality of the broader systems—governance institutions, information environments, coordination mechanisms—that mediate AI’s societal impact.

Structural indicators help answer: Is society equipped to handle AI risks? They track whether institutions can make good decisions, whether information environments support informed debate, and whether coordination mechanisms can address collective action problems.

Key distinctions:

  • Direct metrics (capabilities, safety research) → What AI can do and how safe it is
  • Structural metrics (these) → Whether society can govern AI effectively

Many of these metrics are conceptual or partially measured—they represent important dimensions we should track, even if comprehensive data doesn’t yet exist.


Freedom House “Freedom on the Net” Score

  • Latest (2025): United States remains “Free” but with declining scores
  • Concerns about misinformation ahead of 2024 elections contributing to “unreliable information environment”
  • United Kingdom saw decline due to false information leading to riots in summer 2024
  • Interpretation: Score based on internet freedom, content controls, and users’ rights
  • Source: Freedom on the Net 2025

RSF World Press Freedom Index

  • 2025 Global Average: Economic indicator at “unprecedented, critical low”
  • Global press freedom now classified as “difficult situation” for first time in Index history
  • Disinformation prevalence: 138/180 countries report political actors involved in disinformation campaigns
  • 31 countries report “systematic” disinformation involvement
  • 2024 US-specific: Press freedom violations increased to 49 arrests/charges and 80 assaults on journalists (vs 15 and 45 in 2023)
  • Source: RSF World Press Freedom Index 2025

Trust in Institutions (Edelman Trust Barometer 2024)

  • Business: 63% trust (only trusted institution)
  • Government: Low trust, 42% trust government leaders
  • Media: Actively distrusted
  • Mass/Elite divide on AI: 16-point gap in US (43% high-income vs 27% low-income trust AI)
  • Innovation management: 2:1 margin believe innovation is poorly managed
  • Source: 2024 Edelman Trust Barometer

Conceptual Indicators (Limited Direct Measurement)

Section titled “Conceptual Indicators (Limited Direct Measurement)”

AI-Specific Misinformation Prevalence

  • Conceptual metric: % of AI-related claims in public discourse that are false or misleading
  • Proxy data: 62% of voters primarily concerned (vs 21% excited) about AI (AIPI polling)
  • Elite/public gap: “Large disconnect between elite discourse and what American public wants” - AI Policy Institute
  • Challenge: No systematic tracking of AI misinformation rates
  • Source: AI Policy Institute Polling

World Bank Worldwide Governance Indicators (WGI)

  • Government Effectiveness dimension: Quality of public services, bureaucracy competence, civil service independence, policy credibility
  • Scale: -2.5 to +2.5 (normalized with mean ~0), also mapped to 0-100 scale
  • Latest: 2024 methodology update covering 214 economies, 1996-2023 data
  • Data sources: 35 cross-country sources including household surveys, firm surveys, expert assessments
  • Limitation: “Inputs” focused (institutional capacity) rather than “outputs” (decision quality)
  • Source: World Bank WGI 2024

V-Dem Digital Society Index

  • Coverage: Measures government internet censorship, social media monitoring, online media fractionalization
  • Note: 2024 specific data on information environment not retrieved, but framework exists
  • Source: V-Dem Institute (v-dem.net)

AI Policy Quality Index

  • Conceptual metric: Expert assessment of whether AI policies address actual risks proportionately
  • Current status: No standardized index exists
  • Proxy: Mixed signals—EU AI Act implemented, US executive order, but critiques of regulatory lag

Evidence-Based Policy Rate for AI

  • Conceptual metric: % of major AI policy decisions informed by rigorous evidence
  • Challenge: Would require systematic policy analysis across jurisdictions
  • Current: Anecdotal evidence suggests variable quality

3. Elite vs Public Opinion Divergence on AI

Section titled “3. Elite vs Public Opinion Divergence on AI”

Expert vs Public Trust Gap (Pew Research 2024)

  • Finding: “Experts are far more positive and enthusiastic about AI than the public”
  • Methodology: 5,410 US adults (Aug 2024) vs 1,013 AI experts (Aug-Oct 2024)
  • Experts: Identified via authors/presenters at 21 AI conferences in 2023-2024
  • Source: Pew Research: Public and AI Experts

AI Policy Institute Polling (2024)

  • Development pace preference: 72% prefer slowing AI development vs 8% prefer speeding up
  • Risk vs excitement: 62% primarily concerned vs 21% primarily excited
  • Catastrophic risk belief: 86% believe AI could accidentally cause catastrophic event
  • Liability: 73% believe AI companies should be held liable for harm
  • Regulation preference: 67% think AI models’ power should be restricted
  • Elite disconnect quote: “Large disconnect between elite discourse or discourse in labs and what American public wants” - Daniel Colson, AIPI Executive Director
  • Source: AIPI Polling

Trust Gap in AI Companies (Edelman 2024)

  • Technology sector vs AI innovation: 26-point gap (76% trust tech sector vs 50% trust AI)
  • AI company trust decline: From 62% (5 years ago) to 54% (2024)
  • Rejection willingness: 43% will actively reject AI products if innovation poorly managed
  • Source: Edelman Trust Barometer 2024 - AI Insights

Magnitude: Large and growing gap between expert optimism and public concern Direction: Public more risk-focused; experts more capability-focused Policy implication: Democratic deficit if AI development primarily shaped by technologists


4. Time from AI Risk Identification to Policy Response

Section titled “4. Time from AI Risk Identification to Policy Response”

EU AI Act Timeline (Response to GPT-class models)

  • GPT-3 release: June 2020
  • EU AI Act proposal: April 2021 (10 months after GPT-3)
  • GPT-4 release: March 2023
  • EU AI Act agreement: December 2023 (9 months after GPT-4)
  • AI Act signed: June 2024
  • Entered force: August 2024
  • GPAI provisions applicable: August 2025 (29 months after GPT-4)
  • Full applicability: August 2026
  • Interpretation: ~2.5 years from GPT-4 to enforceable rules on GPAI models
  • Source: EU AI Act Implementation Timeline

US Executive Order on AI

  • GPT-4 release: March 2023
  • Executive Order 14110: October 30, 2023 (7 months after GPT-4)
  • Limitation: Executive order, not legislation; limited enforceability
  • Source: Biden Administration AI Executive Order

AI Safety Institutes

  • UK AISI announced: November 2023 (Bletchley Park AI Safety Summit)
  • US AISI operational: Early 2024
  • AISI Network launched: May 2024 (Seoul AI Summit)
  • First AISI Network meeting: November 2024 (San Francisco)
  • Lag interpretation: ~8-20 months from GPT-4 to safety institute operations
  • Source: AISI International Network

Average Policy Lag Time

  • Conceptual metric: Median time from risk becoming evident to enforceable policy
  • Challenge: Defining “risk becomes evident” vs “risk exists”
  • Current estimate: 1-3 years for major risks based on available cases
  • Comparison: Aviation safety regulations often follow major accidents within months

5. Coordination Failure Rate on AI Governance

Section titled “5. Coordination Failure Rate on AI Governance”

G7 Hiroshima AI Process Code of Conduct

  • Status: Adopted but “provides little guidance” on implementation
  • Critique: “Staffed by diplomats who lack depth of in-house technical expertise”
  • Implementation gap: Code instructs to “identify, evaluate, mitigate risks” without how-to guidance
  • Source: CSIS: G7 Hiroshima AI Process

OECD AI Principles (2019, updated 2024)

  • Adherents: 47 countries including EU
  • Compliance mechanism: None (non-binding)
  • Monitoring: AI Policy Observatory tracks implementation but no enforcement
  • Implementation rate: Variable—no systematic tracking of adherence
  • Source: OECD AI Principles 2024 Update

International AI Safety Institute Network

  • Members (Nov 2024): 10 countries/regions (Australia, Canada, EU, France, Japan, Kenya, Korea, Singapore, UK, US)
  • Challenges identified:
    • Confidentiality and security concerns
    • Legal incompatibilities between national mandates
    • Varying technical capacities
    • Global South institutes risk becoming “token members”
    • Most institutes still hiring/setting priorities as of 2024
  • Coordination body: None yet (recommended but not established)
  • Success metric: Too early to assess
  • Source: AISI Network Analysis

Coordination Success Rate

  • Conceptual metric: % of identified coordination problems that achieve multilateral solutions
  • Current status: Low coordination success on binding agreements
  • Examples of failure:
    • No binding international compute governance
    • No global model registry
    • Fragmented incident reporting systems
    • Limited cross-border enforcement
  • Examples of partial success:
    • AISI Network formation
    • OECD Principles (soft coordination)
    • G7/G20 discussions ongoing

Race-to-the-Bottom Index

  • Conceptual metric: Evidence of jurisdictions weakening standards to attract AI companies
  • Current: Anecdotal concerns but no systematic measurement
  • Source: International Governance of AI

6. Democratic vs Authoritarian AI Adoption Rates

Section titled “6. Democratic vs Authoritarian AI Adoption Rates”

AI Surveillance Adoption

  • China’s market dominance: Exports AI surveillance to “nearly twice as many countries as United States”
  • Chinese surveillance camera market: Hikvision + Dahua = 34% global market share (2024)
  • Global reach: PRC-sourced AI surveillance in 80+ countries (authoritarian and democratic)
  • China’s domestic deployment: Over half the world’s 1 billion surveillance cameras located in China
  • Source: Global Expansion of AI Surveillance

Export Patterns

  • China’s bias: “Significant bias in exporting to autocratic regimes”
  • Huawei “Safe City” agreements (2009-2018): 70%+ involved countries rated “partly free” or “not free” by Freedom House
  • Nuance: “China is exporting surveillance tech to liberal democracies as much as targeting authoritarian markets”
  • Impact finding: Mature democracies did not experience erosion when importing surveillance AI; weak democracies exhibited backsliding regardless of supplier
  • Source: Data-Centric Authoritarianism

Authoritarian Advantage Factors

  • China’s structural advantages for AI surveillance:
    • Lax data privacy laws
    • Government involvement in production/research
    • Large population for training data
    • Societal acceptance of state surveillance
    • Strong AI industrial sectors
  • Source: AI and Authoritarian Governments

Democratic vs Authoritarian AI Capability Gap

  • Conceptual metric: Relative AI capability development in democracies vs autocracies
  • Proxy: US vs China capability race
    • US: 40 notable AI models (2024) vs China: 15 models
    • US private investment: $109.1B vs China: $9.3B
    • But China’s DeepSeek/Qwen/Kimi “closing the gap on reasoning and coding”
  • Interpretation: US maintains edge but China rapidly improving
  • Source: State of AI Report 2025

7. Concentration of AI Capability (Herfindahl Index)

Section titled “7. Concentration of AI Capability (Herfindahl Index)”

Enterprise LLM Market Share (2024-2025)

  • Anthropic: 32% usage share, 40% revenue share
  • OpenAI: 25% usage share, 27% revenue share (down from 50% in 2023)
  • Google: 20% usage share
  • Meta (Llama): 9%
  • DeepSeek: 1%
  • Approximate HHI: ~2,500 (0.32² + 0.25² + 0.20² + 0.09² + 0.01²) × 10,000 ≈ 2,050-2,500
  • Interpretation: “Moderate concentration” (HHI 1,500-2,500); top 3 control ~77%
  • Source: 2025 State of Generative AI in Enterprise - Menlo Ventures

Frontier Model Development Concentration

  • US dominance: 40 notable models (2024) vs China: 15, Europe: 3
  • Competition assessment: “OpenAI retains narrow lead at frontier, but competition intensified”
  • China status: “Credible #2” with DeepSeek, Qwen, Kimi
  • Source: Stanford AI Index 2025

Investment/Funding Concentration

  • Foundation model funding (2025): $80B (40% of all global AI funding)
  • OpenAI + Anthropic: 14% of all global venture investment across all sectors
  • Big Tech backing: “Interconnected web of 90+ partnerships” among Google, Apple, Microsoft, Meta, Amazon, Nvidia
  • Regulatory concern: UK CMA and US FTC investigating concentration via partnerships/investments
  • Source: Big Tech’s Cloud Oligopoly

Compute Concentration

  • Conceptual metric: HHI for GPU/training compute access
  • Challenge: Private compute capacity not publicly reported
  • Known: Nvidia dominance in AI chips; hyperscaler concentration (AWS, Azure, GCP)
  • Implication: Capability concentration may exceed market share concentration

Talent Concentration

  • Conceptual metric: % of top AI researchers at small number of organizations
  • Challenge: Defining “top researchers” and tracking mobility
  • Proxy: Conference authorship concentration, hiring trends

WEF Global Risks Report 2024 - Resilience Assessment

  • Key finding: “Weakened economies and societies may only require smallest shock to edge past tipping point of resilience”
  • Current crises eroding resilience: COVID-19 aftermath, Russia-Ukraine war “exposed cracks in societies”
  • Long-term erosion: “Decades of investment in human development slowly being chipped away”
  • Conflict risk: “Corroding societal resilience risk creating conflict contagion”
  • Source: WEF Global Risks Report 2024

Economic Disruption Preparedness

  • Social safety nets: Vary widely by country (unemployment insurance, retraining programs)
  • Financial instruments: Insurance, catastrophe bonds, public risk pools
  • Challenge: No unified “AI disruption resilience” score exists

Digital Literacy and Misinformation Resilience

  • Recommendation: “Digital literacy campaigns on misinformation and disinformation”
  • Current: No systematic measurement of population-level AI/digital literacy
  • Proxy: General digital skills indices exist but not AI-specific

Institutional Adaptive Capacity

  • Indicators: R&D investment in climate modeling/energy transition (analogous to AI preparedness)
  • Infrastructure resilience: Building codes, disaster preparedness
  • Limitation: No AI-specific resilience metrics

Labor Market Adaptability Index

  • Conceptual metric: How quickly workers can reskill/transition as AI automates tasks
  • Proxy data: Historical adjustment rates to automation, education system responsiveness
  • Challenge: AI may disrupt faster than historical automation

Democratic Resilience to AI-Driven Polarization

  • Conceptual metric: Ability of democratic institutions to function under AI-amplified disinformation
  • Current concerns: Misinformation in 2024 elections (US, UK)
  • No systematic tracking: Would require longitudinal study

AI Incident Database (AIID)

  • Total incidents: 2,000+ documented incidents (as of 2024)
  • Coverage: “Intelligent systems causing safety, fairness, or other real-world problems”
  • Growth: From 1,200+ reports to 2,000+ (rapid increase)
  • Limitation: Voluntary reporting, variable severity, unclear baseline
  • Source: AI Incident Database

AIAAIC Repository

  • Start date: June 2019
  • Coverage: “Incidents and controversies driven by AI, algorithms, automation”
  • Goal: “Systematically documenting incidents where AI systems cause or contribute to harms”
  • Scope: Broader than AIID—includes technical failures and social impacts
  • Source: AIAAIC Repository

OECD AI Incidents Monitor (AIM)

  • Launch: Part of OECD AI Policy Observatory
  • Focus: Policy-relevant cases aligned with governance interests
  • Collaboration: Partnership on AI, Center for Advancement of Trustworthy AI
  • Limitation: More selective than AIAAIC (policy focus vs comprehensive coverage)
  • Source: OECD AIM

Incident Rate per AI System

  • Conceptual metric: Incidents per 1,000 or 10,000 deployed AI systems
  • Challenge: Unknown denominator—no comprehensive count of deployed systems
  • Current: Absolute incident counts rising, but unclear if rate rising

Severity Distribution

  • Available: Incident databases categorize by harm type (safety, fairness, rights)
  • Missing: Standardized severity scales across databases
  • Incompatibility: “Both databases have vastly different and incompatible structures”
  • Source: Standardised Schema for AI Incident Databases

Baseline Comparison

  • Question: Are AI incident rates high compared to other technologies at similar maturity?
  • Challenge: No established baseline or reference class
  • Aviation analogy: Aviation incident rates well-tracked, declining over time—AI lacks comparable infrastructure

AI Lab Support for Near-Miss Reporting

  • Strong agreement: 76% strongly agree, 20% somewhat agree
  • Statement: “AGI labs should report accidents and near misses to appropriate state actors and other AGI labs”
  • Source mechanism: AI incident database
  • Source: EA Forum: Incident Reporting for AI Safety

US Executive Order 14110

  • Provision: Addressed “safety” and “rights” protections
  • Limitation: Not comprehensive near-miss framework
  • State-level: New York State bill would require incident reporting to Attorney General (safety incidents only)
  • Source: Designing Incident Reporting Systems

EU AI Act Incident Reporting

  • Requirement: Single incident reporting requirement
  • Definition: Includes both “rights incidents” and “safety incidents”
  • Limitation: Does not explicitly distinguish near-misses from harms
  • Source: EU AI Act

Proposed Framework Properties (Shrishak 2023)

  1. Voluntary reporting: Essential for capturing near-misses not covered by mandatory serious incident reporting
  2. Non-punitive: Consensus that self-reporting should not lead to punishment since no harm occurred
  3. Accessible: Low barriers to submission
  4. Actionable: Information useful for other developers

Actual Near-Miss Reporting Rate

  • Conceptual metric: % of near-miss events that get reported to databases or regulators
  • Current estimate: Unknown, likely very low
  • Challenge: “Current systems fail to capture numerous near-miss incidents that narrowly avoid accidents”
  • Comparison: Aviation near-miss reporting well-established; AI has no equivalent system yet
  • Source: Developing Near-Miss Reporting System

Culture Gap

  • Aviation standard: Open, non-punitive reporting is norm
  • AI current state: “Lack of comprehensive and reliable data regarding frequency and characteristics”
  • Needed shift: “Building culture of safety for AI requires understanding failure modes, which starts with reporting past incidents”

MetricStatusData QualityUpdate Frequency
Information environment qualityMeasuredHigh (Freedom House, RSF)Annual
Institutional decision-makingProxyMedium (WGI covers general governance, not AI-specific)Annual
Elite/public opinion divergenceMeasuredMedium (multiple polls, varying methods)Quarterly-Annual
Policy response timeMeasuredHigh (specific cases documented)Case-by-case
Coordination failure rateConceptualLow (qualitative assessments only)Ad hoc
Democratic vs authoritarian adoptionMeasuredMedium (surveillance tech tracked, general AI capabilities less clear)Annual
AI capability concentration (HHI)MeasuredMedium (market share known, compute concentration estimated)Quarterly-Annual
Societal resilienceConceptualLow (framework exists, no AI-specific index)Annual (WEF)
AI incident rateMeasuredMedium (absolute counts good, rates unclear due to denominator problem)Continuous
Near-miss reporting rateConceptualVery low (frameworks proposed, actual reporting minimal)Not measured

  1. Denominator problems: Incident rates require knowing # of deployed systems (unknown)
  2. Counterfactuals: Measuring “coordination failure rate” requires knowing what coordination was possible
  3. Lag indicators: Most metrics (incidents, trust, governance quality) are lagging, not leading
  4. Attribution: Hard to isolate AI’s contribution to institutional quality or societal resilience
  5. Standardization: Different databases use incompatible schemas (incidents, governance)
  1. No unified resilience metric: Individual components exist but no composite “AI disruption resilience score”
  2. Weak coordination metrics: Qualitative assessments dominate; no quantitative coordination success rate
  3. Missing baselines: Few comparisons to other technologies at similar development stages
  4. Democratic processes: No metrics for how democratic institutions specifically handle AI (vs general governance)

High-value additions:

  • Standardized AI incident severity scale
  • Near-miss reporting infrastructure and culture-building
  • Democratic resilience to AI-specific challenges (not just general governance)
  • Coordination success metrics (track multilateral agreements, implementation rates)
  • AI-specific institutional capacity assessment (beyond general WGI)

For Risk Assessment:

  • Low trust + weak institutions + high elite/public gap = governance failure more likely
  • Rising incidents + low near-miss reporting = learning from failures inadequate
  • High concentration + weak coordination = race dynamics and power concentration risks

For Forecasting:

  • Policy lag times (1-3 years) inform timeline expectations for future risks
  • Trust trends predict regulatory pressure and public backlash likelihood
  • Coordination challenges suggest multilateral solutions face high barriers

For Intervention:

  • Improving near-miss reporting culture = high-leverage, low-cost
  • Building institutional AI literacy = addresses decision-making quality
  • Bridging elite/public gap = essential for democratic legitimacy
  1. Correlation ≠ causation: Weak governance may cause AI risks OR AI risks may weaken governance
  2. Selection effects: Reported incidents overrepresent visible, Western, English-language cases
  3. Gaming: Once metrics are targets, they can be manipulated (Goodhart’s Law)
  4. Aggregation: Composite indices hide important variation across dimensions