Skip to content

Public Opinion & Awareness

📋Page Status
Quality:72 (Good)
Importance:42.5 (Reference)
Words:2.6k
Backlinks:2
Structure:
📊 3📈 0🔗 14📚 064%Score: 8/15
LLM Summary:Comprehensive tracking of public AI risk awareness (12.3% mention x-risk unprompted in 2023, up 78% from 2022) and concern (50% more concerned than excited in 2025, up from 37% in 2021), with data tables comparing public vs. expert views and trust levels across institutions. While awareness is rising rapidly, specific x-risk understanding remains low despite 59% supporting extinction risk mitigation as priority.

Public opinion shapes the political feasibility of AI governance and influences the trajectory of AI development through democratic pressure, consumer behavior, and social norms. This page tracks key metrics on public awareness, concern, trust, and literacy regarding AI risks and capabilities.

Key Finding: While AI awareness is nearly universal (95%+ in the US), specific awareness of existential risk remains low (~12% mention it unprompted), though general concern is rising rapidly (from 37% in 2021 to 50% in 2025).


1. Population Aware of AI Existential Risk

Section titled “1. Population Aware of AI Existential Risk”
  • 12.3% of US adults mention AI as a potential cause of human extinction (unprompted)

    • Source: Rethink Priorities survey, April 2023 (n=300)
    • +78% increase from 6.9% in December 2022
    • Growth attributed to ChatGPT media coverage surge
  • 43% are very/somewhat concerned about AI causing the end of humanity

    • Source: YouGov, June 2025
    • 16% very concerned, 27% somewhat concerned
    • Up from 37% in March 2025
  • 59% of US adults support prioritizing mitigating extinction risk from AI

    • Source: Rethink Priorities online poll, 2023
    • 26% disagree
    • Among those who disagree: 36% cite “other priorities,” 23% say “not extinction,” 18% say “not yet”
  • 4% select AI as the most likely cause of human extinction
    • Ranks below nuclear war (42%), climate change, pandemics, and asteroids
    • Source: Rethink Priorities nationally-representative survey
  • AI researchers estimate median 5% chance of human extinction from AI
    • Source: AI Impacts survey of 2,778 researchers, 2023
    • ~40% of researchers indicate >10% chance of catastrophic outcomes
    • Unchanged from 2022 survey (same 5% median)

2. Population Concerned About AI Safety (General)

Section titled “2. Population Concerned About AI Safety (General)”
  • 50% of Americans are more concerned than excited about AI (2025)

    • Source: Pew Research, survey of 5,023 adults, June 9-15, 2025
    • +35% increase from 37% in 2021
    • Only 11% are more excited than concerned
  • 57% rate societal risks of AI as high

    • Source: Pew Research, August 2024 (n=5,410)
    • vs. 25% who rate benefits as high
  • 47% believe AI effects on society will be negative

    • Source: YouGov, June 2025
    • +38% increase from 34% in December 2024
    • Steady upward trend: 34% → 40% → 41% → 47% over 6 months
  • 31% of Americans trust AI (2% fully, 29% somewhat)

    • Source: Gallup, 2025 (n=3,128)
    • 60% do not trust AI to make fair, unbiased decisions
    • Only 8% consider themselves “very knowledgeable” about AI
  • Trust is declining: 25% say trust decreased in past year

    • vs. 21% increased, 47% unchanged
    • Source: YouGov, 2025
  • 18% would trust AI to make decisions/take actions

    • vs. 53% would not
    • Even among Gen Z: 43% mistrustful vs. 26% trustful

Specific Concerns (US Public vs. AI Experts)

Section titled “Specific Concerns (US Public vs. AI Experts)”
ConcernPublicAI Experts
Less human connection57%37%
Job loss56%25%
More concerned than excited51%15%
Negative impact next 20 years83%44%

Source: Pew Research, August 2024

Most concerned countries (>50% more concerned than excited):

  • United States
  • Italy
  • Australia
  • Brazil
  • Greece

Least concerned:

  • South Korea (16%)
  • China (72% trust AI - highest globally)

Source: Pew Research Global Survey, October 2025


  • 38,787 articles on AI from 12 major English newspapers (2010-2023)

    • Source: Ittefaq et al. (2024), analysis of 12-country media coverage
    • Dramatic increase in recent years, especially post-ChatGPT
  • 24,827+ articles on generative AI specifically (Jan 2018 - Nov 2023)

    • Source: Landscape of Generative AI in Global News (2024)
    • Sharp spike after ChatGPT launch (November 2022)
  • 2023: 39 publications
  • 2024: 106 publications (+172% year-over-year)
  • 2020-2022: ~20-25 publications/year (plateau)

Source: Systematic bibliometric analysis (2016-2024)

  • 21% negative coverage
  • 13% positive coverage
  • 66% neutral coverage

Source: 12-country analysis (2010-2023)

  • UK and US media have become more critical over time
  • Progressive outlets (NYT, The Guardian) “going slightly negative about AI each year”
  • Increased references to risks and concerns

  • +250% year-over-year increase in “artificial intelligence” searches (UK, 2023)

    • Source: Think with Google, 2023
  • Peak interest: October 2023 for “AI-powered search”

    • Stabilized in first half of 2024
    • Source: Statista, Google Trends data
  • Peak: Week ending March 3, 2024 (score: 100)
    • Surge from mid-February to early March 2024
    • Source: Statista, Google Trends
  • Searches evolving from curiositypractical application
  • “How to AI” queries growing faster than basic informational queries
  • Hands-on queries dominating over definitional searches
  • #1 globally: “Gemini” (Google AI assistant)
  • AI-generated content: Barbie AI, action figures, Ghibli-style art
  • Indicates mainstream adoption of AI tools

Source: Google Year in Search 2025


  • 79% of Americans don’t trust companies to use AI responsibly

    • Source: Bentley-Gallup Survey, 2025
  • 47% globally confident AI companies protect personal data (2024)

    • Down from 50% in 2023 (-6%)
    • Source: Ipsos, cited in Stanford AI Index 2025
InstitutionTrust Level
Employers71%
Universities67%
Large tech companies61%
Start-ups51%

Source: McKinsey US employee survey, Oct-Nov 2024

  • Global average: Only 46% willing to trust AI systems

    • Source: KPMG Global AI Trust Study (n=48,000 across 47 countries), 2025
  • Advanced economies: Trust drops to 39%

    • Trust declining in wealthy nations, rising in emerging economies
  • Perceived trustworthiness: 63% → 56% (-11%)
  • Willingness to rely on AI: 52% → 43% (-17%)
  • Worried about AI: 49% → 62% (+27%)

Source: Global trust surveys (2022-2024)

  • Highest trust: China (72%), India, Nigeria
  • Lowest trust: US (32%), Australia, Ireland, Netherlands (under 33%)

Source: Edelman Trust Barometer 2025

  • AI users: 46% trust AI
  • Non-users: 23% trust AI
  • Trust doubles with usage

Source: Gallup, 2025


  • 62% of US public have little/no confidence in government to regulate AI
    • Source: Pew Research, August 2024 (n=5,410)
    • 53% of AI experts also lack confidence
  • 80% say government should maintain safety rules even if AI develops more slowly

    • Source: Gallup/SCSP survey, 2025 (n=3,128)
    • Only 9% prioritize speed over safety
    • Bipartisan support: 88% Democrats, 79% Republicans/Independents
  • 71% believe regulation is needed (41% “much more”, 30% “somewhat more”)

    • Source: YouGov, 2025
    • Up from 64% in December 2024 (+11%)
  • 70% globally believe national/international AI regulation is needed

    • Source: KPMG Global AI Trust Study, 2025
  • 29% of US consumers believe current regulations are sufficient
    • 72% say more regulation needed
    • 81% would trust AI more if laws/policies were in place
  • ~50% of Americans agree government should have major role in AI oversight (end of 2024)
    • Source: Ipsos/Stanford AI Index, 2024

  • 69% support a 6-month pause on some AI development
    • 41% strongly support
    • 28% somewhat support
    • 13% oppose (4% strongly, 9% somewhat)
    • 18% not sure

Source: YouGov poll, April 3, 2023 (n=20,810 US adults)

Support for Government-Enforced Moratorium

Section titled “Support for Government-Enforced Moratorium”
  • Average 73% support (Yes + Maybe)

    • 39% definite “Yes” (average)
    • Peak support after media intervention: 54% (CNBC survey), 54% (CNN survey)
  • Government enforcement: Average 69% support (Yes + Maybe)

    • 35% definite “Yes”
    • Peak: 52% (CNBC), 44% (CNN)

Source: EA Forum study, April 2023 (n=300, multiple survey conditions)

Context: Future of Life Institute Open Letter

Section titled “Context: Future of Life Institute Open Letter”
  • March 2023: Open letter calling for 6-month pause on systems >GPT-4
  • 30,000+ signatures including Yoshua Bengio, Stuart Russell, Elon Musk
  • Impact: Generated “renewed urgency within governments,” normalized expressing AI fears
  • Reality: No pause occurred; investments in large models continued

  • 67% globally say they have “good understanding” of AI (2024)
    • Source: Ipsos AI Monitor, 32-country survey
GenerationGood Understanding
Gen Z72%
Millennials71%
Gen X~60-65% (est.)
Baby Boomers58%

Source: Ipsos AI Monitor 2024

  • 98% have heard about AI
  • 39% report using AI
  • BUT: When asked about 6 common AI products, 99% have used at least one
    • 83% used 4+ AI products
    • Indicates severe awareness gap about what counts as AI

Source: Gallup, 2025

  • 34.13% highly familiar with AI

  • 50.08% somewhat familiar

  • 15.79% know nothing about AI

  • Among those “familiar”:

    • Only 13.73% actually understand AI processes
    • 57.8% have some understanding
    • 28.47% just know the term

Source: Survey of 800 Americans, 2024

University Students (4 Asian/African nations)

Section titled “University Students (4 Asian/African nations)”
  • Average AI literacy: 2.98 out of 5 (moderate skill level)
  • Significant disparities by:
    • Nationality
    • Field of study (technical > non-technical)
    • Academic degree level
  • No significant difference by:
    • Gender
    • Age

Source: Comparative transnational survey (n=1,800), 2024

  • Persistent global gender gap in GenAI usage
  • Male students show more trust in AI than female students (UK & Poland study, 2024)
  • Younger adults (under 45) more open: 36% expect positive impact vs. 19% of 45+
  • Rural populations perceive higher privacy/safety risks
  • Urban populations more accepting of AI
  • Functional reliability concerns differ significantly
  • 58% of students don’t feel they have sufficient AI knowledge/skills
  • 48% don’t feel prepared for AI-enabled workplace
  • Despite high usage rates

Source: DEC Global AI Student Survey 2024


9. Accuracy of Public Beliefs About AI Capabilities

Section titled “9. Accuracy of Public Beliefs About AI Capabilities”

Top misconceptions identified (Survey of 800 Americans, 2024):

  1. “AI is fully autonomous and self-learning” (50% need clarification)
  2. “AI makes decisions without any errors” (50% need clarification)
  • Public tends to both overestimate and underestimate AI capabilities
  • Overestimate: Autonomy, reasoning, consciousness
  • Underestimate: Current practical capabilities, scope of existing AI use
  • Awareness gap: 39% report using AI, but 99% actually use AI-enabled products
  • 61% unaware they use AI regularly
  • Common products not recognized as AI:
    • Navigation apps
    • Streaming recommendation engines
    • Social media algorithms
    • Weather forecasting
    • Online shopping personalization

Source: Gallup, 2025

  • Experts 3x more optimistic about AI impact (56% vs. 17%)
  • Experts 4x more excited than public (47% vs. 11%)
  • Largest perception gaps:
    • Long-term societal impact
    • Job market effects
    • Loss of human connection

Source: Pew Research, August 2024

  • 83.4% of Americans concerned about AI spreading misinformation in 2024 election
    • Source: US public opinion survey, August 2023

What Americans want to learn more about (2024 survey):

  1. 57.05%: Accuracy of AI-generated results
  2. 56.96%: Data security when using AI
  3. 48.22%: How AI makes decisions
  • Limited research specifically quantifying accuracy of public beliefs
  • Most studies focus on awareness and attitudes, not correctness
  • Significant need for systematic assessment of belief accuracy

  • Concern increased 35% in 4 years (37% → 50%)
  • Acceleration in 2024-2025: 34% → 47% in 6 months
  • ChatGPT (Nov 2022) identified as major inflection point

2. High Pause Support but Low X-Risk Awareness

Section titled “2. High Pause Support but Low X-Risk Awareness”
  • 69% support development pause
  • Only 12% mention existential risk unprompted
  • Suggests concern is about near-term harms, not extinction
  • Trust in AI systems: declining
  • Trust in companies: declining (50% → 47%)
  • Trust in government to regulate: low (62% lack confidence)
  • But: Experience builds trust (46% vs. 23%)
  • 99% use AI products
  • 39% think they use AI
  • 60-point awareness gap about everyday AI
  • 3x gap in optimism about long-term impact
  • 4x gap in excitement vs. concern
  • Suggests communication challenge for AI safety advocates
  • Emerging markets: High trust, high optimism (China 72% trust)
  • Advanced economies: Low trust, high concern (US 32% trust)
  • 45-point trust gap between China and US
  • 70-80% want government to prioritize safety over speed
  • Bipartisan consensus (88% Dems, 79% Reps)
  • But low confidence government can deliver (62% skeptical)

  • “AI” vs. “artificial intelligence” vs. “machine learning” elicits different responses
  • “Existential risk” vs. “very bad outcomes” vs. “human extinction” varies widely
  • Media exposure immediately before survey significantly affects responses
  • People don’t recognize AI in everyday products
  • Self-reported usage dramatically underestimates actual usage
  • Complicates measuring “literacy” vs. “awareness”
  • Opinions shift rapidly with news cycles
  • ChatGPT caused immediate awareness spike
  • Media interventions show immediate effect (EA Forum study)
  • Online panels vs. representative samples
  • US-centric data (most surveys)
  • Limited longitudinal tracking with consistent methodology
  • Does experience increase trust, or do trusting people seek experience?
  • Does media coverage increase concern, or does concern drive coverage?
  • Difficult to establish causal mechanisms



December 24, 2025

Note: This page synthesizes data from multiple surveys conducted 2023-2025. Survey methodologies, sample sizes, and question wordings vary significantly. Numbers should be interpreted as indicative trends rather than precise measurements. For specific use cases, consult original sources for methodology details.