← Back to incidents
Google AI Overviews Generated Dangerous Health Advice from Reddit Satirical Posts
HighGoogle's AI Overviews feature generated dangerous health advice including eating rocks and using glue on pizza, sourcing information from satirical Reddit posts without quality filtering.
Category
Hallucination
Industry
Technology
Status
Ongoing
Date Occurred
May 15, 2024
Date Reported
May 23, 2024
Jurisdiction
US
AI Provider
Google
Model
Gemini
Application Type
embedded
Harm Type
physical
Human Review in Place
No
Litigation Filed
No
searchgooglegeminihealth_misinformationredditsatirical_contentsource_quality
Full Description
In May 2024, Google launched AI Overviews at its annual I/O developer conference as a new search feature that provides AI-generated summaries at the top of search results. The feature, powered by Google's Gemini AI model, was designed to give users quick answers to their queries by synthesizing information from across the web. Shortly after the public rollout on May 15, 2024, users began sharing screenshots of bizarre and potentially dangerous AI-generated responses on social media platforms. The incident gained widespread attention when multiple users documented the AI providing harmful health advice, including recommendations to eat rocks for minerals and to use glue as a pizza topping adhesive.
The technical failure occurred within Google's AI Overviews system, which uses the Gemini large language model to process and synthesize web content into concise summaries. The system's information retrieval and ranking algorithms failed to properly evaluate source credibility, treating satirical Reddit posts with the same authority as legitimate health resources. Specifically, the AI cited a years-old satirical Reddit comment suggesting people should "eat at least one small rock per day" for geological minerals, and another joke post recommending non-toxic glue to prevent pizza cheese from sliding. The model's training on diverse internet content without sufficient quality filtering mechanisms allowed it to mistake obvious satire and parody for factual information, particularly problematic given the health and safety implications of such advice.
The incident created significant reputational damage for Google as screenshots of the dangerous advice went viral across Twitter, Reddit, and other social media platforms within hours of discovery. While no specific cases of physical harm from following the AI's advice were documented, health experts and AI safety researchers expressed serious concerns about the potential for users, particularly children or vulnerable populations, to act on such recommendations. The incident amplified existing criticisms about the rushed deployment of AI features in search results and raised questions about Google's content moderation and fact-checking processes. Media coverage was extensive, with major news outlets highlighting the failure as an example of AI systems' inability to distinguish between credible and satirical sources.
Google responded within 24 hours by acknowledging the issues and implementing technical fixes to reduce such problematic responses. The company issued statements explaining that the examples represented uncommon edge-case queries and that the vast majority of AI Overviews functioned as intended. Google engineers deployed updates to improve the system's ability to identify satirical content and implemented additional quality filters for health-related queries. The company also adjusted the feature's sensitivity settings to reduce the likelihood of generating responses based on low-quality or obviously satirical sources, though specific technical details of these fixes were not disclosed publicly.
The incident highlighted broader industry challenges regarding AI safety and the deployment of large language models in consumer-facing applications. It demonstrated the difficulty of training AI systems to navigate the nuanced differences between factual information and satirical content, particularly when such content appears on platforms like Reddit where context and tone can be ambiguous. The failure raised questions about the adequacy of testing protocols for AI features before public release and the need for more robust content filtering mechanisms when dealing with health and safety information.
This incident contributed to ongoing policy discussions about AI governance and the responsibility of technology companies to ensure their AI systems do not provide harmful advice. It served as a case study for AI safety researchers examining the challenges of source verification and content quality assessment in large-scale AI systems, particularly those designed to synthesize information from diverse and unvetted internet sources.
Root Cause
The AI system failed to distinguish between satirical Reddit posts and legitimate health advice, treating joke responses as authoritative sources when generating search result summaries.
Mitigation Analysis
Source quality filtering and satirical content detection could have prevented this incident. Human review of health-related AI responses before publication would have caught the obviously dangerous advice. Domain-specific training data curation excluding satirical platforms for medical queries would reduce similar risks.
Lessons Learned
The incident demonstrates the critical importance of source quality assessment and content classification in AI systems that provide health or safety advice to the public.
Sources
Google's AI Overview is flubbing some search results
The Verge · May 24, 2024 · news
Google pauses Gemini AI after it produces inaccurate historical images
BBC News · May 23, 2024 · news