← Back to incidents

Google Gemini 2.0 Flash Generated Hallucinated Medical Dosage Information

High

Google's Gemini 2.0 Flash model generated dangerous hallucinated medical information including incorrect drug dosages and interactions, prompting concerns from healthcare professionals about patient safety risks.

Category
Medical Error
Industry
Healthcare
Status
Reported
Date Occurred
Jan 1, 2025
Date Reported
Jan 15, 2025
Jurisdiction
US
AI Provider
Google
Model
Gemini 2.0 Flash
Application Type
chatbot
Harm Type
physical
Human Review in Place
No
Litigation Filed
No
medical_aihallucinationdrug_informationpatient_safetygooglegeminihealthcare_misinformation

Full Description

In early January 2025, healthcare professionals began documenting concerning instances of Google's newly released Gemini 2.0 Flash model generating medically inaccurate information when users submitted health-related queries. The incidents came to light when medical practitioners tested the model's responses to common medication questions and discovered significant errors in dosage recommendations and drug interaction warnings. The documented cases revealed that Gemini 2.0 Flash was providing specific medication dosages that contradicted established medical guidelines, including recommending potentially dangerous doses for common medications like blood thinners and diabetes medications. In several instances, the model confidently stated incorrect drug interaction information that could lead to serious adverse effects if patients or healthcare providers relied on the AI's responses without verification. Healthcare professionals expressed particular concern about the authoritative tone with which Gemini presented these incorrect medical facts, noting that the model's confident delivery could make the misinformation more believable to users seeking quick medical information. The errors were especially problematic because they involved specific, actionable medical advice rather than general health information, creating direct pathways to potential patient harm. Google's existing medical disclaimers, which warn users that the AI should not be used for medical advice, proved insufficient to prevent the generation of specific pharmaceutical misinformation. The incident highlighted the ongoing challenge of general-purpose large language models being inappropriately used for specialized domains requiring high accuracy and safety standards. The healthcare community's response included calls for stronger regulatory oversight of AI models that can generate medical information, even when not explicitly designed for healthcare applications. Medical professionals emphasized that while disclaimers may provide legal protection, they do not prevent the generation of dangerous misinformation that could influence patient care decisions. The incident underscored the critical need for specialized safeguards when AI models are accessible to the general public for any type of query, as users may not distinguish between general information requests and those requiring domain-specific expertise and regulatory compliance.

Root Cause

Gemini 2.0 Flash's training data and inference patterns generated medically inaccurate information including incorrect dosages and drug interactions, despite medical disclaimers, due to the model's tendency to hallucinate authoritative-sounding but factually incorrect medical information.

Mitigation Analysis

Implementation of specialized medical knowledge bases with real-time validation, mandatory human expert review for all medical queries, and stronger content filtering specifically for pharmaceutical information could have prevented these dangerous hallucinations. Additional safeguards should include integration with FDA-approved drug databases and automatic redirection of medical queries to qualified healthcare platforms.

Lessons Learned

General-purpose AI models require specialized safeguards for high-risk domains like healthcare, and standard disclaimers are insufficient protection against dangerous medical misinformation. The incident demonstrates the need for proactive content filtering and domain-specific validation systems.

Sources