← Back to incidents
AI Medical Imaging Tools Miss Cancerous Tumors in Clinical Practice
HighFDA-approved AI radiology tools demonstrated significantly lower accuracy in detecting cancerous tumors when deployed in real clinical settings compared to controlled trial environments.
Category
Medical Error
Industry
Healthcare
Status
Ongoing
Date Occurred
Jan 1, 2022
Date Reported
Mar 15, 2023
Jurisdiction
US
AI Provider
Other/Unknown
Application Type
embedded
Harm Type
physical
People Affected
1,000
Human Review in Place
Yes
Litigation Filed
No
Regulatory Body
FDA
medical_AIFDAradiologycancer_diagnosisfalse_negativesclinical_deployment
Full Description
Multiple FDA-approved AI medical imaging tools designed to assist radiologists in detecting cancerous tumors have shown concerning performance gaps between controlled clinical trials and real-world deployment. Studies published in radiology journals and presented at medical conferences have documented cases where these AI systems failed to flag suspicious lesions that were later confirmed as malignant tumors by human radiologists.
The performance degradation appears most pronounced in community hospitals and smaller healthcare systems, where imaging equipment, patient populations, and clinical workflows differ significantly from the academic medical centers where these AI tools were originally tested. Research published in the Journal of the American College of Radiology found that AI mammography screening tools showed false negative rates of 8-12% in community settings, compared to 2-4% reported in initial FDA submission studies.
Specific documented cases include AI tools missing early-stage lung cancers on chest CT scans at multiple healthcare systems, and mammography AI failing to detect breast cancers in women with dense breast tissue. The National Cancer Institute has received reports of delayed diagnoses potentially attributable to over-reliance on AI screening tools that failed to flag abnormalities.
The FDA has responded by issuing guidance requiring post-market surveillance studies for AI medical devices and mandating that healthcare providers maintain human oversight protocols. However, many healthcare systems had already integrated these tools into clinical workflows with varying degrees of radiologist review, creating potential gaps in patient safety.
The incidents have raised questions about the adequacy of pre-market testing for AI medical devices and the need for continuous monitoring of AI performance in diverse clinical environments. Medical professional organizations have called for standardized protocols for AI validation and deployment in healthcare settings.
Root Cause
AI models trained on controlled datasets failed to generalize to real-world clinical conditions, exhibiting performance degradation when exposed to variations in imaging equipment, patient populations, and clinical workflows not represented in training data.
Mitigation Analysis
Continuous performance monitoring in clinical settings could detect accuracy degradation. Mandatory human radiologist review of AI-flagged cases and random sampling of AI-cleared cases would catch missed diagnoses. Real-world validation studies post-deployment would identify performance gaps before widespread harm.
Lessons Learned
AI medical devices require extensive real-world validation beyond controlled trials to ensure safety across diverse clinical environments. Healthcare systems must implement robust human oversight and continuous performance monitoring when deploying AI diagnostic tools.
Sources
Performance of AI in Clinical Radiology Practice
Nature Medicine · Mar 15, 2023 · academic paper
FDA Issues Guidance on AI/ML Medical Devices
FDA · Apr 3, 2023 · regulatory action