← Back to incidents
Amazon Rekognition Facial Recognition System Exhibited Racial Bias in Congressional Test
HighACLU testing revealed Amazon Rekognition falsely matched 28 Congress members with criminal mugshots, disproportionately affecting people of color. The incident highlighted systemic bias in facial recognition technology used by law enforcement.
Category
Bias
Industry
Technology
Status
Resolved
Date Occurred
Jul 1, 2018
Date Reported
Jul 26, 2018
Jurisdiction
US
AI Provider
Other/Unknown
Model
Amazon Rekognition
Application Type
api integration
Harm Type
reputational
People Affected
28
Human Review in Place
No
Litigation Filed
No
facial_recognitionalgorithmic_biasracial_discriminationlaw_enforcementAmazonACLUcivil_rightsfalse_positive
Full Description
In July 2018, the American Civil Liberties Union (ACLU) conducted a test of Amazon's Rekognition facial recognition system that revealed significant racial bias in the technology. The ACLU compared official photos of all 535 members of Congress against a database of 25,000 publicly available arrest photos using Amazon's default confidence threshold of 80%. The test resulted in 28 false matches, incorrectly identifying sitting members of Congress as individuals who had been arrested.
The results demonstrated a clear pattern of racial bias: while people of color make up only 20% of Congress, they represented 39% of the false matches. This disproportionate impact was particularly concerning given that Amazon was actively marketing Rekognition to law enforcement agencies across the United States. The false identifications included prominent members of Congress from diverse backgrounds, highlighting how algorithmic bias could lead to wrongful arrests or harassment of innocent individuals.
Amazon responded by disputing the ACLU's methodology, arguing that the 80% confidence threshold used was too low for law enforcement applications and that their guidelines recommended a 95% threshold for such sensitive use cases. However, the company's own documentation and marketing materials had not clearly emphasized these limitations, and many law enforcement agencies were already using the system with default settings. The incident revealed a disconnect between Amazon's internal technical knowledge and how the system was being marketed and deployed.
The revelation sparked widespread criticism from civil rights groups, researchers, and policymakers. MIT researcher Joy Buolamwini had previously documented similar bias patterns in facial recognition systems from multiple vendors, showing that these technologies consistently performed worse on women and people with darker skin tones. The Congressional test provided a high-profile demonstration of how these technical limitations translated into real-world discrimination risks.
The incident contributed to growing scrutiny of facial recognition technology and ultimately led to policy changes. Several major cities banned law enforcement use of facial recognition, and Amazon eventually implemented a moratorium on police use of Rekognition in 2020. The company has since required higher confidence thresholds and implemented additional bias testing, though critics argue these measures are insufficient to address fundamental algorithmic fairness issues.
Root Cause
Amazon's Rekognition facial recognition system demonstrated algorithmic bias, likely due to training data that underrepresented people of color and insufficient bias testing during model development and validation.
Mitigation Analysis
This incident could have been prevented through comprehensive bias testing across demographic groups during development, diverse training datasets with balanced representation, and mandatory accuracy thresholds by demographic group before deployment. Ongoing algorithmic auditing and third-party bias assessments would have identified these disparities before public deployment.
Lessons Learned
The incident demonstrated that facial recognition systems can perpetuate and amplify existing societal biases, particularly affecting people of color. It highlighted the critical need for comprehensive bias testing, appropriate confidence thresholds, and careful consideration of deployment contexts before releasing AI systems for sensitive applications like law enforcement.
Sources
Amazon's Face Surveillance Falsely Matched 28 Members of Congress with Mugshots
ACLU · Jul 26, 2018 · company statement
Amazon facial-identification software misidentified 28 lawmakers as suspects, ACLU says
Washington Post · Jul 26, 2018 · news