← Back to incidents

Amazon Rekognition Facial Recognition System Exhibited Racial Bias in Congressional Test

High

ACLU testing revealed Amazon Rekognition falsely matched 28 Congress members with criminal mugshots, disproportionately affecting people of color. The incident highlighted systemic bias in facial recognition technology used by law enforcement.

Category
Bias
Industry
Technology
Status
Resolved
Date Occurred
Jul 1, 2018
Date Reported
Jul 26, 2018
Jurisdiction
US
AI Provider
Other/Unknown
Model
Amazon Rekognition
Application Type
api integration
Harm Type
reputational
People Affected
28
Human Review in Place
No
Litigation Filed
No
facial_recognitionalgorithmic_biasracial_discriminationlaw_enforcementAmazonACLUcivil_rightsfalse_positive

Full Description

In July 2018, the American Civil Liberties Union (ACLU) conducted a test of Amazon's Rekognition facial recognition system that revealed significant racial bias in the technology. The ACLU compared official photos of all 535 members of Congress against a database of 25,000 publicly available arrest photos using Amazon's default confidence threshold of 80%. The test resulted in 28 false matches, incorrectly identifying sitting members of Congress as individuals who had been arrested. The results demonstrated a clear pattern of racial bias: while people of color make up only 20% of Congress, they represented 39% of the false matches. This disproportionate impact was particularly concerning given that Amazon was actively marketing Rekognition to law enforcement agencies across the United States. The false identifications included prominent members of Congress from diverse backgrounds, highlighting how algorithmic bias could lead to wrongful arrests or harassment of innocent individuals. Amazon responded by disputing the ACLU's methodology, arguing that the 80% confidence threshold used was too low for law enforcement applications and that their guidelines recommended a 95% threshold for such sensitive use cases. However, the company's own documentation and marketing materials had not clearly emphasized these limitations, and many law enforcement agencies were already using the system with default settings. The incident revealed a disconnect between Amazon's internal technical knowledge and how the system was being marketed and deployed. The revelation sparked widespread criticism from civil rights groups, researchers, and policymakers. MIT researcher Joy Buolamwini had previously documented similar bias patterns in facial recognition systems from multiple vendors, showing that these technologies consistently performed worse on women and people with darker skin tones. The Congressional test provided a high-profile demonstration of how these technical limitations translated into real-world discrimination risks. The incident contributed to growing scrutiny of facial recognition technology and ultimately led to policy changes. Several major cities banned law enforcement use of facial recognition, and Amazon eventually implemented a moratorium on police use of Rekognition in 2020. The company has since required higher confidence thresholds and implemented additional bias testing, though critics argue these measures are insufficient to address fundamental algorithmic fairness issues.

Root Cause

Amazon's Rekognition facial recognition system demonstrated algorithmic bias, likely due to training data that underrepresented people of color and insufficient bias testing during model development and validation.

Mitigation Analysis

This incident could have been prevented through comprehensive bias testing across demographic groups during development, diverse training datasets with balanced representation, and mandatory accuracy thresholds by demographic group before deployment. Ongoing algorithmic auditing and third-party bias assessments would have identified these disparities before public deployment.

Lessons Learned

The incident demonstrated that facial recognition systems can perpetuate and amplify existing societal biases, particularly affecting people of color. It highlighted the critical need for comprehensive bias testing, appropriate confidence thresholds, and careful consideration of deployment contexts before releasing AI systems for sensitive applications like law enforcement.