Privacy and civil rights advocates have called on Amazon to stop marketing its Rekognition service because of worries about discrimination against minorities. Some Amazon investors have also asked the company to stop out of fear that it makes Amazon vulnerable to lawsuits.
The researchers said that in their tests, Amazon's technology labeled darker-skinned women as men 31 percent of the time. Lighter-skinned women were misidentified 7 percent of the time. Darker-skinned men had a 1 percent error rate, while lighter-skinned men had none.
Artificial intelligence can mimic the biases of their human creators as they make their way into everyday life. The new study, released late Thursday, warns of the potential of abuse and threats to privacy and civil liberties from facial-detection technology.
Matt Wood, general manager of artificial intelligence with Amazon's cloud-computing unit, said the study uses a "facial analysis" and not "facial recognition" technology. Wood said facial analysis "can spot faces in videos or images and assign generic attributes such as wearing glasses; recognition is a different technique by which an individual face is matched to faces in videos and images."
In a Friday post on the Medium website, MIT Media Lab researcher Joy Buolamwini responded that companies should check all systems that analyze human faces for bias.
"If you sell one system that has been shown to have bias on human faces, it is doubtful your other face-based products are also completely bias free," she wrote.
Amazon's reaction shows that it isn't taking the "really grave concerns revealed by this study seriously," said Jacob Snow, an attorney with the American Civil Liberties Union.
Wood said Amazon has updated its technology since the study and done its own analysis with "zero false positive matches."
Amazon's website credits Rekognition for helping the Washington County Sheriff Office in Oregon speed up how long it took to identify suspects from hundreds of thousands of photo records.