• Home
  • Apps
  • Apps News
  • Amazon 'Rekognition' Face Detection Technology Shows Bias, Researchers Say

Amazon 'Rekognition' Face-Detection Technology Shows Bias, Researchers Say

Amazon 'Rekognition' Face-Detection Technology Shows Bias, Researchers Say
Highlights
  • Rekognition often misidentifies women, especially those with darker skin
  • Amazon called on to stop marketing its Rekognition service
  • Amazon tech labelled darker-skinned women as men 31 percent of the time
Advertisement

Facial-detection technology that Amazon is marketing to law enforcement often misidentifies women, particularly those with darker skin, according to researchers from MIT and the University of Toronto.

Privacy and civil rights advocates have called on Amazon to stop marketing its Rekognition service because of worries about discrimination against minorities. Some Amazon investors have also asked the company to stop out of fear that it makes Amazon vulnerable to lawsuits.

The researchers said that in their tests, Amazon's technology labelled darker-skinned women as men 31 percent of the time. Lighter-skinned women were misidentified 7 percent of the time. Darker-skinned men had a 1 percent error rate, while lighter-skinned men had none.

Artificial intelligence can mimic the biases of their human creators as they make their way into everyday life. The new study, released late Thursday, warns of the potential of abuse and threats to privacy and civil liberties from facial-detection technology.

Matt Wood, general manager of artificial intelligence with Amazon's cloud-computing unit, said the study uses a "facial analysis" and not "facial recognition" technology. Wood said facial analysis "can spot faces in videos or images and assign generic attributes such as wearing glasses; recognition is a different technique by which an individual face is matched to faces in videos and images."

In a Friday post on the Medium website, MIT Media Lab researcher Joy Buolamwini responded that companies should check all systems that analyse human faces for bias.

"If you sell one system that has been shown to have bias on human faces, it is doubtful your other face-based products are also completely bias free," she wrote.

Amazon's reaction shows that it isn't taking the "really grave concerns revealed by this study seriously," said Jacob Snow, an attorney with the American Civil Liberties Union.

Buolamwini and Inioluwa Deborah Raji of the University of Toronto said they studied Amazon's technology because the company has marketed it to law enforcement. Raji's LinkedIn account says she is currently a research mentee for artificial intelligence at Google, which competes with Amazon in offering cloud-computing services.

Buolamwini and Raji say Microsoft and IBM have improved their facial-recognition technology since researchers discovered similar problems in a May 2017 study. Their second study, which included Amazon, was done in August 2018. Their paper will be presented on Monday at an artificial intelligence conference in Honolulu.

Wood said Amazon has updated its technology since the study and done its own analysis with "zero false positive matches."

Amazon's website credits Rekognition for helping the Washington County Sheriff Office in Oregon speed up how long it took to identify suspects from hundreds of thousands of photo records.

Comments

For the latest tech news and reviews, follow Gadgets 360 on X, Facebook, WhatsApp, Threads and Google News. For the latest videos on gadgets and tech, subscribe to our YouTube channel. If you want to know everything about top influencers, follow our in-house Who'sThat360 on Instagram and YouTube.

Mark Zuckerberg Plans to Integrate WhatsApp, Instagram, and Facebook Messenger: Report
Facebook Allowed 'Friendly Fraud' to Profit From Kids, Internal Memos Show
Share on Facebook Gadgets360 Twitter Share Tweet Snapchat Share Reddit Comment google-newsGoogle News
 
 

Advertisement

Follow Us

Advertisement

© Copyright Red Pixels Ventures Limited 2024. All rights reserved.
Trending Products »
Latest Tech News »