FB pixel

Chincotech tackles racial bias in facial recognition systems

 

Tokyo-based software company Chincotech has announced the development of a multi-racial facial recognition system to provide superior accuracy than traditional systems, which often have unacceptably high error rates for non-white individuals.

In tests of facial recognition systems by M.I.T. Media Lab Researcher Joy Buolamwini, gender was misidentified for less than 1 percent of lighter-skinned males, and up to 7 percent of lighter-skinned females, The New York Times reports. The same systems misidentified the gender of up to 12 percent of darker-skinned males, and a shocking 35 percent of darker-skinned females.

The datasets used to test facial recognition systems may be contributing to the problem, as one widely-used collection of images is estimated to be more than 75 percent male and more than 80 percent white. Haverford College computer scientist Sorelle Friedler, a reviewing editor on Buolamwini’s research paper (PDF), said that experts have suspected that the performance of facial recognition systems depends on the population being considered, and that the research is the first to empirically confirm the suspicion.

The paper, written by Buolamwini and Microsoft researcher Timnit Gebru, studied facial analysis systems from Microsoft, IBM, and Megvii.

Chincotech is combatting this challenge with a 3D transforming face algorithm that continuously learns multi-racial characteristics to accurately identify people in 2D pictures.

“Our tests have proved that this technique coupled with a system that is taught to learn the difference between races and you have a system that delivers significantly more accurate results,” said Chincotech Head Software Development Engineer Paul Rashford.

Buolamwini has given a TED Talk on coded bias, and advocates for algorithmic accountability as a founder of the Algorithmic Justice League.

As previously reported, University of Surrey researchers developed a multi-racial facial recognition system last year which delivers more accurate results than are typical.

This post was updated at 9:22am on July 27, 2021, to clarify that the Gender Shades study tests facial analysis algorithms, not identification algorithms.

Article Topics

 |   |   |   |   | 

Latest Biometrics News

 

Understanding of what #SafeDPI is, how to achieve it creeps forward

If a government spends millions of dollars on an identity system or any other kind of digital public infrastructure that…

 

Financial firms beef up fraud prevention with biometrics and FIDO standards

Globally, financial companies are moving to strengthen their digital security and identity protocols, leveraging biometrics, FIDO standards and cryptography to…

 

Building trust in the age of digital identity: why cyber resilience must come first

By Nathalie Gosset, VP Identity and Biometric Solutions at Thales Trust is the invisible infrastructure of the digital world. Without…

 

Biometric ticketing, IDV sweeps across Brazilian stadiums under mandate

Brazil has mandated face biometrics for use in large stadiums, a landmark move for the widespread implementation of the technology….

 

China’s supreme court releases facial recognition violation cases in crackdown

China’s highest court has upheld the need for stronger protection of personal information, emphasizing to judges the need to maintain…

 

Privacy doesn’t have to cost us great online services

By Andrew Black, Managing Director ConnectID and Sujeet Rana, Chief Digital Officer NAB For years, we accepted an implicit trade-off…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Biometric Market Analysis

Most Viewed This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events