FB pixel

CMU-Africa researchers develop biometric PAD bias reduction methods and metrics

CMU-Africa researchers develop biometric PAD bias reduction methods and metrics
 

A team of researchers at Carnegie Mellon University Africa (CMU-Africa) in Rwanda have developed a new approach to making biometric presentation attack detection (PAD) resistant to demographic bias.

Their paper on “Fairness-aware face presentation attack detection using local binary patterns: Bridging skin tone bias in biometric systems” focusses on demographic bias affecting African people, and has been published by the Journal of Cybersecurity and Privacy.

The researchers chose local binary patterns (LBPs), a type of visual descriptor used in computer vision for classifying subjects, as a basis for their PAD system in part due to their effectiveness in resource-constrained environments. They used a large dataset compiled by the Chinese Academy of Sciences (CASIA-SURF CeFA) to train, test and validate the algorithms they developed.

The innovation they present is the use of an “ethnicity-aware preprocessing approach.” This involves the targeted application of adaptive brightness settings and gamma correction optimization for the reflective properties of different skin tones, addressing a long-recognized challenge in image quality.

“Current methods address fairness through post hoc corrections or data augmentation, but they fail to address the fundamental challenge of ensuring optimal image quality across different skin tones before feature extraction,” the researchers explained in the paper.

Following feature extraction, the researchers used an SGD (stochastic gradient descent) classifier and balanced class weighting to prevent “systemic bias toward the majority class without requiring data reduction or augmentation.” And group-specific threshold optimization were implemented to limit the equal error rate (EER).

The researchers also developed a statistical framework for evaluating the fairness of PAD algorithms, which includes three novel methods. A coefficient of Variation (CoV) analysis provides a method for evaluating the trade-off between security and fairness. McNemar’s statistical significance testing indicates whether performance differences found are statistically significant or noise. And bootstrap confidence intervals quantify the uncertainty built into performance estimates by resampling with 1,000 iterations to generate 95 percent confidence intervals across ethnic groups.

The changes reduced the difference in accuracy between African and East Asian subjects from just over 3 percent to 0.75 percent, according to the paper.

“The methodology demonstrates that demographic fairness can be secured through algorithmic revisions instead of being reliant on massive data alteration or special-purpose hardware, which makes fair PAD systems easier to use more routinely in different contexts of deployment,” the CMU-Africa researchers conclude.

Related Posts

Article Topics

 |   |   |   |   |   | 

Latest Biometrics News

 

Face biometrics use cases outnumbered only by important considerations

With face biometrics now used regularly in many different sectors and areas of life, stakeholders are asking questions about a…

 

Biometric Update Podcast explores identification at scale using browser fingerprinting

“Browser fingerprinting is this idea that modern browsers are so complex.” So says Valentin Vasilyev, Chief Technology Officer of Fingerprint,…

 

Passkeys now pervasive but passwords persist in enterprise authentication

Passkeys are here; now about those passwords. Specifically, passkeys are now prevalent in the enterprise, the FIDO Alliance says, with…

 

Pornhub returns to UK, but only for iOS users who verify age with Apple

In the UK, “wanker” is not typically a term of endearment. However, the case may be different for Pornhub, which…

 

Europol operated ‘shadow’ IT systems without data safeguards: Report

Europol has operated secret data analysis platforms containing large amounts of personal information, such as identity documents, without the security…

 

EU pushes AI Act deadlines for high-risk systems, including biometrics

The EU has reached a provisional agreement on changes to the AI Act that postpone rules on high-risk AI systems,…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Biometric Market Analysis and Buyer's Guides

Most Viewed This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events