FB pixel

John Howard reflects on DHS biometrics evaluations and the AI testing landscape

Former MdTF Chief Data Scientist highlights work on bias, standards
John Howard reflects on DHS biometrics evaluations and the AI testing landscape
 

When John Howard began working for SAIC at the Maryland Test Facility, biometrics was an emerging technology supported by a nascent industry. When he announced his departure near the end of January, the MdTF and Howard personally were long-established as central figures in proving the science and setting the standards behind facial recognition and other biometrics.

Howard’s lengthy list of contributions as senior principal scientist with SAIC ranges from helping set up and scale assessments of operational biometrics deployments in projects like the DHS Biometric Technology Rally, RIVTD and RIVR, to research and standard development around demographic fairness.

Reflecting on highlights from his time with MdTF in an email interview with Biometric Update, Howard says, “Throughout my career, one of my biggest areas of research has been into demographic effects (‘bias’) and how to test complex AI systems in operations.”

He points out that a system that always fails for every user group could be considered equitable, so usefulness is a function of both overall accuracy and fairness.

“Looking ahead, I believe this work will lay the foundation for how we — as a society and as individual companies – reason about 1) the operational realities of AI systems, 2) deep fakes, and 3) effective AI and data policies.”

Howard’s next project is Sensus AI, the independent testing, assessment, audit and advisory firm he co-founded and will lead as CEO.

Howard makes clear that his departure from MdTF does not reflect any diminished role for the lab in facial recognition policy and development.

“The U.S. Government has a significant role in guiding the future of facial recognition technologies, and government laboratories are going to be critical in providing expertise on the development, evaluation, and deployment of those technologies in the government space,” he says.

Evaluating evaluations

One of Howard’s last projects with MdTF was a paper on face biometrics image quality evaluation tool OFIQ. It suggested that the open-source tool is very limited in its usefulness to DHS, and its quality filter “did not substantially reduce error rates.”

Christoph Busch of Hochschule Darmstadt pushed back on those findings somewhat during an OFIQ user group meeting, suggesting that the criteria of the assessment is not as applicable to other users, like eu-LISA.

Howard does not entirely agree, but sees the debate as an example of the need to apply appropriate expertise to get real-world applications right.

“Airports and other large venues are going to face challenges as they introduce face recognition programs because they are complex environments,” he points out. “For most high-throughput deployments of face recognition systems, you don’t want to discard usable face samples. Our paper showed that is what OFIQ does. Every implementation has its subtleties, eu-LISA included, and that highlights the importance of expert consultation when operationalizing these technologies.”

Consequences of success

When asked about how helpful the biometrics testing and evaluation ecosystem is when it comes to evaluating vendor claims and the applicability of technology to specific use cases, Howard says that it “could do better.”

“Right now, there’s no shortage of people eager to build and deploy AI across all kinds of applications. What is missing, especially in the private sector, is a sufficient number of professionals focused on independent AI test, evaluation and auding. I believe that in the next four years, every organization is going to need independent, qualified, AI test, evaluation — much like organizations do financial audits today. That’s not the model we are currently operating under but I hope we move in that direction.”

As AI and biometrics adoption is followed by enhanced scrutiny and then governance, Howard also sees a need for realistic transparency.

“The industry should be clear to the public that these systems make mistakes,” he writes. “They are constantly getting better and honestly the progress has been remarkable but that success can sometimes lead people to stretch these systems into ever more applications. A system that makes very few errors in one use case can produce an astronomical number of mistakes in another.”

Related Posts

Article Topics

 |   |   |   |   |   |   | 

Latest Biometrics News

 

Australia plans biometric liveness detection refresh for national digital ID

Australia plans to contract a biometric liveness detection capability to support the country’s national digital ID and protect it against…

 

Deepfake threats exploiting the trust inside corporate systems

New York-based AI security company Reality Defender is warning businesses that deepfake threats have moved beyond isolated fraud schemes and…

 

Under AMLA, 95% false positives become a regulator’s problem

By Max Irwin, Regional Vice President EU, Shufti By the end of the day on 22 April 2026, around forty…

 

Sri Lanka defines trust boundaries ahead of digital ID rollout

Sri Lanka’s Unique Digital ID (SL-UDI project is placing trust architecture at the center of its rollout, with officials emphasizing…

 

Biometrics demand holds firm across core and emerging use cases

A UK court ruling that live facial recognition use by police does not violate human rights could have major implications…

 

ADVP and NO2ID back DVS framework from opposing perspectives

The UK’s Digital Verification Service (DVS) trust framework is drawing support from both industry and long-time critics of centralized identity…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Biometric Market Analysis and Buyer's Guides

Most Viewed This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events