FB pixel

Microsoft restricts facial recognition services, sunsets facial analysis

Microsoft restricts facial recognition services, sunsets facial analysis
 

Microsoft has released their Responsible AI Standard, a framework to guide the company’s work in artificial intelligence. Calling it “an important step in our journey to develop better, more trustworthy AI,” Natasha Crampton, Microsoft’s chief responsible AI officer, said the framework puts people at the center of system design decisions and aims to steer them toward better and more equitable outcomes. In a post on Microsoft’s website, Crampton said AI development needs to respect values like privacy, inclusiveness and accountability.

To that end, Microsoft said it will retire from its Azure Face recognition service facial analysis software designed to identify age, gender, emotional states and other qualities, citing concerns about bias and inaccuracy.

Said Crampton, “experts inside and outside the company have highlighted the lack of scientific consensus on the definition of ‘emotions,’ the challenges in how inferences generalize across use cases, regions, and demographics, and the heightened privacy concerns around this type of capability.”

Additional changes will be made to the company’s more traditional biometric systems, including facial recognition, which will now be limited to managed services customers and partners, and restricted to use-cases that have been pre-defined as acceptable. Users will be required to adhere to a code of conduct and follow guardrails to prevent misuse.

In its release, Microsoft emphasized the growing belief among AI observers that globally, laws to regulate uses of AI need to keep up with technological development. Crampton said the company recognizes its responsibility to act. “We believe that we need to work towards ensuring AI systems are responsible by design.”

The U.S.-based tech giant also joined others in halting sales of facial recognition to law enforcement agencies in 2020 in the absence of federal regulation.

The potential risks associated with AI have been in headlines recently, after Blake Lemoine, an AI engineer Google, was put on leave for claiming that the company’s AI language modeling system, LaMDA, had become sentient.

A document with the Responsible AI Standard’s General Guidelines can be accessed here.

Article Topics

 |   |   |   |   |   |   |   |   | 

Latest Biometrics News

 

Growing role of biometrics in everyday life demands urgent deepfake response

Biometrics are becoming more entrenched a couple of market segments, though not as fast as some would like. The top…

 

PNG expands mandatory digital ID to businesses taking gov’t contracts

The government of Papua New Guinea is making its national digital ID a mandatory form of authentication for all business…

 

Imply reaches face biometrics milestone at tech-forward Arena da Baixada

Imply Tecnologia’s facial recognition model has enabled more than 1 million accesses at Arena da Baixada, the home of Club…

 

Following IPO, ROC is investing in homegrown security for US market

In February, Colorado-based biometrics and vision AI provider ROC closed the first big biometrics IPO of 2026, raising just over…

 

Jumio expanding biometric reusable digital identity across LatAm

Following a launch in Brazil last year, U.S.-based Jumio is expanding its face biometrics-based reusable digital identity product, selfie.DONE, across…

 

Denmark imposes age checks to restrict social media to kids under 15

Welcome two more Europeans nations to the global age assurance legislation party. The Danish government is moving ahead with an…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Biometric Market Analysis and Buyer's Guides

Most Viewed This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events