FB pixel

Voice deepfakes are inkblots for consumers – they’re threats or entertainment

Pindrop survey shows split, Identt EAB presentation shows detection difficulty
Voice deepfakes are inkblots for consumers – they’re threats or entertainment
 

A recent marketing survey funded by a voice biometrics firm has found that a majority of respondents are concerned about the threat of deepfakes and voice clones.

The company, Pindrop, held a webinar to discuss the survey. Consumers surveyed were more likely to report feeling negative about voice cloning and deepfakes, but totals are not hugely lopsided. Synthedia performed the research.

Voice deepfakes can be detected with software, but the technology is still maturing, as Identt AI Technical Leader for Biometrics Piotr Kawa explained in another webinar to members of the EAB (European Association for Biometrics).

Consumers’ opinions are well-grounded, according to Pindrop webinar panelist Bret Kinsella, CEO and research director of trade publisher Voicebot.ai.

“The level of awareness (among consumers) is higher than I thought it’d be,” Kinsella said. Awareness is not the only insight from the survey that caught his eye.

Exactly 22.3 percent of consumer surveyed about deepfakes said they felt extremely positive about use of the software. And precisely 22.3 percent of people asked the same question said they felt extremely negative.

When surveyors asked consumers about voice clones as a concept, about 18.8 percent saw maximum upside. More people were extremely negative – 21.6 percent – about voice clones.

Among the positives some respondents saw in both voice clones and deepfakes was improved entertainment. Not surprisingly, people who were more concerned saw the negative possibility of impersonations and other problems.

Social media is where most people encounter doctored video and audio. In descending order: YouTube, Tik Tok, Instagram and Facebook. After that it’s movies and news publications.

Kinsella said that is a problem because it is harder to detect deepfakes and voice clones when someone is distracted.

This matches with research Kawa referred to, published back in 2021, that found only 80 percent of participants in a study were able to correctly identify the authenticity of content they were shown. Detection algorithms set on the same task were right 95 percent of the time. Subsequent studies are no more reassuring.

Building resources to meet generalizing challenges

Kawa began the latest EAB lunch talk with an overview of speech synthesis, and the impact generative AI has had on the field. A variety of commercial SaaS and open-source tools for speech synthesis are now widely available, making it “pretty easy” to synthesize speech, according to Kawa.

He differentiated between text-to-speech (TTS) and voice conversion, in which one person is made to sound like another. Either can be used to carry out audio deepfake attacks.

Deepfake detection methods today mostly relying on deep learning algorithms developed by biometrics researchers and are mostly based on finding artefacts left by synthetic speech algorithms. Kawa lists over a dozen, divided between models based on raw audio, front-end based models, algorithmic front-ends and embedding-based front ends for self-supervised learning.

The number of datasets to train them on has also increased rapidly, particularly over the last two years, according to Kawa.

Audio deepfake detection faces a major challenge in generalization. Kawa demonstrated how models tend to do a good job of detecting deepfakes that are created using the same techniques as the dataset the detection model was trained on. For those made with different techniques, however, performance is poor.

Larger training databases that include deepfakes made with various techniques, along with data augmentation techniques, can improve the detection results, but introducing variables like more background noise can make fakes more difficult to detect.

Kawa concluded with a review of open problems in deepfake detection, including generalization and creating models that can run quickly on consumer-grade electronics.

Article Topics

 |   |   |   |   |   | 

Latest Biometrics News

 

Biometrics race for the borders

Biometrics to ease border crossings are a major theme of the week among Biometric Update’s most-read articles of the week….

 

US election likely to be a missed opportunity to advance digital ID policy

The 2024 U.S. election represents an opportunity for social dialogue around digital identity policy in the wake of a series…

 

India to pilot Digi Yatra for foreign nationals in 2025

India is planning an international pilot project for June 2025 that will see the introduction of facial recognition technology beyond…

 

Papua New Guinea advances digital ID, wallet and govt platform to pilot

Papua New Guinea has stood up a new digital ID, wallet and online government platform, and plans to pilot them…

 

UK police organized crime unit seeks new facial recognition software

The UK’s main law enforcement agency against organized crime is looking into new facial recognition solutions, as the country doubles…

 

The EUDI Wallet was not meant for age assurance: AVPA

The European Union should not look at the EU Digital Identity (EUDI) Wallet as an age-assurance solution to keep minors…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Most Read This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events