FB pixel

Deepfake threat still stirring concern about election integrity

Fears grow around ‘cognitive hacking that changes your relationship with reality’
Deepfake threat still stirring concern about election integrity
 

The specter of deepfakes continues to hover over the various elections of 2024, and even if little actual deepfake-driven disruption has materialized, anyone who has done a Google image search recently will know how much fake content has inundated the digital world.

In response, Microsoft has lent support to an organization looking to develop an open technical standard for establishing the provenance of digital content and identifying AI-generated assets. In Brazil, the DFRLab and the Federal University of Rio de Janeiro’s NetLab UFRJ are testing “various methodologies to examine the landscape of research methods to help identify electoral deepfakes” that could impact elections. Meanwhile in California, a U.S. district judge has paused a law that allows anyone to sue for damages over election deepfakes, suggesting it might violate the Constitution.

Microsoft offers ‘Content Credentials’ to fight deepfake threat

A post by Vanessa Ho in Microsoft’s Building AI Responsibly series says that, as deepfakes become easier to create, it is committed to supporting a “more trustworthy information ecosystem with responsible AI tools and practices.”

Andrew Jenks is director of media provenance for Microsoft, and chair of the Coalition for Content Provenance and Authenticity (C2PA), which Microsoft co-founded. “The repercussions of deepfakes can be incredibly severe,” he says. “They’re a form of cognitive hacking that changes your relationship with reality and how you think about the world.”

Generative AI is spurring a rise in disinformation. Identity theft and political interference are well within the reach of fraudsters in an ecosystem in which uncredited and unsourced content circulates freely.

Microsoft, which has previously pushed for AI deepfake fraud to be made illegal, is currently previewing an application that allows creators and publishers to add “Content Credentials” to their work – or certified metadata that cryptographically attaches details such as who made the content, when it was made and whether its creation involved AI, to serve as a kind of invisible trust stamp.

“Content Credentials provide an important layer of transparency, whether or not AI was involved, to help people make more informed decisions about content they share and consume online,” Jenks says. “As it becomes easier to identify content sourcing and history, people may become more skeptical of material that lacks specific provenance information.”

Identifying deepfakes still a laborious process for Brazilian researchers

Deepfake content has been on the minds of researchers at the DFRLab and the Federal University of Rio de Janeiro’s NetLab UFRJ, as Brazilians headed to the polls this week for municipal elections. For DFRLab, Beatriz Farrugia explores “the challenges of identifying deepfakes ahead of the 2024 Brazil election” in a case study that illustrates difficulties researchers and others face in attempting to identify generative content, and looks at new regulations that “codify rules stating that any campaign content generated or edited by AI must feature a disclaimer acknowledging the use of AI.”

A thorough analysis of platform compliance in identifying deepfakes and AI generated content leads to the following conclusion: “there are limitations in available research methodologies to monitor the proliferation of AI-generated imagery at scale. Due to these current limitations in platform search capabilities, identifying AI-generated content remains a laborious process, requiring ongoing and often manual monitoring, whether by the research community, electoral officials, or the public at large.”

California deepfake law a ‘blunt tool that hinders humorous expression,’ says judge

In a notably humorless ruling on the mechanics of humor, U.S. U.S. District Judge John A. Mendez has granted a preliminary injunction blocking a California law that gives people the right to sue for damages caused by deepfakes, saying it likely violates the First Amendment.

AP quotes the ruling, in which Mendez which “most of AB 2839 acts as a hammer instead of a scalpel, serving as a blunt tool that hinders humorous expression and unconstitutionally stifles the free and unfettered exchange of ideas which is so vital to American democratic debate.”

The law has only been in place for a month, and was accompanied by two other bills putting tough restrictions on the use of AI to create false images or videos in political ads.

Public Citizen, which supported the law, has issued a statement from its co-president Robert Weissman, who says “the court’s decision misses the fundamental problem with deepfakes, which is not simply that they make false claims but that they show candidates saying or doing things that the candidates did not say or do. The court suggests that a targeted candidate can just respond with counter speech – but that is not true, where the candidate has to ask the public not to believe their eyes and ears.”

He notes that the court “recognized that labeling requirements, if narrowly tailored, could pass constitutional muster,” and says “there’s nothing about the First Amendment that ties our hands in addressing fraud and a here-and-now threat to democracy.”

Related Posts

Article Topics

 |   |   |   |   |   |   |   |   | 

Latest Biometrics News

 

Facial age estimation spoof, VPN bypass claims called into question

Reports of the defeat of facial age estimation technology may be greatly exaggerated, and UK children have not flocked to…

 

Reddit users’ questions expose major shortcoming in age assurance effort

 “Am I the only one that’s confused about how they’re going to confirm if you’re a certain age?” So asks…

 

AI fraud threat continues to spur deepfake detection integration, investment, development

Reality Defender and 1Kosmos have announced a strategic partnership that will see the deepfake detection firm integrate its real-time deepfake…

 

iProov, Aware, Paravision power airport biometric boarding pilots at MCO

Airports across the Americas are accelerating their shift to biometric identity systems, with Orlando, Houston and Oklahoma City all rolling…

 

EU and Canada agree to collaborate on digital ID mutual recognition, pilots

Representatives of the European Union and Canada emerged from the meeting of the EU-Canada Digital Partnership Council on Monday with…

 

Ireland plants flag for child online safety as it prepares to assume EU presidency

Ireland may soon follow its neighbors in imposing age assurance requirements on adult content platforms. The Times says members of…

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Biometric Market Analysis

Most Viewed This Week

Featured Company

Biometrics Insight, Opinion

Digital ID In-Depth

Biometrics White Papers

Biometrics Events