New Chrome extension aims to detect deepfake AI-generated voices—but it’s far from perfect

Deepfakes have already been a problem in the 2024 presidential election—and could potentially become a bigger one as Nov. 5 draws near. But a new Chrome browser extension claims it can help people determine which audio clips they hear are legitimately from the candidates (and others) and which are fake.

The Hiya Deepfake Voice Detector uses artificial intelligence to determine if the voice on screen is legitimate or faked. The company claims the tool has a 99% accuracy rate and says it can verify (or debunk) audio in just a few seconds.  

[Screenshot: Hiya]

“Our models are trained to detect subtle audio artifacts unique to AI-generated voices—imperceptible to the human ear but identifiable by machine learning algorithms—and just a second of audio is enough to detect their presence,” Patchen Noelke, vice president of marketing for Hiya, tells Fast Company.

I put the free tool to the test and found that while it certainly does catch some deepfakes, its accuracy varied.

Testing the browser extension

Founded in 2015 as part of online directory Whitepages, Hiya is in the business of providing services that screen calls for spam and fraud. Today, the company has more than 450 million active users, according to its Website.

After installing the Hiya’s extension from the Chrome Webstore and registering, users open the tool in their browser and click “Start analyzing” when they’re watching a video or listening to an audio clip whose origin is in doubt. Users are limited to 20 queries per day. 

While it’s certainly helpful to have any sort of tool that identifies even a percentage of the deepfake video and audio clips (others, though not browser extensions, include Pindrop Security, AI or Not and AI Voice Detector), my own tests with the Hiya Google Chrome app found some notable holes.

[Screenshot: Hiya]

Using real and deepfake soundbites that appeared in a Washington Post story, I had the tool examine both a real and fake clip of Kamala Harris. To its credit, it did call out the deepfake, but the tool said it was uncertain about the authentic clip of Harris.

An AI-generated audio file of Trump was judged as likely authentic, with an authenticity score of 88 out of 100. And a confirmed deepfake of Ukrainian president Volodymyr Zelensky instructing his troops to lay down their weapons got an even higher authenticity score of 90.

Deepfakes of Mark Zuckerberg and an AI created video that featured a voice actor imitating "> Morgan Freeman were also wrongly deemed authentic, with the fake Zuckerberg earning a score of 79 and the fake Freeman getting a 72.

[Screenshot: Hiya]

A "> deepfake of Kim Joo-Ha, a news anchor on Korean television channel MBN (which was presented by the network last year), was spotted immediately as a fake, however, earning a score of just 1.

Noelke admits the deepfake detector has some limitations. Freeman and Zelensky fakes were made with voice actors rather than AI, which is why they weren’t detected, he says. The score on the Trump deepfake changed depending on how much audio was sampled, Noelke notes. (“When scoring the entire file instead of analyzing it in 4-second chunks, the overall score is 63%, placing it in the inconclusive category,” he says.) And the Harris deepfake had loud background music that interfered with scoring, which is why the tool was uncertain.

“The way we designed the Chrome extension is it gives you a confidence score and you do have to judge for yourself,” he says. “There are some things that make it harder [like] the quality of the sound and background music and other things can affect it. It can definitely fail. It’s definitely not 100%— and that’s why we deliver the results the way we [do].”

The app also only analyzes the audio portion of a video, so if a politician, celebrity, or anyone else’s image is digitally inserted into a video where they do not actually appear, it cannot flag that as inauthentic. We tested this on a "> Back to the Future clip that swaps Tom Holland and Robert Downey Jr. for Michael J. Fox and Christopher Lloyd.

Deepfakes and the election

AI has already thrown a wrench into the 2024 presidential election. In January, a deepfake robocall that reproduced the voice of President Joe Biden urged Democrats not to vote in the New Hampshire primary. Since then, there have been false images that show everything from Taylor Swift endorsing Trump to Kamala Harris in a communist uniform to Donald Trump dancing with an underage girl. And Elon Musk, this summer, shared a deepfake that mimicked the voice of Harris claiming to be a “diversity hire” and saying she didn’t know “the first thing about running a country.”

[Screenshot: Hiya]

Musk, when criticized, later said the video was intended as satire.

Several states have enacted laws regulating election deepfakes, including California, Texas, and Florida. But only three of the eight swing states that are expected to help decide the 2024 presidential election have laws on the books—Michigan, Wisconsin and Arizona. (Legislation is pending in North Carolina and Pennsylvania, but is unlikely to pass before the election.)

The Federal Communications Commission (FCC) quickly made AI-generated voices in robocalls illegal, but federal officials have been slower to act on other mediums. And when it comes to online content, which is where these most frequently spread, that continues to be the wild west. The FCC says it does not regulate online content. 

Update, October 17, 2024: This story has been updated with a response from Noelke about the specific audio clips tested.

https://www.fastcompany.com/91211729/new-chrome-extension-detect-ai-generated-voices-deepfakes?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Vytvorené 4mo | 17. 10. 2024, 21:50:04


Ak chcete pridať komentár, prihláste sa

Ostatné príspevky v tejto skupine

5 time-saving Alexa commands you’re probably not using yet

Even if you’re a regular Alexa user, there’s a good chance you haven’t discovered some of its most efficient features.

Actually, strike that: There’s a good chance you’

25. 2. 2025, 7:50:02 | Fast company - tech
Why today’s youth need more math, logic, and grammar skills

The Fast Company Impact Council is a private membership community of influential leaders, experts, executives, and entrepreneurs who share their insights with our audience. Members pay annual

25. 2. 2025, 3:10:10 | Fast company - tech
Here are crypto’s biggest heists after Bybit’s $1.5 billion hack

Cryptocurrency exchange Bybit said last week hackers had stolen digital tokens worth around $1.5 billion, in what researchers called the biggest crypto heist of all time.

Bybit CEO Ben Z

24. 2. 2025, 22:30:07 | Fast company - tech
‘We are never going to stop existing’: Hunter Schafer called out Trump’s passport policy on TikTok

“I had a bit of a harsh reality check today, and felt like it’s important to share with whoever is listening,” model and actress Hunter Schafer said in an eight-minute

24. 2. 2025, 20:20:06 | Fast company - tech
Anthropic’s new Claude AI model can decide between speed and deep thinking

Anthropic released on Monday its Claude 3.7 Sonnet model, which it says returns results faster and can show the user the “chain of thought” it follows to reach an answer. This latest model also po

24. 2. 2025, 20:20:05 | Fast company - tech
Ai2’s Ali Farhadi advocates for open-source AI models. Here’s why

A year before Elon Musk helped start OpenAI in San Francisco, philanthropist and Microsoft cofounder Paul Allen already had established his own nonprofit

24. 2. 2025, 17:50:07 | Fast company - tech