Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Report 5518

Associated Incidents

Incident 114123 Report
Purported AI Voice Cloning Used to Impersonate Secretary of State Marco Rubio

Loading...
How a deepfake of Marco Rubio exposed the alarming ease of AI voice scams
fortune.com · 2025

An audio deepfake impersonating Secretary of State Marco Rubio contacted foreign ministers, a U.S. governor, and a member of Congress with AI-generated voicemails mimicking his voice, according to a senior U.S. official and a State Department cable dated July 3. 

There's no public evidence that any of the recipients of the messages, reportedly designed to extract sensitive information or gain account access, were fooled by the scam. But the incident is the latest high-profile example of how easy---and alarmingly convincing---AI voice scams have become.

With just 15 to 30 seconds of someone's speech uploaded to services like Eleven Labs, Speechify and Respeecher, it's now possible to type out any message and have it read aloud in their voice. Keep in mind, these are tools used perfectly legitimately for a host of things from accessibility to content creation -- but like many AI technologies, can be misused by bad actors.

The threat of deepfakes has escalated

AI-generated deepfakes aren't new, particularly of C-suite leaders and public officials, but they are becoming a bigger problem. Eight months ago, I reported that more than half of chief information security officers (CISOs) surveyed ranked video and audio deepfakes as a growing concern. That threat has only escalated. A new study by Surfshark found that in the first half of 2025 alone, deepfake-related incidents surged to 580---nearly four times as many as in all of 2024 (150 incidents), and dramatically higher than the 64 incidents reported between 2017 and 2023. Losses from deepfake fraud have also skyrocketed, reaching $897 million cumulatively, with $410 million of that in just the first half of 2025. The most common scheme: impersonating public figures to promote fraudulent investments, which has already resulted in $401 million in losses.

"Deepfakes have evolved into real, active cybersecurity threats," Aviad Mizrachi, CTO and cofounder of software security company Frontegg, told me by email. "We're already seeing AI-generated video calls successfully trick employees into authorizing multimillion-dollar payments. These attacks are happening now, and it's a scam that is becoming alarmingly easy for a hacker to deploy."

Part of the problem, Mizrachi added, is that traditional authentication methods---usernames, passwords, one-time codes, and authenticator apps---weren't designed for a world where a scammer can clone your voice or face in seconds. That's because these scams don't necessarily involve breaking into an account---they rely on tricking a real person into handing over credentials or authorizing actions themselves.

"Those traditional security measures to check the identity of an individual obviously don't work anymore," he said, adding that most cybersecurity teams still overlook deepfakes---and that's the vulnerability attackers exploit. A convincing fake voice on a voice mail message or video call can persuade someone to bypass normal procedures or approve a wire transfer, even if all the authentication tools are technically in place.

To guard against that kind of deception, Mizrachi said, organizations need to deploy stronger security tools that rely on physical devices---like a smartphone or hardware security key---to prove someone's identity. These tools, known as FIDO2 or WebAuthn passkeys, are far harder for hackers to fake or phish. And beyond device checks, smart verification systems can also monitor behavioral signals---like typing speed, location, or login habits---to spot anomalies that a cloned voice can't imitate. Those extra layers make it much harder for a deepfake attack to succeed.

Margaret Cunningham, director of security and AI strategy at security firm Darktrace, said that the impersonation attempt of Rubio demonstrates just how easily generative AI can be used to launch convincing, targeted social engineering attacks.

"This threat didn't fail because it was poorly crafted---it failed because it missed the right moment of human vulnerability," she said. "People often don't make decisions in calm, focused conditions. They respond while multitasking, under pressure, and guided by what feels familiar. In those moments, a trusted voice or official-looking message can easily bypass caution."

Deepfakes have impacted democracies around the world

Generative AI has also dramatically lowered the barrier to entry when it comes to media manipulation---making it faster, cheaper, and more scalable than ever before. And it is impacting democracies around the world: A recent New York Times report found that AI-powered deepfakes has transformed elections in at least 50 countries around the world, by using them to demean or defame opponents. 

"This is the new frontier for influence operations," Leah Siskind, AI research fellow at the Foundation for Defense of Democracies, told me. "We've seen other instances of deepfakes of senior government officials used to gain access to personal accounts, but leveraging AI to influence diplomatic relationships and decision-making is a dangerous escalation. This is an urgent national security issue with serious diplomatic ramifications."

Read the Source

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd