×
AI voice clones fool humans with just 4 minutes of training
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

New research from Queen Mary University of London reveals that AI voice clones created with just four minutes of audio recordings are now indistinguishable from real human voices to average listeners. The study demonstrates how sophisticated consumer-grade AI voice technology has become, raising significant concerns about fraud, misinformation, and the potential for voice cloning scams.

What you should know: Researchers tested people’s ability to distinguish between real voices and AI-generated clones using readily available technology.
• The study used 40 synthetic AI voices and 40 human voice clones created with ElevenLabs’ consumer tool, requiring roughly four minutes of recordings per clone.
• When 28 participants were asked to identify which voices were real versus AI-generated, they could not reliably distinguish between voice clones and authentic human recordings.
• “The process required minimal expertise, only a few minutes of voice recordings, and almost no money,” said Nadine Lavan, who led the research at Queen Mary University of London.

The big picture: AI voice synthesis has evolved beyond the robotic quality that once made synthetic speech easily detectable.
• While participants could somewhat distinguish entirely synthetic voices from human ones, voice clones that replicate specific people’s voices have reached near-perfect realism.
• Unlike AI-generated images that sometimes appear “hyper-realistic” compared to real photos, AI voice clones matched human recordings without exceeding their perceived authenticity.
• Participants rated AI-generated voices as more dominant and trustworthy than their human counterparts.

Why this matters: The technology’s accessibility and sophistication create both opportunities and serious risks across multiple sectors.
• Positive applications include improved accessibility tools, educational resources, and enhanced communication experiences through high-quality synthetic voices.
• However, the research highlights growing concerns about copyright infringement, misinformation campaigns, and increasingly sophisticated fraud schemes.
• Voice cloning scams, where criminals impersonate family members or celebrities to manipulate victims, are becoming more difficult to detect.

What they’re saying: Experts acknowledge both the promise and peril of ultra-realistic voice synthesis.
• “The ability to generate realistic voices at scale opens up exciting opportunities,” Lavan noted, pointing to applications in accessibility, education, and communication.
• She warned that the findings reveal “many worrying ethical implications in areas like copyright infringement, the ability to spread misinformation, and fraud.”
• The study adds to mounting evidence that AI voices are “quickly becoming impossible to detect,” despite industry efforts to implement protective guardrails.

People Can’t Distinguish AI Voice Clones From Actual Humans Anymore

Recent News

Survey: 73% of CIOs now use Macs primarily for AI processing

Nearly all CIOs expect to increase Mac investments over the next two years.

Pennsylvania TECH360 conference shows businesses how to deploy AI beyond the hype

Leaders navigate the AI implementation gap with industry-specific strategies.