×
GreekEnglish

×
  • Politics
  • Diaspora
  • World
  • Lifestyle
  • Travel
  • Culture
  • Sports
  • Cooking
Saturday
06
Dec 2025
weather symbol
Athens 12°C
  • Home
  • Politics
  • Economy
  • World
  • Diaspora
  • Lifestyle
  • Travel
  • Culture
  • Sports
  • Mediterranean Cooking
  • Weather
Contact follow Protothema:
Powered by Cloudevo
> technology

Voice Cloning: A new form of AI-assisted fraud sweeps the US and is coming to Europe

What is "vishing" or "voice phishing"? - Three seconds of speech are enough for scammers to steal your voice. Are there ways to avoid falling into their trap?

Newsroom October 20 10:44

Imagine this: It’s just past midnight, and your mother’s number appears on your phone screen. You answer, and on the other end, you hear her voice—it’s unmistakably her, and the caller ID confirms it’s her number—crying out, “Let me go!” “Help!” and other phrases in panic. Then a man takes over, telling you that there was a break-in at her house, but they found nothing valuable, so they’re holding her hostage. He demands that you send a sum of money immediately—before hanging up—or they will kill her.

Until now, phone scams followed a familiar pattern, mostly targeting specific victims. For example, someone would call pretending to be a police officer and tell the victim that a relative had been involved in a serious car accident. To avoid jail, the victim would be asked to pay a large sum of money. While the majority of victims have been elderly, a recent case in Zografou involved a 13-year-old who handed over €2,500 in cash, jewelry, and watches to scammers. If you think you’re unlikely to be scammed, you might want to reconsider, as scammers are now leveraging Artificial Intelligence (AI) to create voice clones, introducing a terrifying new type of fraud.

Three seconds of speech. That’s all an AI application needs to clone your voice—or that of a loved one. After that, to carry out “vishing” (voice phishing), all that’s needed is access to personal information, contacts, SIM card cloning (already widespread), and programming an AI app to deliver the message—using the right tone, of course. The scam has now become so convincing that even the most cautious individuals can be tricked. It’s even easier when scammers have access to social media, videos, and other personal data posted online.

How easy is it for your voice or that of a loved one to be stolen? Consider a brief phone call or a short video posted on TikTok, Facebook, or other social platforms where users share countless videos every day.

“Millions of Targets”

Though relatively new, this type of phone scam has gained popularity in the US and Asia, and it’s now coming to Europe. Last week, the UK’s Starling Bank issued a warning that “millions of people could fall victim to scams using AI programs to clone their voices.”

In the US, this fraud technique gained widespread attention when it was revealed that voice cloning technology was used to trick thousands of Democrats by making them receive calls from a fake voice of Joe Biden, urging them to abstain from voting. In South Korea, a case shocked the nation when a businessman was convinced he was on a video call with his company’s executives—who were, in fact, criminals—causing him to lose $25 million.

AI-powered voice impersonation is making phone scams more believable and dangerous than ever, says Steve Grobman, CTO at cybersecurity firm McAfee. Globally, one in four people has either fallen victim to or knows someone who has been scammed using AI, with 77% of victims losing money, according to McAfee’s “Beware the Artificial Imposter” report.

Cybersecurity experts have found that 6.5 billion fraudulent calls are made worldwide each quarter, with the average American receiving 12 scam calls per month. The “modernization” of this criminal industry was inevitable. Breachsense CEO Josh Amisav explains that with the advancement of AI text-to-speech technology, fraud techniques will become significantly more sophisticated. When combined with leaked or stolen data, scammers can impersonate someone the victim knows, adding personal details to increase the scam’s credibility. These scams have grown so prevalent that an Arizona senator proposed legislation to treat AI as a weapon when used to commit crimes, in order to impose harsher penalties on scammers.

It doesn’t take much for scammers to get into the audio deepfake business—just a few hundred euros. Several companies offer AI and deep learning applications that can create voice replicas of such high quality that they can fool anyone. These technologies were originally developed for legitimate uses, like movie production, audiobooks in different languages, and lending celebrity voices to advertisements or other creative applications. For example, a restaurant chain in the US uses the cloned voice of famous football player Keith Byars to take orders, and a South Korean company offers to recreate the voices of deceased loved ones for their families. On the positive side, the technology is also used for health purposes, such as the Voice Keeper app, which helps people suffering from conditions like throat cancer, Parkinson’s, or ALS to preserve their voices. But it was only a matter of time before this technology was misused.

The Scenarios Used

With voice mimicry technology, fraudsters now have a wider range of scenarios to exploit in order to achieve their ultimate goal—extracting as much money as possible from their victims.

The most common scam scenario in the US involves fake kidnappings. Scammers often target families with young children, convincing them their child has been kidnapped and that they must pay a ransom. These payments are frequently made via anonymous apps or even in person, as in real kidnappings. Initially, large sums (over $1 million) are demanded, but the criminals often settle for much smaller amounts (like $50,000).

In other cases, the stolen voice is used to extract money in different ways. Using AI, scammers call a friend or family member, who hears their “loved one” claim they were in a car accident in a remote area. “I’m hurt and need surgery. I’ve lost everything—wallet, cards, phone—and the hospital is asking for X amount to proceed because insurance won’t cover it. Please send the money to this account immediately; I’ll repay you once I’m out of surgery.”

Sometimes, “professional” voices are used to convince victims to transfer large sums of money quickly. Who wouldn’t panic if they heard their accountant say they must pay a large tax bill to avoid confiscation or jail? Or if a lawyer called for something similar? Scammers even target business accountants, using the voice of a company’s CEO to authorize payments or fulfill fake obligations.

Banks are on high alert, not only to block suspicious transactions but also because AI is becoming more convincing in mimicking human voices. This makes it increasingly likely that criminals will use it to gain access to victims’ bank accounts, perfecting existing scams and bypassing current security measures. That’s why customers are now required to confirm personal details like tax IDs and identity numbers during phone calls with their banks.

Similarly, voice cloning could be used—like the case involving fake Joe Biden calls—to spread misinformation or attempt to manipulate previously secure processes, such as elections. But that’s a whole different category of concern.

Keys to Protection

Experts suggest that while apps capable of detecting scam calls or AI usage are evolving rapidly, “there will never be a silver bullet.” The best protection is for individuals to be cautious and proactive, ensuring it’s practically impossible for them to fall victim.

>Related articles

From MAGA to Make Europe Great Again, with support for patriotic parties and a “stop” on mass immigration – How to stop the onslaught of China

Billionaire Andrej Babis reappointed Prime Minister of the Czech Republic on Tuesday

US Ultimatum to the EU: Take on the majority of NATO spending by 2027

The most important step is recognizing scam calls. As with current fraud schemes, being cautious, avoiding disclosing personal information, and safeguarding data that’s often exposed online are crucial. “If you suspect it’s a scam, it’s better to hang up and stop the conversation.”

Perhaps the most practical advice is to use a family password. Experts recommend that families agree on one or more words or phrases to be used in emergencies. These passwords should never be shared with third parties or posted on social media. They should also be words or phrases that no one outside the family would know.

This way, grandparents, parents, or children can be sure that the person on the other end of the phone isn’t really their loved one saying, “I’m being held hostage,” or “I’m in serious trouble and need money immediately.”

Ask me anything

Explore related questions

#AI#cloning#europe#fraud#US
> More technology

Follow en.protothema.gr on Google News and be the first to know all the news

See all the latest News from Greece and the World, the moment they happen, at en.protothema.gr

> Latest Stories

History has treated her unfairly”: The 400-year mystery surrounding Shakespeare’s wife and son

December 6, 2025

Clash between two professors over a female student: Vulgar flyers, phone calls for “dates,” and slashed tires

December 6, 2025

Alexandros Grigoropoulos: 17 years since the murder that marked an entire generation

December 6, 2025

Enipeas River overflows: 112 issues evacuation order for settlement near Farsala – Four areas on alert

December 6, 2025

Papastavrou: The ministerial meeting of the Greece, Cyprus, Israel and the USA group in Washington in April

December 5, 2025

European Commission handbook depicts the East Aegean islands and the Dodecanese as Turkish

December 5, 2025

Anger in Cyprus over the UN Secretary General’s envoy: She described the occupied territories as the “Turkish” side of Cyprus

December 5, 2025

From MAGA to Make Europe Great Again, with support for patriotic parties and a “stop” on mass immigration – How to stop the onslaught of China

December 5, 2025
All News

> Sports

World Cup 2026: The time of the draw in Washington

FIFA is organizing a dazzling ceremony at the Kennedy Center in Washington for the World Cup draw, with the participation of heads of state and global stars – it begins at 19:00 and will be broadcast by ERT2 Sports

December 5, 2025

Greek Cup: Schedule of the final matchday of the League Phase

December 5, 2025

Nikos Boudouris returns to PAOK as the club’s new general manager

December 5, 2025

Is Olympiacos Moving to Cyprus

December 5, 2025

Panathinaikos FC: “There is absolutely no approach from anyone regarding the acquisition of the share package”

December 5, 2025
Homepage
PERSONAL DATA PROTECTION POLICY COOKIES POLICY TERM OF USE
Powered by Cloudevo
Copyright © 2025 Πρώτο Θέμα