Tips for Better Security Now That AI Clones Our Voices 

Why Trust Techopedia

Spotify announces the ability to use artificial intelligence to translate podcasts - using the podcaster's unique voice - and third-party services now allow us to quickly impersonate someone else. But in the rush to bring exciting tools to market, we open a new Pandora's box of foundational trust and security issues.

Spotify recently announced an AI-driven Voice Translation feature that enables subscribers to hear global podcasts in their native language while retaining the podcaster’s original voice style. Partnering with big names like Dax Shepard and Lex Fridman, the pilot starts with Spanish, French, and German translations. The move aims to break language barriers and deepen listener engagement worldwide. 

Elsewhere, OpenAI also revealed that ChatGPT will soon be able to see, hear, and speak, allowing users to have voice conversations with ChatGPT.

But this is just the beginning. Have you ever fantasized about crooning in the vocal stylings of your musical idol? Innovator and experimental musician Holly Herndon has turned that dream into a near reality by training an AI model to replicate her distinctive voice. At a recent TED Talk, she revealed Holly+, an AI-enabled tool that gives users the power to sing in her voice. 

In a live demo that captured the audience’s attention, artist PHER used two microphones. One picked up his natural voice, while the other transformed it to sound like Holly’s. The effect was like a captivating duet when he sang into both mics. Depending on your perspective, this could be seen as a groundbreaking musical experiment or just digital mimicry.

Move over deepfake videos. The stage is set for AI voice cloning to dominate conversations in the months ahead. However, as someone who adopts a vigilant stance on technology’s unintended consequences, entrusting big tech with the capability to duplicate our vocal identity opens up another Pandora’s box of concerns that seem to follow every new AI solution. 

From Viral Pranks to Security Concerns

Earlier this year, Social media influencer Mia Dio recently hit the headlines for a TikTok video that showed her using AI voice-cloning technology to catch her boyfriend, Billy, cheating. In the viral video, Mia used a cloned voice of Billy to speak with his friend and inquire about his activities from the previous night. The friend seemingly confirmed Billy’s infidelity. However, Mia later revealed the entire scenario was a staged prank involving all parties. The video sparked conversations about the ease and ethical implications of voice cloning, as Mia pointed out that she used the ElevenLabs software to convincingly mimic Billy’s voice for just $4


The rise of voice cloning technology threatens to undermine one of the most fundamental aspects of personal and financial security: trust. The euphoria surrounding voice-based banking’s convenience and accessibility is dampened by high-profile incidents, such as a UK energy firm CEO being defrauded out of €220,000. 

More recently, technologist Rob Williams made a shockingly realistic AI clone of his voice in less than a minute, using free tools anyone can access online. The voice clone was so convincing it even tricked his wife. This recording offers a timely reminder that if a loved one suddenly calls you verbally asking you to transfer some money, don’t just trust your ears—consider setting up a ‘safe word’ for extra security.

Voice Cloning vs. Voice Banking: The Security Struggle Within Fintech

Voice-cloning technologies like VALL-E and ElevenLabs are not just hobbyist tools; they are potent utilities that can create near-perfect replicas of human voices in seconds. The ramifications extend beyond individual fraud cases. A single convincing deepfake audio could initiate cascading failures in a security protocol, instigate stock market upheavals, or disrupt diplomatic relations. The ease and speed of mimicking voices are accelerators for nefarious activities, including disinformation campaigns that can have widespread societal impacts.

While voice-based banking systems are growing at an impressive rate, the corresponding advancements in security protocols have not kept pace. Many financial institutions are still grappling with the intricacies of multi-factor authentication and voice biometrics, which are far from foolproof against sophisticated AI voice-cloning attacks.

This lapse in security exposes customers to financial risks and compromises the integrity of voice-enabled systems, which could result in a crisis of confidence among users. The banking industry arguably got distracted by shiny new tech and providing frictionless experiences. But they must temper its innovation drive with an equally robust emphasis on fortifying security measures. As we navigate this promising but perilous landscape, it becomes crucial for consumers and businesses alike to prioritize security vigilance over the dazzle of cutting-edge features.

Security Tips To Prevent AI Voice Cloning

In the evolving cybercrime landscape, AI voice cloning stands out as an increasingly prevalent method for carrying out scams. To protect yourself, always be cautious of unsolicited calls that request personal or financial information. If you feel pressured by high-stakes emotional appeals, this is a red flag—hang up and contact the person or entity directly using a known phone number.

Be proactive by creating a unique “family password” that only trusted members know, providing an immediate verification step in case of suspicious activity. For an added layer of caution, practice asking detailed questions that only the genuine caller can answer. Technology alone isn’t a foolproof safety net; staying informed about the latest scam techniques and a healthy dose of skepticism will provide the best defense in a world where our voices can be convincingly mimicked.

The Bottom Line

As AI continues to capture headlines and our imaginations, it’s crucial to remember that big tech’s obsession with moving fast and breaking things isn’t new. But what is different now are the stakes involved. This isn’t just about tweaking algorithms that curate our social media feeds or fuel our online echo chambers; it’s a question of foundational trust that touches the core of our personal and financial lives. 

If we fail to manage and regulate this brave new world, we risk crafting a digital dystopia where the lines between trust and deception blur and where we find ourselves questioning not just half of what we see but also every sound we hear.


Related Reading

Related Terms

Neil C. Hughes
Senior Technology Writer
Neil C. Hughes
Senior Technology Writer

Neil is a freelance tech journalist with 20 years of experience in IT. He’s the host of the popular Tech Talks Daily Podcast, picking up a LinkedIn Top Voice for his influential insights in tech. Apart from Techopedia, his work can be found on INC, TNW, TechHQ, and Cybernews. Neil's favorite things in life range from wandering the tech conference show floors from Arizona to Armenia to enjoying a 5-day digital detox at Glastonbury Festival and supporting Derby County.  He believes technology works best when it brings people together.