Robocalls made using AI-generated voices are now illegal in the US, according to a new ruling by the Federal Communications Commission (FCC).
The move comes in the wake of a significant rise in AI voice cloning scams, Spokeo reports. In these fairly sophisticated scams, criminals use AI-generated voices to impersonate politicians, celebrities or even close family members with the ultimate goal of convincing victims to comply with some fraudulent request, like sending them cash.
Here's what you need to know about AI voice scams, including how they work and, most importantly, how to protect yourself.
AI scams leverage artificial intelligence to mimic human behavior, language, interactions and even decision-making processes. This then enables scammers to execute various deceptive schemes aimed at defrauding individuals and organizations.
These scams can take several forms, including:
A specific type of AI voice scam that has been making a lot of headlines recently involves criminals cloning the voice of a family member and then using it to convince a loved one to send the scammer money.
This is a relatively elaborate attack that involves several phases.
First, the attacker finds a voice sample of the person they are trying to impersonate. They can get it from a variety of sources, including the person's social media or from public interviews they've participated in. Recent advances in fake voice generation technology means that sometimes, all a scammer needs is just a few seconds of audio recording.
The scammer then feeds this sample into a voice cloning tool which creates a digital replica of the voice capable of producing speech that sounds remarkably similar to the original. Next, the criminal uses the AI-generated voice to record audio stating that they are in a difficult situation and need money.
They might fabricate scenarios such as being involved in an accident and needing cash to settle with the other driver, getting arrested and requiring money for bail, or being stranded in an unfamiliar or dangerous area and needing funds to purchase a plane ticket home.
The deceitful voice message is then sent to the intended victim, typically via voicemail. The voice can be so persuasive that many recipients will send money without any hesitation after hearing it.
While it can be difficult to differentiate a fake voice from a real one, there are several proactive measures you can take to protect yourself from becoming a victim of an AI voice cloning scam.
Verify the caller's identity
If you receive a suspicious or unexpected voicemail, particularly one that involves money, take a moment to verify the caller through other means. For example, reach out to the person of interest using a different number you've previously used to communicate with them. You can also ask them to call you back at a number you know is theirs or trust.
Treat all urgent requests with suspicion
Scammers will always try to create a sense of urgency with their fraudulent requests. They want you to overlook your better judgment and make a hasty decision. If the caller urges you to act fast, that's a red flag.
Trust your instincts
If something feels off, trust your instincts. It's better to err on the side of caution and take steps to verify the authenticity of a call than to become the victim of a scam.
Educate yourself about AI voice scams
Knowledge and awareness are crucial defenses against AI scams. Stay updated with all the latest trends in AI scams. Share your knowledge or experiences with others to build a communal defense against these scams.
Limit personal information sharing
Another way to avoid falling prey to AI voice scammers is to limit the amount of personal information you share online. Remember that scammers can use personal details they've taken from social media and other online sources to make their impersonations more believable.
Check for anomalies in the voice message
Even though AI-voice clones can be highly convincing, they might still exhibit subtle anomalies or inconsistencies. For example, pay attention to unusual pauses, odd tones and cadences, or background noises that don't fit the context of the message.
Report suspected scams
If you encounter what you believe to be an AI voice or deepfake scam, report it to the appropriate authorities, such as the Federal Trade Commission (FTC) or your local consumer protection agency. Reporting can help authorities track the perpetrators and thus combat these scams.
Understanding how AI voice cloning scams work and implementing some of the preventative measures outlined here can reduce the risk for you and those around you.
If you've already fallen victim, the best thing to do is take immediate action. Report the matter to your bank and the appropriate authorities. You might still be able to recover your funds, though it's not guaranteed. Sharing your experience with others can also help raise awareness, making it harder for scammers to find new victims.
Looking ahead, reverse phone lookup tools like Spokeo can offer protection against AI scammers. If someone leaves you a voicemail claiming to be a family member in an emergency, a quick reverse phone lookup through Spokeo can reveal the real owner of the number. This can then help you determine whether the request is legitimate, or if you're dealing with a scammer.
This story was produced by Spokeo and reviewed and distributed by Stacker Media.