Phone scams have been around since practically the invention of the landline. But there’s a dangerous new element making scam calls even more convincing: the use of deepfake voice technology.
A deepfake can be any type of media — audio, video, image— that has been digitally manipulated, typically for the purpose of misleading people or changing the original meaning. A voice deepfake, sometimes known as AI voice cloning, uses artificial intelligence to construct spoken sentences that sound like people saying something they did not actually say. This voice track can even be placed underneath real video footage, making it appear as if a person has endorsed or carries an opinion contrary to their own. The coupling of real video and false audio can make these deepfakes more challenging to detect and flag.
Deepfake voice scams aren’t new, but AI is making them more real and easier to create. In the past, deepfakes required a significant amount of time and technical expertise. Today, there are more than a dozen AI voice cloning tools freely available on the internet. These tools require only a basic level of expertise to use.
RELATED: See how our Chief Product Officer Faked the Voice of Hiya's CEO
Companies are beginning to take notice of this trend. In a survey conducted by the identity verification firm Regula, 37% of worldwide organizations have experienced voice deepfake identity fraud, and 82% or organizations think that voice deepfakes will be a growing threat in the next two years.
According to a 2023 study conducted by researchers from the Department of Security and Crime Science at University College London in the UK, humans cannot reliably detect speech deepfakes. In an experiment conducted with more than 500 participants, individuals listened to real and fake audio clips and attempted to differentiate between them. The result: listeners could only correctly identify voice deepfakes 73% of the time.
The study concluded, “As speech synthesis algorithms improve and become more realistic, we can expect the detection task to become harder. The difficulty of detecting speech deepfakes confirms their potential for misuse and signals that defenses against this threat are needed.”
The battle against deepfakes and AI voice scams needs to be fought on multiple fronts. Governments play a role too. In August, the European Union’s AI Act was enacted. It’s the world’s first major law regulating the use of AI, and it governs the way companies develop, use and apply AI.
According to NBC News, “The legislation applies a risk-based approach to regulating AI, which means that different applications of the technology are regulated differently depending on the level of risk they pose to society.” Examples of high-risk AI systems include autonomous vehicles, medical devices, loan decisioning systems, educational scoring, and remote biometric identification systems.
The United States is also drafting laws to protect against impersonations and deepfakes. Earlier this year, the US Federal Trade Commission sought public comment on a proposal that would prohibit the impersonation of individuals, which would extend protections of a new rule on government and business impersonation.
According to the FTC, “The agency is taking this action in light of surging complaints around impersonation fraud, as well as public outcry about the harms caused to consumers and to impersonated individuals. Emerging technology — including AI-generated deepfakes — threatens to turbocharge this scourge, and the FTC is committed to using all of its tools to detect, deter, and halt impersonation fraud.”
Even US states are passing legislation to combat deepfakes. Recently, California Governor Gavin Newsom signed into law Assembly Bill 2839 (AB 2839). The law aims to address the rise of AI-generated disinformation ahead of 2024 elections, stating that AB 2839 is crucial for protecting the integrity of elections and restoring voter trust.
Although scams carried out using deepfake speech synthesis pose a danger for both individuals and business, not all is lost. Fortunately, there is a way to identify deepfakes. While fraudsters are using AI for sinister purposes, the same technology can also be used for good.
Earlier this month, Hiya acquired Locus.ai, the industry leader in deepfake detection solutions. The combination of Loccus.ai’s voice intelligence technology and the Hiya’s Adaptive AI fraud prevention system, provides businesses and carriers with the most complete fraud call protection available across the industry today. This versatile solution supports multiple languages, formats, and platforms, including video recordings and live calls across all devices.
Loccus.ai’s deepfake voice detection technology is now available as Hiya AI Voice Detection. Businesses and carriers can immediately integrate AI voice detection into their own services and applications to enhance consumer trust and security. With Hiya, it’s never been easier to identify deepfake audio in personal communications apps, news and social media content, video conferencing contact center platforms, and more.