Deepfakes rewrite the cybersecurity playbook
Fakes aren’t just for April Fool’s anymore. They’re coming to a boardroom near you, with the potential to destabilize your businesses.

A chief executive officer (CEO) calls the head of finance and asks for a large deposit into a supplier’s account.
A senior executive shows up on a company video call and leads a discussion around product pipeline and IP.
A customer verifies her identity via facial recognition and makes a large securities trade.
What do these scenarios have in common? They’re examples of modern-day deepfake scams. The principal players—CEO, senior executive, and customer—were not real. They were created using generative artificial intelligence (AI) and deployed using techniques like voice-based phishing (vishing) and Zoombombing.
Simply put, deepfakes are sophisticated forgeries (video, audio, imagery, and voice) generated through cutting-edge AI techniques like Generative Adversarial Networks (GANs) that manipulate existing visual and auditory content, crafting seemingly genuine yet entirely fabricated material. Beyond the obvious concerns in social media and entertainment, deepfakes have become a pressing issue for businesses globally, as we detail in our new report, “Fake content is becoming a real problem.”
Business leaders are taking note.
A recent KPMG survey of 300 executives revealed that a staggering 92 percent of them are significantly worried about the risks posed by deepfakes. And a survey of cybersecurity professionals found that two-thirds of companies had experienced a deepfake-related incident over the previous year.
Despite the awareness, though, most companies have been slow to act. One survey found that only 29 percent of organizations have taken specific actions to protect themselves, and 46 percent are still formulating their plans.
Meantime, the risks aren’t small or isolated. In 2021, the FBI predicted that “malicious actors almost certainly will leverage synthetic content for cyber and foreign influence operations in the next 12 to 18 months.” Indeed, the consequences could be vast, with financial, reputational, and geopolitical impacts.
The time is now to double-down on security and address deepfake risks head on. How can you protect yourself and your business? Here’s what leaders need to know.
The rise and risk of deepfakes
The wide-ranging promise of generative Al (GenAI) is genuinely exciting. However, the technology underlying these advances is now also available to those who would use it for mischief—or worse. Three factors are colliding to heighten the risks:
Factor #2: Rise of remote work
Work from home has heightened the risk to organizations. Home-based employees—who work in silos and are often disconnected from the heartbeat of the organization—are particularly vulnerable to deepfake attacks. According to one survey, the average impact of a data breach is $1 million higher when a remote working arrangement is involved.
The FBI warns that deepfakes are now being used by bad actors to apply for remote information technology (IT) and programming positions in an effort to gain access to personal, financial, or other proprietary information.
Factor #3: Human fallibility
Employees are well aware of email- and text-based phishing scams at this point, yet many are still buying gift cards when the CEO-bot texts them with a request. A recent study found that 54 percent of employees fell for a phishing scam in the previous year because the email looked legit; more than half said they followed through because the request came from a senior executive.
The lesson: Humans are unpredictable and fallible. As deepfakes become more sophisticated, the stakes are getting higher by the day. Annual training modules aren’t enough. Cybersecurity in the deepfakes age requires persistent personal engagement, behavior reinforcement, easy-to-use internal controls, and—to the extent possible—the elimination of human decision-making.
A new security posture for the deepfakes age
With deepfakes poised to destabilize organizations financially and operationally, and current authentication capabilities largely insufficient to deal with the steady advance of deepfake technology, enterprises will need to adopt new processes, tools, and strategies to better secure their systems, data, and infrastructure.
Think of it as a cloud-based zero-trust future, in which employees no longer need to be “on network” through persistent VPN connections. Access will become conditional, depending on the person and role. In this new reality, security professionals need to remain vigilant. Here are a few steps organizations can take to implement this new security posture.
1
2
3
4
5
Today’s rapid pace of innovation often results in security being left behind in business transformation efforts: for example, 4 in 10 global IT leaders do not believe that their organization’s security measures have kept pace with their digital transformation initiatives.
Organizations have lagged in their ability to predict, detect, and respond to these threats in an automated manner. The key for security professionals is understanding—and keeping up with—the pace and speed at which cybercriminals use these tools.
A deep response
The smartest businesses don’t just manage cyber risk. They use it as a source of growth and competitive edge. Technology makes many things possible, but what’s possible isn’t always safe.
Your cybersecurity function must build resilience and trust as threats grow in volume and sophistication, and as new technology becomes essential for meeting the needs of your customers, employees, suppliers, and society.
To learn more about the threats posed by deepfakes, and how you can create a resilient and trusted digital-first enterprise in the face of these evolving threats, explore our additional insights below.
Explore more insights and opportunities:

Fake content is becoming a real problem
Widespread availability of sophisticated computing technology and AI enables virtually anyone to create highly realistic fake content.

KPMG generative AI survey report: Cybersecurity
An exclusive KPMG survey examines four areas where this remarkable technology shows great promise.

Synthetic identity fraud
A $6 billion problem
Meet our team

Subscribe to receive the KPMG Opportunity (In)sight Newsletter
Turn insight into opportunity with unique perspectives and actionable insights addressing the burning issues atop the C-suite agenda. Delivered monthly.