5 Deepfake Scams That Threaten Enterprises

An AI performing deepfake technology.
Picture: metamorworks/Adobe Inventory

A brand new report from Forrester is cautioning enterprises to be looking out for five deepfake scams that may wreak havoc. The deepfake scams are fraud, inventory worth manipulation, status and model, worker expertise and HR, and amplification.

Deepfake is a functionality that makes use of AI know-how to create artificial video and audio content material that might be used to impersonate somebody, the report’s writer, Jeff Pollard, a vice chairman and principal analyst at Forrester, advised TechRepublic.

The distinction between deepfake and generative AI is that, with the latter, you sort in a immediate to ask a query, and it probabilistically returns a solution, Pollard stated. Deepfake “…leverages AI … however it’s designed to provide video or audio content material versus written solutions or responses that a big language mannequin” returns.

Deepfake scams focusing on enterprises

These are the 5 deepfake scams detailed by Forrester.


Deepfake applied sciences can clone faces and voices, and these methods are used to authenticate and authorize exercise, in response to Forrester.

“Utilizing deepfake know-how to clone and impersonate a person will result in fraudulent monetary transactions victimizing people, however it’ll additionally occur within the enterprise,” the report famous.

One instance of fraud can be impersonating a senior govt to authorize wire transfers to criminals.

“This situation already exists in the present day and can enhance in frequency quickly,” the report cautioned.

Pollard referred to as this probably the most prevalent sort of deepfake “… as a result of it has the shortest path to monetization.”

Inventory worth manipulation

Newsworthy occasions could cause inventory costs to fluctuate, resembling when a widely known govt departs from a publicly traded firm. A deepfake of this kind of announcement might trigger shares to expertise a brief worth decline, and this might have the ripple impact of impacting worker compensation and the corporate’s capacity to obtain financing, the Forrester report stated.

Repute and model

It’s very straightforward to create a false social media submit of “… a distinguished govt utilizing offensive language, insulting clients, blaming companions, and making up details about your services or products,” Pollard stated. This situation creates a nightmare for boards and PR groups, and the report famous that “… it’s all too straightforward to artificially create this situation in the present day.”

This might harm the corporate’s model, Pollard stated, including that “… it’s, frankly, virtually inconceivable to stop.”

Worker expertise and HR

One other “damning” situation is when one worker creates a deepfake utilizing nonconsensual pornographic content material utilizing the likeness of one other worker and circulating it. This could wreak havoc on that worker’s psychological well being and threaten their profession and can “…virtually actually end in litigation,” the report said.

The motivation is somebody considering it’s humorous or on the lookout for revenge, Pollard stated. It’s the rip-off that scares corporations probably the most as a result of it’s “… probably the most regarding or pernicious long run as a result of it’s probably the most tough to stop,” he stated. “It goes in opposition to any typical worker conduct.”


Deepfakes can be utilized to unfold different deepfake content material. Forrester likened this to bots that disseminate content material, “… however as a substitute of giving these bots usernames and submit histories, we give them faces and feelings,” the report stated. These deepfakes may be used to create reactions to an authentic deepfake that was designed to break an organization’s model, so it’s probably seen by a broader viewers.

Organizations’ finest defenses in opposition to deepfakes

Pollard reiterated you could’t forestall deepfakes, which may be simply created by downloading a podcast, for instance, after which cloning an individual’s voice to make them say one thing they didn’t really say.

“There are step-by-step directions for anybody to do that (the flexibility to clone an individual’s voice) technically,” he famous. However one of many defenses in opposition to this “… is to not say and do terrible issues.”

Additional, if the corporate has a historical past of being reliable, genuine, reliable and clear, “… it is going to be tough for individuals to consider suddenly you’re as terrible as a video would possibly make you look like,” he stated. “However if in case you have a observe document of not caring about privateness, it’s not arduous to make a video of an govt…” saying one thing damaging.

There are instruments that provide integrity, verification and traceability to point that one thing isn’t artificial, Pollard added, resembling FakeCatcher from Intel. “It appears to be like at … blood stream within the pixels within the video to determine what somebody’s considering when this was recorded.”

However Pollard issued a word of pessimism about detection instruments, saying they “… evolve after which adversaries get round them after which they must evolve once more. It’s the age-old story with cybersecurity.”

He burdened that deepfakes aren’t going to go away, so organizations have to suppose proactively concerning the risk that they may turn into a goal. Deepfakes will occur, he stated.

“Don’t make the primary time you’re serious about this when it occurs. You need to rehearse this and perceive it so precisely what to do when it occurs,” he stated. “It doesn’t matter if it’s true – it issues if it’s believed sufficient for me to share it.”

And a closing reminder from Pollard: “That is the web. Every thing lives without end.”