Deepfakes: For now women, not democracy, are the main victims

Deepfakes pose a serious threat to democracy in the long run but women are likely to suffer first, a new study says.

Deepfake tech is mainly being used for porn and women are the main victims Deepfakes pose a serious threat to democracy in the long run but women are likely to suffer first, a new study says.

While the 2020 US presidential elections have lawmakers on edge over AI-generated fake videos, a new study by Netherlands-based deepfake-detection outfit Deeptrace shows that the main victims today are women. 

Reddit, the site where the term deepfake was coined, in early 2018 banned deepfake porn or 'involuntary porn', which largely relies on generative adversarial networks (GANs) to insert the face of a woman into existing pornographic material. 

Despite Reddit's influence on internet culture the ban hasn't stopped the emergence of deepfake porn. New apps like DeepNude – an app that quickly undressed a woman in an image – have since popped up as techies experiment with GANs. The problem is that the apps can be used to harass and intimidate women.

SEE: Transgender employees in tech: Why this "progressive" industry has more work to do to achieve true gender inclusivity (TechRepublic cover story)    

Although there's the obvious threat that deepfakes pose to democratic processes, the tech is also being applied in unexpected ways, such as CEO fraud using synthesized voice – a hugely costly type of fraud that has traditionally played out over email.  

According to Deeptrace, deepfake videos have exploded in the past year, rising from 8,000 in December 2018 to 14,678 today. And not surprisingly for the internet, nearly all of the material is pornography, which accounts for 96% of the deepfake videos it's found online. The fake videos have been viewed 134 million times. 

The numbers suggest deepfake porn is still niche but also growing quickly. Additionally, 90% of the fake content depicted women from the US, UK, and Canada, while 2% represented women from South Korea and 2% depicted women from Taiwan.

"Deepfake pornography is a phenomenon that exclusively targets and harms women," the company notes. 

That small number of non-pornographic deepfake videos it analyzed on YouTube mostly contained (61%) synthesized male subjects. 

According to Henry Ajder, a researcher at Deeptrace, currently most of the deepfake porn involves famous women. But he reckons the threat to all women is likely to increase as it becomes less computationally expensive to create deepfakes. 

"As the generative technologies that support deepfakes become increasingly commodified, it is highly likely more private individuals will be targeted," Ajder told ZDNet. 

"This commodification process will make the technology significantly more accessible to bad actors and almost certainly increase the quantity of deepfakes being produced."

As for the political threat, there actually aren't that many cases where deepfakes have changed a political outcome.

SEE: Google's war on deepfakes: As election looms, it shares ton of AI-faked videos  

The only two cited in the report happened in Gabon and Malaysia. The case in Malaysia generated a sex video involving a cabinet minister, while the incident affecting Gabon involved a video released by the government of its president Ali Bongo Ondimba after he'd suffered a stroke.  

Ajder thinks the political threat is currently in the "near future" but argues that the idea of a deepfake is already destabilizing political processes.

"I'd say deepfakes, cybersecurity, and political applications pose very serious near-term risks that we need to prepare for now, even if they aren't causing utter havoc right now in the way some people may think," he said. 

More on deepfakes and security

  • Facebook, Microsoft: We'll pay out $10m for tech to spot deepfake videos
  • Forget email: Scammers use CEO voice 'deepfakes' to con workers into wiring cash
  • 'Deepfake' app Zao sparks major privacy concerns in China
  • AI, quantum computing and 5G could make criminals more dangerous than ever, warn police
  • Samsung uses AI to transform photos into talking head videos
  • Facebook's fact-checkers train AI to detect "deep fake" videos
  • The lurking danger of deepfakes TechRepublic
  • These deepfakes of Bill Hader are absolutely terrifying CNET