Deepfakes and Fake News Pose a Growing Threat to Democracy: Experts

In 2018, he co-authored an article, titled, “Fake news is counterfeit news,” with Northeastern philosophy professor Kay Mathiesen. The article looked at the threat to democracy and knowledge that fake news poses and sought to define the concept. Two years later, he wrote an article about deepfakes, “The Epistemic Threat of Deepfakes,” in which he concluded that deepfakes can lead to false beliefs, undermine the justification for true beliefs, and prevent people from acquiring true beliefs.

Fallis argues both fake news and deepfakes have the negative effect of delegitimizing real news. He says they decrease the amount of true information available, reduce consumers’ trust in authentic media, and put an added burden on fact-checkers to authenticate the vast amount of content online.

“In the case of fake news, you’re creating this online presence that’s supposed to look like a legitimate news site,” Fallis says. “Similarly, in the case of deepfakes, you’re creating video and audio that are also supposed to look like legit media.”

Additionally, in combination with tools used to collect individual users’ personal information en masse, deepfakes also can be used maliciously to target large audiences and manipulate them by playing on their ingrained biases, Fallis says.

“It may not just be this one killer technology,” he says. “It’s not like deepfakes are going to be the one thing that takes us over the cliff. It’s a whole suite of potentially problematic technology.”

Increased political division has a similar impact on the way in which people interpret fake news, where users are clearly seeking out and accepting information that’s compatible with their prior biases, notes Northeastern political science and computer sciences professor David Lazer. However, it’s unclear just how much one drops their critical-thinking skills when encountering media that reinforces their worldview.

“Certainly, we’ve seen an increased polarization in the public’s opinions, and that’s clearly one of the factors that may be at play with the spread of misinformation,” Lazer says. “It’s quite plausible that the political polarization and spread of misinformation are going hand in hand, but that’s an area of needed research.”

Director of Northeastern’s Lazer Lab, which conducts research on social influence and networks, Lazer’s studies focus primarily on the proliferation of misinformation on social media. In 2019, he co-authored a study on the prevalence of fake news on Twitter during the 2016 presidential election cycle.

Deepfake technology is also “quite relevant” to his studies, Lazer says, but there needs to be more research on the different types of misinformation, how they spread, and their psychological impact on consumers of media. The rise in political polarization and its impact on the consumption of media is also a high-priority area of study, he adds.

“We can certainly say over the last 40 years there has been increased polarization of many kinds, and that’s concerning,” Lazer says.

Beyond the issue of users failing to question the deepfakes they come across if the content confirms their existing worldview, the technology poses other significant concerns.

One of the most problematic uses of the technology is when an individual’s likeness, typically a woman’s, is manipulated and put on a sexually explicit video, making it appear as if the individual they are targeting is participating in the sexual activity, says Marc Berkman, the executive director of the Organization for Social Media Safety, a nonprofit dedicated to making social media safe through advocacy and education.

Additionally, as in the case of the deepfake of Zelenskyy, the world is witnessing the technology’s political impact, Berkman says. Deepfakes can potentially interfere with democratic elections and be used as propaganda to sow division and doubt, he says.

Fallis and Berkman emphasize the importance of users cultivating critical-thinking skills when venturing online. One way for people to protect themselves against deepfakes is to engage in safe social-media use: Approach content, particularly news, with a critical eye. 

The Organization for Social Media Safety is currently supporting media training in public schools, helping children understand news sources so they can take a non-partisan approach to evaluating and understanding the credibility of content.

“It’s incredibly important for our democracy to understand what is real and what is not,” Berkman says. “Limiting time on social media to healthy amounts is also important, so people can avoid deepfakes used for propaganda purposes.”

However, Fallis and Berkman note, individual efforts can’t replace structural change in businesses and governments aimed at combating the proliferation of this potentially dangerous technology. 

Social-media giants, like Facebook, have adopted policies vowing to remove deepfakes from their platforms if they meet certain criteria, and some state governments, like California’s, have adopted laws imposing civil liability on creators of intentionally harmful deepfakes.

In California, Berkman says, his organization is working on getting a state law passed that would also impose criminal punishment on the creators of malicious pornographic deepfakes, with the hope that this kind of law expands to other states and that the federal government adopts similar legislation.

Jackson Cote is a contributor to News@Northeastern. This article was originally posted to News@Northeaster.