Battling the “Pandemic of Misinformation”

Mainstream media coverage has added to the problem, analysts say. At many major news outlets, reporters and editors with no medical or public health training were reassigned to cover the unfolding pandemic and are scrambling to get up to speed with complex scientific terminology, methodologies, and research, and then identify, as well as vet, a roster of credible sources. Because many are not yet knowledgeable enough to report critically and authoritatively on the science, they can sometimes lean too heavily on traditional journalism values like balance, novelty, and conflict. In doing so, they lift up outlier and inaccurate counterarguments and hypotheses, unnecessarily muddying the water.

“That’s a huge challenge,” said Ashish Jha, K.T. Li Professor of Global Health and Director of the Harvard Global Health Institute, during an April 24 talk about COVID-19 misinformation hosted by the Technology and Social Change Research Project at the Shorenstein Center for Media, Politics and Public Policy.

“What I have found is a remarkable degree of consensus among people who understand the science of this disease around what the fundamental issues are and then disagreements about trade-offs and policies,” said Jha, who is a frequent commentator on news programs. “The idea of covering the science in a two-sided way on areas where there really isn’t any disagreement has struck me as very, very odd, and it keeps coming up over and over again.”

Then there is the problem of political bias. This has been especially true at right-leaning media outlets, which have largely repeated news angles and viewpoints promoted by the White House and the president on the progress of the pandemic and the efficacy of the administration’s response, boosting unproven COVID-19 treatments and exaggerating the availability of testing and safety equipment and prospects for speedy vaccine development.

Tara Setmayer, a spring 2020 Resident Fellow at the Institute of Politics and former Republican Party communications director, said what’s coming from Fox News and other pro-Trump media goes well beyond misinformation. Whether downplaying the views of government experts on COVID-19’s lethality, blaming China or philanthropist Bill Gates for its spread, or cheering shutdown protests funded by Republican political groups, it’s all part of “an active disinformation campaign,” she said, aimed at deflecting the president’s responsibility as he wages a reelection campaign.

But turning around those who buy into false information is not as simple as piercing epistemic bubbles with facts, said Christopher Robichaud, senior lecturer in ethics and public policy at Harvard Kennedy School (HKS) who teaches the Gen Ed course “Ignorance, Lies, Hogwash and Humbug: The Value of Truth and Knowledge in Democracies.”

Over time, bubble dwellers can become cocooned in a media echo chamber that not only feeds faulty information to audiences, but anticipates criticisms in order to “prebut” potential counterarguments that audience members may encounter from outsiders, much the way cult leaders do.

“It’s not enough to introduce new pieces of evidence. You have to break through their strategies to diminish that counterevidence, and that’s a much harder thing to do than merely exposing people to different perspectives,” he said.

While Facebook, Twitter, and YouTube have all recently ramped up efforts to take down COVID-19 misinformation following public outcry, social media platforms “fall short” when it comes to curbing the flow, said Joan Donovan, who leads the Technology and Social Change Project at HKS.

Since the national shift to remote work, many social media firms are relying more heavily on artificial intelligence to patrol misinformation on their platforms, instead of human moderators, who tend to be more effective, said Donovan. So many users suddenly searching and posting about one specific topic can “signal jam search algorithms, which cannot tell the difference usually between truth and lies.”

These firms are reluctant to spark a regulatory backlash by policing their platforms too tightly and angering one or both political parties.

“So they are careful to take action on content that is deemed immediately harmful (like posts that say to drink chemicals), but are reticent to enforce moderation on calls for people to break the stay-at-home orders,” said Donovan.

Viswanath said public health officials cannot, and should not, chase down and debunk every bit of misinformation or conspiracy theory, lest the attention lends them some credence. The public needs to more closely scrutinize and be “much more skeptical” about what they’re reading and hearing, particularly online, and not try to keep up with the very latest COVID-19 research. “You don’t need to know everything,” he said.

Putting the onus entirely on the public, however, is “unfair and it won’t work,” said Viswanath. Institutions, like social media platforms, have to take more responsibility for what’s out there.

Christina Pazzanese is Harvard staff writer. This  article  is published courtesy of the Harvard Gazette, Harvard University’s official newspaper.