The right and left may not agree on what constitutes misinformation, but both would like to see less of it on social media. And as the world faces the third year of the COVID-19 pandemic, the threat that medical misinformation poses to public health remains real. Companies such as Twitter and Facebook have a stake in cleaning up their platforms — without relying on censoring or fact-checking.
Censoring can engender distrust when social media companies expunge posts or delete accounts without explanation. It can even raise the profile of those who’ve been “canceled.”
And fact-checking isn’t a good solution for complex scientific concepts. That’s because science is not a set of immutable facts, but a system of inquiry that constructs provisional theories based on imperfect data.
A recent post on Politifact illustrates the problem. The claim at issue: a meme circulating on Facebook that viruses evolve to be less virulent. Politifact deemed it false, but Purdue University virologist David Sanders disagrees. “I would say that it actually is true that viruses do tend to evolve to be less harmful to their host,” he told me, though it’s a process than can take decades — or even centuries — from the time a new virus jumps from animal to human host.
Sanders said Politifact had conflated virulence with other things, such as resistance to drugs. When a complex issue is still a matter of scientific uncertainty and debate, rating it “true” or “false” doesn’t work very well.
Another limitation of fact-checking: There’s so much dubious content floating around Facebook and Twitter that human fact checkers can only get to a miniscule fraction. Consumers may wrongly assume what’s left over has been reviewed and is reliable.
“It’s not necessarily that (users) don’t care about accuracy. But instead, it’s that the social media context just distracts them and they forget to think about whether it’s accurate or not before they decide to share it,” said David Rand, professor of management science and cognitive sciences at MIT.
A recent study, published in the journal Nature, found that people improved the accuracy of their sharing when first asked to rate the accuracy of a headline. The idea was that this would shift people’s attention toward accuracy, which people say they believe is important even as they share things based on how popular they’re likely to be.
Facebook and Twitter could harness crowdsourcing to elevate the stories most likely to be true. “You could use that to inform your ranking to correspond to the actual accuracy,” Gordon Pennycook of the University of Regina in Canada said. “In a certain sense, that’s taking it out of the hands of the third parties and give it back to the people.”
Instead, algorithms on platforms like Twitter, YouTube, Instagram and Facebook are structured to suppress learning and feed people an informational junk food diet that reinforces existing beliefs and biases.
“What we are exposed to on social media is strongly affected by our own pre-existing opinions,” said Filippo Menczer, a professor at the Center for Complex Networks and Systems Research at Indiana University. And that’s one reason seemingly apolitical medical topics become politicized. “Political entities have an interest in using whatever people are paying attention to — for example, a health crisis — to manipulate people,” Menczer added.
The “people are getting dumber” myth has been embraced on both the political right and left. We’re not getting dumber. We’re all struggling to understand a complex, fractured world. Censorship and even fact-checking social media won’t solve that problem. To do that, platforms can change the system, giving users more power over what they see.
Faye Flam is a Bloomberg Opinion columnist and host of the podcast “Follow the Science.”