Menu
Aeon
DonateNewsletter
SIGN IN

Photo by Ivan Alvarado/Reuters

i

The misinformation virus

Lies and distortions don’t just afflict the ignorant. The more you know, the more vulnerable you can be to infection

by Elitsa Dermendzhiyska + BIO

Photo by Ivan Alvarado/Reuters

There’s a new virus in town and it’s not fooling around. You can catch it through face-to-face contact or digitally – that is, via a human or bot. Few of us possess immunity, some are even willing hosts; and, despite all we’ve learned about it, this virus is proving more cunning and harder to eradicate than anyone could have expected.

Misinformation isn’t new, of course. Fake news was around even before the invention of the printing press, although the first large-scale journalistic sham occurred in 1835, when the New York Sun published six articles announcing the discovery of life on the Moon (specifically, unicorns, bat men and bipedal beavers). Consider, too, early modern witch hunts, or those colonial myths that depicted slaves as a different species; the back-and-forth volleys of anti-Jewish and anti-German propaganda during the world wars, McCarthyism’s Red Scare, even communism’s utopian narratives. History teems with deceit.

What’s different today is the speed, scope and scale of misinformation, enabled by technology. Online media has given voice to previously marginalised groups, including peddlers of untruth, and has supercharged the tools of deception at their disposal. The transmission of falsehoods now spans a viral cycle in which AI, professional trolls and our own content-sharing activities help to proliferate and amplify misleading claims. These new developments have come on the heels of rising inequality, falling civic engagement and fraying social cohesion – trends that render us more susceptible to demagoguery. Just as alarming, a growing body of research over the past decade is casting doubt on our ability – even our willingness – to resist misinformation in the face of corrective evidence.

The classic experiments to correct misinformation date to the late 1980s. Subjects were given news briefs from the scene of a fictional warehouse fire, one of which mentions a closet with volatile materials – cans of oil paint and gas cylinders – others report ‘thick, oily smoke’, ‘sheets of flames’ and ‘toxic fumes’ that put the firefighters’ lives at risk. A further brief cites the police investigator on the case stating that the closet was, in fact, empty, before the report ends with the fire finally put out.

Having read the briefs, subjects had to answer a series of questions meant to probe their grasp of the correction made by the police investigator. It seems a simple test yet, across a multitude of studies, people repeatedly fail it. In one experiment, as many as 90 per cent of the subjects linked the fire’s toxic nature or intensity to the cans of oil paint and gas cylinders, despite none being found in the closet. More surprisingly, when asked directly, most of these participants readily acknowledged the empty closet. Researchers have reported similar results many times, including using blatantly direct retractions (‘there were no cans of paint or gas cylinders’). Yet no matter how clear the correction, typically more than half of subjects’ references to the original misinformation persist. What’s remarkable is that people appear to cling to the falsehood while knowing it to be false. This suggests that, even if successfully debunked, myths can still creep into our judgments and colour our decisions – an outcome referred to in the literature as ‘the continued influence effect’.

Why does this happen? According to Jason Reifler, professor of political science at the University of Exeter, we tend to take incoming information at face value, ‘because the existence of human society is predicated on the ability of people to interact and [on] expectations of good faith.’ Moreover, myths can take on subtle, crafty forms that feign legitimacy, making them hard to expose without careful analysis or fact checks. This means that those of us too dazed by the job of living to exert an extra mental effort can easily succumb to deception. And once a falsehood has slipped in and become encoded in memory – even weakly – it can prove remarkably sticky and resistant to correction.

One of the most common explanations for the continued influence effect puts it down to a gap in our mental model, or the story we tell ourselves about what happened. If the myth fits the ‘logic’ of events, its retraction leaves a hole, and the cogs of the story no longer click into place. We need the cans of oil paint and the gas cylinders: what would otherwise explain the billows of smoke and the force of the blaze? Remove the volatile materials from the closet, and the causal chain of events in our head unravels. If we aren’t to lose coherence, it makes sense to hold on to both the actual fact and the fitting falsehood – but keep them separate, compartmentalised, so that they don’t clash. This might be why, as studies show, we could be well aware of the truth, yet still allow the myth to creep in elsewhere and corrupt tangential judgments.

Older people might be particularly vulnerable to misinformation that’s repeated when retracted

Another reason why misinformation resists correction is repetition. Once something gets repeated often enough – sensational claims on social media; urban legends passed from one bored timewaster to another – it can trick us into taking it as true merely because of its familiarity. The illusory truth effect, as it’s known, suggests that the easier to process and more familiar something is, the more likely we are to believe it. Which is exactly what repeating a misleading claim does – getting it to go down smooth by strengthening the neural pathways linked to it.

This can pose a challenge for corrections that work by repeating the original misinformation. Consider, this retraction to a myth prone to ensnare hopeful new mothers: ‘Listening to Mozart will not boost your child’s IQ.’ The tiny ‘not’ mid-sentence is all that sets the myth and its correction apart – and it’s easy to imagine that as time passes and memory fades, that ‘not’ will wash away, leaving Mozart’s symphonies and smarter babies linked together in memory, and making the myth more familiar. Could this cause the correction to fail or even backfire?

In 2017, Stephan Lewandowsky, a cognitive scientist at the University of Bristol, and two colleagues from the University of Western Australia set out to investigate this possibility. They measured their test subjects’ beliefs in 20 myths and 20 facts, then corrected the myths in a way that repeated them twice more. Right away, as well as 30 minutes later, the correction significantly reduced subjects’ beliefs in – and references to – the false statements. However, after only a week, belief ratings crept back up to almost double their corrected levels.

Because memory declines with age, older people might be particularly vulnerable to misinformation that’s repeated when retracted. Indeed, in a similar study with older adults, Lewandowsky’s team found that, after three weeks, subjects aged over 65 ended up re-remembering most of the successfully corrected myths as facts. Again, though, no backfire effects occurred – that is, where the correction actually increases belief in the myth – and, despite some contrary earlier evidence, researchers now believe such effects to be rare, if they exist at all. And although repeated mentions of a myth can strengthen it, one repetition during correction seems safe and even desirable as it makes the myth more pliant by activating it in memory.

In recent years, as misinformation has wormed its way into large swathes of society, scientists have been looking for the most effective methods to counter it. Recently, Lewandowsky spearheaded The Debunking Handbook 2020, an online collection of best practice by 22 of the most active researchers in the field. The contributors nominated more than 50 relevant findings and more than 30 practical recommendations, rating them on their importance and the strength of the available evidence. To successfully debunk a myth, the authors conclude, it helps to provide an alternative causal explanation to fill the mental gap that retracting the myth could leave. Counterarguments work too, as they point out the inconsistencies contained in the myth, allowing people to resolve the clash between the true and the false statement. Another strategy is to evoke suspicion about the source of the misinformation. For example, you might be more critical of government officials who reject human-caused global warming if you suspect vested business interests behind the denialist claims.

The most vaccine-hesitant subjects ended up even less willing to vaccinate than they were before the study

Some researchers, however, question the practical significance of debunking strategies devised in a lab. As Reifler put it to me: ‘Are the “positive” effects of interventions we are seeing real effects, in terms of creating long-lasting changes in the accuracy of people’s beliefs? Or is this just a momentary “I’m accepting it because I was told it – and you could’ve told me the exact opposite thing and I’d have believed that?”’ In a world where both media and online platforms have turned into hotbeds of misinformation, Reifler’s question sounds especially urgent. John Cook, a climate change communication researcher at George Mason University in Virginia, told me: ‘I could develop the perfect message that debunks the myth completely. And, even if I could get that message to the right person, what happens if they just go home and turn on Fox News and get five hours of misinformation thrown at them? That particular message will be wiped out.’

It can get worse. Suppose the perfect message does find a person in need of disabusing, and even succeeds in fixing their false beliefs: will that person’s attitudes and behaviour change accordingly? If you tell people that 97 per cent of climate scientists agree about the reality of global warming, studies show that you’ll likely increase their perception of expert consensus on the subject. But whether this greater awareness translates into action – say, support for carbon-reduction policies – remains unclear. The evidence is mixed, and the question has sparked ‘substantial debate and disagreement’ among researchers, says James Druckman, professor of political science at Northwestern University in Illinois. Yet even in studies that do find a knock-on effect on intentions, that effect is small. In other words, you can deliver the facts to people, you can even get them to accept those facts – and it still might not change a thing.

One worrying demonstration of this possibility comes from the realm of vaccines. In a 2016 study, Reifler worked with the political scientist Brendan Nyhan at Dartmouth College in New Hampshire, testing two approaches to debunk the myth that flu vaccines actually cause the flu – a myth partly responsible for low vaccination rates and thousands of preventable deaths from seasonal influenza in the US. One subject group saw official corrective materials from the US Centers for Disease Control and Prevention, while another group received information about the risks of not vaccinating. This latter group showed no change in myth beliefs, whereas in the correction group the myth beliefs substantially declined, even among the most sceptical subjects. It seemed that the correction had worked – and brilliantly. But what ultimately interested Reifler was less the participants’ beliefs and more their intentions to vaccinate – and, across the sample, these didn’t budge at all. The most vaccine-hesitant subjects ended up even less willing to vaccinate than they were before the study.

When I talked to Reifler, he couldn’t name any research that showed that communicating the safety of vaccines (or highlighting the dangers of refusing them) had a positive effect on people’s intentions to vaccinate. At this point in our interview, I faltered. It just seemed too absurd that in a matter of life and death, information potentially key to survival could still be ignored. I asked Reifler if he found this disappointing. He said he was used to it. ‘My entire career is filled with work that is depressing. My usual joke is that if it’s bad for the world, it’s probably good for my research.’

To fully grasp the pernicious nature of the misinformation virus, we need to reconsider the innocence of the host. It’s easy to see ourselves as victims of deception by malicious actors. It’s also tempting to think of being misinformed as something that happens to other people – some unnamed masses, easily swayed by demagoguery and scandal. ‘The problem is that people are sheep,’ one friend said to me. I’ve heard this sentiment echoed time and again by others, the implication always being that they and I were not like those other, misinformed people. No: we were educated, had been taught to think, immune to dupery. But, as it turns out, misinformation doesn’t prey only on the ignorant: sometimes, those who seem least vulnerable to the virus can prove its keenest hosts, and even handmaidens.

Startling evidence for this possibility comes from Dan M Kahan, professor of law and psychology at Yale University who has been studying how ordinary people evaluate complex societal risks. One strand of his research is trying to shed light on the sometimes dramatic disparity between public opinion and scientific evidence. Together with a small group of researchers, in 2010 Kahan set out to demystify this disparity in relation to global warming. At the time, despite widespread consensus among climate scientists, only 57 per cent of Americans believed that there was solid evidence for global warming, and just 35 per cent saw climate change as a serious problem. ‘Never have human societies known so much about mitigating the dangers they face but agreed so little about what they collectively know,’ Kahan wrote.

One standard explanation, which Kahan calls the ‘science comprehension thesis’, holds that people have insufficient grasp of science, and are unlikely to engage in the deliberate, rational thinking needed to digest these often complex issues. It’s a plausible explanation, yet Kahan suspected that it doesn’t tell the whole story.

Asking for people’s take on climate change is also to ask them who they are and what they value

In the 2010 study, published in Nature in 2012, Kahan and his collaborators measured subjects’ science literacy and numeracy, and plotted those against the participants’ perceived risk of global warming. If the science comprehension thesis was right, then the more knowledgeable the subjects, the more they’d converge towards the scientific consensus. Surprisingly, however, the data revealed that those who scored high on hierarchy and individualism – the hallmark values of a conservative outlook – exhibited the opposite pattern: as their science literacy and numeracy increased, their concern for climate change actually declined. What explains this seeming paradox?

Kahan argues that rather than being a simple matter of intelligence or critical thinking, the question of global warming triggers deeply held personal beliefs. In a way, asking for people’s take on climate change is also to ask them who they are and what they value. For conservatives to accept the risk of global warming means to also accept the need for drastic cuts to carbon emissions – an idea utterly at odds with the hierarchical, individualistic values at the core of their identity, which, by rejecting climate change, they seek to protect. Kahan found similar polarisation over social issues that impinge on identity, such as gun control, nuclear energy and fracking, but not over more identity-neutral subjects such as GMO foods and artificial sweeteners. In cases where identity-protective motivations play a key role, people tend to seek and process information in biased ways that conform to their prior beliefs. They might pay attention only to sources they agree with and ignore divergent views. Or they might believe congruent claims without a moment’s thought, but spare no effort finding holes in incongruent statements: the brightest climate-change deniers were simply better than their peers at counter-arguing evidence they didn’t like.

This hints at a vexing conclusion: that the most knowledgeable among us can be more, not less, susceptible to misinformation if it feeds into cherished beliefs and identities. And though most available research points to a conservative bias, liberals are by no means immune.

In a 2003 study, Geoffrey Cohen, then a professor of psychology at Yale, now at Stanford University, asked subjects to evaluate a government-funded job-training programme to help the poor. All subjects were liberal, so naturally the vast majority (76 per cent) favoured the policy. However, if subjects were told that Democrats didn’t support the programme, the results completely reversed: this time, 71 per cent opposed it. Cohen replicated this outcome in a series of influential studies, with both liberal and conservative participants. He showed that subjects would support policies that strongly contradict their own political beliefs if they think that others like them supported those policies. Despite the social influence, obvious to an outsider, participants remained blind to it, and attributed their preferences to objective criteria and personal ideology. This would come as no surprise to social psychologists, who have long attested to the power of the group over the individual, yet most of us would doubtless flinch at the whiff of conformity and the suggestion that our thoughts and actions might not be entirely our own.

For Kahan, though, conformity to group beliefs makes sense. Since each individual has only negligible impact on collective decisions, it’s sensible to focus on optimising one’s social ties instead. Belonging to a community is, after all, a vital source of self-worth, not to mention health, even survival. Socially rejected or isolated people face heightened risks of many diseases as well as early death. Seen from this perspective, then, the impulse to fit our beliefs and behaviours to those of our social groups, even when they clash with our own, is, Kahan argues, ‘exceedingly rational’. Ironically, however, rational individual choices can have irrational collective consequences. As tribal attachments prevail, emotions trump evidence, and the ensuing disagreement chokes off action on important social issues.

Recently, public disagreement has spilled over to the idea of truth itself. The term ‘post-truth’ became the Oxford Dictionaries Word of the Year in 2016, and came to characterise that year’s US presidential election and the Brexit referendum. In a 2017 paper, Lewandowsky argued that we’ve gone ‘beyond misinformation’: ‘The post-truth problem is not a blemish on the mirror,’ he wrote. ‘The problem is that the mirror is a window into an alternative reality.’ In this other reality, marked by the global rise of populism, lies have morphed into an expression of identity, a form of group membership. In the US, the UK, Germany, Austria, Italy, Poland, Brazil and India, populists have captured a growing disenchantment with the status quo by pitting ‘the people’ against ‘the elites’, and attacking so-called elitist values – education, evidence, expertise.

In the populist story, lying takes on the trappings of anti-establishmentarianism, undermining truth as a social norm. This is the misinformation virus at its most diabolical: a point where health (in this case, of the body politic) ceases to matter – as was so graphically demonstrated during the storming of the US Capitol this January – and the host consents to being infected. (The one good thing to come out of that ‘insurrection’ is that tough action was swiftly taken against the peddlers of misinformation with Twitter banning the then president Donald Trump and suspending thousands of QAnon-related accounts.)

It’s probably easier to change what we think others think than what we ourselves do

It’s easy to despair over all the cognitive quirks, personal biases and herd instincts that can strip our defences against the ever-evolving misinformation machinery. I certainly did. Then, I found Elizabeth Levy Paluck. She is a psychologist at Princeton University who studies prejudice reduction – a field in which a century of research appears to have produced many theories but few practical results. In 2006, she led an ambitious project to reduce ethnic hostilities in the Democratic Republic of Congo. She blended a number of prominent theories to create a ‘cocktail of treatments’: a radio drama, in which characters from different communities modelled cooperation and mutual trust; a talk show whose host read audience letters replete with messages of tolerance, and who encouraged listeners to put themselves in the shoes of outgroup members. Nothing worked. After a year of broadcasting, prejudice remained as entrenched as ever.

For Paluck, this was ‘an empirical and theoretical puzzle’, prompting her to wonder if beliefs might be the wrong variable to target. So she turned to social norms, reasoning that it’s probably easier to change what we think others think than what we ourselves do. In 2012, Paluck tested a new approach to reducing student conflict in 56 middle schools in New Jersey. Contrary to popular belief, some evidence suggests that, far from being the product of a few aggressive kids, harassment is a school-wide social norm, perpetuated through action and inaction, by bullies, victims and onlookers. Bullying persists because it’s considered typical and even desirable, while speaking up is seen as wrong. So how do you shift a culture of conflict? Through social influence, Paluck hypothesised: you seed supporters of a new norm and let them transmit it among their peers. In some schools, Paluck had a group of students publicly endorse and model anti-bullying behaviours, and the schools saw a significant decline in reported conflicts – 30 per cent on average, and as much as 60 per cent when groups had higher shares of well-connected model students.

I’ve wondered recently if, like school violence, misinformation is becoming part of the culture, if it persists because some of us actively partake in it, and some merely stand by and allow it to continue. If that’s the case, then perhaps we ought to worry less about fixing people’s false beliefs and focus more on shifting those social norms that make it OK to create, spread, share and tolerate misinformation. Paluck shows one way to do this in practice – highly visible individual action reaching critical mass; another way could entail tighter regulation of social media platforms. And our own actions matter, too. As the Scottish biologist D’Arcy Wentworth Thompson said in 1917, ‘everything is what it is because it got that way’. We are, each and every one of us, precariously perched between our complicity in the world as it is and our capacity to make it what it can be.