Skip to main content Accessibility help
×
Hostname: page-component-76fb5796d-x4r87 Total loading time: 0 Render date: 2024-04-26T19:00:15.513Z Has data issue: false hasContentIssue false

8 - Misinformation and Its Correction

Published online by Cambridge University Press:  24 August 2020

Nathaniel Persily
Affiliation:
Stanford University, California
Joshua A. Tucker
Affiliation:
New York University

Summary

Although political misinformation is not a new phenomenon, the topic has received renewed attention in recent years, in conjunction with sweeping changes in the contemporary media environment. As the Internet and, particularly, social media become an increasingly common source for political information, citizens receive more and more of their news in an uncontrolled and minimally regulated setting where misinformation may easily spread. However, even if the sources of misinformation have fundamentally changed, best practices for correcting misinformation have not. While many of the pieces cited in this chapter do not focus explicitly on the Internet or social media, these works can still inform scholarly understanding of how to correct misinformation on these platforms. The cognitive processes we highlight are likely to translate to the digital realm and are thus crucial to understand when developing prescriptions for social media-based misinformation. Nevertheless, we also spotlight a number of recent studies that examine methods for correcting misinformation in the context of social media.

Type
Chapter
Information
Social Media and Democracy
The State of the Field, Prospects for Reform
, pp. 163 - 198
Publisher: Cambridge University Press
Print publication year: 2020
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This content is Open Access and distributed under the terms of the Creative Commons Attribution licence CC-BY-NC-ND 4.0 https://creativecommons.org/cclicenses/

Fake news is big news. From the diffusion of rumors and conspiracies in the United States to the spread of disinformation by Russian troll farms, misinformation is a hot topic among academics and journalists alike. How can we understand and correct such misinformation? A logical starting point is to fight fiction with fact. Indeed, many proposed solutions to the problem of misinformation assume that the proper remedy is merely to provide more information. In this view, if citizens were only better informed, misinformation would lose its power. However, ample research suggests that the answer is not so simple. Misinformation may continue to endure post-correction for several reasons. First, corrections are rarely able to fully eliminate reliance on misinformation in later judgments. Even when people recall hearing a retraction, the original misinformation may still influence their attitudes and beliefs (what is known as the continued influence effect). Worse yet, people may come to believe in misinformation even more strongly post-correction. In particular, retractions that run counter to individuals’ prior attitudes may bolster beliefs in the original misinformation (what are known as worldview backfire effects). These worldview backfire effects have their roots in directionally motivated reasoning; individuals process misinformation and corrections through the lens of their preexisting beliefs and partisan attachments, so they may actively dispute corrections that contradict their broader worldviews.

Although political misinformation is not a new phenomenon, the topic has received renewed attention in recent years, in conjunction with sweeping changes in the contemporary media environment. As the Internet and, particularly, social media become an increasingly common source for political information (Reference Shearer and MatsaShearer and Matsa 2018), citizens receive more and more of their news in an uncontrolled and minimally regulated setting where misinformation may easily spread (Reference Vosoughi, Roy and AralVosoughi, Roy, and Aral 2018). Validating these concerns, numerous studies of “fake news” spotlight social media platforms, including both Facebook and Twitter, as the primary incubators of misinformation during the 2016 US presidential election (e.g., Reference Allcott and GentzkowAllcott and Gentzkow 2017; Reference Guess, Nyhan and ReiflerGuess, Nyhan, and Reifler 2020). However, even if the sources of misinformation have fundamentally changed, best practices for correcting misinformation have not. While many of the pieces cited in this chapter do not focus explicitly on the Internet or social media, these works can still inform scholarly understanding of how to correct misinformation on these platforms. The cognitive processes we highlight are likely to translate to the digital realm and are thus crucial to understand when developing prescriptions for social media–based misinformation. Nevertheless, we also spotlight a number of recent studies that examine methods for correcting misinformation in the context of social media.

Previous Review Pieces

Several excellent review articles have already greatly enriched our knowledge of misinformation and its correction. Each is a valuable resource for deeper reading on this subject. In the interest of not rehashing existing work, we have made a conscious choice to showcase topics not already covered in these reviews. However, for the benefit of the reader, we summarize the primary takeaways from each piece and preview how we build on the groundwork they laid. First, Reference Lewandowsky, Ecker, Seifert, Schwarz and CookLewandowsky et al. (2012) provide a comprehensive summary of the literature on misinformation and its correction. In particular, they delve into the psychological roots of the continued influence effect and backfire effects and recommend appropriate interventions for practitioners seeking to mitigate these effects. However, since the article’s publication in 2012, the field has evolved in notable ways – especially regarding the existence and magnitude of different types of backfire effects. Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker (2018) thus provide an updated summary of the literature and offer several new strategies for effectively correcting misinformation. We pick up where these two articles leave off; we discuss newer research on both the continued influence effect and backfire effects and suggest ways for future work to continue to flesh out these topics in even greater detail.

Second, Reference Flynn, Nyhan and ReiflerFlynn, Nyhan, and Reifler (2017) offer a more recent review of the misinformation literature, with a specific focus on the relationship between directionally motivated reasoning and political misperceptions. In their view, individuals’ preexisting beliefs strongly affect their responses to corrections, such that individuals with different partisan or ideological leanings may respond to the same political facts in profoundly different ways. Importantly, the authors document a number of individual and contextual moderators of directionally motivated reasoning that predispose certain subsets of the population to be more vulnerable to worldview backfire effects. However, motivated reasoning is not the sole reason why misinformation persists over time. As studies of the continued influence effect demonstrate, individuals may continue to hold misinformed beliefs post-correction, even in the absence of strong prior attitudes. As such, we take a closer look at the full range of psychological mechanisms that may impede attempts to correct misinformation.

Finally, Reference Tucker, Guess and BarberaTucker et al. (2018) focus on the interplay of social media and political polarization in enabling the spread of misinformation, with particular emphasis on the specific actors who manufacture misinformation. However, though they present a wide-ranging and thorough analysis of the contemporary research on the production of digital misinformation, these authors devote substantially less attention to best practices for correction. As a complement to this work, we discuss recent research on strategies to correct misinformation appearing on social media platforms, including Facebook and Twitter.

Organization of the Chapter

In this chapter, we synthesize recent work on misinformation and its correction. Knowledge of this subject is still rapidly developing, and many questions remain unanswered and unresolved.Footnote 1 Here, we pay particular attention to one of these important questions: Why does misinformation persist even after it has been corrected? To this end, we first provide a definition of misinformation and specify the core criteria that help to discriminate between the many related concepts in this area. Second, we discuss two key perspectives on the perseverance of misinformation post-correction: backfire effects and the continued influence effect. Third, we outline a number of individual and contextual moderators that might make certain individuals or groups especially susceptible to misinformation. Finally, we conclude with a series of recommendations for future research.

Defining Misinformation: Mapping Key Criteria

To understand how best to tackle the problem of misinformation, it is essential to first define what this term means. However, scholarly notions of what constitutes misinformation often differ significantly across works and across disciplines. These definitions are highly variable, ranging from simple statements about the misleading nature of misinformation to commentaries on the motivation for the spread of misinformation. Some scholars broadly characterize misinformation as false information; for example, Reference FetzerFetzer (2004) defines it as “false, mistaken, or misleading information” (p. 231), and Reference BerinskyBerinsky (2017) defines it as “information that is factually unsubstantiated” (p. 242). Other scholars take a more restricted view, contrasting the term with other concepts, such as disinformation. For instance, Reference WardleWardle (2018) argues that misinformation is “information that is false, but not intended to cause harm” (p. 5), whereas disinformation is “false information that is deliberately created or disseminated with the express purpose to cause harm” (p. 4). Finally, a third approach emphasizes the temporal nature of misinformation processing, arguing that misinformation’s primary feature is that it is first presented as true but later revealed to be false. Reference Ecker, Lewandowsky, Cheung and MayberyEcker et al. (2015) state that misinformation is “information that is initially presented as factual but subsequently corrected” (p. 102). Similarly, Reference Lewandowsky, Ecker, Seifert, Schwarz and CookLewandowsky et al. (2012) define misinformation as “any piece of information that is initially processed as valid but that is subsequently retracted or corrected” (pp. 124–125). In this sense, information only becomes misinformation when it is first believed and later corrected, separating misinformation from other false information that goes unrebutted.

Compounding the problem is the fact that the term “misinformation” is often confounded with other similar concepts. For instance, as noted in the previous paragraph, some authors attempt to draw a line between misinformation and disinformation, or “information that is false and deliberately created to harm a person, social group, organization, or country” (Reference Wardle and DerakhshanWardle and Derakhshan 2017, p. 20). Other scholars speak of misperceptions, or “cases in which people’s beliefs about factual matters are not supported by clear evidence and expert opinion” (Reference Nyhan and ReiflerNyhan and Reifler 2010, p. 305). Still others make reference to conspiracy theories, which offer unconventional explanations of the causes of events in terms of the “significant causal agency of a relatively small group of persons – the conspirators – acting in secret” (Reference KeeleyKeeley 1999, p. 116; see also Reference Oliver and WoodOliver and Wood 2014). Similar to, though broader in scope than, conspiracy theories are political rumors, which are “unverified stories or information statements people share with one another” (Reference Weeks and GarrettWeeks and Garrett 2014, p. 402). Finally, since the 2016 US presidential election, there has been much talk of fake news, which shares many similarities with disinformation but differs in its presentation. In particular, recent work defines fake news as “fabricated information that mimics news media content in form but not in organizational process or intent” (Reference Lazer, Baum and BenklerLazer et al. 2018, p. Reference Lazer, Baum and Benkler1094).

The multitude of definitions of misinformation speaks to the need for clarity on what exactly we, as a scholarly community, mean when we talk about misinformation. In an attempt to provide such structure, we compiled a wide variety of definitions of misinformation and related terms. Looking for common threads, we identified four overarching criteria for differentiating types of misinformation.Footnote 2 First, we found that different definitions of misinformation place more or less emphasis on the truth value of the information – that is, whether the information has been proven to be untrue or whether it is merely unsubstantiated. Second, we noted that definitions of misinformation vary in their area of focus, particularly whether they emphasize the effects of false information versus false beliefs. Third, we found that scholars distinguish forms of misinformation based on their format, including whether or not the presentation of the information is designed to resemble traditional news sources. Finally, we noted differences in the perceived intentions of the actors who spread misinformation, in terms of their level of awareness that the information was false.

Four Key Criteria

First, truth value: All forms of misinformation, at least to some degree, rest on shaky factual foundations. That is, all misinformation is in some way inaccurate. In some cases, misinformation is characterized by a lack of conclusive evidence to support a particular position, whereas, in others, it involves statements that run counter to mainstream consensus or expert opinion. However, the extent to which information is untrue varies across forms of misinformation; some subtypes may be definitively false (e.g., disinformation or fake news), whereas others may be merely misleading or unverified (e.g., political rumors).

Second, the area of focus: It is important to separate the presence of false information (misinformation) from the endorsement of false beliefs (misperceptions). This distinction is valuable because, as Reference ThorsonThorson (2015) highlights, misperceptions are not exclusively caused by misinformation. Even if individuals only encounter true information, they may still arrive at inaccurate beliefs for other reasons, such as cognitive biases or misinterpretation of available facts. In this sense, the appropriate tools for correction may depend heavily on whether false beliefs are the clear product of misinformation or if they instead originate via other channels.

Third, format: Different types of misinformation may be presented in different ways. In some cases, misinformation may be embedded within otherwise accurate reports, whereas, in other cases, it may exist as standalone content. This is especially relevant to the study of fake news, or fabricated articles that imitate the appearance of traditional news stories (Reference Allcott and GentzkowAllcott and Gentzkow 2017; Reference Lazer, Baum and BenklerLazer et al. 2018). Fake news is a form of disinformation, as it is spread despite being known to be false, but it may be distinguished from other types of disinformation by its unique format – namely, its emulation of legitimate media outlets (Reference Pennycook and RandPennycook and Rand 2018). In addition to fake news, recent work also looks beyond textual forms of misinformation to other types of media, including manipulated images and videos (Reference Kasra, Shen and O’BrienKasra, Shen, and O’Brien 2016; Reference Schwarz, Newman and LeachSchwarz, Newman, and Leach 2016; Reference Shen, Kasra, Pan, Bassett, Malloch and O’BrienShen et al. 2019).

Finally, intentionality: Does the person transmitting misinformation sincerely believe it to be true or are they aware that it is false? By most accounts, this is the primary means of distinguishing between misinformation and disinformation (for a review, see Reference WardleWardle 2018). On the one hand, misinformation may circulate without any intent to deceive. For instance, in the wake of breaking news events, people increasingly turn to the Internet, and especially social media, for real-time updates. As new information is released in a piecemeal fashion, individuals may inadvertently propagate information that later turns out to be false (Reference Nyhan and ReiflerNyhan and Reifler 2015a; Reference Zubiaga, Liakata, Procter, Wong Sak Hoi and TolmieZubiaga et al. 2016). On the other hand, disinformation is false or inaccurate information that is deliberately distributed despite its inaccuracy (Reference StahlStahl 2006; Reference Born and EdgingtonBorn and Edgington 2017). People may choose to share fictitious stories, even when they recognize that these stories are untrue. Why might people knowingly promulgate false information? One answer relates to the disseminators’ motivations; although misinformation is typically not designed to advance a particular agenda, disinformation is often spread in service of concrete goals. For instance, fake news is often designed to go viral on social media (Reference Pennycook and RandPennycook and Rand 2018; Reference Tandoc, Lim and LingTandoc, Lim, and Ling 2018), enabling rapid transmission of highly partisan content and offering a reliable stream of advertising revenue (Reference Tucker, Guess and BarberaTucker et al. 2018). In practice, however, determining a person or group’s intentions is extremely difficult. It is hard to uncover people’s “ground truth” beliefs about the veracity of a piece of information, and it is even harder to ascertain their underlying motivations. That said, recognizing the range of motivations for spreading misinformation is valuable, even if these motivations are hard to disentangle in the wild.

For the purposes of this chapter, we consider “misinformation” an umbrella term under which many associated concepts are subsumed. Moving forward, we recommend misinformation as the default term to use, unless explicitly referring to one of these more specific constructs. Given the difficulty of proving the motivations underlying the spread of false information, we adopt an intent-agnostic approach; we make no assumptions about what compels individuals or groups to broadcast misinformation. Instead, we take the view that misinformation – in all of its forms – may have a considerable, harmful impact on people’s beliefs and behavior. As such, in the discussion that follows, we cite examples of corrective strategies targeted at all different types of misinformation.

Responses to Corrections: Continued Influence and Backfire Effects

Detailing types of information is not a mere technical exercise. A well-functioning democratic society does not necessarily need to be guided by fully informed citizens, but an environment rife with misinformation can easily derail democracy. An uninformed citizenry is arguably far less pernicious than a misinformed citizenry (Reference Kuklinski, Quirk, Jerit, Schweider and RichKuklinski et al. 2000); as Reference Hochschild and EinsteinHochschild and Einstein (2015) write, “people’s unwillingness or inability to use relevant facts in their political choices may be frustrating, but people’s willingness to use mistaken factual claims in their voting and public engagement is actually dangerous to a democratic polity” (p. 14). When the public holds misinformed beliefs, this can not only affect their individual attitudes and behaviors but also shape large-scale policy outcomes (e.g., health care reform, see Reference NyhanNyhan 2010; Reference BerinskyBerinsky 2017). Correcting misinformation is therefore a worthy goal; but how can it best be accomplished?

Previous research suggests that not all corrections are effective in reducing individuals’ reliance on misinformation. There are two pathways through which misinformation might continue to shape attitudes and behaviors post-correction: the continued influence effect and backfire effects. Engrained in the former is the notion that corrections are somewhat, but not entirely, effective at dispelling misinformation. More concerning, however, are the latter, in which corrections not only fail to reduce but actually strengthen beliefs in the original misinformation. Neither of these phenomena offers a particularly sanguine take on the ability to curtail the spread of misinformation. However, each offers its own unique predictions about the most promising avenues for corrections. We begin by reviewing the extant literature on backfire effects and then turn to the continued influence effect.

Backfire Effects

Providing factual corrections of misinformation may, under certain circumstances, only make things worse. Specifically, retractions that challenge people’s worldviews may entrench beliefs in the original misinformation. This phenomenon is known as a backfire effect or, more precisely, a worldview backfire effect.Footnote 3 Reference Nyhan and ReiflerNyhan and Reifler (2010) sounded the first alarm bells about the possibility of these worldview backfire effects. Across a series of studies, they found that, when certain subjects were presented with factual corrections that contradicted their political beliefs, they responded by becoming more, rather than less, wedded to their previous misperceptions. Since their highly influential piece was published, concerns about worldview backfire effects have taken hold both in popular media and in academic circles. This widespread interest has spawned an entire line of work dedicated to elucidating the psychological mechanisms that drive these effects.

Worldview backfire effects can be understood as a product of directionally motivated reasoning (for a comprehensive review, see Reference Flynn, Nyhan and ReiflerFlynn et al. 2017). According to theories of motivated reasoning, individuals are motivated to process information in ways that align with their ultimate goals (Reference KundaKunda 1990). In particular, individuals must balance several competing impulses, including directional goals (to attain a desired outcome) and accuracy goals (to reach the correct conclusion). Worldview backfire effects transpire when directional motivations take precedence over accuracy goals – a frequent occurrence in the realm of politics (Reference Lodge and TaberLodge and Taber 2013).

Two complementary processes are at the heart of these effects. First, confirmation bias: Individuals tend to seek out and interpret new information in ways that validate their preexisting views. Along these lines, individuals also tend to perceive congenial information as more credible or persuasive than opposing evidence (Reference Guess and CoppockGuess and Coppock 2018; Reference Khanna and SoodKhanna and Sood 2018). Second, disconfirmation bias: When exposed to ideologically dissonant information, individuals will call to mind opposing arguments (counterarguing).Footnote 4 In combination, these two processes can cultivate worldview backfire effects; when individuals are confronted with a correction that contradicts their past beliefs, they will act to both discount the correction and bolster their prior views.

Several studies have investigated the potential for worldview backfire effects in the context of misinformation. Although Nyhan and Reifler issued the earliest warnings about this phenomenon, it has since been reproduced across other settings. First, worldview backfire effects have been tied to message presentation, with individuals most resistant to message framing that contradicts their broader worldviews (Reference ZhouZhou 2016). Second, worldview backfire effects have been linked to source cues. For instance, several studies find that Republicans are averse to corrections from Democratic elites (Reference BerinskyBerinsky 2017) or nonpartisan fact-checking sources (Reference Holman and LayHolman and Lay 2019). Finally, worldview backfire effects extend to the behavioral realm; across multiple studies, exposure to pro-vaccine corrections decreased future vaccination intentions among those already hesitant to get vaccinated (Reference Skurnik, Yoon and SchwarzSkurnik, Yoon, and Schwarz 2007; Reference Nyhan, Reifler, Richey and FreedNyhan et al. 2014; Reference Nyhan and ReiflerNyhan and Reifler 2015b; but see Reference HaglinHaglin 2017).

Empirical studies have also taught us about the mechanisms that undergird worldview backfire effects. Consistent with a motivated reasoning perspective, worldview backfire effects appear rooted in counterarguing. In one experiment, Reference Schaffner and RocheSchaffner and Roche (2017) examine differences in survey response times following the release of the October 2012 jobs report, which announced a sharp decrease in the unemployment rate under the Obama administration. They find that those Republicans who took longer to provide estimates of the unemployment rate after the report’s release were less accurate in their responses, suggesting that worldview backfire effects may arise out of deliberate, effortful processes. However, more work beyond this initial study is certainly needed to isolate the mechanisms that underlie worldview backfire effects.

Avoiding Worldview Backfire Effects

In light of mounting concerns about the potential for worldview backfire effects, scholars have explored several tactics for correcting misinformation while circumventing these effects. Although many routes to correction are possible, all designed to counteract directionally motivated reasoning, we summarize here two main subcategories of these corrections focused on source credibility and worldview affirmation.

First, the source of misinformation – as well as its correction – may have a profound impact on responses to corrections. When evaluating the accuracy of a claim, individuals rely heavily on source cues (Reference Schwarz, Newman and LeachSchwarz et al. 2016), which signal a source’s expertise or trustworthiness. In terms of expertise, if sources are depicted as authorities on a given subject, they are likely to be deemed more credible (Reference Vraga and BodeVraga and Bode 2017). In fact, expert consensus is considered a key “gateway belief” that can override directional impulses. For example, communicating the broad scientific agreement about climate change reduces partisan differences in climate change attitudes (Reference van der Linden, Leiserowitz, Feinberg and Maibachvan der Linden et al. 2015; Reference van der Linden, Leiserowitz, Rosenthal and Maibachvan der Linden et al. 2017; Reference Druckman and McGrathDruckman and McGrath 2019; but see Reference Kahan, Jenkins‐Smith and BramanKahan, Jenkins‐Smith, and Braman 2011). However, the trustworthiness of a source seems to matter even more than expertise when countering misinformation (Reference McGinnies and WardMcGinnies and Ward 1980; Reference Guillory and GeraciGuillory and Geraci 2013). People are more likely to view sources as trustworthy if they share similar traits. As a result, corrections that are attributed to an in-group member (e.g., a leader of one’s preferred party) may be more effective than those credited to an out-group member (e.g., an opposing partisan, see Reference Swire, Berinsky, Lewandowsky and EckerSwire, Berinsky et al. 2017). Furthermore, though corrections are most frequently issued by elites, individuals are also receptive to corrections from members of their social circles (Reference Margolin, Hannak and WeberMargolin, Hannak, and Weber 2018; Reference Vraga and BodeVraga and Bode 2018), who may not be experts but may still be deemed trustworthy. Finally, trustworthiness is a function of one’s perceived stake in an issue. Recent research on “unlikely sources” (Reference BerinskyBerinsky 2017; Reference Benegal and ScruggsBenegal and Scruggs 2018; Reference Wintersieck, Fridkin and KenneyWintersieck, Fridkin, and Kenney 2018; Reference Holman and LayHolman and Lay 2019) indicates that corrections are most persuasive when they come from sources who stand to benefit from the spread of misinformation (e.g., a Democratic politician or left-leaning publication correcting a fellow Democrat).

Second, where possible, corrections should be tailored to their target audience: the subset of people for whom these corrections would feel most threatening. Framing corrections to be consonant with, rather than antagonistic to, this group’s values and worldviews may thus be a successful corrective strategy (Reference Kahan, Braman, Monahan, Callahan and PetersKahan et al. 2010; Reference Feinberg and WillerFeinberg and Willer 2015; for reviews, see Reference Lewandowsky, Ecker, Seifert, Schwarz and CookLewandowsky et al. 2012; Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker 2018). In a similar vein, some scholars propose using self-affirmation exercises (Reference Cohen, Aronson and SteeleCohen, Aronson, and Steele 2000; Reference Cohen, Sherman, Bastardi, Hsu, McGoey and RossCohen et al. 2007) to subdue directional motivations (Reference Trevors, Muis, Pekrun, Sinatra and WinneTrevors et al. 2016; Reference Carnahan, Hao, Jiang and LeeCarnahan et al. 2018); if people feel validated in their global self-worth, corrections that impugn their political views may provoke a less defensive response (but see Reference Nyhan and ReiflerNyhan and Reifler 2019).

Backlash Against Worldview Backfire Effects

The previous discussion presumes that worldview backfire effects are not only dangerous but prevalent. Recently, however, there has been backlash against the very notion of worldview backfire effects, with some suggesting they are extremely rare in practice. Reference Wood and PorterWood and Porter (2019) attempt to detect worldview backfire effects across a wide array of divisive issues and find little evidence to support their existence, even when using language identical to Reference Nyhan and ReiflerNyhan and Reifler (2010). When examining several highly polarized issues, including gun control and capital punishment, Reference Guess and CoppockGuess and Coppock (2018) likewise fail to uncover any worldview backfire effects in response to counter-attitudinal information. Instead, individuals seem to accommodate novel information into their later assessments of issues, even if that information runs counter to their beliefs (see also Reference Porter, Wood and KirbyPorter, Wood, and Kirby 2018). However, even if corrections largely improve belief accuracy, these messages seem to have little impact on individuals’ subsequent attitudes, evaluations of politicians, or policy preferences (Reference Swire, Berinsky, Lewandowsky and EckerSwire, Berinsky et al. 2017; Reference Aird, Ecker, Swire, Berinsky and LewandowskyAird et al. 2018; Reference Nyhan, Porter, Reifler and WoodNyhan et al. 2019; Reference Porter, Wood and BahadorPorter, Wood, and Bahador 2019; Reference Barrera, Guriev, Henry and ZhuravskayaBarrera et al. 2020).

What accounts for the discrepancies in results across studies? The answer may be both theoretical and methodological. First, there is a large body of work on the theoretic side. Some scholars have suggested that worldview backfire effects are more likely when corrections necessitate attitude change versus only pertain to a single, specific event (Reference Ecker, Lewandowsky, Fenton and MartinEcker et al. 2014; Reference Ecker and AngEcker and Ang 2019). Why are people better able to absorb ideologically dissonant corrections for one-off events? To answer this question, Reference Ecker and AngEcker and Ang (2019) draw on stereotype subtyping theory (Reference Richards and HewstoneRichards and Hewstone 2001). According to this theory, individuals possess stereotypes about the customary behavior of different groups (e.g., members of a political party). Subtyping is a common response when group members act in ways that flout a well-established stereotype; rather than forming a new stereotype, individuals instead label contrary cases as exceptions to the broader rule. If misinformation pertains to a single, isolated event, individuals may thus be able to internalize disconfirming corrections without altering their deep-seated worldviews. It is much harder, however, to dismiss general patterns of behavior as anomalous. As such, people are more likely to resist corrections that, on acceptance, would require a large-scale shift in their core beliefs.

Reference Redlawsk, Civettini and EmmersonRedlawsk, Civettini, and Emmerson (2010) provide a somewhat different perspective on the boundaries of worldview backfire effects. They posit an “affective tipping point” at which individuals cease to engage in motivated reasoning and instead revise their beliefs to be more accurate – in other words, the point at which individuals pivot from directional to accuracy goals. As people encounter more and more disconfirming information, they may reach a critical threshold at which they are no longer motivated to defend their previous views. In this view, worldview backfire effects will occur until enough contradictory evidence accumulates. After this point, individuals will begin to rationally update their beliefs in response to corrections rather than double down on their previously misinformed views. This theoretical account generates somewhat divergent predictions from Ecker and Ang. For both sets of authors, general cases of misinformation should provoke heightened discomfort. However, if enough contrary evidence comes to light, Redlawsk and colleagues anticipate a diminished likelihood of worldview backfire effects, whereas Ecker and Ang seem to predict the exact opposite. Further work may be needed to adjudicate between these two explanations. In particular, efforts to pinpoint the precise location of this tipping point may prove fruitful.

Methodological differences may play a role as well. Worldview backfire effects may not be immediately apparent post-correction. Instead, they may only emerge after some time has elapsed (Reference Peter and KochPeter and Koch 2016; Reference Pluviano, Watt and Della SalaPluviano, Watt, and Della Sala 2017). However, most research on worldview backfire effects just measures the effect of corrections after a short distraction task. In contrast, studies that do incorporate lengthy time delays (e.g., Reference BerinskyBerinsky 2017; Reference Swire, Berinsky, Lewandowsky and EckerSwire, Berinsky et al. 2017) find that the benefits of corrections quickly dissipate. Worldview backfire effects may therefore only be visible after a delay. Studies of these effects should therefore aim to measure responses at multiple points in time. In addition, worldview backfire effects are more probable for high-salience issues where individuals have strong prior attitudes (Reference Flynn, Nyhan and ReiflerFlynn et al. 2017). Nevertheless, the deep-rooted nature of these issues may limit the range of effect sizes that a single experimental manipulation can elicit. As a result, worldview backfire effects may be especially hard to detect for highly polarized issues – the very issues where we would expect the most pervasive effects.

On a related note, it is essential to come to some consensus regarding what, exactly, we consider a “backfire effect.” In particular, worldview backfire effects may be an artifact of the baseline against which they are measured. Scholars generally define worldview backfire effects as cases where the presentation of both misinformation and its correction is worse than presenting misinformation uncorrected. Yet when considering the deleterious effects of misinformation in society, a more expansive definition may be appropriate. Worldview backfire effects are commonly measured experimentally by comparing respondents who were exposed to misinformation to respondents who were exposed to both misinformation and corrections. However, when examining information that has already spread through society – beliefs about President Obama’s citizenship, for example – a better baseline might be people’s beliefs if they had not been reexposed to misinformation as part of an experiment. If providing a correction to misinformation is worse than providing no information at all, strategies for mitigating misinformation may require substantial adjustment.

Continued Influence Effect

The ubiquity of worldview backfire effects remains an open question. However, even if these effects are overblown, valid concerns about the unintended consequences of corrections remain. In particular, the format in which corrections are delivered may bolster beliefs in misinformation, even in the absence of worldview backfire effects. A near-universal finding in the misinformation literature is that, even after its correction, misinformation continues to influence people’s attitudes and beliefs (for a review, see Reference Walter and TukachinskyWalter and Tukachinsky 2019). This is known as the continued influence effect (Reference Wilkes and LeatherbarrowWilkes and Leatherbarrow 1988; Reference Johnson and SeifertJohnson and Seifert 1994). Importantly, people may correctly recall a retraction yet still use outdated misinformation when reasoning about an event. From this perspective, corrections can partially reduce misperceptions but cannot fully eliminate reliance on misinformation in later judgments.

Why does misinformation linger post-correction? Scholars suggest two potential reasons for the continued influence effect. First, according to the mental model theory, individuals construct models of external events in their heads, which they continuously update as new information becomes available (Reference Johnson and SeifertJohnson and Seifert 1994; Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker 2018). However, retractions often threaten the internal coherence of these models (Reference Gordon, Brooks, Quadflieg, Ecker and LewandowskyGordon et al. 2017; Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker 2018). As a result, even if individuals explicitly recall corrections, they may nevertheless continue to invoke misinformation until a plausible alternative takes its place. Numerous studies find that corrections are more effective when they contain alternative causal accounts rather than just negate the original misinformation (Reference Johnson and SeifertJohnson and Seifert 1994; Reference Ecker, Lewandowsky and TangEcker, Lewandowsky, and Tang 2010; Reference Nyhan and ReiflerNyhan and Reifler 2015a; but see Reference Ecker, Lewandowsky, Cheung and MayberyEcker et al. 2015).

Secondly, the continued influence effect can be understood through dual-process theory. Dual-process theory distinguishes between two types of memory retrieval: automatic and strategic. Automatic processing is fast and unconscious, whereas strategic processing is deliberate and effortful. In addition, automatic processing is relatively acontextual, distilling information down only to its most essential properties, whereas strategic processing is required to retrieve specific details about a piece of information (Reference Ecker, Lewandowsky, Swire and ChangEcker et al. 2011). As a result, individuals may be able to remember a piece of misinformation but not recall relevant features, such as its source or perceived accuracy (Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker 2018). In this view, the continued influence effect constitutes a form of retrieval failure; misinformation is automatically retrieved, but its retraction is not. This emphasis on automatic versus strategic processing is also consistent with an online processing model of misinformation (Reference Lodge and TaberLodge and Taber 2013; Reference ThorsonThorson 2016). According to this model, initial misinformation is encoded with a stronger affective charge than its correction, meaning that misinformation will continue to dominate subsequent evaluations until individuals engage in the strategic processing necessary to explicitly recall a correction.

Most direct studies of the continued influence effect use variants of the same research design, based on the “warehouse fire” script (Reference Wilkes and LeatherbarrowWilkes and Leatherbarrow 1988; Reference Johnson and SeifertJohnson and Seifert 1994). In this scenario, the cause of a fire is initially attributed to volatile chemicals stored in a closet, but the closet is later revealed to have been empty. Subsequent studies have adapted this narrative to other contexts, such as police reports or political misconduct, but all follow a similar format in which information about a breaking news event is relayed over a series of short messages. Experimenters randomly assign some subjects to read a critical piece of misinformation (e.g., the presence of flammable materials) as well as its retraction (e.g., the empty closet). However, this communication technique is arguably ill-suited to the study of political misinformation. First, many of these studies present misinformation and corrections as coming from the same source. However, in the realm of politics, the sources most likely to issue corrections may be the ones least likely to spread the misinformation in the first place. Second, the sequencing of messages may not accurately mimic how individuals encounter information in the real world, where the temporal distance may be either much shorter (instantaneous, if people see a correction before or concurrently with the original misinformation) or much longer (if corrections are issued at a later date). Finally, these studies usually rely on fictional scenarios that do not implicate social identities or prior attitudes (but see Reference Ecker, Lewandowsky, Fenton and MartinEcker et al. 2014), both of which may increase the likelihood of the continued influence effect.

Accordingly, recent work has sought to investigate the presence of the continued influence effect in the political domain. Most notably, Reference ThorsonThorson (2016) introduces the concept of “belief echoes,” a version of the continued influence effect focused on attitudes rather than causal inferences. According to her theory, misinformation may continue to influence political attitudes through two separate processes. First, automatic belief echoes develop as a byproduct of online processing. Even when individuals accept corrections as true, misinformation may still be automatically activated, thereby continuing to affect attitudes outside of conscious awareness. Deliberative belief echoes, on the other hand, occur when individuals assume that the existence of one piece of negative information – even if it is known to be false – increases the likelihood that other relevant negative information is true (a “where there’s smoke, there’s fire” philosophy). Together, these automatic and deliberative belief echoes may contribute to the perpetuation of misinformation post-correction.

Familiarity Backfire Effects

The continued influence effect suggests that corrections are somewhat, though not entirely, effective in reducing belief in misinformation. In fact, contrary to worldview backfire effects, the continued influence effect does not require the existence of strong prior attitudes. However, backfire effects might occur even in the absence of worldview threat. In particular, corrections that repeat misinformation may amplify its influence, constituting an alternate form of backlash known as familiarity backfire effects. Of note, within the political science literature, the term “backfire effect” almost exclusively refers to worldview backfire effects. However, familiarity backfire effects are a much more common area of focus within the psychology literature. To avoid confusion, we treat these concepts as separate phenomena.

Familiarity backfire effects involve cases in which retractions increase, rather than reduce, reliance on misinformation by making misinformation feel more familiar. These effects are primarily studied in the context of repetition. In particular, familiarity backfire effects are considered the product of the illusory truth effect, wherein “repeated statements are easier to process, and subsequently perceived to be more truthful, than new statements” (Reference Fazio, Brashier, Payne and MarshFazio et al. 2015, p. 993). The illusory truth effect operates through a series of complementary psychological mechanisms. First, repeating information strengthens its encoding in memory, enabling easier retrieval later on (for reviews, see Reference Lewandowsky, Ecker, Seifert, Schwarz and CookLewandowsky et al. 2012; Reference Peter and KochPeter and Koch 2016). Second, the difficulty with which information is processed influences its perceived authenticity. This is tied to the metacognitive experience of “processing fluency” (Reference Schwarz, Sanna, Skurnik and YoonSchwarz, et al. 2007); information that is easier to process feels more familiar, and familiarity is a key criterion by which individuals judge accuracy (Reference Alter and OppenheimerAlter and Oppenheimer 2009). Accordingly, if individuals have repeated contact with a piece of misinformation, they may perceive it as more credible than if they encounter it only once, regardless of its content.

The illusory truth effect is of particular concern in regard to misinformation correction, given the standard format of corrections. In particular, as part of the debunking process, most corrections directly reference the original misinformation. For instance, the commonly employed “myths vs. facts” strategy involves repeating misinformation (the “myth”) while simultaneously discrediting it (the “fact”). As such, repeated exposure to misinformation – even during its correction – may activate the familiarity heuristic and therefore enhance the perceived accuracy of misinformation. Indeed, familiarity backfire effects have been detected across numerous studies of this specific correction style (Reference Schwarz, Sanna, Skurnik and YoonSchwarz et al. 2007; Reference Peter and KochPeter and Koch 2016; but see Reference Cameron, Roloff and FriesemaCameron et al. 2013).

Familiarity backfire effects may be especially prominent after a time delay. Though individuals are typically able to differentiate fact from fiction immediately after viewing a correction, they may soon forget the details of the correction and retain only the gist of the original misinformation. For example, Reference Skurnik, Yoon and SchwarzSkurnik et al. (2007) find that subjects were able to distinguish between myths and facts about the flu vaccine right after reading an informational flyer but, after only a short break, were significantly more likely to mistake myths for facts than the reverse. In addition, in a study of healthcare reform, Reference BerinskyBerinsky (2017) notes that the effectiveness of corrections faded rapidly over time, with subjects exposed to corrections no more likely than those in a control group to reject a rumor about “death panels” after just a week. Even if corrections are initially able to reduce misperceptions, their benefits may be short-lived.

Familiarity backfire effects are also likely to be relatively universal, as the illusory truth effect is largely robust across individuals and situations. Even when they have prior knowledge about a subject, individuals tend to rate repeated statements as truer than new statements (Reference Fazio, Brashier, Payne and MarshFazio et al. 2015). In addition, the illusory truth effect is only modestly associated with dispositional skepticism (Reference DiFonzo, Beckstead, Stupak and WaldersDiFonzo et al. 2016) and is uncorrelated with several psychological traits, such as analytical thinking and need for closure, that are otherwise connected to the processing of misinformation (Reference De keersmaecker, Dunning and PennycookDe keersmaecker et al. 2020). Finally, the illusory truth effect appears independent of motivated reasoning; across both politically consistent and discordant statements, repeated exposure corresponds to higher accuracy ratings (Reference Pennycook, Cannon and RandPennycook, Cannon, and Rand 2018).

Not all scholars, though, have found evidence of familiarity backfire effects. Although most scholars acknowledge that familiarity affects the processing of corrections, some dispute the negative relationship between repetition of misinformation and belief accuracy (e.g., Reference Swire, Ecker and LewandowskySwire, Ecker, and Lewandowsky 2017; Reference Pennycook, Cannon and RandPennycook et al. 2018). In fact, Reference Ecker, Hogan and LewandowskyEcker, Hogan, and Lewandowsky (2017) find that retractions that include reminders of the original misinformation are more effective than retractions without this repetition.Footnote 5 They attribute these results to the benefits of coactivating misinformation and corrections (see also Reference Swire, Ecker, Southwell, Thorson and ShebleSwire and Ecker 2018). When misinformation and its correction are summoned simultaneously, individuals are better able to detect discrepancies between the original misinformation and the factual evidence. This “conflict detection” expedites the knowledge revision process, leading to more efficient belief updating. In light of these contradictory findings, it remains unclear how concerned we should be about familiarity backfire effects when correcting misinformation. However, we discuss a number of strategies in the following section to minimize the risk of these effects, regardless of their prevalence.

Avoiding Familiarity Backfire Effects

What strategies exist to correct misinformation while evading familiarity backfire effects? The most obvious solution is to focus on the correction without alluding to the original misinformation. However, several of the pieces cited in this chapter suggest that avoiding repetition is not a magic bullet; at times, providing details about a piece of misinformation can aid in the correction process. Moreover, even if avoiding repetition is the goal, this may not always be possible. In many cases, misinformation is published by one source and corrected by another. Rather than just affirm the facts, corrections may need to invoke the original misinformation in order to provide proper contextualization. Instead of avoiding repetition of misinformation, it may thus be more valuable to focus on reiterating corrections, as a means of increasing the familiarity of accurate information (Reference Ecker, Lewandowsky, Swire and ChangEcker et al. 2011).

Furthermore, processing fluency is not solely a function of repetition (Reference Schwarz, Sanna, Skurnik and YoonSchwarz et al. 2007). On the whole, information that is easier to process will be perceived as more familiar (and therefore more valid). Consequently, corrections may be more successful when they are less cognitively taxing. For example, visual corrections may be easier to digest than long-form fact-checking articles (Reference Alter and OppenheimerAlter and Oppenheimer 2009; Reference Schwarz, Newman and LeachSchwarz et al. 2016). Previous studies using photographs (Reference Garrett, Nisbet and LynchGarrett, Nisbet, and Lynch 2013), infographics (Reference Nyhan and ReiflerNyhan and Reifler 2019), and videos (Reference Young, Jamieson, Poulsen and GoldringYoung et al. 2018) largely corroborate this hypothesis (but see Reference Nyhan, Reifler, Richey and FreedNyhan et al. 2014). Similarly, corrections that employ simple words or grammatical structures may be more decipherable than linguistically complex corrections (Reference Alter and OppenheimerAlter and Oppenheimer 2009). The readability of corrections is thus another important consideration for future research to explore. Finally, it may be optimal for corrections to combine multiple approaches. For instance, corrections that pair pithy images (e.g., PolitiFact’s Truth-O-Meter) with accompanying descriptive text may be especially effective (Reference Amazeen, Thorson, Muddiman and GravesAmazeen et al. 2016).

Victims of Misinformation: Moderators of Misinformation and Its Correction

Overall, misinformation appears both pervasive and difficult to correct once it spreads. However, not all misinformation is created equal, nor are all individuals equally susceptible to its influence. Thus, it is important to examine which groups are most likely to be affected by misinformation in society. In the sections that follow, we outline several factors – both individual and contextual – that may affect the persistence of misinformation among certain groups, by making individuals either more likely to believe misinformation or more resistant to its correction.

Individual Factors

We first discuss several individual-level moderators of receptiveness to misinformation and responsiveness to corrections. These factors may be bifurcated into two strands: those that are explicitly political in nature and those that reflect more fundamental personal or psychological orientations.

Political Factors

Two main political factors contribute to the nature and severity of misinformation effects: political sophistication and ideology. One of the most frequently studied moderators of correction effectiveness is political sophistication, which includes aspects of political knowledge, engagement, and education (for a review, see Reference Flynn, Nyhan and ReiflerFlynn et al. 2017). At first glance, more politically sophisticated individuals should be less susceptible to misinformation than less-informed citizens, as they can draw on their superior knowledge to discern fact from fiction. Along these lines, Reference BerinskyBerinsky (2012) finds that more politically engaged individuals are, on the whole, more likely than others to reject political rumors. However, he also finds that politically sophisticated Republicans are more likely to accept rumors about Democrats, suggesting that political knowledge does not entirely inoculate individuals against misinformation.

In fact, belief in misinformation may actually be more prevalent within this more educated and engaged group. Recent research finds that individuals who are more politically active and engaged are more likely to share misinformation via social media, thereby contributing to the spread of misinformation to other members of the public (Reference Valenzuela, Halpern, Katz and MirandaValenzuela et al. 2019). Moreover, politically sophisticated individuals may be more resistant to corrections. In general, politically sophisticated individuals tend to evince the strongest directional motivations (Reference Lodge and TaberLodge and Taber 2013), corresponding to greater endorsement of misinformation that reinforces their prior beliefs (Reference Nyhan, Reifler and UbelNyhan, Reifler, and Ubel 2013; Reference Miller, Saunders and FarhartMiller, Saunders, and Farhart 2016; Reference Jardina and TraugottJardina and Traugott 2019). Reference Nyhan and ReiflerNyhan and Reifler (2010) propose two mechanisms by which this might be the case: a biased information search and biased information processing. First, politically sophisticated individuals may be more likely to selectively consume ideologically consistent media (confirmation bias), thereby filtering out the sources most likely to publish attitude-incongruent corrections. Second, when encountering attitude-incongruent corrections, politically sophisticated individuals may be best equipped to counterargue against these corrections (disconfirmation bias).

The most politically sophisticated individuals seem the least amenable to corrections when misinformation supports their preexisting beliefs. As a result, corrections may fail to reduce and may even enhance belief in misinformation among this small but consequential group. From this perspective, political sophistication is a crucial determinant of responses to misinformation and its correction. Highly sophisticated partisans have both the motivation and the expertise to discount corrections that run counter to their predispositions. Furthermore, less engaged citizens are unlikely to be exposed to corrections in the first place. When considering solutions to the spread of misinformation, the standard prescription is merely to provide more information. However, this heightened susceptibility to misinformation among the most informed citizens exposes the limits to this approach; when individuals are knowledgeable about and involved in politics, this engagement may ironically engender the strongest opposition to corrections. Thus, a more informed populace may not be a panacea if corrections continue to heighten directional motivations.

Political Ideology and Partisanship

An active debate in the misinformation literature concerns potential asymmetries in responses to misinformation based on political ideology and partisan identification (for a review, see Reference Swire, Ecker and LewandowskySwire, Berinsky et al. 2017). Specifically, some scholars claim that conservatives and Republicans are especially vulnerable to misinformation. In a widely cited article, Reference Jost, Glaser, Kruglanski and SullowayJost et al. (2003) catalog a laundry list of predictors of conservatism (e.g., close-mindedness, intolerance of ambiguity), many of which could engender openness to misinformation and resistance to corrections. In a later piece, Reference Jost, van der Linden, Panagopoulos and HardinJost et al. (2018) highlight several other factors associated with conservatism, including an emphasis on in-group consensus and homogeneous social networks, that may give rise to “echo chambers” in which misinformation can easily spread (see also Reference Nam, Jost and Van BavelNam, Jost, and Van Bavel 2013; Reference Ecker and AngEcker and Ang 2019). Taken together, these pieces paint a picture of conservatives as resistant to change, averse to uncertainty, and drawn to one-sided information environments – all of which might predispose those on the right to favor misinformation, relative to their moderate or liberal counterparts.

These theoretical expectations have some empirical backing. Recent research finds that, during the 2016 election, Republicans were more likely than Democrats to read and share fake news (Reference Grinberg, Joseph, Friedland, Swire-Thompson and LazerGrinberg et al. 2019; Reference Guess, Nagler and TuckerGuess, Nagler, and Tucker 2019; Reference Guess, Nyhan and ReiflerGuess et al. 2020). Furthermore, ideology and partisanship are associated with differences in responses to corrections. For example, Reference Nyhan and ReiflerNyhan and Reifler (2010) report evidence of ideological asymmetry in responses to corrections. Although they find that, regardless of partisan leaning, corrections were generally less effective when they were attitude-incongruent, worldview backfire effects were visible for Republicans but not Democrats (see also Reference Ecker and AngEcker and Ang 2019).Footnote 6 These individual-level differences may be exacerbated by system-wide differences in conservative versus liberal media. Although misinformation originates in both liberal and conservative circles, the insular nature of the conservative media ecosystem may be more conducive to the spread of misinformation (Reference Faris, Roberts, Etling, Bourassa, Zuckerman and BenklerFaris et al. 2017; see also Barberá, Chapter 3, this volume), and conservative media sources are more likely than liberal sites to dismiss or otherwise derogate nonpartisan fact-checkers (Reference Iannucci and AdairIannucci and Adair 2017). Finally, these system-wide differences also extend to individual behavior. In an analysis of tweets about the 2012 presidential election, Reference Shin and ThorsonShin and Thorson (2017) find that Republicans retweeted or replied much less frequently to fact-checking sites than Democrats – and their replies tended to be more acrimonious. Similarly, across both Facebook and Twitter, Reference Amazeen, Vargo and HoppAmazeen, Vargo, and Hopp (2018) find that liberal-leaning individuals tend to be more likely than others to share fact-checking information.

However, this emphasis on the psychological profiles of political conservatives is not without controversy. Kahan and colleagues contend that motivated reasoning is not a uniquely right-wing phenomenon. Instead, all individuals are motivated to express and maintain beliefs similar to those of other members of their identity groups (the “cultural cognition thesis,” e.g., Reference Kahan, Jenkins‐Smith and BramanKahan et al. 2011; Reference KahanKahan 2013). In line with this perspective, several recent works suggest that liberals are not, in fact, immune to the effects of misinformation (Reference Aird, Ecker, Swire, Berinsky and LewandowskyAird et al. 2018; Reference Guess, Nagler and TuckerGuess et al. 2019). Across numerous fields, ranging from science to politics, both conservatives and liberals evince similar levels of motivated reasoning (Reference Nisbet, Cooper and GarrettNisbet, Cooper, and Garrett 2015; Reference Meirick and BessarabovaMeirick and Bessarabova 2016; Reference Frimer, Skitka and MotylFrimer, Skitka, and Motyl 2017; Reference Ecker, Hogan and LewandowskySwire, Ecker et al. 2017; Reference Ditto, Liu and ClarkDitto et al. 2019). While conservatives may disproportionately display the motivational tendencies associated with belief in misinformation, these proclivities do not necessarily translate to behavioral differences.

While political knowledge has been firmly established as a key moderator of misinformation effects, via its relationship to directionally motivated reasoning, the jury is still out regarding the role of political ideology and partisanship. Although conservatives and Republicans may, under certain conditions, be more sensitive to misinformation than others, this divide may be overstated. Are observed cases of ideological asymmetry a function of deeply rooted psychological traits, or do they instead reflect systematic differences in conservative versus liberal media environments (or in the misinformation itself)? Future work should continue to grapple with this tricky distinction.

Personal and Psychological Factors

Misinformation, however, is not contained to the political sphere. More basic personal and psychological factors may predispose certain individuals to champion misinformation and disavow corrections across domains. We highlight four of these potential moderators, namely age, analytical thinking, need for closure, and psychological reactance.Footnote 7 Scholars highlight age as a key demographic variable influencing both exposure and responses to misinformation. Several recent studies find that older adults are more likely than others to share fake news stories on social media (Reference Grinberg, Joseph, Friedland, Swire-Thompson and LazerGrinberg et al. 2019; Reference Guess, Nagler and TuckerGuess et al. 2019). However, other work finds that old age is also associated with greater sharing of fact-checks on social media (Reference Amazeen, Vargo and HoppAmazeen et al. 2018), suggesting that older cohorts may engage differently with political content on social media, relative to their younger counterparts.

Scholars have also identified analytical thinking, or a person’s capacity to override gut feelings and intuitions, as another determinant of their responses to misinformation. In this sense, individuals who are more prone to careful, deliberate processing of information (or “cognitive reflection”) seem to be less susceptible to misinformation. Analytical thinking is associated with reduced belief in conspiracy theories (Reference Swami, Voracek, Stieger, Tran and FurnhamSwami et al. 2014) and increased accuracy in judging fake news headlines (Reference Pennycook and RandPennycook and Rand 2018; Reference Bronstein, Pennycook, Bear, Rand and CannonBronstein et al. 2019; Reference Pennycook, Bear, Collins and RandPennycook and Rand 2020). Furthermore, highly analytical individuals are more willing than others to adjust their attitudes post-correction, even after controlling for a host of other variables (Reference De keersmaecker and RoetsDe keersmaecker and Roets 2017; see also Reference Pennycook and RandTappin, Pennycook, and Rand 2018). While most studies conceptualize analytical thinking as a dispositional trait, recent work suggests that interventions designed to encourage greater deliberation may also prove an effective tool for correcting misinformation (Reference Bago, Rand and PennycookBago, Rand, and Pennycook 2020).

Need for closure may also shape an individual’s susceptibility to misinformation. Need for closure refers to “the expedient desire for any firm belief on a given topic, as opposed to confusion and uncertainty” (Reference Jost, Glaser, Kruglanski and SullowayJost et al. 2003, p. 348, italics in original). This motivation fosters two main behavioral inclinations: the propensity to seize on readily available information and the tendency to cling to previous information (Reference Jost, Glaser, Kruglanski and SullowayJost et al. 2003; Reference Meirick and BessarabovaMeirick and Bessarabova 2016; Reference De keersmaecker, Dunning and PennycookDe keersmaecker et al. 2020). Consequently, individuals with a high need for closure may be more trusting of initial misinformation, which provides closure through explaining the causes of events, and more resistant to corrections, which may sow feelings of confusion and uncertainty (Reference Rapp and SalovichRapp and Salovich 2018). Need for closure, however, is primarily used as a control variable in studies of misinformation and is rarely the main construct of interest. Indeed, the few studies connecting a need for closure to misinformation focus solely on the endorsement, rather than correction, of misinformation (e.g., Reference Leman and CinnirellaLeman and Cinnirella 2013; Reference Moulding, Nix-Carnell and SchnabelMoulding et al. 2016; Reference Marchlewska, Cichocka and KossowskaMarchlewska, Cichocka, and Kossowska 2018). Nevertheless, need for closure may also moderate the effectiveness of corrections. For instance, individuals with a high need for closure may be especially vulnerable to the continued influence effect; if these individuals are less acceptant of gaps in their mental models of an event, they may be more likely to retain misinformation in the absence of plausible alternative explanations. Moving forward, future research should continue to probe the extent to which a high need for closure predisposes certain individuals to disregard corrections.

Finally, high levels of psychological reactance may trigger backfire effects by stimulating counterarguing. Psychological reactance occurs when individuals perceive a threat to their intellectual or behavioral freedoms, such as when they feel strong pressure to adopt a certain attitude or belief (Reference Sensenig and BrehmSensenig and Brehm 1968). In short, many people do not like being told what or how to think. As a result, they may actively defy corrections that seem overly authoritative (Reference Garrett, Nisbet and LynchGarrett et al. 2013; Reference Weeks and GarrettWeeks and Garrett 2014). Misperceptions may thus be even more difficult to remedy for individuals who eschew conformity. Indeed, across countries, anti-vaccination attitudes are significantly and positively correlated with psychological reactance (Reference Hornsey, Harris and FieldingHornsey, Harris, and Fielding 2018). Moreover, several studies document a link between psychological reactance and resistance to climate change messaging (Reference Nisbet, Cooper and GarrettNisbet et al. 2015; Reference Ma, Dixon and HmielowskiMa, Dixon, and Hmielowski 2019). A deeper focus on psychological reactance may therefore help reconcile previously perplexing findings in the misinformation literature. Some accounts of the continued influence effect posit that individuals continue to endorse misinformation because they do not believe corrections to be true (Reference Guillory and GeraciGuillory and Geraci 2013). This tendency may be heightened among those with a contrarian streak. In addition, several scholars caution against providing too many corrections (“overkill” backfire effects, see Reference Cook and LewandowskyCook and Lewandowsky 2011; Reference Lewandowsky, Ecker, Seifert, Schwarz and CookLewandowsky et al. 2012; Reference Ecker, Lewandowsky, Jayawardana and MladenovicEcker et al. 2019). The purported perils of overcorrection may have their roots in psychological reactance (Reference Shu and CarlsonShu and Carlson 2014); inundating people with a surfeit of corrections may provoke feelings of reactance, particularly among those already liable to reject consensus views.

Contextual Factors

Along with individual-level moderators of misinformation effects, contextual factors may play an important role in guiding responses to misinformation and its correction. These variables include the content of misinformation as well as the environments in which misinformation is consumed and corrected.

Content-Based Factors

The actual substance of misinformation – including its subject matter and tone – is an important determinant of its correctability. First, corrections may be differentially effective across issue areas. For example, in a meta-analysis of studies of misinformation correction, Reference Walter and MurphyWalter and Murphy (2018) find that corrections are more effective for health-focused misinformation than for political and scientific misinformation. Second, misinformation may vary in its affective content. Negatively valenced misinformation tends to be more durable than positive or neutral misinformation (Reference Forgas, Laham and VargasForgas, Laham, and Vargas 2005; Reference Guillory and GeraciGuillory and Geraci 2016; but see Reference Mirandola and ToffaliniMirandola and Toffalini 2016). Moreover, the emotions that misinformation arouses may also influence its persistence (Reference Vosoughi, Roy and AralVosoughi et al. 2018). In particular, Reference WeeksWeeks (2015) finds that feelings of anger tend to encourage directionally motivated processing of corrections, whereas feelings of anxiety tend to reduce partisan differences in responses to corrections. However, misinformation does not seem to inspire these emotions in equal measure. Text analysis of comments on Facebook posts containing misinformation finds that responses to misinformation are more frequently characterized by anger as opposed to anxiety (Reference BarfarBarfar 2019).

Environmental Factors

How people encounter misinformation may also influence both their contact with and their responses to corrections. Although misinformation is an age-old problem, the topic has garnered attention in recent years due to concerns about how the Internet – and especially social media – might extend its reach. Many producers of misinformation use social media sites as their main means of disseminating misinformation (Reference Tucker, Guess and BarberaTucker et al. 2018). Reflecting this fact, several recent studies emphasize the role of social networking sites, including Facebook and Twitter, in amplifying exposure to fake news content (Reference Allcott and GentzkowAllcott and Gentzkow 2017; Reference Allcott, Gentzkow and YuAllcott, Gentzkow, and Yu 2019; Reference Guess, Nyhan and ReiflerGuess, Nyhan, and Reifler 2020). However, some work suggests that exposure to fake news on social media is limited to only a small subset of the population (Reference Grinberg, Joseph, Friedland, Swire-Thompson and LazerGrinberg et al. 2019; Reference Guess, Nagler and TuckerGuess et al. 2019), and others find that social media use is only weakly associated with the endorsement of false information (Reference GarrettGarrett 2019).

Even if misinformation may propagate easily via social media, these platforms may be essential to combating its spread. After all, social media can spread corrections in addition to misinformation (Reference VragaVraga 2019). Much work focuses on efforts by social media sites to prevent the spread of misinformation or other harmful rhetoric in the first place (for reviews, see Guess and Lyons, Chapter 2, and Siegel, Chapter 4, this volume). However, social media platforms can also play an active role in correcting misinformation after the fact. To this end, scholars have studied the effectiveness of two types of social media–based corrections: algorithmic and social corrections. Some social media sites have built-in functionalities that can be deployed to combat misinformation. For example, Reference Bode and VragaBode and Vraga (2015, Reference Bode and Vraga2018) focus on Facebook’s “related stories” feature, which recommends relevant articles underneath shared links, and find that fact-checking articles publicized through this system may be effective in increasing belief accuracy – especially on issues where individuals do not possess strong prior attitudes. Another proposed form of algorithmic correction relies on “crowdsourced” data on the trustworthiness of different news outlets to decrease the likelihood that individuals will encounter posts from unreliable sources (Reference Pennycook and RandPennycook and Rand 2019). In addition to these algorithmic corrections, other social media users (e.g., Facebook friends or Twitter followers) can intervene to provide corrections (Reference Vraga and BodeVraga and Bode 2018). These social corrections may be especially effective, as individuals are more likely to accept corrections from people they already know (Reference Friggeri, Adamic, Eckles and ChengFriggeri et al. 2014; Reference Margolin, Hannak and WeberMargolin et al. 2018).

However, some scholars caution about the potential for social media to undermine the correction of misinformation. The “social” nature of social media may increase levels of exposure to misinformation, as individuals are more likely to read news that has been shared or endorsed by members of their social networks (Reference Messing and WestwoodMessing and Westwood 2014; Reference AnspachAnspach 2017). The nature of the social media environment may also inhibit corrections of misinformation; Reference Jun, Meng and JoharJun, Meng, and Johar (2017) warn that people are less likely to fact-check statements in social settings – a form of “virtual bystander effect.” Furthermore, even if corrections circulate on social media, individuals may be more attentive to user comments on these posts than to the actual fact-checking messages themselves. If these comments distort or otherwise misrepresent corrections, individuals may not become better informed, despite their exposure to fact-checking information (Reference Anspach and CarlsonAnspach and Carlson 2018).

Finally, and most importantly, corrective efforts on social media may have unintended consequences. Given the difficulties of correcting misinformation postexposure, many scholars recommend preemptive interventions designed to induce skepticism prior to misinformation exposure (Reference Ecker, Lewandowsky and TangEcker et al. 2010; Reference Peter and KochPeter and Koch 2016; Reference Cook, Lewandowsky and EckerCook, Lewandowsky, and Ecker 2017). Specifically, some scholars recommend training individuals to detect and resist misinformation by highlighting the techniques commonly deployed by creators of misinformation (Reference Roozenbeek and van der LindenRoozenbeek and van der Linden 2019a, Reference Roozenbeek and van der Linden2019b). Social networking sites have adopted similar models. For example, after the 2016 US presidential election, Facebook rolled out a new system to flag potentially inaccurate stories as disputed or false. Warning labels of this sort may be effective in reducing the sharing of flagged stories (Reference MenaMena 2019). However, false stories that go undetected by this system may be viewed as more accurate than they would have were the system never put in place (Reference Pennycook, Bear, Collins and RandPennycook et al. 2020). Similarly, general warnings about the potentially misleading nature of social media posts may decrease beliefs in the accuracy of true headlines (Reference Clayton, Blair and BusamClayton et al. 2019), suggesting that corrections issued on social media might inadvertently erode trust in credible media content.

Exposure to Fact Checks

Only a small subset of the population will likely encounter both misinformation and corrections. On the misinformation side, while some types of misinformation are widespread (e.g., the birther movement), many remain fringe beliefs. Despite rampant fears about “fake news,” fake news sites during the 2016 and 2018 elections received the bulk of their traffic from a very small set of highly partisan consumers (Reference Grinberg, Joseph, Friedland, Swire-Thompson and LazerGrinberg et al. 2019; Reference Guess, Nagler and TuckerGuess et al. 2019, Reference Guess, Nyhan and Reifler2020). On the corrections side, a limited number of people view a limited number of corrections. Relatively few people ever visit professional fact-checking sites, such as PolitiFact or Factcheck.org, without external prompting; the public appreciates fact-checking in theory but shows little interest in practice (Nyhan and Reifler 2015c). These low levels of engagement are exacerbated by patterns of selective exposure to and sharing of fact-checking messages on social media platforms (Reference Shin and ThorsonShin and Thorson 2017; Reference Zollo, Bessi and Del VicarioZollo et al. 2017; Reference Hameleers and van der MeerHameleers and van der Meer 2020), as partisans tend to seek out and share fact checks that reinforce their prior attitudes. If highly engaged members of the public cherry-pick favorable fact-checking messages to share with others, those exposed to these messages may observe only a narrow, unrepresentative slice of the available set of corrections.

Finally, even if individuals do take the initiative to visit fact-checking sites, these sites frequently choose to cover markedly different topics. In fact, even when their coverage does overlap, fact-checking organizations often reach diametrically opposed conclusions about the factual basis for a given piece of information (Reference Marietta, Barker and BowserMarietta, Barker, and Bowser 2015). These potential discrepancies are consequential, as several studies of fact-checking messages find that the content of these messages (e.g., affirming or refuting information) matters more than their source (e.g., Fox News, MSNBC, or PolitiFact) in increasing belief accuracy (Reference WintersieckWintersieck 2017; Reference Wintersieck, Fridkin and KenneyWintersieck et al. 2018).

Conclusion

Within both popular media and academia, concerns abound regarding the prevalence and persistence of misinformation. In an age where misinformation can diffuse rapidly via the Internet and social media, it is more imperative than ever to think creatively about how best to debunk misinformation. Although misinformation may take many forms – ranging from political rumors to disinformation – each of these forms presents a potential threat to democracy by distorting attitudes, behavior, and public policy. Definitional concerns should therefore take a backseat to mitigating the harmful effects of misinformation. Given the potential dangers of misinformation, devising effective strategies for correction is crucial, yet previous prescriptions have often come up short.

In this review, we have discussed two phenomena that may contribute to the durability of misinformation post-correction: the continued influence effect and backfire effects. Though scholars have found evidence that each of these processes undermines the effectiveness of corrections, recent works have cast doubt on their pervasiveness. In light of these findings, several areas merit further research. First, although worldview backfire effects may be less widespread than originally thought, the existence of these effects remains an open question. Efforts to isolate the conditions, both theoretical and methodological, under which worldview backfire effects are most likely to occur may help to resolve this ongoing debate. Similarly, though scholars frequently discourage the repetition of misinformation within corrections, more recent studies have cast doubt on the prevalence of familiarity backfire effects. Given that traditional methods of correction often cite the original misinformation, understanding whether and how this repetition might undercut their effectiveness is important. In particular, clarifying the conditions under which repetition is a benefit versus a hindrance may yield practical recommendations for improving the success of fact-checking sites. Finally, misinformation does not affect all individuals equally, nor is all misinformation equally persuasive. Continuing to identify these places of heterogeneity may enable more active targeting of corrections to those subgroups where misinformation is most likely to take root.

Footnotes

1 Indeed, though we attempt to provide a comprehensive review of the literature on misinformation correction, the field is moving so fast that this review may soon be out of date.

2 We are certainly not the first to propose such a typology (see Reference Born and EdgingtonBorn and Edgington 2017; Reference Tucker, Guess and BarberaTucker et al. 2018; Reference WardleWardle 2018). However, we take a more comprehensive view than many of these previous works in that we seek to integrate a larger number of related concepts into a common theoretical framework.

4 Counterarguing typically involves generating arguments to dispute a correction. Inverting this process, Reference Chan, Jones, Hall Jamieson and AlbarracínChan et al. (2017) also find that corrections are generally less effective when people are asked to record arguments in favor of the original misinformation.

5 As they note, however, their experimental design includes only a short distraction task (30 minutes) separating the presentation of misinformation and its correction from measurement of their dependent variables. Although this time interval is consistent with previous studies (e.g., Reference Skurnik, Yoon and SchwarzSkurnik, Yoon, and Schwarz 2007), it is possible that their results would be different after a longer delay.

6 This observed asymmetry, however, cannot be definitively ascribed to individual-level differences across ideological groups. For instance, there may be qualitative differences between conservative- and liberal-leaning misinformation that make the former stickier. In particular, Reference Nyhan and ReiflerNyhan and Reifler’s (2010) experiments rely on actual examples of misinformation (stem cell research and weapons of mass destruction in Iraq). While this approach has the benefit of greater external validity, these cases may diverge in notable ways beyond their ideological slant (e.g., issue salience or importance). Studies that focus on fabricated misinformation, rather than real-world rumors, may thus be better suited to identifying potential partisan or ideological asymmetries.

7 Of course, these factors may be correlated with political sophistication and ideology and may therefore be at the root of some of the empirical regularities cited in the “Political Factors” section.

References

Aird, M. J., Ecker, U. K. H., Swire, B., Berinsky, A. J., & Lewandowsky, S. (2018). Does truth matter to voters? The effects of correcting political misinformation in an Australian sample. Royal Society Open Science, 5(12), 180593. https://doi.org/10.1098/rsos.180593CrossRefGoogle Scholar
Allcott, H., & Gentzkow, M. (2017). Social media and fake news in the 2016 election. Journal of Economic Perspectives, 31(2), 211236. https://doi.org/10.1257/jep.31.2.211Google Scholar
Allcott, H., Gentzkow, M., & Yu, C. (2019). Trends in the diffusion of misinformation on social media. Research & Politics, 6(2),. https://doi.org/10.1177/2053168019848554Google Scholar
Alter, A. L., & Oppenheimer, D. M. (2009). Uniting the tribes of fluency to form a metacognitive nation. Personality and Social Psychology Review, 13(3), 219235. https://doi.org/10.1177/1088868309341564Google Scholar
Amazeen, M. A., Thorson, E., Muddiman, A., & Graves, L. (2016). Correcting political and consumer misperceptions: The effectiveness and effects of rating scale versus contextual correction formats. Journalism & Mass Communication Quarterly, 95(1), 2848. https://doi.org/10.1177/1077699016678186Google Scholar
Amazeen, M. A., Vargo, C. J., & Hopp, T. (2018). Reinforcing attitudes in a gatewatching news era: Individual-level antecedents to sharing fact-checks on social media. Communication Monographs, 86(1), 112132. https://doi.org/10.1080/03637751.2018.1521984CrossRefGoogle Scholar
Anspach, N. M. (2017). The new personal influence: How our Facebook friends influence the news we read. Political Communication, 34(4), 590606. https://doi.org/10.1080/10584609.2017.1316329Google Scholar
Anspach, N. M., & Carlson, T. N. (2018). What to believe? Social media commentary and belief in misinformation. Political Behavior, 122. https://doi.org/10.1007/s11109-018-9515-zGoogle Scholar
Bago, B., Rand, D., & Pennycook, G. (2020). Fake news, fast and slow: Deliberation reduces belief in false (but not true) news headlines. Journal of Experimental Psychology: General. https://doi.org/10.1037/xge0000729CrossRefGoogle Scholar
Barfar, A. (2019). Cognitive and affective responses to political disinformation in Facebook. Computers in Human Behavior, 101, 173179. https://doi.org/10.1016/j.chb.2019.07.026Google Scholar
Barrera, O. D., Guriev, S. M., Henry, E., & Zhuravskaya, E. (2020). Facts, alternative facts, and fact checking in times of post-truth politics. Journal of Public Economics, 182, 104123. https://doi.org/10.1016/j.jpubeco.2019.104123Google Scholar
Benegal, S. D., & Scruggs, L. A. (2018). Correcting misinformation about climate change: The impact of partisanship in an experimental setting. Climatic Change, 148(1–2), 6180. https://doi.org/10.1007/s10584-018-2192-4Google Scholar
Berinsky, A. J. (2012). Rumors, truths, and reality: A study of political misinformation. Working Paper, Massachusetts Institute of Technology.Google Scholar
Berinsky, A. J.(2017). Rumors and health care reform: Experiments in political misinformation. British Journal of Political Science, 47(2), 241262. https://doi.org/10.1017/S0007123415000186CrossRefGoogle Scholar
Bode, L., & Vraga, E. K. (2015). In related news, that was wrong: The correction of misinformation through related stories functionality in social media: In related news. Journal of Communication, 65(4), 619638. https://doi.org/10.1111/jcom.12166CrossRefGoogle Scholar
Bode, L., & Vraga, E. K.(2018). See something, say something: Correction of global health misinformation on social media. Health Communication, 33(9), 11311140. https://doi.org/10.1080/10410236.2017.1331312CrossRefGoogle ScholarPubMed
Born, K., & Edgington, N. (2017). Analysis of Philanthropic Opportunities to Mitigate the Disinformation/Propaganda Problem. Hewlett Foundation report. https://hewlett.org/wp-content/uploads/2017/11/Hewlett-Disinformation-Propaganda-Report.pdfGoogle Scholar
Bronstein, M. V., Pennycook, G., Bear, A., Rand, D. G., & Cannon, T. D. (2019). Belief in fake news is associated with delusionality, dogmatism, religious fundamentalism, and reduced analytic thinking. Journal of Applied Research in Memory and Cognition, 8(1), 108117. https://doi.org/10.1016/j.jarmac.2018.09.005Google Scholar
Cameron, K. A., Roloff, M. E., Friesema, E. M. et al. (2013). Patient knowledge and recall of health information following exposure to “facts and myths” message format variations. Patient Education and Counseling, 92(3), 381387. https://doi.org/10.1016/j.pec.2013.06.017Google Scholar
Carnahan, D., Hao, Q., Jiang, X., & Lee, H. (2018). Feeling fine about being wrong: The influence of self-affirmation on the effectiveness of corrective information. Human Communication Research, 44(3), 274298. https://doi.org/10.1093/hcr/hqy001Google Scholar
Chan, M. S., Jones, C. R., Hall Jamieson, K., & Albarracín, D. (2017). Debunking: A meta-analysis of the psychological efficacy of messages countering misinformation. Psychological Science, 28(11), 15311546. https://doi.org/10.1177/0956797617714579CrossRefGoogle ScholarPubMed
Clayton, K., Blair, S., Busam, J. A. et al. (2019). Real solutions for fake news? Measuring the effectiveness of general warnings and fact-check tags in reducing belief in false stories on social media. Political Behavior, 123. https://doi.org/10.1007/s11109-019-09533-0Google Scholar
Cohen, G. L., Aronson, J., & Steele, C. M. (2000). When beliefs yield to evidence: Reducing biased evaluation by affirming the self. Personality and Social Psychology Bulletin, 26(9), 11511164. https://doi.org/10.1177/01461672002611011Google Scholar
Cohen, G. L., Sherman, D. K., Bastardi, A., Hsu, L., McGoey, M., & Ross, L. (2007). Bridging the partisan divide: Self-affirmation reduces ideological closed-mindedness and inflexibility in negotiation. Journal of Personality and Social Psychology, 93(3), 415430. https://doi.org/10.1037/0022-3514.93.3.415Google Scholar
Cook, J., & Lewandowsky, S. (2011). The Debunking Handbook. St. Lucia: University of Queensland. http://sks.to/debunkGoogle Scholar
Cook, J., Lewandowsky, S., & Ecker, U. K. H. (2017). Neutralizing misinformation through inoculation: Exposing misleading argumentation techniques reduces their influence. PLoS ONE, 12(5), e0175799. https://doi.org/10.1371/journal.pone.0175799CrossRefGoogle ScholarPubMed
De keersmaecker, J., & Roets, A. (2017). “Fake news”: Incorrect, but hard to correct. The role of cognitive ability on the impact of false information on social impressions. Intelligence, 65, 107110. https://doi.org/10.1016/j.intell.2017.10.005Google Scholar
De keersmaecker, J., Dunning, D., Pennycook, G. et al. (2020). Investigating the robustness of the illusory truth effect across individual differences in cognitive ability, need for cognitive closure, and cognitive style. Personality and Social Psychology Bulletin, 46(2), 204215. https://doi.org/10.1177/0146167219853844Google Scholar
DiFonzo, N., Beckstead, J. W., Stupak, N., & Walders, K. (2016). Validity judgments of rumors heard multiple times: The shape of the truth effect. Social Influence, 11(1), 2239. https://doi.org/10.1080/15534510.2015.1137224Google Scholar
Ditto, P. H., Liu, B. S., Clark, C. J. et al. (2019). At least bias is bipartisan: A meta-analytic comparison of partisan bias in liberals and conservatives. Perspectives on Psychological Science, 14(2), 273291.Google Scholar
Druckman, J. N., & McGrath, M. C. (2019). The evidence for motivated reasoning in climate change preference formation. Nature Climate Change, 9(2), 111119. https://doi.org/10.1038/s41558-018-0360-1Google Scholar
Ecker, U. K. H., & Ang, L. C. (2019). Political attitudes and the processing of misinformation corrections. Political Psychology, 40(2), 241260. https://doi.org/10.1111/pops.12494CrossRefGoogle Scholar
Ecker, U. K. H., Hogan, J. L., & Lewandowsky, S. (2017). Reminders and repetition of misinformation: Helping or hindering its retraction? Journal of Applied Research in Memory and Cognition, 6(2), 185192. https://doi.org/10.1016/j.jarmac.2017.01.014CrossRefGoogle Scholar
Ecker, U. K. H., Lewandowsky, S., Cheung, C. S. C., & Maybery, M. T. (2015). He did it! She did it! No, she did not! Multiple causal explanations and the continued influence of misinformation. Journal of Memory and Language, 85, 101115. https://doi.org/10.1016/j.jml.2015.09.002Google Scholar
Ecker, U. K. H., Lewandowsky, S., Fenton, O., & Martin, K. (2014). Do people keep believing because they want to? Preexisting attitudes and the continued influence of misinformation. Memory & Cognition, 42(2), 292304. https://doi.org/10.3758/s13421-013-0358-xGoogle Scholar
Ecker, U. K. H., Lewandowsky, S., Jayawardana, K., & Mladenovic, A. (2019). Refutations of equivocal claims: No evidence for an ironic effect of counterargument number. Journal of Applied Research in Memory and Cognition, 8(1), 98107. https://doi.org/10.1016/j.jarmac.2018.07.005Google Scholar
Ecker, U. K. H., Lewandowsky, S., Swire, B., & Chang, D. (2011). Correcting false information in memory: Manipulating the strength of misinformation encoding and its retraction. Psychonomic Bulletin & Review, 18(3), 570578. https://doi.org/10.3758/s13423-011-0065-1Google Scholar
Ecker, U. K. H., Lewandowsky, S., & Tang, D. T. W. (2010). Explicit warnings reduce but do not eliminate the continued influence of misinformation. Memory & Cognition, 38(8), 10871100. https://doi.org/10.3758/MC.38.8.1087CrossRefGoogle Scholar
Faris, R., Roberts, H., Etling, B., Bourassa, N., Zuckerman, E., & Benkler, Y. (2017). Partisanship, propaganda, and disinformation: Online media and the 2016 U.S. presidential election. Berkman Klein Center for Internet & Society at Harvard University research publication. https://dash.harvard.edu/handle/1/33759251Google Scholar
Fazio, L. K., Brashier, N. M., Payne, B. K., & Marsh, E. J. (2015). Knowledge does not protect against illusory truth. Journal of Experimental Psychology: General, 144(5), 9931002. https://doi.org/10.1037/xge0000098CrossRefGoogle Scholar
Feinberg, M., & Willer, R. (2015). From gulf to bridge: When do moral arguments facilitate political influence? Personality and Social Psychology Bulletin, 41(12), 16651681. https://doi.org/10.1177/0146167215607842Google Scholar
Fetzer, J. H. (2004). Disinformation: The use of false information. Minds & Machines, 14(2), 231240.Google Scholar
Flynn, D. J., Nyhan, B., & Reifler, J. (2017). The nature and origins of misperceptions: Understanding false and unsupported beliefs about politics. Political Psychology, 38(S1), 127150. https://doi.org/10.1111/pops.12394Google Scholar
Forgas, J. P., Laham, S. M., & Vargas, P. T. (2005). Mood effects on eyewitness memory: Affective influences on susceptibility to misinformation. Journal of Experimental Social Psychology, 41(6), 574588. https://doi.org/10.1016/j.jesp.2004.11.005Google Scholar
Friggeri, A., Adamic, L. A., Eckles, D., & Cheng, J. (2014). Rumor cascades. In Proceedings of the Eighth International AAAI Conference on Weblogs and Social Media (pp. 101110). Palo Alto, CA: AAAI Press.Google Scholar
Frimer, J. A., Skitka, L. J., & Motyl, M. (2017). Liberals and conservatives are similarly motivated to avoid exposure to one another’s opinions. Journal of Experimental Social Psychology, 72, 112. https://doi.org/10.1016/j.jesp.2017.04.003Google Scholar
Garrett, R. K. (2019). Social media’s contribution to political misperceptions in U.S. presidential elections. PLoS ONE, 14(3), e0213500. https://doi.org/10.1371/journal.pone.0213500Google Scholar
Garrett, R. K., Nisbet, E. C., & Lynch, E. K. (2013). Undermining the corrective effects of media-based political fact checking? The role of contextual cues and naïve theory. Journal of Communication, 63(4), 617637. https://doi.org/10.1111/jcom.12038CrossRefGoogle Scholar
Gordon, A., Brooks, J. C. W., Quadflieg, S., Ecker, U. K. H., & Lewandowsky, S. (2017). Exploring the neural substrates of misinformation processing. Neuropsychologia, 106, 216224. https://doi.org/10.1016/j.neuropsychologia.2017.10.003Google Scholar
Grinberg, N., Joseph, K., Friedland, L., Swire-Thompson, B., & Lazer, D. (2019). Fake news on Twitter during the 2016 U.S. presidential election. Science, 363(6425), 374378.Google Scholar
Guess, A., & Coppock, A. (2018). Does counter-attitudinal information cause backlash? Results from three large survey experiments. British Journal of Political Science, 119. https://doi.org/10.1017/S0007123418000327Google Scholar
Guess, A., Nagler, J., & Tucker, J. (2019). Less than you think: Prevalence and predictors of fake news dissemination on Facebook. Science Advances, 5(1), eaau4586.Google Scholar
Guess, A., Nyhan, B., & Reifler, J. (2020). Exposure to untrustworthy websites in the 2016 US election. Nature Human Behaviour, 19. https://doi.org/10.1038/s41562-020-0833-xCrossRefGoogle Scholar
Guillory, J. J., & Geraci, L. (2013). Correcting erroneous inferences in memory: The role of source credibility. Journal of Applied Research in Memory and Cognition, 2(4), 201209. https://doi.org/10.1016/j.jarmac.2013.10.001CrossRefGoogle Scholar
Guillory, J. J., & Geraci, L. (2016). The persistence of erroneous information in memory: The effect of valence on the acceptance of corrected information. Applied Cognitive Psychology, 30(2), 282288. https://doi.org/10.1002/acp.3183Google Scholar
Haglin, K. (2017). The limitations of the backfire effect. Research & Politics, 4(3), 15. https://doi.org/10.1177/2053168017716547Google Scholar
Hameleers, M., & van der Meer, T. G. L. A. (2020). Misinformation and polarization in a high-choice media environment: How effective are political fact-checkers? Communication Research, 47(2), 227250.Google Scholar
Hart, P. S., & Nisbet, E. C. (2012). Boomerang effects in science communication: How motivated reasoning and identity cues amplify opinion polarization about climate mitigation policies. Communication Research, 39(6), 701723. https://doi.org/10.1177/0093650211416646Google Scholar
Hochschild, J. L., & Einstein, K. L. (2015). Do Facts Matter? Information and Misinformation in American Politics (1st ed.). Norman: University of Oklahoma Press.Google Scholar
Holman, M. R., & Lay, J. C. (2019). They see dead people (voting): Correcting misperceptions about voter fraud in the 2016 U.S. presidential election. Journal of Political Marketing, 18(1–2), 3168. https://doi.org/10.1080/15377857.2018.1478656Google Scholar
Hornsey, M. J., Harris, E. A., & Fielding, K. S. (2018). The psychological roots of anti-vaccination attitudes: A 24-nation investigation. Health Psychology, 37(4), 307315. https://doi.org/10.1037/hea0000586Google Scholar
Iannucci, R., & Adair, B. (2017). Heroes or Hacks: The Partisan Divide Over Fact-Checking. Duke Reporters’ Lab report. https://drive.google.com/file/d/0BxoyrEbZxrAMNm9HV2tvcXFma1U/view?usp=embed_facebookGoogle Scholar
Jardina, A., & Traugott, M. (2019). The genesis of the birther rumor: Partisanship, racial attitudes, and political knowledge. The Journal of Race, Ethnicity, and Politics, 4(1), 6080. https://doi.org/10.1017/rep.2018.25Google Scholar
Johnson, H. M., & Seifert, C. M. (1994). Sources of the continued influence effect: When misinformation in memory affects later inferences. Journal of Experimental Psychology: Learning, Memory, and Cognition, 20(6), 14201436. https://doi.org/10.1037/0278-7393.20.6.1420Google Scholar
Jost, J. T., Glaser, J., Kruglanski, A. W., & Sulloway, F. J. (2003). Political conservatism as motivated social cognition. Psychological Bulletin, 129(3), 339375. https://doi.org/10.1037/0033-2909.129.3.339Google Scholar
Jost, J. T., van der Linden, S., Panagopoulos, C., & Hardin, C. D. (2018). Ideological asymmetries in conformity, desire for shared reality, and the spread of misinformation. Current Opinion in Psychology, 23, 7783. https://doi.org/10.1016/j.copsyc.2018.01.003Google Scholar
Jun, Y., Meng, R., & Johar, G. V. (2017). Perceived social presence reduces fact-checking. Proceedings of the National Academy of Sciences, 114(23), 59765981. https://doi.org/10.1073/pnas.1700175114Google Scholar
Kahan, D. M. (2013). Ideology, motivated reasoning, and cognitive reflection. Judgment and Decision Making, 8(4), 407424.Google Scholar
Kahan, D. M., Braman, D., Monahan, J., Callahan, L., & Peters, E. (2010). Cultural cognition and public policy: The case of outpatient commitment laws. Law and Human Behavior, 34(2), 118140. https://doi.org/10.1007/s10979-008-9174-4CrossRefGoogle ScholarPubMed
Kahan, D. M., Jenkins‐Smith, H., & Braman, D. (2011). Cultural cognition of scientific consensus. Journal of Risk Research, 14(2), 147174. https://doi.org/10.1080/13669877.2010.511246Google Scholar
Kasra, M., Shen, C., & O’Brien, J. (2016). Seeing is believing: Do people fail to identify fake images on the Web? Paper presented at AoIR 2016: The 17th Annual Conference of the Association of Internet Researchers, October 5–8, Berlin, Germany.Google Scholar
Keeley, B. L. (1999). Of conspiracy theories. The Journal of Philosophy, 96(3), 109126. https://doi.org/10.2307/2564659Google Scholar
Khanna, K., & Sood, G. (2018). Motivated responding in studies of factual learning. Political Behavior, 40(1), 79101. https://doi.org/10.1007/s11109-017-9395-7Google Scholar
Kuklinski, J. H., Quirk, P. J., Jerit, J., Schweider, D., & Rich, R. F. (2000). Misinformation and the currency of democratic citizenship. The Journal of Politics, 62(3), 790816.CrossRefGoogle Scholar
Kunda, Z. (1990). The case for motivated reasoning. Psychological Bulletin, 108(3), 480498.Google Scholar
Lazer, D. M. J., Baum, M. A., Benkler, Y. et al. (2018). The science of fake news. Science, 359(6380), 10941096. https://doi.org/10.1126/science.aao2998Google Scholar
Leman, P. J., & Cinnirella, M. (2013). Beliefs in conspiracy theories and the need for cognitive closure. Frontiers in Psychology, 4(378), 110. https://doi.org/10.3389/fpsyg.2013.00378Google Scholar
Lewandowsky, S., Ecker, U. K. H., Seifert, C. M., Schwarz, N., & Cook, J. (2012). Misinformation and its correction: Continued influence and successful debiasing. Psychological Science in the Public Interest, 13(3), 106131. https://doi.org/10.1177/1529100612451018Google Scholar
Lodge, M., & Taber, C. S. (2013). The Rationalizing Voter. Cambridge: Cambridge University Press.Google Scholar
Ma, Y., Dixon, G., & Hmielowski, J. D. (2019). Psychological reactance from reading basic facts on climate change: The role of prior views and political identification. Environmental Communication, 13(1), 7186. https://doi.org/10.1080/17524032.2018.1548369Google Scholar
Marchlewska, M., Cichocka, A., & Kossowska, M. (2018). Addicted to answers: Need for cognitive closure and the endorsement of conspiracy beliefs. European Journal of Social Psychology, 48(2), 109117. https://doi.org/10.1002/ejsp.2308CrossRefGoogle Scholar
Margolin, D. B., Hannak, A., & Weber, I. (2018). Political fact-checking on Twitter: When do corrections have an effect? Political Communication, 35(2), 196219. https://doi.org/10.1080/10584609.2017.1334018CrossRefGoogle Scholar
Marietta, M., Barker, D. C., & Bowser, T. (2015). Fact-checking polarized politics: Does the fact-check industry provide consistent guidance on disputed realities? The Forum, 13(4), 577596. https://doi.org/10.1515/for-2015-0040Google Scholar
McGinnies, E., & Ward, C. D. (1980). Better liked than right: Trustworthiness and expertise as factors in credibility. Personality and Social Psychology Bulletin, 6(3), 467472.Google Scholar
Meirick, P. C., & Bessarabova, E. (2016). Epistemic factors in selective exposure and political misperceptions on the right and left: Epistemic factors in news use and misperceptions. Analyses of Social Issues and Public Policy, 16(1), 3668. https://doi.org/10.1111/asap.12101Google Scholar
Mena, P. (2019). Cleaning up social media: The effect of warning labels on likelihood of sharing false news on Facebook. Policy & Internet. https://doi.org/10.1002/poi3.214Google Scholar
Messing, S., & Westwood, S. J. (2014). Selective exposure in the age of social media: Endorsements trump partisan source affiliation when selecting news online. Communication Research, 41(8), 10421063. https://doi.org/10.1177/0093650212466406Google Scholar
Miller, J. M., Saunders, K. L., & Farhart, C. E. (2016). Conspiracy endorsement as motivated reasoning: The moderating roles of political knowledge and trust. American Journal of Political Science, 60(4), 824844. https://doi.org/10.1111/ajps.12234Google Scholar
Mirandola, C., & Toffalini, E. (2016). Arousal – but not valence – reduces false memories at retrieval. PLoS ONE, 11(3). https://doi.org/10.1371/journal.pone.0148716Google Scholar
Moulding, R., Nix-Carnell, S., Schnabel, A. et al. (2016). Better the devil you know than a world you don’t? Intolerance of uncertainty and worldview explanations for belief in conspiracy theories. Personality and Individual Differences, 98, 345354. https://doi.org/10.1016/j.paid.2016.04.060Google Scholar
Nam, H. H., Jost, J. T., & Van Bavel, J. J. (2013). “Not for all the tea in China!” Political ideology and the avoidance of dissonance-arousing situations. PLoS ONE, 8(4), e59837. https://doi.org/10.1371/journal.pone.0059837Google Scholar
Nisbet, E. C., Cooper, K. E., & Garrett, R. K. (2015). The partisan brain: How dissonant science messages lead conservatives and liberals to (dis)trust science. The ANNALS of the American Academy of Political and Social Science, 658(1), 3666. https://doi.org/10.1177/0002716214555474Google Scholar
Nyhan, B. (2010). Why the “death panel” myth wouldn’t die: Misinformation in the health care reform debate. The Forum, 8(1). https://doi.org/10.2202/1540-8884.1354Google Scholar
Nyhan, B., Porter, E., Reifler, J., & Wood, T. J. (2019). Taking fact-checks literally but not seriously? The effects of journalistic fact-checking on factual beliefs and candidate favorability. Political Behavior. https://doi.org/10.1007/s11109-019-09528-xGoogle Scholar
Nyhan, B., & Reifler, J. (2010). When corrections fail: The persistence of political misperceptions. Political Behavior, 32(2), 303330. https://doi.org/10.1007/s11109-010-9112-2Google Scholar
Nyhan, B., & Reifler, J.(2015a). Displacing misinformation about events: An experimental test of causal corrections. Journal of Experimental Political Science, 2(1), 8193. https://doi.org/10.1017/XPS.2014.22CrossRefGoogle Scholar
Nyhan, B., & Reifler, J.(2015b). Does correcting myths about the flu vaccine work? An experimental evaluation of the effects of corrective information. Vaccine, 33(3), 459464. https://doi.org/10.1016/j.vaccine.2014.11.017Google Scholar
Nyhan, B., & Reifler, J.(2015c). Estimating fact-checking’s effects: Evidence from a long-term experiment during campaign 2014. Working Paper, American Press Institute.Google Scholar
Nyhan, B., & Reifler, J.(2019). The roles of information deficits and identity threat in the prevalence of misperceptions. Journal of Elections, Public Opinion, and Parties, 29(2), 222244. https://doi.org/10.1080/17457289.2018.1465061Google Scholar
Nyhan, B., Reifler, J., Richey, S., & Freed, G. L. (2014). Effective messages in vaccine promotion: A randomized trial. Pediatrics, 133(4), e835e842. https://doi.org/10.1542/peds.2013-2365Google Scholar
Nyhan, B., Reifler, J., & Ubel, P. A. (2013). The hazards of correcting myths about health care reform. Medical Care, 51(2), 127132. https://doi.org/10.1097/MLR.0b013e318279486bGoogle Scholar
Oliver, J. E., & Wood, T. J. (2014). Conspiracy theories and the paranoid style(s) of mass opinion. American Journal of Political Science, 58(4), 952966. https://doi.org/10.1111/ajps.12084Google Scholar
Pennycook, G., Bear, A., Collins, E. T., & Rand, D. G. (2020). The implied truth effect: Attaching warnings to a subset of fake news stories increases perceived accuracy of stories without warnings. Management Science. https://doi.org/10.1287/mnsc.2019.3478Google Scholar
Pennycook, G., Cannon, T. D., & Rand, D. G. (2018). Prior exposure increases perceived accuracy of fake news. Journal of Experimental Psychology: General, 147(12), 18651880. https://doi.org/10.1037/xge0000465Google Scholar
Pennycook, G., & Rand, D. G. (2018). Lazy, not biased: Susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition, 188, 3950. https://doi.org/10.1016/j.cognition.2018.06.011Google Scholar
Pennycook, G., & Rand, D. G.(2019). Fighting misinformation on social media using crowdsourced judgments of news source quality. Proceedings of the National Academy of Sciences, 116(7), 25212526. https://doi.org/10.1073/pnas.1806781116Google Scholar
(2020). Who falls for fake news? The roles of bullshit receptivity, overclaiming, familiarity, and analytic thinking. Journal of Personality, 88(2), 185200. https://doi.org/10.1111/jopy.12476Google Scholar
Peter, C., & Koch, T. (2016). When debunking scientific myths fails (and when it does not): The backfire effect in the context of journalistic coverage and immediate judgments as prevention strategy. Science Communication, 38(1), 325. https://doi.org/10.1177/1075547015613523Google Scholar
Pluviano, S., Watt, C., & Della Sala, S. (2017). Misinformation lingers in memory: Failure of three pro-vaccination strategies. PLoS ONE, 12(7), 115. https://doi.org/10.1371/journal.pone.0181640Google Scholar
Porter, E., Wood, T. J., & Bahador, B. (2019). Can presidential misinformation on climate change be corrected? Evidence from Internet and phone experiments. Research & Politics, 6(3). https://doi.org/10.1177/2053168019864784Google Scholar
Porter, E., Wood, T. J., & Kirby, D. (2018). Sex trafficking, Russian infiltration, birth certificates, and pedophilia: A survey experiment correcting fake news. Journal of Experimental Political Science, 5(2), 159164. https://doi.org/10.1017/XPS.2017.32Google Scholar
Rapp, D. N., & Salovich, N. A. (2018). Can’t we just disregard fake news? The consequences of exposure to inaccurate information. Policy Insights from the Behavioral and Brain Sciences, 5(2), 232239.Google Scholar
Redlawsk, D. P., Civettini, A. J. W., & Emmerson, K. M. (2010). The affective tipping point: Do motivated reasoners ever “get it”? Political Psychology, 31(4), 563593. https://doi.org/10.1111/j.1467-9221.2010.00772.xGoogle Scholar
Richards, Z., & Hewstone, M. (2001). Subtyping and subgrouping: Processes for the prevention and promotion of stereotype change. Personality and Social Psychology Review, 5(1), 5273. https://doi.org/10.1207/S15327957PSPR0501_4Google Scholar
Roozenbeek, J., & van der Linden, S. (2019a). Fake news game confers psychological resistance against online misinformation. Palgrave Communications, 5(1), 110. https://doi.org/10.1057/s41599-019-0279-9Google Scholar
Roozenbeek, J., & van der Linden, S.(2019b). The fake news game: Actively inoculating against the risk of misinformation. Journal of Risk Research, 22(5), 570580. https://doi.org/10.1080/13669877.2018.1443491Google Scholar
Schaffner, B. F., & Roche, C. (2017). Misinformation and motivated reasoning. Public Opinion Quarterly, 81(1), 86110. https://doi.org/10.1093/poq/nfw043Google Scholar
Schwarz, N., Newman, E., & Leach, W. (2016). Making the truth stick & the myths fade: Lessons from cognitive psychology. Behavioral Science & Policy, 2(1), 8595. https://doi.org/10.1353/bsp.2016.0009Google Scholar
Schwarz, N., Sanna, L. J., Skurnik, I., & Yoon, C. (2007). Metacognitive experiences and the intricacies of setting people straight: Implications for debiasing and public information campaigns. Advances in Experimental Social Psychology, 39, 127161. https://doi.org/10.1016/S0065-2601(06)39003-XGoogle Scholar
Sensenig, J., & Brehm, J. W. (1968). Attitude change from an implied threat to attitudinal freedom. Journal of Personality and Social Psychology, 8(4, Pt.1), 324330. https://doi.org/10.1037/h0021241Google Scholar
Shearer, E., & Matsa, K. E. (2018). News Use Across Social Media Platforms 2018. Pew Research Center report. https://www.journalism.org/2018/09/10/news-use-across-social-media-platforms-2018/Google Scholar
Shen, C., Kasra, M., Pan, W., Bassett, G. A., Malloch, Y., & O’Brien, J. F. (2019). Fake images: The effects of source, intermediary, and digital media literacy on contextual assessment of image credibility online. New Media & Society, 21(2), 438463.Google Scholar
Shin, J., & Thorson, K. (2017). Partisan selective sharing: The biased diffusion of fact-checking messages on social media: Sharing fact-checking messages on social media. Journal of Communication, 67(2), 233255. https://doi.org/10.1111/jcom.12284Google Scholar
Shu, S. B., & Carlson, K. A. (2014). When three charms but four alarms: Identifying the optimal number of claims in persuasion settings. Journal of Marketing, 78(1), 127139. https://doi.org/10.1509/jm.11.0504Google Scholar
Skurnik, I., Yoon, C., & Schwarz, N. (2007). Myths & Facts” about the flu: Health education campaigns can reduce vaccination intentions. http://webuser.bus.umich.edu/yoonc/research/Papers/Skurnik%5FYoon%5FSchwarz%5F2005%5FMyths%5FFacts%5FFlu%5FHealth%5FEducation%5FCampaigns%5FJAMA.pdfGoogle Scholar
Stahl, B. C. (2006). On the difference or equality of information, misinformation, and disinformation: A critical research perspective. Informing Science: The International Journal of an Emerging Transdiscipline, 9, 8396. https://doi.org/10.28945/473Google Scholar
Swami, V., Voracek, M., Stieger, S., Tran, U. S., & Furnham, A. (2014). Analytic thinking reduces belief in conspiracy theories. Cognition, 133(3), 572585. https://doi.org/10.1016/j.cognition.2014.08.006Google Scholar
Swire, B., Berinsky, A. J., Lewandowsky, S., & Ecker, U. K. H. (2017). Processing political misinformation: Comprehending the Trump phenomenon. Royal Society Open Science, 4(3), 160802. https://doi.org/10.1098/rsos.160802Google Scholar
Swire, B., & Ecker, U. (2018). Misinformation and its correction: Cognitive mechanisms and recommendations for mass communication. In Southwell, B., Thorson, E. A., & Sheble, L. (Eds.), Misinformation and Mass Audiences (pp. 195211). Austin, TX: University of Texas Press. https://utpress.utexas.edu/books/southwell-thorson-sheble-misinformation-and-mass-audiencesGoogle Scholar
Swire, B., Ecker, U. K. H., & Lewandowsky, S. (2017). The role of familiarity in correcting inaccurate information. Journal of Experimental Psychology: Learning, Memory, and Cognition, 43(12), 19481961. https://doi.org/10.1037/xlm0000422Google Scholar
Tandoc, E. C., Lim, Z. W., & Ling, R. (2018). Defining “fake news”: A typology of scholarly definitions. Digital Journalism, 6(2), 137153. https://doi.org/10.1080/21670811.2017.1360143Google Scholar
Tappin, B. M., Pennycook, G., & Rand, D. (2018). Rethinking the link between cognitive sophistication and identity-protective bias in political belief formation. PsyArXiv.org. https://doi.org/10.31234/osf.io/yuzfjGoogle Scholar
Thorson, E.(2016) Belief echoes: The persistent effects of corrected misinformation. Political Communication, 33(3), 460480. https://doi.org/10.1080/10584609.2015.1102187Google Scholar
Trevors, G. J., Muis, K. R., Pekrun, R., Sinatra, G. M., & Winne, P. H. (2016). Identity and epistemic emotions during knowledge revision: A potential account for the backfire effect. Discourse Processes, 53(5–6), 339370. https://doi.org/10.1080/0163853X.2015.1136507Google Scholar
Tucker, J., Guess, A., Barbera, P. et al. (2018). Social media, political polarization, and political disinformation: A review of the scientific literature. SSRN. https://doi.org/10.2139/ssrn.3144139Google Scholar
Valenzuela, S., Halpern, D., Katz, J. E., & Miranda, J. P. (2019). The paradox of participation versus misinformation: Social media, political engagement, and the spread of misinformation. Digital Journalism, 7(6), 802823. https://doi.org/10.1080/21670811.2019.1623701Google Scholar
van der Linden, S. L., Leiserowitz, A. A., Feinberg, G. D., & Maibach, E. W. (2015). The scientific consensus on climate change as a gateway belief: Experimental evidence. PLoS ONE, 10(2), 18. https://doi.org/10.1371/journal.pone.0118489Google Scholar
van der Linden, S. L., Leiserowitz, A., Rosenthal, S., & Maibach, E. (2017). Inoculating the public against misinformation about climate change. Global Challenges, 1(2), 17. https://doi.org/10.1002/gch2.201600008Google Scholar
Vosoughi, S., Roy, D., & Aral, S. (2018). The spread of true and false news online. Science, 359(6380), 11461151. https://doi.org/10.1126/science.aap9559Google Scholar
Vraga, E. K. (2019). What can I do? How to use social media to improve democratic society. Political Communication, 36(2), 315323. https://doi.org/10.1080/10584609.2019.1610620Google Scholar
Vraga, E. K., & Bode, L. (2017). Using expert sources to correct health misinformation in social media. Science Communication, 39(5), 621645. https://doi.org/10.1177/1075547017731776Google Scholar
Vraga, E. K., & Bode, L.(2018). I do not believe you: How providing a source corrects health misperceptions across social media platforms. Information, Communication & Society, 21(10), 13371353. https://doi.org/10.1080/1369118X.2017.1313883CrossRefGoogle Scholar
Walter, N., & Murphy, S. T. (2018). How to unring the bell: A meta-analytic approach to correction of misinformation. Communication Monographs, 85(3), 423441. https://doi.org/10.1080/03637751.2018.1467564Google Scholar
Walter, N., & Tukachinsky, R. (2019). A meta-analytic examination of the continued influence of misinformation in the face of correction: How powerful is it, why does it happen, and how to stop it? Communication Research, 47(2), 155177. https://doi.org/10.1177/0093650219854600Google Scholar
Wardle, C. (2018). Information Disorder: The Essential Glossary. Harvard Kennedy School Shorenstein Center on Media, Politics, and Public Policy. https://firstdraftnews.org/wp-content/uploads/2018/07/infoDisorder_glossary.pdfGoogle Scholar
Wardle, C., & Derakhshan, H. (2017). Information Disorder: Toward an Interdisciplinary Framework for Research and Policymaking. Council of Europe Report No. DGI(2017)09. https://rm.coe.int/information-disorder-toward-an-interdisciplinary-framework-for-researc/168076277cGoogle Scholar
Weeks, B. E. (2015). Emotions, partisanship, and misperceptions: How anger and anxiety moderate the effect of partisan bias on susceptibility to political misinformation. Journal of Communication, 65(4), 699719. https://doi.org/10.1111/jcom.12164Google Scholar
Weeks, B. E., & Garrett, R. K. (2014). Electoral consequences of political rumors: Motivated reasoning, candidate rumors, and vote choice during the 2008 U.S. presidential election. International Journal of Public Opinion Research, 26(4), 401422. https://doi.org/10.1093/ijpor/edu005Google Scholar
Wilkes, A. L., & Leatherbarrow, M. (1988). Editing episodic memory following the identification of error. The Quarterly Journal of Experimental Psychology Section A, 40(2), 361387. https://doi.org/10.1080/02724988843000168Google Scholar
Wintersieck, A., Fridkin, K., & Kenney, P. (2018). The message matters: The influence of fact-checking on evaluations of political messages. Journal of Political Marketing, 128. https://doi.org/10.1080/15377857.2018.1457591Google Scholar
Wintersieck, A. L. (2017). Debating the truth: The impact of fact-checking during electoral debates. American Politics Research, 45(2), 304331. https://doi.org/10.1177/1532673X16686555Google Scholar
Wood, T., & Porter, E. (2019). The elusive backfire effect: Mass attitudes’ steadfast factual adherence. Political Behavior, 41(1), 135163. https://doi.org/10.1007/s11109-018-9443-yGoogle Scholar
Young, D. G., Jamieson, K. H., Poulsen, S., & Goldring, A. (2018). Fact-checking effectiveness as a function of format and tone: Evaluating FactCheck.org and FlackCheck.org. Journalism & Mass Communication Quarterly, 95(1), 4975. https://doi.org/10.1177/1077699017710453Google Scholar
Zhou, J. (2016). Boomerangs versus javelins: How polarization constrains communication on climate change. Environmental Politics, 25(5), 788811. https://doi.org/10.1080/09644016.2016.1166602Google Scholar
Zollo, F., Bessi, A., Del Vicario, M. et al. (2017). Debunking in a world of tribes. PLoS ONE, 12(7), e0181821. https://doi.org/10.1371/journal.pone.0181821Google Scholar
Zubiaga, A., Liakata, M., Procter, R., Wong Sak Hoi, G., & Tolmie, P. (2016). Analysing how people orient to and spread rumours in social media by looking at conversational threads. PLoS ONE, 11(3), 129. https://doi.org/10.1371/journal.pone.0150989Google Scholar

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×