Skip to main content
Log in

Real Solutions for Fake News? Measuring the Effectiveness of General Warnings and Fact-Check Tags in Reducing Belief in False Stories on Social Media

  • Original Paper
  • Published:
Political Behavior Aims and scope Submit manuscript

Abstract

Social media has increasingly enabled “fake news” to circulate widely, most notably during the 2016 U.S. presidential campaign. These intentionally false or misleading stories threaten the democratic goal of a well-informed electorate. This study evaluates the effectiveness of strategies that could be used by Facebook and other social media to counter false stories. Results from a pre-registered experiment indicate that false headlines are perceived as less accurate when people receive a general warning about misleading information on social media or when specific headlines are accompanied by a “Disputed” or “Rated false” tag. Though the magnitudes of these effects are relatively modest, they generally do not vary by whether headlines were congenial to respondents’ political views. In addition, we find that adding a “Rated false” tag to an article headline lowers its perceived accuracy more than adding a “Disputed” tag (Facebook’s original approach) relative to a control condition. Finally, though exposure to the “Disputed” or “Rated false” tags did not affect the perceived accuracy of unlabeled false or true headlines, exposure to a general warning decreased belief in the accuracy of true headlines, suggesting the need for further research into how to most effectively counter false news without distorting belief in true information.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Notes

  1. “Fake news” has many definitions and is frequently used in imprecise or confusing ways. Moreover, the debate over the meaning of the term and related concepts raises epistemological issues that are beyond the scope of this paper (e.g., speaker intent; see Wardle and Derakhshan 2017). We therefore employ “false news” as an alternative term throughout this paper, which define as described above (“factually dubious content that imitates the format of journalism but is produced with no regard for accuracy or fairness”; see Lazer et al. 2018). This approach is consistent with the practices of various news and social media sources (e.g., Oremus 2017) and is intended to avoid unnecessary confusion.

  2. Pennycook and Rand (2017), which we had not seen at the time of pre-registration, also considers this question.

  3. We pre-registered an additional research question about the effects of exposure to a general warning and/or to a “Disputed” or “Rated false” tag on respondents’ self-reported likelihood of “liking” and sharing the headlines on Facebook. The results of this analysis are presented in Online Appendix B.

  4. A minority of studies conclude that MTurk samples are not externally valid (e.g., Krupnikov and Levine 2014). For example, participants on MTurk tend to skew liberal and young. Moreover, the underrepresentation of conservatives and older participants may suggest that these participants differ from other conservatives or older individuals in the general population. However, numerous studies find that experimental treatment effect estimates typically generalize from MTurk to national probability samples, suggesting these problems are rare (e.g., Berinsky et al. 2012; Coppock 2016; Horton et al. 2011; Mullinix et al. 2015). Finally, our MTurk sample is externally valid in the sense that it is made up disproportionately of frequent users of the Internet—precisely the group who may be most likely to encounter false news (Pennycook and Rand 2018a). We thus conclude that respondents from MTurk constitute a valid sample for testing our hypotheses, though replication on representative samples would of course be desirable.

  5. The pilot study tested the effects of “Disputed” and “Rated false” tags only on perceived accuracy and likelihood of liking/sharing for six false news headlines. The results of this study were similar to our main analysis, and are available upon request.

  6. As in most studies, we cannot know how much false news respondents were exposed to during the 2016 presidential election and its aftermath (e.g., Allcott and Gentzkow 2017). While it would be useful to measure this quantity, our main interest is the effect of warnings and tags on belief accuracy when they encounter false news. In addition, the auxiliary measure of misperception belief mentioned above does allow us to test whether individuals who are susceptible to believing false news respond differently to warnings and tags than those who are not. We find no consistent evidence of such heterogeneity in exploratory analyses reported in Online Appendix C. Scholars should collect data on individuals’ exposure to false news and explore treatment effect heterogeneity by this variable directly in future research.

  7. A possible concern is that asking respondents to rate political statements for accuracy could have primed them to be particularly alert to clues that the treatment articles could be deceptive in nature. However, Pennycook et al. (2017) and Pennycook and Rand (2017) did not ask respondents to rate any statements for accuracy before their experiment and also found that tagged false news headlines were rated as less accurate that untagged ones, suggesting that the tags reduce the perceived accuracy of false headlines independently of a possible priming effect.

  8. Some of these articles were originally used in Pennycook et al. (2017), which examined the effect of prior exposure to false news headlines on the perceived accuracy of false news. Others were taken from Silverman (2016), a compilation of the most widely shared false news articles during the 2016 election. The original sources of the false news articles were dubious websites that had intentionally created them for profit.

  9. The true headlines that were tested were taken from actual mainstream news sources and were not intended to be explicitly pro- or anti-Trump, though respondent interpretations of them may differ.

  10. A potential concern is that highly attentive MTurk respondents saw these accuracy questions as an attention check rather than a measure of sincere belief and responded accordingly. However, previous research has found that the effect of corrections to misinformation were almost identical among samples of MTurk workers and Morning Consult poll respondents (Nyhan et al. 2017) and provides limited and inconsistent evidence of demand effects in survey experiments (Mummolo and Peterson 2018).

  11. de Leeuw et al. (2015) find that excluding “don’t know” options but allowing respondents to skip questions in online surveys (as we did) reduces missing data and increases reliability in online surveys relative to the inclusion of a “don’t know” option, and suggest using “don’t know” options only when there is a theoretical reason to do so. We also opt to exclude the “don’t know” option to encourage compatibility between our study and others in the field that examine belief in false news and other forms of political misinformation (e.g., Pennycook et al. 2017; Pennycook and Rand 2017).

  12. Our preregistration did not offer hypotheses about the correlates of false news belief, but see Pennycook and Rand (2018b), which finds that individuals who have a tendency to ascribe profundity to randomly generated sentences and who overstate their level of knowledge are more likely to perceive false news as accurate. Those who engage in analytic thinking are less susceptible.

  13. All results are virtually identical when estimated using ordered probit instead. See Online Appendix C.

  14. We do not include respondent fixed effects, which were incorrectly specified in the pre-registration (they cannot be estimated due to multicollinearity). However, we show in Online Appendix C that our primary results are consistent when estimated in a model that includes random effects by respondent.

  15. The estimates reported here refer to the effects of each treatment alone independent of any moderators, with all other manipulations set at 0. We estimate models that include interactive terms below.

  16. The effects on perceived accuracy reported in Tables 35 are consistent when non-Facebook users are excluded from the sample in exploratory analyses (see Online Appendix C).

  17. A typo in the pre-registration statement to this effect instead mistakenly stated we would exclude “pure independents.” The results below again exclude pure controls but equivalent results including those respondents are provided in Online Appendix C. We do not include respondents with no opinion of Trump in that model because there were so few (n = 4).

  18. Pennycook and Rand (2018a) similarly find that “the correlation between CRT [Cognitive Reflection Test scores] and perceived accuracy is unrelated to how closely the headline aligns with the participant’s ideology... Our findings therefore suggest that susceptibility to fake news is driven more by lazy thinking than it is by partisan bias per se.” Similarly, Porter et al. (2018) find minimal differences between ideological groups in their willingness to accept false news headlines.

  19. We conducted an additional exploratory analysis to test whether the effects of political congeniality were altered by a participant’s political knowledge. Consistent with previous research, we found that high political knowledge was associated with a lower belief in false news stories regardless of the article’s slant. However, we did not find convincing evidence that high political knowledge meaningfully changed a specific warning’s effect on belief in false news headlines. Results for this exploratory analysis are included in Online Appendix C (Table C16).

  20. Headlines viewed by respondents in the “Disputed” or “Rated false” conditions before exposure to the first tag are also excluded (spillover is impossible for participants who are not yet treated).

  21. This difference in effect size could be partially attributable to respondents being aware that their ability to discern true from false headlines was under scrutiny, since they had previously been asked to rate political statements as true or false at the beginning of our survey.

References

Download references

Acknowledgements

We thank the Dartmouth College Office of Undergraduate Research for generous funding support. We are also grateful to Ro’ee Levy and David Rand for helpful comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Brendan Nyhan.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary material 1 (PDF 16849 KB)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Clayton, K., Blair, S., Busam, J.A. et al. Real Solutions for Fake News? Measuring the Effectiveness of General Warnings and Fact-Check Tags in Reducing Belief in False Stories on Social Media. Polit Behav 42, 1073–1095 (2020). https://doi.org/10.1007/s11109-019-09533-0

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11109-019-09533-0

Keywords

Navigation