skip to main content
10.1145/3184558.3188728acmotherconferencesArticle/Chapter ViewAbstractPublication PageswwwConference Proceedingsconference-collections
research-article
Free Access

Misleading or Falsification: Inferring Deceptive Strategies and Types in Online News and Social Media

Published:23 April 2018Publication History

ABSTRACT

Deceptive information in online news and social media has had dramatic effect on our society in recent years. This study is the first to gain deeper insights into writers' intent behind digital misinformation by analyzing psycholinguistic signals: moral foundations and connotations extracted from different types of deceptive news ranging from strategic disinformation to propaganda and hoaxes. To ensure consistency of our findings and generalizability across domains, we experiment with data from: (1) confirmed cases of disinformation in news summaries, (2) propaganda, hoax, and disinformation news pages, and (3) social media news. We first contrast lexical markers of biased language, syntactic and stylistic signals, and connotations across deceptive news types including disinformation, propaganda, and hoaxes, and deceptive strategies including misleading or falsification. We then incorporate these insights to build machine learning and deep learning predictive models to infer deception strategies and deceptive news types. Our experimental results demonstrate that unlike earlier work on deception detection, content combined with biased language markers, moral foundations, and connotations leads to better predictive performance of deception strategies compared to syntactic and stylistic signals (as reported in earlier work on deceptive reviews). Falsification strategy is easier to identify than misleading strategy. Disinformation is more difficult to predict than to propaganda or hoaxes. Deceptive news types (disinformation, propaganda, and hoaxes), unlike deceptive strategies (falsification and misleading), are more salient, and thus easier to identify in tweets than in news reports. Finally, our novel connotation analysis across deception types provides deeper understanding of writers' perspectives and therefore reveals the intentions behind digital misinformation.

References

  1. Darren Scott Appling, Erica J. Briscoe, and Clayton J. Hutto. 2015. Discriminative Models for Predicting Deception Strategies Proceedings of WWW. 947--952. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Eytan Bakshy, Jake M Hofman, Winter A Mason, and Duncan J Watts. 2011. Everyone's an influencer: quantifying influence on twitter Proceedings of WSDM. 65--74. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Erica J Briscoe, D Scott Appling, and Heather Hayes. 2014. Cues to deception in social media communications. International Conference on System Sciences. 1435--1443. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. David B Buller, Judee K Burgoon, JA Daly, and JM Wiemann. 1994. Deception: Strategic and nonstrategic communication. Strategic interpersonal communication (1994), 191--223.Google ScholarGoogle Scholar
  5. Dallas Card, Amber E Boydstun, Justin H Gross, Philip Resnik, and Noah A Smith. 2015. The media frames corpus: Annotations of frames across issues Proceedings of ACL. 438--444.Google ScholarGoogle Scholar
  6. Carlos Castillo, Marcelo Mendoza, and Barbara Poblete. 2011. Information credibility on Twitter. In Proceedings of WWW. 675--684. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Meeyoung Cha, Hamed Haddadi, Fabricio Benevenuto, and P Krishna Gummadi. 2010. Measuring user influence in Twitter: The million follower fallacy. Proceedings of ICWSM, Vol. 10, 10--17 (2010), 30.Google ScholarGoogle ScholarCross RefCross Ref
  8. Justin Cheng, Cristian Danescu-Niculescu-Mizil, Jure Leskovec, and Michael Bernstein. 2017. Anyone Can Become a Troll. American Scientist, Vol. 105, 3 (2017), 152.Google ScholarGoogle ScholarCross RefCross Ref
  9. Eunsol Choi, Chenhao Tan, Lillian Lee, Cristian Danescu-Niculescu-Mizil, and Jennifer Spindel. 2012. Hedge detection as a lens on framing in the GMO debates: A position paper Proceedings of the Workshop on Extra-Propositional Aspects of Meaning in Computational Linguistics. 70--79. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Giovanni Luca Ciampaglia, Prashant Shiralkar, Luis M Rocha, Johan Bollen, Filippo Menczer, and Alessandro Flammini. 2015. Computational fact checking from knowledge networks. PloS one, Vol. 10, 6 (2015), e0128193.Google ScholarGoogle ScholarCross RefCross Ref
  11. Niall J Conroy, Victoria L Rubin, and Yimin Chen. 2015. Automatic deception detection: methods for finding fake news. Proceedings of the Association for Information Science and Technology, Vol. 52, 1 (2015), 1--4. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Bella M DePaulo, James J Lindsay, Brian E Malone, Laura Muhlenbruck, Kelly Charlton, and Harris Cooper. 2003. Cues to deception. Psychological bulletin Vol. 129, 1 (2003), 74.Google ScholarGoogle Scholar
  13. Paul Ekman and Wallace V Friesen. 1969. Nonverbal leakage and clues to deception. Psychiatry, Vol. 32, 1 (1969), 88--106.Google ScholarGoogle ScholarCross RefCross Ref
  14. Song Feng, Ritwik Banerjee, and Yejin Choi. 2012 a. Syntactic stylometry for deception detection. In Proceedings of ACL. 171--175. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Song Feng, Longfei Xing, Anupam Gogar, and Yejin Choi. 2012 b. Distributional Footprints of Deceptive Product Reviews. Proceedings of ICWSM Vol. 12 (2012), 98--105.Google ScholarGoogle Scholar
  16. Emilio Ferrara. 2017. Disinformation and social bot operations in the run up to the 2017 French presidential election. (2017).Google ScholarGoogle Scholar
  17. Eileen Fitzpatrick and Joan Bachenko. 2012. Building a data collection for deception research. Proceedings of the Workshop on Computational Approaches to Deception Detection. 31--38. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Fortune. 2016. Fake Bloomberg news report drives Twitter stock up 8%. http://fortune.com/2015/07/14/fake-twitter-bloomberg-report/. (2016). Accessed: 2016--12--12.Google ScholarGoogle Scholar
  19. Maria Glenski, Ellyn Ayton, Dustin Arendt, and Svitlana Volkova. 2017. Fishing for Clickbaits in Social Images and Texts with Linguistically-Infused Neural Network Models. arXiv preprint arXiv:1710.06390 (2017).Google ScholarGoogle Scholar
  20. Jesse Graham, Jonathan Haidt, and Brian A Nosek. 2009. Liberals and conservatives rely on different sets of moral foundations. Journal of Personality and Social Psychology, Vol. 96, 5 (2009), 1029.Google ScholarGoogle ScholarCross RefCross Ref
  21. Guardian. 2015. Woman dies after taking "diet pills" bought over internet. https://www.theguardian.com/society/2015/apr/21/woman-dies-after-taking-diet-pills-bought-over-internet. (2015).Google ScholarGoogle Scholar
  22. Jonathan Haidt and Jesse Graham. 2007. When morality opposes justice: Conservatives have moral intuitions that liberals may not recognize. Social Justice Research Vol. 20, 1 (2007), 98--116.Google ScholarGoogle ScholarCross RefCross Ref
  23. Jonathan Haidt and Craig Joseph. 2004. Intuitive ethics: How innately prepared intuitions generate culturally variable virtues. Daedalus, Vol. 133, 4 (2004), 55--66.Google ScholarGoogle ScholarCross RefCross Ref
  24. Joan B. Hooper. 1975. On assertive predicates. In Syntax and Semantics, bibfieldeditorJ. Kimball (Ed.), Vol. Vol. 4. 91--124.Google ScholarGoogle Scholar
  25. Lee Howell. 2013. Digital wildfires in a hyperconnected world. WEF Report (2013).Google ScholarGoogle Scholar
  26. Ken Hyland. 2005. Metadiscourse. Wiley Online Library.Google ScholarGoogle Scholar
  27. Fang Jin, Wei Wang, Liang Zhao, Edward Dougherty, Yang Cao, Chang-Tien Lu, and Naren Ramakrishnan. 2014. Misinformation propagation in the age of Twitter. Computer, Vol. 47, 12 (2014), 90--94. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Lauri Karttunen. 1971. Implicative verbs. Language (1971), 340--358.Google ScholarGoogle Scholar
  29. Paul Kiparsky and Carol Kiparsky. 1968. Fact. Indiana University.Google ScholarGoogle Scholar
  30. Srijan Kumar, Robert West, and Jure Leskovec. 2016. Disinformation on the Web: Impact, Characteristics, and Detection of Wikipedia Hoaxes Proceedings of WWW. 591--602. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Kenton Lee, Yoav Artzi, Yejin Choi, and Luke Zettlemoyer. 2015. Event Detection and Factuality Assessment with Non-Expert Supervision. Proceedings of EMNLP (2015).Google ScholarGoogle ScholarCross RefCross Ref
  32. Newton Lee. 2014. Misinformation and Disinformation. Facebook Nation. Springer, 169--188.Google ScholarGoogle Scholar
  33. Kristina Lerman and Rumi Ghosh. 2010. Information contagion: An empirical study of the spread of news on Digg and Twitter social networks. Proceedings of ICWSM (2010), 90--97.Google ScholarGoogle ScholarCross RefCross Ref
  34. Cristian Lumezanu, Nick Feamster, and Hans Klein. 2012. # bias: Measuring the tweeting behavior of propagandists. (2012).Google ScholarGoogle Scholar
  35. Feamster Lumezanu and H Klein. 2012. Measuring the tweeting behavior of propagandists. Proceedings of ICWSM.Google ScholarGoogle Scholar
  36. Rada Mihalcea and Carlo Strapparava. 2009. The lie detector: Explorations in the automatic recognition of deceptive language Proceedings of the ACL-IJCNLP. 309--312. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Tanushree Mitra, Graham P Wright, and Eric Gilbert. 2017. A parsimonious language model of social media credibility across disparate events Proceedings of CSCW. 126--145. Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Bjarke Mønsted, Piotr Sapie.zy'nski, Emilio Ferrara, and Sune Lehmann. 2017. Evidence of complex contagion of information in social media: An experiment using Twitter bots. PLOS ONE, Vol. 12, 9 (09. 2017), 1--12.Google ScholarGoogle ScholarCross RefCross Ref
  39. Matthew L Newman, James W Pennebaker, Diane S Berry, and Jane M Richards. 2003. Lying words: Predicting deception from linguistic styles. Personality and social psychology bulletin Vol. 29, 5 (2003), 665--675.Google ScholarGoogle Scholar
  40. NYTimes. 2016. Google and Facebook Take Aim at Fake News Sites. http://www.nytimes.com/2016/11/15/technology/google-will-ban-websites-that-host-fake-news-from-using-its-ad-service.html. (2016).Google ScholarGoogle Scholar
  41. Myle Ott, Yejin Choi, Claire Cardie, and Jeffrey T Hancock. 2011. Finding deceptive opinion spam by any stretch of the imagination Proceedings of ACL-HLT. 309--319. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research Vol. 12 (2011), 2825--2830. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. James W Pennebaker, Martha E Francis, and Roger J Booth. 2001. Linguistic inquiry and word count: LIWC 2001. Mahway: Lawrence Erlbaum Associates Vol. 71 (2001), 2001.Google ScholarGoogle Scholar
  44. Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In Proceedings EMNLP. 1532--1543.Google ScholarGoogle ScholarCross RefCross Ref
  45. Verónica Pérez-Rosas, Bennett Kleinberg, Alexandra Lefevre, and Rada Mihalcea. 2017. Automatic Detection of Fake News. arXiv preprint arXiv:1708.07104 (2017).Google ScholarGoogle Scholar
  46. Verónica Pérez-Rosas and Rada Mihalcea. 2015. Experiments in Open Domain Deception Detection. Proceedings of EMNLP (2015), 1120--1125.Google ScholarGoogle ScholarCross RefCross Ref
  47. Slav Petrov. 2016. Announcing SyntaxNet: The World's Most Accurate Parser Goes Open Source. Google Research Blog, May Vol. 12 (2016), 2016.Google ScholarGoogle Scholar
  48. Vahed Qazvinian, Emily Rosengren, Dragomir R Radev, and Qiaozhu Mei. 2011. Rumor has it: Identifying misinformation in microblogs Proceedings of EMNLP. 1589--1599. Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. Hannah Rashkin, Eric Bell, Yejin Choi, and Svitlana Volkova. 2017 a. Multilingual Connotation Frames: A Case Study on Social Media for Targeted Sentiment Analysis and Forecast. In Proceedings of ACL, Vol. Vol. 2. 459--464.Google ScholarGoogle Scholar
  50. Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, and Yejin Choi. 2017 b. Truth of varying shades: Analyzing language in fake news and political fact-checking Proceedings of EMNLP. 2921--2927.Google ScholarGoogle Scholar
  51. Hannah Rashkin, Sameer Singh, and Yejin Choi. 2016. Connotation Frames: A Data-Driven Investigation. Proceedings of ACL. 311--321.Google ScholarGoogle ScholarCross RefCross Ref
  52. Marta Recasens, Cristian Danescu-Niculescu-Mizil, and Dan Jurafsky. 2013. Linguistic Models for Analyzing and Detecting Biased Language Proceedings of ACL. 1650--1659.Google ScholarGoogle Scholar
  53. Victoria L Rubin. 2010. On deception and deception detection: Content analysis of computer-mediated stated beliefs. Proceedings of the American Society for Information Science and Technology, Vol. 47, 1 (2010), 1--10. Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. Victoria L Rubin, Yimin Chen, and Niall J Conroy. 2015. Deception detection for news: three types of fakes. Proceedings of the Association for Information Science and Technology, Vol. 52, 1 (2015), 1--4. Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. Victoria L Rubin, Niall J Conroy, Yimin Chen, and Sarah Cornwell. 2016. Fake News or Truth Using Satirical Cues to Detect Potentially Misleading News Proceedings of NAACL-HLT. 7--17.Google ScholarGoogle Scholar
  56. Kate Starbird. 2017. Examining the Alternative Media Ecosystem Through the Production of Alternative Narratives of Mass Shooting Events on Twitter. Proceedings of ICWSM. 230--239.Google ScholarGoogle ScholarCross RefCross Ref
  57. Bruno Takahashi, Edson C Tandoc, and Christine Carmichael. 2015. Communicating on Twitter during a disaster: An analysis of tweets during Typhoon Haiyan in the Philippines. Computers in Human Behavior Vol. 50 (2015), 392--398. Google ScholarGoogle ScholarDigital LibraryDigital Library
  58. Marcella Tambuscio, Giancarlo Ruffo, Alessandro Flammini, and Filippo Menczer. 2015. Fact-checking effect on viral hoaxes: A model of misinformation spread in social networks Proceedings of WWW. 977--982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  59. Michail Tsikerdekis and Sherali Zeadally. 2014. Online deception in social media. Commun. ACM Vol. 57, 9 (2014), 72--80. Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. Onur Varol, Emilio Ferrara, Clayton A Davis, Filippo Menczer, and Alessandro Flammini. 2017. Online human-bot interactions: Detection, estimation, and characterization Proceedings of ICWSM. 280--289.Google ScholarGoogle Scholar
  61. Svitlana Volkova, Kyle Shaffer, Jin Yean Jang, and Nathan Hodas. 2017. Separating Facts from Fiction: Linguistic Models to Classify Suspicious and Trusted News Posts on Twitter. In Proceedings of ACL. 647--653.Google ScholarGoogle ScholarCross RefCross Ref
  62. William Yang Wang. 2017. "Liar, Liar Pants on Fire": A New Benchmark Dataset for Fake News Detection Proceedings of ACL. 422--426.Google ScholarGoogle Scholar
  63. Miron Zuckerman, Richard Koestner, and Robert Driver. 1981. Beliefs about cues associated with deception. Journal of Nonverbal Behavior Vol. 6, 2 (1981), 105--114.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Misleading or Falsification: Inferring Deceptive Strategies and Types in Online News and Social Media

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        WWW '18: Companion Proceedings of the The Web Conference 2018
        April 2018
        2023 pages
        ISBN:9781450356404

        Copyright © 2018 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        International World Wide Web Conferences Steering Committee

        Republic and Canton of Geneva, Switzerland

        Publication History

        • Published: 23 April 2018

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate1,899of8,196submissions,23%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format