ABSTRACT
Deceptive information in online news and social media has had dramatic effect on our society in recent years. This study is the first to gain deeper insights into writers' intent behind digital misinformation by analyzing psycholinguistic signals: moral foundations and connotations extracted from different types of deceptive news ranging from strategic disinformation to propaganda and hoaxes. To ensure consistency of our findings and generalizability across domains, we experiment with data from: (1) confirmed cases of disinformation in news summaries, (2) propaganda, hoax, and disinformation news pages, and (3) social media news. We first contrast lexical markers of biased language, syntactic and stylistic signals, and connotations across deceptive news types including disinformation, propaganda, and hoaxes, and deceptive strategies including misleading or falsification. We then incorporate these insights to build machine learning and deep learning predictive models to infer deception strategies and deceptive news types. Our experimental results demonstrate that unlike earlier work on deception detection, content combined with biased language markers, moral foundations, and connotations leads to better predictive performance of deception strategies compared to syntactic and stylistic signals (as reported in earlier work on deceptive reviews). Falsification strategy is easier to identify than misleading strategy. Disinformation is more difficult to predict than to propaganda or hoaxes. Deceptive news types (disinformation, propaganda, and hoaxes), unlike deceptive strategies (falsification and misleading), are more salient, and thus easier to identify in tweets than in news reports. Finally, our novel connotation analysis across deception types provides deeper understanding of writers' perspectives and therefore reveals the intentions behind digital misinformation.
- Darren Scott Appling, Erica J. Briscoe, and Clayton J. Hutto. 2015. Discriminative Models for Predicting Deception Strategies Proceedings of WWW. 947--952. Google ScholarDigital Library
- Eytan Bakshy, Jake M Hofman, Winter A Mason, and Duncan J Watts. 2011. Everyone's an influencer: quantifying influence on twitter Proceedings of WSDM. 65--74. Google ScholarDigital Library
- Erica J Briscoe, D Scott Appling, and Heather Hayes. 2014. Cues to deception in social media communications. International Conference on System Sciences. 1435--1443. Google ScholarDigital Library
- David B Buller, Judee K Burgoon, JA Daly, and JM Wiemann. 1994. Deception: Strategic and nonstrategic communication. Strategic interpersonal communication (1994), 191--223.Google Scholar
- Dallas Card, Amber E Boydstun, Justin H Gross, Philip Resnik, and Noah A Smith. 2015. The media frames corpus: Annotations of frames across issues Proceedings of ACL. 438--444.Google Scholar
- Carlos Castillo, Marcelo Mendoza, and Barbara Poblete. 2011. Information credibility on Twitter. In Proceedings of WWW. 675--684. Google ScholarDigital Library
- Meeyoung Cha, Hamed Haddadi, Fabricio Benevenuto, and P Krishna Gummadi. 2010. Measuring user influence in Twitter: The million follower fallacy. Proceedings of ICWSM, Vol. 10, 10--17 (2010), 30.Google ScholarCross Ref
- Justin Cheng, Cristian Danescu-Niculescu-Mizil, Jure Leskovec, and Michael Bernstein. 2017. Anyone Can Become a Troll. American Scientist, Vol. 105, 3 (2017), 152.Google ScholarCross Ref
- Eunsol Choi, Chenhao Tan, Lillian Lee, Cristian Danescu-Niculescu-Mizil, and Jennifer Spindel. 2012. Hedge detection as a lens on framing in the GMO debates: A position paper Proceedings of the Workshop on Extra-Propositional Aspects of Meaning in Computational Linguistics. 70--79. Google ScholarDigital Library
- Giovanni Luca Ciampaglia, Prashant Shiralkar, Luis M Rocha, Johan Bollen, Filippo Menczer, and Alessandro Flammini. 2015. Computational fact checking from knowledge networks. PloS one, Vol. 10, 6 (2015), e0128193.Google ScholarCross Ref
- Niall J Conroy, Victoria L Rubin, and Yimin Chen. 2015. Automatic deception detection: methods for finding fake news. Proceedings of the Association for Information Science and Technology, Vol. 52, 1 (2015), 1--4. Google ScholarDigital Library
- Bella M DePaulo, James J Lindsay, Brian E Malone, Laura Muhlenbruck, Kelly Charlton, and Harris Cooper. 2003. Cues to deception. Psychological bulletin Vol. 129, 1 (2003), 74.Google Scholar
- Paul Ekman and Wallace V Friesen. 1969. Nonverbal leakage and clues to deception. Psychiatry, Vol. 32, 1 (1969), 88--106.Google ScholarCross Ref
- Song Feng, Ritwik Banerjee, and Yejin Choi. 2012 a. Syntactic stylometry for deception detection. In Proceedings of ACL. 171--175. Google ScholarDigital Library
- Song Feng, Longfei Xing, Anupam Gogar, and Yejin Choi. 2012 b. Distributional Footprints of Deceptive Product Reviews. Proceedings of ICWSM Vol. 12 (2012), 98--105.Google Scholar
- Emilio Ferrara. 2017. Disinformation and social bot operations in the run up to the 2017 French presidential election. (2017).Google Scholar
- Eileen Fitzpatrick and Joan Bachenko. 2012. Building a data collection for deception research. Proceedings of the Workshop on Computational Approaches to Deception Detection. 31--38. Google ScholarDigital Library
- Fortune. 2016. Fake Bloomberg news report drives Twitter stock up 8%. http://fortune.com/2015/07/14/fake-twitter-bloomberg-report/. (2016). Accessed: 2016--12--12.Google Scholar
- Maria Glenski, Ellyn Ayton, Dustin Arendt, and Svitlana Volkova. 2017. Fishing for Clickbaits in Social Images and Texts with Linguistically-Infused Neural Network Models. arXiv preprint arXiv:1710.06390 (2017).Google Scholar
- Jesse Graham, Jonathan Haidt, and Brian A Nosek. 2009. Liberals and conservatives rely on different sets of moral foundations. Journal of Personality and Social Psychology, Vol. 96, 5 (2009), 1029.Google ScholarCross Ref
- Guardian. 2015. Woman dies after taking "diet pills" bought over internet. https://www.theguardian.com/society/2015/apr/21/woman-dies-after-taking-diet-pills-bought-over-internet. (2015).Google Scholar
- Jonathan Haidt and Jesse Graham. 2007. When morality opposes justice: Conservatives have moral intuitions that liberals may not recognize. Social Justice Research Vol. 20, 1 (2007), 98--116.Google ScholarCross Ref
- Jonathan Haidt and Craig Joseph. 2004. Intuitive ethics: How innately prepared intuitions generate culturally variable virtues. Daedalus, Vol. 133, 4 (2004), 55--66.Google ScholarCross Ref
- Joan B. Hooper. 1975. On assertive predicates. In Syntax and Semantics, bibfieldeditorJ. Kimball (Ed.), Vol. Vol. 4. 91--124.Google Scholar
- Lee Howell. 2013. Digital wildfires in a hyperconnected world. WEF Report (2013).Google Scholar
- Ken Hyland. 2005. Metadiscourse. Wiley Online Library.Google Scholar
- Fang Jin, Wei Wang, Liang Zhao, Edward Dougherty, Yang Cao, Chang-Tien Lu, and Naren Ramakrishnan. 2014. Misinformation propagation in the age of Twitter. Computer, Vol. 47, 12 (2014), 90--94. Google ScholarDigital Library
- Lauri Karttunen. 1971. Implicative verbs. Language (1971), 340--358.Google Scholar
- Paul Kiparsky and Carol Kiparsky. 1968. Fact. Indiana University.Google Scholar
- Srijan Kumar, Robert West, and Jure Leskovec. 2016. Disinformation on the Web: Impact, Characteristics, and Detection of Wikipedia Hoaxes Proceedings of WWW. 591--602. Google ScholarDigital Library
- Kenton Lee, Yoav Artzi, Yejin Choi, and Luke Zettlemoyer. 2015. Event Detection and Factuality Assessment with Non-Expert Supervision. Proceedings of EMNLP (2015).Google ScholarCross Ref
- Newton Lee. 2014. Misinformation and Disinformation. Facebook Nation. Springer, 169--188.Google Scholar
- Kristina Lerman and Rumi Ghosh. 2010. Information contagion: An empirical study of the spread of news on Digg and Twitter social networks. Proceedings of ICWSM (2010), 90--97.Google ScholarCross Ref
- Cristian Lumezanu, Nick Feamster, and Hans Klein. 2012. # bias: Measuring the tweeting behavior of propagandists. (2012).Google Scholar
- Feamster Lumezanu and H Klein. 2012. Measuring the tweeting behavior of propagandists. Proceedings of ICWSM.Google Scholar
- Rada Mihalcea and Carlo Strapparava. 2009. The lie detector: Explorations in the automatic recognition of deceptive language Proceedings of the ACL-IJCNLP. 309--312. Google ScholarDigital Library
- Tanushree Mitra, Graham P Wright, and Eric Gilbert. 2017. A parsimonious language model of social media credibility across disparate events Proceedings of CSCW. 126--145. Google ScholarDigital Library
- Bjarke Mønsted, Piotr Sapie.zy'nski, Emilio Ferrara, and Sune Lehmann. 2017. Evidence of complex contagion of information in social media: An experiment using Twitter bots. PLOS ONE, Vol. 12, 9 (09. 2017), 1--12.Google ScholarCross Ref
- Matthew L Newman, James W Pennebaker, Diane S Berry, and Jane M Richards. 2003. Lying words: Predicting deception from linguistic styles. Personality and social psychology bulletin Vol. 29, 5 (2003), 665--675.Google Scholar
- NYTimes. 2016. Google and Facebook Take Aim at Fake News Sites. http://www.nytimes.com/2016/11/15/technology/google-will-ban-websites-that-host-fake-news-from-using-its-ad-service.html. (2016).Google Scholar
- Myle Ott, Yejin Choi, Claire Cardie, and Jeffrey T Hancock. 2011. Finding deceptive opinion spam by any stretch of the imagination Proceedings of ACL-HLT. 309--319. Google ScholarDigital Library
- F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research Vol. 12 (2011), 2825--2830. Google ScholarDigital Library
- James W Pennebaker, Martha E Francis, and Roger J Booth. 2001. Linguistic inquiry and word count: LIWC 2001. Mahway: Lawrence Erlbaum Associates Vol. 71 (2001), 2001.Google Scholar
- Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In Proceedings EMNLP. 1532--1543.Google ScholarCross Ref
- Verónica Pérez-Rosas, Bennett Kleinberg, Alexandra Lefevre, and Rada Mihalcea. 2017. Automatic Detection of Fake News. arXiv preprint arXiv:1708.07104 (2017).Google Scholar
- Verónica Pérez-Rosas and Rada Mihalcea. 2015. Experiments in Open Domain Deception Detection. Proceedings of EMNLP (2015), 1120--1125.Google ScholarCross Ref
- Slav Petrov. 2016. Announcing SyntaxNet: The World's Most Accurate Parser Goes Open Source. Google Research Blog, May Vol. 12 (2016), 2016.Google Scholar
- Vahed Qazvinian, Emily Rosengren, Dragomir R Radev, and Qiaozhu Mei. 2011. Rumor has it: Identifying misinformation in microblogs Proceedings of EMNLP. 1589--1599. Google ScholarDigital Library
- Hannah Rashkin, Eric Bell, Yejin Choi, and Svitlana Volkova. 2017 a. Multilingual Connotation Frames: A Case Study on Social Media for Targeted Sentiment Analysis and Forecast. In Proceedings of ACL, Vol. Vol. 2. 459--464.Google Scholar
- Hannah Rashkin, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, and Yejin Choi. 2017 b. Truth of varying shades: Analyzing language in fake news and political fact-checking Proceedings of EMNLP. 2921--2927.Google Scholar
- Hannah Rashkin, Sameer Singh, and Yejin Choi. 2016. Connotation Frames: A Data-Driven Investigation. Proceedings of ACL. 311--321.Google ScholarCross Ref
- Marta Recasens, Cristian Danescu-Niculescu-Mizil, and Dan Jurafsky. 2013. Linguistic Models for Analyzing and Detecting Biased Language Proceedings of ACL. 1650--1659.Google Scholar
- Victoria L Rubin. 2010. On deception and deception detection: Content analysis of computer-mediated stated beliefs. Proceedings of the American Society for Information Science and Technology, Vol. 47, 1 (2010), 1--10. Google ScholarDigital Library
- Victoria L Rubin, Yimin Chen, and Niall J Conroy. 2015. Deception detection for news: three types of fakes. Proceedings of the Association for Information Science and Technology, Vol. 52, 1 (2015), 1--4. Google ScholarDigital Library
- Victoria L Rubin, Niall J Conroy, Yimin Chen, and Sarah Cornwell. 2016. Fake News or Truth Using Satirical Cues to Detect Potentially Misleading News Proceedings of NAACL-HLT. 7--17.Google Scholar
- Kate Starbird. 2017. Examining the Alternative Media Ecosystem Through the Production of Alternative Narratives of Mass Shooting Events on Twitter. Proceedings of ICWSM. 230--239.Google ScholarCross Ref
- Bruno Takahashi, Edson C Tandoc, and Christine Carmichael. 2015. Communicating on Twitter during a disaster: An analysis of tweets during Typhoon Haiyan in the Philippines. Computers in Human Behavior Vol. 50 (2015), 392--398. Google ScholarDigital Library
- Marcella Tambuscio, Giancarlo Ruffo, Alessandro Flammini, and Filippo Menczer. 2015. Fact-checking effect on viral hoaxes: A model of misinformation spread in social networks Proceedings of WWW. 977--982. Google ScholarDigital Library
- Michail Tsikerdekis and Sherali Zeadally. 2014. Online deception in social media. Commun. ACM Vol. 57, 9 (2014), 72--80. Google ScholarDigital Library
- Onur Varol, Emilio Ferrara, Clayton A Davis, Filippo Menczer, and Alessandro Flammini. 2017. Online human-bot interactions: Detection, estimation, and characterization Proceedings of ICWSM. 280--289.Google Scholar
- Svitlana Volkova, Kyle Shaffer, Jin Yean Jang, and Nathan Hodas. 2017. Separating Facts from Fiction: Linguistic Models to Classify Suspicious and Trusted News Posts on Twitter. In Proceedings of ACL. 647--653.Google ScholarCross Ref
- William Yang Wang. 2017. "Liar, Liar Pants on Fire": A New Benchmark Dataset for Fake News Detection Proceedings of ACL. 422--426.Google Scholar
- Miron Zuckerman, Richard Koestner, and Robert Driver. 1981. Beliefs about cues associated with deception. Journal of Nonverbal Behavior Vol. 6, 2 (1981), 105--114.Google ScholarCross Ref
Index Terms
- Misleading or Falsification: Inferring Deceptive Strategies and Types in Online News and Social Media
Recommendations
Believability and Harmfulness Shape the Virality of Misleading Social Media Posts
WWW '23: Proceedings of the ACM Web Conference 2023Misinformation on social media presents a major threat to modern societies. While previous research has analyzed the virality across true and false social media posts, not every misleading post is necessarily equally viral. Rather, misinformation has ...
Socio-Emotional Computational Analysis of Propaganda Campaigns on Social Media Users in the Middle East
WWW '23 Companion: Companion Proceedings of the ACM Web Conference 2023Society has been significantly impacted by social media platforms in almost every aspect of their life. This impact has been effectively formulating people’s global mindsets and opinions on political, economic, and social events. Such waves of opinion ...
Social and Political Event Analysis based on Rich Media
MM '18: Proceedings of the 26th ACM international conference on MultimediaThis tutorial aims to provide a comprehensive overview on the applications of rich social media data for real world social and political event analysis, which is a new emerging topic in multimedia research. We will discuss the recent evolution of social ...
Comments