Leveraging Community and Author Context to Explain the Performance and Bias of Text-Based Deception Detection Models

Deceptive news posts shared in online communities can be detected with NLP models, and much recent research has focused on the development of such models. In this work, we use characteristics of online communities and authors — the context of how and where content is posted — to explain the performance of a neural network deception detection model and identify sub-populations who are disproportionately affected by model accuracy or failure. We examine who is posting the content, and where the content is posted to. We find that while author characteristics are better predictors of deceptive content than community characteristics, both characteristics are strongly correlated with model performance. Traditional performance metrics such as F1 score may fail to capture poor model performance on isolated sub-populations such as specific authors, and as such, more nuanced evaluation of deception detection models is critical.

[1]  Bu-Sung Lee,et al.  Unsupervised rumor detection based on users' behaviors using neural networks , 2017, Pattern Recognit. Lett..

[2]  Jintao Li,et al.  Rumor Detection with Hierarchical Social Attention Network , 2018, CIKM.

[3]  Maria Glenski,et al.  Political Bias and Factualness in News Sharing Across more then 100, 000 Online Communities , 2021, ICWSM.

[4]  Ingmar Weber,et al.  Racial Bias in Hate Speech and Abusive Language Detection Datasets , 2019, Proceedings of the Third Workshop on Abusive Language Online.

[5]  Qiang Zhang,et al.  Reply-Aided Detection of Misinformation via Bayesian Deep Learning , 2019, WWW.

[6]  Deepayan Bhowmik,et al.  Fake News Identification on Twitter with Hybrid CNN and RNN Models , 2018, SMSociety.

[7]  V. S. Subrahmanian,et al.  An Army of Me: Sockpuppets in Online Discussion Communities , 2017, WWW.

[8]  E. Hargittai,et al.  THE PARTICIPATION DIVIDE: Content creation and sharing in the digital age1 , 2008 .

[9]  Wei Gao,et al.  From Retweet to Believability: Utilizing Trust to Identify Rumor Spreaders on Twitter , 2017, ASONAM.

[10]  Victoria L. Rubin,et al.  Fake News or Truth? Using Satirical Cues to Detect Potentially Misleading News , 2016 .

[11]  Suresh Venkatasubramanian,et al.  Runaway Feedback Loops in Predictive Policing , 2017, FAT.

[12]  Luo Si,et al.  Rumor Detection by Exploiting User Credibility Information, Attention and Multi-task Learning , 2019, ACL.

[13]  Trevor van Mierlo The 1% Rule in Four Digital Health Social Networks: An Observational Study , 2014, Journal of medical Internet research.

[14]  Hany Farid,et al.  The accuracy, fairness, and limits of predicting recidivism , 2018, Science Advances.

[15]  Jean-Charles Delvenne,et al.  The anatomy of Reddit: An overview of academic research , 2017, Dynamics On and Of Complex Networks III.

[16]  Christos Faloutsos,et al.  REV2: Fraudulent User Prediction in Rating Platforms , 2018, WSDM.

[17]  Pascale Fung,et al.  Reducing Gender Bias in Abusive Language Detection , 2018, EMNLP.

[18]  Eric Gilbert,et al.  A Parsimonious Language Model of Social Media Credibility Across Disparate Events , 2017, CSCW.

[19]  Jeremy Blackburn,et al.  The Pushshift Reddit Dataset , 2020, ICWSM.

[20]  Svitlana Volkova,et al.  Separating Facts from Fiction: Linguistic Models to Classify Suspicious and Trusted News Posts on Twitter , 2017, ACL.

[21]  Eunsol Choi,et al.  Truth of Varying Shades: Analyzing Language in Fake News and Political Fact-Checking , 2017, EMNLP.

[22]  Tim Weninger,et al.  Consumers and Curators: Browsing and Voting Patterns on Reddit , 2017, IEEE Transactions on Computational Social Systems.

[23]  Tim Weninger,et al.  How Humans Versus Bots React to Deceptive and Trusted News Sources: A Case Study of Active Users , 2018, 2018 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM).

[24]  William Yang Wang “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection , 2017, ACL.

[25]  Huan Liu,et al.  Hierarchical Propagation Networks for Fake News Detection: Investigation and Exploitation , 2019, ICWSM.

[26]  Ellyn Ayton,et al.  Explaining Multimodal Deceptive News Prediction Models , 2019, ICWSM.

[27]  Wei Gao,et al.  Detecting Rumors from Microblogs with Recurrent Neural Networks , 2016, IJCAI.

[28]  Yejin Choi,et al.  The Risk of Racial Bias in Hate Speech Detection , 2019, ACL.

[29]  Seth Flaxman,et al.  European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..

[30]  Alexandra Chouldechova,et al.  Fair prediction with disparate impact: A study of bias in recidivism prediction instruments , 2016, Big Data.

[31]  Arkaitz Zubiaga,et al.  Discourse-aware rumour stance classification in social media using sequential classifiers , 2017, Inf. Process. Manag..