Crossmod: A Cross-Community Learning-based System to Assist Reddit Moderators

In this paper, we introduce a novel sociotechnical moderation system for Reddit called Crossmod. Through formative interviews with 11 active moderators from 10 different subreddits, we learned about the limitations of currently available automated tools, and how a new system could extend their capabilities. Developed out of these interviews, Crossmod makes its decisions based on cross-community learning---an approach that leverages a large corpus of previous moderator decisions via an ensemble of classifiers. Finally, we deployed Crossmod in a controlled environment, simulating real-time conversations from two large subreddits with over 10M subscribers each. To evaluate Crossmod's moderation recommendations, 4 moderators reviewed comments scored by Crossmod that had been drawn randomly from existing threads. Crossmod achieved an overall accuracy of 86% when detecting comments that would be removed by moderators, with high recall (over 87.5%). Additionally, moderators reported that they would have removed 95.3% of the comments flagged by Crossmod; however, 98.3% of these comments were still online at the time of this writing (i.e., not removed by the current moderation system). To the best of our knowledge, Crossmod is the first open source, AI-backed sociotechnical moderation system to be designed using participatory methods.

[1]  Eric Gilbert,et al.  Tweety Holmes: A Browser Extension for Abusive Twitter Profile Detection , 2018, CSCW Companion.

[2]  J. Hassard,et al.  Actor Network Theory and After , 1999 .

[3]  David Jurgens,et al.  A Just and Comprehensive Strategy for Using NLP to Address Online Abuse , 2019, ACL.

[4]  Erik Brynjolfsson,et al.  Understanding the Digital Economy: Data, Tools, and Research , 2000 .

[5]  Tomas Mikolov,et al.  Bag of Tricks for Efficient Text Classification , 2016, EACL.

[6]  R. Geiger Bot-based collective blocklists in Twitter: the counterpublic moderation of harassment in a networked public space , 2016 .

[7]  P. Resnick,et al.  Building Successful Online Communities: Evidence-Based Social Design , 2012 .

[8]  P. Gummett The Social Construction of Technological Systems: New Directions in the Sociology and History of Technology , 1988 .

[9]  Martin J. Riedl,et al.  Book review: Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions that Shape Social Media , 2019, New Media Soc..

[10]  Joel R. Tetreault,et al.  Abusive Language Detection in Online User Content , 2016, WWW.

[11]  Lawrence Lessig,et al.  Code and Other Laws of Cyberspace , 1999 .

[12]  Eric Horvitz,et al.  Principles of mixed-initiative user interfaces , 1999, CHI '99.

[13]  Geoff Kaufman,et al.  "At Least the Pizzas You Make Are Hot": Norms, Values, and Abrasive Humor on the Subreddit r/RoastMe , 2018, ICWSM.

[14]  Nicholas Diakopoulos,et al.  Algorithmic Accountability , 2015 .

[15]  T. Gillespie Governance of and by platforms , 2016 .

[16]  Eric Gilbert,et al.  The Bag of Communities: Identifying Abusive Behavior Online with Preexisting Internet Data , 2017, CHI.

[17]  Tomas Mikolov,et al.  Enriching Word Vectors with Subword Information , 2016, TACL.

[18]  Jacob Eisenstein,et al.  You Can't Stay Here , 2017 .

[19]  Michael S. Bernstein,et al.  4chan and /b/: An Analysis of Anonymity and Ephemerality in a Large Online Community , 2011, ICWSM.

[20]  Munmun De Choudhury,et al.  Prevalence and Psychological Effects of Hateful Speech in Online College Communities , 2019, WebSci.

[21]  Tom Feltwell,et al.  "Could You Define That in Bot Terms"?: Requesting, Creating and Using Bots on Reddit , 2017, CHI.

[22]  Paul Resnick,et al.  Slash(dot) and burn: distributed moderation in a large online conversation space , 2004, CHI.

[23]  Clifford Nass,et al.  Normative influences on thoughtful online participation , 2011, CHI.

[24]  Paul Resnick,et al.  Reputation systems , 2000, CACM.

[25]  Cliff Lampe,et al.  Classification and Its Consequences for Online Harassment , 2017, Proc. ACM Hum. Comput. Interact..

[26]  Caroline Haythornthwaite,et al.  Studying Online Social Networks , 2006, J. Comput. Mediat. Commun..

[27]  Eric Gilbert,et al.  Human-Machine Collaboration for Content Regulation , 2019, ACM Trans. Comput. Hum. Interact..

[28]  Mark S. Ackerman,et al.  The Intellectual Challenge of CSCW: The Gap Between Social Requirements and Technical Feasibility , 2000, Hum. Comput. Interact..

[29]  Sarah T. Roberts Commercial Content Moderation: Digital Laborers' Dirty Work , 2016 .

[30]  Dong Wang,et al.  Click-through Prediction for Advertising in Twitter Timeline , 2015, KDD.

[31]  Enid Mumford,et al.  Socio-technical Design: An Unfulfilled Promise or a Future Opportunity , 2000, Organizational and Social Perspectives on IT.

[32]  Tal Z. Zarsky,et al.  The Trouble with Algorithmic Decisions , 2016 .

[33]  Munmun De Choudhury,et al.  #thyghgapp: Instagram Content Moderation and Lexical Variation in Pro-Eating Disorder Communities , 2016, CSCW.

[34]  Steve Sawyer,et al.  Sociotechnical Approaches to the Study of Information Systems , 2014, Computing Handbook, 3rd ed..

[35]  J. Preece,et al.  Online communities: focusing on sociability and usability , 2002 .

[36]  M. Choudhury,et al.  "This Post Will Just Get Taken Down": Characterizing Removed Pro-Eating Disorder Social Media Content , 2016, CHI.

[37]  Rob Kling,et al.  IT and organizational change in digital economies: a socio-technical approach , 1999, CSOC.

[38]  Rebecca Gray,et al.  Understanding User Beliefs About Algorithmic Curation in the Facebook News Feed , 2015, CHI.

[39]  Florence March,et al.  2016 , 2016, Affair of the Heart.

[40]  Zahra Ashktorab,et al.  Designing Cyberbullying Mitigation and Prevention Solutions through Participatory Design With Teenagers , 2016, CHI.

[41]  Nicholas Diakopoulos,et al.  Cooooooooooooooollllllllllllll!!!!!!!!!!!!!! Using Word Lengthening to Detect Sentiment in Microblogs , 2011, EMNLP.

[42]  Motahhare Eslami Understanding and Designing around Users' Interaction with Hidden Algorithms in Sociotechnical Systems , 2017, CSCW Companion.

[43]  Eric Gilbert,et al.  The Internet's Hidden Rules , 2018, Proceedings of the ACM on Human-Computer Interaction.

[44]  A. Bruckman,et al.  Online Harassment and Content Moderation , 2018 .

[45]  John Pavlopoulos,et al.  Deep Learning for User Comment Moderation , 2017, ALW@ACL.

[46]  Yair Zick,et al.  Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems , 2016, 2016 IEEE Symposium on Security and Privacy (SP).

[47]  Paul Resnick,et al.  Trust among strangers in internet transactions: Empirical analysis of eBay' s reputation system , 2002, The Economics of the Internet and E-commerce.

[48]  Sarah T. Roberts Behind the screen: the hidden digital labor of commercial content moderation , 2014 .

[49]  Amy X. Zhang,et al.  Making Sense of Group Chat through Collaborative Tagging and Summarization , 2018, Proc. ACM Hum. Comput. Interact..

[50]  Casey Fiesler,et al.  Reddit Rules! Characterizing an Ecosystem of Governance , 2018, ICWSM.

[51]  Eric Gilbert,et al.  Hybrid Approaches to Detect Comments Violating Macro Norms on Reddit , 2019, ArXiv.

[52]  David R. Karger,et al.  Squadbox: A Tool to Combat Email Harassment Using Friendsourced Moderation , 2018, CHI.

[53]  Elizabeth F. Churchill,et al.  Profanity use in online communities , 2012, CHI.

[54]  Geoff Kaufman,et al.  Moderator engagement and community development in the age of algorithms , 2019, New Media Soc..

[55]  W. Bijker The social construction of bakelite: toward a theory of invention , 1987 .

[56]  Elizabeth D. Mynatt,et al.  Just awful enough: the functional dysfunction of the something awful forums , 2014, CHI.

[57]  Bryce Glass,et al.  Building Web Reputation Systems , 2010 .

[58]  Lucas Dixon,et al.  Ex Machina: Personal Attacks Seen at Scale , 2016, WWW.