Examining factors associated with Twitter account suspension following the 2020 U.S. presidential election

Online social media enables mass-level, transparent, and democratized discussion on numerous socio-political issues. Due to such openness, these platforms often endure manipulation and misinformation - leading to negative impacts. To prevent such harmful activities, platform moderators employ countermeasures to safeguard against actors violating their rules. However, the correlation between publicly outlined policies and employed action is less clear to general people. In this work, we examine violations and subsequent moderation related to the 2020 U.S. President Election discussion on Twitter, a popular micro-blogging site. We focus on quantifying plausible reasons for the suspension, drawing on Twitter's rules and policies by identifying suspended users (Case) and comparing their activities and properties with (yet) non-suspended (Control) users. Using a dataset of 240M election-related tweets made by 21M unique users, we observe that Suspended users violate Twitter's rules at a higher rate (statistically significant) than Control users across all the considered aspects - hate speech, offensiveness, spamming, and civic integrity. Moreover, through the lens of Twitter's suspension mechanism, we qualitatively examine the targeted topics for manipulation.

[1]  Fernando Diaz,et al.  CrisisLex: A Lexicon for Collecting and Filtering Microblogged Communications in Crises , 2014, ICWSM.

[2]  Padmini Srinivasan,et al.  A Postmortem of Suspended Twitter Accounts in the 2016 U.S. Presidential Election , 2019, 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM).

[3]  B. Lewis,et al.  Ethical research standards in a world of big data , 2014, F1000Research.

[4]  Eric Gilbert,et al.  Still out there: Modeling and Identifying Russian Troll Accounts on Twitter , 2019, WebSci.

[5]  Emilio Ferrara,et al.  Characterizing social media manipulation in the 2020 U.S. presidential election , 2020, First Monday.

[6]  Emily Bienvenue Computational propaganda: political parties, politicians, and political manipulation on social media , 2020, International Affairs.

[7]  Emilio Ferrara,et al.  Disinformation and Social Bot Operations in the Run Up to the 2017 French Presidential Election , 2017, First Monday.

[8]  Emilio Ferrara,et al.  Social Bots Distort the 2016 US Presidential Election Online Discussion , 2016, First Monday.

[9]  Eric P. Xing,et al.  Sparse Additive Generative Models of Text , 2011, ICML.

[10]  L. Stewart,et al.  Acting the Part: Examining Information Operations Within #BlackLivesMatter Discourse , 2018, Proc. ACM Hum. Comput. Interact..

[11]  Leonardo Neves,et al.  CEAM: The Effectiveness of Cyclic and Ephemeral Attention Models of User Behavior on Social Platforms , 2021, ICWSM.

[12]  Abdullah Mueen,et al.  On Twitter Purge: A Retrospective Analysis of Suspended Users , 2020, WWW.

[13]  Chao Yang,et al.  CATS: Characterizing automation of Twitter spammers , 2013, 2013 Fifth International Conference on Communication Systems and Networks (COMSNETS).

[14]  Kristina Lerman,et al.  Characterizing the 2016 Russian IRA influence campaign , 2018, Social Network Analysis and Mining.

[15]  Homero Gil de Zúñiga,et al.  Social Media Use for News and Individuals' Social Capital, Civic Engagement and Political Participation , 2012, J. Comput. Mediat. Commun..

[16]  K. Schulz,et al.  Case-control studies: research in reverse , 2002, The Lancet.

[17]  Dawn Xiaodong Song,et al.  Suspended accounts in retrospect: an analysis of twitter spam , 2011, IMC '11.

[18]  Gianluca Stringhini,et al.  Disinformation Warfare: Understanding State-Sponsored Trolls on Twitter and Their Influence on the Web , 2018, WWW.

[19]  Karl Aberer,et al.  Comparing Events Coverage in Online News and Social Media: The Case of Climate Change , 2015, ICWSM.

[20]  Koustuv Saha,et al.  Causal Factors of Effective Psychosocial Outcomes in Online Mental Health Communities , 2020, medRxiv.

[21]  Dominic Lewinski,et al.  Russian Troll Account Classification with Twitter and Facebook Data , 2021, ArXiv.

[22]  Gianluca Stringhini,et al.  Large Scale Crowdsourcing and Characterization of Twitter Abusive Behavior , 2018, ICWSM.

[23]  Abdullah Mueen,et al.  BotCamp: Bot-driven Interactions in Social Campaigns , 2019, WWW.

[24]  A. Bruns,et al.  The use of Twitter hashtags in the formation of ad hoc publics , 2011 .

[25]  Kristina Lerman,et al.  Analyzing the Digital Traces of Political Manipulation: The 2016 Russian Interference Twitter Campaign , 2018, 2018 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM).

[26]  Ingmar Weber,et al.  Automated Hate Speech Detection and the Problem of Offensive Language , 2017, ICWSM.

[27]  Leonardo Neves,et al.  AdverTiming Matters: Examining User Ad Consumption for Effective Ad Allocations on Social Media , 2021, CHI.

[28]  Virgílio A. F. Almeida,et al.  Detecting Spammers on Twitter , 2010 .