Abusive Language Detection in Online Conversations by Combining Content- and Graph-Based Features

In recent years, online social networks have allowed worldwide users to meet and discuss. As guarantors of these communities, the administrators of these platforms must prevent users from adopting inappropriate behaviors. This verification task, mainly done by humans, is more and more difficult due to the ever growing amount of messages to check. Methods have been proposed to automatize this moderation process, mainly by providing approaches based on the textual content of the exchanged messages. Recent work has also shown that characteristics derived from the structure of conversations, in the form of conversational graphs, can help detecting these abusive messages. In this paper, we propose to take advantage of both sources of information by proposing fusion methods integrating content-and graph-based features. Our experiments on raw chat logs show that the content of the messages, but also of their dynamics within a conversation contain partially complementary information, allowing performance improvements on an abusive message classification task with a final F-measure of 93.26%.

[1]  Paul Mutton,et al.  Inferring and visualizing social networks on Internet relay chat , 2004, Proceedings. Eighth International Conference on Information Visualisation, 2004. IV 2004..

[2]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[3]  Albert Ali Salah,et al.  Automatic analysis and identification of verbal aggression and abusive behaviors for online social games , 2015, Comput. Hum. Behav..

[4]  Ellen Spertus,et al.  Smokey: Automatic Recognition of Hostile Messages , 1997, AAAI/IAAI.

[5]  Georges Linarès,et al.  Graph-Based Features for Automatic Online Abuse Detection , 2017, SLSP.

[6]  John Pavlopoulos,et al.  Deep Learning for User Comment Moderation , 2017, ALW@ACL.

[7]  Helen Yannakoudakis,et al.  Neural Character-based Composition Models for Abuse Detection , 2018, ALW.

[8]  Henry Lieberman,et al.  Modeling the Detection of Textual Cyberbullying , 2011, The Social Mobile Web.

[9]  Georges Linarès,et al.  Conversational Networks for Automatic Online Moderation , 2019, IEEE Transactions on Computational Social Systems.

[10]  Leonardo Vidal Batista,et al.  Texture Classification Using the Lempel-Ziv-Welch Algorithm , 2004, SBIA.

[11]  Brian D. Davison,et al.  Detection of Harassment on Web 2.0 , 2009 .

[12]  Ying Chen,et al.  Detecting Offensive Language in Social Media to Protect Adolescent Online Safety , 2012, 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Confernece on Social Computing.

[13]  Lucas Dixon,et al.  Ex Machina: Personal Attacks Seen at Scale , 2016, WWW.

[14]  Djamel A. Zighed,et al.  Extracting Social Networks to Understand Interaction , 2011, 2011 International Conference on Advances in Social Networks Analysis and Mining.

[15]  Bülent Yener,et al.  A Tool for Internet Chatroom Surveillance , 2004, ISI.

[16]  Radha Poovendran,et al.  Deceiving Google's Perspective API Built for Detecting Toxic Comments , 2017, ArXiv.

[17]  Gábor Csárdi,et al.  The igraph software package for complex network research , 2006 .

[18]  Georges Linarès,et al.  Impact Of Content Features For Automatic Online Abuse Detection , 2017, CICLing.