Vulnerable community identification using hate speech detection on social media

Abstract With the rapid development in mobile computing and Web technologies, online hate speech has been increasingly spread in social network platforms since it's easy to post any opinions. Previous studies confirm that exposure to online hate speech has serious offline consequences to historically deprived communities. Thus, research on automated hate speech detection has attracted much attention. However, the role of social networks in identifying hate-related vulnerable community is not well investigated. Hate speech can affect all population groups, but some are more vulnerable to its impact than others. For example, for ethnic groups whose languages have few computational resources, it is a challenge to automatically collect and process online texts, not to mention automatic hate speech detection on social media. In this paper, we propose a hate speech detection approach to identify hatred against vulnerable minority groups on social media. Firstly, in Spark distributed processing framework, posts are automatically collected and pre-processed, and features are extracted using word n-grams and word embedding techniques such as Word2Vec. Secondly, deep learning algorithms for classification such as Gated Recurrent Unit (GRU), a variety of Recurrent Neural Networks (RNNs), are used for hate speech detection. Finally, hate words are clustered with methods such as Word2Vec to predict the potential target ethnic group for hatred. In our experiments, we use Amharic language in Ethiopia as an example. Since there was no publicly available dataset for Amharic texts, we crawled Facebook pages to prepare the corpus. Since data annotation could be biased by culture, we recruit annotators from different cultural backgrounds and achieved better inter-annotator agreement. In our experimental results, feature extraction using word embedding techniques such as Word2Vec performs better in both classical and deep learning-based classification algorithms for hate speech detection, among which GRU achieves the best result. Our proposed approach can successfully identify the Tigre ethnic group as the highly vulnerable community in terms of hatred compared with Amhara and Oromo. As a result, hatred vulnerable group identification is vital to protect them by applying automatic hate speech detection model to remove contents that aggravate psychological harm and physical conflicts. This can also encourage the way towards the development of policies, strategies, and tools to empower and protect vulnerable communities.

[1]  Ika Alfina,et al.  Hate speech detection in the Indonesian language: A dataset and preliminary study , 2017, 2017 International Conference on Advanced Computer Science and Information Systems (ICACSIS).

[2]  B. Turner,et al.  Vulnerability And Human Rights , 2006 .

[3]  Ingmar Weber,et al.  Understanding Abuse: A Typology of Abusive Language Detection Subtasks , 2017, ALW@ACL.

[4]  Walter Daelemans,et al.  Automatic Detection and Prevention of Cyberbullying , 2015 .

[5]  Stephen Lindsay,et al.  Special topic: Designing for and with vulnerable people , 2014, INTR.

[6]  Njagi Dennis Gitari,et al.  A Lexicon-based Approach for Hate Speech Detection , 2015, MUE 2015.

[7]  Andrew Jakubowicz,et al.  Racism and the Affordances of the Internet , 2017 .

[8]  Pak Ming Fan,et al.  Constructing knowledge representation from lecture videos through multimodal analysis , 2013 .

[9]  Felice Dell'Orletta,et al.  Overview of the EVALITA 2018 Hate Speech Detection Task , 2018, EVALITA@CLiC-it.

[10]  Ana-Maria Bliuc,et al.  ‘It’s okay to be racist’: moral disengagement in online discussions of racist incidents in Australia , 2016 .

[11]  Anand Rajaraman,et al.  Mining of Massive Datasets , 2011 .

[12]  Gatot Yudoko,et al.  Technological solution for vulnerable communities: How does its approach matter? , 2014 .

[13]  Pete Burnap,et al.  Us and them: identifying cyber hate on Twitter across multiple protected characteristics , 2016, EPJ Data Science.

[14]  Maarit Makinen,et al.  Social Media and Postelection Crisis in Kenya , 2008 .

[15]  Walid Magdy,et al.  Abusive Language Detection on Arabic Social Media , 2017, ALW@ACL.

[16]  Andreas Reis,et al.  Public Health Ethics: Global Cases, Practice, and Context -- Public Health Ethics: Cases Spanning the Globe , 2016 .

[17]  Susan Dodds,et al.  Why bioethics needs a concept of vulnerability , 2012 .

[18]  Justin W. Patchin,et al.  Offline Consequences of Online Victimization , 2007 .

[19]  Felice Dell'Orletta,et al.  Hate Me, Hate Me Not: Hate Speech Detection on Facebook , 2017, ITASEC.

[20]  Mai ElSherief,et al.  Hate Lingo: A Target-based Linguistic Analysis of Hate Speech in Social Media , 2018, ICWSM.

[21]  Vasudeva Varma,et al.  Deep Learning for Hate Speech Detection in Tweets , 2017, WWW.

[22]  Arjan Durresi,et al.  Techniques for Collecting data in Social Networks , 2014, 2014 17th International Conference on Network-Based Information Systems.

[23]  Yang Xiang,et al.  A Two Phase Deep Learning Model for Identifying Discrimination from Tweets , 2016, EDBT.

[24]  Jenq-Haur Wang,et al.  Social Network Hate Speech Detection for Amharic Language , 2018 .

[25]  Chartes C. Roseman Troublesome Reflection: Racism as the Blind Spot in the Scientific Critique of Race , 2014 .

[26]  Dirk Hovy,et al.  Hateful Symbols or Hateful People? Predictive Features for Hate Speech Detection on Twitter , 2016, NAACL.

[27]  Shaun W. Lawson,et al.  Troubling Vulnerability: Designing with LGBT Young People's Ambivalence Towards Hate Crime Reporting , 2018, CHI.

[28]  Derek Ruths,et al.  A Web of Hate: Tackling Hateful Speech in Online Social Spaces , 2017, ArXiv.

[29]  Corinthias P. M. Sianipar,et al.  Technological solution for vulnerable communities: Questioning the sustainability of Appropriate Technology , 2015 .

[30]  Björn Ross,et al.  Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis , 2016, ArXiv.

[31]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[32]  Derek Ruths,et al.  Vectors for Counterspeech on Twitter , 2017, ALW@ACL.

[33]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[34]  Joanna Perry,et al.  Ireland in an International Comparative Context , 2017 .

[35]  Alexander F. Gelbukh,et al.  Aggression Detection in Social Media: Using Deep Neural Networks, Data Augmentation, and Pseudo Labeling , 2018, TRAC@COLING 2018.

[36]  C. Barry Hoffmaster,et al.  What Does Vulnerability Mean? , 2006, The Hastings Center report.

[37]  Robert E. Goodin,et al.  Protecting the vulnerable : a reanalysis of our social responsibilities , 1985 .

[38]  Yuzhou Wang,et al.  Locate the Hate: Detecting Tweets against Blacks , 2013, AAAI.

[39]  Tomoaki Ohtsuki,et al.  Hate Speech on Twitter: A Pragmatic Approach to Collect Hateful and Offensive Expressions and Perform Hate Speech Detection , 2018, IEEE Access.

[40]  Florencia Luna,et al.  Elucidating the Concept of Vulnerability: Layers Not Labels , 2009 .

[41]  Mauro Conti,et al.  All You Need is "Love": Evading Hate Speech Detection , 2018, ArXiv.

[42]  Sérgio Nunes,et al.  A Survey on Automatic Detection of Hate Speech in Text , 2018, ACM Comput. Surv..

[43]  Stephen Wilkinson,et al.  European Textbook on Ethics in Research , 2010 .

[44]  Martha Albertson Fineman,et al.  The Vulnerable Subject: Anchoring Equality in the Human Condition , 2008 .