Towards Continuous Automatic Audits of Social Media Adaptive Behavior and its Role in Misinformation Spreading

In this paper, we argue for continuous and automatic auditing of social media adaptive behavior and outline its key characteristics and challenges. We are motivated by the spread of online misinformation, which has recently been fueled by opaque recommendations on social media platforms. Although many platforms have declared to take steps against the spread of misinformation, the effectiveness of such measures must be assessed independently. To this end, independent organizations and researchers carry out audits to quantitatively assess platform recommendation behavior and its effects (e.g., filter bubble creation tendencies). The audits are typically based on agents simulating the user behavior and collecting platform reactions (e.g., recommended items). The downside of such auditing is the cost related to the interpretation of collected data (here, some auditors are advancing automatic annotation). Furthermore, social media platforms are dynamic and ever-changing (algorithms change, concepts drift, new content appears). Therefore, audits need to be performed continuously. This further increases the need for automated data annotation. Regarding the data annotation, we argue for the application of weak supervision, semi-supervised learning, and human-in-the-loop techniques.

[1]  Paul Covington,et al.  Deep Neural Networks for YouTube Recommendations , 2016, RecSys.

[2]  Mark Crovella,et al.  How YouTube Leads Privacy-Seeking Users Away from Reliable Information , 2020, UMAP.

[3]  C. Basch,et al.  Preventive Behaviors Conveyed on YouTube to Mitigate Transmission of COVID-19: Cross-Sectional Study , 2020, JMIR public health and surveillance.

[4]  P. Návrat,et al.  Monant : Universal and Extensible Platform for Monitoring , Detection and Mitigation of Antisocial Behaviour , 2019 .

[5]  Pável Calado,et al.  Combining Similarity Features and Deep Representation Learning for Stance Detection in the Context of Checking Fake News , 2018, ACM J. Data Inf. Qual..

[6]  Robi Polikar,et al.  Incremental Learning of Concept Drift in Nonstationary Environments , 2011, IEEE Transactions on Neural Networks.

[7]  Qiang Zhang,et al.  From Stances' Imbalance to Their HierarchicalRepresentation and Detection , 2019, WWW.

[8]  Krishna P. Gummadi,et al.  Quantifying Search Bias: Investigating Sources of Bias for Political Searches in Social Media , 2017, CSCW.

[9]  Tanushree Mitra,et al.  Measuring Misinformation in Video Search Platforms: An Audit Study on YouTube , 2020, Proc. ACM Hum. Comput. Interact..

[10]  Christopher J. Lynch,et al.  Big data, agents, and machine learning: towards a data-driven agent-based modeling approach , 2018, SpringSim.

[11]  Petros Iosifidis,et al.  The battle to end fake news: A qualitative content analysis of Facebook announcements on how it combats disinformation , 2020, International Communication Gazette.

[12]  Xuezhi Wang,et al.  Relevant Document Discovery for Fact-Checking Articles , 2018, WWW.

[13]  Fausto Giunchiglia,et al.  Human-in-the-loop handling of knowledge drift , 2021, Data Mining and Knowledge Discovery.

[14]  Miklos A. Vasarhelyi,et al.  Continuous Online Auditing: A Program of Research , 1999, J. Inf. Syst..

[15]  Anna Kawakami,et al.  The Media Coverage of the 2020 US Presidential Election Candidates through the Lens of Google's Top Stories , 2020, ICWSM.

[16]  Christopher Ré,et al.  Snorkel: Rapid Training Data Creation with Weak Supervision , 2017, Proc. VLDB Endow..

[17]  Li Wei,et al.  Recommending what video to watch next: a multitask ranking system , 2019, RecSys.

[18]  Karrie Karahalios,et al.  Auditing Algorithms : Research Methods for Detecting Discrimination on Internet Platforms , 2014 .

[19]  Nitesh V. Chawla,et al.  The Tesserae Project: Large-Scale, Longitudinal, In Situ, Multimodal Sensing of Information Workers , 2019, CHI Extended Abstracts.

[20]  Pedro O. S. Vaz de Melo,et al.  Facebook Ads Monitor: An Independent Auditing System for Political Ads on Facebook , 2020, WWW.

[21]  Eric D. Ragan,et al.  Soliciting Human-in-the-Loop User Feedback for Interactive Machine Learning Reduces User Trust and Impressions of Model Accuracy , 2020, HCOMP.

[22]  Antisocial Media: How Facebook Disconnects Us and Undermines Democracy , 2018 .