Proposing an Interactive Audit Pipeline for Visual Privacy Research

In an ideal world, deployed machine learning models will enhance our society. We hope that those models will provide unbiased and ethical decisions that will benefit everyone. However, this is not always the case; issues arise during the data preparation process throughout the steps leading to the models’ deployment. The continued use of biased datasets and biased processes will adversely damage communities and increase the cost to fix the problem later. In this work, we walk through the decision making process that a researcher should consider before, during, and after a system deployment to understand the broader impacts of their research in the community. Throughout this paper, we discuss fairness, privacy, and ownership issues in the machine learning pipeline, assert the need for a responsible human-over-the-loop methodology to bring accountability into machine learning pipeline, and finally, reflect on the need to explore research agendas that have harmful societal impacts. We examine visual privacy research and draw lessons that can apply broadly to artificial intelligence. Our goal is to provide a systematic analysis of the machine learning pipeline for visual privacy and bias issues. With this pipeline, we hope to raise stakeholder (e.g., researchers, modelers, corporations) awareness as these issues propagate in the various machine learning phases.

[1]  G. Sapiro,et al.  Learning to Identify While Failing to Discriminate , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).

[2]  Anna Cinzia Squicciarini,et al.  Toward Automated Multiparty Privacy Conflict Detection , 2018, CIKM.

[3]  David J. Crandall,et al.  Enhancing Lifelogging Privacy by Detecting Screens , 2016, CHI.

[4]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[5]  Christan Earl Grant,et al.  A Framework for Interactive t-SNE Clustering , 2016, Int. J. Softw. Informatics.

[6]  Adel Said Elmaghraby,et al.  Cyber security challenges in Smart Cities: Safety, security and privacy , 2014, Journal of advanced research.

[7]  Heinrich Hußmann,et al.  You Can't Watch This!: Privacy-Respectful Photo Browsing on Smartphones , 2016, CHI.

[8]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[9]  Jonathon S. Hare,et al.  Privacy-aware image classification and search , 2012, SIGIR '12.

[10]  Jianping Fan,et al.  Automatic Privacy Prediction to Accelerate Social Image Sharing , 2017, 2017 IEEE Third International Conference on Multimedia Big Data (BigMM).

[11]  Cornelia Caragea,et al.  Image Privacy Prediction Using Deep Neural Networks , 2019, ACM Trans. Web.

[12]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[13]  Xiting Wang,et al.  Towards better analysis of machine learning models: A visual analytics perspective , 2017, Vis. Informatics.

[14]  Timofey S. Chernov,et al.  MIDV-500: A Dataset for Identity Documents Analysis and Recognition on Mobile Devices in Video Stream , 2018, Computer Optics.

[15]  Lakshminarayanan Subramanian,et al.  Cryptagram: photo privacy for online social media , 2013, COSN '13.

[16]  Steven M. Bellovin,et al.  The Failure of Online Social Network Privacy Settings , 2011 .

[17]  Yan Liang,et al.  [Research paper] formalizing interruptible algorithms for human over-the-loop analytics , 2017, 2017 IEEE International Conference on Big Data (Big Data).

[18]  Alexei A. Efros,et al.  Unbiased look at dataset bias , 2011, CVPR 2011.

[19]  David J. Crandall,et al.  Sensitive Lifelogs: A Privacy Analysis of Photos from Wearable Cameras , 2015, CHI.

[20]  Alessandro Acquisti,et al.  Information revelation and privacy in online social networks , 2005, WPES '05.

[21]  Federico Vegetti,et al.  What You Can Scrape and What Is Right to Scrape: A Proposal for a Tool to Collect Public Facebook Data , 2020, Social Media + Society.

[22]  Jasmine DeHart,et al.  Social Media and the Scourge of Visual Privacy , 2020, Inf..

[23]  Hongxin Hu,et al.  Blur vs. Block: Investigating the Effectiveness of Privacy-Enhancing Obfuscation for Images , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[24]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[25]  Carlos D. Castillo,et al.  An All-In-One Convolutional Neural Network for Face Analysis , 2016, 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017).

[26]  David Maxwell Chickering,et al.  ModelTracker: Redesigning Performance Analysis Tools for Machine Learning , 2015, CHI.

[27]  Kush R. Varshney,et al.  Optimized Pre-Processing for Discrimination Prevention , 2017, NIPS.

[28]  Jerry Alan Fails,et al.  Interactive machine learning , 2003, IUI '03.

[29]  Timnit Gebru,et al.  Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018, FAT.

[30]  Hongxin Hu,et al.  Effectiveness and Users' Experience of Obfuscation as a Privacy-Enhancing Technology for Sharing Photos , 2017, Proc. ACM Hum. Comput. Interact..

[31]  Aitor Almeida,et al.  Smart cities survey: Technologies, application domains and challenges for the cities of the future , 2019, Int. J. Distributed Sens. Networks.

[32]  Jasmine DeHart,et al.  Visual Content Privacy Leaks on Social Media Networks , 2018, ArXiv.

[33]  Michael Zimmer and Katharina Kinder-Kurlanda (eds), Internet Research Ethics for the Social Age: New Challenges, Cases, and Contexts , 2018, European Journal of Communication.

[34]  S. Krishnan,et al.  Security, Privacy and Steganographic Analysis of FaceApp and TikTok , 2020 .

[35]  Jose M. Such,et al.  Photo Privacy Conflicts in Social Media: A Large-scale Empirical Study , 2017, CHI.

[36]  David R. Zemmels,et al.  Sharing of Digital Visual Media: Privacy Concerns and Trust Among Young People , 2015 .

[37]  Anthony D. Miyazaki,et al.  Protecting children's privacy online: How parental mediation strategies affect website safeguard effectiveness , 2008 .

[38]  Balachander Krishnamurthy,et al.  Characterizing privacy in online social networks , 2008, WOSN '08.

[39]  Maya Cakmak,et al.  Power to the People: The Role of Humans in Interactive Machine Learning , 2014, AI Mag..

[40]  Yunfeng Zhang,et al.  AI Fairness 360: An extensible toolkit for detecting and mitigating algorithmic bias , 2019, IBM Journal of Research and Development.

[41]  David Danks,et al.  Algorithmic Bias in Autonomous Systems , 2017, IJCAI.

[42]  Krista A. Ehinger,et al.  SUN database: Large-scale scene recognition from abbey to zoo , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[43]  Harini Suresh,et al.  A Framework for Understanding Unintended Consequences of Machine Learning , 2019, ArXiv.

[44]  Andrei Barbu,et al.  Measuring Social Biases in Grounded Vision and Language Embeddings , 2020, NAACL.

[45]  Antonio Torralba,et al.  Ieee Transactions on Pattern Analysis and Machine Intelligence 1 80 Million Tiny Images: a Large Dataset for Non-parametric Object and Scene Recognition , 2022 .

[46]  Inioluwa Deborah Raji,et al.  Saving Face: Investigating the Ethical Concerns of Facial Recognition Auditing , 2020, AIES.

[47]  Yuriy Brun,et al.  Fairness testing: testing software for discrimination , 2017, ESEC/SIGSOFT FSE.

[48]  Mario Fritz,et al.  Automatically Detecting Bystanders in Photos to Reduce Privacy Risks , 2020, 2020 IEEE Symposium on Security and Privacy (SP).

[49]  Olga Russakovsky,et al.  Towards Fairness in Visual Recognition: Effective Strategies for Bias Mitigation , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[50]  Aditya G. Parameswaran,et al.  A Human-in-the-loop Perspective on AutoML: Milestones and the Road Ahead , 2019, IEEE Data Eng. Bull..

[51]  David S. Rosenblum,et al.  What Anyone Can Know: The Privacy Risks of Social Networking Sites , 2007, IEEE Security & Privacy.

[52]  Alexandros André Chaaraoui,et al.  Visual privacy protection methods: A survey , 2015, Expert Syst. Appl..

[53]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[54]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[55]  M. Zimmer “But the data is already public”: on the ethics of research in Facebook , 2010, Ethics and Information Technology.

[56]  E. A. vanZoonen Privacy concerns in smart cities , 2016 .

[57]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[58]  Chi Lin,et al.  VizWiz-Priv: A Dataset for Recognizing the Presence and Purpose of Private Visual Information in Images Taken by Blind People , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[59]  David Darling,et al.  Identification of Subjects and Bystanders in Photos with Feature-Based Machine Learning , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[60]  Considerations for Designing Private and Inexpensive Smart Cities , 2020 .

[61]  Mariella Dimiccoli,et al.  Mitigating Bystander Privacy Concerns in Egocentric Activity Recognition with Deep Learning and Intentional Image Degradation , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[62]  Vinay Uday Prabhu,et al.  Multimodal datasets: misogyny, pornography, and malignant stereotypes , 2021, ArXiv.

[63]  Sherali Zeadally,et al.  Bystanders' Privacy , 2017, IT Professional.

[64]  Jerry Alan Fails,et al.  A design tool for camera-based interaction , 2003, CHI '03.

[65]  Jonathon S. Hare,et al.  PicAlert!: a system for privacy-aware image classification and retrieval , 2012, CIKM.

[66]  Trevor Darrell,et al.  Women also Snowboard: Overcoming Bias in Captioning Models , 2018, ECCV.

[67]  Ang Li,et al.  HideMe: Privacy-Preserving Photo Sharing on Social Networks , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[68]  Leiser Silva,et al.  Legality and Ethics of Web Scraping , 2018, AMCIS.

[69]  T. Hellström,et al.  Bias in machine learning - what is it good for? , 2020, NeHuAI@ECAI.

[70]  Niels Bantilan,et al.  Themis-ml: A Fairness-Aware Machine Learning Interface for End-To-End Discrimination Discovery and Mitigation , 2017, ArXiv.

[71]  Cornelia Caragea,et al.  Image Privacy Prediction Using Deep Features , 2016, AAAI.