Progressive disclosure: empirically motivated approaches to designing effective transparency

As we increasingly delegate important decisions to intelligent systems, it is essential that users understand how algorithmic decisions are made. Prior work has often taken a technocentric approach to transparency. In contrast, we explore empirical user-centric methods to better understand user reactions to transparent systems. We assess user reactions to transparency in two studies. In Study 1, users anticipated that a more transparent system would perform better, but retracted this evaluation after experience with the system. Qualitative data suggest this arose because transparency is distracting and undermines simple heuristics users form about system operation. Study 2 explored these effects in depth, suggesting that users may benefit from initially simplified feedback that hides potential system errors and assists users in building working heuristics about system operation. We use these findings to motivate new progressive disclosure principles for transparency in intelligent systems.

[1]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[2]  C. F. Kao,et al.  The efficient assessment of need for cognition. , 1984, Journal of personality assessment.

[3]  Engin Bozdag,et al.  Bias in algorithmic filtering and personalization , 2013, Ethics and Information Technology.

[4]  Fabio Paternò,et al.  Human Computer Interaction with Mobile Devices , 2002, Lecture Notes in Computer Science.

[5]  Tony Doyle,et al.  Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy , 2017, Inf. Soc..

[6]  Steve Whittaker,et al.  What Are You Hiding? Algorithmic Transparency and User Perceptions , 2018, AAAI Spring Symposia.

[7]  Clifford Nass,et al.  The media equation - how people treat computers, television, and new media like real people and places , 1996 .

[8]  Daniel S. Weld,et al.  The challenge of crafting intelligible intelligence , 2018, Commun. ACM.

[9]  Andrea Bunt,et al.  Are explanations always important?: a study of deployed, low-cost intelligent interactive systems , 2012, IUI '12.

[10]  John Nerbonne,et al.  The Secret Life of Pronouns. What Our Words Say About Us , 2014, Lit. Linguistic Comput..

[11]  S. Hart,et al.  Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .

[12]  Mark W. Newman,et al.  Learning from a learning thermostat: lessons for intelligent systems for the home , 2013, UbiComp.

[13]  D. Hilton Conversational processes and causal explanation. , 1990 .

[14]  Trevor Darrell,et al.  Grounding Visual Explanations , 2018, ECCV.

[15]  Karrie Karahalios,et al.  The Illusion of Control: Placebo Effects of Control Settings , 2018, CHI.

[16]  Anind K. Dey,et al.  Investigating intelligibility for uncertain context-aware applications , 2011, UbiComp '11.

[17]  Mi Zhang,et al.  MyBehavior: automatic personalized health feedback from user behaviors and preferences using smartphones , 2015, UbiComp.

[18]  Lei Huang,et al.  Sentence-level Emotion Classification with Label and Context Dependence , 2015, ACL.

[19]  Marilyn A. Walker,et al.  Learning Lexico-Functional Patterns for First-Person Affect , 2017, ACL.

[20]  Karrie Karahalios,et al.  First I "like" it, then I hide it: Folk Theories of Social Feeds , 2016, CHI.

[21]  Steve Whittaker,et al.  What Does All This Data Mean for My Future Mood? Actionable Analytics and Targeted Reflection for Emotional Well-Being , 2017, Hum. Comput. Interact..

[22]  Wojciech Samek,et al.  Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..

[23]  I. Ajzen The theory of planned behavior , 1991 .

[24]  Weng-Keen Wong,et al.  Fixing the program my computer learned: barriers for end users, challenges for the machine , 2009, IUI.

[25]  Lars Vedel Kessing,et al.  Supporting disease insight through data analysis: refinements of the monarca self-assessment system , 2013, UbiComp.

[26]  Tara Matthews,et al.  Can an Algorithm Know the "Real You"?: Understanding People's Reactions to Hyper-personal Analytics Systems , 2015, CHI.

[27]  Steve Whittaker,et al.  Dice in the Black Box: User Experiences with an Inscrutable Algorithm , 2018, AAAI Spring Symposia.

[28]  Bongshin Lee,et al.  Understanding self-reflection: how people reflect on personal data through visual data exploration , 2017, PervasiveHealth.

[29]  Cathy O'Neil,et al.  Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy , 2016, Vikalpa: The Journal for Decision Makers.

[30]  Lloyd H. Nakatani,et al.  Soft machines: A philosophy of user-computer interface design , 1983, CHI '83.

[31]  Anind K. Dey,et al.  Sensors Know When to Interrupt You in the Car: Detecting Driver Interruptibility Through Monitoring of Peripheral Interactions , 2015, CHI.

[32]  Henriette Cramer,et al.  "Play PRBLMS": Identifying and Correcting Less Accessible Content in Voice Interfaces , 2018, CHI.

[33]  Anind K. Dey,et al.  Assessing demand for intelligibility in context-aware applications , 2009, UbiComp.

[34]  Mark Bilandzic,et al.  Bringing Transparency Design into Practice , 2018, IUI.

[35]  Seth Flaxman,et al.  European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..

[36]  C. Nass,et al.  Machines and Mindlessness , 2000 .

[37]  Steve Whittaker,et al.  On Being Told How We Feel , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[38]  E. Schegloff Repair After Next Turn: The Last Structurally Provided Defense of Intersubjectivity in Conversation , 1992, American Journal of Sociology.

[39]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[40]  N. McGlynn Thinking fast and slow. , 2014, Australian veterinary journal.

[41]  Weng-Keen Wong,et al.  Principles of Explanatory Debugging to Personalize Interactive Machine Learning , 2015, IUI.

[42]  Regina A. Pomranky,et al.  The role of trust in automation reliance , 2003, Int. J. Hum. Comput. Stud..

[43]  Jure Leskovec,et al.  Interpretable & Explorable Approximations of Black Box Models , 2017, ArXiv.

[44]  René F. Kizilcec How Much Information?: Effects of Transparency on Trust in an Algorithmic Interface , 2016, CHI.

[45]  Siobhan Chapman Logic and Conversation , 2005 .

[46]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[47]  Jun Zhao,et al.  'It's Reducing a Human Being to a Percentage': Perceptions of Justice in Algorithmic Decisions , 2018, CHI.

[48]  John T. Cacioppo,et al.  The Elaboration Likelihood Model of Persuasion , 1986, Advances in Experimental Social Psychology.

[49]  Cory Maloney Mathematics as a Tool of Manipulation in Modern Society. Review of the book by Cathy O’Neil «Weapons of Math Destruction. How Big Data Increases Inequality and Threatens Democracy» , 2017 .

[50]  Nitesh V. Chawla,et al.  MOOC Dropout Prediction: Lessons Learned from Making Pipelines Interpretable , 2017, WWW.

[51]  Jimeng Sun,et al.  RetainVis: Visual Analytics with Interpretable and Interactive Recurrent Neural Networks on Electronic Medical Records , 2018, IEEE Transactions on Visualization and Computer Graphics.

[52]  N. Moray,et al.  Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. , 1996, Ergonomics.

[53]  D. Kahneman Thinking, Fast and Slow , 2011 .

[54]  J. Pennebaker,et al.  Language use of depressed and depression-vulnerable college students , 2004 .

[55]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[56]  Konrad Tollmar,et al.  Health Mashups: Presenting Statistical Patterns between Wellbeing Data and Context in Natural Language to Promote Behavior Change , 2013, TCHI.

[57]  Jeremy P. Birnholtz,et al.  How People Form Folk Theories of Social Media Feeds and What it Means for How We Study Self-Presentation , 2018, CHI.

[58]  Steve Whittaker,et al.  Echoes from the past: how technology mediated reflection improves well-being , 2013, CHI.

[59]  Martin F. Porter,et al.  An algorithm for suffix stripping , 1997, Program.

[60]  Wanda Pratt,et al.  Understanding quantified-selfers' practices in collecting and exploring personal data , 2014, CHI.

[61]  Enzo Grossi,et al.  Development and validation of the short version of the Psychological General Well-Being Index (PGWB-S) , 2006, Health and quality of life outcomes.

[62]  Jenna Wiens,et al.  Machine Learning for Healthcare: On the Verge of a Major Shift in Healthcare Epidemiology , 2018, Clinical infectious diseases : an official publication of the Infectious Diseases Society of America.

[63]  Sarvapali D. Ramchurn,et al.  Seeing (Movement) is Believing: The Effect of Motion on Perception of Automatic Systems Performance , 2021, Hum. Comput. Interact..

[64]  J. Pennebaker,et al.  The Secret Life of Pronouns , 2003, Psychological science.

[65]  Akane Sano,et al.  SNAPSHOT Expose : Stage Based and Social Theory Based Applications to Reduce Stress and Improve Wellbeing , 2016 .

[66]  K. Borgwardt,et al.  Machine Learning in Medicine , 2015, Mach. Learn. under Resour. Constraints Vol. 3.

[67]  Mohan S. Kankanhalli,et al.  Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda , 2018, CHI.

[68]  V. Braun,et al.  Using thematic analysis in psychology , 2006 .

[69]  Mary L. Cummings,et al.  Automation Bias in Intelligent Time Critical Decision Support Systems , 2004 .

[70]  H. Garfinkel Studies in Ethnomethodology , 1968 .

[71]  Enzo Grossi,et al.  Psychological General Well-Being Index--Short Version , 2017 .

[72]  Eric Harslem,et al.  Designing the STAR User Interface , 1987, ECICS.

[73]  Rachael Tatman,et al.  Gender and Dialect Bias in YouTube’s Automatic Captions , 2017, EthNLP@EACL.

[74]  Eli Pariser,et al.  The Filter Bubble: What the Internet Is Hiding from You , 2011 .

[75]  Steve Whittaker,et al.  Mood modeling: accuracy depends on active logging and reflection , 2018, Personal and Ubiquitous Computing.

[76]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[77]  Anind K. Dey,et al.  Why and why not explanations improve the intelligibility of context-aware intelligent systems , 2009, CHI.

[78]  W. Keith Edwards,et al.  Intelligibility and Accountability: Human Considerations in Context-Aware Systems , 2001, Hum. Comput. Interact..

[79]  Ira Assent,et al.  Explaining Outliers by Subspace Separability , 2013, 2013 IEEE 13th International Conference on Data Mining.

[80]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[81]  Eli Pariser FILTER BUBBLE: Wie wir im Internet entmündigt werden , 2012 .

[82]  Daniel McDuff,et al.  AffectAura: an intelligent system for emotional memory , 2012, CHI.

[83]  Bernt Schiele,et al.  Towards improving trust in context-aware systems by displaying system confidence , 2005, Mobile HCI.

[84]  L J Skitka,et al.  Automation bias: decision making and performance in high-tech cockpits. , 1997, The International journal of aviation psychology.

[85]  Karrie Karahalios,et al.  Communicating Algorithmic Process in Online Behavioral Advertising , 2018, CHI.

[86]  John M. Carroll,et al.  Training wheels in a user interface , 1984, CACM.

[87]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[88]  Klaus-Robert Müller,et al.  "What is relevant in a text document?": An interpretable machine learning approach , 2016, PloS one.

[89]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[90]  Johannes Gehrke,et al.  Intelligible models for classification and regression , 2012, KDD.