Oopsy-daisy: failure stories in quantitative evaluation studies for visualizations

Designing, conducting, and interpreting evaluation studies with human participants is challenging. While researchers in cognitive psychology, social science, and human-computer interaction view competence in evaluation study methodology a key job skill, it is only recently that visualization researchers have begun to feel the need to learn this skill as well. Acquiring such competence is a lengthy and difficult process fraught with much trial and error. Recent work on patterns for visualization evaluation is now providing much-needed best practices for how to evaluate a visualization technique with human participants. However, negative examples of evaluation methods that fail, yield no usable results, or simply do not work are still missing, mainly because of the difficulty and lack of incentive for publishing negative results or failed research. In this paper, we take the position that there are many good ideas with the best intentions for how to evaluate a visualization tool that simply do not work. We call upon the community to help collect these negative examples in order to show the other side of the coin: what not to do when trying to evaluate visualization.

[1]  Tamara Munzner,et al.  A Nested Model for Visualization Design and Validation , 2009, IEEE Transactions on Visualization and Computer Graphics.

[2]  Ralph Johnson,et al.  design patterns elements of reusable object oriented software , 2019 .

[3]  M. Sheelagh T. Carpendale,et al.  Grounded evaluation of information visualizations , 2008, BELIV.

[4]  Lorrie Faith Cranor,et al.  Are your participants gaming the system?: screening mechanical turk workers , 2010, CHI.

[5]  Chaomei Chen,et al.  Empirical studies of information visualization: a meta-analysis , 2000, Int. J. Hum. Comput. Stud..

[6]  David D. Redell,et al.  An evaluation of the ninth SOSP submissions or how (and how not) to write a good systems paper , 1983, ACM SIGOPS Oper. Syst. Rev..

[7]  Jacob Cohen,et al.  A power primer. , 1992, Psychological bulletin.

[8]  Panagiotis G. Ipeirotis Analyzing the Amazon Mechanical Turk marketplace , 2010, XRDS.

[9]  Sung-Hee Kim,et al.  How to filter out random clickers in a crowdsourcing-based study? , 2012, BELIV '12.

[10]  Tamara Munzner,et al.  Process and Pitfalls in Writing Information Visualization Research Papers , 2008, Information Visualization.

[11]  Ben Shneiderman,et al.  Strategies for evaluating information visualization tools: multi-dimensional in-depth long-term case studies , 2006, BELIV '06.

[12]  Catherine Plaisant,et al.  The challenge of information visualization evaluation , 2004, AVI.

[13]  Kasper Hornbæk,et al.  Some Whys and Hows of Experiments in Human-Computer Interaction , 2013, Found. Trends Hum. Comput. Interact..

[14]  Niklas Elmqvist,et al.  Patterns for visualization evaluation , 2012, BELIV '12.

[15]  Priti Shah,et al.  Bar and Line Graph Comprehension: An Interaction of Top-Down and Bottom-Up Processes , 2011, Top. Cogn. Sci..

[16]  Jeffrey Heer,et al.  Strategies for crowdsourcing social data analysis , 2012, CHI.

[17]  Jesse J. Chandler,et al.  Inside the Turk , 2014 .

[18]  Jean-Daniel Fekete,et al.  Task taxonomy for graph visualization , 2006, BELIV '06.

[19]  Marina Daecher,et al.  Experimental Human Computer Interaction A Practical Guide With Visual Examples , 2016 .

[20]  Jean-Daniel Fekete,et al.  A Principled Way of Assessing Visualization Literacy , 2014, IEEE Transactions on Visualization and Computer Graphics.

[21]  Chris North,et al.  Information Visualization , 2008, Lecture Notes in Computer Science.

[22]  M. Sheelagh T. Carpendale,et al.  Empirical Studies in Information Visualization: Seven Scenarios , 2012, IEEE Transactions on Visualization and Computer Graphics.

[23]  Tamara Munzner,et al.  Increasing the utility of quantitative empirical studies for meta-analysis , 2008, BELIV '08.

[24]  Sung-Hee Kim,et al.  Does an Eye Tracker Tell the Truth about Visualizations?: Findings while Investigating Visualizations for Decision Making , 2012, IEEE Transactions on Visualization and Computer Graphics.

[25]  PlaisantCatherine,et al.  Empirical Studies in Information Visualization , 2012 .

[26]  David S. Ebert,et al.  Applied visual analytics for economic decision-making , 2008, 2008 IEEE Symposium on Visual Analytics Science and Technology.

[27]  Thomas J. Mowbray,et al.  AntiPatterns: Refactoring Software, Architectures, and Projects in Crisis , 1998 .

[28]  M. Sheelagh T. Carpendale,et al.  Evaluating Information Visualizations , 2008, Information Visualization.