A Systematic Review on the Practice of Evaluating Visualization

We present an assessment of the state and historic development of evaluation practices as reported in papers published at the IEEE Visualization conference. Our goal is to reflect on a meta-level about evaluation in our community through a systematic understanding of the characteristics and goals of presented evaluations. For this purpose we conducted a systematic review of ten years of evaluations in the published papers using and extending a coding scheme previously established by Lam et al. [2012]. The results of our review include an overview of the most common evaluation goals in the community, how they evolved over time, and how they contrast or align to those of the IEEE Information Visualization conference. In particular, we found that evaluations specific to assessing resulting images and algorithm performance are the most prevalent (with consistently 80-90% of all papers since 1997). However, especially over the last six years there is a steady increase in evaluation methods that include participants, either by evaluating their performances and subjective feedback or by evaluating their work practices and their improved analysis and reasoning capabilities using visual tools. Up to 2010, this trend in the IEEE Visualization conference was much more pronounced than in the IEEE Information Visualization conference which only showed an increasing percentage of evaluation through user performance and experience testing. Since 2011, however, also papers in IEEE Information Visualization show such an increase of evaluations of work practices and analysis as well as reasoning using visual tools. Further, we found that generally the studies reporting requirements analyses and domain-specific work practices are too informally reported which hinders cross-comparison and lowers external validity.

[1]  Leif D. Nelson,et al.  Data from Paper “False-Positive Psychology: Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant” , 2014 .

[2]  Jonathan W. Decker,et al.  Evaluation of Multivariate Visualization on a Multivariate Task , 2012, IEEE Transactions on Visualization and Computer Graphics.

[3]  Tamara Munzner,et al.  Design Study Methodology: Reflections from the Trenches and the Stacks , 2012, IEEE Transactions on Visualization and Computer Graphics.

[4]  Frans Vos,et al.  WYSIWYP: What You See Is What You Pick , 2012, IEEE Transactions on Visualization and Computer Graphics.

[5]  Klaus Mueller,et al.  Human Computation in Visualization: Using Purpose Driven Games for Robust Evaluation of Visualization Algorithms , 2012, IEEE Transactions on Visualization and Computer Graphics.

[6]  Michael Gleicher,et al.  Why ask why?: considering motivation in visualization evaluation , 2012, BELIV '12.

[7]  Daniel F. Keefe,et al.  Toward mixed method evaluations of scientific visualizations and design process as an evaluation tool , 2012, BELIV '12.

[8]  M. Sheelagh T. Carpendale,et al.  Empirical Studies in Information Visualization: Seven Scenarios , 2012, IEEE Transactions on Visualization and Computer Graphics.

[9]  Valerio Pascucci,et al.  Topology Verification for Isosurface Extraction , 2012, IEEE Transactions on Visualization and Computer Graphics.

[10]  Edward Cutrell,et al.  "Yours is better!": participant response bias in HCI , 2012, CHI.

[11]  Martin Schmettow,et al.  Sample size in usability studies , 2012, Commun. ACM.

[12]  A. Ynnerman,et al.  Multi-Touch Table System for Medical Visualization: Application to Orthopedic Surgery Planning , 2011, IEEE Transactions on Visualization and Computer Graphics.

[13]  R. Kirby,et al.  GPU-Based Interactive Cut-Surface Extraction From High-Order Finite Element Fields , 2011, IEEE Transactions on Visualization and Computer Graphics.

[14]  Jason Dykes,et al.  BallotMaps: Detecting Name Bias in Alphabetically Ordered Ballot Papers , 2011, IEEE Transactions on Visualization and Computer Graphics.

[15]  Hans Hagen,et al.  Collaborative visualization: Definition, challenges, and research agenda , 2011, Inf. Vis..

[16]  G. Cumming Understanding the New Statistics: Effect Sizes, Confidence Intervals, and Meta-Analysis , 2011 .

[17]  Andreas Butz,et al.  Information visualization evaluation in large companies: Challenges, experiences and recommendations , 2011, Inf. Vis..

[18]  Kate E Decleene,et al.  Publication Manual of the American Psychological Association , 2011 .

[19]  Barry A. T. Brown,et al.  Into the wild: challenges and opportunities for field trial methods , 2011, CHI.

[20]  Charl P. Botha,et al.  Articulated Planar Reformation for Change Visualization in Small Animal Imaging , 2010, IEEE Transactions on Visualization and Computer Graphics.

[21]  Eduard Gröller,et al.  World Lines , 2010, IEEE Transactions on Visualization and Computer Graphics.

[22]  G. Kindlmann,et al.  Superquadric Glyphs for Symmetric Second-Order Tensors , 2010, IEEE Transactions on Visualization and Computer Graphics.

[23]  Camilla Forsell,et al.  A Guide to Scientific Evaluation in Information Visualization , 2010, 2010 14th International Conference Information Visualisation.

[24]  David L Donoho,et al.  An invitation to reproducible computational research. , 2010, Biostatistics.

[25]  Robert Michael Kirby,et al.  Verifiable Visualization for Isosurface Extraction , 2009, IEEE Transactions on Visualization and Computer Graphics.

[26]  Stefan Bruckner,et al.  BrainGazer - Visual Queries for Neurobiology Research , 2009, IEEE Transactions on Visualization and Computer Graphics.

[27]  Tamara Munzner,et al.  MizBee: A Multiscale Synteny Browser , 2009, IEEE Transactions on Visualization and Computer Graphics.

[28]  Tamara Munzner,et al.  A Nested Model for Visualization Design and Validation , 2009, IEEE Transactions on Visualization and Computer Graphics.

[29]  B. Shneiderman,et al.  Integrating Statistics and Visualization for Exploratory Power: From Long-Term Case Studies to Design Guidelines , 2009, IEEE Computer Graphics and Applications.

[30]  M. Orne Demand Characteristics and the Concept of Quasi-Controls1 , 2009 .

[31]  Sharan B. Merriam,et al.  Qualitative Research: A Guide to Design and Implementation , 2009 .

[32]  Hans-Peter Seidel,et al.  Estimating Crossing Fibers: A Tensor Decomposition Approach , 2008, IEEE Transactions on Visualization and Computer Graphics.

[33]  Cullen D. Jackson,et al.  Using Visual Design Experts in Critique-Based Evaluation of 2D Vector Visualization Methods , 2008, IEEE Transactions on Visualization and Computer Graphics.

[34]  Daniel F. Keefe,et al.  Scientific Sketching for Collaborative VR Visualization Design , 2008, IEEE Transactions on Visualization and Computer Graphics.

[35]  Fritz Drury,et al.  Visualization Criticism , 2008, IEEE Computer Graphics and Applications.

[36]  William Buxton,et al.  Usability evaluation considered harmful (some of the time) , 2008, CHI.

[37]  M. Sheelagh T. Carpendale,et al.  Evaluating Information Visualizations , 2008, Information Visualization.

[38]  Robert Kosara,et al.  Visualization Criticism - The Missing Link Between Information Visualization and Art , 2007, 2007 11th International Conference Information Visualization (IV '07).

[39]  Martin Isenburg,et al.  Fast and Efficient Compression of Floating-Point Data , 2006, IEEE Transactions on Visualization and Computer Graphics.

[40]  B. Green,et al.  Writing narrative literature reviews for peer-reviewed journals: secrets of the trade. , 2006, Journal of chiropractic medicine.

[41]  Jarke J. van Wijk,et al.  Views on Visualization , 2006, IEEE Transactions on Visualization and Computer Graphics.

[42]  Keith Andrews,et al.  Evaluating information visualisations , 2006, BELIV '06.

[43]  Ben Shneiderman,et al.  Strategies for evaluating information visualization tools: multi-dimensional in-depth long-term case studies , 2006, BELIV '06.

[44]  Ben Shneiderman,et al.  Knowledge discovery in high-dimensional data: case studies and a user survey for the rank-by-feature framework , 2006, IEEE Transactions on Visualization and Computer Graphics.

[45]  J. V. Wijk The value of visualization , 2005, VIS 05. IEEE Visualization, 2005..

[46]  Melanie Tory,et al.  Evaluating Visualizations: Do Expert Reviews Work? , 2005, IEEE Computer Graphics and Applications.

[47]  T.L.J. Ferris,et al.  Participatory IT Design: Designing for Business and Workplace Realities , 2005, IEEE Transactions on Professional Communication.

[48]  David H. Laidlaw,et al.  Artistic Collaboration in Designing VR Visualizations , 2005, IEEE Computer Graphics and Applications.

[49]  J. Tinsley Oden,et al.  Verification and validation in computational engineering and science: basic concepts , 2004 .

[50]  Chris R. Johnson Top Scientific Visualization Research Problems , 2004, IEEE Computer Graphics and Applications.

[51]  Catherine Plaisant,et al.  The challenge of information visualization evaluation , 2004, AVI.

[52]  Daniel Acevedo Feliz,et al.  Designer-critiqued comparison of 2D vector visualization methods: a pilot study , 2003, SIGGRAPH '03.

[53]  Victoria Interrante,et al.  User Studies: Why, How, and When? , 2003, IEEE Computer Graphics and Applications.

[54]  Graham J Hole,et al.  How to Design and Report Experiments , 2002 .

[55]  R. Barker Bausell,et al.  Power analysis for experimental research , 2002 .

[56]  Yvonne Rogers,et al.  Interaction Design: Beyond Human-Computer Interaction , 2002 .

[57]  Michael Lewis,et al.  Evaluating visualizations: using a taxonomic guide , 2000, Int. J. Hum. Comput. Stud..

[58]  Chaomei Chen,et al.  Empirical studies of information visualization: a meta-analysis , 2000, Int. J. Hum. Comput. Stud..

[59]  Mary Czerwinski,et al.  Empirical evaluation of information visualizations: an introduction , 2000, Int. J. Hum. Comput. Stud..

[60]  Ben Shneiderman,et al.  Designing the User Interface: Strategies for Effective Human-Computer Interaction , 1998 .

[61]  J. McGrath,et al.  Methodology matters: doing research in the behavioral and social sciences , 1995 .

[62]  Al Globus,et al.  Evaluation of visualization software , 1995, COMG.

[63]  Jacob Cohen The earth is round (p < .05) , 1994 .

[64]  Charbel Farhat,et al.  Validation, verification and evaluation , 1994, Proceedings Visualization '94.

[65]  Jakob Nielsen,et al.  Heuristic evaluation of user interfaces , 1990, CHI '90.

[66]  Gary James Jason,et al.  The Logic of Scientific Discovery , 1988 .

[67]  M. J. Intons-Peterson,et al.  Imagery paradigms: how vulnerable are they to experimenters' expectations? , 1983, Journal of experimental psychology. Human perception and performance.

[68]  Daniel A. Keim,et al.  Mastering the Information Age - Solving Problems with Visual Analytics , 2010 .

[69]  Robert Michael Kirby,et al.  The Need for Verifiable Visualization , 2008, IEEE Computer Graphics and Applications.

[70]  T. Rhyne,et al.  Toward measuring visualization insight , 2006, IEEE Computer Graphics and Applications.

[71]  Melanie Tory,et al.  Human factors in visualization research , 2004, IEEE Transactions on Visualization and Computer Graphics.

[72]  Rex B. Kline,et al.  What's Wrong With Statistical Tests--And Where We Go From Here. , 2004 .

[73]  Klaus Krippendorff,et al.  Content Analysis: An Introduction to Its Methodology , 1980 .