Versus - A tool for evaluating visualizations and image quality using a 2AFC methodology

Abstract Novel visualization methods and strategies are necessary to cope with the deluge of datasets present in any scientific field to make discoveries and find answers to previously unanswered questions. These methods and strategies should not only present scientific findings as images in a concise way but also need to be effective and expressive, which often remain untested. Here, we present Versus, a tool to enable easy image quality assessment and image ranking, utilizing a two-alternative forced choice methodology (2AFC) and an efficient ranking algorithm based on a binary search. The tool provides a systematic way of setting up evaluation experiments via the web without the necessity to install any additional software or require any programming skills. Furthermore, Versus can easily interface with crowdsourcing platforms, such as Amazon’s Mechanical Turk, or can be used as a stand-alone system to carry out evaluations with experts. We demonstrate the use of Versus by means of an image evaluation study, aiming to determine if hue, saturation, brightness, and texture are good indicators of uncertainty in three-dimensional protein structures. Drawing from the power of crowdsourcing, we argue that there is demand and also great potential for this tool to become a standard for simple and fast image evaluations, with the aim to test the effectiveness and expressiveness of scientific visualizations.

[1]  Matthew A. Hibbs,et al.  Visualization of omics data for systems biology , 2010, Nature Methods.

[2]  Fabian A. Buske,et al.  Aquaria: simplifying discovery and insight from protein structures , 2015, Nature Methods.

[3]  M. Sheelagh T. Carpendale,et al.  Grounded evaluation of information visualizations , 2008, BELIV.

[4]  Gerrit C. van der Veer,et al.  BREF: BRowser-based Evaluation Framework , 2017, Conference on Designing Interactive Systems.

[5]  Steven Franconeri,et al.  Ranking Visualizations of Correlation Using Weber's Law , 2014, IEEE Transactions on Visualization and Computer Graphics.

[6]  Chris North,et al.  Toward measuring visualization insight , 2006, IEEE Computer Graphics and Applications.

[7]  Timo Ropinski,et al.  A crowdsourcing system for integrated and reproducible evaluation in scientific visualization , 2016, 2016 IEEE Pacific Visualization Symposium (PacificVis).

[8]  M. Sheelagh T. Carpendale,et al.  Empirical Studies in Information Visualization: Seven Scenarios , 2012, IEEE Transactions on Visualization and Computer Graphics.

[9]  Ivan V. Bajic,et al.  A platform for subjective image quality evaluation on mobile devices , 2016, 2016 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE).

[10]  Olivier Chapuis,et al.  Touchstone: exploratory design of experiments , 2007, CHI.

[11]  Adam Finkelstein,et al.  Perceptual models of viewpoint preference , 2011, TOGS.

[12]  Wouter Meulemans,et al.  Map LineUps: Effects of spatial structure on graphical inference , 2017, IEEE Transactions on Visualization and Computer Graphics.

[13]  M. Sheelagh T. Carpendale,et al.  Evaluating Information Visualizations , 2008, Information Visualization.

[14]  P E Bourne,et al.  The Protein Data Bank. , 2002, Nucleic acids research.

[15]  James P. Ahrens,et al.  ETK: An Evaluation Toolkit for Visualization User Studies , 2017, EuroVis.

[16]  Catherine Plaisant,et al.  The challenge of information visualization evaluation , 2004, AVI.

[17]  María Martín,et al.  Activities at the Universal Protein Resource (UniProt) , 2013, Nucleic Acids Res..

[18]  Harri Siirtola,et al.  Using Gaze Data in Evaluating Interactive Visualizations , 2009, HCIV.

[19]  D. Amnon Silverstein,et al.  Efficient method for paired comparison , 2001, J. Electronic Imaging.

[20]  Jeffrey Heer,et al.  Crowdsourcing graphical perception: using mechanical turk to assess visualization design , 2010, CHI.

[21]  Rafal Mantiuk,et al.  Comparison of Four Subjective Methods for Image Quality Assessment , 2012, Comput. Graph. Forum.

[22]  W. Cleveland,et al.  Graphical Perception: Theory, Experimentation, and Application to the Development of Graphical Methods , 1984 .

[23]  Radu Jianu,et al.  GraphUnit: Evaluating Interactive Graph Visualizations Using Crowdsourcing , 2015, Comput. Graph. Forum.

[24]  Krzysztof Z. Gajos,et al.  TurkServer: Enabling Synchronous and Longitudinal Online Experiments , 2012, HCOMP@AAAI.

[25]  Fernando González-Ladrón-de-Guevara,et al.  Towards an integrated crowdsourcing definition , 2012, J. Inf. Sci..

[26]  Bongshin Lee,et al.  Information Visualization Evaluation Using Crowdsourcing , 2018, Comput. Graph. Forum.

[27]  Ivan Viola,et al.  Two-Level Approach to Efficient Visualization of Protein Dynamics , 2007, IEEE Transactions on Visualization and Computer Graphics.

[28]  Phuoc Tran-Gia,et al.  Anatomy of a Crowdsourcing Platform - Using the Example of Microworkers.com , 2011, 2011 Fifth International Conference on Innovative Mobile and Internet Services in Ubiquitous Computing.

[29]  Julian Heinrich,et al.  Evaluating the Effectiveness of Color to Convey Alignment Quality in Macromolecular Structures , 2015, 2015 Big Data Visual Analytics (BDVA).

[30]  Dominique Brodbeck,et al.  Research directions in data wrangling: Visualizations and transformations for usable and credible data , 2011, Inf. Vis..

[31]  Wolfgang Aigner,et al.  EvalBench: A Software Library for Visualization Evaluation , 2013, Comput. Graph. Forum.

[32]  Conrad C. Huang,et al.  UCSF Chimera—A visualization system for exploratory research and analysis , 2004, J. Comput. Chem..

[33]  Yang Liu,et al.  Somewhere Over the Rainbow: An Empirical Assessment of Quantitative Colormaps , 2018, CHI.

[34]  Timo Ropinski,et al.  Inviwo - An extensible, multi-purpose visualization framework , 2015, 2015 IEEE Scientific Visualization Conference (SciVis).

[35]  Anton J. Enright,et al.  Visualizing genome and systems biology: technologies, tools, implementation techniques and trends, past, present and future , 2015, GigaScience.

[36]  Victoria Interrante,et al.  User Studies: Why, How, and When? , 2003, IEEE Computer Graphics and Applications.

[37]  Eric Schenk,et al.  Crowdsourcing: What can be Outsourced to the Crowd, and Why ? , 2009 .

[38]  Steven Franconeri,et al.  Perception of Average Value in Multiclass Scatterplots , 2013, IEEE Transactions on Visualization and Computer Graphics.

[39]  Bang Wong,et al.  Visual Representation of Scientific Information , 2011, Science Signaling.

[40]  David H. Laidlaw,et al.  Thoughts on User Studies: Why, How and When , 1993 .

[41]  J. Heinrich,et al.  Evaluating Viewpoint Entropy for Ribbon Representation of Protein Structure , 2016, Comput. Graph. Forum.

[42]  Andreas Butz,et al.  Information visualization evaluation in large companies: Challenges, experiences and recommendations , 2011, Inf. Vis..