WHOSE - A Tool for Whole-Session Analysis in IIR

One of the main challenges in Interactive Information Retrieval (IIR) evaluation is the development and application of re-usable tools that allow researchers to analyze search behavior of real users in different environments and different domains, but with comparable results. Furthermore, IIR recently focuses more on the analysis of whole sessions, which includes all user interactions that are carried out within a session but also across several sessions by the same user. Some frameworks have already been proposed for the evaluation of controlled experiments in IIR, but yet no framework is available for interactive evaluation of search behavior from real-world information retrieval (IR) systems with real users. In this paper we present a framework for whole-session evaluation that can also utilize these uncontrolled data sets. The logging component can easily be integrated into real-world IR systems for generating and analyzing new log data. Furthermore, due to a supplementary mapping it is also possible to analyze existing log data. For every IR system different actions and filters can be defined. This allows system operators and researchers to use the framework for the analysis of user search behavior in their IR systems and to compare it with others. Using a graphical user interface they have the possibility to interactively explore the data set from a broad overview down to individual sessions.

[1]  Ben Shneiderman,et al.  The eyes have it: a task by data type taxonomy for information visualizations , 1996, Proceedings 1996 IEEE Symposium on Visual Languages.

[2]  Ryen W. White,et al.  Modeling and analysis of cross-session search tasks , 2011, SIGIR.

[3]  Nicholas J. Belkin,et al.  Personalization of search results using interaction behaviors in search sessions , 2012, SIGIR '12.

[4]  Norbert Fuhr,et al.  A probability ranking principle for interactive information retrieval , 2008, Information Retrieval.

[5]  Ellen M. Voorhees,et al.  TREC: Experiment and Evaluation in Information Retrieval (Digital Libraries and Electronic Publishing) , 2005 .

[6]  Jeffrey Heer,et al.  What did they do? understanding clickstreams with the WebQuilt visualization system , 2002, AVI '02.

[7]  Jacek Gwizdka,et al.  A User-Centered Experiment and Logging Framework for Interactive Information Retrieval , 2009, UIIR@SIGIR.

[8]  Jean Scholtz,et al.  VISVIP: 3D visualization of paths through web sites , 1999, Proceedings. Tenth International Workshop on Database and Expert Systems Applications. DEXA 99.

[9]  Rosie Jones,et al.  Beyond the session timeout: automatic hierarchical segmentation of search topics in query logs , 2008, CIKM '08.

[10]  Kwan-Liu Ma,et al.  Visual analysis of massive web session data , 2012, IEEE Symposium on Large Data Analysis and Visualization (LDAV).

[11]  Tamara Munzner,et al.  Session Viewer: Visual Exploratory Analysis of Web Session Logs , 2007, 2007 IEEE Symposium on Visual Analytics Science and Technology.

[12]  J. E. Pitkow,et al.  WebViz: A tool for WWW access log analysis , 1994 .

[13]  Susan T. Dumais,et al.  Evaluation Challenges and Directions for Information-Seeking Support Systems , 2009, Computer.

[14]  Yang Song,et al.  Evaluating the effectiveness of search task trails , 2012, WWW.

[15]  Norbert Fuhr,et al.  Quantitative Analysis of Search Sessions Enhanced by Gaze Tracking with Dynamic Areas of Interest , 2012, TPDL.

[16]  Leif Azzopardi,et al.  SCAMP: a tool for conducting interactive information retrieval experiments , 2012, IIiX.

[17]  Elaine Toms,et al.  WiIRE: the Web interactive information retrieval experimentation system prototype , 2004, Inf. Process. Manag..

[18]  Elaine Toms,et al.  Building a Common Framework for IIR Evaluation , 2013, CLEF.

[19]  Nicholas J. Belkin,et al.  On the evaluation of interactive information retrieval systems , 2010 .

[20]  Norbert Fuhr,et al.  ezDL: An Interactive Search and Evaluation System , 2012, OSIR@SIGIR.

[21]  Marcia J. Bates,et al.  Where should the person stop and the information search interface start? , 1990, Inf. Process. Manag..

[22]  Ben Carterette,et al.  Evaluating multi-query sessions , 2011, SIGIR.

[23]  Kalervo Järvelin,et al.  Time drives interaction: simulating sessions in diverse searching environments , 2012, SIGIR '12.

[24]  Steve Fox,et al.  Evaluating implicit measures to improve web search , 2005, TOIS.