Characterizing the usability of interactive applications through query log analysis

People routinely rely on Internet search engines to support their use of interactive systems: they issue queries to learn how to accomplish tasks, troubleshoot problems, and otherwise educate themselves on products. Given this common behavior, we argue that search query logs can usefully augment traditional usability methods by revealing the primary tasks and needs of a product's user population. We term this use of search query logs CUTS - characterizing usability through search. In this paper, we introduce CUTS and describe an automated process for harvesting, ordering, labeling, filtering, and grouping search queries related to a given product. Importantly, this data set can be assembled in minutes, is timely, has a high degree of ecological validity, and is arguably less prone to self-selection bias than data gathered via traditional usability methods. We demonstrate the utility of this approach by applying it to a number of popular software and hardware systems.

[1]  Ricardo A. Baeza-Yates,et al.  Extracting semantic relations from query logs , 2007, KDD '07.

[2]  Charles L. A. Clarke,et al.  Characterizing large-scale use of a direct manipulation application in the wild , 2010, Graphics Interface.

[3]  Efthimis N. Efthimiadis,et al.  Analyzing and evaluating query reformulation strategies in web search logs , 2009, CIKM.

[4]  Ji-Rong Wen,et al.  WWW 2007 / Track: Search Session: Personalization A Largescale Evaluation and Analysis of Personalized Search Strategies ABSTRACT , 2022 .

[5]  Scott R. Klemmer,et al.  Example-centric programming: integrating web search into the development environment , 2010, CHI.

[6]  Elizabeth D. Mynatt,et al.  Supporting experimentation with Side-Views , 2002, CACM.

[7]  Jakob Nielsen,et al.  Heuristic evaluation of user interfaces , 1990, CHI '90.

[8]  Daniel E. Rose,et al.  Understanding user goals in web search , 2004, WWW '04.

[9]  George A. Miller,et al.  WordNet: A Lexical Database for English , 1995, HLT.

[10]  Matthew Richardson,et al.  Learning about the world through long-term query logs , 2008, TWEB.

[11]  Wagner Meira,et al.  Rank-preserving two-level caching for scalable search engines , 2001, SIGIR '01.

[12]  Carolyn Watters,et al.  A field study characterizing Web-based information-seeking tasks , 2007 .

[13]  Ziv Bar-Yossef,et al.  Mining search engine query logs via suggestion sampling , 2008, Proc. VLDB Endow..

[14]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[15]  Eser Kandogan,et al.  Field studies of computer system administrators: analysis of system management tools and practices , 2004, CSCW.

[16]  David F. Redmiles,et al.  Extracting usability information from user interface events , 2000, CSUR.

[17]  Robin Jeffries,et al.  Undo and erase events as indicators of usability problems , 2009, CHI.

[18]  Jeremy Ginsberg,et al.  Detecting influenza epidemics using search engine query data , 2009, Nature.

[19]  Andrei Broder,et al.  A taxonomy of web search , 2002, SIGF.

[20]  Anne Aula,et al.  How does search behavior change as search becomes more difficult? , 2010, CHI.

[21]  Anselm L. Strauss,et al.  Basics of qualitative research : techniques and procedures for developing grounded theory , 1998 .

[22]  Lillian Lee,et al.  Opinion Mining and Sentiment Analysis , 2008, Found. Trends Inf. Retr..

[23]  Scott E. Hudson,et al.  Dynamic detection of novice vs. skilled use without a task model , 2007, CHI.

[24]  Philip J. Guo,et al.  Two studies of opportunistic programming: interleaving web foraging, learning, and writing code , 2009, CHI.