Using Clicks as Implicit Judgments: Expectations Versus Observations

Clickthrough data has been the subject of increasing popularity as an implicit indicator of user feedback. Previous analysis has suggested that user click behaviour is subject to a quality bias--that is, users click at different rank positions when viewing effective search results than when viewing less effective search results. Based on this observation, it should be possible to use click data to infer the quality of the underlying search system. In this paper we carry out a user study to systematically investigate how click behaviour changes for different levels of search system effectiveness as measured by information retrieval performance metrics. Our results show that click behaviour does not vary systematically with the quality of search results. However, click behaviour does vary significantly between individual users, and between search topics. This suggests that using direct click behaviour--click rank and click frequency--to infer the quality of the underlying search system is problematic. Further analysis of our user click data indicates that the correspondence between clicks in a search result list and subsequent confirmation that the clicked resource is actually relevant is low. Using clicks as an implicit indication of relevance should therefore be done with caution.

[1]  Susan T. Dumais,et al.  Learning user interaction models for predicting web search result preferences , 2006, SIGIR.

[2]  Filip Radlinski,et al.  Active exploration for learning rankings from clickthrough data , 2007, KDD '07.

[3]  Charles L. A. Clarke,et al.  Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval , 2007, SIGIR 2007.

[4]  Donna K. Harman,et al.  The TREC Test Collections , 2005 .

[5]  Nick Craswell,et al.  Random walks on the click graph , 2007, SIGIR.

[6]  Falk Scholer,et al.  User performance versus precision measures for simple search tasks , 2006, SIGIR.

[7]  José Luis Vicedo González,et al.  TREC: Experiment and evaluation in information retrieval , 2007, J. Assoc. Inf. Sci. Technol..

[8]  Gobinda G. Chowdhury,et al.  TREC: Experiment and Evaluation in Information Retrieval , 2007 .

[9]  Hugh E. Williams,et al.  Fast generation of result snippets in web search , 2007, SIGIR.

[10]  Thorsten Joachims,et al.  Accurately Interpreting Clickthrough Data as Implicit Feedback , 2017 .

[11]  Thorsten Joachims,et al.  Optimizing search engines using clickthrough data , 2002, KDD.

[12]  Kotagiri Ramamohanarao,et al.  Long-Term Learning for Web Search Engines , 2002, PKDD.

[13]  Falk Scholer,et al.  Examining the Pseudo-Standard Web Search Engine Results Page , 2006, Aust. J. Intell. Inf. Process. Syst..

[14]  James Allan,et al.  When will information retrieval be "good enough"? , 2005, SIGIR '05.

[15]  Filip Radlinski,et al.  Query chains: learning to rank from implicit feedback , 2005, KDD '05.

[16]  Ellen M. Voorhees,et al.  Retrieval System Evaluation , 2005 .

[17]  Thorsten Joachims,et al.  Accurately interpreting clickthrough data as implicit feedback , 2005, SIGIR '05.

[18]  Filip Radlinski,et al.  Evaluating the accuracy of implicit feedback from clicks and query reformulations in Web search , 2007, TOIS.

[19]  Susan T. Dumais,et al.  Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval , 2004, SIGIR 2004.

[20]  Peter Bailey,et al.  Engineering a multi-purpose test collection for Web retrieval experiments , 2003, Inf. Process. Manag..

[21]  Ellen M. Voorhees,et al.  TREC: Experiment and Evaluation in Information Retrieval (Digital Libraries and Electronic Publishing) , 2005 .

[22]  Alistair Moffat,et al.  Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval , 2005, SIGIR 2005.

[23]  Eric Brill,et al.  Improving web search ranking by incorporating user behavior information , 2006, SIGIR.

[24]  Steve Fox,et al.  Evaluating implicit measures to improve web search , 2005, TOIS.