Bayesian Ranker Comparison Based on Historical User Interactions
暂无分享,去创建一个
[1] ChengXiang Zhai,et al. Evaluation of methods for relative comparison of retrieval systems based on clickthroughs , 2009, CIKM.
[2] Christopher M. Bishop,et al. Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .
[3] Christos Faloutsos,et al. Tailoring click models to user goals , 2009, WSCD '09.
[4] Cyril W. Cleverdon,et al. Aslib Cranfield research project - Factors determining the performance of indexing systems; Volume 1, Design; Part 2, Appendices , 1966 .
[5] Ron Kohavi,et al. Controlled experiments on the web: survey and practical guide , 2009, Data Mining and Knowledge Discovery.
[6] Chao Liu,et al. Efficient multiple-click models in web search , 2009, WSDM '09.
[7] Cyril W. Cleverdon,et al. Factors determining the performance of indexing systems , 1966 .
[8] Ben Carterette,et al. System effectiveness, user models, and user utility: a conceptual framework for investigation , 2011, SIGIR.
[9] Mark Sanderson,et al. Test Collection Based Evaluation of Information Retrieval Systems , 2010, Found. Trends Inf. Retr..
[10] Tao Qin,et al. LETOR: A benchmark collection for research on learning to rank for information retrieval , 2010, Information Retrieval.
[11] Filip Radlinski,et al. Relevance and Effort: An Analysis of Document Utility , 2014, CIKM.
[12] Filip Radlinski,et al. Large-scale validation and analysis of interleaved search evaluation , 2012, TOIS.
[13] Filip Radlinski,et al. On caption bias in interleaving experiments , 2012, CIKM '12.
[14] Olivier Chapelle,et al. A dynamic bayesian network click model for web search ranking , 2009, WWW '09.
[15] Thorsten Joachims,et al. Interactively optimizing information retrieval systems as a dueling bandits problem , 2009, ICML '09.
[16] Filip Radlinski,et al. Comparing the sensitivity of information retrieval metrics , 2010, SIGIR.
[17] Katja Hofmann,et al. Reusing historical interaction data for faster online learning to rank for IR , 2013, DIR.
[18] M. de Rijke,et al. Multileaved Comparisons for Fast Online Evaluation , 2014, CIKM.
[19] Thorsten Joachims,et al. Optimizing search engines using clickthrough data , 2002, KDD.
[20] Milad Shokouhi,et al. Expected browsing utility for web search evaluation , 2010, CIKM.
[21] M. de Rijke,et al. Click model-based information retrieval metrics , 2013, SIGIR.
[22] Benjamin Piwowarski,et al. A user browsing model to predict search engine click data from past observations. , 2008, SIGIR '08.
[23] Filip Radlinski,et al. How does clickthrough data reflect retrieval quality? , 2008, CIKM '08.
[24] Olivier Chapelle,et al. Expected reciprocal rank for graded relevance , 2009, CIKM.
[25] Hang Li,et al. AdaRank: a boosting algorithm for information retrieval , 2007, SIGIR.
[26] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[27] Katja Hofmann,et al. Evaluating aggregated search using interleaving , 2013, CIKM.