Sensitive and Scalable Online Evaluation with Theoretical Guarantees
暂无分享,去创建一个
[1] M. de Rijke,et al. Click Models for Web Search , 2015, Click Models for Web Search.
[2] Filip Radlinski,et al. Optimized interleaving for online retrieval evaluation , 2013, WSDM.
[3] Yue Gao,et al. Learning more powerful test statistics for click-based retrieval evaluation , 2010, SIGIR.
[4] Filip Radlinski,et al. Online Evaluation for Information Retrieval , 2016, Found. Trends Inf. Retr..
[5] Filip Radlinski,et al. How does clickthrough data reflect retrieval quality? , 2008, CIKM '08.
[6] Tao Qin,et al. LETOR: Benchmark Dataset for Research on Learning to Rank for Information Retrieval , 2007 .
[7] David Hawking,et al. Overview of the TREC 2003 Web Track , 2003, TREC.
[8] Filip Radlinski,et al. Large-scale validation and analysis of interleaved search evaluation , 2012, TOIS.
[9] M. de Rijke,et al. Multileave Gradient Descent for Fast Online Learning to Rank , 2016, WSDM.
[10] Katja Hofmann,et al. A probabilistic method for inferring preferences from clicks , 2011, CIKM '11.
[11] Ingemar J. Cox,et al. An Improved Multileaving Algorithm for Online Ranker Evaluation , 2016, SIGIR.
[12] Ron Kohavi,et al. Online controlled experiments at large scale , 2013, KDD.
[13] Tao Qin,et al. Introducing LETOR 4.0 Datasets , 2013, ArXiv.
[14] M. de Rijke,et al. Multileaved Comparisons for Fast Online Evaluation , 2014, CIKM.
[15] Shinichi Nakajima,et al. Global analytic solution of fully-observed variational Bayesian matrix factorization , 2013, J. Mach. Learn. Res..
[16] Mark Sanderson,et al. Test Collection Based Evaluation of Information Retrieval Systems , 2010, Found. Trends Inf. Retr..
[17] Yi Chang,et al. Yahoo! Learning to Rank Challenge Overview , 2010, Yahoo! Learning to Rank Challenge.
[18] M. de Rijke,et al. A Comparative Analysis of Interleaving Methods for Aggregated Search , 2015, TOIS.
[19] Craig MacDonald,et al. Generalized Team Draft Interleaving , 2015, CIKM.
[20] Maarten de Rijke,et al. Probabilistic Multileave Gradient Descent , 2016, ECIR.
[21] Yisong Yue,et al. Beyond position bias: examining result attractiveness as a source of presentation bias in clickthrough data , 2010, WWW '10.
[22] Thorsten Joachims,et al. Optimizing search engines using clickthrough data , 2002, KDD.
[23] Salvatore Orlando,et al. Fast Ranking with Additive Ensembles of Oblivious and Non-Oblivious Regression Trees , 2016, ACM Trans. Inf. Syst..
[24] Thorsten Joachims,et al. Evaluating Retrieval Performance Using Clickthrough Data , 2003, Text Mining.
[25] Thorsten Joachims,et al. Unbiased Evaluation of Retrieval Quality using Clickthrough Data , 2002 .
[26] Ron Kohavi,et al. Controlled experiments on the web: survey and practical guide , 2009, Data Mining and Knowledge Discovery.
[27] Thorsten Joachims,et al. Interactively optimizing information retrieval systems as a dueling bandits problem , 2009, ICML '09.
[28] ChengXiang Zhai,et al. Evaluation of methods for relative comparison of retrieval systems based on clickthroughs , 2009, CIKM.
[29] Katja Hofmann,et al. Fidelity, Soundness, and Efficiency of Interleaved Comparison Methods , 2013, TOIS.
[30] Ben Carterette,et al. Million Query Track 2007 Overview , 2008, TREC.
[31] Chao Liu,et al. Efficient multiple-click models in web search , 2009, WSDM '09.
[32] M. de Rijke,et al. Probabilistic Multileave for Online Retrieval Evaluation , 2015, SIGIR.