The goal of the Redundancy, Diversity, and Interdependent Document Relevance workshop was to explore how ranking, performance assessment and learning to rank can move beyond the assumption that the relevance of a document is independent of other documents. In particular, the workshop focussed on three themes: the effect of redundancy on information retrieval utility (for example, minimizing the wasted effort of users who must skip redundant information), the role of diversity (for example, for mitigating the risk of misinterpreting ambiguous queries), and algorithms for set-level optimization (where the quality of a set of retrieved documents is not simply the sum of its parts). This workshop built directly upon the Beyond Binary Relevance: Preferences, Diversity and Set-Level Judgments workshop at SIGIR 2008 [3], shifting focus to address the questions left open by the discussions and results from that workshop. As such, it was the first workshop to explicitly focus on the related research challenges of redundancy, diversity, and interdependent relevance – all of which require novel performance measures, learning methods, and evaluation techniques. The workshop program committee consisted of 15 researchers from academia and industry, with experience in IR evaluation, machine learning, and IR algorithmic design. Over 40 people attended the workshop. This report aims to summarize the workshop, and also to systematize common themes and key concepts so as to encourage research in the three workshop themes. It contains our attempt to summarize and organize the topics that came up in presentations as well as in discussions, pulling out common elements. Many audience members contributed, yet due to the free-flowing discussion, attributing all the observations to particular audience members is unfortunately impossible. Not all audience members would necessarily agree with the views presented, but we do attempt to present a consensus view as far as possible.
[1]
William W. Cohen,et al.
Beyond independent relevance: methods and evaluation metrics for subtopic retrieval
,
2003,
SIGIR.
[2]
ChengXiang Zhai,et al.
Active feedback in ad hoc information retrieval
,
2005,
SIGIR '05.
[3]
David Hawking,et al.
Evaluation by comparing result sets in context
,
2006,
CIKM '06.
[4]
John D. Lafferty,et al.
A risk minimization framework for information retrieval
,
2006,
Inf. Process. Manag..
[5]
Xiaojin Zhu,et al.
Improving Diversity in Ranking using Absorbing Random Walks
,
2007,
NAACL.
[6]
Ben Carterette,et al.
Beyond binary relevance: preferences, diversity, and set-level judgments
,
2008,
SIGF.
[7]
Charles L. A. Clarke,et al.
Novelty and diversity in information retrieval evaluation
,
2008,
SIGIR '08.
[8]
Thorsten Joachims,et al.
Predicting diverse subsets using structural SVMs
,
2008,
ICML '08.
[9]
Sreenivas Gollapudi,et al.
Diversifying search results
,
2009,
WSDM '09.
[10]
Sreenivas Gollapudi,et al.
An axiomatic approach for result diversification
,
2009,
WWW '09.
[11]
Dafna Shahaf,et al.
Turning down the noise in the blogosphere
,
2009,
KDD.
[12]
Charles L. A. Clarke,et al.
An Effectiveness Measure for Ambiguous and Underspecified Queries
,
2009,
ICTIR.