University of Glasgow at TREC 2011: Experiments with Terrier in Crowdsourcing, Microblog, and Web Tracks

In TREC 2011, we focus on tackling the new challenges proposed by the pilot Crowdsourcing and Microblog tracks, using our Terrier Information Retrieval Platform. Meanwhile, we continue to build upon our novel xQuAD framework and data-driven ranking approaches within Terrier to achieve effective and efficient ranking for the TREC Web track. In particular, the aim of our Microblog track participation is the development of a learning to rank approach for filtering within a tweet ranking environment, where tweets are ranked in reverse chronological order. In the Crowdsourcing track, we work to achieve a closer integration between the crowdsourcing marketplaces that are used for relevance assessment, and Terrier, which produces the document rankings to be assessed. Moreover, we focus on generating relevance assessments quickly and at a minimal cost. For the Web track, we enhance the data-driven learning support within Terrier by proposing a novel framework for the fast computation of document features for learning to rank.

[1]  Giorgio Gambosi,et al.  FUB, IASI-CNR and University of Tor Vergata at TREC 2008 Blog Track , 2008, TREC.

[2]  Craig MacDonald,et al.  Explicit Search Result Diversification through Sub-queries , 2010, ECIR.

[3]  Howard R. Turtle,et al.  Query Evaluation: Strategies and Optimizations , 1995, Inf. Process. Manag..

[4]  Iadh Ounis,et al.  Incorporating term dependency in the dfr framework , 2007, SIGIR.

[5]  Aniket Kittur,et al.  Crowdsourcing user studies with Mechanical Turk , 2008, CHI.

[6]  Donald Metzler,et al.  Automatic feature selection in the markov random field model for information retrieval , 2007, CIKM '07.

[7]  Andrei Z. Broder,et al.  Efficient query evaluation using a two-level retrieval process , 2003, CIKM '03.

[8]  W. Bruce Croft,et al.  A Markov random field model for term dependencies , 2005, SIGIR '05.

[9]  Gianni Amati,et al.  Probability models for information retrieval based on divergence from randomness , 2003 .

[10]  Craig MacDonald,et al.  Intent-aware search result diversification , 2011, SIGIR.

[11]  W. Bruce Croft,et al.  Quality-biased ranking of web documents , 2011, WSDM '11.

[12]  Iadh Ounis,et al.  The Static Absorbing Model for the Web , 2005, J. Web Eng..

[13]  J. Fleiss Measuring nominal scale agreement among many raters. , 1971 .

[14]  Ben He,et al.  Terrier : A High Performance and Scalable Information Retrieval Platform , 2022 .

[15]  Iadh Ounis,et al.  Overview of the TREC 2011 Microblog Track , 2011, TREC.

[16]  Charles L. A. Clarke,et al.  Efficient and effective spam filtering and re-ranking for large web datasets , 2010, Information Retrieval.

[17]  Charles L. A. Clarke,et al.  Overview of the TREC 2011 Web Track , 2011, TREC.

[18]  Craig MacDonald,et al.  Effectiveness beyond the first crawl tier , 2011, CIKM '11.

[19]  R. McCreadie Crowdsourcing Blog Track Top News Judgments at TREC , 2011 .

[20]  Craig MacDonald,et al.  Selectively diversifying web search results , 2010, CIKM.

[21]  Craig MacDonald,et al.  University of Glasgow at WebCLEF 2005: Experiments in per-field Normalisation and Language Specific Stemming , 2005, CLEF.

[22]  Craig MacDonald,et al.  Exploiting query reformulations for web search result diversification , 2010, WWW '10.

[23]  M. J. Box,et al.  Non-linear optimization techniques; , 1969 .

[24]  Rodrygo L. T. Santos,et al.  Diversifying for Multiple Information Needs , 2011 .

[25]  Craig MacDonald,et al.  University of Glasgow at TREC 2009: Experiments with Terrier , 2009, TREC.

[26]  Stephen E. Robertson,et al.  Okapi at TREC-3 , 1994, TREC.

[27]  Stephen E. Robertson,et al.  GatfordCentre for Interactive Systems ResearchDepartment of Information , 1996 .

[28]  Craig MacDonald,et al.  University of Glasgow at TREC 2010: Experiments with Terrier in Blog and Web Tracks , 2010, TREC.