High Quality Expertise Evidence for Expert Search

In an Enterprise setting, an expert search system can assist users with their "expertise need" by suggesting people with relevant expertise to the topic of interest. These systems typically work by associating documentary evidence of expertise to each candidate expert, and then ranking the candidates by the extent to which the documents in their profile are about the query. There are three important factors that affect the retrieval performance of an expert search system - firstly, the selection of the candidate profiles (the documents associated with each candidate), secondly, how the topicality of the documents is measured, and thirdly how the evidence of expertise from the associated documents is combined. In this work, we investigate a new dimension to expert finding, namely whether some documents are better indicators of expertise than others in each candidate's profile. We apply five techniques to predict the quality documents in candidate profiles, which are likely to be good indicators of expertise. The techniques applied include the identification of possible candidate homepages, and of clustering the documents in each profile to determine the candidate's main areas of expertise. The proposed approaches are evaluated on three expert search task from recent TREC Enterprise tracks and provide conclusions.

[1]  Maarten de Rijke,et al.  Finding experts and their eetails in e-mail corpora , 2006, WWW '06.

[2]  Craig MacDonald,et al.  Expertise drift and query expansion in expert search , 2007, CIKM '07.

[3]  ChengXiang Zhai,et al.  Probabilistic Models for Expert Finding , 2007, ECIR.

[4]  Craig MacDonald,et al.  Voting techniques for expert search , 2008, Knowledge and Information Systems.

[5]  Ronald Fagin,et al.  Searching the workplace web , 2003, WWW '03.

[6]  Iadh Ounis,et al.  Combination of Document Priors in Web Information Retrieval , 2007, RIAO.

[7]  Iadh Ounis,et al.  University of Glasgow at TREC 2006: Experiments in Terabyte and Enterprise Tracks with Terrier , 2006, TREC.

[8]  Peter Ingwersen,et al.  Developing a Test Collection for the Evaluation of Integrated Search , 2010, ECIR.

[9]  Craig MacDonald,et al.  Terrier Information Retrieval Platform , 2005, ECIR.

[10]  Gianni Amati,et al.  Frequentist and Bayesian Approach to Information Retrieval , 2006, ECIR.

[11]  David Hawking,et al.  How Valuable is External Link Evidence When Searching Enterprise Webs? , 2004, ADC.

[12]  David Hawking,et al.  Panoptic Expert: Searching for experts not just for documents , 2001 .

[13]  Craig MacDonald,et al.  Expert Search Evaluation by Supporting Documents , 2008, ECIR.

[14]  Craig MacDonald,et al.  Voting for candidates: adapting data fusion techniques for an expert search task , 2006, CIKM '06.

[15]  Nick Craswell,et al.  Overview of the TREC 2005 Enterprise Track , 2005, TREC.

[16]  W. Bruce Croft,et al.  Finding experts in community-based question-answering services , 2005, CIKM '05.

[17]  Djoerd Hiemstra,et al.  The Importance of Prior Probabilities for Entry Page Search , 2002, SIGIR '02.

[18]  Stephen E. Robertson,et al.  Relevance weighting for query independent evidence , 2005, SIGIR '05.

[19]  Craig MacDonald,et al.  Using Relevance Feedback in Expert Search , 2007, ECIR.

[20]  Ben He,et al.  Terrier : A High Performance and Scalable Information Retrieval Platform , 2022 .

[21]  M. de Rijke,et al.  Formal models for expert finding in enterprise corpora , 2006, SIGIR.

[22]  W. Bruce Croft,et al.  Hierarchical Language Models for Expert Finding in Enterprise Corpora , 2008, Int. J. Artif. Intell. Tools.

[23]  Shenghua Bao,et al.  Research on Expert Search at Enterprise Track of TREC 2006 , 2005, TREC.