ELI⇥IR: Expertise Learning & Identification ⇥ Information Retrieval

In today’s knowledge-based economy, having the proper expertise is crucial to resolving many tasks. Expertise Finding (EF) is the area of research concerned with matching available experts to given tasks. As tandard approach is to input at ask description/proposal/paper into an EF system, and receive recommended experts as output. Mostly, EF systems operate either via a content-based approach, which uses the text of the input, as well as the text of the available experts’ profiles to determine a match, and structure-based approaches, which use the inherent relationship between experts, affiliations, papers, etc. (such as is available in citation networks). The majority of methods use one approach (content-based, "C") or the other (structure-based, "S"); though sometimes both approaches are used in tandem (C and S). The underlying data representation is fundamentally different, which makes the methods mutually incompatible. However, in previous work Watanabe et al. [34] achieved good results by converting content-based data to a structure-representation and using a structure-based approach. We posit that the reverse may also hold merit, namely, a content-based approach leveraging structure-based data converted to a content-based representation. We compare our idea to a content only-based approach, demonstrating that our method yields substantially better performance, and thereby substantiating our claim.

[1]  Berthier A. Ribeiro-Neto,et al.  A brief survey of web data extraction tools , 2002, SGMD.

[2]  Richard F. Hartl,et al.  Reviewer Assignment for Scientific Articles using Memetic Algorithms , 2007, Metaheuristics.

[3]  Paul Prekop,et al.  Supporting Knowledge and Expertise Finding within Australia's Defence Science and Technology Organisation , 2007, 2007 40th Annual Hawaii International Conference on System Sciences (HICSS'07).

[4]  Lynn A. Streeter,et al.  Comparing and combining the effectiveness of latent semantic indexing and the ordinary vector space model for information retrieval , 1989, Inf. Process. Manag..

[5]  Mark T. Maybury,et al.  Expert Finding Systems , 2006 .

[6]  Fan Wang,et al.  A Survey on Reviewer Assignment Problem , 2008, IEA/AIE.

[7]  F. Harary,et al.  The cohesiveness of blocks in social networks: Node connectivity and conditional density , 2001 .

[8]  Susan T. Dumais,et al.  The vocabulary problem in human-system communication , 1987, CACM.

[9]  Yiqun Liu,et al.  THUIR at TREC2008: Enterprise Track , 2008 .

[10]  Henry G. Small,et al.  Co-citation in the scientific literature: A new measure of the relationship between two documents , 1973, J. Am. Soc. Inf. Sci..

[11]  Ryen W. White,et al.  Enhancing Expert Finding Using Organizational Hierarchies , 2009, ECIR.

[12]  David Hartvigsen,et al.  The Conference Paper‐Reviewer Assignment Problem* , 1999 .

[13]  M. de Rijke,et al.  Formal models for expert finding in enterprise corpora , 2006, SIGIR.

[14]  M. M. Kessler Bibliographic coupling between scientific papers , 1963 .

[15]  Peter Bailey,et al.  Overview of the TREC 2008 Enterprise Track , 2008, TREC.

[16]  M. Randic,et al.  Resistance distance , 1993 .

[17]  Jakob Nielsen,et al.  Automating the assignment of submitted manuscripts to reviewers , 1992, SIGIR '92.

[18]  Toramatsu Shintani,et al.  A paper recommendation mechanism for the research support system Papits , 2005, International Workshop on Data Engineering Issues in E-Commerce.

[19]  Junyu Niu,et al.  Using Role Determination and Expert Mining in the Enterprise Environment , 2008, TREC.

[20]  Michael J. Pazzani,et al.  Mining for proposal reviewers: lessons learned at the national science foundation , 2006, KDD '06.

[21]  Haiqiang Chen,et al.  Research on Enterprise Track of TREC 2008 , 2007, TREC.

[22]  David Yarowsky,et al.  Taking the load off the conference chairs-towards a digital paper-routing assistant , 1999, EMNLP.

[23]  Jan-Ming Ho,et al.  A Reviewer Recommendation System Based on Collaborative Intelligence , 2009, 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology.

[24]  Dragomir R. Radev,et al.  Blind men and elephants: What do citation summaries tell us about a research article? , 2008, J. Assoc. Inf. Sci. Technol..

[25]  Jianhan Zhu The University College London at TREC 2008 Enterprise Track , 2008, TREC.

[26]  Alfred Kobsa,et al.  Expert-Finding Systems for Organizations: Problem and Domain Analysis and the DEMOIR Approach , 2003, J. Organ. Comput. Electron. Commer..

[27]  Wei Lu,et al.  CSIR at TREC 2008 Expert Search Task: Modeling Expert Evidence in Expert Search , 2008 .

[28]  Manabu Okumura,et al.  Towards Multi-paper Summarization Using Reference Information , 1999, IJCAI.

[29]  M. Hasan,et al.  Using Publications and Domain Knowledge to Build Research Profiles: An Application in Automatic Reviewer Assignment , 2007, 2007 International Conference on Information and Communication Technology.

[30]  Craig MacDonald,et al.  University of Glasgow at TREC 2008: Experiments in Blog, Enterprise, and Relevance Feedback Tracks with Terrier , 2008, TREC.

[31]  Dorothea Wagner,et al.  Approximating Clustering Coefficient and Transitivity , 2005, J. Graph Algorithms Appl..

[32]  M. de Rijke,et al.  Combining Candidate and Document Models for Expert Search , 2008, TREC.