A user study of relevance judgments for E-Discovery

This paper presents a comparative user study that investigates the relevance judgments made by assessors with a law background and assessors without. Four law students and four library and information studies (LIS) students were recruited to judge independently the relevance of 100 documents for each of four requests for production of responsive documents for litigation purposes, two requests for the tobacco Master Settlement Agreement collection and two for the Enron corporate emails collection. Both quantitative and qualitative methods are used to analyze data collected through the relevance judgment task, an entry questionnaire, and an exit interview. Being given the same task guidelines, the LIS student assessors judged relevant documents just as accurately as the law student assessors, while they judged nonrelevant documents slightly less accurately than the law student assessors. In addition, participants achieved moderate to substantial agreement on their relevance judgments. Relevance judgment speed varied significantly among participants, although on average it was about the same for the two groups. Factors influencing the accuracy and the speed of participants' relevance judgments are discussed based on preliminary analysis of qualitative data collected through the exit interviews.

[1]  Douglas W. Oard,et al.  TREC 2006 Legal Track Overview , 2006, TREC.

[2]  Cyril W. Cleverdon The effect of variations in relevance assessments in comparative experimental tests of index languages , 1970 .

[3]  Douglas W. Oard,et al.  Overview of the TREC 2008 Legal Track , 2008, TREC.

[4]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[5]  Peter Bailey,et al.  Relevance assessment: are judges exchangeable and does it matter , 2008, SIGIR '08.

[6]  Dagobert Soergel,et al.  An evidence perspective on topical relevance types and its implications for exploratory and task-based retrieval , 2006, Inf. Res..

[7]  Jianqiang Wang,et al.  TREC 2008 at the University at Buffalo: Legal and Blog Track , 2008, TREC.

[8]  William S. Cooper,et al.  A definition of relevance for information retrieval , 1971, Inf. Storage Retr..

[9]  Ellen M. Voorhees Variations in relevance judgments and the measurement of retrieval effectiveness , 2000, Inf. Process. Manag..

[10]  Patrick Wilson,et al.  Situational relevance , 1973, Inf. Storage Retr..

[11]  Carol L. Barry,et al.  Users' Criteria for Relevance Evaluation: A Cross-situational Comparison , 1998, Inf. Process. Manag..

[12]  Douglas W. Oard,et al.  Overview of the TREC 2009 Legal Track , 2009, TREC.

[13]  Peiling Wang,et al.  A Cognitive Model of Document Use During a Research Project. Study I. Document Selection , 1998, J. Am. Soc. Inf. Sci..

[14]  Jianqiang Wang,et al.  TREC 2009 at the University of Buffalo: Interactive Legal E-Discovery With Enron Emails , 2009, TREC.

[15]  Peiling Wang,et al.  A cognitive model of document use during a research project. Study I. document selection , 1998 .

[16]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .