A Proposal to Extend and Enrich the Scientific Data Curation of Evaluation Campaigns

This paper examines the current way of keeping the data produced during an evaluation campaign of Information Retrieval Systems (IRSs) and highlights some shortenings of it. In particular, the Cranfield methodology has been designed for creating comparable experiments and evaluating the performances of IRS rather than modeling and managing the scientific data produced during an evaluation campaign. The data produced during an evaluation campaign of IRSs are valuable scientific data, and as a consequence, their lineage should be tracked since it allows us to judge the quality and applicability of information for a given use; those data should be enriched progressively adding further analyses and interpretations on them; it should be possibile to cite them and their further elaboration, since this is an effective way forexplicitly mentioningandmakingreferencesto useful information, for improving the cooperation among researchers and to facilitate the transfer of scientific and innovative results from research groups to the industrial sector.

[1]  Editors , 1986, Brain Research Bulletin.

[2]  M. Zeleny Management support systems: Towards integrated knowledge management , 1987 .

[3]  David A. Hull Using statistical testing in the evaluation of retrieval experiments , 1993, SIGIR.

[4]  Tim Berners-Lee,et al.  Universal Resource Identifiers in WWW: A Unifying Syntax for the Expression of Names and Addresses of Objects on the Network as used in the World-Wide Web , 1994, RFC.

[5]  S. Sudarshan,et al.  Data models , 1996, CSUR.

[6]  Cyril Cleverdon,et al.  The Cranfield tests on index language devices , 1997 .

[7]  Justin Zobel,et al.  How reliable are the results of large-scale information retrieval experiments? , 1998, SIGIR '98.

[8]  Roy T. Fielding,et al.  Uniform Resource Identifiers (URI): Generic Syntax , 1998, RFC.

[9]  Ivar Jacobson,et al.  The unified modeling language reference manual , 2010 .

[10]  Ivar Jacobson,et al.  The Unified Modeling Language User Guide , 1998, J. Database Manag..

[11]  W. Bruce Croft Combining Approaches to Information Retrieval , 2002 .

[12]  Ellen M. Voorhees,et al.  Overview of the TREC 2004 Robust Retrieval Track , 2004 .

[13]  William L. Anderson Some challenges and issues in managing, and preserving access to, long-lived collections of digital scientific and technical data , 2004, Data Sci. J..

[14]  Hans-Jörg Schek,et al.  Digital library information-technology infrastructures , 2005, International Journal on Digital Libraries.

[15]  C. Buckley,et al.  Reliable Information Access Final Workshop Report , 2004 .

[16]  Hsin-Hsi Chen,et al.  Overview of CLIR Task at the Fourth NTCIR Workshop , 2004, NTCIR.

[17]  Jan Brase Using Digital Library Techniques - Registration of Scientific Primary Data , 2004, ECDL.

[18]  Donna K. Harman,et al.  The NRRC reliable information access (RIA) workshop , 2004, SIGIR '04.

[19]  Norman Paskin,et al.  Digital Object Identifiers for scientific data , 2005, Data Sci. J..

[20]  Jennifer Widom,et al.  The Lowell database research self-assessment , 2003, CACM.

[21]  Hsin-Hsi Chen,et al.  Overview of CLIR Task at the Sixth NTCIR Workshop , 2005, NTCIR.

[22]  Carol Peters,et al.  CLEF 2005: Ad Hoc Track Overview , 2005, CLEF.

[23]  Giorgio Maria Di Nunzio,et al.  DIRECT: A System for Evaluating Information Access Components of Digital Libraries , 2005, ECDL.

[24]  Giorgio Maria Di Nunzio,et al.  Scientific Data of an Evaluation Campaign: Do We Properly Deal With Them? , 2006, CLEF.

[25]  R. Frost,et al.  From data to wisdom: pathways to successful data management for Australian science. Report of the working group on Data for Science to the Prime Minister's Science, Engineering and Innovation Council (PMSEIC) , 2006 .

[26]  Giorgio Maria Di Nunzio,et al.  Scientific Evaluation of a DLMS: A Service for Evaluating Information Access Components , 2006, ECDL.

[27]  Nicola Ferro,et al.  Queries and Relevance Assessments: The Right Context for the Right Topic , 2006 .

[28]  Kurt Bilde,et al.  En forskningsartikel: This is a test , 2007 .

[29]  José Luis Vicedo González,et al.  TREC: Experiment and evaluation in information retrieval , 2007, J. Assoc. Inf. Sci. Technol..

[30]  Niso ANSI/NISO Z39.88-2004 The OpenURL Framework for Context-Sensitive Services , 2008 .

[31]  Vanessa Murdock Ellen Voorhees and Donna Harman (eds): TREC Experiment and Evaluation in Information Retrieval , 2008, Information Retrieval.