The potential and actual effectiveness of interactive query expansion

In query expansion, terms from a source such as relevance feedback are added to the query. This often improves retrieval effectiveness but results are variable across queries. In interactive query expansion (IQE) the automatically-derived terms are instead offered as suggestions to the searcher, who decides which to add. There is little evidence of whether IQE is likely to be effective over multiple iterations in a large scale retrieval context, or whether inexperienced users can achieve this effectiveness in practice. These experiments address these two questions. A small but significant improvement in potential retrieval effectiveness is found. This is consistent across a range of topics. Inexperienced users’ term selections consistently fail to improve on automatic query expansion, however. It is concluded that interactive query expansion has good potential, particular y for term sources that are porer than relevance feedback. But it may be difficult for searchers to realise this potential without experience or training in term selection and free-text search strategies.

[1]  Nicholas J. Belkin,et al.  A case for interaction: a study of interactive information retrieval behavior and effectiveness , 1996, CHI.

[2]  E. Michael Keen,et al.  Presenting Results of Experimental Retrieval Comparisons , 1997, Inf. Process. Manag..

[3]  Stephen E. Robertson,et al.  Relevance weighting of search terms , 1976, J. Am. Soc. Inf. Sci..

[4]  James Allan,et al.  The effect of adding relevance information in a relevance feedback environment , 1994, SIGIR '94.

[5]  Donna K. Harman,et al.  A failure analysis of the limitation of suffixing in an online environment , 1987, SIGIR '87.

[6]  Peter Willett,et al.  Instruct: a teaching package for experimental methods in information retrieval , 1988 .

[7]  C. J. van Rijsbergen,et al.  Proceedings of the 10th annual international ACM SIGIR conference on Research and development in information retrieval , 1987, SIGIR 1987.

[8]  David A. Hull Stemming algorithms: a case study for detailed evaluation , 1996 .

[9]  Tamas E. Doszkocs,et al.  CITE NLM: natural-language searching in an online catalog , 1983 .

[10]  Valerie Galpin,et al.  Relevance feedback in a public access catalogue for a research library: MUSCAT at the Scott Polar Research Institute Library , 1988 .

[11]  Chris Buckley,et al.  New Retrieval Approaches Using SMART: TREC 4 , 1995, TREC.

[12]  Peter Willett,et al.  The limitations of term co-occurrence data for query expansion in document retrieval systems , 1991, J. Am. Soc. Inf. Sci..

[13]  Ellen M. Voorhees,et al.  Vector Expansion in a Large Collection , 1992, TREC.

[14]  Jack Minker,et al.  An evaluation of query expansion by the addition of clustered terms for a document retrieval system , 1972, Inf. Storage Retr..

[15]  Micheline Hancock-Beaulieu,et al.  An Evaluation of Interactive Query Expansion in an Online Library Catalogue with a Graphical User Interface , 1995, J. Documentation.

[16]  Julie Beth Lovins,et al.  Development of a stemming algorithm , 1968, Mech. Transl. Comput. Linguistics.

[17]  Jaana Kristensen,et al.  Expanding End-Users' Query Statements for Free Text Searching with a Search-Aid Thesaurus , 1993, Inf. Process. Manag..

[18]  Gerard Salton,et al.  Improving retrieval performance by relevance feedback , 1997, J. Am. Soc. Inf. Sci..

[19]  Jose Enrique Araya Interactive query formulation and feedback experiments in information retrieval , 1990 .

[20]  Donna K. Harman,et al.  Relevance feedback revisited , 1992, SIGIR '92.

[21]  Alan F. Smeaton,et al.  The Retrieval Effects of Query Expansion on a Feedback Document Retrieval System , 1983, Comput. J..

[22]  Martin F. Porter,et al.  An algorithm for suffix stripping , 1997, Program.

[23]  Stephen E. Robertson,et al.  Probabilistic models of indexing and searching , 1980, SIGIR '80.