We develop a method for predicting the quality of the passage retrieval component in question answering systems. Since high-quality passages form the basis for accurate answer extraction, our method naturally extends to prediction of an entire system’s effectiveness at extracting a correct answer for a given question. Such prediction of question performance may lead to ways of guiding users in improving questions unlikely to succeed. Our metric is also a necessary research step towards systems that automatically tailor their methods to suit each individual question. Building on previous work on predicting the performance of queries in retrieving documents, we show how to compute the clarity score for questions using passage-based collections. We show that this score is correlated with the average precision in a TREC-9 based system, breakdown the correlation by question type, and discuss example questions. We also study a more general set of questions extracted from a Web log to help make the case for the general usefulness of performance prediction based on question clarity score.
[1]
Peter Bailey,et al.
Engineering a multi-purpose test collection for Web retrieval experiments
,
2003,
Inf. Process. Manag..
[2]
W. Bruce Croft,et al.
Quantifying query ambiguity
,
2002
.
[3]
Jimmy J. Lin,et al.
Web question answering: is more always better?
,
2002,
SIGIR '02.
[4]
Subhabrata Chakraborti,et al.
Nonparametric Statistical Inference
,
2011,
International Encyclopedia of Statistical Science.
[5]
Thomas M. Cover,et al.
Elements of Information Theory
,
2005
.
[6]
Ellen M. Voorhees,et al.
Overview of the TREC-9 Question Answering Track
,
2000,
TREC.
[7]
W. Bruce Croft,et al.
Relevance-Based Language Models
,
2001,
SIGIR '01.
[8]
W. Bruce Croft,et al.
Predicting query performance
,
2002,
SIGIR '02.
[9]
Richard D. Deveaux,et al.
Applied Smoothing Techniques for Data Analysis
,
1999,
Technometrics.
[10]
W. Bruce Croft,et al.
INQUERY System Overview
,
1993,
TIPSTER.