Answer Selection in a Multi-stream Open Domain Question Answering System

Question answering systems aim to meet users’ information needs by returning exact answers in response to a question. Traditional open domain question answering systems are built around a single pipeline architecture. In an attempt to exploit multiple resources as well as multiple answering strategies, systems based on a multi-stream architecture have recently been introduced. Such systems face the challenging problem of having to select a single answer from pools of answers obtained using essentially different techniques. We report on experiments aimed at understanding and evaluating the effect of different options for answer selection in a multi-stream question answering system. We examine the impact of local tiling techniques, assignments of weights to streams based on past performance and/or question type, as well redundancy-based ideas. Our main finding is that redundancy-based ideas in combination with naively learned stream weights conditioned on question type work best, and improve significantly over a number of baselines.

[1]  Sanda M. Harabagiu,et al.  The Structure and Performance of an Open-Domain Question Answering System , 2000, ACL.

[2]  Gilad Mishne,et al.  How Frogs Built the Berlin Wall: A Detailed Error Analysis of a Question Answering System for Dutch , 2003, CLEF.

[3]  W. Bruce Croft Combining Approaches to Information Retrieval , 2002 .

[4]  Robert J. Gaizauskas,et al.  The University of Sheffield TREC 2002 Q&A System , 2002, TREC.

[5]  Dragomir R. Radev,et al.  Question-answering by predictive annotation , 2000, SIGIR '00.

[6]  Edward A. Fox,et al.  Combination of Multiple Searches , 1993, TREC.

[7]  Jennifer Chu-Carroll,et al.  A Multi-Strategy and Multi-Source Approach to Question Answering , 2002, TREC.

[8]  Charles L. A. Clarke,et al.  Statistical Selection of Exact Answers (MultiText Experiments for TREC 2002) , 2002, TREC.

[9]  Martin M. Soubbotin,et al.  Use of Patterns for Detection of Likely Answer Strings: A Systematic Approach , 2002, TREC.

[10]  Carol Peters,et al.  Comparative Evaluation of Multilingual Information Access Systems , 2003, Lecture Notes in Computer Science.

[11]  Eduard H. Hovy,et al.  Question Answering in Webclopedia , 2000, TREC.

[12]  M. de Rijke,et al.  Tequesta: The University of Amsterdam's Textual Question Answering System , 2001, TREC.

[13]  Sanda M. Harabagiu,et al.  Performance Issues and Error Analysis in an Open-Domain Question Answering System , 2002, ACL.

[14]  Bernardo Magnini,et al.  Is It the Right Answer? Exploiting Web Redundancy for Answer Validation , 2002, ACL.

[15]  Djoerd Hiemstra,et al.  Translation Resources, Merging Strategies, and Relevance Feedback for Cross-Language Information Retrieval , 2000, CLEF.

[16]  Brigitte Grau,et al.  The Question Answering System QALC at LIMSI, Experiments in Using Web and WordNet , 2002, TREC.

[17]  Scott Miller,et al.  TREC 2002 QA at BBN: Answer Selection and Confidence Estimation , 2002, TREC.

[18]  Gilad Mishne,et al.  The University of Amsterdam at the TREC 2003 Question Answering Track , 2003, TREC.

[19]  Gilad Mishne,et al.  How frogs built the Berlin Wall , 2004 .

[20]  Carol Peters,et al.  Cross-Language Information Retrieval and Evaluation , 2001, Lecture Notes in Computer Science.

[21]  Eduard H. Hovy,et al.  Offline Strategies for Online Question Answering: Answering Questions Before They Are Asked , 2003, ACL.

[22]  Jimmy J. Lin,et al.  AskMSR: Question Answering Using the Worldwide Web , 2002 .