Answering engineers' questions using semantic annotations

Question–answering (QA) systems have proven to be helpful, especially to those who feel uncomfortable entering keywords, sometimes extended with search symbols such as +, *, and so forth. In developing such systems, the main focus has been on the enhanced retrieval performance of searches, and recent trends in QA systems center on the extraction of exact answers. However, when their usability was evaluated, some users indicated that they found it difficult to accept the answers because of the absence of supporting context and rationale. Current approaches to address this problem include providing answers with linking paragraphs or with summarizing extensions. Both methods are believed to be sufficient to answer questions seeking the names of objects or quantities that have only a single answer. However, neither method addresses the situation when an answer requires the comparison and integration of information appearing in multiple documents or in several places in a single document. This paper argues that coherent answer generation is crucial for such questions, and that the key to this coherence is to analyze texts to a level beyond sentence annotations. To demonstrate this idea, a prototype has been developed based on rhetorical structure theory, and a preliminary evaluation has been carried out. The evaluation indicates that users prefer to see the extended answers that can be generated using such semantic annotations, provided that additional context and rationale information are made available.

[1]  S Ahmed,et al.  SEMANTIC ANNOTATION TO SUPPORT AUTOMATIC TAXONOMY CLASSIFICATION , 2006 .

[2]  Kathleen McKeown,et al.  Cut and Paste Based Text Summarization , 2000, ANLP.

[3]  Ellen M. Voorhees,et al.  Overview of the TREC 2002 Question Answering Track , 2003, TREC.

[4]  A. Knott,et al.  Using Linguistic Phenomena to Motivate a Set of Coherence Relations. , 1994 .

[5]  Task-Based Evaluation of Summary Quality: Describing Relationships between Scientific Papers , 2001 .

[6]  W. E. Bosma Extending Answers using Discourse Structure , 2005 .

[7]  Lili Aunimo,et al.  Question Answering Experiments for Finnish and French , 2005, CLEF.

[8]  Rob H. Bracewell,et al.  From Discourse Analysis to Answering Design Questions , 2004, LSTKM@EKAW.

[9]  Oren Etzioni,et al.  Scaling question answering to the Web , 2001, WWW '01.

[10]  Anna Maria Di Sciullo,et al.  Natural Language Understanding , 2009, SoMeT.

[11]  Daniel Marcu,et al.  An Unsupervised Approach to Recognizing Discourse Relations , 2002, ACL.

[12]  Sanda M. Harabagiu,et al.  Experiments with Interactive Question Answering in Complex Scenarios , 2004, Workshop On Pragmatics Of Question Answering.

[13]  Mark T. Maybury,et al.  Advances in Automatic Text Summarization , 1999 .

[14]  Lili Aunimo,et al.  Question Answering using Semantic Annotation , 2005, CLEF.

[15]  Daniel Marcu,et al.  Finding the WRITE Stuff: Automatic Identification of Discourse Structure in Student Essays , 2003, IEEE Intell. Syst..

[16]  Michael ODonnell,et al.  RSTTool 2.4 - A markup Tool for Rhetorical Structure Theory , 2000, INLG.

[17]  Eduard H. Hovy,et al.  Automated Discourse Generation Using Discourse Structure Relations , 1993, Artif. Intell..

[18]  Stephen E. Robertson,et al.  GatfordCentre for Interactive Systems ResearchDepartment of Information , 1996 .

[19]  Regina Barzilay,et al.  Information Fusion in the Context of Multi-Document Summarization , 1999, ACL.

[20]  Daniel Marcu,et al.  Discourse Trees Are Good Indicators of Importance in Text , 1999 .

[21]  Ralph Grishman,et al.  A Corpus-based Probabilistic Grammar with Only Two Non-terminals , 1995, IWPT.

[22]  Elizabeth D. Liddy,et al.  Finding Answers to Complex Questions , 2004, New Directions in Question Answering.

[23]  Philip Resnik,et al.  Using Information Content to Evaluate Semantic Similarity in a Taxonomy , 1995, IJCAI.

[24]  Kuntz Werner,et al.  Issues as Elements of Information Systems , 1970 .

[25]  Rob H. Bracewell,et al.  DRed and Design Folders: A Way of Capturing, Storing and Passing On Knowledge Generated During Design Projects , 2004, DAC 2004.

[26]  George A. Miller,et al.  Introduction to WordNet: An On-line Lexical Database , 1990 .

[27]  Jianchang Mao,et al.  Enterprise Search: Tough Stuff , 2004, ACM Queue.

[28]  Beatrice Santorini,et al.  Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.

[29]  Enrico Motta,et al.  AquaLog: An Ontology-Portable Question Answering System for the Semantic Web , 2005, ESWC.

[30]  Salim Roukos,et al.  TREC-6 Ad-Hoc Retrieval , 1997, TREC.

[31]  Km Wallace,et al.  IMPROVING DOCUMENT ACCESSIBILITY THROUGH ONTOLOGY-BASED INFORMATION SHARING , 2006 .

[32]  Jimmy J. Lin,et al.  What Makes a Good Answer? The Role of Context in Question Answering , 2003, INTERACT.

[33]  W. Mann,et al.  Rhetorical Structure Theory: looking back and moving ahead , 2006 .

[34]  Rob H. Bracewell,et al.  A TOOL FOR CAPTURING DESIGN RATIONALE , 2003 .

[35]  William C. Mann,et al.  Rhetorical Structure Theory: Toward a functional theory of text organization , 1988 .

[36]  Elizabeth D. Liddy,et al.  Enhanced Text Retrieval Using Natural Language Processing , 2005 .

[37]  Ken M. Wallace,et al.  A Methodology for Creating Ontologies for Engineering Design , 2007, J. Comput. Inf. Sci. Eng..

[38]  Ehud Reiter,et al.  A corpus analysis of discourse relations for Natural Language Generation , 2003 .

[39]  Leila Kosseim,et al.  The Problem of Precision in Restricted-Domain Question Answering. Some Proposed Methods of Improvement , 2004, Conference On Question Answering In Restricted Domains.

[40]  Teruko Mitamura,et al.  Extending the JAVELIN QA System with Domain Semantics ∗ , 2005 .

[41]  Saeema Ahmed,et al.  Encouraging reuse of design knowledge: a method to index knowledge , 2005 .