Semi-Distantly Supervised Neural Model for Generating Compact Answers to Open-Domain Why Questions

This paper proposes a neural network-based method for generating compact answers to open-domain why-questions (e.g., “Why was Mr. Trump elected as the president of the US?”). Unlike factoid question answering methods that provide short text spans as answers, existing work for whyquestion answering have aimed at answering questions by retrieving relatively long text passages, each of which often consists of several sentences, from a text archive. While the actual answer to a why-question may be expressed over several consecutive sentences, these often contain redundant and/or unrelated parts. Such answers would not be suitable for spoken dialog systems and smart speakers such as Amazon Echo, which receive much attention in these days. In this work, we aim at generating non-redundant compact answers to why-questions from answer passages retrieved from a very large web data corpora (4 billion web pages) by an already existing open-domain why-question answering system, using a novel neural network obtained by extending existing summarization methods. We also automatically generate training data using a large number of causal relations automatically extracted from 4 billion web pages by an existing supervised causality recognizer. The data is used to train our neural network, together with manually created training data. Through a series of experiments, we show that both our novel neural network and auto-generated training data improve the quality of the generated answers both in ROUGE score and in a subjective evaluation.

[1]  Ryuichiro Higashinaka,et al.  Corpus-based Question Answering for why-Questions , 2008, IJCNLP.

[2]  J. Fleiss Measuring nominal scale agreement among many raters. , 1971 .

[3]  Junta Mizuno,et al.  WISDOM X, DISAANA and D-SUMM: Large-scale NLP Systems for Analyzing Textual Big Data , 2016, COLING.

[4]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[5]  Jason Weston,et al.  A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.

[6]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[7]  Jong-Hoon Oh,et al.  Why-Question Answering using Intra- and Inter-Sentential Causal Relations , 2013, ACL.

[8]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[9]  Jason Weston,et al.  Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.

[10]  Jong-Hoon Oh,et al.  Multi-Column Convolutional Neural Networks with Causality-Attention for Why-Question Answering , 2017, WSDM.

[11]  Jong-Hoon Oh,et al.  Why Question Answering using Sentiment Analysis and Word Classes , 2012, EMNLP.

[12]  Jian Zhang,et al.  SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.

[13]  Jong-Hoon Oh,et al.  Improving Event Causality Recognition with Multiple Background Knowledge Sources Using Multi-Column Convolutional Neural Networks , 2017, AAAI.

[14]  Yutaka Kidawara,et al.  Toward Future Scenario Generation: Extracting Event Causality Exploiting Semantic Relation, Context, and Association Features , 2014, ACL.

[15]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[16]  Jong-Hoon Oh,et al.  Excitatory or Inhibitory: A New Semantic Orientation Extracts Contradiction and Causality from the Web , 2012, EMNLP.

[17]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[18]  Bowen Zhou,et al.  Improved Representation Learning for Question Answer Matching , 2016, ACL.

[19]  Bowen Zhou,et al.  Attentive Pooling Networks , 2016, ArXiv.

[20]  Ming Zhou,et al.  Neural Question Generation from Text: A Preliminary Study , 2017, NLPCC.

[21]  Hans van Halteren,et al.  Learning to rank for why-question answering , 2011, Information Retrieval.

[22]  Peter Jansen,et al.  Creating Causal Embeddings for Question Answering with Minimal Supervision , 2016, EMNLP.

[23]  Jong-Hoon Oh,et al.  A Semi-Supervised Learning Approach to Why-Question Answering , 2016, AAAI.

[24]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[25]  Balaraman Ravindran,et al.  Diversity driven attention model for query-based abstractive summarization , 2017, ACL.

[26]  Roxana Gîrju,et al.  Automatic Detection of Causal Relations for Question Answering , 2003, ACL 2003.