Generating a Common Question from Multiple Documents using Multi-source Encoder-Decoder Models

Ambiguous user queries in search engines result in the retrieval of documents that often span multiple topics. One potential solution is for the search engine to generate multiple refined queries, each of which relates to a subset of the documents spanning the same topic. A preliminary step towards this goal is to generate a question that captures common concepts of multiple documents. We propose a new task of generating common question from multiple documents and present simple variant of an existing multi-source encoder-decoder framework, called the Multi-Source Question Generator (MSQG). We first train an RNN-based single encoder-decoder generator from (single document, question) pairs. At test time, given multiple documents, the 'Distribute' step of our MSQG model predicts target word distributions for each document using the trained model. The 'Aggregate' step aggregates these distributions to generate a common question. This simple yet effective strategy significantly outperforms several existing baseline models applied to the new task when evaluated using automated metrics and human judgments on the MS-MARCO-QA dataset.

[1]  Eric Chu,et al.  Unsupervised Neural Multi-document Abstractive Summarization , 2018, ArXiv.

[2]  Satoshi Nakamura,et al.  Multi-Source Neural Machine Translation with Missing Data , 2018, NMT@ACL.

[3]  Ben He,et al.  NPRF: A Neural Pseudo Relevance Feedback Framework for Ad-hoc Information Retrieval , 2018, EMNLP.

[4]  W. Bruce Croft,et al.  Asking Clarifying Questions in Open-Domain Information-Seeking Conversations , 2019, SIGIR.

[5]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[6]  Todd M. Gureckis,et al.  Asking and evaluating natural language questions , 2016, CogSci.

[7]  Lihong Li,et al.  Neural Approaches to Conversational AI , 2019, Found. Trends Inf. Retr..

[8]  Yanjun Ma,et al.  Answer-focused and Position-aware Neural Question Generation , 2018, EMNLP.

[9]  Marilyn A. Walker,et al.  Neural Generation of Diverse Questions using Answer Focus, Contextual and Linguistic Features , 2018, INLG.

[10]  Yang Liu,et al.  Modeling Coverage for Neural Machine Translation , 2016, ACL.

[11]  Jindrich Libovický,et al.  Input Combination Strategies for Multi-Source Transformer Decoder , 2018, WMT.

[12]  P J Webros BACKPROPAGATION THROUGH TIME: WHAT IT DOES AND HOW TO DO IT , 1990 .

[13]  Zhiguo Wang,et al.  A Unified Query-based Generative Model for Question Generation and Question Answering , 2017, ArXiv.

[14]  Kyunghyun Cho,et al.  Passage Re-ranking with BERT , 2019, ArXiv.

[15]  Delphine Bernhard,et al.  Query Expansion based on Pseudo Relevance Feedback from Definition Clusters , 2010, COLING.

[16]  Boyang Li,et al.  Multiplicative Representations for Unsupervised Semantic Role Induction , 2016, ACL.

[17]  W. Bruce Croft,et al.  Quary Expansion Using Local and Global Document Analysis , 1996, SIGIR Forum.

[18]  Yejin Choi,et al.  Deep Communicating Agents for Abstractive Summarization , 2018, NAACL.

[19]  Fei Liu,et al.  Adapting the Neural Encoder-Decoder Framework from Single to Multi-Document Summarization , 2018, EMNLP.

[20]  Igor Labutov,et al.  Deep Questions without Deep Understanding , 2015, ACL.

[21]  Xiaofei Wang,et al.  Multi-encoder multi-resolution framework for end-to-end speech recognition , 2018, ArXiv.

[22]  Yue Zhang,et al.  Leveraging Context Information for Natural Question Generation , 2018, NAACL.

[23]  Kevin Knight,et al.  Multi-Source Neural Translation , 2016, NAACL.

[24]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[25]  Xinya Du,et al.  Learning to Ask: Neural Question Generation for Reading Comprehension , 2017, ACL.

[26]  Kai Liu,et al.  Multi-Passage Machine Reading Comprehension with Cross-Passage Answer Verification , 2018, ACL.

[27]  Harris Wu,et al.  Evaluating Web-based Question Answering Systems , 2002, LREC.

[28]  Kyunghyun Cho,et al.  Task-Oriented Query Reformulation with Reinforcement Learning , 2017, EMNLP.

[29]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[30]  Jianfeng Gao,et al.  Image-Grounded Conversations: Multimodal Context for Natural Question and Response Generation , 2017, IJCNLP.

[31]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[32]  John D. Lafferty,et al.  Model-based feedback in the language modeling approach to information retrieval , 2001, CIKM '01.

[33]  Junji Tomita,et al.  Multi-style Generative Reading Comprehension , 2019, ACL.

[34]  Gerard Salton,et al.  The SMART Retrieval System—Experiments in Automatic Document Processing , 1971 .

[35]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[36]  David A. Smith,et al.  Multi-Input Attention for Unsupervised OCR Correction , 2018, ACL.

[37]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[38]  Luke S. Zettlemoyer,et al.  Deep Contextualized Word Representations , 2018, NAACL.

[39]  W. Bruce Croft,et al.  Query expansion using local and global document analysis , 1996, SIGIR '96.

[40]  Ellen M Voorhees Question answering in TREC , 2001, CIKM '01.

[41]  Luo Si,et al.  A Deep Cascade Model for Multi-Document Reading Comprehension , 2018, AAAI.

[42]  Jindrich Libovický,et al.  Attention Strategies for Multi-Source Sequence-to-Sequence Learning , 2017, ACL.

[43]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[44]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[45]  W. Bruce Croft,et al.  Latent concept expansion using markov random fields , 2007, SIGIR.

[46]  Jannis Bulian,et al.  Ask the Right Questions: Active Question Reformulation with Reinforcement Learning , 2017, ICLR.

[47]  Jianfeng Gao,et al.  A Human Generated MAchine Reading COmprehension Dataset , 2018 .

[48]  Jaana Kekäläinen,et al.  Cumulated gain-based evaluation of IR techniques , 2002, TOIS.

[49]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[50]  Ming Zhou,et al.  Question Generation for Question Answering , 2017, EMNLP.

[51]  Yaser Al-Onaizan,et al.  Zero-Resource Translation with Multi-Lingual Neural Machine Translation , 2016, EMNLP.

[52]  Christof Monz,et al.  Ensemble Learning for Multi-Source Neural Machine Translation , 2016, COLING.

[53]  Stephen E. Robertson,et al.  Selecting good expansion terms for pseudo-relevance feedback , 2008, SIGIR '08.

[54]  Hugo Zaragoza,et al.  The Probabilistic Relevance Framework: BM25 and Beyond , 2009, Found. Trends Inf. Retr..