A Compare-Aggregate Model with Embedding Selector for Answer Selection

Answer selection is a challenging task in natural language processing that requires both natural language understanding and word knowledge. At present, most of recent methods draw on insights from attention mechanism to learn the complex semantic relations between questions and answers. Previous remarkable approaches mainly apply general Compare-Aggregate framework. In this paper, we propose a novel Compare-Aggregate framework with embedding selector to solve answer selection task. Unlike previous Compare-Aggregate methods which just use one type of Attention mechanism and lack the use of word vectors at different level, we employ two types of Attention mechanism in a model and add a selector layer to choose a best input for aggregation layer. We evaluate the model on the two answer selection tasks: WikiQA and TrecQA. On the two different datasets, our approach outperforms several strong baselines and achieves state-of-the-art performance.

[1]  Di Wang,et al.  A Long Short-Term Memory Model for Answer Sentence Selection in Question Answering , 2015, ACL.

[2]  Wenpeng Yin,et al.  Why and How to Pay Different Attention to Phrase Alignments of Different Intensities , 2016, ArXiv.

[3]  Nan Yang,et al.  Context-Aware Answer Sentence Selection With Hierarchical Gated Recurrent Neural Networks , 2018, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[4]  Ming-Wei Chang,et al.  Question Answering Using Enhanced Lexical Semantic Models , 2013, ACL.

[5]  Alessandro Moschitti,et al.  Modeling Relational Information in Question-Answer Pairs with Convolutional Neural Networks , 2016, ArXiv.

[6]  Jason Weston,et al.  Key-Value Memory Networks for Directly Reading Documents , 2016, EMNLP.

[7]  Shuohang Wang,et al.  Learning Natural Language Inference with LSTM , 2015, NAACL.

[8]  Daniele Bonadiman,et al.  Convolutional Neural Networks vs. Convolution Kernels: Feature Engineering for Answer Sentence Reranking , 2016, NAACL.

[9]  Si Li,et al.  A Compare-Aggregate Model with Dynamic-Clip Attention for Answer Selection , 2017, CIKM.

[10]  Yan Li,et al.  M$^2$S-Net: Multi-Modal Similarity Metric Learning based Deep Convolutional Network for Answer Selection , 2016, ArXiv.

[11]  Jun Zhao,et al.  Inner Attention based Recurrent Neural Networks for Answer Selection , 2016, ACL.

[12]  Zhiguo Wang,et al.  Bilateral Multi-Perspective Matching for Natural Language Sentences , 2017, IJCAI.

[13]  Yi Yang,et al.  WikiQA: A Challenge Dataset for Open-Domain Question Answering , 2015, EMNLP.

[14]  Noah A. Smith,et al.  Tree Edit Models for Recognizing Textual Entailments, Paraphrases, and Answers to Questions , 2010, NAACL.

[15]  Hinrich Schütze,et al.  Erratum: “ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs” , 2016, Transactions of the Association for Computational Linguistics.

[16]  Shuohang Wang,et al.  A Compare-Aggregate Model for Matching Text Sequences , 2016, ICLR.

[17]  Jimmy J. Lin,et al.  Pairwise Word Interaction Modeling with Deep Neural Networks for Semantic Similarity Measurement , 2016, NAACL.

[18]  Lei Yu,et al.  Deep Learning for Answer Sentence Selection , 2014, ArXiv.

[19]  Bowen Zhou,et al.  Improved Representation Learning for Question Answer Matching , 2016, ACL.

[20]  Bowen Zhou,et al.  ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs , 2015, TACL.

[21]  Phil Blunsom,et al.  Neural Variational Inference for Text Processing , 2015, ICML.

[22]  Jakob Uszkoreit,et al.  A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.

[23]  Zhiguo Wang,et al.  Sentence Similarity Learning by Lexical Decomposition and Composition , 2016, COLING.