Option Attentive Capsule Network for Multi-choice Reading Comprehension

In this paper, we study the problem of multi-choice reading comprehension, which requires a machine to select the correct answer from a set of candidates based on the given passage and question. Most existing approaches focus on designing sophisticated attention to model the interactions of the sequence triplets (passage, question and candidate options), which aims to extract the answer clues from the passage. After this matching stage, a simple pooling operation is usually applied to aggregate the matching results to make final decisions. However, a bottom-up max or average pooling may loss essential information of the evidence clues and ignore the inter relationships of the sentences, especially dealing with complex questions when there are multiple evidence clues. To this end, we propose an option attentive capsule network with dynamic routing to overcome this issue. Instead of pooling, we introduce a capsule aggregating layer to dynamically fuse the information from multiple evidence clues and iteratively refine the matching representation. Furthermore, we design an option attention-based routing policy to focus more on each candidate option when clustering the features of low-level capsules. Experimental results demonstrate that our proposed model achieves state-of-the-art performance on RACE dataset.

[1]  Geoffrey E. Hinton,et al.  Dynamic Routing Between Capsules , 2017, NIPS.

[2]  Shiyu Chang,et al.  A Co-Matching Model for Multi-choice Reading Comprehension , 2018, ACL.

[3]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[4]  Zhen-Hua Ling,et al.  Neural Natural Language Inference Models Enhanced with External Knowledge , 2017, ACL.

[5]  Matthew Richardson,et al.  MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text , 2013, EMNLP.

[6]  Geoffrey E. Hinton,et al.  Transforming Auto-Encoders , 2011, ICANN.

[7]  Wentao Ma,et al.  Convolutional Spatial Attention Model for Reading Comprehension with Multiple-Choice Questions , 2019, AAAI.

[8]  Ruslan Salakhutdinov,et al.  Gated-Attention Readers for Text Comprehension , 2016, ACL.

[9]  Wei Zhang,et al.  Evidence Aggregation for Answer Re-Ranking in Open-Domain Question Answering , 2017, ICLR.

[10]  Mitesh M. Khapra,et al.  ElimiNet: A Model for Eliminating Options for Reading Comprehension with Multiple Choice Questions , 2018, IJCAI.

[11]  Danqi Chen,et al.  A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task , 2016, ACL.

[12]  Wei Wang,et al.  Multi-Granularity Hierarchical Attention Fusion Networks for Reading Comprehension and Question Answering , 2018, ACL.

[13]  Hai Zhao,et al.  Multi-labeled Relation Extraction with Attentive Capsule Network , 2018, AAAI.

[14]  Wei Zhang,et al.  Attention-Based Capsule Networks with Dynamic Routing for Relation Extraction , 2018, EMNLP.

[15]  Guokun Lai,et al.  RACE: Large-scale ReAding Comprehension Dataset From Examinations , 2017, EMNLP.

[16]  Furu Wei,et al.  Hierarchical Attention Flow for Multiple-Choice Reading Comprehension , 2018, AAAI.