RI-Match: Integrating Both Representations and Interactions for Deep Semantic Matching

Existing deep matching methods can be mainly categorized into two kinds, i.e. representation focused methods and interaction focused methods. Representation focused methods usually focus on learning the representation of each sentence, while interaction focused methods typically aim to obtain the representations of different interaction signals. However, both sentence level representations and interaction signals are important for the complex semantic matching tasks. Therefore, in this paper, we propose a new deep learning architecture to combine the merits of both deep matching approaches. Firstly, two kinds of word level matching matrices are constructed based on word identities and word embeddings, to capture both exact and semantic matching signals. Secondly, a sentence level matching matrix is constructed, with each element stands for the interaction between two sentence representations at corresponding positions, generated by a bidirectional long short term memory (Bi-LSTM). In this way, sentence level representations are well captured in the matching process. The above matrices are then fed into a spatial recurrent neural network (RNN), to generate the high level interaction representations. Finally, the matching score is produced by a k-Max pooling and a multilayer perceptron (MLP). Experiments on paraphrasing identification shows that our model outperforms traditional state-of-the art baselines significantly.

[1]  Xueqi Cheng,et al.  Text Matching as Image Recognition , 2016, AAAI.

[2]  Bowen Zhou,et al.  ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs , 2015, TACL.

[3]  Li Zhao,et al.  Attention-based LSTM for Aspect-level Sentiment Classification , 2016, EMNLP.

[4]  Gang Wang,et al.  A Siamese Long Short-Term Memory Architecture for Human Re-identification , 2016, ECCV.

[5]  Xuanjing Huang,et al.  Convolutional Neural Tensor Network Architecture for Community-Based Question Answering , 2015, IJCAI.

[6]  Konrad Schindler,et al.  Learning by Tracking: Siamese CNN for Robust Target Association , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[7]  Xueqi Cheng,et al.  Match-SRNN: Modeling the Recursive Matching Structure with Spatial RNN , 2016, IJCAI.

[8]  Wei Xu,et al.  Bidirectional LSTM-CRF Models for Sequence Tagging , 2015, ArXiv.

[9]  Hang Li,et al.  Convolutional Neural Network Architectures for Matching Natural Language Sentences , 2014, NIPS.

[10]  Zhiguo Wang,et al.  FAQ-based Question Answering via Word Alignment , 2015, ArXiv.

[11]  Phil Blunsom,et al.  A Convolutional Neural Network for Modelling Sentences , 2014, ACL.

[12]  Zhiguo Wang,et al.  Bilateral Multi-Perspective Matching for Natural Language Sentences , 2017, IJCAI.

[13]  Danqi Chen,et al.  Reasoning With Neural Tensor Networks for Knowledge Base Completion , 2013, NIPS.

[14]  T. Munich,et al.  Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks , 2008, NIPS.

[15]  Jimmy J. Lin,et al.  Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks , 2015, EMNLP.

[16]  Zhiguo Wang,et al.  Sentence Similarity Learning by Lexical Decomposition and Composition , 2016, COLING.

[17]  Xueqi Cheng,et al.  A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations , 2015, AAAI.