Multi-Scale Attentive Interaction Networks for Chinese Medical Question Answer Selection

The past few years have witnessed a trend that the deep learning techniques have been increasingly applied in healthcare due to the explosive growth of big data. The online medical community, where users can ask qualified doctors about medical questions with just a few keystrokes and mouse clicks anytime and anywhere, has become quite popular recently. In this paper, we investigate the problem of Chinese medical question answer selection, which is a crucial subtask of automatic question answering and fairly challenging because of its language and domain characteristics. We introduce an end-to-end multi-scale interactive networks framework to address the issue. The framework consists of several multi-scale deep neural layers which extract the deep semantic information of medical text from different levels of granularity, shortcut connections which prevent network degradation problem, and attentive interaction which mines the correlation between questions and answers. To evaluate our framework, we update and expand a dataset called cMedQA v2.0. Experimental results demonstrate that our model outperforms the existing state-of-the-art models with noticeable margins.

[1]  Demis Hassabis,et al.  Mastering the game of Go without human knowledge , 2017, Nature.

[2]  Tripti Dodiya,et al.  Rule Based Architecture for Medical Question Answering System , 2012, SocProS.

[3]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[4]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[5]  Dongyan Zhao,et al.  Hybrid Question Answering over Knowledge Base and Free Text , 2016, COLING.

[6]  Houfeng Wang,et al.  Attentive Interactive Neural Networks for Answer Selection in Community Question Answering , 2017, AAAI.

[7]  Zhaoyun Ding,et al.  Chinese Medical Question Answer Matching Using End-to-End Character-Level Multi-Scale CNNs , 2017 .

[8]  Xuanjing Huang,et al.  Convolutional Neural Tensor Network Architecture for Community-Based Question Answering , 2015, IJCAI.

[9]  Bowen Zhou,et al.  LSTM-based Deep Learning Models for non-factoid answer selection , 2015, ArXiv.

[10]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[11]  Zhiguo Wang,et al.  Bilateral Multi-Perspective Matching for Natural Language Sentences , 2017, IJCAI.

[12]  Xin Zhang,et al.  Multi-level Composite Neural Networks for Medical Question Answer Matching , 2018, 2018 IEEE Third International Conference on Data Science in Cyberspace (DSC).

[13]  Yu Zhang,et al.  Highway long short-term memory RNNS for distant speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[14]  Mohit Bansal,et al.  Shortcut-Stacked Sentence Encoders for Multi-Domain Inference , 2017, RepEval@EMNLP.

[15]  Bowen Zhou,et al.  Attentive Pooling Networks , 2016, ArXiv.

[16]  Bowen Zhou,et al.  Applying deep learning to answer selection: A study and an open task , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).

[17]  Li Li,et al.  Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records , 2016, Scientific Reports.

[18]  Jason Weston,et al.  A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.

[19]  Bram van Ginneken,et al.  A survey on deep learning in medical image analysis , 2017, Medical Image Anal..

[20]  Phil Blunsom,et al.  A Convolutional Neural Network for Modelling Sentences , 2014, ACL.

[21]  Zhiguo Wang,et al.  A Unified Query-based Generative Model for Question Generation and Question Answering , 2017, ArXiv.

[22]  Phil Blunsom,et al.  Teaching Machines to Read and Comprehend , 2015, NIPS.

[23]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[24]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[25]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[26]  Yoshimasa Tsuruoka,et al.  A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks , 2016, EMNLP.

[27]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Rajarshi Das,et al.  Question Answering on Knowledge Bases and Text using Universal Schema and Memory Networks , 2017, ACL.

[29]  Ting Liu,et al.  Attention-over-Attention Neural Networks for Reading Comprehension , 2016, ACL.

[30]  Ali Farhadi,et al.  Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.

[31]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[32]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[33]  Hugo Zaragoza,et al.  The Probabilistic Relevance Framework: BM25 and Beyond , 2009, Found. Trends Inf. Retr..

[34]  Bowen Zhou,et al.  Improved Representation Learning for Question Answer Matching , 2016, ACL.