Accurate classification of tumor text based on attention mechanism and transfer learning

With the increase in the number of electronic medical records, precise classification of medical records can be use to accurately diagnose diseases. In medical texts, there are differences in the number of texts of different diseases, especially for specific diseases and a small number of samples have brought great challenges to medical text classification. In response to this problem, this paper proposes an accurate text classification model base on transfer learning combined with attention mechanism neural network, and finally obtains accurate category text through two-step classification. The model first uses the attention mechanism long-term and short-term cyclic neural network to extract the overall tumor sample from the mass of unbalanced medical record texts through the common characteristics of the tumor medical record, and uses the similarity between diseases to combine with the convolutional neural network through migration learning. The characteristics of each type of tumor disease are migrate to achieve the effect of precise classification training. The precise training model is used to finally realize the precise classification of tumor diseases. The model was tested on a private tumor medical data set, and the results showed that the accuracy of the tumor data set (precision) was increased by 5% compare to the average of the baseline optimal classification model. The increase in accuracy is important for accurate classification significance.

[1]  Wei Shi,et al.  Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification , 2016, ACL.

[2]  Tomas Mikolov,et al.  Bag of Tricks for Efficient Text Classification , 2016, EACL.

[3]  Sebastian Ruder,et al.  Universal Language Model Fine-tuning for Text Classification , 2018, ACL.

[4]  Alex Graves,et al.  Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.

[5]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[6]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[7]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[8]  Yann LeCun,et al.  Very Deep Convolutional Networks for Text Classification , 2016, EACL.

[9]  Bo Huang,et al.  A New Method of Region Embedding for Text Classification , 2018, ICLR.

[10]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[11]  Grigorios Tsoumakas,et al.  Multi-Label Classification: An Overview , 2007, Int. J. Data Warehous. Min..

[12]  Jesús Alcalá-Fdez,et al.  KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework , 2011, J. Multiple Valued Log. Soft Comput..

[13]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[14]  Charles Elkan,et al.  Learning to Diagnose with LSTM Recurrent Neural Networks , 2015, ICLR.