Semi-supervised text classification with deep convolutional neural network using feature fusion approach

Supervised learning algorithms employ labeled training data for classification purposes while obtaining labeled data for large datasets is costly and time consuming. Semi-supervised learning algorithms, on the contrary, use a small set of labeled data and a large set of unlabeled data to improve predication performance and thus may be a good alternative to supervised learning algorithms for large text datasets. Although many semi-supervised learning algorithms have been proposed in the data science literature, most of these algorithms are not feasible for discrete and unstructured text data.This paper aims to improve classification accuracy of semi-supervised learning algorithms applied to text data. To achieve this goal, a novel design for convolutional neural network is employed in a co-training semi-supervised learning algorithm which adds augmented data as the second input of the convolutional neural network to predict labels of text data. we also propose a novel approach for partitioning the dataset into independent views via topic modeling to train independent classifiers. In so doing, neighbour classifiers are found and confident predictions of unlabeled data are fused into labeled data. The prediction accuracy of the combined algorithm is then compared to the state-of-the-art supervised and semi-supervised learning algorithms. Our findings show that the proposed combined algorithm outperforms the supervised and semi-supervised algorithms in terms of prediction accuracy. CCS CONCEPTS• Information systems → Content analysis and feature selection.

[1]  Wenqing Sun,et al.  Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data , 2017, Comput. Medical Imaging Graph..

[2]  Tong Zhang,et al.  Semi-supervised Convolutional Neural Networks for Text Categorization via Region Embedding , 2015, NIPS.

[3]  Yong Shi,et al.  Laplacian twin support vector machine for semi-supervised classification , 2012, Neural Networks.

[4]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[5]  Leif E. Peterson K-nearest neighbor , 2009, Scholarpedia.

[6]  Jun Zhao,et al.  Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.

[7]  Xiaojin Zhu,et al.  --1 CONTENTS , 2006 .

[8]  O. Chapelle,et al.  Semi-Supervised Learning (Chapelle, O. et al., Eds.; 2006) [Book reviews] , 2009, IEEE Transactions on Neural Networks.

[9]  Tong Zhang,et al.  Supervised and Semi-Supervised Text Categorization using LSTM for Region Embeddings , 2016, ICML.

[10]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[11]  Fei-Fei Li,et al.  Large-Scale Video Classification with Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Xiaojun Wan,et al.  Co-Training for Cross-Lingual Sentiment Classification , 2009, ACL.

[13]  Xueqi Cheng,et al.  Adaptive co-training SVM for sentiment classification on tweets , 2013, CIKM.

[14]  Thorsten Joachims,et al.  Counterfactual Learning-to-Rank for Additive Metrics and Deep Models , 2018, ArXiv.

[15]  Hamideh Afsarmanesh,et al.  Semi-supervised self-training for decision tree classifiers , 2017, Int. J. Mach. Learn. Cybern..

[16]  Francisco Herrera,et al.  On the characterization of noise filters for self-training semi-supervised in nearest neighbor classification , 2014, Neurocomputing.

[17]  Hal Daumé,et al.  A Co-training Approach for Multi-view Spectral Clustering , 2011, ICML.

[18]  Phil Blunsom,et al.  A Convolutional Neural Network for Modelling Sentences , 2014, ACL.

[19]  Quoc V. Le,et al.  Semi-supervised Sequence Learning , 2015, NIPS.

[20]  Dacheng Tao,et al.  A Survey on Multi-view Learning , 2013, ArXiv.

[21]  Tong Zhang,et al.  Effective Use of Word Order for Text Categorization with Convolutional Neural Networks , 2014, NAACL.

[22]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[23]  Volkmar Frinken,et al.  Keyword spotting for self-training of BLSTM NN based handwriting recognition systems , 2014, Pattern Recognit..

[24]  Eirini Ntoutsi,et al.  Large Scale Sentiment Learning with Limited Labels , 2017, KDD.