Regression Task on Big Data with Convolutional Neural Network

As one of most widely utilized methods in deep learning, convolutional neural network (CNN) has been proven effective in many machine learning applications, especially in the areas of image understanding and computer vision. However, CNN is mainly used for applications with the approach of classification, while its usage for regression is not well-studied. In this work, we propose a strategy based on CNN with Visual Geometry Group Network (VGG) for image regression task. We have applied this method on images of MNIST processed with labels of continuous number. In our study, the original discrete classes of handwriting numbers are converted into float numbers with respect to normal distribution, thereby the traditional classification task in MNIST becomes a regression one. In our study, different loss functions such as Mean Absolute Error (MAE) and Log-cosh have been applied and validated. Final results generated by model trained with CNN with VGG with 10-fold cross-validation can be obtained, where MAE is less than 0.25, compared to the much higher error of around 3 with the use of other loss functions and convolutional layers. The significantly reduced error suggests the applicability of our proposed method.

[1]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[2]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[3]  C. Willmott,et al.  Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance , 2005 .

[4]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[5]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[6]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[7]  Ron Kohavi,et al.  A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection , 1995, IJCAI.

[8]  Sepp Hochreiter,et al.  Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.

[9]  Antonio Plaza,et al.  A new deep convolutional neural network for fast hyperspectral image classification , 2017, ISPRS Journal of Photogrammetry and Remote Sensing.

[10]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[11]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[12]  Kunihiko Fukushima,et al.  Neocognitron: A Self-Organizing Neural Network Model for a Mechanism of Visual Pattern Recognition , 1982 .

[13]  D. Hubel,et al.  Early Exploration of the Visual Cortex , 1998, Neuron.

[14]  Arthur L. Samuel,et al.  Some studies in machine learning using the game of checkers , 2000, IBM J. Res. Dev..

[15]  Patrice Y. Simard,et al.  Best practices for convolutional neural networks applied to visual document analysis , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..

[16]  Robert Hecht-Nielsen III.3 – Theory of the Backpropagation Neural Network* , 1992 .