A Novel Training Mechanism for Extending Convolutional Neural Network

Convolutional Neural Network (CNN) has obtained great success in the computer vision domain in the recent years. These CNN models adopt deeper neural network architecture to achieve high recognition accuracy, the training costs of time and dataset are dramatically increased. While the recognizing categories are expanded, the CNN architecture needs to be modified, the whole CNN model requires to be retrained. Transfer learning method is adopted to save the training cost by migrating part of learned weights, from the existed CNN model to the target CNN model with expanded recognizing categories. However, the requirement of modifying neural network architecture still consumes huge amount of the training cost. This paper presents a new training mechanism, called Extended Learning, to solve the above problems. By using the proposed Partially Back-Propagation Operation, the CNN model can expand new classification categories without modifying the architecture of the CNN model, the learning weights from previously training results can be retained, the training cost of time and dataset can be reduced accordingly. The experimental result shows that the proposed extended learning method can save 16.7% training image count compared to the transfer learning method, with the target accuracy of 0.75.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[3]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[4]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[5]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[7]  Ivan Laptev,et al.  Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[8]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[10]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[11]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[12]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[13]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .