Will Transfer Learning Enhance ImageNet Classification Accuracy Using ImageNet-Pretrained Models?

The huge impact of Transfer Learning (TL) techniques in many fields was achieved using several state-of-the-art ImageNet-pretrained models. These models have shown great performance improvements on this dataset over the last few years. One of the recently used TL techniques is feature extraction with the help of Feature Concatenation (FC), where the extracted features of multiple pretrained models are concatenated together before training on them, to produce more robust and discriminative feature representations on various classification tasks. However, neither TL nor FC techniques have been tested on the same dataset that initially trained the pretrained models, i.e. ImageNet. Hence, this work provides an investigative study to test the possibility of improving the ImageNet accuracy using the feature extraction approach of TL with the help of FC techniques. The results of this work show that there is no TL technique that can be used with or without FC to increase the accuracy of pretrained models on the original dataset on which they were trained. Even for FC, it cannot produce a more discriminative feature representation for the original data than what the individual models can produce.

[1]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[2]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  D. Hubel,et al.  Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.

[4]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.

[5]  Zhiping Lin,et al.  Deep CNNs for microscopic image classification by exploiting transfer learning and feature concatenation , 2018, 2018 IEEE International Symposium on Circuits and Systems (ISCAS).

[6]  Edoardo Serra,et al.  Convolutional Neural Network Ensemble Fine-Tuning for Extended Transfer Learning , 2018, BigData Congress.

[7]  Trevor Darrell,et al.  DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition , 2013, ICML.

[8]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[9]  Mahmoud Al-Ayyoub,et al.  Determine Bipolar Disorder Level from Patient Interviews Using Bi-LSTM and Feature Fusion , 2018, 2018 Fifth International Conference on Social Networks Analysis, Management and Security (SNAMS).

[10]  Samy Bengio,et al.  Links between perceptrons, MLPs and SVMs , 2004, ICML.

[11]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Terrence J. Sejnowski,et al.  Parallel Networks that Learn to Pronounce English Text , 1987, Complex Syst..

[13]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[14]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[15]  Stefan Carlsson,et al.  CNN Features Off-the-Shelf: An Astounding Baseline for Recognition , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[16]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[17]  Quoc V. Le,et al.  Do Better ImageNet Models Transfer Better? , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Bernhard Schölkopf,et al.  Estimating the Support of a High-Dimensional Distribution , 2001, Neural Computation.

[19]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[20]  Geoffrey E. Hinton Learning Translation Invariant Recognition in Massively Parallel Networks , 1987, PARLE.

[21]  J. Orbach Principles of Neurodynamics. Perceptrons and the Theory of Brain Mechanisms. , 1962 .

[22]  Mahmoud Al-Ayyoub,et al.  Performance study of augmentation techniques for HEp2 CNN classification , 2018, 2018 9th International Conference on Information and Communication Systems (ICICS).

[23]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[24]  Jian Sun,et al.  Identity Mappings in Deep Residual Networks , 2016, ECCV.

[25]  Chih-Jen Lin,et al.  A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.

[26]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).