Semi-random deep neural networks for near real-time target classification

In recent years deep neural networks have shown great advances in image processing tasks. For modern datasets, these networks require long training times due to backpropagation, high amount of computational resources for weight updates, and memory intensive weight storage. Exploiting randomness during the training of deep neural networks can mitigate these concerns by reducing the computational costs without sacrificing network performance. However, a fully randomized network has limitations for real-time target classification as it leads to poor performance. Therefore we are motivated in using semi-random deep neural networks to exploit random fixed weights. In this paper, we demonstrate that semi-random deep neural networks can achieve near real-time training with comparable accuracies to conventional deep neural networks models. We find that these networks are enhanced by the usage of skip connections and train rapidly at the cost of dense memory usage. With greater memory resources available, these networks can train on larger datasets at a fraction of the training time costs. These semi-random deep neural network architectures open up an avenue for further research in utilizing random fixed weights in neural networks.

[1]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[2]  Guang-Bin Huang,et al.  Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[3]  Yujun Zeng,et al.  Traffic Sign Recognition Using Extreme Learning Classifier with Deep Convolutional Features , 2015 .

[4]  Kaiming He,et al.  Mask R-CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[5]  Ruslan Salakhutdinov,et al.  Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.

[6]  Nicolas Pinto,et al.  Beyond simple features: A large-scale feature search approach to unconstrained face recognition , 2011, Face and Gesture 2011.

[7]  Herbert Jaeger,et al.  The''echo state''approach to analysing and training recurrent neural networks , 2001 .

[8]  James Demmel,et al.  ImageNet Training in Minutes , 2017, ICPP.

[9]  Le Zhang,et al.  Visual Tracking With Convolutional Random Vector Functional Link Network , 2017, IEEE Transactions on Cybernetics.

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  Geoffrey E. Hinton,et al.  Large scale distributed neural network training through online distillation , 2018, ICLR.

[12]  Jin Chen,et al.  A hybrid convolutional neural networks with extreme learning machine for WCE image classification , 2015, 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[13]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[14]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[15]  H. D. Block The perceptron: a model for brain functioning. I , 1962 .

[16]  Henry Markram,et al.  On the computational power of circuits of spiking neurons , 2004, J. Comput. Syst. Sci..

[17]  Xinyi Yang,et al.  Deep Convolutional Extreme Learning Machine and Its Application in Handwritten Digit Classification , 2016, Comput. Intell. Neurosci..

[18]  AI Koan,et al.  Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning , 2008, NIPS.

[19]  Junmo Kim,et al.  Deep Pyramidal Residual Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[21]  Luca Antiga,et al.  Automatic differentiation in PyTorch , 2017 .

[22]  Yoshua Bengio,et al.  BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 , 2016, ArXiv.

[23]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Hanan Samet,et al.  Training Quantized Nets: A Deeper Understanding , 2017, NIPS.

[25]  P. N. Suganthan,et al.  A comprehensive evaluation of random vector functional link networks , 2016, Inf. Sci..

[26]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[27]  Dejan J. Sobajic,et al.  Neural-net computing and the intelligent control of systems , 1992 .

[28]  Haipeng Wang,et al.  Application of deep-learning algorithms to MSTAR data , 2015, 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS).

[29]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[30]  Alexander J. Smola,et al.  Parallelized Stochastic Gradient Descent , 2010, NIPS.

[31]  Gang Sun,et al.  Squeeze-and-Excitation Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[32]  Eric R. Keydel,et al.  MSTAR extended operating conditions: a tutorial , 1996, Defense, Security, and Sensing.

[33]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Bernard Widrow,et al.  The No-Prop algorithm: A new learning algorithm for multilayer neural networks , 2013, Neural Networks.

[35]  Zhenghao Chen,et al.  On Random Weights and Unsupervised Feature Learning , 2011, ICML.