Deep Convolutional Neural Network with Dilated Convolution Using Small Size Dataset

Since Alex Krizhevsky won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2012 competition by building a very intelligent deep convolutional neural network (D-CNNs), more and more researchers have been engaged in the research and development of deep convolutional neural network (D-CNNs). However, recent researches on deep convolutional neural networks are mostly based on ImageNet datasets. The network model based on such a large dataset is mostly blind to increase the number of network layers, ignoring that most data sets in application are far from the order of magnitude of ImageNet datasets. Such deep networks tend to perform poorly in small datasets (CIFAR-10), since deep models are easy to overfitting. In this paper, we’ve applied some of the more efficient methods that have been proposed in recent years to traditional deep convolutional neural networks. We proposed a modified Alex network and used this model to fit CIFAR-10. By adding Batch Normalization, using Dilated Convolution and replacing Fully Connected layer (FC) with Global Average Pooling (GAP), we achieved 8.6% error rate on CIFAR-10 without severe overfitting. Our results show that the deep CNN can be used to fit small datasets with proper modifications and the results are much better than before.

[1]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[2]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[4]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[5]  Vladlen Koltun,et al.  Multi-Scale Context Aggregation by Dilated Convolutions , 2015, ICLR.

[6]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[7]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[8]  Yann LeCun,et al.  Regularization of Neural Networks using DropConnect , 2013, ICML.

[9]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[10]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[11]  Kaiming He,et al.  Deep Residual Learning for Image Recognition Supplementary Materials , 2016 .

[12]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[13]  Wangmeng Zuo,et al.  Learning Deep CNN Denoiser Prior for Image Restoration , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).