Differential convolutional neural network

Convolutional neural networks with strong representation ability of deep structures have ever increasing popularity in many research areas. The main difference of Convolutional Neural Networks with respect to existing similar artificial neural networks is the inclusion of the convolutional part. This inclusion directly increases the performance of artificial neural networks. This fact has led to the development of many different convolutional models and techniques. In this work, a novel convolution technique named as Differential Convolution and updated error back-propagation algorithm is proposed. The proposed technique aims to transfer feature maps containing directional activation differences to the next layer. This implementation takes the idea of how convolved features change on the feature map into consideration. In a sense, this process adapts the mathematical differentiation operation into the convolutional process. Proposed improved back propagation algorithm also considers neighborhood activation errors. This property increases the classification performance without changing the number of filters. Four different experiment sets were performed to observe the performance and the adaptability of the differential convolution technique. In the first experiment set utilization of the differential convolution on a traditional convolutional neural network structure made a performance boost up to 55.29% for the test accuracy. In the second experiment set differential convolution adaptation raised the top1 and top5 test accuracies of AlexNet by 5.3% and 4.75% on ImageNet dataset. In the third experiment set differential convolution utilized model outperformed all compared convolutional structures. In the fourth experiment set, the Differential VGGNet model obtained by adapting proposed differential convolution technique performed 93.58% and 75.06% accuracy values for CIFAR10 and CIFAR100 datasets, respectively. The accuracy values of the Differential NIN model containing differential convolution operation were 92.44% and 72.65% for the same datasets. In these experiment sets, it was observed that the differential convolution technique outperformed both traditional convolution and other compared convolution techniques. In addition, easy adaptation of the proposed technique to different convolutional structures and its efficiency demonstrate that popular deep learning models may be improved with differential convolution.

[1]  Cordelia Schmid,et al.  Semi-Local Affine Parts for Object Recognition , 2004, BMVC.

[2]  Antonio Criminisi,et al.  Object categorization by learned universal visual dictionary , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[3]  Xiang Zhang,et al.  OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks , 2013, ICLR.

[4]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[5]  Qiang Chen,et al.  Network In Network , 2013, ICLR.

[6]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[7]  Leon O. Chua,et al.  Cellular neural networks: applications , 1988 .

[8]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[9]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[10]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  George E. Dahl Deep Learning Approaches to Problems in Speech Recognition, Computational Chemistry, and Natural Language Text Processing , 2015 .

[12]  Patrick Mäder,et al.  Plant species classification using flower images—A comparative study of local feature representations , 2017, PloS one.

[13]  Oskar Söderkvist,et al.  Computer Vision Classification of Leaves from Swedish Trees , 2001 .

[14]  Cordelia Schmid,et al.  A maximum entropy framework for part-based texture and object recognition , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[15]  Katja Markert,et al.  Learning Models for Object Recognition from Natural Language Descriptions , 2009, BMVC.

[16]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[17]  Quoc V. Le,et al.  Tiled convolutional neural networks , 2010, NIPS.

[18]  Ronald Davis,et al.  Neural networks and deep learning , 2017 .

[19]  Jan-Olof Eklundh,et al.  Joint visual vocabulary for animal classification , 2008, 2008 19th International Conference on Pattern Recognition.

[20]  Gregory K. Wallace,et al.  The JPEG still picture compression standard , 1992 .

[21]  Lawrence D. Jackel,et al.  Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.

[22]  Guang-Zhong Yang,et al.  Deep Learning for Health Informatics , 2017, IEEE Journal of Biomedical and Health Informatics.

[23]  Andrew Zisserman,et al.  A Visual Vocabulary for Flower Classification , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[24]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[25]  T Chandrakumar,et al.  Classifying Diabetic Retinopathy using Deep Learning Architecture , 2016 .

[26]  Vladlen Koltun,et al.  Multi-Scale Context Aggregation by Dilated Convolutions , 2015, ICLR.

[27]  Marc Teboulle,et al.  A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..