Should Kernels Be Trained in CNN?- a Paradigm of AG-Net

The Convolutional Neural Network constantly updates the weights of kernels to learn the feature representation, which makes the computational cost unaffordable. This work first proposed a Randomized Convolution Kernel with a kernel group to extract the multidimensional feature of each pixel. An AG-Net is then constructed, which can generate a layer containing multiple Gaussian Mixture Models to replace the convolutional layer. There are several Randomized Convolution Kernels in AG-Net to generate several multidimensional feature sets according to different multidimensional features. And each multidimensional feature set gets a Gaussian Mixture Model with Adaptive Resonance Theory. In training, each input is mapped by the Gaussian Mixture Models and the kernel sets. Then a fully-connected layer is used for high-level reasoning. Experiments show that the weights of kernels can be random, and the feature maps based on the similarity of pixels in multidimensional features can be well used in image processing.

[1]  Shuicheng Yan,et al.  Drop an Octave: Reducing Spatial Redundancy in Convolutional Neural Networks With Octave Convolution , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[2]  Frank L. Lewis,et al.  Optimal control using adaptive resonance theory and Q-learning , 2019, Neurocomputing.

[3]  Gang Wang,et al.  Learning fine-grained features via a CNN Tree for Large-scale Classification , 2015, Neurocomputing.

[4]  Stefan Carlsson,et al.  CNN Features Off-the-Shelf: An Astounding Baseline for Recognition , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[5]  Ting Liu,et al.  Recent advances in convolutional neural networks , 2015, Pattern Recognit..

[6]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[7]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.

[8]  S. Grossberg,et al.  ART 2: self-organization of stable category recognition codes for analog input patterns. , 1987, Applied optics.

[9]  Lixin Gao,et al.  Accelerating Expectation-Maximization Algorithms with Frequent Updates , 2012, 2012 IEEE International Conference on Cluster Computing.

[10]  Stephen Grossberg,et al.  Fuzzy ART: Fast stable learning and categorization of analog patterns by an adaptive resonance system , 1991, Neural Networks.

[11]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[12]  Ivan Laptev,et al.  Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  Jian Sun,et al.  Convolutional neural networks at constrained time cost , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Sim Heng Ong,et al.  Colour image segmentation using the self-organizing map and adaptive resonance theory , 2005, Image Vis. Comput..

[15]  Bo Yang,et al.  Generative learning for imbalanced data using the Gaussian mixed model , 2019, Appl. Soft Comput..

[16]  François Chollet,et al.  Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).