Dendrite morphological neurons trained by stochastic gradient descent

Dendrite morphological neurons are a type of artificial neural network that work with min and max operators instead of algebraic products. These morphological operators allow each dendrite to build a hyper-box in classification N-dimensional space. In contrast with classical perceptrons, these simple geometrical representations, hyper-boxes, allow the proposal of training methods based on heuristics without using of an optimisation method. In literature, it has been claimed that these heuristics-based trainings have advantages: there are no convergence problems, perfect classification can always be reached and training is performed in only one epoch. This paper shows that these assumed advantages come with a cost: these heuristics increase classification errors in the test set because they are not optimal and learning generalisation is poor. To solve these problems, we introduce a novel method to train dendrite morphological neurons based on stochastic gradient descent for classification tasks, using these heuristics just for initialisation of learning parameters. We add a softmax layer to the neural architecture for calculating gradients and also propose and evaluate four different methods to initialise the dendrite parameters. Experiments are performed based on several real and synthetic datasets. Results show that we can enhance the testing accuracy in comparison with solely heuristics-based training methods. This approach reaches competitive performance with respect to other popular machine learning algorithms. Our code developed in Matlab is available online.

[1]  Aboul Ella Hassanien,et al.  Soft Computing Models in Industrial and Environmental Applications, 6th International Conference SOCO 2011, 6-8 April, 2011, Salamanca, Spain , 2011, SOCO.

[2]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[3]  Gerhard X. Ritter,et al.  Noise Masking for Pattern Recall Using a Single Lattice Matrix Auto-Associative Memory , 2006, 2006 IEEE International Conference on Fuzzy Systems.

[4]  Jennifer L. Davidson,et al.  Morphology neural networks: An introduction with applications , 1993 .

[5]  Peter Sussner,et al.  An introduction to morphological perceptrons with competitive learning , 2009, 2009 International Joint Conference on Neural Networks.

[6]  Randall K Weinstein,et al.  Architectures for high-performance FPGA implementations of neural models , 2006, Journal of neural engineering.

[7]  Jennifer L. Davidson,et al.  Template learning in morphological neural nets , 1991, Optics & Photonics.

[8]  J. van Leeuwen,et al.  Neural Networks: Tricks of the Trade , 2002, Lecture Notes in Computer Science.

[9]  Gerhard X. Ritter,et al.  Lattice algebra approach to single-neuron computation , 2003, IEEE Trans. Neural Networks.

[10]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[11]  Peter Sussner,et al.  An introduction to morphological neural networks , 1996, Proceedings of 13th International Conference on Pattern Recognition.

[12]  Gabriella Sanniti di Baja,et al.  Pattern recognition : 5th Mexican Conference, MCPR 2013, Querétaro, Mexico, June 26-29, 2013, proceedings , 2013 .

[13]  Gerhard X. Ritter,et al.  Orthonormal Basis Lattice Neural Networks , 2006, 2006 IEEE International Conference on Fuzzy Systems.

[14]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[15]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[16]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[17]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[18]  David S. Broomhead,et al.  Multivariable Functional Interpolation and Adaptive Networks , 1988, Complex Syst..

[19]  Gerhard X. Ritter,et al.  Learning In Lattice Neural Networks that Employ Dendritic Computing , 2006, FUZZ-IEEE.

[20]  Montavon,et al.  [Lecture Notes in Computer Science] Neural Networks: Tricks of the Trade Volume 7700 || Deep Learning via Semi-supervised Embedding , 2012 .

[21]  Hermann Ney,et al.  Cross-entropy vs. squared error training: a theoretical and experimental comparison , 2013, INTERSPEECH.

[22]  Peter Sussner,et al.  Morphological perceptron learning , 1998, Proceedings of the 1998 IEEE International Symposium on Intelligent Control (ISIC) held jointly with IEEE International Symposium on Computational Intelligence in Robotics and Automation (CIRA) Intell.

[23]  D. Broomhead,et al.  Radial Basis Functions, Multi-Variable Functional Interpolation and Adaptive Networks , 1988 .

[24]  G. Kane Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol 1: Foundations, vol 2: Psychological and Biological Models , 1994 .

[25]  Juan Humberto Sossa Azuela,et al.  New Radial Basis Function Neural Network Architecture for Pattern Classification: First Results , 2014, CIARP.

[26]  Juan Humberto Sossa Azuela,et al.  Modified Dendrite Morphological Neural Network Applied to 3D Object Recognition , 2013, MCPR.

[27]  Gerhard X. Ritter,et al.  Recursion and Feedback in Image Algebra , 1991 .

[28]  Douglas Kline,et al.  Revisiting squared-error and cross-entropy functions for training neural network classifiers , 2005, Neural Computing & Applications.

[29]  Manuel Graña,et al.  Optimal Hyperbox Shrinking in Dendritic Computing Applied to Alzheimer's Disease Detection in MRI , 2011, SOCO.

[30]  Peter Sussner,et al.  Constructive Morphological Neural Networks: Some Theoretical Aspects and Experimental Results in Classification , 2009, Constructive Neural Networks.

[31]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[32]  Gerhard X. Ritter,et al.  Theory of morphological neural networks , 1990, Photonics West - Lasers and Applications in Science and Engineering.

[33]  Gerhard X. Ritter,et al.  Computational Intelligence Based on Lattice Theory , 2007, Studies in Computational Intelligence.

[34]  Gerhard X. Ritter,et al.  Two lattice metrics dendritic computing for pattern recognition , 2014, 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE).

[35]  Peter Sussner,et al.  Morphological perceptrons with competitive learning: Lattice-theoretical framework and constructive learning algorithm , 2011, Inf. Sci..

[36]  Juan Humberto Sossa Azuela,et al.  Efficient training for dendrite morphological neural networks , 2014, Neurocomputing.

[37]  Ricardo de A. Araújo A morphological perceptron with gradient-based learning for Brazilian stock market forecasting , 2012, Neural Networks.

[38]  Gerhard X. Ritter,et al.  Reconstruction of Patterns from Noisy Inputs Using Morphological Associative Memories , 2003, Journal of Mathematical Imaging and Vision.