Random neural networks for deep learning

The random neural network (RNN) is a mathematical model for an “integrate and fire” spiking network that closely resembles the stochastic behaviour of neurons in mammalian brains. Since its proposal in 1989, there have been numerous investigations into the RNN’s applications and learning algorithms. Deep learning (DL) has achieved great success in machine learning, but there has been no research into the properties of the RNN for DL to combine their power. This thesis intends to bridge the gap between RNNs and DL, in order to provide a powerful tool that not only achieves state-of-the-art performance as a DL algorithm but can also be deployed in energy-efficient neomorphic hardware. Based on the RNN function approximator proposed by Gelenbe in 1999, the approximation capability of the RNN is investigated and an efficient classifier is developed. By combining the RNN, DL and non-negative matrix factorisation, new shallow and multilayer non-negative autoencoders are developed. The autoencoders are tested on typical image datasets and real-world datasets from different domains, and the test results yield the desired high learning accuracy. The concept of dense nuclei/clusters is examined, using RNN theory as a basis. In dense nuclei, neurons may interconnect via soma-to-soma interactions and conventional synaptic connections. A mathematical model of the dense nuclei is proposed and the transfer function can be deduced. A multi-layer architecture of the dense nuclei is constructed for DL, whose value is demonstrated by experiments on multi-channel datasets and server-state classification in cloud servers. The power of the standard RNN for DL is investigated. The ability of the RNN with only positive parameters to conduct image convolution operations is demonstrated. A multi-layer architecture of the RNN (MLRNN) is developed, which achieves comparable or better classification at a lower computation cost than conventional DL methods. Finally, a theoretical study into the MLRNN for DL is presented. Based on the layer-output analyses, the MLRNN is shown to be a universal function approximator. The effects of the layer number on the learning capability and high-level representation extraction are analysed. A hypothesis for transforming the DL problem into a moment-learning problem is also presented.

[1]  Erol Gelenbe,et al.  Adaptive Dispatching of Tasks in the Cloud , 2015, IEEE Transactions on Cloud Computing.

[2]  A. Gutierrez-Galvez,et al.  Signal and Data Processing for Machine Olfaction and Chemical Sensing: A Review , 2012, IEEE Sensors Journal.

[3]  Marc Teboulle,et al.  A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..

[4]  Ah Chung Tsoi,et al.  Universal Approximation Using Feedforward Neural Networks: A Survey of Some Existing Methods, and Some New Results , 1998, Neural Networks.

[5]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[6]  Erol Gelenbe,et al.  Random Neural Networks with Multiple Classes of Signals , 1999, Neural Computation.

[7]  Erol Gelenbe,et al.  Deep Learning with Dense Random Neural Networks , 2017, ICMMI.

[8]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[9]  Andreas Stafylopatis,et al.  Training the random neural network using quasi-Newton methods , 2000, Eur. J. Oper. Res..

[10]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[11]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[12]  Xin Zhang,et al.  End to End Learning for Self-Driving Cars , 2016, ArXiv.

[13]  Yuan Xu,et al.  Orthogonal Polynomials of Several Variables , 2014, 1701.02709.

[14]  B. Kröse,et al.  Bayesian Activity Recognition in Residence for Elders , 2007 .

[15]  J. Orbach Principles of Neurodynamics. Perceptrons and the Theory of Brain Mechanisms. , 1962 .

[16]  Marc-Oliver Gewaltig,et al.  NEST (NEural Simulation Tool) , 2007, Scholarpedia.

[17]  Erol Gelenbe,et al.  Learning in the Recurrent Random Neural Network , 1992, Neural Computation.

[18]  Georgios Loukas,et al.  A Denial of Service Detector based on Maximum Likelihood Detection and the Random Neural Network , 2007, Comput. J..

[19]  Erol Gelenbe,et al.  Video quality and traffic QoS in learning-based subsampled and receiver-interpolated video sequences , 2000, IEEE Journal on Selected Areas in Communications.

[20]  John H. Holland,et al.  Tests on a cell assembly theory of the action of the brain, using a large digital computer , 1956, IRE Trans. Inf. Theory.

[21]  Dimiter R. Avresky,et al.  Machine Learning for Achieving Self-* Properties and Seamless Execution of Applications in the Cloud , 2015, 2015 IEEE Fourth Symposium on Network Cloud Computing and Applications (NCCA).

[22]  Sylvain Arlot,et al.  A survey of cross-validation procedures for model selection , 2009, 0907.4728.

[23]  Erol Gelenbe,et al.  Deep Learning in Multi-Layer Architectures of Dense Nuclei , 2016, ArXiv.

[24]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[25]  Jooyoung Park,et al.  Approximation and Radial-Basis-Function Networks , 1993, Neural Computation.

[26]  Erol Gelenbe,et al.  Random network learning and image compression , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[27]  Taskin Koçak,et al.  Survey of random neural network applications , 2000, Eur. J. Oper. Res..

[28]  Erol Gelenbe,et al.  Random neural network texture model , 2000, Electronic Imaging.

[29]  Martin A. Riedmiller,et al.  A direct adaptive method for faster backpropagation learning: the RPROP algorithm , 1993, IEEE International Conference on Neural Networks.

[30]  E Gelenbe,et al.  Oscillatory corticothalamic response to somatosensory input. , 1998, Bio Systems.

[31]  Peter L. Bartlett,et al.  Neural Network Learning - Theoretical Foundations , 1999 .

[32]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[33]  Tony R. Martinez,et al.  Heterogeneous radial basis function networks , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).

[34]  H. A. Guvenir,et al.  A supervised machine learning algorithm for arrhythmia analysis , 1997, Computers in Cardiology 1997.

[35]  Terrence J. Sejnowski,et al.  Analysis of hidden units in a layered network trained to classify sonar targets , 1988, Neural Networks.

[36]  Erol Gelenbe,et al.  Area-based results for mine detection , 2000, IEEE Trans. Geosci. Remote. Sens..

[37]  Billur Barshan,et al.  Comparative study on classifying human activities with miniature inertial and magnetic sensors , 2010, Pattern Recognit..

[38]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[39]  Patrik O. Hoyer,et al.  Non-negative sparse coding , 2002, Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing.

[40]  Steve B. Furber,et al.  The SpiNNaker Project , 2014, Proceedings of the IEEE.

[41]  Nitesh V. Chawla,et al.  SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..

[42]  Guang-Bin Huang,et al.  Extreme Learning Machine for Multilayer Perceptron , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[43]  Zhang Yunong Weights and structure determination of Chebyshev-polynomial neural networks for pattern classification , 2012 .

[44]  Andrew S. Cassidy,et al.  TrueHappiness: Neuromorphic emotion recognition on TrueNorth , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[45]  Jooyoung Park,et al.  Universal Approximation Using Radial-Basis-Function Networks , 1991, Neural Computation.

[46]  Xiaojin Zhu,et al.  --1 CONTENTS , 2006 .

[47]  David Rogers,et al.  Extended-Connectivity Fingerprints , 2010, J. Chem. Inf. Model..

[48]  Romain Brette,et al.  Neuroinformatics Original Research Article Brian: a Simulator for Spiking Neural Networks in Python , 2022 .

[49]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[50]  Herman H. Goldstine,et al.  Preliminary discussion of the logical design of an electronic computing instrument (1946) , 1989 .

[51]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[52]  Ferhan Pekergin,et al.  Dynamical random neural network approach to the traveling salesman problem , 1993, Proceedings of IEEE Systems Man and Cybernetics Conference - SMC.

[53]  R. Huerta,et al.  Calibration transfer and drift counteraction in chemical sensor arrays using Direct Standardization , 2016 .

[54]  L. C. Kasun,et al.  Representational Learning with Extreme Learning Machine for Big Data Liyanaarachchi , 2022 .

[55]  H. Sebastian Seung,et al.  Learning the parts of objects by non-negative matrix factorization , 1999, Nature.

[56]  Kevin Baker,et al.  Classification of radar returns from the ionosphere using neural networks , 1989 .

[57]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[58]  Dongsheng Guo,et al.  Cross-validation based weights and structure determination of Chebyshev-polynomial neural networks for pattern classification , 2014, Pattern Recognit..

[59]  Martin T. Hagan,et al.  Comparison of Stochastic Global Optimization Methods to Estimate Neural Network Weights , 2007, Neural Processing Letters.

[60]  Steve B. Furber,et al.  Efficient SpiNNaker simulation of a heteroassociative memory using the Neural Engineering Framework , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[61]  Stelios Timotheou Nonnegative Least Squares Learning for the Random Neural Network , 2008, ICANN.

[62]  Erol Gelenbe,et al.  Cognitive Packet Network for Bilateral Asymmetric Connections , 2014, IEEE Transactions on Industrial Informatics.

[63]  O. Mangasarian,et al.  Multisurface method of pattern separation for medical diagnosis applied to breast cytology. , 1990, Proceedings of the National Academy of Sciences of the United States of America.

[64]  Simon Haykin,et al.  GradientBased Learning Applied to Document Recognition , 2001 .

[65]  Volkan Atalay,et al.  Texture Classification and Retrieval Using the Random Neural Network Model , 2006, 6th IEEE Southwest Symposium on Image Analysis and Interpretation, 2004..

[66]  Balakrishna J. Prabhu,et al.  A resource-sharing game with relative priorities , 2014, Perform. Evaluation.

[67]  Rajat Raina,et al.  Large-scale deep unsupervised learning using graphics processors , 2009, ICML '09.

[68]  Yunong Zhang,et al.  Pruning-included weights and structure determination of 2-input neuronet using Chebyshev polynomials of Class 1 , 2012, Proceedings of the 10th World Congress on Intelligent Control and Automation.

[69]  David M. Reif,et al.  Profiling Chemicals Based on Chronic Toxicity Results from the U.S. EPA ToxRef Database , 2008, Environmental health perspectives.

[70]  Zhijun Zhang,et al.  Weights and structure determination of multiple-input feed-forward neural network activated by Chebyshev polynomials of Class 2 via cross-validation , 2014, Neural Computing and Applications.

[71]  O. Mangasarian,et al.  Pattern Recognition Via Linear Programming: Theory and Application to Medical Diagnosis , 1989 .

[72]  Yi Shen,et al.  Image Segmentation Based on Random Neural Network Model and Gabor Filters , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[73]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[74]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[75]  Erol Gelenbe,et al.  Wafer surface reconstruction from top–down scanning electron microscope images , 2004 .

[76]  Léon Bottou,et al.  The Tradeoffs of Large Scale Learning , 2007, NIPS.

[77]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[78]  Max A. Little,et al.  Exploiting Nonlinear Recurrence and Fractal Scaling Properties for Voice Disorder Detection , 2007 .

[79]  Carlos Soares,et al.  A Comparison of Ranking Methods for Classification Algorithm Selection , 2000, ECML.

[80]  Andrew S. Cassidy,et al.  Cognitive computing systems: Algorithms and applications for networks of neurosynaptic cores , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[81]  Erol Gelenbe,et al.  Non-negative Autoencoder with Simplified Random Neural Network , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[82]  Y. LeCun,et al.  Learning methods for generic object recognition with invariance to pose and lighting , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[83]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[84]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[85]  K. Schittkowski,et al.  NONLINEAR PROGRAMMING , 2022 .

[86]  Erol Gelenbe,et al.  Autonomous search for mines , 1998, Eur. J. Oper. Res..

[87]  Erol Gelenbe,et al.  Single-cell based random neural network for deep learning , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[88]  James P. Bridge,et al.  Machine Learning for First-Order Theorem Proving , 2014, J. Autom. Reason..

[89]  Matthew D. Zeiler ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.

[90]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[91]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[92]  Yiran Chen,et al.  Exploring the optimal learning technique for IBM TrueNorth platform to overcome quantization loss , 2016, 2016 IEEE/ACM International Symposium on Nanoscale Architectures (NANOARCH).

[93]  Rainer Storn,et al.  Differential Evolution – A Simple and Efficient Heuristic for global Optimization over Continuous Spaces , 1997, J. Glob. Optim..

[94]  Nicholas T. Carnevale,et al.  The NEURON Book: Epilogue , 2006 .

[95]  Dimiter R. Avresky,et al.  Proactive Software Rejuvenation Based on Machine Learning Techniques , 2009, CloudComp.

[96]  Alessio Sancetta,et al.  Bernstein Approximations to the Copula Function and Portfolio Optimization , 2001 .

[97]  Billur Barshan,et al.  Human Activity Recognition Using Inertial/Magnetic Sensor Units , 2010, HBU.

[98]  Qiang Yang,et al.  Sensor-Based Abnormal Human-Activity Detection , 2008, IEEE Transactions on Knowledge and Data Engineering.

[99]  Stephen E. Wright,et al.  A simple and elementary proof of the Karush–Kuhn–Tucker theorem for inequality-constrained optimization , 2009, Optim. Lett..

[100]  Billur Barshan,et al.  Recognizing Daily and Sports Activities in Two Open Source Machine Learning Environments Using Body-Worn Sensor Units , 2014, Comput. J..

[101]  Gerardo Rubino,et al.  Levenberg - Marquardt Training Algorithms for Random Neural Networks , 2011, Comput. J..

[102]  Yan-Da Li,et al.  Approximation by random networks with bounded number of layers , 1999, Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No.98TH8468).

[103]  Erol Gelenbe,et al.  Image and video compression , 1998 .

[104]  Yu-Jin Zhang,et al.  Nonnegative Matrix Factorization: A Comprehensive Review , 2013, IEEE Transactions on Knowledge and Data Engineering.

[105]  Pierre Yger,et al.  PyNN: A Common Interface for Neuronal Network Simulators , 2008, Front. Neuroinform..

[106]  Isik Aybay,et al.  A digital neuron realization for the random neural network model , 1997, Proceedings of International Conference on Neural Networks (ICNN'97).

[107]  Zhi-Hong Mao,et al.  Function approximation with spiked random networks , 1999, IEEE Trans. Neural Networks.

[108]  Hadi Larijani,et al.  Evaluating perceived voice quality on packet networks using different random neural network architectures , 2011, Perform. Evaluation.

[109]  C. Ding,et al.  On the Equivalence of Nonnegative Matrix Factorization and K-means - Spectral Clustering , 2005 .

[110]  Erol Gelenbe,et al.  Deep learning with random neural networks , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[111]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[112]  Paul Horton,et al.  A Probabilistic Classification System for Predicting the Cellular Localization Sites of Proteins , 1996, ISMB.

[113]  Erol Gelenbe,et al.  Random Neural Networks with Negative and Positive Signals and Product Form Solution , 1989, Neural Computation.

[114]  Stelios Timotheou,et al.  The Random Neural Network: A Survey , 2010, Comput. J..

[115]  Yu Wang,et al.  Adaptive Inertia Weight Particle Swarm Optimization , 2006, ICAISC.

[116]  Taskin Koçak,et al.  Design and implementation of a random neural network routing engine , 2003, IEEE Trans. Neural Networks.

[117]  Yuxiao Hu,et al.  Learning a Spatially Smooth Subspace for Face Recognition , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[118]  Abhinav Vishnu,et al.  Deep learning for computational chemistry , 2017, J. Comput. Chem..

[119]  Erol Gelenbe,et al.  Design and performance of cognitive packet networks , 2001, Perform. Evaluation.

[120]  Jesús Favela,et al.  Activity Recognition for the Smart Hospital , 2008, IEEE Intelligent Systems.

[121]  Ming Liu,et al.  Sensor-based human activity recognition system with a multilayered model using time series shapelets , 2015, Knowl. Based Syst..

[122]  N. Metropolis,et al.  The Monte Carlo method. , 1949 .

[123]  Erol Gelenbe The first decade of G-networks , 2000, Eur. J. Oper. Res..

[124]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.

[125]  Jesse Hoey,et al.  Sensor-Based Activity Recognition , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[126]  Hai Jin,et al.  Projective Nonnegative Graph Embedding , 2010, IEEE Transactions on Image Processing.

[127]  Tan Ning,et al.  An RBF Neural Network Classifier with Centers,Variances and Weights Directly Determined , 2009 .

[128]  Zhi-Hong Mao,et al.  Function Approximation by Random Neural Networks with a Bounded Number of Layers , 2006 .

[129]  Dharmendra S. Modha,et al.  Backpropagation for Energy-Efficient Neuromorphic Computing , 2015, NIPS.

[130]  Diane J. Cook,et al.  Recognizing independent and joint activities among multiple residents in smart environments , 2010, J. Ambient Intell. Humaniz. Comput..

[131]  Erol Gelenbe,et al.  Video compression with random neural networks , 1996, Proceedings of International Workshop on Neural Networks for Identification, Control, Robotics and Signal/Image Processing.

[132]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[133]  John Pavlus THE SEARCH FOR A NEW MACHINE. , 2015, Scientific American.

[134]  Erol Gelenbe,et al.  Multi-layer neural networks for quality of service oriented server-state classification in cloud servers , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[135]  Eric O. Postma,et al.  Dimensionality Reduction: A Comparative Review , 2008 .

[136]  Luca Maria Gambardella,et al.  Deep, Big, Simple Neural Nets for Handwritten Digit Recognition , 2010, Neural Computation.

[137]  Daniel Gutierrez-Galan,et al.  Multilayer Spiking Neural Network for Audio Samples Classification Using SpiNNaker , 2016, ICANN.

[138]  Allan Pinkus,et al.  Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.

[139]  Lorenzo D'Ambrosio Extension of Bernstein polynomials to infinite dimensional case , 2006, J. Approx. Theory.

[140]  Erol Gelenbe,et al.  Random Neural Networks with Synchronized Interactions , 2008, Neural Computation.

[141]  Andrew S. Cassidy,et al.  Convolutional networks for fast, energy-efficient neuromorphic computing , 2016, Proceedings of the National Academy of Sciences.

[142]  Stelios Timotheou,et al.  A novel weight initialization method for the random neural network , 2009, Neurocomputing.

[143]  D. Dix,et al.  The ToxCast program for prioritizing toxicity testing of environmental chemicals. , 2007, Toxicological sciences : an official journal of the Society of Toxicology.

[144]  Erol Gelenbe,et al.  Random Neural Network Recognition of Shaped Objects in Strong Clutter , 1997, ICANN.

[145]  Andrea Vedaldi,et al.  MatConvNet: Convolutional Neural Networks for MATLAB , 2014, ACM Multimedia.

[146]  Erol Gelenbe,et al.  Learning in the multiple class random neural network , 2002, IEEE Trans. Neural Networks.

[147]  O. Mangasarian,et al.  Robust linear programming discrimination of two linearly inseparable sets , 1992 .

[148]  Taskin Koçak,et al.  Learning in the feed-forward random neural network: A critical review , 2011, Perform. Evaluation.

[149]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[150]  Erol Gelenbe,et al.  Stability of the Random Neural Network Model , 1990, Neural Computation.

[151]  Erol Gelenbe,et al.  Analog hardware implementation of the random neural network model , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.

[152]  Khaled F. Hussain,et al.  Laser intensity vehicle classification system based on random neural network , 2005, ACM-SE 43.

[153]  Erol Gelenbe,et al.  Large scale simulation for human evacuation and rescue , 2012, Comput. Math. Appl..

[154]  D. Perrett,et al.  Recognition of objects and their component parts: responses of single units in the temporal cortex of the macaque. , 1994, Cerebral cortex.

[155]  Chris H. Q. Ding,et al.  Orthogonal nonnegative matrix t-factorizations for clustering , 2006, KDD '06.

[156]  Erol Gelenbe,et al.  Big Data for Autonomic Intercontinental Overlays , 2016, IEEE Journal on Selected Areas in Communications.

[157]  H. White,et al.  Universal approximation using feedforward networks with non-sigmoid hidden layer activation functions , 1989, International 1989 Joint Conference on Neural Networks.

[158]  Fernando Nogueira,et al.  Imbalanced-learn: A Python Toolbox to Tackle the Curse of Imbalanced Datasets in Machine Learning , 2016, J. Mach. Learn. Res..

[159]  Victor C. M. Leung,et al.  Extreme Learning Machines [Trends & Controversies] , 2013, IEEE Intelligent Systems.