Deep belief networks and cortical algorithms: A comparative study for supervised classification

Abstract The failure of shallow neural network architectures in replicating human intelligence led the machine learning community to focus on deep learning, to computationally match human intelligence. The wide availability of increasing computing power coupled with the development of more efficient training algorithms have allowed the implementation of deep learning principles in a manner and span that had not been previously possible. This has led to the inception of deep architectures that capitalize on recent advances in artificial intelligence and insights from cognitive neuroscience to provide better learning solutions. In this paper, we discuss two such algorithms that represent different approaches to deep learning with varied levels of maturity. The more mature but less biologically inspired Deep Belief Network (DBN) and the more biologically grounded Cortical Algorithms (CA) are first introduced to give readers a bird’s eye view of the higher-level concepts that make up these algorithms, as well as some of their technical underpinnings and applications. Their theoretical computational complexity is then derived before comparing their empirical performance on some publicly available classification datasets. Multiple network architectures were compared and showed that CA outperformed DBN on most datasets, with the best network architecture consisting of six hidden layers.

[1]  Ben Goertzel,et al.  A world survey of artificial brain projects, Part I: Large-scale brain simulations , 2010, Neurocomputing.

[2]  Geoffrey E. Hinton,et al.  To recognize shapes, first learn to generate images. , 2007, Progress in brain research.

[3]  Sepp Hochreiter,et al.  Untersuchungen zu dynamischen neuronalen Netzen , 1991 .

[4]  F. Craik,et al.  Cognition through the lifespan: mechanisms of change , 2006, Trends in Cognitive Sciences.

[5]  Teuvo Kohonen,et al.  Self-organized formation of topologically correct feature maps , 2004, Biological Cybernetics.

[6]  N. Geschwind Specializations of the human brain. , 1979, Scientific American.

[7]  Richard Granger,et al.  Novel speech processing mechanism derived from auditory neocortical circuit analysis , 1996, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96.

[8]  Yoshua Bengio,et al.  Gradient Flow in Recurrent Nets: the Difficulty of Learning Long-Term Dependencies , 2001 .

[9]  Sandhya Samarasinghe,et al.  Neural Networks for Applied Sciences and Engineering: From Fundamentals to Complex Pattern Recognition , 2006 .

[10]  Kunihiko Fukushima,et al.  Neocognitron: A hierarchical neural network capable of visual pattern recognition , 1988, Neural Networks.

[11]  Michael C. Anderson,et al.  The Role of Inhibition in Learning , 2008 .

[12]  Geoffrey E. Hinton,et al.  Acoustic Modeling Using Deep Belief Networks , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[13]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[14]  Philipp Slusallek,et al.  Introduction to real-time ray tracing , 2005, SIGGRAPH Courses.

[15]  Yoshua Bengio,et al.  Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.

[16]  Geoffrey E. Hinton Deep belief networks , 2009, Scholarpedia.

[17]  Douglas Eck,et al.  Learning Features from Music Audio with Deep Belief Networks , 2010, ISMIR.

[18]  Isaac Meilijson,et al.  Synaptic Pruning in Development: A Computational Account , 1998, Neural Computation.

[19]  Yoshua Bengio,et al.  Inference for the Generalization Error , 1999, Machine Learning.

[20]  Pascal Vincent,et al.  The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training , 2009, AISTATS.

[21]  John R. Anderson ACT: A simple theory of complex cognition. , 1996 .

[22]  G. Shepherd The Synaptic Organization of the Brain , 1979 .

[23]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[24]  J. Szentágothai The Ferrier Lecture, 1977 The neuron network of the cerebral cortex: a functional interpretation , 1978, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[25]  Ben Goertzel,et al.  A world survey of artificial brain projects, Part II: Biologically inspired cognitive architectures , 2010, Neurocomputing.

[26]  Geoffrey E. Hinton,et al.  3D Object Recognition with Deep Belief Nets , 2009, NIPS.

[27]  Christian Igel,et al.  An Introduction to Restricted Boltzmann Machines , 2012, CIARP.

[28]  M. Aizerman,et al.  Theoretical Foundations of the Potential Function Method in Pattern Recognition Learning , 1964 .

[29]  Jason Weston,et al.  Curriculum learning , 2009, ICML '09.

[30]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[31]  Eduardo Ros,et al.  Modeling the Cerebellar Microcircuit: New Strategies for a Long-Standing Issue , 2016, Front. Cell. Neurosci..

[32]  Nadine Hajj,et al.  Weighted entropy cortical algorithms for isolated Arabic speech recognition , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[33]  Tara N. Sainath,et al.  Making Deep Belief Networks effective for large vocabulary continuous speech recognition , 2011, 2011 IEEE Workshop on Automatic Speech Recognition & Understanding.

[34]  H. Lagercrantz,et al.  The functional architecture of the infant brain as revealed by resting-state fMRI. , 2011, Cerebral cortex.

[35]  Xiaolong Wang,et al.  Active Deep Networks for Semi-Supervised Sentiment Classification , 2010, COLING.

[36]  V. Mountcastle The columnar organization of the neocortex. , 1997, Brain : a journal of neurology.

[37]  Stan Franklin,et al.  THE LIDA ARCHITECTURE: ADDING NEW MODES OF LEARNING TO AN INTELLIGENT, AUTONOMOUS, SOFTWARE AGENT , 2006 .

[38]  Dong Yu,et al.  Context-Dependent Pre-Trained Deep Neural Networks for Large-Vocabulary Speech Recognition , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[39]  Nadine Hajj,et al.  A MapReduce Cortical Algorithms Implementation for Unsupervised Learning of Big Data , 2015, INNS Conference on Big Data.

[40]  John M. DeSesso Functional Anatomy of the Brain , 2009 .

[41]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[42]  Sébastien Rebecchi,et al.  An Introduction to Deep Learning , 2011, ESANN.

[43]  Tomaso Poggio,et al.  CNS: a GPU-based framework for simulating cortically-organized networks , 2010 .

[44]  L. Steinberg Cognitive and affective development in adolescence , 2005, Trends in Cognitive Sciences.

[45]  Dong Yu,et al.  Investigation of full-sequence training of deep belief networks for speech recognition , 2010, INTERSPEECH.

[46]  Robert J. Baron,et al.  The Cerebral Computer: An Introduction To the Computational Structure of the Human Brain , 1987 .

[47]  Yoshua Bengio,et al.  Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.

[48]  T. Kohonen Self-organized formation of topographically correct feature maps , 1982 .

[49]  Paul Smolensky,et al.  Information processing in dynamical systems: foundations of harmony theory , 1986 .

[50]  John Nolte,et al.  The Human Brain An Introduction to Its Functional Anatomy , 2013 .

[51]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[52]  Yann LeCun,et al.  Convolutional networks and applications in vision , 2010, Proceedings of 2010 IEEE International Symposium on Circuits and Systems.

[53]  Jürgen Schmidhuber,et al.  Learning Complex, Extended Sequences Using the Principle of History Compression , 1992, Neural Computation.

[54]  Mikko H. Lipasti,et al.  Cortical columns: Building blocks for intelligent systems , 2009, 2009 IEEE Symposium on Computational Intelligence for Multimedia Signal and Vision Processing.

[55]  Mikko H. Lipasti,et al.  Discovering Cortical Algorithms , 2018, IJCCI.

[56]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[57]  Yoshua Bengio,et al.  Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.

[58]  Geoffrey E. Hinton,et al.  How neural networks learn from experience. , 1992, Scientific American.

[59]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.

[60]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[61]  M. Friedman The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance , 1937 .

[62]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[63]  Tara N. Sainath,et al.  Deep Belief Networks using discriminative features for phone recognition , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[64]  R. O’Reilly,et al.  Computational Explorations in Cognitive Neuroscience: Understanding the Mind by Simulating the Brain , 2000 .

[65]  Derek K. Jones,et al.  Occipito-temporal connections in the human brain. , 2003, Brain : a journal of neurology.

[66]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[67]  Geoffrey E. Hinton A Practical Guide to Training Restricted Boltzmann Machines , 2012, Neural Networks: Tricks of the Trade.

[68]  Jason Weston,et al.  A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.

[69]  J. Hawkins,et al.  On Intelligence , 2004 .

[70]  Indranil Saha,et al.  journal homepage: www.elsevier.com/locate/neucom , 2022 .

[71]  Thomas F. Nugent,et al.  Dynamic mapping of human cortical development during childhood through early adulthood. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[72]  Eugene M. Izhikevich,et al.  Which model to use for cortical spiking neurons? , 2004, IEEE Transactions on Neural Networks.

[73]  Mikko H. Lipasti,et al.  A case for neuromorphic ISAs , 2011, ASPLOS XVI.