A Bio-Inspired Incremental Learning Architecture for Applied Perceptual Problems

We present a biologically inspired architecture for incremental learning that remains resource-efficient even in the face of very high data dimensionalities (>1000) that are typically associated with perceptual problems. In particular, we investigate how a new perceptual (object) class can be added to a trained architecture without retraining, while avoiding the well-known catastrophic forgetting effects typically associated with such scenarios. At the heart of the presented architecture lies a generative description of the perceptual space by a self-organized approach which at the same time approximates the neighborhood relations in this space on a two-dimensional plane. This approximation, which closely imitates the topographic organization of the visual cortex, allows an efficient local update rule for incremental learning even in the face of very high dimensionalities, which we demonstrate by tests on the well-known MNIST benchmark. We complement the model by adding a biologically plausible short-term memory system, allowing it to retain excellent classification accuracy even under incremental learning in progress. The short-term memory is additionally used to reinforce new data statistics by replaying previously stored samples during dedicated “sleep” phases.

[1]  Pierre-Yves Oudeyer,et al.  Incremental local online Gaussian Mixture Regression for imitation learning of multiple tasks , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[2]  Jan Peters,et al.  Local Gaussian process regression for real-time model-based robot control , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[3]  Norman M Weinberger,et al.  The nucleus basalis and memory codes: Auditory cortical plasticity and the induction of specific, associative behavioral memory , 2003, Neurobiology of Learning and Memory.

[4]  Stefan Schaal,et al.  Locally Weighted Projection Regression : An O(n) Algorithm for Incremental Real Time Learning in High Dimensional Space , 2000 .

[5]  Michael M Merzenich,et al.  Perceptual Learning Directs Auditory Cortical Map Reorganization through Top-Down Influences , 2006, The Journal of Neuroscience.

[6]  R. O’Reilly The Division of Labor Between the Neocortex and Hippocampus , 2010 .

[7]  Yoshua Bengio,et al.  An Empirical Investigation of Catastrophic Forgeting in Gradient-Based Neural Networks , 2013, ICLR.

[8]  Roshani Ade,et al.  Incremental Learning From Unbalanced Data with Concept Class, Concept Drift and Missing Features : A Review , 2014 .

[9]  Keiji Tanaka,et al.  Inferotemporal cortex and object vision. , 1996, Annual review of neuroscience.

[10]  T. Kohonen Self-Organized Formation of Correct Feature Maps , 1982 .

[11]  T. Palmeri,et al.  Not just the norm: Exemplar-based models also predict face aftereffects , 2014, Psychonomic Bulletin & Review.

[12]  Huan Liu,et al.  Handling concept drifts in incremental learning with support vector machines , 1999, KDD '99.

[13]  Antoine Bordes,et al.  The Huller: A Simple and Efficient Online SVM , 2005, ECML.

[14]  Olivier Sigaud,et al.  On-line regression algorithms for learning mechanical models of robots: A survey , 2011, Robotics Auton. Syst..

[15]  Teuvo Kohonen,et al.  Self-organized formation of topologically correct feature maps , 2004, Biological Cybernetics.

[16]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[17]  Martin V. Butz,et al.  Computational Complexity of the XCS Classifier System , 2005 .

[18]  Robert C. Wolpert,et al.  A Review of the , 1985 .

[19]  James L. McClelland,et al.  Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.

[20]  Stefan Rüping,et al.  Incremental Learning with Support Vector Machines , 2001, ICDM.

[21]  T. Kohonen Self-organized formation of topographically correct feature maps , 1982 .

[22]  Alexander Gepperth,et al.  Biologically inspired incremental learning for high-dimensional spaces , 2015, 2015 Joint IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob).

[23]  B. McNaughton,et al.  Modeling the spontaneous reactivation of experience‐specific hippocampal cell assembles during sleep , 1996, Hippocampus.

[24]  R. Desimone,et al.  Clustering of perirhinal neurons with similar properties following visual experience in adult monkeys , 2000, Nature Neuroscience.

[25]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[26]  M. Hasselmo,et al.  The effect of learning on the face selective responses of neurons in the cortex in the superior temporal sulcus of the monkey , 2004, Experimental Brain Research.

[27]  Stefan Schaal,et al.  A Library for Locally Weighted Projection Regression , 2008, J. Mach. Learn. Res..

[28]  M. Hasselmo The role of acetylcholine in learning and memory , 2006, Current Opinion in Neurobiology.

[29]  M. Giese,et al.  Norm-based face encoding by single neurons in the monkey inferotemporal cortex , 2006, Nature.