Vector Symbolic Architectures as a Computing Framework for Nanoscale Hardware

This article reviews recent progress in the development of the computing framework Vector Symbolic Architectures (also known as Hyperdimensional Computing). This framework is well suited for implementation in stochastic, nanoscale hardware and it naturally expresses the types of cognitive operations required for Artificial Intelligence (AI). We demonstrate in this article that the ring-like algebraic structure of Vector Symbolic Architectures offers simple but powerful operations on highdimensional vectors that can support all data structures and manipulations relevant in modern computing. In addition, we illustrate the distinguishing feature of Vector Symbolic Architectures, “computing in superposition,” which sets it apart from conventional computing. This latter property opens the door to efficient solutions to the difficult combinatorial search problems inherent in AI applications. Vector Symbolic Architectures are Turing complete, as we show, and we see them acting as a framework for computing with distributed representations in myriad AI settings. This paper serves as a reference for computer architects by illustrating techniques and philosophy of VSAs for distributed computing and relevance to emerging computing hardware, such as neuromorphic computing.

[1]  Christos H. Papadimitriou,et al.  Brain computation by assemblies of neurons , 2019, Proceedings of the National Academy of Sciences.

[2]  A. I. Martyshkin,et al.  Search for a substring of characters using the theory of non-deterministic finite automata and vector-character architecture , 2020, Bulletin of Electrical Engineering and Informatics.

[3]  Dmitri A. Rachkovskij,et al.  Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning , 2001, Neural Computation.

[4]  姜乐 a:b:c≠a÷b÷c , 1994 .

[5]  Douglas Summers-Stay,et al.  A Computational Theory for Life-Long Learning of Semantics , 2018, AGI.

[6]  Ivan Tyukin,et al.  Blessing of dimensionality: mathematical foundations of the statistical physics of data , 2018, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[7]  Pentti Kanerva,et al.  Large Patterns Make Great Symbols: An Example of Learning from Example , 1998, Hybrid Neural Systems.

[8]  Bruno A. Olshausen,et al.  Resonator Networks, 1: An Efficient Solution for Factoring High-Dimensional, Distributed Representations of Data Structures , 2020, Neural Computation.

[9]  F. van der Velde,et al.  Neural blackboard architectures of combinatorial structures in cognition , 2006, Behavioral and Brain Sciences.

[10]  Alexander Legalov,et al.  Associative synthesis of finite state automata model of a controlled object with hyperdimensional computing , 2017, IECON 2017 - 43rd Annual Conference of the IEEE Industrial Electronics Society.

[11]  Okko Johannes Räsänen,et al.  Modeling Dependencies in Multiple Parallel Data Streams with Hyperdimensional Computing , 2014, IEEE Signal Processing Letters.

[12]  Luca Benini,et al.  Integrating event-based dynamic vision sensors with sparse hyperdimensional computing: a low-power accelerator with online learning capability , 2020, ISLPED.

[13]  K. Parhi,et al.  Classification Using Hyperdimensional Computing: A Review , 2020, IEEE Circuits and Systems Magazine.

[14]  Matthew Cook,et al.  Universality in Elementary Cellular Automata , 2004, Complex Syst..

[15]  Peer Neubert,et al.  An Introduction to Hyperdimensional Computing for Robotics , 2019, KI - Künstliche Intelligenz.

[16]  Denis Kleyko,et al.  Integer Echo State Networks: Efficient Reservoir Computing for Digital Hardware , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[17]  Terrence C. Stewart,et al.  A neural representation of continuous space using fractional binding , 2019, CogSci.

[18]  Ian Taylor,et al.  Constructing distributed time-critical applications using cognitive enabled services , 2019, Future Gener. Comput. Syst..

[19]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[20]  Chris Eliasmith,et al.  Biologically Plausible, Human-scale Knowledge Representation , 2016, CogSci.

[21]  Nabil Imam,et al.  Rapid online learning and robust recall in a neuromorphic olfactory circuit , 2019, Nature Machine Intelligence.

[22]  Günther Palm,et al.  Parallel processing for associative and neuronal networks , 1984, Biological Cybernetics.

[23]  Massimo Panella,et al.  Hyperdimensional Computing for Efficient Distributed Classification with Randomized Neural Networks , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).

[24]  Jan M. Rabaey,et al.  A Programmable Hyper-Dimensional Processor Architecture for Human-Centric IoT , 2019, IEEE Journal on Emerging and Selected Topics in Circuits and Systems.

[25]  HyperEmbed: Tradeoffs Between Resources and Performance in NLP Tasks with Hyperdimensional Computing Enabled Embedding of n-gram Statistics , 2019, 2021 International Joint Conference on Neural Networks (IJCNN).

[26]  R. Weale Vision. A Computational Investigation Into the Human Representation and Processing of Visual Information. David Marr , 1983 .

[27]  Geoffrey E. Hinton,et al.  Distributed Representations , 1986, The Philosophy of Artificial Intelligence.

[28]  Dmitri A. Rachkovskij,et al.  Neural Distributed Autoassociative Memories: A Survey , 2017, ArXiv.

[29]  Jan M. Rabaey,et al.  Generalized Learning Vector Quantization for Classification in Randomized Neural Networks and Hyperdimensional Computing , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).

[30]  Michael Beigl,et al.  Collective Communication for Dense Sensing Environments , 2011, 2011 Seventh International Conference on Intelligent Environments.

[31]  P. Kanerva,et al.  Hyperdimensional Computing for Text Classification , 2016 .

[32]  Okko Johannes Räsänen,et al.  Sequence Prediction With Sparse Distributed Hyperdimensional Coding Applied to the Analysis of Mobile Phone Use Patterns , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[33]  Tajana Simunic,et al.  GenieHD: Efficient DNA Pattern Matching Accelerator Using Hyperdimensional Computing , 2020, 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[34]  Ross W. Gayler Vector Symbolic Architectures answer Jackendoff's challenges for cognitive neuroscience , 2004, ArXiv.

[35]  Ross W. Gayler,et al.  Vector symbolic architectures are a viable alternative for Jackendoff's challenges , 2006, Behavioral and Brain Sciences.

[36]  Peter beim Graben,et al.  Vector Symbolic Architectures for Context-Free Grammars , 2020, Cognitive Computation.

[37]  Trevor Bekolay,et al.  A Large-Scale Model of the Functioning Brain , 2012, Science.

[38]  Eric A. Weiss,et al.  The Hyperdimensional Stack Machine , 2018 .

[39]  Luca Benini,et al.  PULP-HD: Accelerating Brain-Inspired High-Dimensional Computing on a Parallel Ultra-Low Power Platform , 2018, 2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC).

[40]  Ah Chung Tsoi,et al.  The Graph Neural Network Model , 2009, IEEE Transactions on Neural Networks.

[41]  Simon D. Levy,et al.  A distributed basis for analogical mapping , 2009 .

[42]  Luca Benini,et al.  Efficient Biosignal Processing Using Hyperdimensional Computing: Network Templates for Combined Learning and Classification of ExG Signals , 2019, Proceedings of the IEEE.

[43]  Fredrik Sandin,et al.  Analogical mapping and inference with binary spatter codes and sparse distributed memory , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[44]  Pentti Kanerva,et al.  Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors , 2009, Cognitive Computation.

[45]  Tajana Simunic,et al.  FELIX: Fast and Energy-Efficient Logic in Memory , 2018, 2018 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[46]  Friedrich T. Sommer,et al.  Variable Binding for Sparse Distributed Representations: Theory and Applications , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[47]  Luc Steels,et al.  A Vector Representation of Fluid Construction Grammar Using Holographic Reduced Representations , 2015, EAPCogSci.

[48]  Günther Palm,et al.  Information capacity in recurrent McCulloch-Pitts networks with sparsely coded memory states , 1992 .

[49]  Evgeny Osipov,et al.  Density Encoding Enables Resource-Efficient Randomly Connected Neural Networks , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[50]  Dmitri A. Rachkovskij,et al.  Building a world model with structure-sensitive sparse binary distributed representations , 2013, BICA 2013.

[51]  Tajana Simunic,et al.  F5-HD: Fast Flexible FPGA-based Framework for Refreshing Hyperdimensional Computing , 2019, FPGA.

[52]  Donald E. Knuth,et al.  Fast Pattern Matching in Strings , 1977, SIAM J. Comput..

[53]  Tony A. Plate,et al.  Holographic reduced representations , 1995, IEEE Trans. Neural Networks.

[54]  Mingguo Zhao,et al.  Towards artificial general intelligence with hybrid Tianjic chip architecture , 2019, Nature.

[55]  J. Fodor,et al.  Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.

[56]  Li Fan,et al.  Summary cache: a scalable wide-area web cache sharing protocol , 2000, TNET.

[57]  Evgeny Osipov,et al.  Recognizing Permuted Words with Vector Symbolic Architectures: A Cambridge Test for Machines , 2016, BICA.

[58]  J. Rabaey,et al.  A wearable biosensing system with in-sensor adaptive machine learning for hand gesture recognition , 2020, Nature Electronics.

[59]  Charles P. Dolan,et al.  Tensor Product Production System: a Modular Architecture and Representation , 1989 .

[60]  Garrick Orchard,et al.  Advancing Neuromorphic Computing With Loihi: A Survey of Results and Outlook , 2021, Proceedings of the IEEE.

[61]  Jan M. Rabaey,et al.  Classification and Recall With Binary Hyperdimensional Computing: Tradeoffs in Choice of Density and Mapping Characteristics , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[62]  Mohsen Imani,et al.  THRIFTY: Training with Hyperdimensional Computing across Flash Hierarchy , 2020, 2020 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[63]  Alessio Burrello,et al.  Hyperdimensional Computing With Local Binary Patterns: One-Shot Learning of Seizure Onset and Identification of Ictogenic Brain Regions Using Short-Time iEEG Recordings , 2020, IEEE Transactions on Biomedical Engineering.

[64]  P. Kanerva,et al.  Permutations as a means to encode order in word space , 2008 .

[65]  Friedrich T. Sommer,et al.  A Theory of Sequence Indexing and Working Memory in Recurrent Neural Networks , 2018, Neural Computation.

[66]  Dmitri A. Rachkovskij,et al.  SIMILARITY‐BASED RETRIEVAL WITH STRUCTURE‐SENSITIVE SPARSE BINARY DISTRIBUTED REPRESENTATIONS , 2012, Comput. Intell..

[67]  Andrew S. Cassidy,et al.  A million spiking-neuron integrated circuit with a scalable communication network and interface , 2014, Science.

[68]  Johan Kwisthout,et al.  On the computational power and complexity of Spiking Neural Networks , 2020, NICE.

[69]  Chris Eliasmith,et al.  Neural Engineering: Computation, Representation, and Dynamics in Neurobiological Systems , 2004, IEEE Transactions on Neural Networks.

[70]  Aditya Joshi,et al.  Language Geometry Using Random Indexing , 2016, QI.

[71]  Jan M. Rabaey,et al.  High-Dimensional Computing as a Nanoscalable Paradigm , 2017, IEEE Transactions on Circuits and Systems I: Regular Papers.

[72]  Evgeny Osipov,et al.  A Hyperdimensional Computing Framework for Analysis of Cardiorespiratory Synchronization During Paced Deep Breathing , 2019, IEEE Access.

[73]  Jan M. Rabaey,et al.  A Robust and Energy-Efficient Classifier Using Brain-Inspired Hyperdimensional Computing , 2016, ISLPED.

[74]  Turlough Neary,et al.  Small Weakly Universal Turing Machines , 2007, FCT.

[75]  E. M. Kussul,et al.  On image texture recognition by associative-projective neurocomputer , 1991 .

[76]  Daswin De Silva,et al.  Integer Self-Organizing Maps for Digital Hardware , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[77]  Torsten Hoefler,et al.  Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. , 2018 .

[78]  Artem Sokolov,et al.  Vector representations for efficient comparison and search for similar strings , 2007 .

[79]  Trevor Cohen,et al.  Reasoning with vectors: A continuous model for fast robust inference , 2015, Log. J. IGPL.

[80]  S. Thorpe Localized versus distributed representations , 1998 .

[81]  Sasu Tarkoma,et al.  Theory and Practice of Bloom Filters for Distributed Systems , 2012, IEEE Communications Surveys & Tutorials.

[82]  Nikita Lyamin,et al.  Dependable MAC Layer Architecture Based on Holographic Data Representation Using Hyper-Dimensional Binary Spatter Codes , 2012, MACOM.

[83]  Denis Kleyko,et al.  End to End Binarized Neural Networks for Text Classification , 2020, SUSTAINLP.

[84]  Anders Holst,et al.  Random indexing of text samples for latent semantic analysis , 2000 .

[85]  Trevor Bekolay,et al.  Nengo: a Python tool for building large-scale functional brain models , 2014, Front. Neuroinform..

[86]  Geoffrey E. Hinton Tensor Product Variable Binding and the Representation of Symbolic Structures in Connectionist Systems , 1991 .

[87]  S. Furber,et al.  To build a brain , 2012, IEEE Spectrum.

[88]  David Reitter,et al.  Holographic Declarative Memory: Distributional Semantics as the Architecture of Memory , 2019, Cogn. Sci..

[89]  Ross W. Gayler,et al.  Multiplicative Binding, Representation Operators & Analogy , 1998 .

[90]  Evgeny Osipov,et al.  On Bidirectional Transitions between Localist and Distributed Representations: The Case of Common Substrings Search Using Vector Symbolic Architecture , 2014, BICA.

[91]  Luca Benini,et al.  In-memory hyperdimensional computing , 2019, Nature Electronics.

[92]  Pentti Kanerva,et al.  Fully Distributed Representation , 1997 .

[93]  József Fiser,et al.  Neural Variability and Sampling-Based Probabilistic Representations in the Visual Cortex , 2016, Neuron.

[94]  Jane Neumann,et al.  Learning the systematic transformation of holographic reduced representations , 2002, Cognitive Systems Research.

[95]  Bart De Moor,et al.  Geometric Analogue of Holographic Reduced Representation , 2007, ArXiv.

[96]  Jan M. Rabaey,et al.  Hyperdimensional computing with 3D VRRAM in-memory kernels: Device-architecture co-design for energy-efficient, error-resilient language recognition , 2016, 2016 IEEE International Electron Devices Meeting (IEDM).

[97]  Denis Kleyko,et al.  Commentaries on "Learning Sensorimotor Control with Neuromorphic Sensors: Toward Hyperdimensional Active Perception" [Science Robotics Vol 4 Issue 30 (2019) 1-10 , 2020, ArXiv.

[98]  Geoffrey E. Hinton,et al.  Distributed representations and nested compositional structure , 1994 .

[99]  Jan M. Rabaey,et al.  Exploring Hyperdimensional Associative Memory , 2017, 2017 IEEE International Symposium on High Performance Computer Architecture (HPCA).

[100]  Tony Plate,et al.  Estimating Analogical Similarity by Dot-Products of Holographic Reduced Representations , 1993, NIPS.

[101]  Luca Benini,et al.  Robust high-dimensional memory-augmented neural networks , 2020, Nature Communications.

[102]  Hun-Seok Kim,et al.  HDM: Hyper-Dimensional Modulation for Robust Low-Power Communications , 2018, 2018 IEEE International Conference on Communications (ICC).

[103]  Wolfgang Maass,et al.  Lower Bounds for the Computational Power of Networks of Spiking Neurons , 1996, Neural Computation.

[104]  Bruno A. Olshausen,et al.  Resonator Networks, 2: Factorization Performance and Capacity Compared to Optimization-Based Methods , 2020, Neural Computation.

[105]  Nikolaos Papakonstantinou,et al.  Hyperdimensional Computing in Industrial Systems: The Use-Case of Distributed Fault Isolation in a Power Plant , 2018, IEEE Access.

[106]  Stephen Wolfram,et al.  A New Kind of Science , 2003, Artificial Life.

[107]  Wolfgang Porod,et al.  Coupled oscillators for computing: A review and perspective , 2020 .

[108]  Pablo Barceló,et al.  On the Turing Completeness of Modern Neural Network Architectures , 2019, ICLR.

[109]  Daswin De Silva,et al.  Trajectory clustering of road traffic in urban environments using incremental machine learning in combination with hyperdimensional computing , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[110]  Trevor Cohen,et al.  Orthogonality and Orthography: Introducing Measured Distance into Semantic Space , 2013, QI.

[111]  Jan M. Rabaey,et al.  Hyperdimensional Computing Exploiting Carbon Nanotube FETs, Resistive RAM, and Their Monolithic 3D Integration , 2018, IEEE Journal of Solid-State Circuits.

[112]  Richard M. Karp,et al.  Efficient Randomized Pattern-Matching Algorithms , 1987, IBM J. Res. Dev..

[113]  Chris Eliasmith,et al.  A Spiking Neuron Model of Serial-Order Recall , 2010 .

[114]  Jan M. Rabaey,et al.  Brain-inspired computing exploiting carbon nanotube FETs and resistive RAM: Hyperdimensional computing case study , 2018, 2018 IEEE International Solid - State Circuits Conference - (ISSCC).

[115]  Terrence C. Stewart,et al.  of the Annual Meeting of the Cognitive Science Society Title Symbolic Reasoning in Spiking Neurons : A Model of the Cortex / Basal Ganglia / Thalamus Loop , 2010 .

[116]  Suraj Bajracharya,et al.  Learning Behavior Hierarchies via High-Dimensional Sensor Projection , 2013, AAAI Workshop: Learning Rich Representations from Low-Level Sensors.

[117]  Hong Wang,et al.  Loihi: A Neuromorphic Manycore Processor with On-Chip Learning , 2018, IEEE Micro.

[118]  Graham Cormode,et al.  An improved data stream summary: the count-min sketch and its applications , 2004, J. Algorithms.

[119]  Yukiko Kikuchi,et al.  Structured sequence processing and combinatorial binding: neurobiologically and computationally informed hypotheses , 2019, Philosophical Transactions of the Royal Society B.

[120]  Friedrich T. Sommer,et al.  Robust computation with rhythmic spike patterns , 2019, Proceedings of the National Academy of Sciences.

[121]  Andrew McCallum,et al.  Energy and Policy Considerations for Deep Learning in NLP , 2019, ACL.

[122]  Tony A. Plate Networks which learn to store variable-length sequences in a fixed set of unit activations , 2007 .

[123]  Michael N. Jones,et al.  Encoding Sequential Information in Semantic Space Models: Comparing Holographic Reduced Representation and Random Permutation , 2015, Comput. Intell. Neurosci..

[124]  Günther Palm,et al.  Memory Capacities for Synaptic and Structural Plasticity G ¨ Unther Palm , 2022 .

[125]  Emmanuel Dupoux,et al.  Holographic String Encoding , 2011, Cogn. Sci..

[126]  Yiannis Aloimonos,et al.  Learning sensorimotor control with neuromorphic sensors: Toward hyperdimensional active perception , 2019, Science Robotics.

[127]  Andrew S. Cassidy,et al.  Merolla communication network and interface A million spiking-neuron integrated circuit with a scalable , 2014 .

[128]  Luca Benini,et al.  Hyperdimensional biosignal processing: A case study for EMG-based hand gesture recognition , 2016, 2016 IEEE International Conference on Rebooting Computing (ICRC).

[129]  Stephen I. Gallant,et al.  Representing Objects, Relations, and Sequences , 2013, Neural Computation.

[130]  Pentti Kanerva Computing with High-Dimensional Vectors , 2019, IEEE Design & Test.

[131]  Dana S. Scott,et al.  Finite Automata and Their Decision Problems , 1959, IBM J. Res. Dev..

[132]  Terrence C. Stewart,et al.  Sentence processing in spiking neurons: A biologically plausible left-corner parser , 2014, CogSci.

[133]  Peter Dayan,et al.  Bayesian retrieval in associative memories with storage errors , 1998, IEEE Trans. Neural Networks.

[134]  Denis Kleyko,et al.  Autoscaling Bloom filter: controlling trade-off between true and false positives , 2017, Neural Computing and Applications.

[135]  Jussi H. Poikonen,et al.  High-dimensional computing with sparse vectors , 2015, 2015 IEEE Biomedical Circuits and Systems Conference (BioCAS).

[136]  Burton H. Bloom,et al.  Space/time trade-offs in hash coding with allowable errors , 1970, CACM.

[137]  Anders S. G. Andrae,et al.  On Global Electricity Usage of Communication Technology: Trends to 2030 , 2015 .

[138]  Valeriy Vyatkin,et al.  Distributed Representation of n-gram Statistics for Boosting Self-organizing Maps with Hyperdimensional Computing , 2019, Ershov Informatics Conference.

[139]  David Van Brackle,et al.  Vector Representation for Sub-Graph Encoding to Resolve Entities , 2016 .

[140]  H. C. LONGUET-HIGGINS,et al.  Non-Holographic Associative Memory , 1969, Nature.