1
|
Kleyko D, Kymn CJ, Thomas A, Olshausen BA, Sommer FT, Frady EP. Principled neuromorphic reservoir computing. Nat Commun 2025; 16:640. [PMID: 39809739 PMCID: PMC11733134 DOI: 10.1038/s41467-025-55832-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Accepted: 12/20/2024] [Indexed: 01/16/2025] Open
Abstract
Reservoir computing advances the intriguing idea that a nonlinear recurrent neural circuit-the reservoir-can encode spatio-temporal input signals to enable efficient ways to perform tasks like classification or regression. However, recently the idea of a monolithic reservoir network that simultaneously buffers input signals and expands them into nonlinear features has been challenged. A representation scheme in which memory buffer and expansion into higher-order polynomial features can be configured separately has been shown to significantly outperform traditional reservoir computing in prediction of multivariate time-series. Here we propose a configurable neuromorphic representation scheme that provides competitive performance on prediction, but with significantly better scaling properties than directly materializing higher-order features as in prior work. Our approach combines the use of randomized representations from traditional reservoir computing with mathematical principles for approximating polynomial kernels via such representations. While the memory buffer can be realized with standard reservoir networks, computing higher-order features requires networks of 'Sigma-Pi' neurons, i.e., neurons that enable both summation as well as multiplication of inputs. Finally, we provide an implementation of the memory buffer and Sigma-Pi networks on Loihi 2, an existing neuromorphic hardware platform.
Collapse
Affiliation(s)
- Denis Kleyko
- Centre for Applied Autonomous Sensor Systems, Örebro University, Örebro, Sweden.
- Intelligent Systems Lab, RISE Research Institutes of Sweden, Kista, Sweden.
| | - Christopher J Kymn
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, CA, USA
| | - Anthony Thomas
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, CA, USA
- Electrical and Computer Engineering, University of California, Davis, CA, USA
| | - Bruno A Olshausen
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, CA, USA
| | - Friedrich T Sommer
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, CA, USA.
- Neuromorphic Computing Lab, Intel, Santa Clara, CA, USA.
| | - E Paxon Frady
- Neuromorphic Computing Lab, Intel, Santa Clara, CA, USA
| |
Collapse
|
2
|
Osipov E, Kahawala S, Haputhanthri D, Kempitiya T, De Silva D, Alahakoon D, Kleyko D. Hyperseed: Unsupervised Learning With Vector Symbolic Architectures. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:6583-6597. [PMID: 36383581 DOI: 10.1109/tnnls.2022.3211274] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Motivated by recent innovations in biologically inspired neuromorphic hardware, this article presents a novel unsupervised machine learning algorithm named Hyperseed that draws on the principles of vector symbolic architectures (VSAs) for fast learning of a topology preserving feature map of unlabeled data. It relies on two major operations of VSA, binding and bundling. The algorithmic part of Hyperseed is expressed within the Fourier holographic reduced representations (FHRR) model, which is specifically suited for implementation on spiking neuromorphic hardware. The two primary contributions of the Hyperseed algorithm are few-shot learning and a learning rule based on single vector operation. These properties are empirically evaluated on synthetic datasets and on illustrative benchmark use cases, IRIS classification, and a language identification task using the n -gram statistics. The results of these experiments confirm the capabilities of Hyperseed and its applications in neuromorphic hardware.
Collapse
|
3
|
Hernández-Cano A, Ni Y, Zou Z, Zakeri A, Imani M. Hyperdimensional computing with holographic and adaptive encoder. Front Artif Intell 2024; 7:1371988. [PMID: 38655269 PMCID: PMC11037243 DOI: 10.3389/frai.2024.1371988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Accepted: 03/18/2024] [Indexed: 04/26/2024] Open
Abstract
Introduction Brain-inspired computing has become an emerging field, where a growing number of works focus on developing algorithms that bring machine learning closer to human brains at the functional level. As one of the promising directions, Hyperdimensional Computing (HDC) is centered around the idea of having holographic and high-dimensional representation as the neural activities in our brains. Such representation is the fundamental enabler for the efficiency and robustness of HDC. However, existing HDC-based algorithms suffer from limitations within the encoder. To some extent, they all rely on manually selected encoders, meaning that the resulting representation is never adapted to the tasks at hand. Methods In this paper, we propose FLASH, a novel hyperdimensional learning method that incorporates an adaptive and learnable encoder design, aiming at better overall learning performance while maintaining good properties of HDC representation. Current HDC encoders leverage Random Fourier Features (RFF) for kernel correspondence and enable locality-preserving encoding. We propose to learn the encoder matrix distribution via gradient descent and effectively adapt the kernel for a more suitable HDC encoding. Results Our experiments on various regression datasets show that tuning the HDC encoder can significantly boost the accuracy, surpassing the current HDC-based algorithm and providing faster inference than other baselines, including RFF-based kernel ridge regression. Discussion The results indicate the importance of an adaptive encoder and customized high-dimensional representation in HDC.
Collapse
Affiliation(s)
- Alejandro Hernández-Cano
- Department of Computer Science, École polytechnique fédérale de Lausanne (EPFL), Lausanne, Switzerland
| | - Yang Ni
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Zhuowen Zou
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Ali Zakeri
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Mohsen Imani
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| |
Collapse
|
4
|
Li X, Wang S. Simple and complex cells revisited: toward a selectivity-invariance model of object recognition. Front Comput Neurosci 2023; 17:1282828. [PMID: 37905187 PMCID: PMC10613527 DOI: 10.3389/fncom.2023.1282828] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 09/19/2023] [Indexed: 11/02/2023] Open
Abstract
This paper presents a theoretical perspective on modeling ventral stream processing by revisiting the computational abstraction of simple and complex cells. In parallel to David Marr's vision theory, we organize the new perspective into three levels. At the computational level, we abstract simple and complex cells into space partitioning and composition in a topological space based on the redundancy exploitation hypothesis of Horace Barlow. At the algorithmic level, we present a hierarchical extension of sparse coding by exploiting the manifold constraint in high-dimensional space (i.e., the blessing of dimensionality). The resulting over-parameterized models for object recognition differ from existing hierarchical models by disentangling the objectives of selectivity and invariance computation. It is possible to interpret our hierarchical construction as a computational implementation of cortically local subspace untangling for object recognition and face representation, which are closely related to exemplar-based and axis-based coding in the medial temporal lobe. At the implementation level, we briefly discuss two possible implementations based on asymmetric sparse autoencoders and divergent spiking neural networks.
Collapse
Affiliation(s)
- Xin Li
- Department of Computer Science, University at Albany, Albany, NY, United States
| | - Shuo Wang
- Department of Radiology, Washington University at St. Louis, St. Louis, MO, United States
| |
Collapse
|
5
|
Kleyko D, Bybee C, Huang PC, Kymn CJ, Olshausen BA, Frady EP, Sommer FT. Efficient Decoding of Compositional Structure in Holistic Representations. Neural Comput 2023; 35:1159-1186. [PMID: 37187162 DOI: 10.1162/neco_a_01590] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Accepted: 02/27/2023] [Indexed: 05/17/2023]
Abstract
We investigate the task of retrieving information from compositional distributed representations formed by hyperdimensional computing/vector symbolic architectures and present novel techniques that achieve new information rate bounds. First, we provide an overview of the decoding techniques that can be used to approach the retrieval task. The techniques are categorized into four groups. We then evaluate the considered techniques in several settings that involve, for example, inclusion of external noise and storage elements with reduced precision. In particular, we find that the decoding techniques from the sparse coding and compressed sensing literature (rarely used for hyperdimensional computing/vector symbolic architectures) are also well suited for decoding information from the compositional distributed representations. Combining these decoding techniques with interference cancellation ideas from communications improves previously reported bounds (Hersche et al., 2021) of the information rate of the distributed representations from 1.20 to 1.40 bits per dimension for smaller codebooks and from 0.60 to 1.26 bits per dimension for larger codebooks.
Collapse
Affiliation(s)
- Denis Kleyko
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A
- Intelligent Systems Laboratory, Research Institutes of Sweden, 16440 Kista, Sweden
| | - Connor Bybee
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A.
| | - Ping-Chen Huang
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A.
| | - Christopher J Kymn
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A.
| | - Bruno A Olshausen
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A.
| | - E Paxon Frady
- Neuromorphic Computing Laboratory, Intel Labs, Santa Clara, CA 95054, U.S.A.
| | - Friedrich T Sommer
- Redwood Center for Theoretical Neuroscience, University of California at Berkeley, Berkeley, CA 94720, U.S.A
- Neuromorphic Computing Laboratory, Intel Labs, Santa Clara, CA 95054, U.S.A.
| |
Collapse
|
6
|
Hiratani N, Sompolinsky H. Optimal Quadratic Binding for Relational Reasoning in Vector Symbolic Neural Architectures. Neural Comput 2023; 35:105-155. [PMID: 36543330 DOI: 10.1162/neco_a_01558] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2022] [Accepted: 09/13/2022] [Indexed: 12/24/2022]
Abstract
Binding operation is fundamental to many cognitive processes, such as cognitive map formation, relational reasoning, and language comprehension. In these processes, two different modalities, such as location and objects, events and their contextual cues, and words and their roles, need to be bound together, but little is known about the underlying neural mechanisms. Previous work has introduced a binding model based on quadratic functions of bound pairs, followed by vector summation of multiple pairs. Based on this framework, we address the following questions: Which classes of quadratic matrices are optimal for decoding relational structures? And what is the resultant accuracy? We introduce a new class of binding matrices based on a matrix representation of octonion algebra, an eight-dimensional extension of complex numbers. We show that these matrices enable a more accurate unbinding than previously known methods when a small number of pairs are present. Moreover, numerical optimization of a binding operator converges to this octonion binding. We also show that when there are a large number of bound pairs, however, a random quadratic binding performs, as well as the octonion and previously proposed binding methods. This study thus provides new insight into potential neural mechanisms of binding operations in the brain.
Collapse
Affiliation(s)
- Naoki Hiratani
- Center for Brain Science, Harvard University, Cambridge MA 02138, U.S.A.
| | - Haim Sompolinsky
- Center for Brain Science, Harvard University, Cambridge MA 02138, U.S.A
- Edmond and Lily Safra Center for Brain Sciences, Hebrew University, Jerusalem 91904, Israel
| |
Collapse
|
7
|
Teeters JL, Kleyko D, Kanerva P, Olshausen BA. On separating long- and short-term memories in hyperdimensional computing. Front Neurosci 2023; 16:867568. [PMID: 36699525 PMCID: PMC9869149 DOI: 10.3389/fnins.2022.867568] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2022] [Accepted: 10/05/2022] [Indexed: 01/12/2023] Open
Abstract
Operations on high-dimensional, fixed-width vectors can be used to distribute information from several vectors over a single vector of the same width. For example, a set of key-value pairs can be encoded into a single vector with multiplication and addition of the corresponding key and value vectors: the keys are bound to their values with component-wise multiplication, and the key-value pairs are combined into a single superposition vector with component-wise addition. The superposition vector is, thus, a memory which can then be queried for the value of any of the keys, but the result of the query is approximate. The exact vector is retrieved from a codebook (a.k.a. item memory), which contains vectors defined in the system. To perform these operations, the item memory vectors and the superposition vector must be the same width. Increasing the capacity of the memory requires increasing the width of the superposition and item memory vectors. In this article, we demonstrate that in a regime where many (e.g., 1,000 or more) key-value pairs are stored, an associative memory which maps key vectors to value vectors requires less memory and less computing to obtain the same reliability of storage as a superposition vector. These advantages are obtained because the number of storage locations in an associate memory can be increased without increasing the width of the vectors in the item memory. An associative memory would not replace a superposition vector as a medium of storage, but could augment it, because data recalled from an associative memory could be used in algorithms that use a superposition vector. This would be analogous to how human working memory (which stores about seven items) uses information recalled from long-term memory (which is much larger than the working memory). We demonstrate the advantages of an associative memory experimentally using the storage of large finite-state automata, which could model the storage and recall of state-dependent behavior by brains.
Collapse
Affiliation(s)
- Jeffrey L. Teeters
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, Berkeley, CA, United States
| | - Denis Kleyko
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, Berkeley, CA, United States
- Intelligent Systems Lab, Research Institutes of Sweden, Kista, Sweden
| | - Pentti Kanerva
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, Berkeley, CA, United States
| | - Bruno A. Olshausen
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, Berkeley, CA, United States
| |
Collapse
|
8
|
Steinberg J, Sompolinsky H. Associative memory of structured knowledge. Sci Rep 2022; 12:21808. [PMID: 36528630 PMCID: PMC9759586 DOI: 10.1038/s41598-022-25708-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2022] [Accepted: 12/02/2022] [Indexed: 12/23/2022] Open
Abstract
A long standing challenge in biological and artificial intelligence is to understand how new knowledge can be constructed from known building blocks in a way that is amenable for computation by neuronal circuits. Here we focus on the task of storage and recall of structured knowledge in long-term memory. Specifically, we ask how recurrent neuronal networks can store and retrieve multiple knowledge structures. We model each structure as a set of binary relations between events and attributes (attributes may represent e.g., temporal order, spatial location, role in semantic structure), and map each structure to a distributed neuronal activity pattern using a vector symbolic architecture scheme.We then use associative memory plasticity rules to store the binarized patterns as fixed points in a recurrent network. By a combination of signal-to-noise analysis and numerical simulations, we demonstrate that our model allows for efficient storage of these knowledge structures, such that the memorized structures as well as their individual building blocks (e.g., events and attributes) can be subsequently retrieved from partial retrieving cues. We show that long-term memory of structured knowledge relies on a new principle of computation beyond the memory basins. Finally, we show that our model can be extended to store sequences of memories as single attractors.
Collapse
Affiliation(s)
- Julia Steinberg
- Department of Physics, Harvard University, Cambridge, MA, 02138, USA.
- Center for Brain Science, Harvard University, Cambridge, MA, 02138, USA.
- Joseph Henry Laboratories of Physics, Princeton University, Princeton, NJ, 08544, USA.
| | - Haim Sompolinsky
- Center for Brain Science, Harvard University, Cambridge, MA, 02138, USA.
- Edmond and Lily Safra Center for Brain Sciences, Hebrew University, 91904, Jerusalem, Israel.
| |
Collapse
|
9
|
Rachkovskij DA. Representation of spatial objects by shift-equivariant similarity-preserving hypervectors. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07619-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
10
|
Kleyko D, Frady EP, Sommer FT. Cellular Automata Can Reduce Memory Requirements of Collective-State Computing. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:2701-2713. [PMID: 34699370 PMCID: PMC9215349 DOI: 10.1109/tnnls.2021.3119543] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Various nonclassical approaches of distributed information processing, such as neural networks, reservoir computing (RC), vector symbolic architectures (VSAs), and others, employ the principle of collective-state computing. In this type of computing, the variables relevant in computation are superimposed into a single high-dimensional state vector, the collective state. The variable encoding uses a fixed set of random patterns, which has to be stored and kept available during the computation. In this article, we show that an elementary cellular automaton with rule 90 (CA90) enables the space-time tradeoff for collective-state computing models that use random dense binary representations, i.e., memory requirements can be traded off with computation running CA90. We investigate the randomization behavior of CA90, in particular, the relation between the length of the randomization period and the size of the grid, and how CA90 preserves similarity in the presence of the initialization noise. Based on these analyses, we discuss how to optimize a collective-state computing model, in which CA90 expands representations on the fly from short seed patterns-rather than storing the full set of random patterns. The CA90 expansion is applied and tested in concrete scenarios using RC and VSAs. Our experimental results show that collective-state computing with CA90 expansion performs similarly compared to traditional collective-state models, in which random patterns are generated initially by a pseudorandom number generator and then stored in a large memory.
Collapse
|
11
|
Abstract
AbstractVector Symbolic Architectures combine a high-dimensional vector space with a set of carefully designed operators in order to perform symbolic computations with large numerical vectors. Major goals are the exploitation of their representational power and ability to deal with fuzziness and ambiguity. Over the past years, several VSA implementations have been proposed. The available implementations differ in the underlying vector space and the particular implementations of the VSA operators. This paper provides an overview of eleven available VSA implementations and discusses their commonalities and differences in the underlying vector space and operators. We create a taxonomy of available binding operations and show an important ramification for non self-inverse binding operations using an example from analogical reasoning. A main contribution is the experimental comparison of the available implementations in order to evaluate (1) the capacity of bundles, (2) the approximation quality of non-exact unbinding operations, (3) the influence of combining binding and bundling operations on the query answering performance, and (4) the performance on two example applications: visual place- and language-recognition. We expect this comparison and systematization to be relevant for development of VSAs, and to support the selection of an appropriate VSA for a particular task. The implementations are available.
Collapse
|