1
|
Osipov E, Kahawala S, Haputhanthri D, Kempitiya T, De Silva D, Alahakoon D, Kleyko D. Hyperseed: Unsupervised Learning With Vector Symbolic Architectures. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:6583-6597. [PMID: 36383581 DOI: 10.1109/tnnls.2022.3211274] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Motivated by recent innovations in biologically inspired neuromorphic hardware, this article presents a novel unsupervised machine learning algorithm named Hyperseed that draws on the principles of vector symbolic architectures (VSAs) for fast learning of a topology preserving feature map of unlabeled data. It relies on two major operations of VSA, binding and bundling. The algorithmic part of Hyperseed is expressed within the Fourier holographic reduced representations (FHRR) model, which is specifically suited for implementation on spiking neuromorphic hardware. The two primary contributions of the Hyperseed algorithm are few-shot learning and a learning rule based on single vector operation. These properties are empirically evaluated on synthetic datasets and on illustrative benchmark use cases, IRIS classification, and a language identification task using the n -gram statistics. The results of these experiments confirm the capabilities of Hyperseed and its applications in neuromorphic hardware.
Collapse
|
2
|
Taler V, Johns B. Using big data to understand bilingual performance in semantic fluency: Findings from the Canadian Longitudinal Study on Aging. PLoS One 2022; 17:e0277660. [PMID: 36441767 PMCID: PMC9704680 DOI: 10.1371/journal.pone.0277660] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Accepted: 11/01/2022] [Indexed: 11/29/2022] Open
Abstract
OBJECTIVES This study aimed to characterize verbal fluency performance in monolinguals and bilinguals using data from the Canadian Longitudinal Study on Aging (CLSA). METHODS A large sample of adults aged 45-85 (n = 12,875) completed a one-minute animal fluency task in English. Participants were English-speaking monolinguals (n = 9,759), bilinguals who spoke English as their first language (L1 bilinguals, n = 1,836), and bilinguals who spoke English as their second language (L2 bilinguals, n = 1,280). Using a distributional modeling approach to quantify the semantic similarity of words, we examined the impact of word frequency and pairwise semantic similarity on performance on this task. RESULTS Overall, L1 bilinguals outperformed monolinguals on the verbal fluency task: they produced more items, and these items were of lower average frequency and semantic similarity. Monolinguals in turn outperformed L2 bilinguals on these measures. The results held across different age groups, educational, and income levels. DISCUSSION These results demonstrate an advantage for bilinguals compared to monolinguals on a category fluency task, when performed in the first language, indicating that, at least in the CLSA sample, bilinguals have superior semantic search capabilities in their first language compared to monolingual speakers of that language.
Collapse
Affiliation(s)
- Vanessa Taler
- School of Psychology, University of Ottawa, Ottawa, Canada,Bruyère Research Institute, Ottawa, Canada,* E-mail:
| | - Brendan Johns
- Department of Psychology, McGill University, Montreal, Canada
| |
Collapse
|
3
|
Kleyko D, Davies M, Frady EP, Kanerva P, Kent SJ, Olshausen BA, Osipov E, Rabaey JM, Rachkovskij DA, Rahimi A, Sommer FT. Vector Symbolic Architectures as a Computing Framework for Emerging Hardware. PROCEEDINGS OF THE IEEE. INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS 2022; 110:1538-1571. [PMID: 37868615 PMCID: PMC10588678] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 10/24/2023]
Abstract
This article reviews recent progress in the development of the computing framework Vector Symbolic Architectures (also known as Hyperdimensional Computing). This framework is well suited for implementation in stochastic, emerging hardware and it naturally expresses the types of cognitive operations required for Artificial Intelligence (AI). We demonstrate in this article that the field-like algebraic structure of Vector Symbolic Architectures offers simple but powerful operations on high-dimensional vectors that can support all data structures and manipulations relevant to modern computing. In addition, we illustrate the distinguishing feature of Vector Symbolic Architectures, "computing in superposition," which sets it apart from conventional computing. It also opens the door to efficient solutions to the difficult combinatorial search problems inherent in AI applications. We sketch ways of demonstrating that Vector Symbolic Architectures are computationally universal. We see them acting as a framework for computing with distributed representations that can play a role of an abstraction layer for emerging computing hardware. This article serves as a reference for computer architects by illustrating the philosophy behind Vector Symbolic Architectures, techniques of distributed computing with them, and their relevance to emerging computing hardware, such as neuromorphic computing.
Collapse
Affiliation(s)
- Denis Kleyko
- Redwood Center for Theoretical Neuroscience at the University of California at Berkeley, CA 94720, USA and also with the Intelligent Systems Lab at Research Institutes of Sweden, 16440 Kista, Sweden
| | - Mike Davies
- Neuromorphic Computing Lab, Intel Labs, Santa Clara, CA 95054, USA
| | - E Paxon Frady
- Neuromorphic Computing Lab, Intel Labs, Santa Clara, CA 95054, USA
| | - Pentti Kanerva
- Redwood Center for Theoretical Neuroscience at the University of California at Berkeley, CA 94720, USA
| | - Spencer J Kent
- Redwood Center for Theoretical Neuroscience at the University of California at Berkeley, CA 94720, USA
| | - Bruno A Olshausen
- Redwood Center for Theoretical Neuroscience at the University of California at Berkeley, CA 94720, USA
| | - Evgeny Osipov
- Department of Computer Science Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden
| | - Jan M Rabaey
- Department of Electrical Engineering and Computer Sciences at the University of California at Berkeley, CA 94720, USA
| | - Dmitri A Rachkovskij
- International Research and Training Center for Information Technologies and Systems, 03680 Kyiv, Ukraine, and with the Department of Computer Science Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden
| | - Abbas Rahimi
- IBM Research - Zurich, 8803 Rüschlikon, Switzerland
| | - Friedrich T Sommer
- Neuromorphic Computing Lab, Intel Labs, Santa Clara, CA 95054, USA and also with the Redwood Center for Theoretical Neuroscience at the University of California at Berkeley, CA 94720, USA
| |
Collapse
|
4
|
Abstract
AbstractVector symbolic architectures (VSA) are a viable approach for the hyperdimensional representation of symbolic data, such as documents, syntactic structures, or semantic frames. We present a rigorous mathematical framework for the representation of phrase structure trees and parse trees of context-free grammars (CFG) in Fock space, i.e. infinite-dimensional Hilbert space as being used in quantum field theory. We define a novel normal form for CFG by means of term algebras. Using a recently developed software toolbox, called FockBox, we construct Fock space representations for the trees built up by a CFG left-corner (LC) parser. We prove a universal representation theorem for CFG term algebras in Fock space and illustrate our findings through a low-dimensional principal component projection of the LC parser state. Our approach could leverage the development of VSA for explainable artificial intelligence (XAI) by means of hyperdimensional deep neural computation.
Collapse
|
5
|
Watkinson N, Givargis T, Joe V, Nicolau A, Veidenbaum A. Detecting COVID-19 Related Pneumonia On CT Scans Using Hyperdimensional Computing. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:3970-3973. [PMID: 34892100 DOI: 10.1109/embc46164.2021.9630898] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Pneumonia is a common complication associated with COVID-19 infections. Unlike common versions of pneumonia that spread quickly through large lung regions, COVID-19 related pneumonia starts in small localized pockets before spreading over the course of several days. This makes the infection more resilient and with a high probability of developing acute respiratory distress syndrome. Because of the peculiar spread pattern, the use of pulmonary computerized tomography (CT) scans was key in identifying COVID-19 infections. Identifying uncommon pulmonary diseases could be a strong line of defense in early detection of new respiratory infection-causing viruses. In this paper we describe a classification algorithm based on hyperdimensional computing for the detection of COVID-19 pneumonia in CT scans. We test our algorithm using three different datasets. The highest reported accuracy is 95.2% with an F1 score of 0.90, and all three models had a precision of 1 (0 false positives).
Collapse
|
6
|
Distributional social semantics: Inferring word meanings from communication patterns. Cogn Psychol 2021; 131:101441. [PMID: 34666227 DOI: 10.1016/j.cogpsych.2021.101441] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2021] [Revised: 10/05/2021] [Accepted: 10/07/2021] [Indexed: 11/20/2022]
Abstract
Distributional models of lexical semantics have proven to be powerful accounts of how word meanings are acquired from the natural language environment (Günther, Rinaldi, & Marelli, 2019; Kumar, 2020). Standard models of this type acquire the meaning of words through the learning of word co-occurrence statistics across large corpora. However, these models ignore social and communicative aspects of language processing, which is considered central to usage-based and adaptive theories of language (Tomasello, 2003; Beckner et al., 2009). Johns (2021) recently demonstrated that integrating social and communicative information into a lexical strength measure allowed for benchmark fits to be attained for lexical organization data, indicating that the social world contains important statistical information for language learning and processing. Through the analysis of the communication patterns of over 330,000 individuals on the online forum Reddit, totaling approximately 55 billion words of text, the findings of the current article demonstrates that social information about word usage allows for unique aspects of a word's meaning to be acquired, providing a new pathway for distributional model development.
Collapse
|
7
|
Hersche M, Lippuner S, Korb M, Benini L, Rahimi A. Near-channel classifier: symbiotic communication and classification in high-dimensional space. Brain Inform 2021; 8:16. [PMID: 34403011 PMCID: PMC8371050 DOI: 10.1186/s40708-021-00138-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2021] [Accepted: 07/29/2021] [Indexed: 12/02/2022] Open
Abstract
Brain-inspired high-dimensional (HD) computing represents and manipulates data using very long, random vectors with dimensionality in the thousands. This representation provides great robustness for various classification tasks where classifiers operate at low signal-to-noise ratio (SNR) conditions. Similarly, hyperdimensional modulation (HDM) leverages the robustness of complex-valued HD representations to reliably transmit information over a wireless channel, achieving a similar SNR gain compared to state-of-the-art codes. Here, we first propose methods to improve HDM in two ways: (1) reducing the complexity of encoding and decoding operations by generating, manipulating, and transmitting bipolar or integer vectors instead of complex vectors; (2) increasing the SNR gain by 0.2 dB using a new soft-feedback decoder; it can also increase the additive superposition capacity of HD vectors up to 1.7\documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\times$$\end{document}× in noise-free cases. Secondly, we propose to combine encoding/decoding aspects of communication with classification into a single framework by relying on multifaceted HD representations. This leads to a near-channel classification (NCC) approach that avoids transformations between different representations and the overhead of multiple layers of encoding/decoding, hence reducing latency and complexity of a wireless smart distributed system while providing robustness against noise and interference from other nodes. We provide a use-case for wearable hand gesture recognition with 5 classes from 64 EMG sensors, where the encoded vectors are transmitted to a remote node for either performing NCC, or reconstruction of the encoded data. In NCC mode, the original classification accuracy of 94% is maintained, even in the channel at SNR of 0 dB, by transmitting 10,000-bit vectors. We remove the redundancy by reducing the vector dimensionality to 2048-bit that still exhibits a graceful degradation: less than 6% accuracy loss is occurred in the channel at − 5 dB, and with the interference from 6 nodes that simultaneously transmit their encoded vectors. In the reconstruction mode, it improves the mean-squared error by up to 20 dB, compared to standard decoding, when transmitting 2048-dimensional vectors.
Collapse
Affiliation(s)
- Michael Hersche
- Integrated Systems Laboratory, ETH Zurich, Zurich, Switzerland. .,IBM Research-Zurich, Zurich, Switzerland.
| | - Stefan Lippuner
- Integrated Systems Laboratory, ETH Zurich, Zurich, Switzerland
| | - Matthias Korb
- Integrated Systems Laboratory, ETH Zurich, Zurich, Switzerland.,Institute of Microelectronics and Integrated Circuits, Bundeswehr University, Munich, Germany
| | - Luca Benini
- Integrated Systems Laboratory, ETH Zurich, Zurich, Switzerland.,Department of Electrical, Electronic, and Information Engineering, University of Bologna, Bologna, Italy
| | | |
Collapse
|
8
|
Taler V, Johns BT, Jones MN. A Large-Scale Semantic Analysis of Verbal Fluency Across the Aging Spectrum: Data From the Canadian Longitudinal Study on Aging. J Gerontol B Psychol Sci Soc Sci 2021; 75:e221-e230. [PMID: 30624721 DOI: 10.1093/geronb/gbz003] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2018] [Accepted: 01/07/2019] [Indexed: 11/12/2022] Open
Abstract
OBJECTIVES The present study aimed to characterize changes in verbal fluency performance across the lifespan using data from the Canadian Longitudinal Study on Aging (CLSA). METHODS We examined verbal fluency performance in a large sample of adults aged 45-85 (n = 12,686). Data are from the Tracking cohort of the CLSA. Participants completed a computer-assisted telephone interview that included an animal fluency task, in which they were asked to name as many animals as they could in 1 min. We employed a computational modeling approach to examine the factors driving performance on this task. RESULTS We found that the sequence of items produced was best predicted by their semantic neighborhood, and that pairwise similarity accounted for most of the variance in participant analyses. Moreover, the total number of items produced declined slightly with age, and older participants produced items of higher frequency and denser semantic neighborhood than younger adults. DISCUSSION These findings indicate subtle changes in the way people perform this task as they age. The use of computational models allowed for a large increase in the amount of variance accounted for in this data set over standard assessment types, providing important theoretical insights into the aging process.
Collapse
Affiliation(s)
- Vanessa Taler
- School of Psychology, University of Ottawa, Canada
- Bruyère Research Institute, Ottawa, Canada
| | - Brendan T Johns
- Department of Communicative Disorders and Sciences, University at Buffalo, New York
| | - Michael N Jones
- Department of Psychological and Brain Sciences, Indiana University, Bloomington
| |
Collapse
|
9
|
The influence of place and time on lexical behavior: A distributional analysis. Behav Res Methods 2019; 51:2438-2453. [DOI: 10.3758/s13428-019-01289-z] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
10
|
Abstract
Psychologists have made substantial progress at developing empirically validated formal expressions of how people perceive, learn, remember, think, and know. In this article, we present an academic search engine for cognitive psychology that leverages computational expressions of human cognition (vector-space models of semantics) to represent and find articles in the psychological record. The method shows how psychological theory can be used to inform and aid the design of psychologically intuitive computer interfaces.
Collapse
|
11
|
Gosmann J, Eliasmith C. Vector-Derived Transformation Binding: An Improved Binding Operation for Deep Symbol-Like Processing in Neural Networks. Neural Comput 2019; 31:849-869. [DOI: 10.1162/neco_a_01179] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
We present a new binding operation, vector-derived transformation binding (VTB), for use in vector symbolic architectures (VSA). The performance of VTB is compared to circular convolution, used in holographic reduced representations (HRRs), in terms of list and stack encoding capacity. A special focus is given to the possibility of a neural implementation by the means of the Neural Engineering Framework (NEF). While the scaling of required neural resources is slightly worse for VTB, it is found to be on par with circular convolution for list encoding and better for encoding of stacks. Furthermore, VTB influences the vector length less, which also benefits a neural implementation. Consequently, we argue that VTB is an improvement over HRRs for neurally implemented VSAs.
Collapse
Affiliation(s)
- Jan Gosmann
- Centre for Theoretical Neuroscience, University of Waterloo, Waterloo, ON N2L 3G1 Canada
| | - Chris Eliasmith
- Centre for Theoretical Neuroscience, University of Waterloo, Waterloo, ON N2L 3G1 Canada
| |
Collapse
|
12
|
Johns BT. Mining a Crowdsourced Dictionary to Understand Consistency and Preference in Word Meanings. Front Psychol 2019; 10:268. [PMID: 30833917 PMCID: PMC6387934 DOI: 10.3389/fpsyg.2019.00268] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2018] [Accepted: 01/28/2019] [Indexed: 12/02/2022] Open
Abstract
Big data approaches to psychology have become increasing popular (Jones, 2017). Two of the main developments of this line of research is the advent of distributional models of semantics (e.g., Landauer and Dumais, 1997), which learn the meaning of words from large text corpora, and the collection of mega datasets of human behavior (e.g., The English lexicon project; Balota et al., 2007). The current article combines these two approaches, with the goal being to understand the consistency and preference that people have for word meanings. This was accomplished by mining a large amount of data from an online, crowdsourced dictionary and analyzing this data with a distributional model. Overall, it was found that even for words that are not an active part of the language environment, there is a large amount of consistency in the word meanings that different people have. Additionally, it was demonstrated that users of a language have strong preferences for word meanings, such that definitions to words that do not conform to people’s conceptions are rejected by a community of language users. The results of this article provides insights into the cultural evolution of word meanings, and sheds light on alternative methodologies that can be used to understand lexical behavior.
Collapse
Affiliation(s)
- Brendan T Johns
- Department of Communicative Disorders and Sciences, University at Buffalo, Buffalo, NY, United States
| |
Collapse
|
13
|
Kleyko D, Rahimi A, Rachkovskij DA, Osipov E, Rabaey JM. Classification and Recall With Binary Hyperdimensional Computing: Tradeoffs in Choice of Density and Mapping Characteristics. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:5880-5898. [PMID: 29993669 DOI: 10.1109/tnnls.2018.2814400] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Hyperdimensional (HD) computing is a promising paradigm for future intelligent electronic appliances operating at low power. This paper discusses tradeoffs of selecting parameters of binary HD representations when applied to pattern recognition tasks. Particular design choices include density of representations and strategies for mapping data from the original representation. It is demonstrated that for the considered pattern recognition tasks (using synthetic and real-world data) both sparse and dense representations behave nearly identically. This paper also discusses implementation peculiarities which may favor one type of representations over the other. Finally, the capacity of representations of various densities is discussed.
Collapse
|
14
|
|
15
|
Controlling Individuals Growth in Semantic Genetic Programming through Elitist Replacement. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2016; 2016:8326760. [PMID: 27057158 PMCID: PMC4707023 DOI: 10.1155/2016/8326760] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/06/2015] [Revised: 09/29/2015] [Accepted: 10/01/2015] [Indexed: 12/05/2022]
Abstract
In 2012, Moraglio and coauthors introduced new genetic operators for Genetic Programming, called geometric semantic genetic operators. They have the very interesting advantage of inducing a unimodal error surface for any supervised learning problem. At the same time, they have the important drawback of generating very large data models that are usually very hard to understand and interpret. The objective of this work is to alleviate this drawback, still maintaining the advantage. More in particular, we propose an elitist version of geometric semantic operators, in which offspring are accepted in the new population only if they have better fitness than their parents. We present experimental evidence, on five complex real-life test problems, that this simple idea allows us to obtain results of a comparable quality (in terms of fitness), but with much smaller data models, compared to the standard geometric semantic operators. In the final part of the paper, we also explain the reason why we consider this a significant improvement, showing that the proposed elitist operators generate manageable models, while the models generated by the standard operators are so large in size that they can be considered unmanageable.
Collapse
|
16
|
Chubala CM, Johns BT, Jamieson RK, Mewhort DJK. Applying an exemplar model to an implicit rule-learning task: Implicit learning of semantic structure. Q J Exp Psychol (Hove) 2016; 69:1049-55. [PMID: 26730987 DOI: 10.1080/17470218.2015.1130068] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
Studies of implicit learning often examine peoples' sensitivity to sequential structure. Computational accounts have evolved to reflect this bias. An experiment conducted by Neil and Higham [Neil, G. J., & Higham, P. A.(2012). Implicit learning of conjunctive rule sets: An alternative to artificial grammars. Consciousness and Cognition, 21, 1393-1400] points to limitations in the sequential approach. In the experiment, participants studied words selected according to a conjunctive rule. At test, participants discriminated rule-consistent from rule-violating words but could not verbalize the rule. Although the data elude explanation by sequential models, an exemplar model of implicit learning can explain them. To make the case, we simulate the full pattern of results by incorporating vector representations for the words used in the experiment, derived from the large-scale semantic space models LSA and BEAGLE, into an exemplar model of memory, MINERVA 2. We show that basic memory processes in a classic model of memory capture implicit learning of non-sequential rules, provided that stimuli are appropriately represented.
Collapse
Affiliation(s)
- Chrissy M Chubala
- a Department of Psychology , University of Manitoba , Winnipeg , MB , Canada
| | - Brendan T Johns
- b Department of Communicative Disorders and Sciences , State University of New York at Buffalo , Buffalo , NY , USA
| | - Randall K Jamieson
- a Department of Psychology , University of Manitoba , Winnipeg , MB , Canada
| | - D J K Mewhort
- c Department of Psychology , Queen's University at Kingston , Kingston , ON , Canada
| |
Collapse
|