1
|
Daruwalla K, Lipasti M. Information bottleneck-based Hebbian learning rule naturally ties working memory and synaptic updates. Front Comput Neurosci 2024; 18:1240348. [PMID: 38818385 PMCID: PMC11137249 DOI: 10.3389/fncom.2024.1240348] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 04/26/2024] [Indexed: 06/01/2024] Open
Abstract
Deep neural feedforward networks are effective models for a wide array of problems, but training and deploying such networks presents a significant energy cost. Spiking neural networks (SNNs), which are modeled after biologically realistic neurons, offer a potential solution when deployed correctly on neuromorphic computing hardware. Still, many applications train SNNs offline, and running network training directly on neuromorphic hardware is an ongoing research problem. The primary hurdle is that back-propagation, which makes training such artificial deep networks possible, is biologically implausible. Neuroscientists are uncertain about how the brain would propagate a precise error signal backward through a network of neurons. Recent progress addresses part of this question, e.g., the weight transport problem, but a complete solution remains intangible. In contrast, novel learning rules based on the information bottleneck (IB) train each layer of a network independently, circumventing the need to propagate errors across layers. Instead, propagation is implicit due the layers' feedforward connectivity. These rules take the form of a three-factor Hebbian update a global error signal modulates local synaptic updates within each layer. Unfortunately, the global signal for a given layer requires processing multiple samples concurrently, and the brain only sees a single sample at a time. We propose a new three-factor update rule where the global signal correctly captures information across samples via an auxiliary memory network. The auxiliary network can be trained a priori independently of the dataset being used with the primary network. We demonstrate comparable performance to baselines on image classification tasks. Interestingly, unlike back-propagation-like schemes where there is no link between learning and memory, our rule presents a direct connection between working memory and synaptic updates. To the best of our knowledge, this is the first rule to make this link explicit. We explore these implications in initial experiments examining the effect of memory capacity on learning performance. Moving forward, this work suggests an alternate view of learning where each layer balances memory-informed compression against task performance. This view naturally encompasses several key aspects of neural computation, including memory, efficiency, and locality.
Collapse
Affiliation(s)
- Kyle Daruwalla
- Cold Spring Harbor Laboratory, Long Island, NY, United States
| | - Mikko Lipasti
- Electrical and Computer Engineering Department, University of Wisconsin-Madison, Madison, WI, United States
| |
Collapse
|
2
|
Agnes EJ, Vogels TP. Co-dependent excitatory and inhibitory plasticity accounts for quick, stable and long-lasting memories in biological networks. Nat Neurosci 2024; 27:964-974. [PMID: 38509348 PMCID: PMC11089004 DOI: 10.1038/s41593-024-01597-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Accepted: 02/08/2024] [Indexed: 03/22/2024]
Abstract
The brain's functionality is developed and maintained through synaptic plasticity. As synapses undergo plasticity, they also affect each other. The nature of such 'co-dependency' is difficult to disentangle experimentally, because multiple synapses must be monitored simultaneously. To help understand the experimentally observed phenomena, we introduce a framework that formalizes synaptic co-dependency between different connection types. The resulting model explains how inhibition can gate excitatory plasticity while neighboring excitatory-excitatory interactions determine the strength of long-term potentiation. Furthermore, we show how the interplay between excitatory and inhibitory synapses can account for the quick rise and long-term stability of a variety of synaptic weight profiles, such as orientation tuning and dendritic clustering of co-active synapses. In recurrent neuronal networks, co-dependent plasticity produces rich and stable motor cortex-like dynamics with high input sensitivity. Our results suggest an essential role for the neighborly synaptic interaction during learning, connecting micro-level physiology with network-wide phenomena.
Collapse
Affiliation(s)
- Everton J Agnes
- Centre for Neural Circuits and Behaviour, University of Oxford, Oxford, UK.
- Biozentrum, University of Basel, Basel, Switzerland.
| | - Tim P Vogels
- Centre for Neural Circuits and Behaviour, University of Oxford, Oxford, UK
- Institute of Science and Technology Austria, Klosterneuburg, Austria
| |
Collapse
|
3
|
Galloni AR, Yuan Y, Zhu M, Yu H, Bisht RS, Wu CTM, Grienberger C, Ramanathan S, Milstein AD. Neuromorphic one-shot learning utilizing a phase-transition material. Proc Natl Acad Sci U S A 2024; 121:e2318362121. [PMID: 38630718 PMCID: PMC11047090 DOI: 10.1073/pnas.2318362121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 03/25/2024] [Indexed: 04/19/2024] Open
Abstract
Design of hardware based on biological principles of neuronal computation and plasticity in the brain is a leading approach to realizing energy- and sample-efficient AI and learning machines. An important factor in selection of the hardware building blocks is the identification of candidate materials with physical properties suitable to emulate the large dynamic ranges and varied timescales of neuronal signaling. Previous work has shown that the all-or-none spiking behavior of neurons can be mimicked by threshold switches utilizing material phase transitions. Here, we demonstrate that devices based on a prototypical metal-insulator-transition material, vanadium dioxide (VO2), can be dynamically controlled to access a continuum of intermediate resistance states. Furthermore, the timescale of their intrinsic relaxation can be configured to match a range of biologically relevant timescales from milliseconds to seconds. We exploit these device properties to emulate three aspects of neuronal analog computation: fast (~1 ms) spiking in a neuronal soma compartment, slow (~100 ms) spiking in a dendritic compartment, and ultraslow (~1 s) biochemical signaling involved in temporal credit assignment for a recently discovered biological mechanism of one-shot learning. Simulations show that an artificial neural network using properties of VO2 devices to control an agent navigating a spatial environment can learn an efficient path to a reward in up to fourfold fewer trials than standard methods. The phase relaxations described in our study may be engineered in a variety of materials and can be controlled by thermal, electrical, or optical stimuli, suggesting further opportunities to emulate biological learning in neuromorphic hardware.
Collapse
Affiliation(s)
- Alessandro R. Galloni
- Department of Neuroscience and Cell Biology, Robert Wood Johnson Medical School, Rutgers, The State University of New Jersey, Piscataway, NJ08854
- Center for Advanced Biotechnology and Medicine, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Yifan Yuan
- Department of Electrical and Computer Engineering, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Minning Zhu
- Department of Electrical and Computer Engineering, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Haoming Yu
- School of Materials Engineering, Purdue University, West Lafayette, IN47907
| | - Ravindra S. Bisht
- Department of Electrical and Computer Engineering, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Chung-Tse Michael Wu
- Department of Electrical and Computer Engineering, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Christine Grienberger
- Department of Neuroscience, Brandeis University, Waltham, MA02453
- Department of Biology and Volen National Center for Complex Systems, Brandeis University, Waltham, MA02453
| | - Shriram Ramanathan
- Department of Electrical and Computer Engineering, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| | - Aaron D. Milstein
- Department of Neuroscience and Cell Biology, Robert Wood Johnson Medical School, Rutgers, The State University of New Jersey, Piscataway, NJ08854
- Center for Advanced Biotechnology and Medicine, Rutgers, The State University of New Jersey, Piscataway, NJ08854
| |
Collapse
|
4
|
Lakshminarasimhan KJ, Xie M, Cohen JD, Sauerbrei BA, Hantman AW, Litwin-Kumar A, Escola S. Specific connectivity optimizes learning in thalamocortical loops. Cell Rep 2024; 43:114059. [PMID: 38602873 PMCID: PMC11104520 DOI: 10.1016/j.celrep.2024.114059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Revised: 01/04/2024] [Accepted: 03/20/2024] [Indexed: 04/13/2024] Open
Abstract
Thalamocortical loops have a central role in cognition and motor control, but precisely how they contribute to these processes is unclear. Recent studies showing evidence of plasticity in thalamocortical synapses indicate a role for the thalamus in shaping cortical dynamics through learning. Since signals undergo a compression from the cortex to the thalamus, we hypothesized that the computational role of the thalamus depends critically on the structure of corticothalamic connectivity. To test this, we identified the optimal corticothalamic structure that promotes biologically plausible learning in thalamocortical synapses. We found that corticothalamic projections specialized to communicate an efference copy of the cortical output benefit motor control, while communicating the modes of highest variance is optimal for working memory tasks. We analyzed neural recordings from mice performing grasping and delayed discrimination tasks and found corticothalamic communication consistent with these predictions. These results suggest that the thalamus orchestrates cortical dynamics in a functionally precise manner through structured connectivity.
Collapse
Affiliation(s)
| | - Marjorie Xie
- Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY 10027, USA
| | - Jeremy D Cohen
- Neuroscience Center, University of North Carolina, Chapel Hill, NC 27559, USA
| | - Britton A Sauerbrei
- Department of Neurosciences, Case Western Reserve University, Cleveland, OH 44106, USA
| | - Adam W Hantman
- Neuroscience Center, University of North Carolina, Chapel Hill, NC 27559, USA
| | - Ashok Litwin-Kumar
- Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY 10027, USA.
| | - Sean Escola
- Department of Psychiatry, Columbia University, New York, NY 10032, USA.
| |
Collapse
|
5
|
Pagkalos M, Makarov R, Poirazi P. Leveraging dendritic properties to advance machine learning and neuro-inspired computing. Curr Opin Neurobiol 2024; 85:102853. [PMID: 38394956 DOI: 10.1016/j.conb.2024.102853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Revised: 02/04/2024] [Accepted: 02/05/2024] [Indexed: 02/25/2024]
Abstract
The brain is a remarkably capable and efficient system. It can process and store huge amounts of noisy and unstructured information, using minimal energy. In contrast, current artificial intelligence (AI) systems require vast resources for training while still struggling to compete in tasks that are trivial for biological agents. Thus, brain-inspired engineering has emerged as a promising new avenue for designing sustainable, next-generation AI systems. Here, we describe how dendritic mechanisms of biological neurons have inspired innovative solutions for significant AI problems, including credit assignment in multi-layer networks, catastrophic forgetting, and high-power consumption. These findings provide exciting alternatives to existing architectures, showing how dendritic research can pave the way for building more powerful and energy efficient artificial learning systems.
Collapse
Affiliation(s)
- Michalis Pagkalos
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece; Department of Biology, University of Crete, Heraklion, 70013, Greece. https://twitter.com/MPagkalos
| | - Roman Makarov
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece; Department of Biology, University of Crete, Heraklion, 70013, Greece. https://twitter.com/_RomanMakarov
| | - Panayiota Poirazi
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece.
| |
Collapse
|
6
|
Fitz H, Hagoort P, Petersson KM. Neurobiological Causal Models of Language Processing. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2024; 5:225-247. [PMID: 38645618 PMCID: PMC11025648 DOI: 10.1162/nol_a_00133] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Accepted: 12/18/2023] [Indexed: 04/23/2024]
Abstract
The language faculty is physically realized in the neurobiological infrastructure of the human brain. Despite significant efforts, an integrated understanding of this system remains a formidable challenge. What is missing from most theoretical accounts is a specification of the neural mechanisms that implement language function. Computational models that have been put forward generally lack an explicit neurobiological foundation. We propose a neurobiologically informed causal modeling approach which offers a framework for how to bridge this gap. A neurobiological causal model is a mechanistic description of language processing that is grounded in, and constrained by, the characteristics of the neurobiological substrate. It intends to model the generators of language behavior at the level of implementational causality. We describe key features and neurobiological component parts from which causal models can be built and provide guidelines on how to implement them in model simulations. Then we outline how this approach can shed new light on the core computational machinery for language, the long-term storage of words in the mental lexicon and combinatorial processing in sentence comprehension. In contrast to cognitive theories of behavior, causal models are formulated in the "machine language" of neurobiology which is universal to human cognition. We argue that neurobiological causal modeling should be pursued in addition to existing approaches. Eventually, this approach will allow us to develop an explicit computational neurobiology of language.
Collapse
Affiliation(s)
- Hartmut Fitz
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Neurobiology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Peter Hagoort
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Neurobiology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Karl Magnus Petersson
- Neurobiology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Faculty of Medicine and Biomedical Sciences, University of Algarve, Faro, Portugal
| |
Collapse
|
7
|
Anisetti VR, Kandala A, Scellier B, Schwarz JM. Frequency Propagation: Multimechanism Learning in Nonlinear Physical Networks. Neural Comput 2024; 36:596-620. [PMID: 38457749 DOI: 10.1162/neco_a_01648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2023] [Accepted: 11/20/2023] [Indexed: 03/10/2024]
Abstract
We introduce frequency propagation, a learning algorithm for nonlinear physical networks. In a resistive electrical circuit with variable resistors, an activation current is applied at a set of input nodes at one frequency and an error current is applied at a set of output nodes at another frequency. The voltage response of the circuit to these boundary currents is the superposition of an activation signal and an error signal whose coefficients can be read in different frequencies of the frequency domain. Each conductance is updated proportionally to the product of the two coefficients. The learning rule is local and proved to perform gradient descent on a loss function. We argue that frequency propagation is an instance of a multimechanism learning strategy for physical networks, be it resistive, elastic, or flow networks. Multimechanism learning strategies incorporate at least two physical quantities, potentially governed by independent physical mechanisms, to act as activation and error signals in the training process. Locally available information about these two signals is then used to update the trainable parameters to perform gradient descent. We demonstrate how earlier work implementing learning via chemical signaling in flow networks (Anisetti, Scellier, et al., 2023) also falls under the rubric of multimechanism learning.
Collapse
Affiliation(s)
| | - Ananth Kandala
- Department of Physics, University of Florida, Gainesville, FL 32611, U.S.A.
| | | | - J M Schwarz
- Physics Department, Syracuse University, Syracuse, NY 13244 U.S.A
- Indian Creek Farm, Ithaca, NY 14850, U.S.A.
| |
Collapse
|
8
|
Barlow BSM, Longtin A, Joós B. Impact on backpropagation of the spatial heterogeneity of sodium channel kinetics in the axon initial segment. PLoS Comput Biol 2024; 20:e1011846. [PMID: 38489374 PMCID: PMC10942053 DOI: 10.1371/journal.pcbi.1011846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Accepted: 01/21/2024] [Indexed: 03/17/2024] Open
Abstract
In a variety of neurons, action potentials (APs) initiate at the proximal axon, within a region called the axon initial segment (AIS), which has a high density of voltage-gated sodium channels (NaVs) on its membrane. In pyramidal neurons, the proximal AIS has been reported to exhibit a higher proportion of NaVs with gating properties that are "right-shifted" to more depolarized voltages, compared to the distal AIS. Further, recent experiments have revealed that as neurons develop, the spatial distribution of NaV subtypes along the AIS can change substantially, suggesting that neurons tune their excitability by modifying said distribution. When neurons are stimulated axonally, computational modelling has shown that this spatial separation of gating properties in the AIS enhances the backpropagation of APs into the dendrites. In contrast, in the more natural scenario of somatic stimulation, our simulations show that the same distribution can impede backpropagation, suggesting that the choice of orthodromic versus antidromic stimulation can bias or even invert experimental findings regarding the role of NaV subtypes in the AIS. We implemented a range of hypothetical NaV distributions in the AIS of three multicompartmental pyramidal cell models and investigated the precise kinetic mechanisms underlying such effects, as the spatial distribution of NaV subtypes is varied. With axonal stimulation, proximal NaV availability dominates, such that concentrating right-shifted NaVs in the proximal AIS promotes backpropagation. However, with somatic stimulation, the models are insensitive to availability kinetics. Instead, the higher activation threshold of right-shifted NaVs in the AIS impedes backpropagation. Therefore, recently observed developmental changes to the spatial separation and relative proportions of NaV1.2 and NaV1.6 in the AIS differentially impact activation and availability. The observed effects on backpropagation, and potentially learning via its putative role in synaptic plasticity (e.g. through spike-timing-dependent plasticity), are opposite for orthodromic versus antidromic stimulation, which should inform hypotheses about the impact of the developmentally regulated subcellular localization of these NaV subtypes.
Collapse
Affiliation(s)
- Benjamin S. M. Barlow
- Department of Physics, University of Ottawa, STEM Complex, 150 Louis-Pasteur Pvt, Ottawa, Ontario, Canada
| | - André Longtin
- Department of Physics, University of Ottawa, STEM Complex, 150 Louis-Pasteur Pvt, Ottawa, Ontario, Canada
- Center for Neural Dynamics and AI, University of Ottawa, Ottawa, Ontario, Canada
- Department of Cellular and Molecular Medicine, Faculty of Medicine, University of Ottawa, Ottawa, Ontario, Canada
| | - Béla Joós
- Department of Physics, University of Ottawa, STEM Complex, 150 Louis-Pasteur Pvt, Ottawa, Ontario, Canada
- Center for Neural Dynamics and AI, University of Ottawa, Ottawa, Ontario, Canada
| |
Collapse
|
9
|
Beninger J, Rossbroich J, Tóth K, Naud R. Functional subtypes of synaptic dynamics in mouse and human. Cell Rep 2024; 43:113785. [PMID: 38363673 DOI: 10.1016/j.celrep.2024.113785] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2023] [Revised: 12/08/2023] [Accepted: 01/27/2024] [Indexed: 02/18/2024] Open
Abstract
Synapses preferentially respond to particular temporal patterns of activity with a large degree of heterogeneity that is informally or tacitly separated into classes. Yet, the precise number and properties of such classes are unclear. Do they exist on a continuum and, if so, when is it appropriate to divide that continuum into functional regions? In a large dataset of glutamatergic cortical connections, we perform model-based characterization to infer the number and characteristics of functionally distinct subtypes of synaptic dynamics. In rodent data, we find five clusters that partially converge with transgenic-associated subtypes. Strikingly, the application of the same clustering method in human data infers a highly similar number of clusters, supportive of stable clustering. This nuanced dictionary of functional subtypes shapes the heterogeneity of cortical synaptic dynamics and provides a lens into the basic motifs of information transmission in the brain.
Collapse
Affiliation(s)
- John Beninger
- Center for Neural Dynamics and Artificial Intelligence, University of Ottawa, Ottawa, ON K1H 8M5, Canada; uOttawa Brain and Mind Research Institute, University of Ottawa, Ottawa, ON K1H 8M5, Canada; Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, ON K1H 8M5, Canada
| | - Julian Rossbroich
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland; Faculty of Science, University of Basel, Basel, Switzerland
| | - Katalin Tóth
- Center for Neural Dynamics and Artificial Intelligence, University of Ottawa, Ottawa, ON K1H 8M5, Canada; uOttawa Brain and Mind Research Institute, University of Ottawa, Ottawa, ON K1H 8M5, Canada; Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, ON K1H 8M5, Canada
| | - Richard Naud
- Center for Neural Dynamics and Artificial Intelligence, University of Ottawa, Ottawa, ON K1H 8M5, Canada; uOttawa Brain and Mind Research Institute, University of Ottawa, Ottawa, ON K1H 8M5, Canada; Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, ON K1H 8M5, Canada; Department of Physics, University of Ottawa, Ottawa, ON K1H 8M5, Canada.
| |
Collapse
|
10
|
Ford AN, Czarny JE, Rogalla MM, Quass GL, Apostolides PF. Auditory Corticofugal Neurons Transmit Auditory and Non-auditory Information During Behavior. J Neurosci 2024; 44:e1190232023. [PMID: 38123993 PMCID: PMC10869159 DOI: 10.1523/jneurosci.1190-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 11/08/2023] [Accepted: 11/29/2023] [Indexed: 12/23/2023] Open
Abstract
Layer 5 pyramidal neurons of sensory cortices project "corticofugal" axons to myriad sub-cortical targets, thereby broadcasting high-level signals important for perception and learning. Recent studies suggest dendritic Ca2+ spikes as key biophysical mechanisms supporting corticofugal neuron function: these long-lasting events drive burst firing, thereby initiating uniquely powerful signals to modulate sub-cortical representations and trigger learning-related plasticity. However, the behavioral relevance of corticofugal dendritic spikes is poorly understood. We shed light on this issue using 2-photon Ca2+ imaging of auditory corticofugal dendrites as mice of either sex engage in a GO/NO-GO sound-discrimination task. Unexpectedly, only a minority of dendritic spikes were triggered by behaviorally relevant sounds under our conditions. Task related dendritic activity instead mostly followed sound cue termination and co-occurred with mice's instrumental licking during the answer period of behavioral trials, irrespective of reward consumption. Temporally selective, optogenetic silencing of corticofugal neurons during the trial answer period impaired auditory discrimination learning. Thus, auditory corticofugal systems' contribution to learning and plasticity may be partially nonsensory in nature.
Collapse
Affiliation(s)
- Alexander N Ford
- Department of Otolaryngology/Head and Neck Surgery, Kresge Hearing Research Institute, Ann Arbor, Michigan 48109
| | - Jordyn E Czarny
- Department of Otolaryngology/Head and Neck Surgery, Kresge Hearing Research Institute, Ann Arbor, Michigan 48109
| | - Meike M Rogalla
- Department of Otolaryngology/Head and Neck Surgery, Kresge Hearing Research Institute, Ann Arbor, Michigan 48109
| | - Gunnar L Quass
- Department of Otolaryngology/Head and Neck Surgery, Kresge Hearing Research Institute, Ann Arbor, Michigan 48109
| | - Pierre F Apostolides
- Department of Otolaryngology/Head and Neck Surgery, Kresge Hearing Research Institute, Ann Arbor, Michigan 48109
- Department of Molecular and Integrative Physiology, University of Michigan Medical School, Ann Arbor, Michigan 48109
| |
Collapse
|
11
|
Ellwood IT. Short-term Hebbian learning can implement transformer-like attention. PLoS Comput Biol 2024; 20:e1011843. [PMID: 38277432 PMCID: PMC10849393 DOI: 10.1371/journal.pcbi.1011843] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 02/07/2024] [Accepted: 01/19/2024] [Indexed: 01/28/2024] Open
Abstract
Transformers have revolutionized machine learning models of language and vision, but their connection with neuroscience remains tenuous. Built from attention layers, they require a mass comparison of queries and keys that is difficult to perform using traditional neural circuits. Here, we show that neurons can implement attention-like computations using short-term, Hebbian synaptic potentiation. We call our mechanism the match-and-control principle and it proposes that when activity in an axon is synchronous, or matched, with the somatic activity of a neuron that it synapses onto, the synapse can be briefly strongly potentiated, allowing the axon to take over, or control, the activity of the downstream neuron for a short time. In our scheme, the keys and queries are represented as spike trains and comparisons between the two are performed in individual spines allowing for hundreds of key comparisons per query and roughly as many keys and queries as there are neurons in the network.
Collapse
Affiliation(s)
- Ian T. Ellwood
- Department of Neurobiology and Behavior, Cornell University, Ithaca, NY, United States of America
| |
Collapse
|
12
|
Capone C, Lupo C, Muratore P, Paolucci PS. Beyond spiking networks: The computational advantages of dendritic amplification and input segregation. Proc Natl Acad Sci U S A 2023; 120:e2220743120. [PMID: 38019856 PMCID: PMC10710097 DOI: 10.1073/pnas.2220743120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2022] [Accepted: 10/11/2023] [Indexed: 12/01/2023] Open
Abstract
The brain can efficiently learn a wide range of tasks, motivating the search for biologically inspired learning rules for improving current artificial intelligence technology. Most biological models are composed of point neurons and cannot achieve state-of-the-art performance in machine learning. Recent works have proposed that input segregation (neurons receive sensory information and higher-order feedback in segregated compartments), and nonlinear dendritic computation would support error backpropagation in biological neurons. However, these approaches require propagating errors with a fine spatiotemporal structure to all the neurons, which is unlikely to be feasible in a biological network. To relax this assumption, we suggest that bursts and dendritic input segregation provide a natural support for target-based learning, which propagates targets rather than errors. A coincidence mechanism between the basal and the apical compartments allows for generating high-frequency bursts of spikes. This architecture supports a burst-dependent learning rule, based on the comparison between the target bursting activity triggered by the teaching signal and the one caused by the recurrent connections, providing support for target-based learning. We show that this framework can be used to efficiently solve spatiotemporal tasks, such as context-dependent store and recall of three-dimensional trajectories, and navigation tasks. Finally, we suggest that this neuronal architecture naturally allows for orchestrating "hierarchical imitation learning", enabling the decomposition of challenging long-horizon decision-making tasks into simpler subtasks. We show a possible implementation of this in a two-level network, where the high network produces the contextual signal for the low network.
Collapse
Affiliation(s)
- Cristiano Capone
- Istituto Nazionale di Fisica Nucleare (INFN), Sezione di Roma, Rome00185, Italy
| | - Cosimo Lupo
- Istituto Nazionale di Fisica Nucleare (INFN), Sezione di Roma, Rome00185, Italy
| | - Paolo Muratore
- Scuola Internazionale Superiore di Studi Avanzati (SISSA), Visual Neuroscience Lab, Trieste34136, Italy
| | | |
Collapse
|
13
|
Makarov R, Pagkalos M, Poirazi P. Dendrites and efficiency: Optimizing performance and resource utilization. Curr Opin Neurobiol 2023; 83:102812. [PMID: 37980803 DOI: 10.1016/j.conb.2023.102812] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 10/19/2023] [Accepted: 10/21/2023] [Indexed: 11/21/2023]
Abstract
The brain is a highly efficient system that has evolved to optimize performance under limited resources. In this review, we highlight recent theoretical and experimental studies that support the view that dendrites make information processing and storage in the brain more efficient. This is achieved through the dynamic modulation of integration versus segregation of inputs and activity within a neuron. We argue that under conditions of limited energy and space, dendrites help biological networks to implement complex functions such as processing natural stimuli on behavioral timescales, performing the inference process on those stimuli in a context-specific manner, and storing the information in overlapping populations of neurons. A global picture starts to emerge, in which dendrites help the brain achieve efficiency through a combination of optimization strategies that balance the tradeoff between performance and resource utilization.
Collapse
Affiliation(s)
- Roman Makarov
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece; Department of Biology, University of Crete, Heraklion, 70013, Greece. https://twitter.com/_RomanMakarov
| | - Michalis Pagkalos
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece; Department of Biology, University of Crete, Heraklion, 70013, Greece. https://twitter.com/MPagkalos
| | - Panayiota Poirazi
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece.
| |
Collapse
|
14
|
Zahid U, Guo Q, Fountas Z. Predictive Coding as a Neuromorphic Alternative to Backpropagation: A Critical Evaluation. Neural Comput 2023; 35:1881-1909. [PMID: 37844326 DOI: 10.1162/neco_a_01620] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2023] [Accepted: 08/01/2023] [Indexed: 10/18/2023]
Abstract
Backpropagation has rapidly become the workhorse credit assignment algorithm for modern deep learning methods. Recently, modified forms of predictive coding (PC), an algorithm with origins in computational neuroscience, have been shown to result in approximately or exactly equal parameter updates to those under backpropagation. Due to this connection, it has been suggested that PC can act as an alternative to backpropagation with desirable properties that may facilitate implementation in neuromorphic systems. Here, we explore these claims using the different contemporary PC variants proposed in the literature. We obtain time complexity bounds for these PC variants, which we show are lower bounded by backpropagation. We also present key properties of these variants that have implications for neurobiological plausibility and their interpretations, particularly from the perspective of standard PC as a variational Bayes algorithm for latent probabilistic models. Our findings shed new light on the connection between the two learning frameworks and suggest that in its current forms, PC may have more limited potential as a direct replacement of backpropagation than previously envisioned.
Collapse
Affiliation(s)
- Umais Zahid
- Huawei Technologies R&D, London N19 3HT, U.K.
| | - Qinghai Guo
- Huawei Technologies R&D, Shenzhen 518129, China
| | | |
Collapse
|
15
|
Francioni V, Tang VD, Brown NJ, Toloza EH, Harnett M. Vectorized instructive signals in cortical dendrites during a brain-computer interface task. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.11.03.565534. [PMID: 37961227 PMCID: PMC10635122 DOI: 10.1101/2023.11.03.565534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/15/2023]
Abstract
Backpropagation of error is the most widely used learning algorithm in artificial neural networks, forming the backbone of modern machine learning and artificial intelligence1,2. Backpropagation provides a solution to the credit assignment problem by vectorizing an error signal tailored to individual neurons. Recent theoretical models have suggested that neural circuits could implement backpropagation-like learning by semi-independently processing feedforward and feedback information streams in separate dendritic compartments3-7. This presents a compelling, but untested, hypothesis for how cortical circuits could solve credit assignment in the brain. We designed a neurofeedback brain-computer interface (BCI) task with an experimenter-defined reward function to evaluate the key requirements for dendrites to implement backpropagation-like learning. We trained mice to modulate the activity of two spatially intermingled populations (4 or 5 neurons each) of layer 5 pyramidal neurons in the retrosplenial cortex to rotate a visual grating towards a target orientation while we recorded GCaMP activity from somas and corresponding distal apical dendrites. We observed that the relative magnitudes of somatic versus dendritic signals could be predicted using the activity of the surrounding network and contained information about task-related variables that could serve as instructive signals, including reward and error. The signs of these putative teaching signals both depended on the causal role of individual neurons in the task and predicted changes in overall activity over the course of learning. These results provide the first biological evidence of a backpropagation-like solution to the credit assignment problem in the brain.
Collapse
Affiliation(s)
- Valerio Francioni
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Vincent D Tang
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Norma J. Brown
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Enrique H.S. Toloza
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Mark Harnett
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| |
Collapse
|
16
|
Halvagal MS, Zenke F. The combination of Hebbian and predictive plasticity learns invariant object representations in deep sensory networks. Nat Neurosci 2023; 26:1906-1915. [PMID: 37828226 PMCID: PMC10620089 DOI: 10.1038/s41593-023-01460-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Accepted: 09/08/2023] [Indexed: 10/14/2023]
Abstract
Recognition of objects from sensory stimuli is essential for survival. To that end, sensory networks in the brain must form object representations invariant to stimulus changes, such as size, orientation and context. Although Hebbian plasticity is known to shape sensory networks, it fails to create invariant object representations in computational models, raising the question of how the brain achieves such processing. In the present study, we show that combining Hebbian plasticity with a predictive form of plasticity leads to invariant representations in deep neural network models. We derive a local learning rule that generalizes to spiking neural networks and naturally accounts for several experimentally observed properties of synaptic plasticity, including metaplasticity and spike-timing-dependent plasticity. Finally, our model accurately captures neuronal selectivity changes observed in the primate inferotemporal cortex in response to altered visual experience. Thus, we provide a plausible normative theory emphasizing the importance of predictive plasticity mechanisms for successful representational learning.
Collapse
Affiliation(s)
- Manu Srinath Halvagal
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland
- Faculty of Science, University of Basel, Basel, Switzerland
| | - Friedemann Zenke
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland.
- Faculty of Science, University of Basel, Basel, Switzerland.
| |
Collapse
|
17
|
Zhang Y, He G, Ma L, Liu X, Hjorth JJJ, Kozlov A, He Y, Zhang S, Kotaleski JH, Tian Y, Grillner S, Du K, Huang T. A GPU-based computational framework that bridges neuron simulation and artificial intelligence. Nat Commun 2023; 14:5798. [PMID: 37723170 PMCID: PMC10507119 DOI: 10.1038/s41467-023-41553-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2022] [Accepted: 09/08/2023] [Indexed: 09/20/2023] Open
Abstract
Biophysically detailed multi-compartment models are powerful tools to explore computational principles of the brain and also serve as a theoretical framework to generate algorithms for artificial intelligence (AI) systems. However, the expensive computational cost severely limits the applications in both the neuroscience and AI fields. The major bottleneck during simulating detailed compartment models is the ability of a simulator to solve large systems of linear equations. Here, we present a novel Dendritic Hierarchical Scheduling (DHS) method to markedly accelerate such a process. We theoretically prove that the DHS implementation is computationally optimal and accurate. This GPU-based method performs with 2-3 orders of magnitude higher speed than that of the classic serial Hines method in the conventional CPU platform. We build a DeepDendrite framework, which integrates the DHS method and the GPU computing engine of the NEURON simulator and demonstrate applications of DeepDendrite in neuroscience tasks. We investigate how spatial patterns of spine inputs affect neuronal excitability in a detailed human pyramidal neuron model with 25,000 spines. Furthermore, we provide a brief discussion on the potential of DeepDendrite for AI, specifically highlighting its ability to enable the efficient training of biophysically detailed models in typical image classification tasks.
Collapse
Affiliation(s)
- Yichen Zhang
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
| | - Gan He
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
| | - Lei Ma
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
- Beijing Academy of Artificial Intelligence (BAAI), Beijing, 100084, China
| | - Xiaofei Liu
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
- School of Information Science and Engineering, Yunnan University, Kunming, 650500, China
| | - J J Johannes Hjorth
- Science for Life Laboratory, School of Electrical Engineering and Computer Science, Royal Institute of Technology KTH, Stockholm, SE-10044, Sweden
| | - Alexander Kozlov
- Science for Life Laboratory, School of Electrical Engineering and Computer Science, Royal Institute of Technology KTH, Stockholm, SE-10044, Sweden
- Department of Neuroscience, Karolinska Institute, Stockholm, SE-17165, Sweden
| | - Yutao He
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
| | - Shenjian Zhang
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
| | - Jeanette Hellgren Kotaleski
- Science for Life Laboratory, School of Electrical Engineering and Computer Science, Royal Institute of Technology KTH, Stockholm, SE-10044, Sweden
- Department of Neuroscience, Karolinska Institute, Stockholm, SE-17165, Sweden
| | - Yonghong Tian
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
- School of Electrical and Computer Engineering, Shenzhen Graduate School, Peking University, Shenzhen, 518055, China
| | - Sten Grillner
- Department of Neuroscience, Karolinska Institute, Stockholm, SE-17165, Sweden
| | - Kai Du
- Institute for Artificial Intelligence, Peking University, Beijing, 100871, China.
| | - Tiejun Huang
- National Key Laboratory for Multimedia Information Processing, School of Computer Science, Peking University, Beijing, 100871, China
- Beijing Academy of Artificial Intelligence (BAAI), Beijing, 100084, China
- Institute for Artificial Intelligence, Peking University, Beijing, 100871, China
| |
Collapse
|
18
|
Liu H, Qin Y, Chen HY, Wu J, Ma J, Du Z, Wang N, Zou J, Lin S, Zhang X, Zhang Y, Wang H. Artificial Neuronal Devices Based on Emerging Materials: Neuronal Dynamics and Applications. ADVANCED MATERIALS (DEERFIELD BEACH, FLA.) 2023; 35:e2205047. [PMID: 36609920 DOI: 10.1002/adma.202205047] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/03/2022] [Revised: 12/02/2022] [Indexed: 06/17/2023]
Abstract
Artificial neuronal devices are critical building blocks of neuromorphic computing systems and currently the subject of intense research motivated by application needs from new computing technology and more realistic brain emulation. Researchers have proposed a range of device concepts that can mimic neuronal dynamics and functions. Although the switching physics and device structures of these artificial neurons are largely different, their behaviors can be described by several neuron models in a more unified manner. In this paper, the reports of artificial neuronal devices based on emerging volatile switching materials are reviewed from the perspective of the demonstrated neuron models, with a focus on the neuronal functions implemented in these devices and the exploitation of these functions for computational and sensing applications. Furthermore, the neuroscience inspirations and engineering methods to enrich the neuronal dynamics that remain to be implemented in artificial neuronal devices and networks toward realizing the full functionalities of biological neurons are discussed.
Collapse
Affiliation(s)
- Hefei Liu
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Yuan Qin
- Center for Power Electronics Systems, Bradley Department of Electrical and Computer Engineering, Virginia Polytechnic Institute and State University, Blacksburg, VA, 24060, USA
| | - Hung-Yu Chen
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jiangbin Wu
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jiahui Ma
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Zhonghao Du
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Nan Wang
- Mork Family Department of Chemical Engineering and Materials Science, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jingyi Zou
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Sen Lin
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Xu Zhang
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Yuhao Zhang
- Center for Power Electronics Systems, Bradley Department of Electrical and Computer Engineering, Virginia Polytechnic Institute and State University, Blacksburg, VA, 24060, USA
| | - Han Wang
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
- Mork Family Department of Chemical Engineering and Materials Science, University of Southern California, Los Angeles, CA, 90089, USA
| |
Collapse
|
19
|
Llobera J, Charbonnier C. Physics-based character animation and human motor control. Phys Life Rev 2023; 46:190-219. [PMID: 37480729 DOI: 10.1016/j.plrev.2023.06.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2023] [Accepted: 06/25/2023] [Indexed: 07/24/2023]
Abstract
Motor neuroscience and physics-based character animation (PBCA) approach human and humanoid control from different perspectives. The primary goal of PBCA is to control the movement of a ragdoll (humanoid or animal) applying forces and torques within a physical simulation. The primary goal of motor neuroscience is to understand the contribution of different parts of the nervous system to generate coordinated movements. We review the functional principles and the functional anatomy of human motor control and the main strategies used in PBCA. We then explore common research points by discussing the functional anatomy and ongoing debates in motor neuroscience from the perspective of PBCA. We also suggest there are several benefits to be found in studying sensorimotor integration and human-character coordination through closer collaboration between these two fields.
Collapse
Affiliation(s)
- Joan Llobera
- Artanim Foundation, 40, chemin du Grand-Puits, 1217 Meyrin - Geneva, Switzerland.
| | - Caecilia Charbonnier
- Artanim Foundation, 40, chemin du Grand-Puits, 1217 Meyrin - Geneva, Switzerland
| |
Collapse
|
20
|
Bernáez Timón L, Ekelmans P, Kraynyukova N, Rose T, Busse L, Tchumatchenko T. How to incorporate biological insights into network models and why it matters. J Physiol 2023; 601:3037-3053. [PMID: 36069408 DOI: 10.1113/jp282755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2022] [Accepted: 08/24/2022] [Indexed: 11/08/2022] Open
Abstract
Due to the staggering complexity of the brain and its neural circuitry, neuroscientists rely on the analysis of mathematical models to elucidate its function. From Hodgkin and Huxley's detailed description of the action potential in 1952 to today, new theories and increasing computational power have opened up novel avenues to study how neural circuits implement the computations that underlie behaviour. Computational neuroscientists have developed many models of neural circuits that differ in complexity, biological realism or emergent network properties. With recent advances in experimental techniques for detailed anatomical reconstructions or large-scale activity recordings, rich biological data have become more available. The challenge when building network models is to reflect experimental results, either through a high level of detail or by finding an appropriate level of abstraction. Meanwhile, machine learning has facilitated the development of artificial neural networks, which are trained to perform specific tasks. While they have proven successful at achieving task-oriented behaviour, they are often abstract constructs that differ in many features from the physiology of brain circuits. Thus, it is unclear whether the mechanisms underlying computation in biological circuits can be investigated by analysing artificial networks that accomplish the same function but differ in their mechanisms. Here, we argue that building biologically realistic network models is crucial to establishing causal relationships between neurons, synapses, circuits and behaviour. More specifically, we advocate for network models that consider the connectivity structure and the recorded activity dynamics while evaluating task performance.
Collapse
Affiliation(s)
- Laura Bernáez Timón
- Institute for Physiological Chemistry, University of Mainz Medical Center, Mainz, Germany
| | - Pierre Ekelmans
- Frankfurt Institute for Advanced Studies, Frankfurt, Germany
| | - Nataliya Kraynyukova
- Institute of Experimental Epileptology and Cognition Research, University of Bonn Medical Center, Bonn, Germany
| | - Tobias Rose
- Institute of Experimental Epileptology and Cognition Research, University of Bonn Medical Center, Bonn, Germany
| | - Laura Busse
- Division of Neurobiology, Faculty of Biology, LMU Munich, Munich, Germany
- Bernstein Center for Computational Neuroscience, Munich, Germany
| | - Tatjana Tchumatchenko
- Institute for Physiological Chemistry, University of Mainz Medical Center, Mainz, Germany
- Institute of Experimental Epileptology and Cognition Research, University of Bonn Medical Center, Bonn, Germany
| |
Collapse
|
21
|
Richards BA, Kording KP. The study of plasticity has always been about gradients. J Physiol 2023; 601:3141-3149. [PMID: 37078235 DOI: 10.1113/jp282747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2022] [Accepted: 04/11/2023] [Indexed: 04/21/2023] Open
Abstract
The experimental study of learning and plasticity has always been driven by an implicit question: how can physiological changes be adaptive and improve performance? For example, in Hebbian plasticity only synapses from presynaptic neurons that were active are changed, avoiding useless changes. Similarly, in dopamine-gated learning synapse changes depend on reward or lack thereof and do not change when everything is predictable. Within machine learning we can make the question of which changes are adaptive concrete: performance improves when changes correlate with the gradient of an objective function quantifying performance. This result is general for any system that improves through small changes. As such, physiology has always implicitly been seeking mechanisms that allow the brain to approximate gradients. Coming from this perspective we review the existing literature on plasticity-related mechanisms, and we show how these mechanisms relate to gradient estimation. We argue that gradients are a unifying idea to explain the many facets of neuronal plasticity.
Collapse
Affiliation(s)
- Blake Aaron Richards
- Mila, Montreal, Quebec, Canada
- School of Computer Science, McGill University, Montreal, Quebec, Canada
- Department of Neurology & Neurosurgery, McGill University, Montreal, Quebec, Canada
- Montreal Neurological Institute, Montreal, Quebec, Canada
- Learning in Machines and Brains Program, CIFAR, Toronto, Ontario, Canada
| | - Konrad Paul Kording
- Learning in Machines and Brains Program, CIFAR, Toronto, Ontario, Canada
- Department of Bioengineering, University of Pennsylvania, Philadelphia, Pennsylvania, USA
- Department of Neuroscience, University of Pennsylvania, Philadelphia, Pennsylvania, USA
| |
Collapse
|
22
|
Levenstein D, Okun M. Logarithmically scaled, gamma distributed neuronal spiking. J Physiol 2023; 601:3055-3069. [PMID: 36086892 PMCID: PMC10952267 DOI: 10.1113/jp282758] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2022] [Accepted: 07/28/2022] [Indexed: 11/08/2022] Open
Abstract
Naturally log-scaled quantities abound in the nervous system. Distributions of these quantities have non-intuitive properties, which have implications for data analysis and the understanding of neural circuits. Here, we review the log-scaled statistics of neuronal spiking and the relevant analytical probability distributions. Recent work using log-scaling revealed that interspike intervals of forebrain neurons segregate into discrete modes reflecting spiking at different timescales and are each well-approximated by a gamma distribution. Each neuron spends most of the time in an irregular spiking 'ground state' with the longest intervals, which determines the mean firing rate of the neuron. Across the entire neuronal population, firing rates are log-scaled and well approximated by the gamma distribution, with a small number of highly active neurons and an overabundance of low rate neurons (the 'dark matter'). These results are intricately linked to a heterogeneous balanced operating regime, which confers upon neuronal circuits multiple computational advantages and has evolutionarily ancient origins.
Collapse
Affiliation(s)
- Daniel Levenstein
- Department of Neurology and NeurosurgeryMcGill UniversityMontrealQCCanada
- MilaMontréalQCCanada
| | - Michael Okun
- Department of Psychology and Neuroscience InstituteUniversity of SheffieldSheffieldUK
| |
Collapse
|
23
|
Wong-Campos JD, Park P, Davis H, Qi Y, Tian H, Itkis DG, Kim D, Grimm JB, Plutkis SE, Lavis L, Cohen AE. Voltage dynamics of dendritic integration and back-propagation in vivo. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.05.25.542363. [PMID: 37292691 PMCID: PMC10245993 DOI: 10.1101/2023.05.25.542363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Neurons integrate synaptic inputs within their dendrites and produce spiking outputs, which then propagate down the axon and back into the dendrites where they contribute to plasticity. Mapping the voltage dynamics in dendritic arbors of live animals is crucial for understanding neuronal computation and plasticity rules. Here we combine patterned channelrhodopsin activation with dual-plane structured illumination voltage imaging, for simultaneous perturbation and monitoring of dendritic and somatic voltage in Layer 2/3 pyramidal neurons in anesthetized and awake mice. We examined the integration of synaptic inputs and compared the dynamics of optogenetically evoked, spontaneous, and sensory-evoked back-propagating action potentials (bAPs). Our measurements revealed a broadly shared membrane voltage throughout the dendritic arbor, and few signatures of electrical compartmentalization among synaptic inputs. However, we observed spike rate acceleration-dependent propagation of bAPs into distal dendrites. We propose that this dendritic filtering of bAPs may play a critical role in activity-dependent plasticity.
Collapse
Affiliation(s)
- J David Wong-Campos
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Pojeong Park
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Hunter Davis
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Yitong Qi
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - He Tian
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Daniel G Itkis
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Doyeon Kim
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
| | - Jonathan B Grimm
- Janelia Research Campus, Howard Hughes Medical Institute, Ashburn, Virginia, USA
| | - Sarah E Plutkis
- Janelia Research Campus, Howard Hughes Medical Institute, Ashburn, Virginia, USA
| | - Luke Lavis
- Janelia Research Campus, Howard Hughes Medical Institute, Ashburn, Virginia, USA
| | - Adam E Cohen
- Department of Chemistry and Chemical Biology, Harvard University, Cambridge, MA, USA
- Department of Physics, Harvard University, Cambridge, MA, USA
| |
Collapse
|
24
|
Aceituno PV, Farinha MT, Loidl R, Grewe BF. Learning cortical hierarchies with temporal Hebbian updates. Front Comput Neurosci 2023; 17:1136010. [PMID: 37293353 PMCID: PMC10244748 DOI: 10.3389/fncom.2023.1136010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2023] [Accepted: 04/25/2023] [Indexed: 06/10/2023] Open
Abstract
A key driver of mammalian intelligence is the ability to represent incoming sensory information across multiple abstraction levels. For example, in the visual ventral stream, incoming signals are first represented as low-level edge filters and then transformed into high-level object representations. Similar hierarchical structures routinely emerge in artificial neural networks (ANNs) trained for object recognition tasks, suggesting that similar structures may underlie biological neural networks. However, the classical ANN training algorithm, backpropagation, is considered biologically implausible, and thus alternative biologically plausible training methods have been developed such as Equilibrium Propagation, Deep Feedback Control, Supervised Predictive Coding, and Dendritic Error Backpropagation. Several of those models propose that local errors are calculated for each neuron by comparing apical and somatic activities. Notwithstanding, from a neuroscience perspective, it is not clear how a neuron could compare compartmental signals. Here, we propose a solution to this problem in that we let the apical feedback signal change the postsynaptic firing rate and combine this with a differential Hebbian update, a rate-based version of classical spiking time-dependent plasticity (STDP). We prove that weight updates of this form minimize two alternative loss functions that we prove to be equivalent to the error-based losses used in machine learning: the inference latency and the amount of top-down feedback necessary. Moreover, we show that the use of differential Hebbian updates works similarly well in other feedback-based deep learning frameworks such as Predictive Coding or Equilibrium Propagation. Finally, our work removes a key requirement of biologically plausible models for deep learning and proposes a learning mechanism that would explain how temporal Hebbian learning rules can implement supervised hierarchical learning.
Collapse
Affiliation(s)
- Pau Vilimelis Aceituno
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
- ETH AI Center, ETH Zurich, Zurich, Switzerland
| | | | - Reinhard Loidl
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Benjamin F. Grewe
- Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
- ETH AI Center, ETH Zurich, Zurich, Switzerland
| |
Collapse
|
25
|
Malakasis N, Chavlis S, Poirazi P. Synaptic turnover promotes efficient learning in bio-realistic spiking neural networks. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.05.22.541722. [PMID: 37292929 PMCID: PMC10245885 DOI: 10.1101/2023.05.22.541722] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
While artificial machine learning systems achieve superhuman performance in specific tasks such as language processing, image and video recognition, they do so use extremely large datasets and huge amounts of power. On the other hand, the brain remains superior in several cognitively challenging tasks while operating with the energy of a small lightbulb. We use a biologically constrained spiking neural network model to explore how the neural tissue achieves such high efficiency and assess its learning capacity on discrimination tasks. We found that synaptic turnover, a form of structural plasticity, which is the ability of the brain to form and eliminate synapses continuously, increases both the speed and the performance of our network on all tasks tested. Moreover, it allows accurate learning using a smaller number of examples. Importantly, these improvements are most significant under conditions of resource scarcity, such as when the number of trainable parameters is halved and when the task difficulty is increased. Our findings provide new insights into the mechanisms that underlie efficient learning in the brain and can inspire the development of more efficient and flexible machine learning algorithms.
Collapse
Affiliation(s)
- Nikos Malakasis
- School of Medicine, University of Crete, Heraklion 70013, Greece
- Institute of Molecular Biology and Biotechnology, Foundation for Research and Technology-Hellas, Heraklion 70013, Greece
| | - Spyridon Chavlis
- Institute of Molecular Biology and Biotechnology, Foundation for Research and Technology-Hellas, Heraklion 70013, Greece
| | - Panayiota Poirazi
- Institute of Molecular Biology and Biotechnology, Foundation for Research and Technology-Hellas, Heraklion 70013, Greece
| |
Collapse
|
26
|
Gillon CJ, Lecoq JA, Pina JE, Ahmed R, Billeh YN, Caldejon S, Groblewski P, Henley TM, Kato I, Lee E, Luviano J, Mace K, Nayan C, Nguyen TV, North K, Perkins J, Seid S, Valley MT, Williford A, Bengio Y, Lillicrap TP, Zylberberg J, Richards BA. Responses of pyramidal cell somata and apical dendrites in mouse visual cortex over multiple days. Sci Data 2023; 10:287. [PMID: 37198203 DOI: 10.1038/s41597-023-02214-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Accepted: 05/05/2023] [Indexed: 05/19/2023] Open
Abstract
The apical dendrites of pyramidal neurons in sensory cortex receive primarily top-down signals from associative and motor regions, while cell bodies and nearby dendrites are heavily targeted by locally recurrent or bottom-up inputs from the sensory periphery. Based on these differences, a number of theories in computational neuroscience postulate a unique role for apical dendrites in learning. However, due to technical challenges in data collection, little data is available for comparing the responses of apical dendrites to cell bodies over multiple days. Here we present a dataset collected through the Allen Institute Mindscope's OpenScope program that addresses this need. This dataset comprises high-quality two-photon calcium imaging from the apical dendrites and the cell bodies of visual cortical pyramidal neurons, acquired over multiple days in awake, behaving mice that were presented with visual stimuli. Many of the cell bodies and dendrite segments were tracked over days, enabling analyses of how their responses change over time. This dataset allows neuroscientists to explore the differences between apical and somatic processing and plasticity.
Collapse
Affiliation(s)
- Colleen J Gillon
- Department of Cell & Systems Biology, University of Toronto, Toronto, Ontario, Canada
- Department of Biological Sciences, University of Toronto Scarborough, Toronto, Ontario, Canada
- Mila, Montréal, Québec, Canada
| | | | - Jason E Pina
- Department of Physics and Astronomy, York University, Toronto, Ontario, Canada
- Centre for Vision Research, York University, Toronto, Ontario, Canada
| | - Ruweida Ahmed
- Allen Institute, MindScope Program, Seattle, WA, USA
| | | | | | | | - Timothy M Henley
- Department of Physics and Astronomy, York University, Toronto, Ontario, Canada
- Centre for Vision Research, York University, Toronto, Ontario, Canada
| | - India Kato
- Allen Institute, MindScope Program, Seattle, WA, USA
| | - Eric Lee
- Allen Institute, MindScope Program, Seattle, WA, USA
| | | | - Kyla Mace
- Allen Institute, MindScope Program, Seattle, WA, USA
| | - Chelsea Nayan
- Allen Institute, MindScope Program, Seattle, WA, USA
| | | | - Kat North
- Allen Institute, MindScope Program, Seattle, WA, USA
| | - Jed Perkins
- Allen Institute, MindScope Program, Seattle, WA, USA
| | - Sam Seid
- Allen Institute, MindScope Program, Seattle, WA, USA
| | | | - Ali Williford
- Allen Institute, MindScope Program, Seattle, WA, USA
| | - Yoshua Bengio
- Mila, Montréal, Québec, Canada
- Département d'informatique et de recherche opérationnelle, Université de Montréal, Montréal, Québec, Canada
- Learning in Machines and Brains Program, Canadian Institute for Advanced Research, Toronto, Ontario, Canada
| | - Timothy P Lillicrap
- DeepMind, Inc, London, UK
- Centre for Computation, Mathematics and Physics in the Life Sciences and Experimental Biology, University College London, London, UK
| | - Joel Zylberberg
- Department of Physics and Astronomy, York University, Toronto, Ontario, Canada.
- Centre for Vision Research, York University, Toronto, Ontario, Canada.
- Learning in Machines and Brains Program, Canadian Institute for Advanced Research, Toronto, Ontario, Canada.
- Vector Institute for Artificial Intelligence, Toronto, Ontario, Canada.
| | - Blake A Richards
- Mila, Montréal, Québec, Canada.
- Learning in Machines and Brains Program, Canadian Institute for Advanced Research, Toronto, Ontario, Canada.
- School of Computer Science, McGill University, Montréal, Québec, Canada.
- Department of Neurology & Neurosurgery, McGill University, Montréal, Québec, Canada.
- Montreal Neurological Institute, McGill University, Montréal, Québec, Canada.
| |
Collapse
|
27
|
Schneider A, Azabou M, McDougall-Vigier L, Parks DF, Ensley S, Bhaskaran-Nair K, Nowakowski T, Dyer EL, Hengen KB. Transcriptomic cell type structures in vivo neuronal activity across multiple timescales. Cell Rep 2023; 42:112318. [PMID: 36995938 PMCID: PMC10539488 DOI: 10.1016/j.celrep.2023.112318] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2022] [Revised: 02/04/2023] [Accepted: 03/15/2023] [Indexed: 03/30/2023] Open
Abstract
Cell type is hypothesized to be a key determinant of a neuron's role within a circuit. Here, we examine whether a neuron's transcriptomic type influences the timing of its activity. We develop a deep-learning architecture that learns features of interevent intervals across timescales (ms to >30 min). We show that transcriptomic cell-class information is embedded in the timing of single neuron activity in the intact brain of behaving animals (calcium imaging and extracellular electrophysiology) as well as in a bio-realistic model of the visual cortex. Further, a subset of excitatory cell types are distinguishable but can be classified with higher accuracy when considering cortical layer and projection class. Finally, we show that computational fingerprints of cell types may be universalizable across structured stimuli and naturalistic movies. Our results indicate that transcriptomic class and type may be imprinted in the timing of single neuron activity across diverse stimuli.
Collapse
Affiliation(s)
- Aidan Schneider
- Department of Biology, Washington University in St. Louis, St. Louis, MO 63130, USA
| | - Mehdi Azabou
- School of Electrical & Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
| | | | - David F Parks
- Department of Biomolecular Engineering, University of California, Santa Cruz, Santa Cruz, CA 95064, USA
| | - Sahara Ensley
- Department of Biology, Washington University in St. Louis, St. Louis, MO 63130, USA
| | - Kiran Bhaskaran-Nair
- Department of Biology, Washington University in St. Louis, St. Louis, MO 63130, USA
| | - Tomasz Nowakowski
- Department of Anatomy, University of California, San Francisco, San Francisco, CA 94158, USA
| | - Eva L Dyer
- School of Electrical & Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA; Coulter Department of Biomedical Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
| | - Keith B Hengen
- Department of Biology, Washington University in St. Louis, St. Louis, MO 63130, USA.
| |
Collapse
|
28
|
Meir Y, Tevet O, Tzach Y, Hodassman S, Gross RD, Kanter I. Efficient shallow learning as an alternative to deep learning. Sci Rep 2023; 13:5423. [PMID: 37080998 PMCID: PMC10119101 DOI: 10.1038/s41598-023-32559-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Accepted: 03/29/2023] [Indexed: 04/22/2023] Open
Abstract
The realization of complex classification tasks requires training of deep learning (DL) architectures consisting of tens or even hundreds of convolutional and fully connected hidden layers, which is far from the reality of the human brain. According to the DL rationale, the first convolutional layer reveals localized patterns in the input and large-scale patterns in the following layers, until it reliably characterizes a class of inputs. Here, we demonstrate that with a fixed ratio between the depths of the first and second convolutional layers, the error rates of the generalized shallow LeNet architecture, consisting of only five layers, decay as a power law with the number of filters in the first convolutional layer. The extrapolation of this power law indicates that the generalized LeNet can achieve small error rates that were previously obtained for the CIFAR-10 database using DL architectures. A power law with a similar exponent also characterizes the generalized VGG-16 architecture. However, this results in a significantly increased number of operations required to achieve a given error rate with respect to LeNet. This power law phenomenon governs various generalized LeNet and VGG-16 architectures, hinting at its universal behavior and suggesting a quantitative hierarchical time-space complexity among machine learning architectures. Additionally, the conservation law along the convolutional layers, which is the square-root of their size times their depth, is found to asymptotically minimize error rates. The efficient shallow learning that is demonstrated in this study calls for further quantitative examination using various databases and architectures and its accelerated implementation using future dedicated hardware developments.
Collapse
Affiliation(s)
- Yuval Meir
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel
| | - Ofek Tevet
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel
| | - Yarden Tzach
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel
| | - Shiri Hodassman
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel
| | - Ronit D Gross
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel
| | - Ido Kanter
- Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel.
- Gonda Interdisciplinary Brain Research Center, Bar-Ilan University, 52900, Ramat-Gan, Israel.
| |
Collapse
|
29
|
Ritvo VJH, Nguyen A, Turk-Browne NB, Norman KA. Differentiation and Integration of Competing Memories: A Neural Network Model. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.04.02.535239. [PMID: 37066178 PMCID: PMC10103961 DOI: 10.1101/2023.04.02.535239] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
What determines when neural representations of memories move together (integrate) or apart (differentiate)? Classic supervised learning models posit that, when two stimuli predict similar outcomes, their representations should integrate. However, these models have recently been challenged by studies showing that pairing two stimuli with a shared associate can sometimes cause differentiation, depending on the parameters of the study and the brain region being examined. Here, we provide a purely unsupervised neural network model that can explain these and other related findings. The model can exhibit integration or differentiation depending on the amount of activity allowed to spread to competitors - inactive memories are not modified, connections to moderately active competitors are weakened (leading to differentiation), and connections to highly active competitors are strengthened (leading to integration). The model also makes several novel predictions - most importantly, that differentiation will be rapid and asymmetric. Overall, these modeling results provide a computational explanation for a diverse set of seemingly contradictory empirical findings in the memory literature, as well as new insights into the dynamics at play during learning.
Collapse
Affiliation(s)
| | - Alex Nguyen
- Princeton Neuroscience Institute, Princeton University
| | | | - Kenneth A. Norman
- Department of Psychology, Princeton University
- Princeton Neuroscience Institute, Princeton University
| |
Collapse
|
30
|
Shervani-Tabar N, Rosenbaum R. Meta-learning biologically plausible plasticity rules with random feedback pathways. Nat Commun 2023; 14:1805. [PMID: 37002222 PMCID: PMC10066328 DOI: 10.1038/s41467-023-37562-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Accepted: 03/21/2023] [Indexed: 04/04/2023] Open
Abstract
Backpropagation is widely used to train artificial neural networks, but its relationship to synaptic plasticity in the brain is unknown. Some biological models of backpropagation rely on feedback projections that are symmetric with feedforward connections, but experiments do not corroborate the existence of such symmetric backward connectivity. Random feedback alignment offers an alternative model in which errors are propagated backward through fixed, random backward connections. This approach successfully trains shallow models, but learns slowly and does not perform well with deeper models or online learning. In this study, we develop a meta-learning approach to discover interpretable, biologically plausible plasticity rules that improve online learning performance with fixed random feedback connections. The resulting plasticity rules show improved online training of deep models in the low data regime. Our results highlight the potential of meta-learning to discover effective, interpretable learning rules satisfying biological constraints.
Collapse
Affiliation(s)
- Navid Shervani-Tabar
- Department of Applied and Computational Mathematics and Statistics, University of Notre Dame, Notre Dame, IN, 46556, USA.
| | - Robert Rosenbaum
- Department of Applied and Computational Mathematics and Statistics, University of Notre Dame, Notre Dame, IN, 46556, USA
| |
Collapse
|
31
|
Mastrogiuseppe F, Hiratani N, Latham P. Evolution of neural activity in circuits bridging sensory and abstract knowledge. eLife 2023; 12:79908. [PMID: 36881019 PMCID: PMC9991064 DOI: 10.7554/elife.79908] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2022] [Accepted: 01/06/2023] [Indexed: 03/08/2023] Open
Abstract
The ability to associate sensory stimuli with abstract classes is critical for survival. How are these associations implemented in brain circuits? And what governs how neural activity evolves during abstract knowledge acquisition? To investigate these questions, we consider a circuit model that learns to map sensory input to abstract classes via gradient-descent synaptic plasticity. We focus on typical neuroscience tasks (simple, and context-dependent, categorization), and study how both synaptic connectivity and neural activity evolve during learning. To make contact with the current generation of experiments, we analyze activity via standard measures such as selectivity, correlations, and tuning symmetry. We find that the model is able to recapitulate experimental observations, including seemingly disparate ones. We determine how, in the model, the behaviour of these measures depends on details of the circuit and the task. These dependencies make experimentally testable predictions about the circuitry supporting abstract knowledge acquisition in the brain.
Collapse
Affiliation(s)
| | - Naoki Hiratani
- Center for Brain Science, Harvard UniversityHarvardUnited States
| | - Peter Latham
- Gatsby Computational Neuroscience Unit, University College LondonLondonUnited Kingdom
| |
Collapse
|
32
|
Tamura K, Yamamoto Y, Kobayashi T, Kuriyama R, Yamazaki T. Discrimination and learning of temporal input sequences in a cerebellar Purkinje cell model. Front Cell Neurosci 2023; 17:1075005. [PMID: 36816857 PMCID: PMC9932327 DOI: 10.3389/fncel.2023.1075005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 01/10/2023] [Indexed: 02/05/2023] Open
Abstract
Introduction Temporal information processing is essential for sequential contraction of various muscles with the appropriate timing and amplitude for fast and smooth motor control. These functions depend on dynamics of neural circuits, which consist of simple neurons that accumulate incoming spikes and emit other spikes. However, recent studies indicate that individual neurons can perform complex information processing through the nonlinear dynamics of dendrites with complex shapes and ion channels. Although we have extensive evidence that cerebellar circuits play a vital role in motor control, studies investigating the computational ability of single Purkinje cells are few. Methods We found, through computer simulations, that a Purkinje cell can discriminate a series of pulses in two directions (from dendrite tip to soma, and from soma to dendrite), as cortical pyramidal cells do. Such direction sensitivity was observed in whatever compartment types of dendrites (spiny, smooth, and main), although they have dierent sets of ion channels. Results We found that the shortest and longest discriminable sequences lasted for 60 ms (6 pulses with 10 ms interval) and 4,000 ms (20 pulses with 200 ms interval), respectively. and that the ratio of discriminable sequences within the region of the interesting parameter space was, on average, 3.3% (spiny), 3.2% (smooth), and 1.0% (main). For the direction sensitivity, a T-type Ca2+ channel was necessary, in contrast with cortical pyramidal cells that have N-methyl-D-aspartate receptors (NMDARs). Furthermore, we tested whether the stimulus direction can be reversed by learning, specifically by simulated long-term depression, and obtained positive results. Discussion Our results show that individual Purkinje cells can perform more complex information processing than is conventionally assumed for a single neuron, and suggest that Purkinje cells act as sequence discriminators, a useful role in motor control and learning.
Collapse
Affiliation(s)
- Kaaya Tamura
- Graduate School of Informatics and Engineering, The University of Electro-Communications, Tokyo, Japan
| | - Yuki Yamamoto
- Graduate School of Medical and Dental Sciences, Tokyo Medical and Dental University, Tokyo, Japan
| | - Taira Kobayashi
- Graduate School of Informatics and Engineering, The University of Electro-Communications, Tokyo, Japan,Graduate School of Sciences and Technology for Innovation, Yamaguchi University, Yamaguchi, Japan
| | - Rin Kuriyama
- Graduate School of Informatics and Engineering, The University of Electro-Communications, Tokyo, Japan
| | - Tadashi Yamazaki
- Graduate School of Informatics and Engineering, The University of Electro-Communications, Tokyo, Japan,*Correspondence: Tadashi Yamazaki ✉
| |
Collapse
|
33
|
Aru J, Drüke M, Pikamäe J, Larkum ME. Mental navigation and the neural mechanisms of insight. Trends Neurosci 2023; 46:100-109. [PMID: 36462993 DOI: 10.1016/j.tins.2022.11.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2022] [Revised: 11/01/2022] [Accepted: 11/07/2022] [Indexed: 12/03/2022]
Abstract
How do new ideas come about? The central hypothesis presented here states that insights might happen during mental navigation and correspond to rapid plasticity at the cellular level. We highlight the differences between neocortical and hippocampal mechanisms of insight. We argue that the suddenness of insight can be related to the sudden emergence of place fields in the hippocampus. According to our hypothesis, insights are supported by a state of mind-wandering that can be tied to the process of combining knowledge pieces during sharp-wave ripples (SWRs). Our framework connects the dots between research on creativity, mental navigation, and specific synaptic plasticity mechanisms in the hippocampus.
Collapse
Affiliation(s)
- Jaan Aru
- Institute of Computer Science, University of Tartu, Tartu, Estonia.
| | - Moritz Drüke
- Institute of Biology, Humboldt University Berlin, Berlin, Germany
| | - Juhan Pikamäe
- Institute of Molecular and Cell Biology, University of Tartu, Tartu, Estonia
| | - Matthew E Larkum
- Institute of Biology, Humboldt University Berlin, Berlin, Germany; Neurocure Center for Excellence, Charité Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
34
|
McFarlan AR, Chou CYC, Watanabe A, Cherepacha N, Haddad M, Owens H, Sjöström PJ. The plasticitome of cortical interneurons. Nat Rev Neurosci 2023; 24:80-97. [PMID: 36585520 DOI: 10.1038/s41583-022-00663-9] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/21/2022] [Indexed: 12/31/2022]
Abstract
Hebb postulated that, to store information in the brain, assemblies of excitatory neurons coding for a percept are bound together via associative long-term synaptic plasticity. In this view, it is unclear what role, if any, is carried out by inhibitory interneurons. Indeed, some have argued that inhibitory interneurons are not plastic. Yet numerous recent studies have demonstrated that, similar to excitatory neurons, inhibitory interneurons also undergo long-term plasticity. Here, we discuss the many diverse forms of long-term plasticity that are found at inputs to and outputs from several types of cortical inhibitory interneuron, including their plasticity of intrinsic excitability and their homeostatic plasticity. We explain key plasticity terminology, highlight key interneuron plasticity mechanisms, extract overarching principles and point out implications for healthy brain functionality as well as for neuropathology. We introduce the concept of the plasticitome - the synaptic plasticity counterpart to the genome or the connectome - as well as nomenclature and definitions for dealing with this rich diversity of plasticity. We argue that the great diversity of interneuron plasticity rules is best understood at the circuit level, for example as a way of elucidating how the credit-assignment problem is solved in deep biological neural networks.
Collapse
Affiliation(s)
- Amanda R McFarlan
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.,Integrated Program in Neuroscience, McGill University, Montréal, Québec, Canada
| | - Christina Y C Chou
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.,Integrated Program in Neuroscience, McGill University, Montréal, Québec, Canada
| | - Airi Watanabe
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.,Integrated Program in Neuroscience, McGill University, Montréal, Québec, Canada
| | - Nicole Cherepacha
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada
| | - Maria Haddad
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.,Integrated Program in Neuroscience, McGill University, Montréal, Québec, Canada
| | - Hannah Owens
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.,Integrated Program in Neuroscience, McGill University, Montréal, Québec, Canada
| | - P Jesper Sjöström
- Centre for Research in Neuroscience, Department of Medicine, The Research Institute of the McGill University Health Centre, Montréal, Québec, Canada.
| |
Collapse
|
35
|
Learning on tree architectures outperforms a convolutional feedforward network. Sci Rep 2023; 13:962. [PMID: 36717568 PMCID: PMC9886946 DOI: 10.1038/s41598-023-27986-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Accepted: 01/11/2023] [Indexed: 02/01/2023] Open
Abstract
Advanced deep learning architectures consist of tens of fully connected and convolutional hidden layers, currently extended to hundreds, are far from their biological realization. Their implausible biological dynamics relies on changing a weight in a non-local manner, as the number of routes between an output unit and a weight is typically large, using the backpropagation technique. Here, a 3-layer tree architecture inspired by experimental-based dendritic tree adaptations is developed and applied to the offline and online learning of the CIFAR-10 database. The proposed architecture outperforms the achievable success rates of the 5-layer convolutional LeNet. Moreover, the highly pruned tree backpropagation approach of the proposed architecture, where a single route connects an output unit and a weight, represents an efficient dendritic deep learning.
Collapse
|
36
|
Pagkalos M, Chavlis S, Poirazi P. Introducing the Dendrify framework for incorporating dendrites to spiking neural networks. Nat Commun 2023; 14:131. [PMID: 36627284 PMCID: PMC9832130 DOI: 10.1038/s41467-022-35747-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2022] [Accepted: 12/22/2022] [Indexed: 01/11/2023] Open
Abstract
Computational modeling has been indispensable for understanding how subcellular neuronal features influence circuit processing. However, the role of dendritic computations in network-level operations remains largely unexplored. This is partly because existing tools do not allow the development of realistic and efficient network models that account for dendrites. Current spiking neural networks, although efficient, are usually quite simplistic, overlooking essential dendritic properties. Conversely, circuit models with morphologically detailed neuron models are computationally costly, thus impractical for large-network simulations. To bridge the gap between these two extremes and facilitate the adoption of dendritic features in spiking neural networks, we introduce Dendrify, an open-source Python package based on Brian 2. Dendrify, through simple commands, automatically generates reduced compartmental neuron models with simplified yet biologically relevant dendritic and synaptic integrative properties. Such models strike a good balance between flexibility, performance, and biological accuracy, allowing us to explore dendritic contributions to network-level functions while paving the way for developing more powerful neuromorphic systems.
Collapse
Affiliation(s)
- Michalis Pagkalos
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece
- Department of Biology, University of Crete, Heraklion, 70013, Greece
| | - Spyridon Chavlis
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece
| | - Panayiota Poirazi
- Institute of Molecular Biology and Biotechnology (IMBB), Foundation for Research and Technology Hellas (FORTH), Heraklion, 70013, Greece.
| |
Collapse
|
37
|
Boven E, Pemberton J, Chadderton P, Apps R, Costa RP. Cerebro-cerebellar networks facilitate learning through feedback decoupling. Nat Commun 2023; 14:51. [PMID: 36599827 DOI: 10.1038/s41467-022-35658-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Accepted: 12/15/2022] [Indexed: 01/06/2023] Open
Abstract
Behavioural feedback is critical for learning in the cerebral cortex. However, such feedback is often not readily available. How the cerebral cortex learns efficiently despite the sparse nature of feedback remains unclear. Inspired by recent deep learning algorithms, we introduce a systems-level computational model of cerebro-cerebellar interactions. In this model a cerebral recurrent network receives feedback predictions from a cerebellar network, thereby decoupling learning in cerebral networks from future feedback. When trained in a simple sensorimotor task the model shows faster learning and reduced dysmetria-like behaviours, in line with the widely observed functional impact of the cerebellum. Next, we demonstrate that these results generalise to more complex motor and cognitive tasks. Finally, the model makes several experimentally testable predictions regarding cerebro-cerebellar task-specific representations over learning, task-specific benefits of cerebellar predictions and the differential impact of cerebellar and inferior olive lesions. Overall, our work offers a theoretical framework of cerebro-cerebellar networks as feedback decoupling machines.
Collapse
Affiliation(s)
- Ellen Boven
- Bristol Computational Neuroscience Unit, Intelligent Systems Labs, SCEEM, Faculty of Engineering, University of Bristol, Bristol, BS8 1TH, UK
- School of Physiology, Pharmacology and Neuroscience, Faculty of Life Sciences, University of Bristol, Bristol, BS8 1TH, UK
| | - Joseph Pemberton
- Bristol Computational Neuroscience Unit, Intelligent Systems Labs, SCEEM, Faculty of Engineering, University of Bristol, Bristol, BS8 1TH, UK
| | - Paul Chadderton
- School of Physiology, Pharmacology and Neuroscience, Faculty of Life Sciences, University of Bristol, Bristol, BS8 1TH, UK
| | - Richard Apps
- School of Physiology, Pharmacology and Neuroscience, Faculty of Life Sciences, University of Bristol, Bristol, BS8 1TH, UK
| | - Rui Ponte Costa
- Bristol Computational Neuroscience Unit, Intelligent Systems Labs, SCEEM, Faculty of Engineering, University of Bristol, Bristol, BS8 1TH, UK.
| |
Collapse
|
38
|
Passos LA, Papa JP, Hussain A, Adeel A. Canonical Cortical Graph Neural Networks and its Application for Speech Enhancement in Audio-Visual Hearing Aids. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.11.081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
39
|
Sawant Y, Kundu JN, Radhakrishnan VB, Sridharan D. A Midbrain Inspired Recurrent Neural Network Model for Robust Change Detection. J Neurosci 2022; 42:8262-8283. [PMID: 36123120 PMCID: PMC9653281 DOI: 10.1523/jneurosci.0164-22.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 07/26/2022] [Accepted: 07/30/2022] [Indexed: 11/21/2022] Open
Abstract
We present a biologically inspired recurrent neural network (RNN) that efficiently detects changes in natural images. The model features sparse, topographic connectivity (st-RNN), closely modeled on the circuit architecture of a "midbrain attention network." We deployed the st-RNN in a challenging change blindness task, in which changes must be detected in a discontinuous sequence of images. Compared with a conventional RNN, the st-RNN learned 9x faster and achieved state-of-the-art performance with 15x fewer connections. An analysis of low-dimensional dynamics revealed putative circuit mechanisms, including a critical role for a global inhibitory (GI) motif, for successful change detection. The model reproduced key experimental phenomena, including midbrain neurons' sensitivity to dynamic stimuli, neural signatures of stimulus competition, as well as hallmark behavioral effects of midbrain microstimulation. Finally, the model accurately predicted human gaze fixations in a change blindness experiment, surpassing state-of-the-art saliency-based methods. The st-RNN provides a novel deep learning model for linking neural computations underlying change detection with psychophysical mechanisms.SIGNIFICANCE STATEMENT For adaptive survival, our brains must be able to accurately and rapidly detect changing aspects of our visual world. We present a novel deep learning model, a sparse, topographic recurrent neural network (st-RNN), that mimics the neuroanatomy of an evolutionarily conserved "midbrain attention network." The st-RNN achieved robust change detection in challenging change blindness tasks, outperforming conventional RNN architectures. The model also reproduced hallmark experimental phenomena, both neural and behavioral, reported in seminal midbrain studies. Lastly, the st-RNN outperformed state-of-the-art models at predicting human gaze fixations in a laboratory change blindness experiment. Our deep learning model may provide important clues about key mechanisms by which the brain efficiently detects changes.
Collapse
Affiliation(s)
- Yash Sawant
- Centre for Neuroscience, Indian Institute of Science, Bangalore 560012, India
| | - Jogendra Nath Kundu
- Department of Computational and Data Sciences, Indian Institute of Science, Bangalore 560012, India
| | | | - Devarajan Sridharan
- Centre for Neuroscience, Indian Institute of Science, Bangalore 560012, India
- Department of Computer Science and Automation, Indian Institute of Science, Bangalore 560012, India
| |
Collapse
|
40
|
Yu C, Du Y, Chen M, Wang A, Wang G, Li E. MAP-SNN: Mapping spike activities with multiplicity, adaptability, and plasticity into bio-plausible spiking neural networks. Front Neurosci 2022; 16:945037. [PMID: 36203801 PMCID: PMC9531034 DOI: 10.3389/fnins.2022.945037] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Accepted: 08/29/2022] [Indexed: 11/26/2022] Open
Abstract
Spiking Neural Networks (SNNs) are considered more biologically realistic and power-efficient as they imitate the fundamental mechanism of the human brain. Backpropagation (BP) based SNN learning algorithms that utilize deep learning frameworks have achieved good performance. However, those BP-based algorithms partially ignore bio-interpretability. In modeling spike activity for biological plausible BP-based SNNs, we examine three properties: multiplicity, adaptability, and plasticity (MAP). Regarding multiplicity, we propose a Multiple-Spike Pattern (MSP) with multiple-spike transmission to improve model robustness in discrete time iterations. To realize adaptability, we adopt Spike Frequency Adaption (SFA) under MSP to reduce spike activities for enhanced efficiency. For plasticity, we propose a trainable state-free synapse that models spike response current to increase the diversity of spiking neurons for temporal feature extraction. The proposed SNN model achieves competitive performances on the N-MNIST and SHD neuromorphic datasets. In addition, experimental results demonstrate that the proposed three aspects are significant to iterative robustness, spike efficiency, and the capacity to extract spikes' temporal features. In summary, this study presents a realistic approach for bio-inspired spike activity with MAP, presenting a novel neuromorphic perspective for incorporating biological properties into spiking neural networks.
Collapse
Affiliation(s)
- Chengting Yu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
- Zhejiang University - University of Illinois at Urbana-Champaign Institute, Zhejiang University, Haining, China
| | - Yangkai Du
- College of Computer Science and Technology, Zhejiang University, Hangzhou, China
| | - Mufeng Chen
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Aili Wang
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
- Zhejiang University - University of Illinois at Urbana-Champaign Institute, Zhejiang University, Haining, China
- *Correspondence: Aili Wang
| | - Gaoang Wang
- Zhejiang University - University of Illinois at Urbana-Champaign Institute, Zhejiang University, Haining, China
| | - Erping Li
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
- Zhejiang University - University of Illinois at Urbana-Champaign Institute, Zhejiang University, Haining, China
| |
Collapse
|
41
|
Small, correlated changes in synaptic connectivity may facilitate rapid motor learning. Nat Commun 2022; 13:5163. [PMID: 36056006 PMCID: PMC9440011 DOI: 10.1038/s41467-022-32646-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2021] [Accepted: 08/08/2022] [Indexed: 11/08/2022] Open
Abstract
Animals rapidly adapt their movements to external perturbations, a process paralleled by changes in neural activity in the motor cortex. Experimental studies suggest that these changes originate from altered inputs (Hinput) rather than from changes in local connectivity (Hlocal), as neural covariance is largely preserved during adaptation. Since measuring synaptic changes in vivo remains very challenging, we used a modular recurrent neural network to qualitatively test this interpretation. As expected, Hinput resulted in small activity changes and largely preserved covariance. Surprisingly given the presumed dependence of stable covariance on preserved circuit connectivity, Hlocal led to only slightly larger changes in activity and covariance, still within the range of experimental recordings. This similarity is due to Hlocal only requiring small, correlated connectivity changes for successful adaptation. Simulations of tasks that impose increasingly larger behavioural changes revealed a growing difference between Hinput and Hlocal, which could be exploited when designing future experiments.
Collapse
|
42
|
Mercier MS, Magloire V, Cornford JH, Kullmann DM. Long-term potentiation in neurogliaform interneurons modulates excitation-inhibition balance in the temporoammonic pathway. J Physiol 2022; 600:4001-4017. [PMID: 35876215 PMCID: PMC9540908 DOI: 10.1113/jp282753] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2021] [Accepted: 07/19/2022] [Indexed: 11/08/2022] Open
Abstract
Apical dendrites of pyramidal neurons integrate information from higher-order cortex and thalamus, and gate signalling and plasticity at proximal synapses. In the hippocampus, neurogliaform cells and other interneurons located within stratum lacunosum-moleculare (SLM) mediate powerful inhibition of CA1 pyramidal neuron distal dendrites. Is the recruitment of such inhibition itself subject to use-dependent plasticity, and if so, what induction rules apply? Here we show that interneurons in mouse SLM exhibit Hebbian NMDA receptor-dependent long-term potentiation (LTP). Such plasticity can be induced by selective optogenetic stimulation of afferents in the temporoammonic pathway from the entorhinal cortex (EC), but not by equivalent stimulation of afferents from the thalamic nucleus reuniens. We further show that theta-burst patterns of afferent firing induces LTP in neurogliaform interneurons identified using neuron-derived neurotrophic factor (Ndnf)-Cre mice. Theta-burst activity of EC afferents led to an increase in disynaptic feed-forward inhibition, but not monosynaptic excitation, of CA1 pyramidal neurons. Activity-dependent synaptic plasticity in SLM interneurons thus alters the excitation-inhibition balance at EC inputs to the apical dendrites of pyramidal neurons, implying a dynamic role for these interneurons in gating CA1 dendritic computations. KEY POINTS: Electrogenic phenomena in distal dendrites of principal neurons in the hippocampus have a major role in gating synaptic plasticity at afferent synapses on proximal dendrites. Apical dendrites also receive powerful feed-forward inhibition, mediated in large part by neurogliaform neurons. Here we show that theta-burst activity in afferents from the entorhinal cortex (EC) induces 'Hebbian' long-term potentiation (LTP) at excitatory synapses recruiting these GABAergic cells. LTP in interneurons innervating apical dendrites increases disynaptic inhibition of principal neurons, thus shifting the excitation-inhibition balance in the temporoammonic (TA) pathway in favour of inhibition, with implications for computations and learning rules in proximal dendrites.
Collapse
Affiliation(s)
- Marion S. Mercier
- UCL Queen Square Institute of NeurologyDepartment of Clinical and Experimental EpilepsyUniversity College LondonLondonUK
| | - Vincent Magloire
- UCL Queen Square Institute of NeurologyDepartment of Clinical and Experimental EpilepsyUniversity College LondonLondonUK
| | - Jonathan H. Cornford
- UCL Queen Square Institute of NeurologyDepartment of Clinical and Experimental EpilepsyUniversity College LondonLondonUK
| | - Dimitri M. Kullmann
- UCL Queen Square Institute of NeurologyDepartment of Clinical and Experimental EpilepsyUniversity College LondonLondonUK
| |
Collapse
|
43
|
Developmental depression-to-facilitation shift controls excitation-inhibition balance. Commun Biol 2022; 5:873. [PMID: 36008708 PMCID: PMC9411206 DOI: 10.1038/s42003-022-03801-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Accepted: 08/04/2022] [Indexed: 11/25/2022] Open
Abstract
Changes in the short-term dynamics of excitatory synapses over development have been observed throughout cortex, but their purpose and consequences remain unclear. Here, we propose that developmental changes in synaptic dynamics buffer the effect of slow inhibitory long-term plasticity, allowing for continuously stable neural activity. Using computational modeling we demonstrate that early in development excitatory short-term depression quickly stabilises neural activity, even in the face of strong, unbalanced excitation. We introduce a model of the commonly observed developmental shift from depression to facilitation and show that neural activity remains stable throughout development, while inhibitory synaptic plasticity slowly balances excitation, consistent with experimental observations. Our model predicts changes in the input responses from phasic to phasic-and-tonic and more precise spike timings. We also observe a gradual emergence of short-lasting memory traces governed by short-term plasticity development. We conclude that the developmental depression-to-facilitation shift may control excitation-inhibition balance throughout development with important functional consequences. Using computational modelling this study proposes that the commonly observed depression-to-facilitation shift across development controls excitation-inhibition balance in the brain.
Collapse
|
44
|
Vafidis P, Owald D, D'Albis T, Kempter R. Learning accurate path integration in ring attractor models of the head direction system. eLife 2022; 11:69841. [PMID: 35723252 PMCID: PMC9286743 DOI: 10.7554/elife.69841] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Accepted: 06/17/2022] [Indexed: 11/13/2022] Open
Abstract
Ring attractor models for angular path integration have received strong experimental support. To function as integrators, head direction circuits require precisely tuned connectivity, but it is currently unknown how such tuning could be achieved. Here, we propose a network model in which a local, biologically plausible learning rule adjusts synaptic efficacies during development, guided by supervisory allothetic cues. Applied to the Drosophila head direction system, the model learns to path-integrate accurately and develops a connectivity strikingly similar to the one reported in experiments. The mature network is a quasi-continuous attractor and reproduces key experiments in which optogenetic stimulation controls the internal representation of heading, and where the network remaps to integrate with different gains in rodents. Our model predicts that path integration requires self-supervised learning during a developmental phase, and proposes a general framework to learn to path-integrate with gain-1 even in architectures that lack the physical topography of a ring.
Collapse
Affiliation(s)
- Pantelis Vafidis
- Computation and Neural Systems, California Institute of Technology, Pasadena, United States
| | - David Owald
- NeuroCure, Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Tiziano D'Albis
- Department of Biology, Humboldt-Universität zu Berlin, Berlin, Germany
| | - Richard Kempter
- Department of Biology, Humboldt-Universität zu Berlin, Berlin, Germany
| |
Collapse
|
45
|
Tozzi A. Bipolar reasoning in feedback pathways. Biosystems 2022; 215-216:104652. [PMID: 35247481 DOI: 10.1016/j.biosystems.2022.104652] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2022] [Revised: 02/26/2022] [Accepted: 02/26/2022] [Indexed: 11/18/2022]
Abstract
Instead of the conventional 0 and 1 values, bipolar reasoning uses -1, 0, +1 to describe double-sided judgements in which neutral elements are halfway between positive and negative evaluations (e.g., "uncertain" lies between "impossible" and "totally sure"). We discuss the state-of-the-art in bipolar logics and recall two medieval forerunners, i.e., William of Ockham and Nicholas of Autrecourt, who embodied a bipolar mode of thought that is eminently modern. Starting from the trivial observation that "once a wheat sheaf is sealed and tied up, the packed down straws display the same orientation", we work up a new theory of the bipolar nature of networks, suggesting that orthodromic (i.e., feedforward, bottom-up) projections might be functionally coupled with antidromic (i.e., feedback, top-down) projections via the mathematical apparatus of presheaves/globular sets. When an entrained oscillation such as a neuronal spike propagates from A to B, changes in B might lead to changes in A, providing unexpected antidromic effects. Our account points towards the methodological feasibility of novel neural networks in which message feedback is guaranteed by backpropagation mechanisms endowed in the same feedforward circuits. Bottom-up/top-down transmission at various coarse-grained network levels provides fresh insights in far-flung scientific fields such as object persistence, memory reinforcement, visual recognition, Bayesian inferential circuits and multidimensional activity of the brain. Implying that axonal stimulation by external sources might backpropagate and modify neuronal electric oscillations, our theory also suggests testable previsions concerning the optimal location of transcranial magnetic stimulation's coils in patients affected by drug-resistant epilepsy.
Collapse
Affiliation(s)
- Arturo Tozzi
- Center for Nonlinear Science, Department of Physics, University of North Texas, Denton, TX, USA, 1155 Union Circle, #311427, Denton, TX, 76203-5017, USA.
| |
Collapse
|
46
|
Galloni AR, Ye Z, Rancz E. Dendritic Domain-Specific Sampling of Long-Range Axons Shapes Feedforward and Feedback Connectivity of L5 Neurons. J Neurosci 2022; 42:3394-3405. [PMID: 35241493 PMCID: PMC9034780 DOI: 10.1523/jneurosci.1620-21.2022] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2021] [Revised: 12/30/2021] [Accepted: 01/05/2022] [Indexed: 11/21/2022] Open
Abstract
Feedforward and feedback pathways interact in specific dendritic domains to enable cognitive functions such as predictive processing and learning. Based on axonal projections, hierarchically lower areas are thought to form synapses primarily on dendrites in middle cortical layers, whereas higher-order areas are thought to target dendrites in layer 1 and in deep layers. However, the extent to which functional synapses form in regions of axodendritic overlap has not been extensively studied. Here, we use viral tracing in the secondary visual cortex of male mice to map brain-wide inputs to thick-tufted layer 5 pyramidal neurons. Furthermore, we provide a comprehensive map of input locations through subcellular optogenetic circuit mapping. We show that input pathways target distinct dendritic domains with far greater specificity than appears from their axonal branching, often deviating substantially from the canonical patterns. Common assumptions regarding the dendrite-level interaction of feedforward and feedback inputs may thus need revisiting.SIGNIFICANCE STATEMENT Perception and learning depend on the ability of the brain to shape neuronal representations across all processing stages. Long-range connections across different hierarchical levels enable diverse sources of contextual information, such as predictions or motivational state, to modify feedforward signals. Assumptions regarding the organization of this hierarchical connectivity have not been extensively verified. Here, we assess the synaptic connectivity of brain-wide projections onto pyramidal neurons in the visual cortex of mice. Using trans-synaptic viral tracing and subcellular optogenetic circuit mapping, we show that functional synapses do not follow the consistent connectivity rule predicted by their axonal branching patterns. These findings highlight the diversity of computational strategies operating throughout cortical networks and may aid in building better artificial networks.
Collapse
Affiliation(s)
- Alessandro R Galloni
- The Francis Crick Institute, London NW1 1AT, United Kingdom
- University College London, London WC1E 6BT, United Kingdom
| | - Zhiwen Ye
- The Francis Crick Institute, London NW1 1AT, United Kingdom
| | - Ede Rancz
- The Francis Crick Institute, London NW1 1AT, United Kingdom
| |
Collapse
|
47
|
Prediction-error neurons in circuits with multiple neuron types: Formation, refinement, and functional implications. Proc Natl Acad Sci U S A 2022; 119:e2115699119. [PMID: 35320037 PMCID: PMC9060484 DOI: 10.1073/pnas.2115699119] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/14/2023] Open
Abstract
An influential idea in neuroscience is that neural circuits do not only passively process sensory information but rather actively compare them with predictions thereof. A core element of this comparison is prediction-error neurons, the activity of which only changes upon mismatches between actual and predicted sensory stimuli. While it has been shown that these prediction-error neurons come in different variants, it is largely unresolved how they are simultaneously formed and shaped by highly interconnected neural networks. By using a computational model, we study the circuit-level mechanisms that give rise to different variants of prediction-error neurons. Our results shed light on the formation, refinement, and robustness of prediction-error circuits, an important step toward a better understanding of predictive processing. Predictable sensory stimuli do not evoke significant responses in a subset of cortical excitatory neurons. Some of those neurons, however, change their activity upon mismatches between actual and predicted stimuli. Different variants of these prediction-error neurons exist, and they differ in their responses to unexpected sensory stimuli. However, it is unclear how these variants can develop and coexist in the same recurrent network and how they are simultaneously shaped by the astonishing diversity of inhibitory interneurons. Here, we study these questions in a computational network model with three types of inhibitory interneurons. We find that balancing excitation and inhibition in multiple pathways gives rise to heterogeneous prediction-error circuits. Dependent on the network’s initial connectivity and distribution of actual and predicted sensory inputs, these circuits can form different variants of prediction-error neurons that are robust to network perturbations and generalize to stimuli not seen during learning. These variants can be learned simultaneously via homeostatic inhibitory plasticity with low baseline firing rates. Finally, we demonstrate that prediction-error neurons can support biased perception, we illustrate a number of functional implications, and we discuss testable predictions.
Collapse
|
48
|
Intrinsic bursts facilitate learning of Lévy flight movements in recurrent neural network models. Sci Rep 2022; 12:4951. [PMID: 35322813 PMCID: PMC8943163 DOI: 10.1038/s41598-022-08953-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2021] [Accepted: 03/09/2022] [Indexed: 11/24/2022] Open
Abstract
Isolated spikes and bursts of spikes are thought to provide the two major modes of information coding by neurons. Bursts are known to be crucial for fundamental processes between neuron pairs, such as neuronal communications and synaptic plasticity. Neuronal bursting also has implications in neurodegenerative diseases and mental disorders. Despite these findings on the roles of bursts, whether and how bursts have an advantage over isolated spikes in the network-level computation remains elusive. Here, we demonstrate in a computational model that not isolated spikes, but intrinsic bursts can greatly facilitate learning of Lévy flight random walk trajectories by synchronizing burst onsets across a neural population. Lévy flight is a hallmark of optimal search strategies and appears in cognitive behaviors such as saccadic eye movements and memory retrieval. Our results suggest that bursting is crucial for sequence learning by recurrent neural networks when sequences comprise long-tailed distributed discrete jumps.
Collapse
|
49
|
|
50
|
Distinct Firing Activities of the Hypothalamic Arcuate Nucleus Neurons to Appetite Hormones. Int J Mol Sci 2022; 23:ijms23052609. [PMID: 35269751 PMCID: PMC8910626 DOI: 10.3390/ijms23052609] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2022] [Revised: 02/22/2022] [Accepted: 02/22/2022] [Indexed: 01/27/2023] Open
Abstract
The hypothalamic arcuate nucleus (Arc) is a central unit that controls the appetite through the integration of metabolic, hormonal, and neuronal afferent inputs. Agouti-related protein (AgRP), proopiomelanocortin (POMC), and dopaminergic neurons in the Arc differentially regulate feeding behaviors in response to hunger, satiety, and appetite, respectively. At the time of writing, the anatomical and electrophysiological characterization of these three neurons has not yet been intensively explored. Here, we interrogated the overall characterization of AgRP, POMC, and dopaminergic neurons using genetic mouse models, immunohistochemistry, and whole-cell patch recordings. We identified the distinct geographical location and intrinsic properties of each neuron in the Arc with the transgenic lines labelled with cell-specific reporter proteins. Moreover, AgRP, POMC, and dopaminergic neurons had different firing activities to ghrelin and leptin treatments. Ghrelin led to the increased firing rate of dopaminergic and AgRP neurons, and the decreased firing rate of POMC. In sharp contrast, leptin resulted in the decreased firing rate of AgRP neurons and the increased firing rate of POMC neurons, while it did not change the firing rate of dopaminergic neurons in Arc. These findings demonstrate the anatomical and physiological uniqueness of three hypothalamic Arc neurons to appetite control.
Collapse
|