1
|
Shabani H, Zrenner E, Rathbun DL, Hosseinzadeh Z. Electrical Input Filters of Ganglion Cells in Wild Type and Degenerating rd10 Mouse Retina as a Template for Selective Electrical Stimulation. IEEE Trans Neural Syst Rehabil Eng 2024; 32:850-864. [PMID: 38294929 DOI: 10.1109/tnsre.2024.3360890] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/02/2024]
Abstract
Bionic vision systems are currently limited by indiscriminate activation of all retinal ganglion cells (RGCs)- despite the dozens of known RGC types which each encode a different visual message. Here, we use spike-triggered averaging to explore how electrical responsiveness varies across RGC types toward the goal of using this variation to create type-selective electrical stimuli. A battery of visual stimuli and a randomly distributed sequence of electrical pulses were delivered to healthy and degenerating (4-week-old rd10) mouse retinas. Ganglion cell spike trains were recorded during stimulation using a 60-channel microelectrode array. Hierarchical clustering divided the recorded RGC populations according to their visual and electrical response patterns. Novel electrical stimuli were presented to assess type-specific selectivity. In healthy retinas, responses fell into 35 visual patterns and 14 electrical patterns. In degenerating retinas, responses fell into 12 visual and 23 electrical patterns. Few correspondences between electrical and visual response patterns were found except for the known correspondence of ON visual type with upward deflecting electrical type and OFF cells with downward electrical profiles. Further refinement of the approach presented here may yet yield the elusive nuances necessary for type-selective stimulation. This study greatly deepens our understanding of electrical input filters in the context of detailed visual response characterization and includes the most complete examination yet of degenerating electrical input filters.
Collapse
|
2
|
Pennington JR, David SV. A convolutional neural network provides a generalizable model of natural sound coding by neural populations in auditory cortex. PLoS Comput Biol 2023; 19:e1011110. [PMID: 37146065 DOI: 10.1371/journal.pcbi.1011110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2022] [Revised: 05/17/2023] [Accepted: 04/17/2023] [Indexed: 05/07/2023] Open
Abstract
Convolutional neural networks (CNNs) can provide powerful and flexible models of neural sensory processing. However, the utility of CNNs in studying the auditory system has been limited by their requirement for large datasets and the complex response properties of single auditory neurons. To address these limitations, we developed a population encoding model: a CNN that simultaneously predicts activity of several hundred neurons recorded during presentation of a large set of natural sounds. This approach defines a shared spectro-temporal space and pools statistical power across neurons. Population models of varying architecture performed consistently and substantially better than traditional linear-nonlinear models on data from primary and non-primary auditory cortex. Moreover, population models were highly generalizable. The output layer of a model pre-trained on one population of neurons could be fit to data from novel single units, achieving performance equivalent to that of neurons in the original fit data. This ability to generalize suggests that population encoding models capture a complete representational space across neurons in an auditory cortical field.
Collapse
Affiliation(s)
- Jacob R Pennington
- Washington State University, Vancouver, Washington, United States of America
| | - Stephen V David
- Oregon Hearing Research Center, Oregon Health and Science University, Oregon, United States of America
| |
Collapse
|
3
|
Sadagopan S, Kar M, Parida S. Quantitative models of auditory cortical processing. Hear Res 2023; 429:108697. [PMID: 36696724 PMCID: PMC9928778 DOI: 10.1016/j.heares.2023.108697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 12/17/2022] [Accepted: 01/12/2023] [Indexed: 01/15/2023]
Abstract
To generate insight from experimental data, it is critical to understand the inter-relationships between individual data points and place them in context within a structured framework. Quantitative modeling can provide the scaffolding for such an endeavor. Our main objective in this review is to provide a primer on the range of quantitative tools available to experimental auditory neuroscientists. Quantitative modeling is advantageous because it can provide a compact summary of observed data, make underlying assumptions explicit, and generate predictions for future experiments. Quantitative models may be developed to characterize or fit observed data, to test theories of how a task may be solved by neural circuits, to determine how observed biophysical details might contribute to measured activity patterns, or to predict how an experimental manipulation would affect neural activity. In complexity, quantitative models can range from those that are highly biophysically realistic and that include detailed simulations at the level of individual synapses, to those that use abstract and simplified neuron models to simulate entire networks. Here, we survey the landscape of recently developed models of auditory cortical processing, highlighting a small selection of models to demonstrate how they help generate insight into the mechanisms of auditory processing. We discuss examples ranging from models that use details of synaptic properties to explain the temporal pattern of cortical responses to those that use modern deep neural networks to gain insight into human fMRI data. We conclude by discussing a biologically realistic and interpretable model that our laboratory has developed to explore aspects of vocalization categorization in the auditory pathway.
Collapse
Affiliation(s)
- Srivatsun Sadagopan
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA; Department of Bioengineering, University of Pittsburgh, Pittsburgh, PA, USA; Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA.
| | - Manaswini Kar
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA
| | - Satyabrata Parida
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA
| |
Collapse
|
4
|
Almasi A, Sun SH, Yunzab M, Jung YJ, Meffin H, Ibbotson MR. How Stimulus Statistics Affect the Receptive Fields of Cells in Primary Visual Cortex. J Neurosci 2022; 42:5198-5211. [PMID: 35610048 PMCID: PMC9236288 DOI: 10.1523/jneurosci.0664-21.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2021] [Revised: 05/08/2022] [Accepted: 05/11/2022] [Indexed: 12/31/2022] Open
Abstract
We studied the changes that neuronal receptive field (RF) models undergo when the statistics of the stimulus are changed from those of white Gaussian noise (WGN) to those of natural scenes (NSs), by fitting the models to multielectrode data recorded from primary visual cortex (V1) of female cats. This allowed the estimation of both a cascade of linear filters on the stimulus, as well as the static nonlinearities that map the output of the filters to the neuronal spike rates. We found that cells respond differently to these two classes of stimuli, with mostly higher spike rates and shorter response latencies to NSs than to WGN. The most striking finding was that NSs resulted in RFs that had additional uncovered filters compared with WGN. This finding was not an artifact of the higher spike rates observed for NSs relative to WGN, but rather was related to a change in coding. Our results reveal a greater extent of nonlinear processing in V1 neurons when stimulated using NSs compared with WGN. Our findings indicate the existence of nonlinear mechanisms that endow V1 neurons with context-dependent transmission of visual information.SIGNIFICANCE STATEMENT This study addresses a fundamental question about the concept of the receptive field (RF): does the encoding of information depend on the context or statistical regularities of the stimulus type? We applied state-of-the-art RF modeling techniques to data collected from multielectrode recordings from cat visual cortex in response to two statistically distinct stimulus types: white Gaussian noise and natural scenes. We find significant differences between the RFs that emerge from our data-driven modeling. Natural scenes result in far more complex RFs that combine multiple features in the visual input. Our findings reveal that different regimes or modes of operation are at work in visual cortical processing depending on the information present in the visual input. The complexity of V1 neural coding appears to be dependent on the complexity of the stimulus. We believe this new finding will have interesting implications for our understanding of the efficient transmission of information in sensory systems, which is an integral assumption of many computational theories (e.g., efficient and predictive coding of sensory processing in the brain).
Collapse
Affiliation(s)
- Ali Almasi
- National Vision Research Institute, Carlton, Victoria 3053, Australia
| | - Shi Hai Sun
- National Vision Research Institute, Carlton, Victoria 3053, Australia
| | - Molis Yunzab
- National Vision Research Institute, Carlton, Victoria 3053, Australia
| | - Young Jun Jung
- National Vision Research Institute, Carlton, Victoria 3053, Australia
| | - Hamish Meffin
- Department of Biomedical Engineering, The University of Melbourne, Parkville, Victoria 3010, Australia
| | - Michael R Ibbotson
- National Vision Research Institute, Carlton, Victoria 3053, Australia
- Department of Optometry and Vision Sciences, The University of Melbourne, Parkville, Victoria 3010, Australia
| |
Collapse
|
5
|
Shabani H, Sadeghi M, Zrenner E, Rathbun DL, Hosseinzadeh Z. Classification of pseudocalcium visual responses from mouse retinal ganglion cells. Vis Neurosci 2021; 38:E016. [PMID: 35548862 DOI: 10.1017/s0952523821000158] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/11/2022]
Abstract
Recently, a detailed catalog of 32 retinal ganglion cell (RGC) visual response patterns in mouse has emerged. However, the 10,000 samples required for this catalog-based on fluorescent signals from a calcium indicator dye-are much harder to acquire from the extracellular spike train recordings underlying our bionic vision research. Therefore, we sought to convert spike trains into pseudocalcium signals so that our data could be directly matched to the 32 predefined, calcium signal-based groups. A microelectrode array (MEA) was used to record spike trains from mouse RGCs of 29 retinas. Visual stimuli were adapted from the Baden et al. study; including moving bars, full-field contrast and temporal frequency chirps, and black-white and UV-green color flashes. Spike train histograms were converted into pseudocalcium traces with an OGB-1 convolution kernel. Response features were extracted using sparse principal components analysis to match each RGC to one of the 32 RGC groups. These responses mapped onto of the 32 previously described groups; however, some of the groups remained unmatched. Thus, adaptation of the Baden et al. methodology for MEA recordings of spike trains instead of calcium recordings was partially successful. Different classification methods, however, will be needed to define clear RGC groups from MEA data for our bionic vision research. Nevertheless, others may pursue a pseudocalcium approach to reconcile spike trains with calcium signals. This work will help to guide them on the limitations and potential pitfalls of such an approach.
Collapse
Affiliation(s)
- H Shabani
- Institute for Ophthalmic Research, Centre for Ophthalmology, Eberhard Karls University, Tübingen, Germany
| | - Mahdi Sadeghi
- Institute for Ophthalmic Research, Centre for Ophthalmology, Eberhard Karls University, Tübingen, Germany
| | - E Zrenner
- Institute for Ophthalmic Research, Centre for Ophthalmology, Eberhard Karls University, Tübingen, Germany
- Werner Reichardt Centre for Integrative Neuroscience (CIN), Tübingen, Germany
| | - D L Rathbun
- Institute for Ophthalmic Research, Centre for Ophthalmology, Eberhard Karls University, Tübingen, Germany
- Department of Ophthalmology, Detroit Institute of Ophthalmology, Henry Ford Health System, Detroit, Michigan
| | - Z Hosseinzadeh
- Department of Molecular and Cellular Mechanisms of Neurodegeneration, Paul Flechsig Institute for Brain Research, University of Leipzig, Leipzig, Germany
| |
Collapse
|
6
|
Bush NE, Solla SA, Hartmann MJZ. Continuous, multidimensional coding of 3D complex tactile stimuli by primary sensory neurons of the vibrissal system. Proc Natl Acad Sci U S A 2021; 118:e2020194118. [PMID: 34353902 PMCID: PMC8364131 DOI: 10.1073/pnas.2020194118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Across all sensory modalities, first-stage sensory neurons are an information bottleneck: they must convey all information available for an animal to perceive and act in its environment. Our understanding of coding properties of primary sensory neurons in the auditory and visual systems has been aided by the use of increasingly complex, naturalistic stimulus sets. By comparison, encoding properties of primary somatosensory afferents are poorly understood. Here, we use the rodent whisker system to examine how tactile information is represented in primary sensory neurons of the trigeminal ganglion (Vg). Vg neurons have long been thought to segregate into functional classes associated with separate streams of information processing. However, this view is based on Vg responses to restricted stimulus sets which potentially underreport the coding capabilities of these neurons. In contrast, the current study records Vg responses to complex three-dimensional (3D) stimulation while quantifying the complete 3D whisker shape and mechanics, thereby beginning to reveal their full representational capabilities. The results show that individual Vg neurons simultaneously represent multiple mechanical features of a stimulus, do not preferentially encode principal components of the stimuli, and represent continuous and tiled variations of all available mechanical information. These results directly contrast with proposed codes in which subpopulations of Vg neurons encode select stimulus features. Instead, individual Vg neurons likely overcome the information bottleneck by encoding large regions of a complex sensory space. This proposed tiled and multidimensional representation at the Vg directly constrains the computations performed by more central neurons of the vibrissotrigeminal pathway.
Collapse
Affiliation(s)
- Nicholas E Bush
- Interdepartmental Neuroscience Program, Northwestern University, Evanston, IL 60208
| | - Sara A Solla
- Department of Physics and Astronomy, Northwestern University, Evanston, IL 60208
- Department of Physiology, Northwestern University, Chicago, IL 60611
| | - Mitra J Z Hartmann
- Department of Biomedical Engineering, Northwestern University, Evanston, IL 60208;
- Department of Mechanical Engineering, Northwestern University, Evanston, IL 60208
| |
Collapse
|
7
|
Keeley SL, Zoltowski DM, Aoi MC, Pillow JW. Modeling statistical dependencies in multi-region spike train data. Curr Opin Neurobiol 2020; 65:194-202. [PMID: 33334641 PMCID: PMC7769979 DOI: 10.1016/j.conb.2020.11.005] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2020] [Revised: 11/10/2020] [Accepted: 11/10/2020] [Indexed: 11/17/2022]
Abstract
Neural computations underlying cognition and behavior rely on the coordination of neural activity across multiple brain areas. Understanding how brain areas interact to process information or generate behavior is thus a central question in neuroscience. Here we provide an overview of statistical approaches for characterizing statistical dependencies in multi-region spike train recordings. We focus on two classes of models in particular: regression-based models and shared latent variable models. Regression-based models describe interactions in terms of a directed transformation of information from one region to another. Shared latent variable models, on the other hand, seek to describe interactions in terms of sources that capture common fluctuations in spiking activity across regions. We discuss the advantages and limitations of each of these approaches and future directions for the field. We intend this review to be an introduction to the statistical methods in multi-region models for computational neuroscientists and experimentalists alike.
Collapse
Affiliation(s)
- Stephen L Keeley
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA
| | - David M Zoltowski
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA
| | - Mikio C Aoi
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA
| | - Jonathan W Pillow
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA.
| |
Collapse
|
8
|
Pennington JR, David SV. Complementary Effects of Adaptation and Gain Control on Sound Encoding in Primary Auditory Cortex. eNeuro 2020; 7:ENEURO.0205-20.2020. [PMID: 33109632 PMCID: PMC7675144 DOI: 10.1523/eneuro.0205-20.2020] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2020] [Revised: 08/15/2020] [Accepted: 09/05/2020] [Indexed: 11/24/2022] Open
Abstract
An important step toward understanding how the brain represents complex natural sounds is to develop accurate models of auditory coding by single neurons. A commonly used model is the linear-nonlinear spectro-temporal receptive field (STRF; LN model). The LN model accounts for many features of auditory tuning, but it cannot account for long-lasting effects of sensory context on sound-evoked activity. Two mechanisms that may support these contextual effects are short-term plasticity (STP) and contrast-dependent gain control (GC), which have inspired expanded versions of the LN model. Both models improve performance over the LN model, but they have never been compared directly. Thus, it is unclear whether they account for distinct processes or describe one phenomenon in different ways. To address this question, we recorded activity of neurons in primary auditory cortex (A1) of awake ferrets during presentation of natural sounds. We then fit models incorporating one nonlinear mechanism (GC or STP) or both (GC+STP) using this single dataset, and measured the correlation between the models' predictions and the recorded neural activity. Both the STP and GC models performed significantly better than the LN model, but the GC+STP model outperformed both individual models. We also quantified the equivalence of STP and GC model predictions and found only modest similarity. Consistent results were observed for a dataset collected in clean and noisy acoustic contexts. These results establish general methods for evaluating the equivalence of arbitrarily complex encoding models and suggest that the STP and GC models describe complementary processes in the auditory system.
Collapse
Affiliation(s)
- Jacob R Pennington
- Department of Mathematics, Washington State University, Vancouver, WA, 98686
| | - Stephen V David
- Department of Otolaryngology, Oregon Health and Science University, Portland, OR, 97239
| |
Collapse
|
9
|
Latimer KW, Rieke F, Pillow JW. Inferring synaptic inputs from spikes with a conductance-based neural encoding model. eLife 2019; 8:47012. [PMID: 31850846 PMCID: PMC6989090 DOI: 10.7554/elife.47012] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2019] [Accepted: 12/17/2019] [Indexed: 01/15/2023] Open
Abstract
Descriptive statistical models of neural responses generally aim to characterize the mapping from stimuli to spike responses while ignoring biophysical details of the encoding process. Here, we introduce an alternative approach, the conductance-based encoding model (CBEM), which describes a mapping from stimuli to excitatory and inhibitory synaptic conductances governing the dynamics of sub-threshold membrane potential. Remarkably, we show that the CBEM can be fit to extracellular spike train data and then used to predict excitatory and inhibitory synaptic currents. We validate these predictions with intracellular recordings from macaque retinal ganglion cells. Moreover, we offer a novel quasi-biophysical interpretation of the Poisson generalized linear model (GLM) as a special case of the CBEM in which excitation and inhibition are perfectly balanced. This work forges a new link between statistical and biophysical models of neural encoding and sheds new light on the biophysical variables that underlie spiking in the early visual pathway.
Collapse
Affiliation(s)
- Kenneth W Latimer
- Department of Physiology and Biophysics, University of Washington, Seattle, United States
| | - Fred Rieke
- Department of Physiology and Biophysics, University of Washington, Seattle, United States
| | - Jonathan W Pillow
- Princeton Neuroscience Institute, Department of Psychology, Princeton University, Princeton, United States
| |
Collapse
|
10
|
Abstract
With modern neurophysiological methods able to record neural activity throughout the visual pathway in the context of arbitrarily complex visual stimulation, our understanding of visual system function is becoming limited by the available models of visual neurons that can be directly related to such data. Different forms of statistical models are now being used to probe the cellular and circuit mechanisms shaping neural activity, understand how neural selectivity to complex visual features is computed, and derive the ways in which neurons contribute to systems-level visual processing. However, models that are able to more accurately reproduce observed neural activity often defy simple interpretations. As a result, rather than being used solely to connect with existing theories of visual processing, statistical modeling will increasingly drive the evolution of more sophisticated theories.
Collapse
Affiliation(s)
- Daniel A. Butts
- Department of Biology and Program in Neuroscience and Cognitive Science, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
11
|
Characterizing and dissociating multiple time-varying modulatory computations influencing neuronal activity. PLoS Comput Biol 2019; 15:e1007275. [PMID: 31513570 PMCID: PMC6759185 DOI: 10.1371/journal.pcbi.1007275] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2018] [Revised: 09/24/2019] [Accepted: 07/18/2019] [Indexed: 11/19/2022] Open
Abstract
In many brain areas, sensory responses are heavily modulated by factors including attentional state, context, reward history, motor preparation, learned associations, and other cognitive variables. Modelling the effect of these modulatory factors on sensory responses has proven challenging, mostly due to the time-varying and nonlinear nature of the underlying computations. Here we present a computational model capable of capturing and dissociating multiple time-varying modulatory effects on neuronal responses on the order of milliseconds. The model’s performance is tested on extrastriate perisaccadic visual responses in nonhuman primates. Visual neurons respond to stimuli presented around the time of saccades differently than during fixation. These perisaccadic changes include sensitivity to the stimuli presented at locations outside the neuron’s receptive field, which suggests a contribution of multiple sources to perisaccadic response generation. Current computational approaches cannot quantitatively characterize the contribution of each modulatory source in response generation, mainly due to the very short timescale on which the saccade takes place. In this study, we use a high spatiotemporal resolution experimental paradigm along with a novel extension of the generalized linear model framework (GLM), termed the sparse-variable GLM, to allow for time-varying model parameters representing the temporal evolution of the system with a resolution on the order of milliseconds. We used this model framework to precisely map the temporal evolution of the spatiotemporal receptive field of visual neurons in the middle temporal area during the execution of a saccade. Moreover, an extended model based on a factorization of the sparse-variable GLM allowed us to disassociate and quantify the contribution of individual sources to the perisaccadic response. Our results show that our novel framework can precisely capture the changes in sensitivity of neurons around the time of saccades, and provide a general framework to quantitatively track the role of multiple modulatory sources over time. The sensory responses of neurons in many brain areas, particularly those in higher prefrontal or parietal areas, are strongly influenced by factors including task rules, attentional state, context, reward history, motor preparation, learned associations, and other cognitive variables. These modulations often occur in combination, or on fast timescales which present a challenge for both experimental and modelling approaches aiming to describe the underlying mechanisms or computations. Here we present a computational model capable of capturing and dissociating multiple time-varying modulatory effects on spiking responses on the order of milliseconds. The model’s performance is evaluated by testing its ability to reproduce and dissociate multiple changes in visual sensitivity occurring in extrastriate visual cortex around the time of rapid eye movements. No previous model is capable of capturing these changes with as fine a resolution as that presented here. Our model both provides specific insight into the nature and time course of changes in visual sensitivity around the time of eye movements, and offers a general framework applicable to a wide variety of contexts in which sensory processing is modulated dynamically by multiple time-varying cognitive or behavioral factors, to understand the neuronal computations underpinning these modulations and make predictions about the underlying mechanisms.
Collapse
|
12
|
Correction: The Equivalence of Information-Theoretic and Likelihood-Based Methods for Neural Dimensionality Reduction. PLoS Comput Biol 2019; 15:e1007139. [PMID: 31199805 PMCID: PMC6568373 DOI: 10.1371/journal.pcbi.1007139] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
|
13
|
Geng K, Shin DC, Song D, Hampson RE, Deadwyler SA, Berger TW, Marmarelis VZ. Multi-Input, Multi-Output Neuronal Mode Network Approach to Modeling the Encoding Dynamics and Functional Connectivity of Neural Systems. Neural Comput 2019; 31:1327-1355. [PMID: 31113305 DOI: 10.1162/neco_a_01204] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
This letter proposes a novel method, multi-input, multi-output neuronal mode network (MIMO-NMN), for modeling encoding dynamics and functional connectivity in neural ensembles such as the hippocampus. Compared with conventional approaches such as the Volterra-Wiener model, linear-nonlinear-cascade (LNC) model, and generalized linear model (GLM), the NMN has several advantages in terms of estimation accuracy, model interpretation, and functional connectivity analysis. We point out the limitations of current neural spike modeling methods, especially the estimation biases caused by the imbalanced class problem when the number of zeros is significantly larger than ones in the spike data. We use synthetic data to test the performance of NMN with a comparison of the traditional methods, and the results indicate the NMN approach could reduce the imbalanced class problem and achieve better predictions. Subsequently, we apply the MIMO-NMN method to analyze data from the human hippocampus. The results indicate that the MIMO-NMN method is a promising approach to modeling neural dynamics and analyzing functional connectivity of multi-neuronal data.
Collapse
Affiliation(s)
- Kunling Geng
- Department of Biomedical Engineering and Biomedical Simulations Resource Center, University of Southern California, Los Angeles, CA, 90089, U.S.A.
| | - Dae C Shin
- Department of Biomedical Engineering and Biomedical Simulations Resource Center, University of Southern California, Los Angeles, CA, 90089, U.S.A.
| | - Dong Song
- Department of Biomedical Engineering and Biomedical Simulations Resource Center, University of Southern California, Los Angeles, CA, 90089, U.S.A.
| | - Robert E Hampson
- Department of Physiology and Pharmacology, Wake Forest School of Medicine, Winston-Salem, NC, 27157, U.S.A.
| | - Samuel A Deadwyler
- Department of Physiology and Pharmacology, Wake Forest School of Medicine, Winston-Salem, NC, 27157, U.S.A.
| | - Theodore W Berger
- Department of Biomedical Engineering and Biomedical Simulations Resource Center, University of Southern California, Los Angeles, CA, 90089, U.S.A.
| | - Vasilis Z Marmarelis
- Department of Biomedical Engineering and Biomedical Simulations Resource Center, University of Southern California, Los Angeles, CA, 90089, U.S.A.
| |
Collapse
|
14
|
Pandarinath C, O'Shea DJ, Collins J, Jozefowicz R, Stavisky SD, Kao JC, Trautmann EM, Kaufman MT, Ryu SI, Hochberg LR, Henderson JM, Shenoy KV, Abbott LF, Sussillo D. Inferring single-trial neural population dynamics using sequential auto-encoders. Nat Methods 2018; 15:805-815. [PMID: 30224673 PMCID: PMC6380887 DOI: 10.1038/s41592-018-0109-9] [Citation(s) in RCA: 325] [Impact Index Per Article: 46.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Accepted: 06/28/2018] [Indexed: 01/28/2023]
Abstract
Neuroscience is experiencing a revolution in which simultaneous recording
of many thousands of neurons is revealing population dynamics that are not
apparent from single-neuron responses. This structure is typically extracted
from trial-averaged data, but deeper understanding requires studying
single-trial phenomena, which is challenging due to incomplete sampling of the
neural population, trial-to-trial variability, and fluctuations in action
potential timing. We introduce Latent Factor Analysis via Dynamical Systems
(LFADS), a deep learning method to infer latent dynamics from single-trial
neural spiking data. LFADS uses a nonlinear dynamical system to infer the
dynamics underlying observed spiking activity and to extract
‘de-noised’ single-trial firing rates. When applied to a variety
of monkey and human motor cortical datasets, LFADS predicts observed behavioral
variables with unprecedented accuracy, extracts precise estimates of neural
dynamics on single trials, infers perturbations to those dynamics that correlate
with behavioral choices, and combines data from non-overlapping recording
sessions spanning months to improve inference of underlying dynamics.
Collapse
Affiliation(s)
- Chethan Pandarinath
- Wallace H. Coulter Department of Biomedical Engineering, Emory University and Georgia Institute of Technology, Atlanta, GA, USA. .,Department of Neurosurgery, Emory University, Atlanta, GA, USA. .,Department of Neurosurgery, Stanford University, Stanford, CA, USA. .,Department of Electrical Engineering, Stanford University, Stanford, CA, USA. .,Stanford Neurosciences Institute, Stanford University, Stanford, CA, USA.
| | - Daniel J O'Shea
- Department of Electrical Engineering, Stanford University, Stanford, CA, USA.,Neurosciences Graduate Program, Stanford University, Stanford, CA, USA
| | - Jasmine Collins
- Google AI, Google Inc., Mountain View, CA, USA.,University of California, Berkeley, Berkeley, CA, USA
| | - Rafal Jozefowicz
- Google AI, Google Inc., Mountain View, CA, USA.,OpenAI, San Francisco, CA, USA
| | - Sergey D Stavisky
- Department of Neurosurgery, Stanford University, Stanford, CA, USA.,Department of Electrical Engineering, Stanford University, Stanford, CA, USA.,Stanford Neurosciences Institute, Stanford University, Stanford, CA, USA.,Neurosciences Graduate Program, Stanford University, Stanford, CA, USA
| | - Jonathan C Kao
- Department of Electrical Engineering, Stanford University, Stanford, CA, USA.,Department of Electrical Engineering, University of California, Los Angeles, Los Angeles, CA, USA
| | - Eric M Trautmann
- Neurosciences Graduate Program, Stanford University, Stanford, CA, USA
| | - Matthew T Kaufman
- Neurosciences Graduate Program, Stanford University, Stanford, CA, USA.,Cold Spring Harbor Laboratory, Cold Spring Harbor, NY, USA
| | - Stephen I Ryu
- Department of Electrical Engineering, Stanford University, Stanford, CA, USA.,Department of Neurosurgery, Palo Alto Medical Foundation, Palo Alto, CA, USA
| | - Leigh R Hochberg
- VA RR&D Center for Neurorestoration and Neurotechnology, Veterans Affairs Medical Center, Providence, RI, USA.,Center for Neurotechnology and Neurorecovery, Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA.,School of Engineering and Carney Institute for Brain Science, Brown University, Providence, RI, USA
| | - Jaimie M Henderson
- Department of Neurosurgery, Stanford University, Stanford, CA, USA.,Stanford Neurosciences Institute, Stanford University, Stanford, CA, USA
| | - Krishna V Shenoy
- Department of Electrical Engineering, Stanford University, Stanford, CA, USA.,Stanford Neurosciences Institute, Stanford University, Stanford, CA, USA.,Department of Neurobiology, Stanford University, Stanford, CA, USA.,Department of Bioengineering, Stanford University, Stanford, CA, USA.,Bio-X Program, Stanford University, Stanford, CA, USA.,Howard Hughes Medical Institute, Stanford University, Stanford, CA, USA
| | - L F Abbott
- Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USA.,Department of Neuroscience, Columbia University, New York, NY, USA.,Department of Physiology and Cellular Biophysics, Columbia University, New York, NY, USA
| | - David Sussillo
- Department of Electrical Engineering, Stanford University, Stanford, CA, USA. .,Stanford Neurosciences Institute, Stanford University, Stanford, CA, USA. .,Google AI, Google Inc., Mountain View, CA, USA.
| |
Collapse
|
15
|
Westö J, May PJC. Describing complex cells in primary visual cortex: a comparison of context and multifilter LN models. J Neurophysiol 2018; 120:703-719. [PMID: 29718805 PMCID: PMC6139451 DOI: 10.1152/jn.00916.2017] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2018] [Revised: 04/30/2018] [Accepted: 04/30/2018] [Indexed: 11/24/2022] Open
Abstract
Receptive field (RF) models are an important tool for deciphering neural responses to sensory stimuli. The two currently popular RF models are multifilter linear-nonlinear (LN) models and context models. Models are, however, never correct, and they rely on assumptions to keep them simple enough to be interpretable. As a consequence, different models describe different stimulus-response mappings, which may or may not be good approximations of real neural behavior. In the current study, we take up two tasks: 1) we introduce new ways to estimate context models with realistic nonlinearities, that is, with logistic and exponential functions, and 2) we evaluate context models and multifilter LN models in terms of how well they describe recorded data from complex cells in cat primary visual cortex. Our results, based on single-spike information and correlation coefficients, indicate that context models outperform corresponding multifilter LN models of equal complexity (measured in terms of number of parameters), with the best increase in performance being achieved by the novel context models. Consequently, our results suggest that the multifilter LN-model framework is suboptimal for describing the behavior of complex cells: the context-model framework is clearly superior while still providing interpretable quantizations of neural behavior. NEW & NOTEWORTHY We used data from complex cells in primary visual cortex to estimate a wide variety of receptive field models from two frameworks that have previously not been compared with each other. The models included traditionally used multifilter linear-nonlinear models and novel variants of context models. Using mutual information and correlation coefficients as performance measures, we showed that context models are superior for describing complex cells and that the novel context models performed the best.
Collapse
Affiliation(s)
- Johan Westö
- Department of Neuroscience and Biomedical Engineering Aalto University , Espoo , Finland
| | - Patrick J C May
- Department of Psychology, Lancaster University , Lancaster , United Kingdom
| |
Collapse
|
16
|
Benjamin AS, Fernandes HL, Tomlinson T, Ramkumar P, VerSteeg C, Chowdhury RH, Miller LE, Kording KP. Modern Machine Learning as a Benchmark for Fitting Neural Responses. Front Comput Neurosci 2018; 12:56. [PMID: 30072887 PMCID: PMC6060269 DOI: 10.3389/fncom.2018.00056] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2017] [Accepted: 06/29/2018] [Indexed: 11/13/2022] Open
Abstract
Neuroscience has long focused on finding encoding models that effectively ask "what predicts neural spiking?" and generalized linear models (GLMs) are a typical approach. It is often unknown how much of explainable neural activity is captured, or missed, when fitting a model. Here we compared the predictive performance of simple models to three leading machine learning methods: feedforward neural networks, gradient boosted trees (using XGBoost), and stacked ensembles that combine the predictions of several methods. We predicted spike counts in macaque motor (M1) and somatosensory (S1) cortices from standard representations of reaching kinematics, and in rat hippocampal cells from open field location and orientation. Of these methods, XGBoost and the ensemble consistently produced more accurate spike rate predictions and were less sensitive to the preprocessing of features. These methods can thus be applied quickly to detect if feature sets relate to neural activity in a manner not captured by simpler methods. Encoding models built with a machine learning approach accurately predict spike rates and can offer meaningful benchmarks for simpler models.
Collapse
Affiliation(s)
- Ari S. Benjamin
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, United States
| | - Hugo L. Fernandes
- Department of Physical Medicine and Rehabilitation, Rehabilitation Institute of Chicago, Northwestern University, Chicago, IL, United States
| | - Tucker Tomlinson
- Department of Physiology, Northwestern University, Chicago, IL, United States
| | - Pavan Ramkumar
- Department of Physical Medicine and Rehabilitation, Rehabilitation Institute of Chicago, Northwestern University, Chicago, IL, United States
- Department of Neurobiology, Northwestern University, Evanston, IL, United States
| | - Chris VerSteeg
- Department of Biomedical Engineering, Northwestern University, Evanston, IL, United States
| | - Raeed H. Chowdhury
- Department of Physiology, Northwestern University, Chicago, IL, United States
- Department of Biomedical Engineering, Northwestern University, Evanston, IL, United States
| | - Lee E. Miller
- Department of Physical Medicine and Rehabilitation, Rehabilitation Institute of Chicago, Northwestern University, Chicago, IL, United States
- Department of Physiology, Northwestern University, Chicago, IL, United States
- Department of Biomedical Engineering, Northwestern University, Evanston, IL, United States
| | - Konrad P. Kording
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, United States
- Department of Neuroscience, University of Pennsylvania, Philadelphia, PA, United States
| |
Collapse
|
17
|
Xing D, Qian C, Li H, Zhang S, Zhang Q, Hao Y, Zheng X, Wu Z, Wang Y, Pan G. Predicting Spike Trains from PMd to M1 Using Discrete Time Rescaling Targeted GLM. IEEE Trans Cogn Dev Syst 2018. [DOI: 10.1109/tcds.2017.2707466] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
18
|
Weber AI, Pillow JW. Capturing the Dynamical Repertoire of Single Neurons with Generalized Linear Models. Neural Comput 2017; 29:3260-3289. [PMID: 28957020 DOI: 10.1162/neco_a_01021] [Citation(s) in RCA: 42] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
A key problem in computational neuroscience is to find simple, tractable models that are nevertheless flexible enough to capture the response properties of real neurons. Here we examine the capabilities of recurrent point process models known as Poisson generalized linear models (GLMs). These models are defined by a set of linear filters and a point nonlinearity and are conditionally Poisson spiking. They have desirable statistical properties for fitting and have been widely used to analyze spike trains from electrophysiological recordings. However, the dynamical repertoire of GLMs has not been systematically compared to that of real neurons. Here we show that GLMs can reproduce a comprehensive suite of canonical neural response behaviors, including tonic and phasic spiking, bursting, spike rate adaptation, type I and type II excitation, and two forms of bistability. GLMs can also capture stimulus-dependent changes in spike timing precision and reliability that mimic those observed in real neurons, and can exhibit varying degrees of stochasticity, from virtually deterministic responses to greater-than-Poisson variability. These results show that Poisson GLMs can exhibit a wide range of dynamic spiking behaviors found in real neurons, making them well suited for qualitative dynamical as well as quantitative statistical studies of single-neuron and population response properties.
Collapse
Affiliation(s)
- Alison I Weber
- Graduate Program in Neuroscience, University of Washington, Seattle, WA 98195, U.S.A.
| | - Jonathan W Pillow
- Princeton Neuroscience Institute, Princeton University, and Department of Psychology, Princeton University, Princeton, NJ 08540, U.S.A.
| |
Collapse
|
19
|
Chen X, Beck JM, Pearson JM. Neuron's eye view: Inferring features of complex stimuli from neural responses. PLoS Comput Biol 2017; 13:e1005645. [PMID: 28827790 PMCID: PMC5578681 DOI: 10.1371/journal.pcbi.1005645] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2016] [Revised: 08/31/2017] [Accepted: 06/17/2017] [Indexed: 11/26/2022] Open
Abstract
Experiments that study neural encoding of stimuli at the level of individual neurons typically choose a small set of features present in the world—contrast and luminance for vision, pitch and intensity for sound—and assemble a stimulus set that systematically varies along these dimensions. Subsequent analysis of neural responses to these stimuli typically focuses on regression models, with experimenter-controlled features as predictors and spike counts or firing rates as responses. Unfortunately, this approach requires knowledge in advance about the relevant features coded by a given population of neurons. For domains as complex as social interaction or natural movement, however, the relevant feature space is poorly understood, and an arbitrary a priori choice of features may give rise to confirmation bias. Here, we present a Bayesian model for exploratory data analysis that is capable of automatically identifying the features present in unstructured stimuli based solely on neuronal responses. Our approach is unique within the class of latent state space models of neural activity in that it assumes that firing rates of neurons are sensitive to multiple discrete time-varying features tied to the stimulus, each of which has Markov (or semi-Markov) dynamics. That is, we are modeling neural activity as driven by multiple simultaneous stimulus features rather than intrinsic neural dynamics. We derive a fast variational Bayesian inference algorithm and show that it correctly recovers hidden features in synthetic data, as well as ground-truth stimulus features in a prototypical neural dataset. To demonstrate the utility of the algorithm, we also apply it to cluster neural responses and demonstrate successful recovery of features corresponding to monkeys and faces in the image set. Many neuroscience experiments begin with a set of reduced stimuli designed to vary only along a small set of variables. Yet many phenomena of interest—natural movies, objects—are not easily parameterized by a small number of dimensions. Here, we develop a novel Bayesian model for clustering stimuli based solely on neural responses, allowing us to discover which latent features of complex stimuli actually drive neural activity. We demonstrate that this model allows us to recover key features of neural responses in a pair of well-studied paradigms.
Collapse
Affiliation(s)
- Xin Chen
- Duke Institute for Brain Sciences, Duke University, Durham, North Carolina, United States of America
- Center for Cognitive Neuroscience, Duke University, Durham, North Carolina, United States of America
| | - Jeffrey M. Beck
- Center for Cognitive Neuroscience, Duke University, Durham, North Carolina, United States of America
- Department of Neurobiology, Duke University Medical Center, Durham, North Carolina, United States of America
| | - John M. Pearson
- Duke Institute for Brain Sciences, Duke University, Durham, North Carolina, United States of America
- Center for Cognitive Neuroscience, Duke University, Durham, North Carolina, United States of America
- * E-mail:
| |
Collapse
|
20
|
Meyer AF, Williamson RS, Linden JF, Sahani M. Models of Neuronal Stimulus-Response Functions: Elaboration, Estimation, and Evaluation. Front Syst Neurosci 2017; 10:109. [PMID: 28127278 PMCID: PMC5226961 DOI: 10.3389/fnsys.2016.00109] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2016] [Accepted: 12/19/2016] [Indexed: 11/13/2022] Open
Abstract
Rich, dynamic, and dense sensory stimuli are encoded within the nervous system by the time-varying activity of many individual neurons. A fundamental approach to understanding the nature of the encoded representation is to characterize the function that relates the moment-by-moment firing of a neuron to the recent history of a complex sensory input. This review provides a unifying and critical survey of the techniques that have been brought to bear on this effort thus far—ranging from the classical linear receptive field model to modern approaches incorporating normalization and other nonlinearities. We address separately the structure of the models; the criteria and algorithms used to identify the model parameters; and the role of regularizing terms or “priors.” In each case we consider benefits or drawbacks of various proposals, providing examples for when these methods work and when they may fail. Emphasis is placed on key concepts rather than mathematical details, so as to make the discussion accessible to readers from outside the field. Finally, we review ways in which the agreement between an assumed model and the neuron's response may be quantified. Re-implemented and unified code for many of the methods are made freely available.
Collapse
Affiliation(s)
- Arne F Meyer
- Gatsby Computational Neuroscience Unit, University College London London, UK
| | - Ross S Williamson
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear InfirmaryBoston, MA, USA; Department of Otology and Laryngology, Harvard Medical SchoolBoston, MA, USA
| | - Jennifer F Linden
- Ear Institute, University College LondonLondon, UK; Department of Neuroscience, Physiology and Pharmacology, University College LondonLondon, UK
| | - Maneesh Sahani
- Gatsby Computational Neuroscience Unit, University College London London, UK
| |
Collapse
|
21
|
Harper NS, Schoppe O, Willmore BDB, Cui Z, Schnupp JWH, King AJ. Network Receptive Field Modeling Reveals Extensive Integration and Multi-feature Selectivity in Auditory Cortical Neurons. PLoS Comput Biol 2016; 12:e1005113. [PMID: 27835647 PMCID: PMC5105998 DOI: 10.1371/journal.pcbi.1005113] [Citation(s) in RCA: 35] [Impact Index Per Article: 3.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2015] [Accepted: 08/22/2016] [Indexed: 11/28/2022] Open
Abstract
Cortical sensory neurons are commonly characterized using the receptive field, the linear dependence of their response on the stimulus. In primary auditory cortex neurons can be characterized by their spectrotemporal receptive fields, the spectral and temporal features of a sound that linearly drive a neuron. However, receptive fields do not capture the fact that the response of a cortical neuron results from the complex nonlinear network in which it is embedded. By fitting a nonlinear feedforward network model (a network receptive field) to cortical responses to natural sounds, we reveal that primary auditory cortical neurons are sensitive over a substantially larger spectrotemporal domain than is seen in their standard spectrotemporal receptive fields. Furthermore, the network receptive field, a parsimonious network consisting of 1-7 sub-receptive fields that interact nonlinearly, consistently better predicts neural responses to auditory stimuli than the standard receptive fields. The network receptive field reveals separate excitatory and inhibitory sub-fields with different nonlinear properties, and interaction of the sub-fields gives rise to important operations such as gain control and conjunctive feature detection. The conjunctive effects, where neurons respond only if several specific features are present together, enable increased selectivity for particular complex spectrotemporal structures, and may constitute an important stage in sound recognition. In conclusion, we demonstrate that fitting auditory cortical neural responses with feedforward network models expands on simple linear receptive field models in a manner that yields substantially improved predictive power and reveals key nonlinear aspects of cortical processing, while remaining easy to interpret in a physiological context.
Collapse
Affiliation(s)
- Nicol S. Harper
- Dept. of Physiology, Anatomy and Genetics (DPAG), Sherrington Building, University of Oxford, United Kingdom
- Institute of Biomedical Engineering, Department of Engineering Science, Old Road Campus Research Building, University of Oxford, Headington, United Kingdom
| | - Oliver Schoppe
- Dept. of Physiology, Anatomy and Genetics (DPAG), Sherrington Building, University of Oxford, United Kingdom
- Bio-Inspired Information Processing, Technische Universität München, Germany
| | - Ben D. B. Willmore
- Dept. of Physiology, Anatomy and Genetics (DPAG), Sherrington Building, University of Oxford, United Kingdom
| | - Zhanfeng Cui
- Institute of Biomedical Engineering, Department of Engineering Science, Old Road Campus Research Building, University of Oxford, Headington, United Kingdom
| | - Jan W. H. Schnupp
- Dept. of Physiology, Anatomy and Genetics (DPAG), Sherrington Building, University of Oxford, United Kingdom
- Department of Biomedical Science, City University of Hong Kong, Kowloon Tong, Hong Kong
| | - Andrew J. King
- Dept. of Physiology, Anatomy and Genetics (DPAG), Sherrington Building, University of Oxford, United Kingdom
| |
Collapse
|
22
|
Agarwal R, Chen Z, Kloosterman F, Wilson MA, Sarma SV. A Novel Nonparametric Approach for Neural Encoding and Decoding Models of Multimodal Receptive Fields. Neural Comput 2016; 28:1356-87. [PMID: 27172447 DOI: 10.1162/neco_a_00847] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Pyramidal neurons recorded from the rat hippocampus and entorhinal cortex, such as place and grid cells, have diverse receptive fields, which are either unimodal or multimodal. Spiking activity from these cells encodes information about the spatial position of a freely foraging rat. At fine timescales, a neuron's spike activity also depends significantly on its own spike history. However, due to limitations of current parametric modeling approaches, it remains a challenge to estimate complex, multimodal neuronal receptive fields while incorporating spike history dependence. Furthermore, efforts to decode the rat's trajectory in one- or two-dimensional space from hippocampal ensemble spiking activity have mainly focused on spike history-independent neuronal encoding models. In this letter, we address these two important issues by extending a recently introduced nonparametric neural encoding framework that allows modeling both complex spatial receptive fields and spike history dependencies. Using this extended nonparametric approach, we develop novel algorithms for decoding a rat's trajectory based on recordings of hippocampal place cells and entorhinal grid cells. Results show that both encoding and decoding models derived from our new method performed significantly better than state-of-the-art encoding and decoding models on 6 minutes of test data. In addition, our model's performance remains invariant to the apparent modality of the neuron's receptive field.
Collapse
Affiliation(s)
- Rahul Agarwal
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, MD 21218, U.S.A.
| | - Zhe Chen
- Department of Psychiatry and Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY 10016, U.S.A.
| | - Fabian Kloosterman
- Neuro-Electronics Research Flanders, Leuven, Belgium; Imec, Leuven, Belgium; and Brain and Cognition Research Unit, KU Leuven 3000, Belgium
| | - Matthew A Wilson
- Picower Institute for Learning and Memory, MIT, Cambridge, MA 02139, U.S.A.
| | - Sridevi V Sarma
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, MD 21218, U.S.A.
| |
Collapse
|
23
|
Thorson IL, Liénard J, David SV. The Essential Complexity of Auditory Receptive Fields. PLoS Comput Biol 2015; 11:e1004628. [PMID: 26683490 PMCID: PMC4684325 DOI: 10.1371/journal.pcbi.1004628] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2015] [Accepted: 10/26/2015] [Indexed: 12/05/2022] Open
Abstract
Encoding properties of sensory neurons are commonly modeled using linear finite impulse response (FIR) filters. For the auditory system, the FIR filter is instantiated in the spectro-temporal receptive field (STRF), often in the framework of the generalized linear model. Despite widespread use of the FIR STRF, numerous formulations for linear filters are possible that require many fewer parameters, potentially permitting more efficient and accurate model estimates. To explore these alternative STRF architectures, we recorded single-unit neural activity from auditory cortex of awake ferrets during presentation of natural sound stimuli. We compared performance of > 1000 linear STRF architectures, evaluating their ability to predict neural responses to a novel natural stimulus. Many were able to outperform the FIR filter. Two basic constraints on the architecture lead to the improved performance: (1) factorization of the STRF matrix into a small number of spectral and temporal filters and (2) low-dimensional parameterization of the factorized filters. The best parameterized model was able to outperform the full FIR filter in both primary and secondary auditory cortex, despite requiring fewer than 30 parameters, about 10% of the number required by the FIR filter. After accounting for noise from finite data sampling, these STRFs were able to explain an average of 40% of A1 response variance. The simpler models permitted more straightforward interpretation of sensory tuning properties. They also showed greater benefit from incorporating nonlinear terms, such as short term plasticity, that provide theoretical advances over the linear model. Architectures that minimize parameter count while maintaining maximum predictive power provide insight into the essential degrees of freedom governing auditory cortical function. They also maximize statistical power available for characterizing additional nonlinear properties that limit current auditory models. Understanding how the brain solves sensory problems can provide useful insight for the development of automated systems such as speech recognizers and image classifiers. Recent developments in nonlinear regression and machine learning have produced powerful algorithms for characterizing the input-output relationship of complex systems. However, the complexity of sensory neural systems, combined with practical limitations on experimental data, make it difficult to apply arbitrarily complex analyses to neural data. In this study we pushed analysis in the opposite direction, toward simpler models. We asked how simple a model can be while still capturing the essential sensory properties of neurons in auditory cortex. We found that substantially simpler formulations of the widely-used spectro-temporal receptive field are able to perform as well as the best current models. These simpler formulations define new basis sets that can be incorporated into state-of-the-art machine learning algorithms for a more exhaustive exploration of sensory processing.
Collapse
Affiliation(s)
- Ivar L. Thorson
- Oregon Hearing Research Center, Oregon Health & Science University, Portland, Oregon, United States of America
| | - Jean Liénard
- Department of Mathematics, Washington State University, Vancouver, Washington, United States of America
| | - Stephen V. David
- Oregon Hearing Research Center, Oregon Health & Science University, Portland, Oregon, United States of America
- * E-mail:
| |
Collapse
|