1
|
Ichwansyah R, Onda K, Egawa J, Matsuo T, Suzuki T, Someya T, Hasegawa I, Kawasaki K. Animacy processing by distributed and interconnected networks in the temporal cortex of monkeys. Front Behav Neurosci 2024; 18:1478439. [PMID: 39735387 PMCID: PMC11671252 DOI: 10.3389/fnbeh.2024.1478439] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2024] [Accepted: 11/21/2024] [Indexed: 12/31/2024] Open
Abstract
Animacy perception, the ability to discern living from non-living entities, is crucial for survival and social interaction, as it includes recognizing abstract concepts such as movement, purpose, and intentions. This process involves interpreting cues that may suggest the intentions or actions of others. It engages the temporal cortex (TC), particularly the superior temporal sulcus (STS) and the adjacent region of the inferior temporal cortex (ITC), as well as the dorsomedial prefrontal cortex (dmPFC). However, it remains unclear how animacy is dynamically encoded over time in these brain areas and whether its processing is distributed or localized. In this study, we addressed these questions by employing a symbolic categorization task involving animate and inanimate objects using natural movie stimuli. Simultaneously, electrocorticography were conducted in both the TC and dmPFC. Time-frequency analysis revealed region-specific frequency representations throughout the observation of the movies. Spatial searchlight decoding analysis demonstrated that animacy processing is represented in a distributed manner. Regions encoding animacy information were found to be dispersed across the fundus and lip of the STS, as well as in the ITC. Next, we examined whether these dispersed regions form functional networks. Independent component analysis revealed that the spatial distribution of the component with the most significant animacy information corresponded with the dispersed regions identified by the spatial decoding analysis. Furthermore, Granger causality analysis indicated that these regions exhibit frequency-specific directional functional connectivity, with a general trend of causal influence from the ITC to STS across multiple frequency bands. Notably, a prominent feedback flow in the alpha band from the ITC to both the ventral bank and fundus of the STS was identified. These findings suggest a distributed and functionally interconnected neural substrate for animacy processing across the STS and ITC.
Collapse
Affiliation(s)
- Rizal Ichwansyah
- Department of Neurophysiology, Niigata University School of Medicine, Niigata, Japan
- Department of Psychiatry, Niigata University School of Medicine, Niigata, Japan
| | - Keigo Onda
- Department of Neurophysiology, Niigata University School of Medicine, Niigata, Japan
- Department of Psychiatry, Niigata University School of Medicine, Niigata, Japan
| | - Jun Egawa
- Department of Psychiatry, Niigata University School of Medicine, Niigata, Japan
| | - Takeshi Matsuo
- Department of Neurosurgery, Tokyo Metropolitan Neurological Hospital, Tokyo, Japan
| | - Takafumi Suzuki
- Center for Information and Neural Networks, National Institute of Information and Communications Technology, Osaka, Japan
| | - Toshiyuki Someya
- Department of Psychiatry, Niigata University School of Medicine, Niigata, Japan
| | - Isao Hasegawa
- Department of Neurophysiology, Niigata University School of Medicine, Niigata, Japan
| | - Keisuke Kawasaki
- Department of Neurophysiology, Niigata University School of Medicine, Niigata, Japan
| |
Collapse
|
2
|
Niu Y, Xiang J, Gao K, Wu J, Sun J, Wang B, Ding R, Dou M, Wen X, Cui X, Zhou M. Multi-Frequency Entropy for Quantifying Complex Dynamics and Its Application on EEG Data. ENTROPY (BASEL, SWITZERLAND) 2024; 26:728. [PMID: 39330063 PMCID: PMC11431093 DOI: 10.3390/e26090728] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2024] [Revised: 07/24/2024] [Accepted: 07/26/2024] [Indexed: 09/28/2024]
Abstract
Multivariate entropy algorithms have proven effective in the complexity dynamic analysis of electroencephalography (EEG) signals, with researchers commonly configuring the variables as multi-channel time series. However, the complex quantification of brain dynamics from a multi-frequency perspective has not been extensively explored, despite existing evidence suggesting interactions among brain rhythms at different frequencies. In this study, we proposed a novel algorithm, termed multi-frequency entropy (mFreEn), enhancing the capabilities of existing multivariate entropy algorithms and facilitating the complexity study of interactions among brain rhythms of different frequency bands. Firstly, utilizing simulated data, we evaluated the mFreEn's sensitivity to various noise signals, frequencies, and amplitudes, investigated the effects of parameters such as the embedding dimension and data length, and analyzed its anti-noise performance. The results indicated that mFreEn demonstrated enhanced sensitivity and reduced parameter dependence compared to traditional multivariate entropy algorithms. Subsequently, the mFreEn algorithm was applied to the analysis of real EEG data. We found that mFreEn exhibited a good diagnostic performance in analyzing resting-state EEG data from various brain disorders. Furthermore, mFreEn showed a good classification performance for EEG activity induced by diverse task stimuli. Consequently, mFreEn provides another important perspective to quantify complex dynamics.
Collapse
Affiliation(s)
- Yan Niu
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Jie Xiang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Kai Gao
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Jinglong Wu
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China;
| | - Jie Sun
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Bin Wang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Runan Ding
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Mingliang Dou
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Xin Wen
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China;
| | - Xiaohong Cui
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China; (Y.N.); (J.X.); (K.G.); (J.S.); (B.W.); (R.D.); (M.D.); (X.C.)
| | - Mengni Zhou
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China;
| |
Collapse
|
3
|
Lützow Holm E, Fernández Slezak D, Tagliazucchi E. Contribution of low-level image statistics to EEG decoding of semantic content in multivariate and univariate models with feature optimization. Neuroimage 2024; 293:120626. [PMID: 38677632 DOI: 10.1016/j.neuroimage.2024.120626] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2024] [Revised: 04/23/2024] [Accepted: 04/24/2024] [Indexed: 04/29/2024] Open
Abstract
Spatio-temporal patterns of evoked brain activity contain information that can be used to decode and categorize the semantic content of visual stimuli. However, this procedure can be biased by low-level image features independently of the semantic content present in the stimuli, prompting the need to understand the robustness of different models regarding these confounding factors. In this study, we trained machine learning models to distinguish between concepts included in the publicly available THINGS-EEG dataset using electroencephalography (EEG) data acquired during a rapid serial visual presentation paradigm. We investigated the contribution of low-level image features to decoding accuracy in a multivariate model, utilizing broadband data from all EEG channels. Additionally, we explored a univariate model obtained through data-driven feature selection applied to the spatial and frequency domains. While the univariate models exhibited better decoding accuracy, their predictions were less robust to the confounding effect of low-level image statistics. Notably, some of the models maintained their accuracy even after random replacement of the training dataset with semantically unrelated samples that presented similar low-level content. In conclusion, our findings suggest that model optimization impacts sensitivity to confounding factors, regardless of the resulting classification performance. Therefore, the choice of EEG features for semantic decoding should ideally be informed by criteria beyond classifier performance, such as the neurobiological mechanisms under study.
Collapse
Affiliation(s)
- Eric Lützow Holm
- National Scientific and Technical Research Council (CONICET), Godoy Cruz 2290, CABA 1425, Argentina; Institute of Applied and Interdisciplinary Physics and Department of Physics, University of Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina.
| | - Diego Fernández Slezak
- National Scientific and Technical Research Council (CONICET), Godoy Cruz 2290, CABA 1425, Argentina; Departamento de Computación, Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina; Instituto de Investigación en Ciencias de la Computación (ICC), CONICET-Universidad de Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina
| | - Enzo Tagliazucchi
- National Scientific and Technical Research Council (CONICET), Godoy Cruz 2290, CABA 1425, Argentina; Institute of Applied and Interdisciplinary Physics and Department of Physics, University of Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina; Latin American Brain Health (BrainLat), Universidad Adolfo Ibáñez, Av. Diag. Las Torres 2640, Peñalolén 7941169, Santiago Región Metropolitana, Chile.
| |
Collapse
|
4
|
Karimi-Rouzbahani H. Evidence for Multiscale Multiplexed Representation of Visual Features in EEG. Neural Comput 2024; 36:412-436. [PMID: 38363657 DOI: 10.1162/neco_a_01649] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 12/01/2023] [Indexed: 02/18/2024]
Abstract
Distinct neural processes such as sensory and memory processes are often encoded over distinct timescales of neural activations. Animal studies have shown that this multiscale coding strategy is also implemented for individual components of a single process, such as individual features of a multifeature stimulus in sensory coding. However, the generalizability of this encoding strategy to the human brain has remained unclear. We asked if individual features of visual stimuli were encoded over distinct timescales. We applied a multiscale time-resolved decoding method to electroencephalography (EEG) collected from human subjects presented with grating visual stimuli to estimate the timescale of individual stimulus features. We observed that the orientation and color of the stimuli were encoded in shorter timescales, whereas spatial frequency and the contrast of the same stimuli were encoded in longer timescales. The stimulus features appeared in temporally overlapping windows along the trial supporting a multiplexed coding strategy. These results provide evidence for a multiplexed, multiscale coding strategy in the human visual system.
Collapse
Affiliation(s)
- Hamid Karimi-Rouzbahani
- Neurosciences Centre, Mater Hospital, Brisbane 4101, Australia
- Queensland Brain Institute, University of Queensland, Brisbane 4067, Australia
- Mater Research Institute, University of Queensland, Brisbane 4101, Australia
| |
Collapse
|
5
|
Ghazaryan G, van Vliet M, Lammi L, Lindh-Knuutila T, Kivisaari S, Hultén A, Salmelin R. Cortical time-course of evidence accumulation during semantic processing. Commun Biol 2023; 6:1242. [PMID: 38066098 PMCID: PMC10709650 DOI: 10.1038/s42003-023-05611-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Accepted: 11/20/2023] [Indexed: 12/18/2023] Open
Abstract
Our understanding of the surrounding world and communication with other people are tied to mental representations of concepts. In order for the brain to recognize an object, it must determine which concept to access based on information available from sensory inputs. In this study, we combine magnetoencephalography and machine learning to investigate how concepts are represented and accessed in the brain over time. Using brain responses from a silent picture naming task, we track the dynamics of visual and semantic information processing, and show that the brain gradually accumulates information on different levels before eventually reaching a plateau. The timing of this plateau point varies across individuals and feature models, indicating notable temporal variation in visual object recognition and semantic processing.
Collapse
Affiliation(s)
- Gayane Ghazaryan
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland.
| | - Marijn van Vliet
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
| | - Lotta Lammi
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
| | - Tiina Lindh-Knuutila
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
| | - Sasa Kivisaari
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
| | - Annika Hultén
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
- Aalto NeuroImaging, Aalto University, P.O. Box 12200, Aalto, FI-00076, Finland
| | - Riitta Salmelin
- Department of Neuroscience and Biomedical Engineering, Aalto University, P.O. Box 12200, FI-00076, Aalto, Finland
- Aalto NeuroImaging, Aalto University, P.O. Box 12200, Aalto, FI-00076, Finland
| |
Collapse
|
6
|
von Seth J, Nicholls VI, Tyler LK, Clarke A. Recurrent connectivity supports higher-level visual and semantic object representations in the brain. Commun Biol 2023; 6:1207. [PMID: 38012301 PMCID: PMC10682037 DOI: 10.1038/s42003-023-05565-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Accepted: 11/09/2023] [Indexed: 11/29/2023] Open
Abstract
Visual object recognition has been traditionally conceptualised as a predominantly feedforward process through the ventral visual pathway. While feedforward artificial neural networks (ANNs) can achieve human-level classification on some image-labelling tasks, it's unclear whether computational models of vision alone can accurately capture the evolving spatiotemporal neural dynamics. Here, we probe these dynamics using a combination of representational similarity and connectivity analyses of fMRI and MEG data recorded during the recognition of familiar, unambiguous objects. Modelling the visual and semantic properties of our stimuli using an artificial neural network as well as a semantic feature model, we find that unique aspects of the neural architecture and connectivity dynamics relate to visual and semantic object properties. Critically, we show that recurrent processing between the anterior and posterior ventral temporal cortex relates to higher-level visual properties prior to semantic object properties, in addition to semantic-related feedback from the frontal lobe to the ventral temporal lobe between 250 and 500 ms after stimulus onset. These results demonstrate the distinct contributions made by semantic object properties in explaining neural activity and connectivity, highlighting it as a core part of object recognition not fully accounted for by current biologically inspired neural networks.
Collapse
Affiliation(s)
- Jacqueline von Seth
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | | | - Lorraine K Tyler
- Department of Psychology, University of Cambridge, Cambridge, UK
- Cambridge Centre for Ageing and Neuroscience (Cam-CAN), University of Cambridge and MRC Cognition and Brain Sciences Unit, Cambridge, UK
| | - Alex Clarke
- Department of Psychology, University of Cambridge, Cambridge, UK.
| |
Collapse
|
7
|
Gruenwald J, Sieghartsleitner S, Kapeller C, Scharinger J, Kamada K, Brunner P, Guger C. Characterization of High-Gamma Activity in Electrocorticographic Signals. Front Neurosci 2023; 17:1206120. [PMID: 37609450 PMCID: PMC10440607 DOI: 10.3389/fnins.2023.1206120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 07/10/2023] [Indexed: 08/24/2023] Open
Abstract
Introduction Electrocorticographic (ECoG) high-gamma activity (HGA) is a widely recognized and robust neural correlate of cognition and behavior. However, fundamental signal properties of HGA, such as the high-gamma frequency band or temporal dynamics of HGA, have never been systematically characterized. As a result, HGA estimators are often poorly adjusted, such that they miss valuable physiological information. Methods To address these issues, we conducted a thorough qualitative and quantitative characterization of HGA in ECoG signals. Our study is based on ECoG signals recorded from 18 epilepsy patients while performing motor control, listening, and visual perception tasks. In this study, we first categorize HGA into HGA types based on the cognitive/behavioral task. For each HGA type, we then systematically quantify three fundamental signal properties of HGA: the high-gamma frequency band, the HGA bandwidth, and the temporal dynamics of HGA. Results The high-gamma frequency band strongly varies across subjects and across cognitive/behavioral tasks. In addition, HGA time courses have lowpass character, with transients limited to 10 Hz. The task-related rise time and duration of these HGA time courses depend on the individual subject and cognitive/behavioral task. Task-related HGA amplitudes are comparable across the investigated tasks. Discussion This study is of high practical relevance because it provides a systematic basis for optimizing experiment design, ECoG acquisition and processing, and HGA estimation. Our results reveal previously unknown characteristics of HGA, the physiological principles of which need to be investigated in further studies.
Collapse
Affiliation(s)
- Johannes Gruenwald
- g.tec medical engineering GmbH, Schiedlberg, Austria
- Institute of Computational Perception, Johannes Kepler University, Linz, Austria
| | - Sebastian Sieghartsleitner
- g.tec medical engineering GmbH, Schiedlberg, Austria
- Institute of Computational Perception, Johannes Kepler University, Linz, Austria
| | | | - Josef Scharinger
- Institute of Computational Perception, Johannes Kepler University, Linz, Austria
| | - Kyousuke Kamada
- Department for Neurosurgery, Asahikawa Medical University, Asahikawa, Japan
- Hokashin Group Megumino Hospital, Sapporo, Japan
| | - Peter Brunner
- National Center for Adaptive Neurotechnologies, Albany, NY, United States
- Department of Neurosurgery, Washington University School of Medicine, St. Louis, MO, United States
| | | |
Collapse
|
8
|
Gu L, Li A, Yang R, Yang J, Pang Y, Qu J, Mei L. Category-specific and category-general neural codes of recognition memory in the ventral visual pathway. Cortex 2023; 164:77-89. [PMID: 37207411 DOI: 10.1016/j.cortex.2023.04.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2022] [Revised: 03/09/2023] [Accepted: 04/03/2023] [Indexed: 05/21/2023]
Abstract
Researchers have identified category-specific brain regions, such as the fusiform face area (FFA) and parahippocampal place area (PPA) in the ventral visual pathway, which respond preferentially to one particular category of visual objects. In addition to their category-specific role in visual object identification and categorization, regions in the ventral visual pathway play critical roles in recognition memory. Nevertheless, it is not clear whether the contributions of those brain regions to recognition memory are category-specific or category-general. To address this question, the present study adopted a subsequent memory paradigm and multivariate pattern analysis (MVPA) to explore category-specific and category-general neural codes of recognition memory in the visual pathway. The results revealed that the right FFA and the bilateral PPA showed category-specific neural patterns supporting recognition memory of faces and scenes, respectively. In contrast, the lateral occipital cortex seemed to carry category-general neural codes of recognition memory. These results provide neuroimaging evidence for category-specific and category-general neural mechanisms of recognition memory in the ventral visual pathway.
Collapse
Affiliation(s)
- Lala Gu
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Aqian Li
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Rui Yang
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Jiayi Yang
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Yingdan Pang
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Jing Qu
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China
| | - Leilei Mei
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, China; School of Psychology, South China Normal University, Guangzhou, China; Guangdong Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, China.
| |
Collapse
|
9
|
Rybář M, Daly I. Neural decoding of semantic concepts: A systematic literature review. J Neural Eng 2022; 19. [PMID: 35344941 DOI: 10.1088/1741-2552/ac619a] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Accepted: 03/27/2022] [Indexed: 11/12/2022]
Abstract
Objective Semantic concepts are coherent entities within our minds. They underpin our thought processes and are a part of the basis for our understanding of the world. Modern neuroscience research is increasingly exploring how individual semantic concepts are encoded within our brains and a number of studies are beginning to reveal key patterns of neural activity that underpin specific concepts. Building upon this basic understanding of the process of semantic neural encoding, neural engineers are beginning to explore tools and methods for semantic decoding: identifying which semantic concepts an individual is focused on at a given moment in time from recordings of their neural activity. In this paper we review the current literature on semantic neural decoding. Approach We conducted this review according to the Preferred Reporting Items for Systematic reviews and Meta-Analysis (PRISMA) guidelines. Specifically, we assess the eligibility of published peer-reviewed reports via a search of PubMed and Google Scholar. We identify a total of 74 studies in which semantic neural decoding is used to attempt to identify individual semantic concepts from neural activity. Results Our review reveals how modern neuroscientific tools have been developed to allow decoding of individual concepts from a range of neuroimaging modalities. We discuss specific neuroimaging methods, experimental designs, and machine learning pipelines that are employed to aid the decoding of semantic concepts. We quantify the efficacy of semantic decoders by measuring information transfer rates. We also discuss current challenges presented by this research area and present some possible solutions. Finally, we discuss some possible emerging and speculative future directions for this research area. Significance Semantic decoding is a rapidly growing area of research. However, despite its increasingly widespread popularity and use in neuroscientific research this is the first literature review focusing on this topic across neuroimaging modalities and with a focus on quantifying the efficacy of semantic decoders.
Collapse
Affiliation(s)
- Milan Rybář
- School of Computer Science and Electronic Engineering, University of Essex, Wivenhoe Park, Colchester, Essex, CO4 3SQ, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| | - Ian Daly
- University of Essex, School of Computer Science and Electronic Engineering, Wivenhoe Park, Colchester, Colchester, Essex, CO4 3SQ, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
| |
Collapse
|
10
|
Voluntary control of semantic neural representations by imagery with conflicting visual stimulation. Commun Biol 2022; 5:214. [PMID: 35304588 PMCID: PMC8933408 DOI: 10.1038/s42003-022-03137-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Accepted: 02/08/2022] [Indexed: 12/04/2022] Open
Abstract
Neural representations of visual perception are affected by mental imagery and attention. Although attention is known to modulate neural representations, it is unknown how imagery changes neural representations when imagined and perceived images semantically conflict. We hypothesized that imagining an image would activate a neural representation during its perception even while watching a conflicting image. To test this hypothesis, we developed a closed-loop system to show images inferred from electrocorticograms using a visual semantic space. The successful control of the feedback images demonstrated that the semantic vector inferred from electrocorticograms became closer to the vector of the imagined category, even while watching images from different categories. Moreover, modulation of the inferred vectors by mental imagery depended asymmetrically on the perceived and imagined categories. Shared neural representation between mental imagery and perception was still activated by the imagery under semantically conflicting perceptions depending on the semantic category. In this study, intracranial EEG recordings show that neural representations of imagined images can still be present in humans even when they are shown conflicting images.
Collapse
|
11
|
Karimi-Rouzbahani H, Woolgar A. When the Whole Is Less Than the Sum of Its Parts: Maximum Object Category Information and Behavioral Prediction in Multiscale Activation Patterns. Front Neurosci 2022; 16:825746. [PMID: 35310090 PMCID: PMC8924472 DOI: 10.3389/fnins.2022.825746] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Accepted: 01/24/2022] [Indexed: 11/19/2022] Open
Abstract
Neural codes are reflected in complex neural activation patterns. Conventional electroencephalography (EEG) decoding analyses summarize activations by averaging/down-sampling signals within the analysis window. This diminishes informative fine-grained patterns. While previous studies have proposed distinct statistical features capable of capturing variability-dependent neural codes, it has been suggested that the brain could use a combination of encoding protocols not reflected in any one mathematical feature alone. To check, we combined 30 features using state-of-the-art supervised and unsupervised feature selection procedures (n = 17). Across three datasets, we compared decoding of visual object category between these 17 sets of combined features, and between combined and individual features. Object category could be robustly decoded using the combined features from all of the 17 algorithms. However, the combination of features, which were equalized in dimension to the individual features, were outperformed across most of the time points by the multiscale feature of Wavelet coefficients. Moreover, the Wavelet coefficients also explained the behavioral performance more accurately than the combined features. These results suggest that a single but multiscale encoding protocol may capture the EEG neural codes better than any combination of protocols. Our findings put new constraints on the models of neural information encoding in EEG.
Collapse
Affiliation(s)
- Hamid Karimi-Rouzbahani
- Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, United Kingdom
- Department of Cognitive Science, Perception in Action Research Centre, Macquarie University, Sydney, NSW, Australia
- Department of Computing, Macquarie University, Sydney, NSW, Australia
| | - Alexandra Woolgar
- Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, United Kingdom
- Department of Cognitive Science, Perception in Action Research Centre, Macquarie University, Sydney, NSW, Australia
| |
Collapse
|
12
|
Nagata K, Kunii N, Shimada S, Fujitani S, Takasago M, Saito N. Spatiotemporal target selection for intracranial neural decoding of abstract and concrete semantics. Cereb Cortex 2022; 32:5544-5554. [PMID: 35169837 PMCID: PMC9753048 DOI: 10.1093/cercor/bhac034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2021] [Revised: 01/18/2022] [Accepted: 01/19/2021] [Indexed: 01/25/2023] Open
Abstract
Decoding the inner representation of a word meaning from human cortical activity is a substantial challenge in the development of speech brain-machine interfaces (BMIs). The semantic aspect of speech is a novel target of speech decoding that may enable versatile communication platforms for individuals with impaired speech ability; however, there is a paucity of electrocorticography studies in this field. We decoded the semantic representation of a word from single-trial cortical activity during an imageability-based property identification task that required participants to discriminate between the abstract and concrete words. Using high gamma activity in the language-dominant hemisphere, a support vector machine classifier could discriminate the 2-word categories with significantly high accuracy (73.1 ± 7.5%). Activities in specific time components from two brain regions were identified as significant predictors of abstract and concrete dichotomy. Classification using these feature components revealed that comparable prediction accuracy could be obtained based on a spatiotemporally targeted decoding approach. Our study demonstrated that mental representations of abstract and concrete word processing could be decoded from cortical high gamma activities, and the coverage of implanted electrodes and time window of analysis could be successfully minimized. Our findings lay the foundation for the future development of semantic-based speech BMIs.
Collapse
Affiliation(s)
- Keisuke Nagata
- Department of Neurosurgery, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan
| | - Naoto Kunii
- Corresponding author: Department of Neurosurgery, The University of Tokyo, 73-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan.
| | - Seijiro Shimada
- Department of Neurosurgery, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan
| | - Shigeta Fujitani
- Department of Neurosurgery, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan
| | - Megumi Takasago
- Department of Neurosurgery, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan
| | - Nobuhito Saito
- Department of Neurosurgery, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8655, Japan
| |
Collapse
|
13
|
|
14
|
Luo S, Rabbani Q, Crone NE. Brain-Computer Interface: Applications to Speech Decoding and Synthesis to Augment Communication. Neurotherapeutics 2022; 19:263-273. [PMID: 35099768 PMCID: PMC9130409 DOI: 10.1007/s13311-022-01190-2] [Citation(s) in RCA: 37] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/16/2022] [Indexed: 01/03/2023] Open
Abstract
Damage or degeneration of motor pathways necessary for speech and other movements, as in brainstem strokes or amyotrophic lateral sclerosis (ALS), can interfere with efficient communication without affecting brain structures responsible for language or cognition. In the worst-case scenario, this can result in the locked in syndrome (LIS), a condition in which individuals cannot initiate communication and can only express themselves by answering yes/no questions with eye blinks or other rudimentary movements. Existing augmentative and alternative communication (AAC) devices that rely on eye tracking can improve the quality of life for people with this condition, but brain-computer interfaces (BCIs) are also increasingly being investigated as AAC devices, particularly when eye tracking is too slow or unreliable. Moreover, with recent and ongoing advances in machine learning and neural recording technologies, BCIs may offer the only means to go beyond cursor control and text generation on a computer, to allow real-time synthesis of speech, which would arguably offer the most efficient and expressive channel for communication. The potential for BCI speech synthesis has only recently been realized because of seminal studies of the neuroanatomical and neurophysiological underpinnings of speech production using intracranial electrocorticographic (ECoG) recordings in patients undergoing epilepsy surgery. These studies have shown that cortical areas responsible for vocalization and articulation are distributed over a large area of ventral sensorimotor cortex, and that it is possible to decode speech and reconstruct its acoustics from ECoG if these areas are recorded with sufficiently dense and comprehensive electrode arrays. In this article, we review these advances, including the latest neural decoding strategies that range from deep learning models to the direct concatenation of speech units. We also discuss state-of-the-art vocoders that are integral in constructing natural-sounding audio waveforms for speech BCIs. Finally, this review outlines some of the challenges ahead in directly synthesizing speech for patients with LIS.
Collapse
Affiliation(s)
- Shiyu Luo
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA.
| | - Qinwan Rabbani
- Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD, USA
| | - Nathan E Crone
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| |
Collapse
|
15
|
Diab MS, Elhosseini MA, El-Sayed MS, Ali HA. Brain Strategy Algorithm for Multiple Object Tracking Based on Merging Semantic Attributes and Appearance Features. SENSORS 2021; 21:s21227604. [PMID: 34833680 PMCID: PMC8625767 DOI: 10.3390/s21227604] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/08/2021] [Revised: 11/10/2021] [Accepted: 11/11/2021] [Indexed: 11/16/2022]
Abstract
The human brain can effortlessly perform vision processes using the visual system, which helps solve multi-object tracking (MOT) problems. However, few algorithms simulate human strategies for solving MOT. Therefore, devising a method that simulates human activity in vision has become a good choice for improving MOT results, especially occlusion. Eight brain strategies have been studied from a cognitive perspective and imitated to build a novel algorithm. Two of these strategies gave our algorithm novel and outstanding results, rescuing saccades and stimulus attributes. First, rescue saccades were imitated by detecting the occlusion state in each frame, representing the critical situation that the human brain saccades toward. Then, stimulus attributes were mimicked by using semantic attributes to reidentify the person in these occlusion states. Our algorithm favourably performs on the MOT17 dataset compared to state-of-the-art trackers. In addition, we created a new dataset of 40,000 images, 190,000 annotations and 4 classes to train the detection model to detect occlusion and semantic attributes. The experimental results demonstrate that our new dataset achieves an outstanding performance on the scaled YOLOv4 detection model by achieving a 0.89 mAP 0.5.
Collapse
Affiliation(s)
- Mai S. Diab
- Faculty of Computer & Artificial Intelligence, Benha University, Benha 13511, Egypt;
- Intoolab Ltd., London WC2H 9JQ, UK
- Correspondence:
| | - Mostafa A. Elhosseini
- Computers Engineering and Control System, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt; (M.A.E.); (H.A.A.)
- College of Computer Science and Engineering in Yanbu, Taibah University, Madinah 46421, Saudi Arabia
| | - Mohamed S. El-Sayed
- Faculty of Computer & Artificial Intelligence, Benha University, Benha 13511, Egypt;
| | - Hesham A. Ali
- Computers Engineering and Control System, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt; (M.A.E.); (H.A.A.)
- Faculty of Artificial Intelligence, Delta University for Science and Technology, Mansoura 35511, Egypt
| |
Collapse
|
16
|
Sato N, Matsumoto R, Shimotake A, Matsuhashi M, Otani M, Kikuchi T, Kunieda T, Mizuhara H, Miyamoto S, Takahashi R, Ikeda A. Frequency-Dependent Cortical Interactions during Semantic Processing: An Electrocorticogram Cross-spectrum Analysis Using a Semantic Space Model. Cereb Cortex 2021; 31:4329-4339. [PMID: 33942078 DOI: 10.1093/cercor/bhab089] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Convergent evidence has demonstrated that semantics are represented by the interaction between a multimodal semantic hub at the anterior temporal lobe (ATL) and other modality-specific association cortical areas. Electrocorticogram (ECoG) recording with high spatiotemporal resolutions is efficient in evaluating such cortical interactions; however, this has not been a focus of preceding studies. The present study evaluated cortical interactions during picture naming using a novel ECoG cross-spectrum analysis, which was formulated from a computational simulation of neuronal networks and combined with a vector space model of semantics. The results clarified three types of frequency-dependent cortical networks: 1) an earlier-period (0.2-0.8 s from stimulus onset) high-gamma-band (90-150 Hz) network with a hub at the posterior fusiform gyrus, 2) a later-period (0.4-1.0 s) beta-band (15-40 Hz) network with multiple hubs at the ventral ATL and posterior middle temporal gyrus, and 3) a pre-articulation theta-band (4-7 Hz) network distributed over widely located cortical regions. These results suggest that frequency-dependent cortical interactions can characterize the underlying processes of semantic cognition, and the beta-band network with a hub at the ventral ATL is especially associated with the formation of semantic representation.
Collapse
Affiliation(s)
- Naoyuki Sato
- Department of Complex and Intelligent Systems, School of Systems Information Science, Future University Hakodate, Hakodate, Hokkaido 041-8655, Japan
| | - Riki Matsumoto
- Division of Neurology, Kobe University Graduate School of Medicine, Kobe, Hyogo 650-0017, Japan.,Department of Neurology, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan
| | - Akihiro Shimotake
- Department of Neurology, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan.,Department of Epilepsy, Movement Disorders and Physiology, Kyoto University Graduate School of Medicine, Sakyo, Kyoto 606-8507, Japan
| | - Masao Matsuhashi
- Department of Epilepsy, Movement Disorders and Physiology, Kyoto University Graduate School of Medicine, Sakyo, Kyoto 606-8507, Japan.,Human Brain Research Center, Kyoto University Graduate School of Medicine, Kyoto 606-8507, Japan
| | - Mayumi Otani
- Department of Neurology, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan
| | - Takayuki Kikuchi
- Department of Neurosurgery, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan
| | - Takeharu Kunieda
- Department of Neurosurgery, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan.,Department of Neurosurgery, Ehime University Graduate School of Medicine, Shizukawa Toon City, Ehime 791-0295, Japan
| | - Hiroaki Mizuhara
- Kyoto University Graduate School of Informatics, Sakyo, Kyoto 606-8501, Japan
| | - Susumu Miyamoto
- Department of Neurosurgery, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan
| | - Ryosuke Takahashi
- Department of Neurology, Kyoto University Graduate School of Medicine, Kyoto University Hospital, Sakyo, Kyoto 606-8507, Japan
| | - Akio Ikeda
- Department of Epilepsy, Movement Disorders and Physiology, Kyoto University Graduate School of Medicine, Sakyo, Kyoto 606-8507, Japan
| |
Collapse
|
17
|
Duan L, Li J, Ji H, Pang Z, Zheng X, Lu R, Li M, Zhuang J. Zero-Shot Learning for EEG Classification in Motor Imagery-Based BCI System. IEEE Trans Neural Syst Rehabil Eng 2020; 28:2411-2419. [PMID: 32986556 DOI: 10.1109/tnsre.2020.3027004] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
A brain-computer interface (BCI) based on motor imagery (MI) translates human intentions into computer commands by recognizing the electroencephalogram (EEG) patterns of different imagination tasks. However, due to the scarcity of MI commands and the long calibration time, using the MI-based BCI system in practice is still challenging. Zero-shot learning (ZSL), which can recognize objects whose instances may not have been seen during training, has the potential to substantially reduce the calibration time. Thus, in this context, we first try to use a new type of motor imagery task, which is a combination of traditional tasks and propose a novel zero-shot learning model that can recognize both known and unknown categories of EEG signals. This is achieved by first learning a non-linear projection from EEG features to the target space and then applying a novelty detection method to differentiate unknown classes from known classes. Applications to a dataset collected from nine subjects confirm the possibility of identifying a new type of motor imagery only using already obtained motor imagery data. Results indicate that the classification accuracy of our zero-shot based method accounts for 91.81% of the traditional method which uses all categories of data.
Collapse
|
18
|
Berezutskaya J, Freudenburg ZV, Ambrogioni L, Güçlü U, van Gerven MAJ, Ramsey NF. Cortical network responses map onto data-driven features that capture visual semantics of movie fragments. Sci Rep 2020; 10:12077. [PMID: 32694561 PMCID: PMC7374611 DOI: 10.1038/s41598-020-68853-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2020] [Accepted: 06/30/2020] [Indexed: 11/08/2022] Open
Abstract
Research on how the human brain extracts meaning from sensory input relies in principle on methodological reductionism. In the present study, we adopt a more holistic approach by modeling the cortical responses to semantic information that was extracted from the visual stream of a feature film, employing artificial neural network models. Advances in both computer vision and natural language processing were utilized to extract the semantic representations from the film by combining perceptual and linguistic information. We tested whether these representations were useful in studying the human brain data. To this end, we collected electrocorticography responses to a short movie from 37 subjects and fitted their cortical patterns across multiple regions using the semantic components extracted from film frames. We found that individual semantic components reflected fundamental semantic distinctions in the visual input, such as presence or absence of people, human movement, landscape scenes, human faces, etc. Moreover, each semantic component mapped onto a distinct functional cortical network involving high-level cognitive regions in occipitotemporal, frontal and parietal cortices. The present work demonstrates the potential of the data-driven methods from information processing fields to explain patterns of cortical responses, and contributes to the overall discussion about the encoding of high-level perceptual information in the human brain.
Collapse
Affiliation(s)
- Julia Berezutskaya
- Brain Center, Department of Neurology and Neurosurgery, University Medical Center Utrecht, Heidelberglaan 100, 3584 CX, Utrecht, The Netherlands.
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, 6525 HR, Nijmegen, The Netherlands.
| | - Zachary V Freudenburg
- Brain Center, Department of Neurology and Neurosurgery, University Medical Center Utrecht, Heidelberglaan 100, 3584 CX, Utrecht, The Netherlands
| | - Luca Ambrogioni
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, 6525 HR, Nijmegen, The Netherlands
| | - Umut Güçlü
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, 6525 HR, Nijmegen, The Netherlands
| | - Marcel A J van Gerven
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, 6525 HR, Nijmegen, The Netherlands
| | - Nick F Ramsey
- Brain Center, Department of Neurology and Neurosurgery, University Medical Center Utrecht, Heidelberglaan 100, 3584 CX, Utrecht, The Netherlands
| |
Collapse
|
19
|
Clarke A. Dynamic activity patterns in the anterior temporal lobe represents object semantics. Cogn Neurosci 2020; 11:111-121. [PMID: 32249714 PMCID: PMC7446031 DOI: 10.1080/17588928.2020.1742678] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2019] [Revised: 02/07/2020] [Indexed: 02/07/2023]
Abstract
The anterior temporal lobe (ATL) is considered a crucial area for the representation of transmodal concepts. Recent evidence suggests that specific regions within the ATL support the representation of individual object concepts, as shown by studies combining multivariate analysis methods and explicit measures of semantic knowledge. This research looks to further our understanding by probing conceptual representations at a spatially and temporally resolved neural scale. Representational similarity analysis was applied to human intracranial recordings from anatomically defined lateral to medial ATL sub-regions. Neural similarity patterns were tested against semantic similarity measures, where semantic similarity was defined by a hybrid corpus-based and feature-based approach. Analyses show that the perirhinal cortex, in the medial ATL, significantly related to semantic effects around 200 to 400 ms, and were greater than more lateral ATL regions. Further, semantic effects were present in low frequency (theta and alpha) oscillatory phase signals. These results provide converging support that more medial regions of the ATL support the representation of basic-level visual object concepts within the first 400 ms, and provide a bridge between prior fMRI and MEG work by offering detailed evidence for the presence of conceptual representations within the ATL.
Collapse
Affiliation(s)
- Alex Clarke
- Department of Psychology, University of Cambridge, Cambridge, UK
| |
Collapse
|
20
|
Semantic and perceptual priming activate partially overlapping brain networks as revealed by direct cortical recordings in humans. Neuroimage 2019; 203:116204. [DOI: 10.1016/j.neuroimage.2019.116204] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2018] [Revised: 08/19/2019] [Accepted: 09/17/2019] [Indexed: 01/20/2023] Open
|
21
|
Bruffaerts R, Tyler LK, Shafto M, Tsvetanov KA, Clarke A. Perceptual and conceptual processing of visual objects across the adult lifespan. Sci Rep 2019; 9:13771. [PMID: 31551468 PMCID: PMC6760174 DOI: 10.1038/s41598-019-50254-5] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2019] [Accepted: 09/02/2019] [Indexed: 12/24/2022] Open
Abstract
Making sense of the external world is vital for multiple domains of cognition, and so it is crucial that object recognition is maintained across the lifespan. We investigated age differences in perceptual and conceptual processing of visual objects in a population-derived sample of 85 healthy adults (24-87 years old) by relating measures of object processing to cognition across the lifespan. Magnetoencephalography (MEG) was recorded during a picture naming task to provide a direct measure of neural activity, that is not confounded by age-related vascular changes. Multiple linear regression was used to estimate neural responsivity for each individual, namely the capacity to represent visual or semantic information relating to the pictures. We find that the capacity to represent semantic information is linked to higher naming accuracy, a measure of task-specific performance. In mature adults, the capacity to represent semantic information also correlated with higher levels of fluid intelligence, reflecting domain-general performance. In contrast, the latency of visual processing did not relate to measures of cognition. These results indicate that neural responsivity measures relate to naming accuracy and fluid intelligence. We propose that maintaining neural responsivity in older age confers benefits in task-related and domain-general cognitive processes, supporting the brain maintenance view of healthy cognitive ageing.
Collapse
Affiliation(s)
- Rose Bruffaerts
- Department of Psychology, University of Cambridge, Cambridge, CB2 3EB, UK
- Laboratory for Cognitive Neurology, Department of Neurosciences, University of Leuven, 3000, Leuven, Belgium
- Neurology Department, University Hospitals Leuven, 3000, Leuven, Belgium
| | - Lorraine K Tyler
- Department of Psychology, University of Cambridge, Cambridge, CB2 3EB, UK.
- Cambridge Centre for Ageing and Neuroscience (Cam-CAN), University of Cambridge and MRC Cognition and Brain Sciences Unit, Cambridge, CB2 7EF, UK.
| | - Meredith Shafto
- Department of Psychology, University of Cambridge, Cambridge, CB2 3EB, UK
| | - Kamen A Tsvetanov
- Department of Psychology, University of Cambridge, Cambridge, CB2 3EB, UK
- Cambridge Centre for Ageing and Neuroscience (Cam-CAN), University of Cambridge and MRC Cognition and Brain Sciences Unit, Cambridge, CB2 7EF, UK
| | - Alex Clarke
- Department of Psychology, University of Cambridge, Cambridge, CB2 3EB, UK
| |
Collapse
|
22
|
Bruffaerts R, De Deyne S, Meersmans K, Liuzzi AG, Storms G, Vandenberghe R. Redefining the resolution of semantic knowledge in the brain: Advances made by the introduction of models of semantics in neuroimaging. Neurosci Biobehav Rev 2019; 103:3-13. [PMID: 31132379 DOI: 10.1016/j.neubiorev.2019.05.015] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2019] [Revised: 05/15/2019] [Accepted: 05/17/2019] [Indexed: 12/12/2022]
Abstract
The boundaries of our understanding of conceptual representation in the brain have been redrawn since the introduction of explicit models of semantics. These models are grounded in vast behavioural datasets acquired in healthy volunteers. Here, we review the most important techniques which have been applied to detect semantic information in neuroimaging data and argue why semantic models are possibly the most valuable addition to the research of semantics in recent years. Using multivariate analysis, predictions based on patient lesion data have been confirmed during semantic processing in healthy controls. Secondly, this new method has given rise to new research avenues, e.g. the detection of semantic processing outside of the temporal cortex. As a future line of work, the same research strategy could be useful to study neurological conditions such as the semantic variant of primary progressive aphasia, which is characterized by pathological semantic processing.
Collapse
Affiliation(s)
- Rose Bruffaerts
- Laboratory for Cognitive Neurology, Department of Neurosciences, KU Leuven, Belgium; Neurology Department, University Hospitals Leuven, 3000 Leuven, Belgium.
| | - Simon De Deyne
- Laboratory of Experimental Psychology, Humanities and Social Sciences Group, KU Leuven, Belgium
| | - Karen Meersmans
- Laboratory for Cognitive Neurology, Department of Neurosciences, KU Leuven, Belgium
| | | | - Gert Storms
- Laboratory of Experimental Psychology, Humanities and Social Sciences Group, KU Leuven, Belgium
| | - Rik Vandenberghe
- Laboratory for Cognitive Neurology, Department of Neurosciences, KU Leuven, Belgium; Neurology Department, University Hospitals Leuven, 3000 Leuven, Belgium
| |
Collapse
|
23
|
Rabbani Q, Milsap G, Crone NE. The Potential for a Speech Brain-Computer Interface Using Chronic Electrocorticography. Neurotherapeutics 2019; 16:144-165. [PMID: 30617653 PMCID: PMC6361062 DOI: 10.1007/s13311-018-00692-2] [Citation(s) in RCA: 31] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/30/2022] Open
Abstract
A brain-computer interface (BCI) is a technology that uses neural features to restore or augment the capabilities of its user. A BCI for speech would enable communication in real time via neural correlates of attempted or imagined speech. Such a technology would potentially restore communication and improve quality of life for locked-in patients and other patients with severe communication disorders. There have been many recent developments in neural decoders, neural feature extraction, and brain recording modalities facilitating BCI for the control of prosthetics and in automatic speech recognition (ASR). Indeed, ASR and related fields have developed significantly over the past years, and many lend many insights into the requirements, goals, and strategies for speech BCI. Neural speech decoding is a comparatively new field but has shown much promise with recent studies demonstrating semantic, auditory, and articulatory decoding using electrocorticography (ECoG) and other neural recording modalities. Because the neural representations for speech and language are widely distributed over cortical regions spanning the frontal, parietal, and temporal lobes, the mesoscopic scale of population activity captured by ECoG surface electrode arrays may have distinct advantages for speech BCI, in contrast to the advantages of microelectrode arrays for upper-limb BCI. Nevertheless, there remain many challenges for the translation of speech BCIs to clinical populations. This review discusses and outlines the current state-of-the-art for speech BCI and explores what a speech BCI using chronic ECoG might entail.
Collapse
Affiliation(s)
- Qinwan Rabbani
- Department of Electrical Engineering, The Johns Hopkins University Whiting School of Engineering, Baltimore, MD, USA.
| | - Griffin Milsap
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Nathan E Crone
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| |
Collapse
|
24
|
Ibayashi K, Kunii N, Matsuo T, Ishishita Y, Shimada S, Kawai K, Saito N. Decoding Speech With Integrated Hybrid Signals Recorded From the Human Ventral Motor Cortex. Front Neurosci 2018; 12:221. [PMID: 29674950 PMCID: PMC5895763 DOI: 10.3389/fnins.2018.00221] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2017] [Accepted: 03/20/2018] [Indexed: 11/13/2022] Open
Abstract
Restoration of speech communication for locked-in patients by means of brain computer interfaces (BCIs) is currently an important area of active research. Among the neural signals obtained from intracranial recordings, single/multi-unit activity (SUA/MUA), local field potential (LFP), and electrocorticography (ECoG) are good candidates for an input signal for BCIs. However, the question of which signal or which combination of the three signal modalities is best suited for decoding speech production remains unverified. In order to record SUA, LFP, and ECoG simultaneously from a highly localized area of human ventral sensorimotor cortex (vSMC), we fabricated an electrode the size of which was 7 by 13 mm containing sparsely arranged microneedle and conventional macro contacts. We determined which signal modality is the most capable of decoding speech production, and tested if the combination of these signals could improve the decoding accuracy of spoken phonemes. Feature vectors were constructed from spike frequency obtained from SUAs and event-related spectral perturbation derived from ECoG and LFP signals, then input to the decoder. The results showed that the decoding accuracy for five spoken vowels was highest when features from multiple signals were combined and optimized for each subject, and reached 59% when averaged across all six subjects. This result suggests that multi-scale signals convey complementary information for speech articulation. The current study demonstrated that simultaneous recording of multi-scale neuronal activities could raise decoding accuracy even though the recording area is limited to a small portion of cortex, which is advantageous for future implementation of speech-assisting BCIs.
Collapse
Affiliation(s)
- Kenji Ibayashi
- Department of Neurosurgery, The University of Tokyo Hospital, Tokyo, Japan
| | - Naoto Kunii
- Department of Neurosurgery, The University of Tokyo Hospital, Tokyo, Japan
| | - Takeshi Matsuo
- Department of Neurosurgery, Tokyo Metropolitan Neurological Hospital, Tokyo, Japan
| | - Yohei Ishishita
- Department of Neurosurgery, The University of Tokyo Hospital, Tokyo, Japan
| | - Seijiro Shimada
- Department of Neurosurgery, The University of Tokyo Hospital, Tokyo, Japan
| | - Kensuke Kawai
- Department of Neurosurgery, Jichi Medical University, Tochigi, Japan
| | - Nobuhito Saito
- Department of Neurosurgery, The University of Tokyo Hospital, Tokyo, Japan
| |
Collapse
|
25
|
Kapeller C, Ogawa H, Schalk G, Kunii N, Coon WG, Scharinger J, Guger C, Kamada K. Real-time detection and discrimination of visual perception using electrocorticographic signals. J Neural Eng 2018; 15:036001. [PMID: 29359711 DOI: 10.1088/1741-2552/aaa9f6] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/30/2023]
Abstract
OBJECTIVE Several neuroimaging studies have demonstrated that the ventral temporal cortex contains specialized regions that process visual stimuli. This study investigated the spatial and temporal dynamics of electrocorticographic (ECoG) responses to different types and colors of visual stimulation that were presented to four human participants, and demonstrated a real-time decoder that detects and discriminates responses to untrained natural images. APPROACH ECoG signals from the participants were recorded while they were shown colored and greyscale versions of seven types of visual stimuli (images of faces, objects, bodies, line drawings, digits, and kanji and hiragana characters), resulting in 14 classes for discrimination (experiment I). Additionally, a real-time system asynchronously classified ECoG responses to faces, kanji and black screens presented via a monitor (experiment II), or to natural scenes (i.e. the face of an experimenter, natural images of faces and kanji, and a mirror) (experiment III). Outcome measures in all experiments included the discrimination performance across types based on broadband γ activity. MAIN RESULTS Experiment I demonstrated an offline classification accuracy of 72.9% when discriminating among the seven types (without color separation). Further discrimination of grey versus colored images reached an accuracy of 67.1%. Discriminating all colors and types (14 classes) yielded an accuracy of 52.1%. In experiment II and III, the real-time decoder correctly detected 73.7% responses to face, kanji and black computer stimuli and 74.8% responses to presented natural scenes. SIGNIFICANCE Seven different types and their color information (either grey or color) could be detected and discriminated using broadband γ activity. Discrimination performance maximized for combined spatial-temporal information. The discrimination of stimulus color information provided the first ECoG-based evidence for color-related population-level cortical broadband γ responses in humans. Stimulus categories can be detected by their ECoG responses in real time within 500 ms with respect to stimulus onset.
Collapse
Affiliation(s)
- C Kapeller
- Guger Technologies OG, Graz, Austria. Department of Computational Perception, Johannes Kepler University, Linz, Austria
| | | | | | | | | | | | | | | |
Collapse
|
26
|
Caceres CA, Roos MJ, Rupp KM, Milsap G, Crone NE, Wolmetz ME, Ratto CR. Feature Selection Methods for Zero-Shot Learning of Neural Activity. Front Neuroinform 2017; 11:41. [PMID: 28690513 PMCID: PMC5481359 DOI: 10.3389/fninf.2017.00041] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2016] [Accepted: 06/07/2017] [Indexed: 11/13/2022] Open
Abstract
Dimensionality poses a serious challenge when making predictions from human neuroimaging data. Across imaging modalities, large pools of potential neural features (e.g., responses from particular voxels, electrodes, and temporal windows) have to be related to typically limited sets of stimuli and samples. In recent years, zero-shot prediction models have been introduced for mapping between neural signals and semantic attributes, which allows for classification of stimulus classes not explicitly included in the training set. While choices about feature selection can have a substantial impact when closed-set accuracy, open-set robustness, and runtime are competing design objectives, no systematic study of feature selection for these models has been reported. Instead, a relatively straightforward feature stability approach has been adopted and successfully applied across models and imaging modalities. To characterize the tradeoffs in feature selection for zero-shot learning, we compared correlation-based stability to several other feature selection techniques on comparable data sets from two distinct imaging modalities: functional Magnetic Resonance Imaging and Electrocorticography. While most of the feature selection methods resulted in similar zero-shot prediction accuracies and spatial/spectral patterns of selected features, there was one exception; A novel feature/attribute correlation approach was able to achieve those accuracies with far fewer features, suggesting the potential for simpler prediction models that yield high zero-shot classification accuracy.
Collapse
Affiliation(s)
- Carlos A Caceres
- Applied Physics Laboratory, Johns Hopkins UniversityLaurel, MD, United States
| | - Matthew J Roos
- Applied Physics Laboratory, Johns Hopkins UniversityLaurel, MD, United States
| | - Kyle M Rupp
- Department of Biomedical Engineering, Johns Hopkins UniversityBaltimore, MD, United States
| | - Griffin Milsap
- Department of Biomedical Engineering, Johns Hopkins UniversityBaltimore, MD, United States
| | - Nathan E Crone
- Department of Neurology, Johns Hopkins MedicineBaltimore, MD, United States
| | - Michael E Wolmetz
- Applied Physics Laboratory, Johns Hopkins UniversityLaurel, MD, United States
| | - Christopher R Ratto
- Applied Physics Laboratory, Johns Hopkins UniversityLaurel, MD, United States
| |
Collapse
|