1
|
Xu W, Liao P, Cao M, White DJ, Lyu B, Gao JH. Facilitating cognitive neuroscience research with 80-sensor optically pumped magnetometer magnetoencephalography (OPM-MEG). Neuroimage 2025; 311:121182. [PMID: 40180002 DOI: 10.1016/j.neuroimage.2025.121182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2024] [Revised: 02/28/2025] [Accepted: 03/31/2025] [Indexed: 04/05/2025] Open
Abstract
Recent advancements in optically pumped magnetometer magnetoencephalography (OPM-MEG) make it a promising alternative to conventional SQUID-MEG systems. Nonetheless, as reported in the literature, current OPM-MEG systems are often constrained by a limited number of sampling points, which restricts their capability to match the full-head coverage offered by SQUID-MEG systems. Additionally, whether OPM-MEG can deliver results comparable to SQUID-MEG in practical cognitive neuroscience applications remains largely unexplored. In this study, we introduce a high-density, full-head coverage OPM-MEG system with 80 sensors and systematically compare the performance of OPM-MEG and SQUID-MEG, from sensor- to source-level analysis, across various classic cognitive tasks. Our results demonstrate that visual and auditory evoked fields captured using OPM-MEG align closely with those obtained from SQUID-MEG. Furthermore, steady-state visual evoked field and finger-tapping-induced beta power change recorded with OPM-MEG are accurately localized to corresponding brain regions, with activation centers highly congruent to those observed with SQUID-MEG. For resting-state recordings, the two modalities exhibit similar power distributions, functional connectomes, and microstate clusters. These findings indicate that the 80-sensor OPM-MEG system provides spatial and temporal characteristics comparable to those of traditional SQUID-MEG. Thus, our study offers empirical evidence supporting the efficacy of high-density OPM-MEG and suggests that OPM-MEG, with dense sampling capability, represents a compelling alternative to conventional SQUID-MEG, facilitating further exploration of human cognition.
Collapse
Affiliation(s)
- Wei Xu
- Center for MRI Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing, 100871, China; Changping Laboratory, Beijing, 102206, China
| | - Pan Liao
- Center for MRI Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing, 100871, China; Changping Laboratory, Beijing, 102206, China
| | - Miao Cao
- Centre for Mental Health & Brain Sciences, Swinburne University of Technology, Hawthorn, VIC, 3122, Australia
| | - David J White
- Centre for Mental Health & Brain Sciences, Swinburne University of Technology, Hawthorn, VIC, 3122, Australia
| | | | - Jia-Hong Gao
- Center for MRI Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing, 100871, China; Changping Laboratory, Beijing, 102206, China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy Ion Physics, School of Physics, Peking University, Beijing, 100871, China; McGovern Institute for Brain Research, Peking University, Beijing, 100871, China; National Biomedical Imaging Center, Peking University, Beijing, 100871, China.
| |
Collapse
|
2
|
Sun Y, Chen X, Liu B, Liang L, Wang Y, Gao S, Gao X. Signal acquisition of brain-computer interfaces: A medical-engineering crossover perspective review. FUNDAMENTAL RESEARCH 2025; 5:3-16. [PMID: 40166113 PMCID: PMC11955058 DOI: 10.1016/j.fmre.2024.04.011] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2023] [Revised: 04/01/2024] [Accepted: 04/07/2024] [Indexed: 04/02/2025] Open
Abstract
Brain-computer interface (BCI) technology represents a burgeoning interdisciplinary domain that facilitates direct communication between individuals and external devices. The efficacy of BCI systems is largely contingent upon the progress in signal acquisition methodologies. This paper endeavors to provide an exhaustive synopsis of signal acquisition technologies within the realm of BCI by scrutinizing research publications from the last ten years. Our review synthesizes insights from both clinical and engineering viewpoints, delineating a comprehensive two-dimensional framework for understanding signal acquisition in BCIs. We delineate nine discrete categories of technologies, furnishing exemplars for each and delineating the salient challenges pertinent to these modalities. This review furnishes researchers and practitioners with a broad-spectrum comprehension of the signal acquisition landscape in BCI, and deliberates on the paramount issues presently confronting the field. Prospective enhancements in BCI signal acquisition should focus on harmonizing a multitude of disciplinary perspectives. Achieving equilibrium between signal fidelity, invasiveness, biocompatibility, and other pivotal considerations is imperative. By doing so, we can propel BCI technology forward, bolstering its effectiveness, safety, and dependability, thereby contributing to an auspicious future for human-technology integration.
Collapse
Affiliation(s)
- Yike Sun
- Department of Biomedical Engineering, Tsinghua University, Beijing 100084, China
| | - Xiaogang Chen
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China
| | - Bingchuan Liu
- Department of Biomedical Engineering, Tsinghua University, Beijing 100084, China
| | - Liyan Liang
- Center for Intellectual Property and Innovation Development, China Academy of Information and Communications Technology, Beijing 100161, China
| | - Yijun Wang
- Institute of Semiconductors, Chinese Academy of Sciences, Beijing 100083, China
| | - Shangkai Gao
- Department of Biomedical Engineering, Tsinghua University, Beijing 100084, China
| | - Xiaorong Gao
- Department of Biomedical Engineering, Tsinghua University, Beijing 100084, China
| |
Collapse
|
3
|
Tang C, Gao T, Wang G, Chen B. Coherence-based channel selection and Riemannian geometry features for magnetoencephalography decoding. Cogn Neurodyn 2024; 18:3535-3548. [PMID: 39712116 PMCID: PMC11655792 DOI: 10.1007/s11571-024-10085-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2023] [Revised: 01/03/2024] [Accepted: 02/02/2024] [Indexed: 12/24/2024] Open
Abstract
Magnetoencephalography (MEG) records the extremely weak magnetic fields on the surface of the scalp through highly sensitive sensors. Multi-channel MEG data provide higher spatial and temporal resolution when measuring brain activities, and can be applied for brain-computer interfaces as well. However, a large number of channels leads to high computational complexity and can potentially impact decoding accuracy. To improve the accuracy of MEG decoding, this paper proposes a new coherence-based channel selection method that effectively identifies task-relevant channels, reducing the presence of noisy and redundant information. Riemannian geometry is then used to extract effective features from selected channels of MEG data. Finally, MEG decoding is achieved by training a support vector machine classifier with the Radial Basis Function kernel. Experiments were conducted on two public MEG datasets to validate the effectiveness of the proposed method. The results from Dataset 1 show that Riemannian geometry achieves higher classification accuracy (compared to common spatial patterns and power spectral density) in the single-subject visual decoding task. Moreover, coherence-based channel selection significantly (P = 0.0002) outperforms the use of all channels. Moving on to Dataset 2, the results reveal that coherence-based channel selection is also significantly (P <0.0001) superior to all channels and channels around C3 and C4 in cross-session mental imagery decoding tasks. Additionally, the proposed method outperforms state-of-the-art methods in motor imagery tasks.
Collapse
Affiliation(s)
- Chao Tang
- National Key Laboratory of Human–Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an, 710049 China
| | - Tianyi Gao
- National Key Laboratory of Human–Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an, 710049 China
| | - Gang Wang
- Key Laboratory of Biomedical Information Engineering of Ministry of Education, Xi’an Jiaotong University, Xi’an, 710049 China
| | - Badong Chen
- National Key Laboratory of Human–Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an, 710049 China
| |
Collapse
|
4
|
Zhang M, Tang E, Ding H, Zhang Y. Artificial Intelligence and the Future of Communication Sciences and Disorders: A Bibliometric and Visualization Analysis. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:4369-4390. [PMID: 39418583 DOI: 10.1044/2024_jslhr-24-00157] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/19/2024]
Abstract
PURPOSE As artificial intelligence (AI) takes an increasingly prominent role in health care, a growing body of research is being dedicated to its application in the investigation of communication sciences and disorders (CSD). This study aims to provide a comprehensive overview, serving as a valuable resource for researchers, developers, and professionals seeking to comprehend the evolving landscape of AI in CSD research. METHOD We conducted a bibliometric analysis of AI-based research in the discipline of CSD published up to December 2023. Utilizing the Web of Science and Scopus databases, we identified 15,035 publications, with 4,375 meeting our inclusion criteria. Based on the bibliometric data, we examined publication trends and patterns, characteristics of research activities, and research hotspot tendencies. RESULTS From 1985 onwards, there has been a consistent annual increase in publications, averaging 16.51%, notably surging from 2012 to 2023. The primary communication disorders studied include autism, aphasia, dysarthria, Parkinson's disease, and Alzheimer's disease. Noteworthy AI models instantiated in CSD research encompass support vector machine, convolutional neural network, and hidden Markov model, among others. CONCLUSIONS Compared to AI applications in other fields, the adoption of AI in CSD has lagged slightly behind. While CSD studies primarily use classical machine learning techniques, there is a growing trend toward the integration of deep learning methods. AI technology offers significant benefits for both research and clinical practice in CSD, but it also presents certain challenges. Moving forward, collaboration among technological, research, and clinical domains is essential to empower researchers and speech-language pathologists to effectively leverage AI technology for the study, diagnosis, assessment, and rehabilitation of CSD. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.27162564.
Collapse
Affiliation(s)
- Minyue Zhang
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- National Research Centre for Language and Well-being, Shanghai, China
| | - Enze Tang
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, China
| | - Hongwei Ding
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- National Research Centre for Language and Well-being, Shanghai, China
| | - Yang Zhang
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Twin Cities, Minneapolis
- Masonic Institute for the Developing Brain, University of Minnesota, Twin Cities, Minneapolis
| |
Collapse
|
5
|
Rehman M, Anwer H, Garay H, Alemany-Iturriaga J, Díez IDLT, Siddiqui HUR, Ullah S. Decoding Brain Signals from Rapid-Event EEG for Visual Analysis Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2024; 24:6965. [PMID: 39517862 PMCID: PMC11548637 DOI: 10.3390/s24216965] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/13/2024] [Revised: 10/02/2024] [Accepted: 10/28/2024] [Indexed: 11/16/2024]
Abstract
The perception and recognition of objects around us empower environmental interaction. Harnessing the brain's signals to achieve this objective has consistently posed difficulties. Researchers are exploring whether the poor accuracy in this field is a result of the design of the temporal stimulation (block versus rapid event) or the inherent complexity of electroencephalogram (EEG) signals. Decoding perceptive signal responses in subjects has become increasingly complex due to high noise levels and the complex nature of brain activities. EEG signals have high temporal resolution and are non-stationary signals, i.e., their mean and variance vary overtime. This study aims to develop a deep learning model for the decoding of subjects' responses to rapid-event visual stimuli and highlights the major factors that contribute to low accuracy in the EEG visual classification task.The proposed multi-class, multi-channel model integrates feature fusion to handle complex, non-stationary signals. This model is applied to the largest publicly available EEG dataset for visual classification consisting of 40 object classes, with 1000 images in each class. Contemporary state-of-the-art studies in this area investigating a large number of object classes have achieved a maximum accuracy of 17.6%. In contrast, our approach, which integrates Multi-Class, Multi-Channel Feature Fusion (MCCFF), achieves a classification accuracy of 33.17% for 40 classes. These results demonstrate the potential of EEG signals in advancing EEG visual classification and offering potential for future applications in visual machine models.
Collapse
Affiliation(s)
- Madiha Rehman
- Institute of Computer Science, Khwaja Fareed University of Engineering & Information Technology, Rahim Yar Khan 64200, Pakistan; (H.A.); (H.u.R.S.); (S.U.)
| | - Humaira Anwer
- Institute of Computer Science, Khwaja Fareed University of Engineering & Information Technology, Rahim Yar Khan 64200, Pakistan; (H.A.); (H.u.R.S.); (S.U.)
| | - Helena Garay
- Universidad Europea del Atlantico, Isabel Torres 21, 39011 Santander, Spain;
- Universidade Internacional do Cuanza, Cuito EN 250, Angola
- Universidad de La Romana, Edificio G&G, C/Héctor René Gil, Esquina C/Francisco Castillo Marquez, La Romana 22000, Dominican Republic;
| | - Josep Alemany-Iturriaga
- Universidad de La Romana, Edificio G&G, C/Héctor René Gil, Esquina C/Francisco Castillo Marquez, La Romana 22000, Dominican Republic;
- Facultad de Ciencias Sociales y Humanidades, Universidad Europea del Atlántico, Isabel Torres 21, 39011 Santander, Spain
- Departamento de Ciencias de Lenguaje, Educación y Comunicaciones, Universidad Internacional Iberoamericana Arecibo, Arecibo, PR 00613, USA
| | - Isabel De la Torre Díez
- Department of Signal Theory, Communications and Telematics Engineering, University of Valladolid, 47011 Valladolid, Spain;
| | - Hafeez ur Rehman Siddiqui
- Institute of Computer Science, Khwaja Fareed University of Engineering & Information Technology, Rahim Yar Khan 64200, Pakistan; (H.A.); (H.u.R.S.); (S.U.)
| | - Saleem Ullah
- Institute of Computer Science, Khwaja Fareed University of Engineering & Information Technology, Rahim Yar Khan 64200, Pakistan; (H.A.); (H.u.R.S.); (S.U.)
| |
Collapse
|
6
|
Cheng H, He K, Li C, Ma X, Zheng F, Xu W, Liao P, Yang R, Li D, Qin L, Na S, Lyu B, Gao JH. Wireless optically pumped magnetometer MEG. Neuroimage 2024; 300:120864. [PMID: 39322096 DOI: 10.1016/j.neuroimage.2024.120864] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Revised: 09/04/2024] [Accepted: 09/23/2024] [Indexed: 09/27/2024] Open
Abstract
The current magnetoencephalography (MEG) systems, which rely on cables for control and signal transmission, do not fully realize the potential of wearable optically pumped magnetometers (OPM). This study presents a significant advancement in wireless OPM-MEG by reducing magnetization in the electronics and developing a tailored wireless communication protocol. Our protocol effectively eliminates electromagnetic interference, particularly in the critical frequency bands of MEG signals, and accurately synchronizes the acquisition and stimulation channels with the host computer's clock. We have successfully achieved single-channel wireless OPM-MEG measurement and demonstrated its reliability by replicating three well-established experiments: The alpha rhythm, auditory evoked field, and steady-state visual evoked field in the human brain. Our prototype wireless OPM-MEG system not only streamlines the measurement process but also represents a major step forward in the development of wearable OPM-MEG applications in both neuroscience and clinical research.
Collapse
Affiliation(s)
- Hao Cheng
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China
| | - Kaiyan He
- Changping Laboratory, Beijing 102206, PR China
| | - Congcong Li
- Changping Laboratory, Beijing 102206, PR China
| | - Xiao Ma
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy lon Physics, School of Physics, Peking University, Beijing 100871, PR China
| | - Fufu Zheng
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy lon Physics, School of Physics, Peking University, Beijing 100871, PR China
| | - Wei Xu
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China
| | - Pan Liao
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China
| | - Rui Yang
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy lon Physics, School of Physics, Peking University, Beijing 100871, PR China
| | - Dongxu Li
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy lon Physics, School of Physics, Peking University, Beijing 100871, PR China
| | - Lang Qin
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China
| | - Shuai Na
- National Biomedical Imaging Center, Peking University, Beijing 100871, PR China
| | | | - Jia-Hong Gao
- Center for MRl Research, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing 100871, PR China; Changping Laboratory, Beijing 102206, PR China; Beijing City Key Lab for Medical Physics and Engineering, Institution of Heavy lon Physics, School of Physics, Peking University, Beijing 100871, PR China; PKU-IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China; National Biomedical Imaging Center, Peking University, Beijing 100871, PR China; Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei 230088, PR China.
| |
Collapse
|
7
|
Wu H, Wang R, Ma Y, Liang X, Liu C, Yu D, An N, Ning X. Decoding N400m Evoked Component: A Tutorial on Multivariate Pattern Analysis for OP-MEG Data. Bioengineering (Basel) 2024; 11:609. [PMID: 38927845 PMCID: PMC11200846 DOI: 10.3390/bioengineering11060609] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2024] [Revised: 05/26/2024] [Accepted: 06/11/2024] [Indexed: 06/28/2024] Open
Abstract
Multivariate pattern analysis (MVPA) has played an extensive role in interpreting brain activity, which has been applied in studies with modalities such as functional Magnetic Resonance Imaging (fMRI), Magnetoencephalography (MEG) and Electroencephalography (EEG). The advent of wearable MEG systems based on optically pumped magnetometers (OPMs), i.e., OP-MEG, has broadened the application of bio-magnetism in the realm of neuroscience. Nonetheless, it also raises challenges in temporal decoding analysis due to the unique attributes of OP-MEG itself. The efficacy of decoding performance utilizing multimodal fusion, such as MEG-EEG, also remains to be elucidated. In this regard, we investigated the impact of several factors, such as processing methods, models and modalities, on the decoding outcomes of OP-MEG. Our findings indicate that the number of averaged trials, dimensionality reduction (DR) methods, and the number of cross-validation folds significantly affect the decoding performance of OP-MEG data. Additionally, decoding results vary across modalities and fusion strategy. In contrast, decoder type, resampling frequency, and sliding window length exert marginal effects. Furthermore, we introduced mutual information (MI) to investigate how information loss due to OP-MEG data processing affect decoding accuracy. Our study offers insights for linear decoding research using OP-MEG and expand its application in the fields of cognitive neuroscience.
Collapse
Affiliation(s)
- Huanqi Wu
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Ruonan Wang
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Yuyu Ma
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Xiaoyu Liang
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Changzeng Liu
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Dexin Yu
- Shandong Key Laboratory for Magnetic Field-Free Medicine and Functional Imaging, Institute of Magnetic Field-Free Medicine and Functional Imaging, Shandong University, 27 South Shanda Rd., Licheng Dist., Jinan 250100, China;
| | - Nan An
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
| | - Xiaolin Ning
- Key Laboratory of Ultra-Weak Magnetic Field Measurement Technology, Ministry of Education, School of Instrumentation and Optoelectronic Engineering, Beihang University, 37 Xueyuan Rd., Haidian Dist., Beijing 100083, China; (H.W.); (R.W.); (Y.M.); (X.L.); (C.L.)
- Hangzhou Institute of National Extremely-Weak Magnetic Field Infrastructure, 465 Binan Rd., Binjiang Dist., Hangzhou 310000, China
- Shandong Key Laboratory for Magnetic Field-Free Medicine and Functional Imaging, Institute of Magnetic Field-Free Medicine and Functional Imaging, Shandong University, 27 South Shanda Rd., Licheng Dist., Jinan 250100, China;
- Hefei National Laboratory, Gaoxin Dist., Hefei 230093, China
| |
Collapse
|
8
|
Anastasopoulou I, Cheyne DO, van Lieshout P, Johnson BW. Decoding kinematic information from beta-band motor rhythms of speech motor cortex: a methodological/analytic approach using concurrent speech movement tracking and magnetoencephalography. Front Hum Neurosci 2024; 18:1305058. [PMID: 38646159 PMCID: PMC11027130 DOI: 10.3389/fnhum.2024.1305058] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2023] [Accepted: 02/26/2024] [Indexed: 04/23/2024] Open
Abstract
Introduction Articulography and functional neuroimaging are two major tools for studying the neurobiology of speech production. Until now, however, it has generally not been feasible to use both in the same experimental setup because of technical incompatibilities between the two methodologies. Methods Here we describe results from a novel articulography system dubbed Magneto-articulography for the Assessment of Speech Kinematics (MASK), which is technically compatible with magnetoencephalography (MEG) brain scanning systems. In the present paper we describe our methodological and analytic approach for extracting brain motor activities related to key kinematic and coordination event parameters derived from time-registered MASK tracking measurements. Data were collected from 10 healthy adults with tracking coils on the tongue, lips, and jaw. Analyses targeted the gestural landmarks of reiterated utterances/ipa/ and /api/, produced at normal and faster rates. Results The results show that (1) Speech sensorimotor cortex can be reliably located in peri-rolandic regions of the left hemisphere; (2) mu (8-12 Hz) and beta band (13-30 Hz) neuromotor oscillations are present in the speech signals and contain information structures that are independent of those present in higher-frequency bands; and (3) hypotheses concerning the information content of speech motor rhythms can be systematically evaluated with multivariate pattern analytic techniques. Discussion These results show that MASK provides the capability, for deriving subject-specific articulatory parameters, based on well-established and robust motor control parameters, in the same experimental setup as the MEG brain recordings and in temporal and spatial co-register with the brain data. The analytic approach described here provides new capabilities for testing hypotheses concerning the types of kinematic information that are encoded and processed within specific components of the speech neuromotor system.
Collapse
Affiliation(s)
| | - Douglas Owen Cheyne
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
- Hospital for Sick Children Research Institute, Toronto, ON, Canada
| | - Pascal van Lieshout
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
| | | |
Collapse
|
9
|
Simistira Liwicki F, Gupta V, Saini R, De K, Abid N, Rakesh S, Wellington S, Wilson H, Liwicki M, Eriksson J. Bimodal electroencephalography-functional magnetic resonance imaging dataset for inner-speech recognition. Sci Data 2023; 10:378. [PMID: 37311807 DOI: 10.1038/s41597-023-02286-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Accepted: 06/01/2023] [Indexed: 06/15/2023] Open
Abstract
The recognition of inner speech, which could give a 'voice' to patients that have no ability to speak or move, is a challenge for brain-computer interfaces (BCIs). A shortcoming of the available datasets is that they do not combine modalities to increase the performance of inner speech recognition. Multimodal datasets of brain data enable the fusion of neuroimaging modalities with complimentary properties, such as the high spatial resolution of functional magnetic resonance imaging (fMRI) and the temporal resolution of electroencephalography (EEG), and therefore are promising for decoding inner speech. This paper presents the first publicly available bimodal dataset containing EEG and fMRI data acquired nonsimultaneously during inner-speech production. Data were obtained from four healthy, right-handed participants during an inner-speech task with words in either a social or numerical category. Each of the 8-word stimuli were assessed with 40 trials, resulting in 320 trials in each modality for each participant. The aim of this work is to provide a publicly available bimodal dataset on inner speech, contributing towards speech prostheses.
Collapse
Affiliation(s)
- Foteini Simistira Liwicki
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden.
| | - Vibha Gupta
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | - Rajkumar Saini
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | - Kanjar De
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | - Nosheen Abid
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | - Sumit Rakesh
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | | | - Holly Wilson
- University of Bath, Department of Computer Science, Bath, UK
| | - Marcus Liwicki
- Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Intelligent Systems LAB, Luleå, Sweden
| | - Johan Eriksson
- Umeå University, Department of Integrative Medical Biology (IMB) and Umeå Center for Functional Brain Imaging (UFBI), Umeå, Sweden
| |
Collapse
|
10
|
Ganesh A, Cervantes AJ, Kennedy PR. Slow Firing Single Units Are Essential for Optimal Decoding of Silent Speech. Front Hum Neurosci 2022; 16:874199. [PMID: 35992944 PMCID: PMC9382878 DOI: 10.3389/fnhum.2022.874199] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Accepted: 06/14/2022] [Indexed: 11/18/2022] Open
Abstract
The motivation of someone who is locked-in, that is, paralyzed and mute, is to find relief for their loss of function. The data presented in this report is part of an attempt to restore one of those lost functions, namely, speech. An essential feature of the development of a speech prosthesis is optimal decoding of patterns of recorded neural signals during silent or covert speech, that is, speaking “inside the head” with output that is inaudible due to the paralysis of the articulators. The aim of this paper is to illustrate the importance of both fast and slow single unit firings recorded from an individual with locked-in syndrome and from an intact participant speaking silently. Long duration electrodes were implanted in the motor speech cortex for up to 13 years in the locked-in participant. The data herein provide evidence that slow firing single units are essential for optimal decoding accuracy. Additional evidence indicates that slow firing single units can be conditioned in the locked-in participant 5 years after implantation, further supporting their role in decoding.
Collapse
Affiliation(s)
- Ananya Ganesh
- Neural Signals Inc., Neural Prostheses Laboratory, Duluth, GA, United States
| | | | - Philip R. Kennedy
- Neural Signals Inc., Neural Prostheses Laboratory, Duluth, GA, United States
- *Correspondence: Philip R. Kennedy
| |
Collapse
|
11
|
Simistira Liwicki F, Gupta V, Saini R, De K, Liwicki M. Rethinking the Methods and Algorithms for Inner Speech Decoding and Making Them Reproducible. NEUROSCI 2022; 3:226-244. [PMID: 39483370 PMCID: PMC11523721 DOI: 10.3390/neurosci3020017] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2022] [Accepted: 04/12/2022] [Indexed: 11/03/2024] Open
Abstract
This study focuses on the automatic decoding of inner speech using noninvasive methods, such as Electroencephalography (EEG). While inner speech has been a research topic in philosophy and psychology for half a century, recent attempts have been made to decode nonvoiced spoken words by using various brain-computer interfaces. The main shortcomings of existing work are reproducibility and the availability of data and code. In this work, we investigate various methods (using Convolutional Neural Network (CNN), Gated Recurrent Unit (GRU), Long Short-Term Memory Networks (LSTM)) for the detection task of five vowels and six words on a publicly available EEG dataset. The main contributions of this work are (1) subject dependent vs. subject-independent approaches, (2) the effect of different preprocessing steps (Independent Component Analysis (ICA), down-sampling and filtering), and (3) word classification (where we achieve state-of-the-art performance on a publicly available dataset). Overall we achieve a performance accuracy of 35.20% and 29.21% when classifying five vowels and six words, respectively, in a publicly available dataset, using our tuned iSpeech-CNN architecture. All of our code and processed data are publicly available to ensure reproducibility. As such, this work contributes to a deeper understanding and reproducibility of experiments in the area of inner speech detection.
Collapse
Affiliation(s)
- Foteini Simistira Liwicki
- Embedded Intelligent Systems LAB, Machine Learning, Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden; (V.G.); (R.S.); (K.D.); (M.L.)
| | - Vibha Gupta
- Embedded Intelligent Systems LAB, Machine Learning, Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden; (V.G.); (R.S.); (K.D.); (M.L.)
| | - Rajkumar Saini
- Embedded Intelligent Systems LAB, Machine Learning, Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden; (V.G.); (R.S.); (K.D.); (M.L.)
| | - Kanjar De
- Embedded Intelligent Systems LAB, Machine Learning, Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden; (V.G.); (R.S.); (K.D.); (M.L.)
| | - Marcus Liwicki
- Embedded Intelligent Systems LAB, Machine Learning, Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 97187 Luleå, Sweden; (V.G.); (R.S.); (K.D.); (M.L.)
| |
Collapse
|
12
|
Proix T, Delgado Saa J, Christen A, Martin S, Pasley BN, Knight RT, Tian X, Poeppel D, Doyle WK, Devinsky O, Arnal LH, Mégevand P, Giraud AL. Imagined speech can be decoded from low- and cross-frequency intracranial EEG features. Nat Commun 2022; 13:48. [PMID: 35013268 PMCID: PMC8748882 DOI: 10.1038/s41467-021-27725-3] [Citation(s) in RCA: 46] [Impact Index Per Article: 15.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2021] [Accepted: 12/03/2021] [Indexed: 01/19/2023] Open
Abstract
Reconstructing intended speech from neural activity using brain-computer interfaces holds great promises for people with severe speech production deficits. While decoding overt speech has progressed, decoding imagined speech has met limited success, mainly because the associated neural signals are weak and variable compared to overt speech, hence difficult to decode by learning algorithms. We obtained three electrocorticography datasets from 13 patients, with electrodes implanted for epilepsy evaluation, who performed overt and imagined speech production tasks. Based on recent theories of speech neural processing, we extracted consistent and specific neural features usable for future brain computer interfaces, and assessed their performance to discriminate speech items in articulatory, phonetic, and vocalic representation spaces. While high-frequency activity provided the best signal for overt speech, both low- and higher-frequency power and local cross-frequency contributed to imagined speech decoding, in particular in phonetic and vocalic, i.e. perceptual, spaces. These findings show that low-frequency power and cross-frequency dynamics contain key information for imagined speech decoding.
Collapse
Affiliation(s)
- Timothée Proix
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland.
| | - Jaime Delgado Saa
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland
| | - Andy Christen
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland
| | - Stephanie Martin
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland
| | - Brian N Pasley
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, USA
| | - Robert T Knight
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, USA
- Department of Psychology, University of California, Berkeley, Berkeley, USA
| | - Xing Tian
- Division of Arts and Sciences, New York University Shanghai, Shanghai, China
- Shanghai Key Laboratory of Brain Functional Genomics (Ministry of Education), School of Psychology and Cognitive Science, East China Normal University, Shanghai, China
- NYU-ECNU Institute of Brain and Cognitive Science at NYU Shanghai, Shanghai, China
| | - David Poeppel
- Department of Psychology, New York University, New York, NY, USA
- Ernst Strüngmann Institute for Neuroscience, Frankfurt, Germany
| | - Werner K Doyle
- Department of Neurology, New York University Grossman School of Medicine, New York, NY, USA
| | - Orrin Devinsky
- Department of Neurology, New York University Grossman School of Medicine, New York, NY, USA
| | - Luc H Arnal
- Institut de l'Audition, Institut Pasteur, INSERM, F-75012, Paris, France
| | - Pierre Mégevand
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland
- Division of Neurology, Geneva University Hospitals, Geneva, Switzerland
| | - Anne-Lise Giraud
- Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, Geneva, Switzerland
| |
Collapse
|
13
|
Dash D, Ferrari P, Babajani-Feremi A, Borna A, Schwindt PDD, Wang J. Magnetometers vs Gradiometers for Neural Speech Decoding. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:6543-6546. [PMID: 34892608 DOI: 10.1109/embc46164.2021.9630489] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Neural speech decoding aims at providing natural rate communication assistance to patients with locked-in state (e.g. due to amyotrophic lateral sclerosis, ALS) in contrast to the traditional brain-computer interface (BCI) spellers which are slow. Recent studies have shown that Magnetoencephalography (MEG) is a suitable neuroimaging modality to study neural speech decoding considering its excellent temporal resolution that can characterize the fast dynamics of speech. Gradiometers have been the preferred choice for sensor space analysis with MEG, due to their efficacy in noise suppression over magnetometers. However, recent development of optically pumped magnetometers (OPM) based wearable-MEG devices have shown great potential in future BCI applications, yet, no prior study has evaluated the performance of magnetometers in neural speech decoding. In this study, we decoded imagined and spoken speech from the MEG signals of seven healthy participants and compared the performance of magnetometers and gradiometers. Experimental results indicated that magnetometers also have the potential for neural speech decoding, although the performance was significantly lower than that obtained with gradiometers. Further, we implemented a wavelet based denoising strategy that improved the performance of both magnetometers and gradiometers significantly. These findings reconfirm that gradiometers are preferable in MEG based decoding analysis but also provide the possibility towards the use of magnetometers (or OPMs) for the development of the next-generation speech-BCIs.
Collapse
|
14
|
Lee DY, Lee M, Lee SW. Decoding Imagined Speech Based on Deep Metric Learning for Intuitive BCI Communication. IEEE Trans Neural Syst Rehabil Eng 2021; 29:1363-1374. [PMID: 34255630 DOI: 10.1109/tnsre.2021.3096874] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Imagined speech is a highly promising paradigm due to its intuitive application and multiclass scalability in the field of brain-computer interfaces. However, optimal feature extraction and classifiers have not yet been established. Furthermore, retraining still requires a large number of trials when new classes are added. The aim of this study is (i) to increase the classification performance for imagined speech and (ii) to apply a new class using a pretrained classifier with a small number of trials. We propose a novel framework based on deep metric learning that learns the distance by comparing the similarity between samples. We also applied the instantaneous frequency and spectral entropy used for speech signals to electroencephalography signals during imagined speech. The method was evaluated on two public datasets (6-class Coretto DB and 5-class BCI Competition DB). We achieved a 6-class accuracy of 45.00 ± 3.13% and a 5-class accuracy of 48.10 ± 3.68% using the proposed method, which significantly outperformed state-of-the-art methods. Additionally, we verified that the new class could be detected through incremental learning with a small number of trials. As a result, the average accuracy is 44.50 ± 0.26% for Coretto DB and 47.12 ± 0.27% for BCI Competition DB, which shows similar accuracy to baseline accuracy without incremental learning. Our results have shown that the accuracy can be greatly improved even with a small number of trials by selecting appropriate features from imagined speech. The proposed framework could be directly used to help construct an extensible intuitive communication system based on brain-computer interfaces.
Collapse
|
15
|
Panachakel JT, Ramakrishnan AG. Decoding Covert Speech From EEG-A Comprehensive Review. Front Neurosci 2021; 15:642251. [PMID: 33994922 PMCID: PMC8116487 DOI: 10.3389/fnins.2021.642251] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2020] [Accepted: 03/18/2021] [Indexed: 11/13/2022] Open
Abstract
Over the past decade, many researchers have come up with different implementations of systems for decoding covert or imagined speech from EEG (electroencephalogram). They differ from each other in several aspects, from data acquisition to machine learning algorithms, due to which, a comparison between different implementations is often difficult. This review article puts together all the relevant works published in the last decade on decoding imagined speech from EEG into a single framework. Every important aspect of designing such a system, such as selection of words to be imagined, number of electrodes to be recorded, temporal and spatial filtering, feature extraction and classifier are reviewed. This helps a researcher to compare the relative merits and demerits of the different approaches and choose the one that is most optimal. Speech being the most natural form of communication which human beings acquire even without formal education, imagined speech is an ideal choice of prompt for evoking brain activity patterns for a BCI (brain-computer interface) system, although the research on developing real-time (online) speech imagery based BCI systems is still in its infancy. Covert speech based BCI can help people with disabilities to improve their quality of life. It can also be used for covert communication in environments that do not support vocal communication. This paper also discusses some future directions, which will aid the deployment of speech imagery based BCI for practical applications, rather than only for laboratory experiments.
Collapse
Affiliation(s)
- Jerrin Thomas Panachakel
- Medical Intelligence and Language Engineering Laboratory, Department of Electrical Engineering, Indian Institute of Science, Bangalore, India
| | | |
Collapse
|