1
|
Liu R, Hu Y, Wu J, Wong KC, Huang ZA, Huang YA, Chen Tan K. Dynamic Graph Representation Learning for Spatio-Temporal Neuroimaging Analysis. IEEE TRANSACTIONS ON CYBERNETICS 2025; 55:1121-1134. [PMID: 40031724 DOI: 10.1109/tcyb.2025.3531657] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Neuroimaging analysis aims to reveal the information-processing mechanisms of the human brain in a noninvasive manner. In the past, graph neural networks (GNNs) have shown promise in capturing the non-Euclidean structure of brain networks. However, existing neuroimaging studies focused primarily on spatial functional connectivity, despite temporal dynamics in complex brain networks. To address this gap, we propose a spatio-temporal interactive graph representation framework (STIGR) for dynamic neuroimaging analysis that encompasses different aspects from classification and regression tasks to interpretation tasks. STIGR leverages a dynamic adaptive-neighbor graph convolution network to capture the interrelationships between spatial and temporal dynamics. To address the limited global scope in graph convolutions, a self-attention module based on Transformers is introduced to extract long-term dependencies. Contrastive learning is used to adaptively contrast similarities between adjacent scanning windows, modeling cross-temporal correlations in dynamic graphs. Extensive experiments on six public neuroimaging datasets demonstrate the competitive performance of STIGR across different platforms, achieving state-of-the-art results in classification and regression tasks. The proposed framework enables the detection of remarkable temporal association patterns between regions of interest based on sequential neuroimaging signals, offering medical professionals a versatile and interpretable tool for exploring task-specific neurological patterns. Our codes and models are available at https://github.com/77YQ77/STIGR/.
Collapse
|
2
|
Wu C, Yao B, Zhang X, Li T, Wang J, Pu J. The Application of Entropy in Motor Imagery Paradigms of Brain-Computer Interfaces. Brain Sci 2025; 15:168. [PMID: 40002501 PMCID: PMC11853529 DOI: 10.3390/brainsci15020168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2024] [Revised: 02/05/2025] [Accepted: 02/06/2025] [Indexed: 02/27/2025] Open
Abstract
Background: In motor imagery brain-computer interface (MI-BCI) research, electroencephalogram (EEG) signals are complex and nonlinear. This complexity and nonlinearity render signal processing and classification challenging when employing traditional linear methods. Information entropy, with its intrinsic nonlinear characteristics, effectively captures the dynamic behavior of EEG signals, thereby addressing the limitations of traditional methods in capturing linear features. However, the multitude of entropy types leads to unclear application scenarios, with a lack of systematic descriptions. Methods: This study conducted a review of 63 high-quality research articles focused on the application of entropy in MI-BCI, published between 2019 and 2023. It summarizes the names, functions, and application scopes of 13 commonly used entropy measures. Results: The findings indicate that sample entropy (16.3%), Shannon entropy (13%), fuzzy entropy (12%), permutation entropy (9.8%), and approximate entropy (7.6%) are the most frequently utilized entropy features in MI-BCI. The majority of studies employ a single entropy feature (79.7%), with dual entropy (9.4%) and triple entropy (4.7%) being the most prevalent combinations in multiple entropy applications. The incorporation of entropy features can significantly enhance pattern classification accuracy (by 8-10%). Most studies (67%) utilize public datasets for classification verification, while a minority design and conduct experiments (28%), and only 5% combine both methods. Conclusions: Future research should delve into the effects of various entropy features on specific problems to clarify their application scenarios. As research methodologies continue to evolve and advance, entropy features are poised to play a significant role in a wide array of fields and contexts.
Collapse
Affiliation(s)
- Chengzhen Wu
- School of Life Sciences, Tiangong University, Tianjin 300387, China;
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China; (B.Y.); (X.Z.); (T.L.)
| | - Bo Yao
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China; (B.Y.); (X.Z.); (T.L.)
- Tianjin Key Laboratory of Neuromodulation and Neurorepair, Tianjin 300192, China
| | - Xin Zhang
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China; (B.Y.); (X.Z.); (T.L.)
| | - Ting Li
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China; (B.Y.); (X.Z.); (T.L.)
| | - Jinhai Wang
- School of Life Sciences, Tiangong University, Tianjin 300387, China;
| | - Jiangbo Pu
- Institute of Biomedical Engineering, Chinese Academy of Medical Sciences and Peking Union Medical College, Tianjin 300192, China; (B.Y.); (X.Z.); (T.L.)
- Tianjin Key Laboratory of Neuromodulation and Neurorepair, Tianjin 300192, China
| |
Collapse
|
3
|
Nareklishvili M, Geitle M. Deep Ensemble Transformers for Dimensionality Reduction. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2091-2102. [PMID: 38294917 DOI: 10.1109/tnnls.2024.3357621] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/02/2024]
Abstract
We propose deep ensemble transformers (DETs), a fast, scalable approach for dimensionality reduction problems. This method leverages the power of deep neural networks and employs cascade ensemble techniques as its fundamental feature extraction tool. To handle high-dimensional data, our approach employs a flexible number of intermediate layers sequentially. These layers progressively transform the input data into decision tree predictions. To further enhance prediction performance, the output from the final intermediate layer is fed through a feed-forward neural network architecture for final prediction. We derive an upper bound of the disparity between the generalization error and the empirical error and demonstrate that it converges to zero. This highlights the generalizability of our method to parameter estimation and feature selection problems. In our experimental evaluations, DETs outperform existing models in terms of prediction accuracy, representation learning ability, and computational time. Specifically, the method achieves over 95% accuracy in gene expression data and can be trained on average 50% faster than traditional artificial neural networks (ANNs).
Collapse
|
4
|
Cui W, Xiang Y, Wang Y, Yu T, Liao XF, Hu B, Li Y. Deep Multiview Module Adaption Transfer Network for Subject-Specific EEG Recognition. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2917-2930. [PMID: 38252578 DOI: 10.1109/tnnls.2024.3350085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/24/2024]
Abstract
Transfer learning is one of the popular methods to solve the problem of insufficient data in subject-specific electroencephalogram (EEG) recognition tasks. However, most existing approaches ignore the difference between subjects and transfer the same feature representations from source domain to different target domains, resulting in poor transfer performance. To address this issue, we propose a novel subject-specific EEG recognition method named deep multiview module adaption transfer (DMV-MAT) network. First, we design a universal deep multiview (DMV) network to generate different types of discriminative features from multiple perspectives, which improves the generalization performance by extensive feature sets. Second, module adaption transfer (MAT) is designed to evaluate each module by the feature distributions of source and target samples, which can generate an optimal weight sharing strategy for each target subject and promote the model to learn domain-invariant and domain-specific features simultaneously. We conduct extensive experiments in two EEG recognition tasks, i.e., motor imagery (MI) and seizure prediction, on four datasets. Experimental results demonstrate that the proposed method achieves promising performance compared with the state-of-the-art methods, indicating a feasible solution for subject-specific EEG recognition tasks. Implementation codes are available at https://github.com/YangLibuaa/DMV-MAT.
Collapse
|
5
|
Gu H, Chen T, Ma X, Zhang M, Sun Y, Zhao J. CLTNet: A Hybrid Deep Learning Model for Motor Imagery Classification. Brain Sci 2025; 15:124. [PMID: 40002457 PMCID: PMC11852626 DOI: 10.3390/brainsci15020124] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2024] [Revised: 01/15/2025] [Accepted: 01/24/2025] [Indexed: 02/27/2025] Open
Abstract
BACKGROUND Brain-computer interface (BCI) technology opens up new avenues for human-machine interaction and rehabilitation by connecting the brain to machines. Electroencephalography (EEG)-based motor imagery (MI) classification is a key component of BCI technology, which is capable of translating neural activity in the brain into commands for controlling external devices. Despite the great potential of BCI technology, the challenges of extracting and decoding brain signals limit its wide application. METHODS To address this challenge, this study proposes a novel hybrid deep learning model, CLTNet, which focuses on solving the feature extraction problem to improve the classification of MI-EEG signals. In the preliminary feature extraction stage, CLTNet uses a convolutional neural network (CNN) to extract time series, channel, and spatial features of EEG signals to obtain important local information. In the deep feature extraction stage, the model combines the long short-term memory (LSTM) network and the Transformer module to capture time-series data and global dependencies in the EEG. The LSTM explains the dynamics of the brain activity, while the Transformer's self-attention mechanism reveals the global features of the time series. Ultimately, the CLTNet model classifies motor imagery EEG signals through a fully connected layer. RESULTS The model achieved an average accuracy of 83.02% and a Kappa value of 0.77 on the BCI IV 2a dataset, and 87.11% and a Kappa value of 0.74 on the BCI IV 2b dataset, both of which outperformed the traditional methods. CONCLUSIONS The innovation of the CLTNet model is that it integrates multiple network architectures, which offers a more comprehensive understanding of the characteristics of the EEG signals during motor imagery, providing a more comprehensive perspective and establishing a new benchmark for future research in this area.
Collapse
Affiliation(s)
- He Gu
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
| | - Tingwei Chen
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
| | - Xiao Ma
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
| | - Mengyuan Zhang
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
| | - Yan Sun
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
| | - Jian Zhao
- College of Computer Science and Technology, Changchun University, Changchun 130022, China; (H.G.); (T.C.); (X.M.); (M.Z.); (Y.S.)
- Key Laboratory of Intelligent Rehabilitation and Barrier-Free Access for the Disabled, Ministry of Education, Changchun 130022, China
- Jilin Provincial Key Laboratory of Human Health State Identification and Function Enhancement, Changchun 130022, China
- Jilin Rehabilitation Equipment and Technology Engineering Research Center for the Disabled, Changchun 130022, China
| |
Collapse
|
6
|
Li X, Yang Z, Tu X, Wang J, Huang J. MFRC-Net: Multi-Scale Feature Residual Convolutional Neural Network for Motor Imagery Decoding. IEEE J Biomed Health Inform 2025; 29:224-234. [PMID: 39316474 DOI: 10.1109/jbhi.2024.3467090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/26/2024]
Abstract
Motor imagery (MI) decoding is the basis of external device control via electroencephalogram (EEG). However, the majority of studies prioritize enhancing the accuracy of decoding methods, often overlooking the magnitude and computational resource demands of deep learning models. In this study, we propose a novel lightweight Multi-Scale Feature Residual Convolutional Neural Network (MFRC-Net). MFRC-Net primarily consists of two blocks: temporal multi-scale residual convolution blocks and cross-domain dual-stream spatial convolution blocks. The former captures dynamic changes in EEG signals across various time scales through multi-scale grouped convolution and backbone temporal convolution skip connections; the latter improves local spatial feature extraction and calibrates feature mapping through the introduction of cross-domain spatial filtering layers. Furthermore, by specifically optimizing the loss function, MFRC-Net effectively reduces sensitivity to outliers. Experiment results on the BCI Competition IV 2a dataset and the SHU dataset demonstrate that, with a parameter size of only 13 K, MFRC-Net achieves accuracy of 85.1% and 69.3%, respectively, surpassing current state-of-the-art models. The integration of temporal multi-scale residual convolution blocks and cross-domain dual-stream spatial convolution blocks in lightweight models significantly boosts performance, as evidenced by ablation studies and visualizations.
Collapse
|
7
|
Zhi H, Yu T, Gu Z, Lin Z, Che L, Li Y, Yu Z. Supervised Contrastive Learning-Based Domain Generalization Network for Cross-Subject Motor Decoding. IEEE Trans Biomed Eng 2025; 72:401-412. [PMID: 39046861 DOI: 10.1109/tbme.2024.3432934] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/27/2024]
Abstract
Developing an electroencephalogram (EEG)-based motor imagery and motor execution (MI/ME) decoding system that is both highly accurate and calibration-free for cross-subject applications remains challenging due to domain shift problem inherent in such scenario. Recent research has increasingly embraced transfer learning strategies, especially domain adaptation techniques. However, domain adaptation becomes impractical when the target subject data is either difficult to obtain or unknown. To address this issue, we propose a supervised contrastive learning-based domain generalization network (SCLDGN) for cross-subject MI/ME decoding. Firstly, the feature encoder is purposefully designed to learn the EEG discriminative feature representations. Secondly, the domain alignment based on deep correlation alignment constrains the representations distance across various domains to learn domain-invariant features. In addition, the class regularization block is proposed, where the supervised contrastive learning with domain-agnostic mixup is established to learn the class-relevant features and achieve class-level alignment. Finally, in the latent space, clusters of domain-agnostic representations from the same class are mapped closer together. Consequently, SCLDGN is capable of learning domain-invariant and class-relevant discriminative representations, which are essential for effective cross-subject decoding. Extensive experiments conducted on six MI/ME datasets demonstrate the effectiveness of the proposed method in comparison with other state-of-the-art approaches. Furthermore, ablation study and visualization analyses explain the generalization mechanism of the proposed method and also show neurophysiologically meaningful patterns related to MI/ME.
Collapse
|
8
|
Wang H, Qi Y, Yao L, Wang Y, Farina D, Pan G. A Human-Machine Joint Learning Framework to Boost Endogenous BCI Training. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:17534-17548. [PMID: 37647178 DOI: 10.1109/tnnls.2023.3305621] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/01/2023]
Abstract
Brain-computer interfaces (BCIs) provide a direct pathway from the brain to external devices and have demonstrated great potential for assistive and rehabilitation technologies. Endogenous BCIs based on electroencephalogram (EEG) signals, such as motor imagery (MI) BCIs, can provide some level of control. However, mastering spontaneous BCI control requires the users to generate discriminative and stable brain signal patterns by imagery, which is challenging and is usually achieved over a very long training time (weeks/months). Here, we propose a human-machine joint learning framework to boost the learning process in endogenous BCIs, by guiding the user to generate brain signals toward an optimal distribution estimated by the decoder, given the historical brain signals of the user. To this end, we first model the human-machine joint learning process in a uniform formulation. Then a human-machine joint learning framework is proposed: 1) for the human side, we model the learning process in a sequential trial-and-error scenario and propose a novel "copy/new" feedback paradigm to help shape the signal generation of the subject toward the optimal distribution and 2) for the machine side, we propose a novel adaptive learning algorithm to learn an optimal signal distribution along with the subject's learning process. Specifically, the decoder reweighs the brain signals generated by the subject to focus more on "good" samples to cope with the learning process of the subject. Online and psuedo-online BCI experiments with 18 healthy subjects demonstrated the advantages of the proposed joint learning process over coadaptive approaches in both learning efficiency and effectiveness.
Collapse
|
9
|
Cai S, Li P, Li H. A Bio-Inspired Spiking Attentional Neural Network for Attentional Selection in the Listening Brain. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:17387-17397. [PMID: 37585329 DOI: 10.1109/tnnls.2023.3303308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/18/2023]
Abstract
Humans show a remarkable ability in solving the cocktail party problem. Decoding auditory attention from the brain signals is a major step toward the development of bionic ears emulating human capabilities. Electroencephalography (EEG)-based auditory attention detection (AAD) has attracted considerable interest recently. Despite much progress, the performance of traditional AAD decoders remains to be improved, especially in low-latency settings. State-of-the-art AAD decoders based on deep neural networks generally lack the intrinsic temporal coding ability in biological networks. In this study, we first propose a bio-inspired spiking attentional neural network, denoted as BSAnet, for decoding auditory attention. BSAnet is capable of exploiting the temporal dynamics of EEG signals using biologically plausible neurons and an attentional mechanism. Experiments on two publicly available datasets confirm the superior performance of BSAnet over other state-of-the-art systems across various evaluation conditions. Moreover, BSAnet imitates realistic brain-like information processing, through which we show the advantage of brain-inspired computational models.
Collapse
|
10
|
Chowdhury RS, Bose S, Ghosh S, Konar A. Attention Induced Dual Convolutional-Capsule Network (AIDC-CN): A deep learning framework for motor imagery classification. Comput Biol Med 2024; 183:109260. [PMID: 39426071 DOI: 10.1016/j.compbiomed.2024.109260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Revised: 09/08/2024] [Accepted: 10/06/2024] [Indexed: 10/21/2024]
Abstract
In recent times, Electroencephalography (EEG)-based motor imagery (MI) decoding has garnered significant attention due to its extensive applicability in healthcare, including areas such as assistive robotics and rehabilitation engineering. Nevertheless, the decoding of EEG signals presents considerable challenges owing to their inherent complexity, non-stationary characteristics, and low signal-to-noise ratio. Notably, deep learning-based classifiers have emerged as a prominent focus for addressing the EEG signal decoding process. This study introduces a novel deep learning classifier named the Attention Induced Dual Convolutional-Capsule Network (AIDC-CN) with the specific aim of accurately categorizing various motor imagination class labels. To enhance the classifier's performance, a dual feature extraction approach leveraging spectrogram and brain connectivity networks has been employed, diversifying the feature set in the classification task. The main highlights of the proposed AIDC-CN classifier includes the introduction of a dual convolution layer to handle the brain connectivity and spectrogram features, addition of a novel self-attention module (SAM) to accentuate the relevant parts of the convolved spectrogram features, introduction of a new cross-attention module (CAM) to refine the outputs obtained from the dual convolution layers and incorporation of a Gaussian Error Linear Unit (GELU) based dynamic routing algorithm to strengthen the coupling among the primary and secondary capsule layers. Performance analysis undertaken on four public data sets depict the superior performance of the proposed model with respect to the state-of-the-art techniques. The code for this model is available at https://github.com/RiteshSurChowdhury/AIDC-CN.
Collapse
Affiliation(s)
- Ritesh Sur Chowdhury
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India
| | - Shirsha Bose
- Department of Informatics, Technical University of Munich, Munich, Bavaria 85748, Germany
| | - Sayantani Ghosh
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India
| | - Amit Konar
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India.
| |
Collapse
|
11
|
An S, Kim S, Chikontwe P, Park SH. Dual Attention Relation Network With Fine-Tuning for Few-Shot EEG Motor Imagery Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:15479-15493. [PMID: 37379192 DOI: 10.1109/tnnls.2023.3287181] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/30/2023]
Abstract
Recently, motor imagery (MI) electroencephalography (EEG) classification techniques using deep learning have shown improved performance over conventional techniques. However, improving the classification accuracy on unseen subjects is still challenging due to intersubject variability, scarcity of labeled unseen subject data, and low signal-to-noise ratio (SNR). In this context, we propose a novel two-way few-shot network able to efficiently learn how to learn representative features of unseen subject categories and classify them with limited MI EEG data. The pipeline includes an embedding module that learns feature representations from a set of signals, a temporal-attention module to emphasize important temporal features, an aggregation-attention module for key support signal discovery, and a relation module for final classification based on relation scores between a support set and a query signal. In addition to the unified learning of feature similarity and a few-shot classifier, our method can emphasize informative features in support data relevant to the query, which generalizes better on unseen subjects. Furthermore, we propose to fine-tune the model before testing by arbitrarily sampling a query signal from the provided support set to adapt to the distribution of the unseen subject. We evaluate our proposed method with three different embedding modules on cross-subject and cross-dataset classification tasks using brain-computer interface (BCI) competition IV 2a, 2b, and GIST datasets. Extensive experiments show that our model significantly improves over the baselines and outperforms existing few-shot approaches.
Collapse
|
12
|
Keutayeva A, Fakhrutdinov N, Abibullaev B. Compact convolutional transformer for subject-independent motor imagery EEG-based BCIs. Sci Rep 2024; 14:25775. [PMID: 39468119 PMCID: PMC11519587 DOI: 10.1038/s41598-024-73755-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 09/20/2024] [Indexed: 10/30/2024] Open
Abstract
Motor imagery electroencephalography (EEG) analysis is crucial for the development of effective brain-computer interfaces (BCIs), yet it presents considerable challenges due to the complexity of the data and inter-subject variability. This paper introduces EEGCCT, an application of compact convolutional transformers designed specifically to improve the analysis of motor imagery tasks in EEG. Unlike traditional approaches, EEGCCT model significantly enhances generalization from limited data, effectively addressing a common limitation in EEG datasets. We validate and test our models using the open-source BCI Competition IV datasets 2a and 2b, employing a Leave-One-Subject-Out (LOSO) strategy to ensure subject-independent performance. Our findings demonstrate that EEGCCT not only outperforms conventional models like EEGNet in standard evaluations but also achieves better performance compared to other advanced models such as Conformer, Hybrid s-CViT, and Hybrid t-CViT, while utilizing fewer parameters and achieving an accuracy of 70.12%. Additionally, the paper presents a comprehensive ablation study that includes targeted data augmentation, hyperparameter optimization, and architectural improvements.
Collapse
Affiliation(s)
- Aigerim Keutayeva
- Institute of Smart Systems and Artificial Intelligence (ISSAI), Nazarbayev University, Astana, 010000, Kazakhstan.
| | - Nail Fakhrutdinov
- Department of Computer Science, Nazarbayev University, Astana, 010000, Kazakhstan
| | - Berdakh Abibullaev
- Department of Robotics Engineering, Nazarbayev University, Astana, 010000, Kazakhstan
| |
Collapse
|
13
|
Liao L, Lu J, Wang L, Zhang Y, Gao D, Wang M. CT-Net: an interpretable CNN-Transformer fusion network for fNIRS classification. Med Biol Eng Comput 2024; 62:3233-3247. [PMID: 38816665 DOI: 10.1007/s11517-024-03138-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 05/23/2024] [Indexed: 06/01/2024]
Abstract
Functional near-infrared spectroscopy (fNIRS), an optical neuroimaging technique, has been widely used in the field of brain activity recognition and brain-computer interface. Existing works have proposed deep learning-based algorithms for the fNIRS classification problem. In this paper, a novel approach based on convolutional neural network and Transformer, named CT-Net, is established to guide the deep modeling for the classification of mental arithmetic (MA) tasks. We explore the effect of data representations, and design a temporal-level combination of two raw chromophore signals to improve the data utilization and enrich the feature learning of the model. We evaluate our model on two open-access datasets and achieve the classification accuracy of 98.05% and 77.61%, respectively. Moreover, we explain our model by the gradient-weighted class activation mapping, which presents a high consistent between the contributing value of features learned by the model and the mapping of brain activity in the MA task. The results suggest the feasibility and interpretability of CT-Net for decoding MA tasks.
Collapse
Affiliation(s)
- Lingxiang Liao
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Jingqing Lu
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Lutao Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Yongqing Zhang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Dongrui Gao
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Manqing Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China.
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, 611731, China.
| |
Collapse
|
14
|
Xia X, Shi Y, Li P, Liu X, Liu J, Men H. FBANet: An Effective Data Mining Method for Food Olfactory EEG Recognition. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:13550-13560. [PMID: 37220050 DOI: 10.1109/tnnls.2023.3269949] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
At present, the sensory evaluation of food mostly depends on artificial sensory evaluation and machine perception, but artificial sensory evaluation is greatly interfered with by subjective factors, and machine perception is difficult to reflect human feelings. In this article, a frequency band attention network (FBANet) for olfactory electroencephalogram (EEG) was proposed to distinguish the difference in food odor. First, the olfactory EEG evoked experiment was designed to collect the olfactory EEG, and the preprocessing of olfactory EEG, such as frequency division, was completed. Second, the FBANet consisted of frequency band feature mining and frequency band feature self-attention, in which frequency band feature mining can effectively mine multiband features of olfactory EEG with different scales, and frequency band feature self-attention can integrate the extracted multiband features and realize classification. Finally, compared with other advanced models, the performance of the FBANet was evaluated. The results show that FBANet was better than the state-of-the-art techniques. In conclusion, FBANet effectively mined the olfactory EEG data information and distinguished the differences between the eight food odors, which proposed a new idea for food sensory evaluation based on multiband olfactory EEG analysis.
Collapse
|
15
|
Li D, Wang J, Xu J, Fang X, Ji Y. Cross-Channel Specific-Mutual Feature Transfer Learning for Motor Imagery EEG Signals Decoding. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:13472-13482. [PMID: 37220058 DOI: 10.1109/tnnls.2023.3269512] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
In recent years, with the rapid development of deep learning, various deep learning frameworks have been widely used in brain-computer interface (BCI) research for decoding motor imagery (MI) electroencephalogram (EEG) signals to understand brain activity accurately. The electrodes, however, record the mixed activities of neurons. If different features are directly embedded in the same feature space, the specific and mutual features of different neuron regions are not considered, which will reduce the expression ability of the feature itself. We propose a cross-channel specific-mutual feature transfer learning (CCSM-FT) network model to solve this problem. The multibranch network extracts the specific and mutual features of brain's multiregion signals. Effective training tricks are used to maximize the distinction between the two kinds of features. Suitable training tricks can also improve the effectiveness of the algorithm compared with novel models. Finally, we transfer two kinds of features to explore the potential of mutual and specific features to enhance the expressive power of the feature and use the auxiliary set to improve identification performance. The experimental results show that the network has a better classification effect in the BCI Competition IV-2a and the HGD datasets.
Collapse
|
16
|
Mahalungkar SP, Shrivastava R, Angadi S. A brief survey on human activity recognition using motor imagery of EEG signals. Electromagn Biol Med 2024; 43:312-327. [PMID: 39425602 DOI: 10.1080/15368378.2024.2415089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 10/06/2024] [Indexed: 10/21/2024]
Abstract
Human being's biological processes and psychological activities are jointly connected to the brain. So, the examination of human activity is more significant for the well-being of humans. There are various models for brain activity detection considering neuroimaging for attaining decreased time requirement, increased control commands, and enhanced accuracy. Motor Imagery (MI)-based Brain-Computer Interface (BCI) systems create a way in which the brain can interact with the environment by processing Electroencephalogram (EEG) signals. Human Activity Recognition (HAR) deals with identifying the physiological activities of human beings based on sensory signals. This survey reviews the different methods available for HAR based on MI-EEG signals. A total of 50 research articles based on HAR from EEG signals are considered in this survey. This survey discusses the challenges faced by various techniques for HAR. Moreover, the papers are assessed considering various parameters, techniques, publication year, performance metrics, utilized tools, employed databases, etc. There were many techniques developed to solve the problem of HAR and they are classified as Machine Learning (ML) and Deep Learning (DL)models. At last, the research gaps and limitations of the techniques were discussed that contribute to developing an effective HAR.
Collapse
Affiliation(s)
- Seema Pankaj Mahalungkar
- Department of Computer Science and Engineering, Mansarovar Global University, Bhopal, Madhya Pradesh, India
- Computer Science and Engineering, Nutan College of Engineering and Research, Talegaon Dabhade, Pune, India
| | - Rahul Shrivastava
- School of Computer Science and Engineering, VIT Bhopal University, Bhopal, Madhya Pradesh, India
| | - Sanjeevkumar Angadi
- Computer Science and Engineering, Nutan College of Engineering and Research, Talegaon Dabhade, Pune, India
| |
Collapse
|
17
|
Lee MH, Shomanov A, Begim B, Kabidenova Z, Nyssanbay A, Yazici A, Lee SW. EAV: EEG-Audio-Video Dataset for Emotion Recognition in Conversational Contexts. Sci Data 2024; 11:1026. [PMID: 39300129 PMCID: PMC11413008 DOI: 10.1038/s41597-024-03838-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 08/29/2024] [Indexed: 09/22/2024] Open
Abstract
Understanding emotional states is pivotal for the development of next-generation human-machine interfaces. Human behaviors in social interactions have resulted in psycho-physiological processes influenced by perceptual inputs. Therefore, efforts to comprehend brain functions and human behavior could potentially catalyze the development of AI models with human-like attributes. In this study, we introduce a multimodal emotion dataset comprising data from 30-channel electroencephalography (EEG), audio, and video recordings from 42 participants. Each participant engaged in a cue-based conversation scenario, eliciting five distinct emotions: neutral, anger, happiness, sadness, and calmness. Throughout the experiment, each participant contributed 200 interactions, which encompassed both listening and speaking. This resulted in a cumulative total of 8,400 interactions across all participants. We evaluated the baseline performance of emotion recognition for each modality using established deep neural network (DNN) methods. The Emotion in EEG-Audio-Visual (EAV) dataset represents the first public dataset to incorporate three primary modalities for emotion recognition within a conversational context. We anticipate that this dataset will make significant contributions to the modeling of the human emotional process, encompassing both fundamental neuroscience and machine learning viewpoints.
Collapse
Affiliation(s)
- Min-Ho Lee
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adai Shomanov
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Balgyn Begim
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Zhuldyz Kabidenova
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Aruna Nyssanbay
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adnan Yazici
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Seong-Whan Lee
- Korea University, Department of Artificial Intelligence, Seoul, 02841, Republic of Korea.
| |
Collapse
|
18
|
Sultana M, Perdikis S. Automatic Feature Selection for Sensorimotor Rhythms Brain-Computer Interface Fusing Expert and Data-Driven Knowledge. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3422-3431. [PMID: 39250352 DOI: 10.1109/tnsre.2024.3456591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/11/2024]
Abstract
Early brain-computer interface (BCI) systems were mainly based on prior neurophysiological knowledge coupled with feedback training, while state-of-the-art interfaces rely on data-driven, machine learning (ML)-oriented methods. Despite the advances in BCI that ML can be credited with, the performance of BCI solutions is still not up to the mark, posing a major barrier to the widespread use of this technology. This paper proposes a novel, automatic feature selection method for BCI able to leverage both data-dependent and expert knowledge to suppress noisy features and highlight the most relevant ones thanks to a fuzzy logic (FL) system. Our approach exploits the capability of FL to increase the reliability of decision-making by fusing heterogeneous information channels while maintaining transparency and simplicity. We show that our method leads to significant improvement in classification accuracy, feature stability and class bias when applied to large motor imagery or attempt datasets including end-users with motor disabilities. We postulate that combining data-driven methods with knowledge derived from neuroscience literature through FL can enhance the performance, explainability, and learnability of BCIs.
Collapse
|
19
|
Rong F, Yang B, Guan C. Decoding Multi-Class Motor Imagery From Unilateral Limbs Using EEG Signals. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3399-3409. [PMID: 39236133 DOI: 10.1109/tnsre.2024.3454088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/07/2024]
Abstract
The EEG is a widely utilized neural signal source, particularly in motor imagery-based brain-computer interface (MI-BCI), offering distinct advantages in applications like stroke rehabilitation. Current research predominantly concentrates on the bilateral limbs paradigm and decoding, but the use scenarios for stroke rehabilitation are typically for unilateral upper limbs. There is a significant challenge to decoding unilateral MI of multitasks due to the overlapped spatial neural activities of the tasks. This study aims to formulate a novel MI-BCI experimental paradigm for unilateral limbs with multitasks. The paradigm encompasses four imagined movement directions: top-bottom, left-right, top right-bottom left, and top left-bottom right. Forty-six healthy subjects participated in this experiment. Commonly used machine learning techniques, such as FBCSP, EEGNet, deepConvNet, and FBCNet, were employed for evaluation. To improve decoding accuracy, we propose an MVCA method that introduces temporal convolution and attention mechanism to effectively capture temporal features from multiple perspectives. With the MVCA model, we have achieved 40.6% and 64.89% classification accuracies for the four-class and two-class scenarios (top right-bottom left and top left-bottom right), respectively. Conclusion: This is the first study demonstrating that motor imagery of multiple directions in unilateral limbs can be decoded. In particular, decoding two directions, right top to left bottom and left top to right bottom, provides the best accuracy, which sheds light on future studies. This study advances the development of the MI-BCI paradigm, offering preliminary evidence for the feasibility of decoding multiple directional information from EEG. This, in turn, enhances the dimensions of MI control commands.
Collapse
|
20
|
Rajpura P, Cecotti H, Kumar Meena Y. Explainable artificial intelligence approaches for brain-computer interfaces: a review and design space. J Neural Eng 2024; 21:041003. [PMID: 39029500 DOI: 10.1088/1741-2552/ad6593] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Accepted: 07/19/2024] [Indexed: 07/21/2024]
Abstract
Objective.This review paper provides an integrated perspective of Explainable Artificial Intelligence (XAI) techniques applied to Brain-Computer Interfaces (BCIs). BCIs use predictive models to interpret brain signals for various high-stake applications. However, achieving explainability in these complex models is challenging as it compromises accuracy. Trust in these models can be established by incorporating reasoning or causal relationships from domain experts. The field of XAI has emerged to address the need for explainability across various stakeholders, but there is a lack of an integrated perspective in XAI for BCI (XAI4BCI) literature. It is necessary to differentiate key concepts like explainability, interpretability, and understanding, often used interchangeably in this context, and formulate a comprehensive framework.Approach.To understand the need of XAI for BCI, we pose six key research questions for a systematic review and meta-analysis, encompassing its purposes, applications, usability, and technical feasibility. We employ the PRISMA methodology-preferred reporting items for systematic reviews and meta-analyses to review (n = 1246) and analyse (n = 84) studies published in 2015 and onwards for key insights.Main results.The results highlight that current research primarily focuses on interpretability for developers and researchers, aiming to justify outcomes and enhance model performance. We discuss the unique approaches, advantages, and limitations of XAI4BCI from the literature. We draw insights from philosophy, psychology, and social sciences. We propose a design space for XAI4BCI, considering the evolving need to visualise and investigate predictive model outcomes customised for various stakeholders in the BCI development and deployment lifecycle.Significance.This paper is the first to focus solely on reviewing XAI4BCI research articles. This systematic review and meta-analysis findings with the proposed design space prompt important discussions on establishing standards for BCI explanations, highlighting current limitations, and guiding the future of XAI in BCI.
Collapse
Affiliation(s)
- Param Rajpura
- Human-AI Interaction (HAIx) Lab, Indian Institute of Technology Gandhinagar, Gandhinagar, India
| | - Hubert Cecotti
- Department of Computer Science, California State University, Fresno, CA, United States of America
| | - Yogesh Kumar Meena
- Human-AI Interaction (HAIx) Lab, Indian Institute of Technology Gandhinagar, Gandhinagar, India
| |
Collapse
|
21
|
Luo Y, Mu W, Wang L, Wang J, Wang P, Gan Z, Zhang L, Kang X. An EEG channel selection method for motor imagery based on Fisher score and local optimization. J Neural Eng 2024; 21:036030. [PMID: 38842111 DOI: 10.1088/1741-2552/ad504a] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 05/24/2024] [Indexed: 06/07/2024]
Abstract
Objective. Multi-channel electroencephalogram (EEG) technology in brain-computer interface (BCI) research offers the advantage of enhanced spatial resolution and system performance. However, this also implies that more time is needed in the data processing stage, which is not conducive to the rapid response of BCI. Hence, it is a necessary and challenging task to reduce the number of EEG channels while maintaining decoding effectiveness.Approach. In this paper, we propose a local optimization method based on the Fisher score for within-subject EEG channel selection. Initially, we extract the common spatial pattern characteristics of EEG signals in different bands, calculate Fisher scores for each channel based on these characteristics, and rank them accordingly. Subsequently, we employ a local optimization method to finalize the channel selection.Main results. On the BCI Competition IV Dataset IIa, our method selects an average of 11 channels across four bands, achieving an average accuracy of 79.37%. This represents a 6.52% improvement compared to using the full set of 22 channels. On our self-collected dataset, our method similarly achieves a significant improvement of 24.20% with less than half of the channels, resulting in an average accuracy of 76.95%.Significance. This research explores the importance of channel combinations in channel selection tasks and reveals that appropriately combining channels can further enhance the quality of channel selection. The results indicate that the model selected a small number of channels with higher accuracy in two-class motor imagery EEG classification tasks. Additionally, it improves the portability of BCI systems through channel selection and combinations, offering the potential for the development of portable BCI systems.
Collapse
Affiliation(s)
- Yangjie Luo
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Wei Mu
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Lu Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Junkongshuai Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Pengchao Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Zhongxue Gan
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
| | - Lihua Zhang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
| | - Xiaoyang Kang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
- Yiwu Research Institute of Fudan University, Yiwu City, People's Republic of China
- Research Center for Intelligent Sensing, Zhejiang Lab, Hangzhou, People's Republic of China
| |
Collapse
|
22
|
Miao M, Yang Z, Sheng Z, Xu B, Zhang W, Cheng X. Multi-source deep domain adaptation ensemble framework for cross-dataset motor imagery EEG transfer learning. Physiol Meas 2024; 45:055024. [PMID: 38772402 DOI: 10.1088/1361-6579/ad4e95] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2023] [Accepted: 05/21/2024] [Indexed: 05/23/2024]
Abstract
Objective. Electroencephalography (EEG) is an important kind of bioelectric signal for measuring physiological activities of the brain, and motor imagery (MI) EEG has significant clinical application prospects. Convolutional neural network has become a mainstream algorithm for MI EEG classification, however lack of subject-specific data considerably restricts its decoding accuracy and generalization performance. To address this challenge, a novel transfer learning (TL) framework using auxiliary dataset to improve the MI EEG classification performance of target subject is proposed in this paper.Approach. We developed a multi-source deep domain adaptation ensemble framework (MSDDAEF) for cross-dataset MI EEG decoding. The proposed MSDDAEF comprises three main components: model pre-training, deep domain adaptation, and multi-source ensemble. Moreover, for each component, different designs were examined to verify the robustness of MSDDAEF.Main results. Bidirectional validation experiments were performed on two large public MI EEG datasets (openBMI and GIST). The highest average classification accuracy of MSDDAEF reaches 74.28% when openBMI serves as target dataset and GIST serves as source dataset. While the highest average classification accuracy of MSDDAEF is 69.85% when GIST serves as target dataset and openBMI serves as source dataset. In addition, the classification performance of MSDDAEF surpasses several well-established studies and state-of-the-art algorithms.Significance. The results of this study show that cross-dataset TL is feasible for left/right-hand MI EEG decoding, and further indicate that MSDDAEF is a promising solution for addressing MI EEG cross-dataset variability.
Collapse
Affiliation(s)
- Minmin Miao
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| | - Zhong Yang
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
| | - Zhenzhen Sheng
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| | - Baoguo Xu
- School of Instrument Science and Engineering, Southeast University, Nanjing, People's Republic of China
| | - Wenbin Zhang
- College of Computer Science and Software Engineering, Hohai University, Nanjing, Jiangsu Province, People's Republic of China
| | - Xinmin Cheng
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| |
Collapse
|
23
|
Akuthota S, K R, Ravichander J. Artifact removal and motor imagery classification in EEG using advanced algorithms and modified DNN. Heliyon 2024; 10:e27198. [PMID: 38560190 PMCID: PMC10980936 DOI: 10.1016/j.heliyon.2024.e27198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Revised: 02/21/2024] [Accepted: 02/26/2024] [Indexed: 04/04/2024] Open
Abstract
This paper presents an advanced approach for EEG artifact removal and motor imagery classification using a combination of Four Class Iterative Filtering and Filter Bank Common Spatial Pattern Algorithm with a Modified Deep Neural Network (DNN) classifier. The research aims to enhance the accuracy and reliability of BCI systems by addressing the challenges posed by EEG artifacts and complex motor imagery tasks. The methodology begins by introducing FCIF, a novel technique for ocular artifact removal, utilizing iterative filtering and filter banks. FCIF's mathematical formulation allows for effective artifact mitigation, thereby improving the quality of EEG data. In tandem, the FC-FBCSP algorithm is introduced, extending the Filter Bank Common Spatial Pattern approach to handle four-class motor imagery classification. The Modified DNN classifier enhances the discriminatory power of the FC-FBCSP features, optimizing the classification process. The paper showcases a comprehensive experimental setup, featuring the utilization of BCI Competition IV Dataset 2a & 2b. Detailed preprocessing steps, including filtering and feature extraction, are presented with mathematical rigor. Results demonstrate the remarkable artifact removal capabilities of FCIF and the classification prowess of FC-FBCSP combined with the Modified DNN classifier. Comparative analysis highlights the superiority of the proposed approach over baseline methods and the method achieves the mean accuracy of 98.575%.
Collapse
Affiliation(s)
- Srinath Akuthota
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| | - RajKumar K
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| | - Janapati Ravichander
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| |
Collapse
|
24
|
Li W, Li H, Sun X, Kang H, An S, Wang G, Gao Z. Self-supervised contrastive learning for EEG-based cross-subject motor imagery recognition. J Neural Eng 2024; 21:026038. [PMID: 38565100 DOI: 10.1088/1741-2552/ad3986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 03/20/2024] [Indexed: 04/04/2024]
Abstract
Objective. The extensive application of electroencephalography (EEG) in brain-computer interfaces (BCIs) can be attributed to its non-invasive nature and capability to offer high-resolution data. The acquisition of EEG signals is a straightforward process, but the datasets associated with these signals frequently exhibit data scarcity and require substantial resources for proper labeling. Furthermore, there is a significant limitation in the generalization performance of EEG models due to the substantial inter-individual variability observed in EEG signals.Approach. To address these issues, we propose a novel self-supervised contrastive learning framework for decoding motor imagery (MI) signals in cross-subject scenarios. Specifically, we design an encoder combining convolutional neural network and attention mechanism. In the contrastive learning training stage, the network undergoes training with the pretext task of data augmentation to minimize the distance between pairs of homologous transformations while simultaneously maximizing the distance between pairs of heterologous transformations. It enhances the amount of data utilized for training and improves the network's ability to extract deep features from original signals without relying on the true labels of the data.Main results. To evaluate our framework's efficacy, we conduct extensive experiments on three public MI datasets: BCI IV IIa, BCI IV IIb, and HGD datasets. The proposed method achieves cross-subject classification accuracies of 67.32%, 82.34%, and 81.13%on the three datasets, demonstrating superior performance compared to existing methods.Significance. Therefore, this method has great promise for improving the performance of cross-subject transfer learning in MI-based BCI systems.
Collapse
Affiliation(s)
- Wenjie Li
- Tianjin International Engineering Institute, Tianjin University, Tianjin 300072, People's Republic of China
| | - Haoyu Li
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Xinlin Sun
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Huicong Kang
- Department of Neurology, Shanxi Bethune Hospital, Shanxi Academy of Medical Science, Tongji Shanxi Hospital, Third Hospital of Shanxi Medical University, Taiyuan 030000, People's Republic of China
- Department of Neurology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430000, People's Republic of China
| | - Shan An
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Guoxin Wang
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Zhongke Gao
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| |
Collapse
|
25
|
Srimadumathi V, Ramasubba Reddy M. Classification of Motor Imagery EEG signals using high resolution time-frequency representations and convolutional neural network. Biomed Phys Eng Express 2024; 10:035025. [PMID: 38513274 DOI: 10.1088/2057-1976/ad3647] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2023] [Accepted: 03/21/2024] [Indexed: 03/23/2024]
Abstract
A Motor Imagery (MI) based Brain Computer Interface (BCI) system aims to provide neuro-rehabilitation for the motor disabled people and patients with brain injuries (e.g., stroke patients) etc. The aim of this work is to classify the left and right hand MI tasks by utilizing the occurrence of event related desynchronization and synchronization (ERD\ERS) in the Electroencephalogram (EEG) during these tasks. This study proposes to use a set of Complex Morlet Wavelets (CMW) having frequency dependent widths to generate high-resolution time-frequency representations (TFR) of the MI EEG signals present in the channels C3 and C4. A novel method for the selection of the value of number of cycles relative to the center frequency of the CMW is studied here for extracting the MI task features. The generated TFRs are given as input to a Convolutional neural network (CNN) for classifying them into left or right hand MI tasks. The proposed framework attains a classification accuracy of 82.2% on the BCI Competition IV dataset 2a, showing that the TFRs generated in this work give a higher classification accuracy than the baseline methods and other existing algorithms.
Collapse
Affiliation(s)
- V Srimadumathi
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology, Madras, 600036, India
| | - M Ramasubba Reddy
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology, Madras, 600036, India
| |
Collapse
|
26
|
Han J, Gu X, Yang GZ, Lo B. Noise-Factorized Disentangled Representation Learning for Generalizable Motor Imagery EEG Classification. IEEE J Biomed Health Inform 2024; 28:765-776. [PMID: 38010934 DOI: 10.1109/jbhi.2023.3337072] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
Motor Imagery (MI) Electroencephalography (EEG) is one of the most common Brain-Computer Interface (BCI) paradigms that has been widely used in neural rehabilitation and gaming. Although considerable research efforts have been dedicated to developing MI EEG classification algorithms, they are mostly limited in handling scenarios where the training and testing data are not from the same subject or session. Such poor generalization capability significantly limits the realization of BCI in real-world applications. In this paper, we proposed a novel framework to disentangle the representation of raw EEG data into three components, subject/session-specific, MI-task-specific, and random noises, so that the subject/session-specific feature extends the generalization capability of the system. This is realized by a joint discriminative and generative framework, supported by a series of fundamental training losses and training strategies. We evaluated our framework on three public MI EEG datasets, and detailed experimental results show that our method can achieve superior performance by a large margin compared to current state-of-the-art benchmark algorithms.
Collapse
|
27
|
Park D, Park H, Kim S, Choo S, Lee S, Nam CS, Jung JY. Spatio-Temporal Explanation of 3D-EEGNet for Motor Imagery EEG Classification Using Permutation and Saliency. IEEE Trans Neural Syst Rehabil Eng 2023; 31:4504-4513. [PMID: 37934650 DOI: 10.1109/tnsre.2023.3330922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2023]
Abstract
Recently, convolutional neural network (CNN)-based classification models have shown good performance for motor imagery (MI) brain-computer interfaces (BCI) using electroencephalogram (EEG) in end-to-end learning. Although a few explainable artificial intelligence (XAI) techniques have been developed, it is still challenging to interpret the CNN models for EEG-based BCI classification effectively. In this research, we propose 3D-EEGNet as a 3D CNN model to improve both the explainability and performance of MI EEG classification. The proposed approach exhibited better performances on two MI EEG datasets than the existing EEGNet, which uses a 2D input shape. The MI classification accuracies are improved around 1.8% and 6.1% point in average on the datasets, respectively. The permutation-based XAI method is first applied for the reliable explanation of the 3D-EEGNet. Next, to find a faster XAI method for spatio-temporal explanation, we design a novel technique based on the normalized discounted cumulative gain (NDCG) for selecting the best among a few saliency-based methods due to their higher time complexity than the permutation-based method. Among the saliency-based methods, DeepLIFT was selected because the NDCG scores indicated its results are the most similar to the permutation-based results. Finally, the fast spatio-temporal explanation using DeepLIFT provides deeper understanding for the classification results of the 3D-EEGNet and the important properties in the MI EEG experiments.
Collapse
|
28
|
Deny P, Cheon S, Son H, Choi KW. Hierarchical Transformer for Motor Imagery-Based Brain Computer Interface. IEEE J Biomed Health Inform 2023; 27:5459-5470. [PMID: 37578918 DOI: 10.1109/jbhi.2023.3304646] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2023]
Abstract
In this paper, we propose a novel transformer-based classification algorithm for the brain computer interface (BCI) using a motor imagery (MI) electroencephalogram (EEG) signal. To design the MI classification algorithm, we apply an up-to-date deep learning model, the transformer, that has revolutionized the natural language processing (NLP) and successfully widened its application to many other domains such as the computer vision. Within a long MI trial spanning a few seconds, the classification algorithm should give more attention to the time periods during which the intended motor task is imagined by the subject without any artifact. To achieve this goal, we propose a hierarchical transformer architecture that consists of a high-level transformer (HLT) and a low-level transformer (LLT). We break down a long MI trial into a number of short-term intervals. The LLT extracts a feature from each short-term interval, and the HLT pays more attention to the features from more relevant short-term intervals by using the self-attention mechanism of the transformer. We have done extensive tests of the proposed scheme on four open MI datasets, and shown that the proposed hierarchical transformer excels in both the subject-dependent and subject-independent tests.
Collapse
|
29
|
Guo L, Yu T, Zhao S, Li X, Liao X, Li Y. CLEP: Contrastive Learning for Epileptic Seizure Prediction Using a Spatio-Temporal-Spectral Network. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3915-3926. [PMID: 37796668 DOI: 10.1109/tnsre.2023.3322275] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/07/2023]
Abstract
Seizure prediction of epileptic preictal period through electroencephalogram (EEG) signals is important for clinical epilepsy diagnosis. However, recent deep learning-based methods commonly employ intra-subject training strategy and need sufficient data, which are laborious and time-consuming for a practical system and pose a great challenge for seizure predicting. Besides, multi-domain characterizations, including spatio-temporal-spectral dependencies in an epileptic brain are generally neglected or not considered simultaneously in current approaches, and this insufficiency commonly leads to suboptimal seizure prediction performance. To tackle the above issues, in this paper, we propose Contrastive Learning for Epileptic seizure Prediction (CLEP) using a Spatio-Temporal-Spectral Network (STS-Net). Specifically, the CLEP learns intrinsic epileptic EEG patterns across subjects by contrastive learning. The STS-Net extracts multi-scale temporal and spectral representations under different rhythms from raw EEG signals. Then, a novel triple attention layer (TAL) is employed to construct inter-dimensional interaction among multi-domain features. Moreover, a spatio dynamic graph convolution network (sdGCN) is proposed to dynamically model the spatial relationships between electrodes and aggregate spatial information. The proposed CLEP-STS-Net achieves a sensitivity of 96.7% and a false prediction rate of 0.072/h on the CHB-MIT scalp EEG database. We also validate the proposed method on clinical intracranial EEG (iEEG) database from our Xuanwu Hospital of Capital Medical University, and the predicting system yielded a sensitivity of 95%, a false prediction rate of 0.087/h. The experimental results outperform the state-of-the-art studies which validate the efficacy of our method. Our code is available at https://github.com/LianghuiGuo/CLEP-STS-Net.
Collapse
|
30
|
Zhang D, Li H, Xie J, Li D. MI-DAGSC: A domain adaptation approach incorporating comprehensive information from MI-EEG signals. Neural Netw 2023; 167:183-198. [PMID: 37659115 DOI: 10.1016/j.neunet.2023.08.008] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Revised: 07/24/2023] [Accepted: 08/06/2023] [Indexed: 09/04/2023]
Abstract
Non-stationarity of EEG signals leads to high variability between subjects, making it challenging to directly use data from other subjects (source domain) for the classifier in the current subject (target domain). In this study, we propose MI-DAGSC to address domain adaptation challenges in EEG-based motor imagery (MI) decoding. By combining domain-level information, class-level information, and inter-sample structure information, our model effectively aligns the feature distributions of source and target domains. This work is an extension of our previous domain adaptation work MI-DABAN (Li et al., 2023). Based on MI-DABAN, MI-DAGSC designs Sample-Feature Blocks (SFBs) and Graph Convolution Blocks (GCBs) to focus on intra-sample and inter-sample information. The synergistic integration of SFBs and GCBs enable the model to capture comprehensive information and understand the relationship between samples, thus improving representation learning. Furthermore, we introduce a triplet loss to enhance the alignment and compactness of feature representations. Extensive experiments on real EEG datasets demonstrate the effectiveness of MI-DAGSC, confirming that our method makes a valuable contribution to the MI-EEG decoding. Moreover, it holds great potential for various applications in brain-computer interface systems and neuroscience research. And the code of the proposed architecture in this study is available under https://github.com/zhangdx21/MI-DAGSC.
Collapse
Affiliation(s)
- Dongxue Zhang
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China.
| | - Huiying Li
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China.
| | - Jingmeng Xie
- Xi'an Jiaotong University, College of Electronic information, Xi'an, Shanxi Province, China.
| | - Dajun Li
- Jilin Provincial People's Hospital, Changchun, Jilin Province, China
| |
Collapse
|
31
|
Cui J, Yuan L, Wang Z, Li R, Jiang T. Towards best practice of interpreting deep learning models for EEG-based brain computer interfaces. Front Comput Neurosci 2023; 17:1232925. [PMID: 37663037 PMCID: PMC10470463 DOI: 10.3389/fncom.2023.1232925] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 07/24/2023] [Indexed: 09/05/2023] Open
Abstract
Introduction As deep learning has achieved state-of-the-art performance for many tasks of EEG-based BCI, many efforts have been made in recent years trying to understand what have been learned by the models. This is commonly done by generating a heatmap indicating to which extent each pixel of the input contributes to the final classification for a trained model. Despite the wide use, it is not yet understood to which extent the obtained interpretation results can be trusted and how accurate they can reflect the model decisions. Methods We conduct studies to quantitatively evaluate seven different deep interpretation techniques across different models and datasets for EEG-based BCI. Results The results reveal the importance of selecting a proper interpretation technique as the initial step. In addition, we also find that the quality of the interpretation results is inconsistent for individual samples despite when a method with an overall good performance is used. Many factors, including model structure and dataset types, could potentially affect the quality of the interpretation results. Discussion Based on the observations, we propose a set of procedures that allow the interpretation results to be presented in an understandable and trusted way. We illustrate the usefulness of our method for EEG-based BCI with instances selected from different scenarios.
Collapse
Affiliation(s)
- Jian Cui
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
| | - Liqiang Yuan
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Zhaoxiang Wang
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
| | - Ruilin Li
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Tianzi Jiang
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
- Brainnetome Center, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
32
|
Dong Y, Wen X, Gao F, Gao C, Cao R, Xiang J, Cao R. Subject-Independent EEG Classification of Motor Imagery Based on Dual-Branch Feature Fusion. Brain Sci 2023; 13:1109. [PMID: 37509039 PMCID: PMC10377689 DOI: 10.3390/brainsci13071109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2023] [Revised: 07/13/2023] [Accepted: 07/19/2023] [Indexed: 07/30/2023] Open
Abstract
A brain computer interface (BCI) system helps people with motor dysfunction interact with the external environment. With the advancement of technology, BCI systems have been applied in practice, but their practicability and usability are still greatly challenged. A large amount of calibration time is often required before BCI systems are used, which can consume the patient's energy and easily lead to anxiety. This paper proposes a novel motion-assisted method based on a novel dual-branch multiscale auto encoder network (MSAENet) to decode human brain motion imagery intentions, while introducing a central loss function to compensate for the shortcomings of traditional classifiers that only consider inter-class differences and ignore intra-class coupling. The effectiveness of the method is validated on three datasets, namely BCIIV2a, SMR-BCI and OpenBMI, to achieve zero calibration of the MI-BCI system. The results show that our proposed network displays good results on all three datasets. In the case of subject-independence, the MSAENet outperformed the other four comparison methods on the BCIIV2a and SMR-BCI datasets, while achieving F1_score values as high as 69.34% on the OpenBMI dataset. Our method maintains better classification accuracy with a small number of parameters and short prediction times, and the method achieves zero calibration of the MI-BCI system.
Collapse
Affiliation(s)
- Yanqing Dong
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Xin Wen
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Fang Gao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Chengxin Gao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Ruochen Cao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Jie Xiang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China
| | - Rui Cao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| |
Collapse
|
33
|
Nam H, Kim JM, Choi W, Bak S, Kam TE. The effects of layer-wise relevance propagation-based feature selection for EEG classification: a comparative study on multiple datasets. Front Hum Neurosci 2023; 17:1205881. [PMID: 37342822 PMCID: PMC10277566 DOI: 10.3389/fnhum.2023.1205881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 05/17/2023] [Indexed: 06/23/2023] Open
Abstract
Introduction The brain-computer interface (BCI) allows individuals to control external devices using their neural signals. One popular BCI paradigm is motor imagery (MI), which involves imagining movements to induce neural signals that can be decoded to control devices according to the user's intention. Electroencephalography (EEG) is frequently used for acquiring neural signals from the brain in the fields of MI-BCI due to its non-invasiveness and high temporal resolution. However, EEG signals can be affected by noise and artifacts, and patterns of EEG signals vary across different subjects. Therefore, selecting the most informative features is one of the essential processes to enhance classification performance in MI-BCI. Methods In this study, we design a layer-wise relevance propagation (LRP)-based feature selection method which can be easily integrated into deep learning (DL)-based models. We assess its effectiveness for reliable class-discriminative EEG feature selection on two different publicly available EEG datasets with various DL-based backbone models in the subject-dependent scenario. Results and discussion The results show that LRP-based feature selection enhances the performance for MI classification on both datasets for all DL-based backbone models. Based on our analysis, we believe that it can broad its capability to different research domains.
Collapse
Affiliation(s)
| | | | | | | | - Tae-Eui Kam
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| |
Collapse
|
34
|
García-Murillo DG, Álvarez-Meza AM, Castellanos-Dominguez CG. KCS-FCnet: Kernel Cross-Spectral Functional Connectivity Network for EEG-Based Motor Imagery Classification. Diagnostics (Basel) 2023; 13:diagnostics13061122. [PMID: 36980430 PMCID: PMC10046910 DOI: 10.3390/diagnostics13061122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2023] [Revised: 02/25/2023] [Accepted: 03/02/2023] [Indexed: 03/18/2023] Open
Abstract
This paper uses EEG data to introduce an approach for classifying right and left-hand classes in Motor Imagery (MI) tasks. The Kernel Cross-Spectral Functional Connectivity Network (KCS-FCnet) method addresses these limitations by providing richer spatial-temporal-spectral feature maps, a simpler architecture, and a more interpretable approach for EEG-driven MI discrimination. In particular, KCS-FCnet uses a single 1D-convolutional-based neural network to extract temporal-frequency features from raw EEG data and a cross-spectral Gaussian kernel connectivity layer to model channel functional relationships. As a result, the functional connectivity feature map reduces the number of parameters, improving interpretability by extracting meaningful patterns related to MI tasks. These patterns can be adapted to the subject’s unique characteristics. The validation results prove that introducing KCS-FCnet shallow architecture is a promising approach for EEG-based MI classification with the potential for real-world use in brain–computer interface systems.
Collapse
|
35
|
Chen J, Wang D, Yi W, Xu M, Tan X. Filter bank sinc-convolutional network with channel self-attention for high performance motor imagery decoding. J Neural Eng 2023; 20. [PMID: 36763992 DOI: 10.1088/1741-2552/acbb2c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2022] [Accepted: 02/10/2023] [Indexed: 02/12/2023]
Abstract
Objective.Motor Imagery Brain-Computer Interface (MI-BCI) is an active Brain-Computer Interface (BCI) paradigm focusing on the identification of motor intention, which is one of the most important non-invasive BCI paradigms. In MI-BCI studies, deep learning-based methods (especially lightweight networks) have attracted more attention in recent years, but the decoding performance still needs further improving.Approach.To solve this problem, we designed a filter bank structure with sinc-convolutional layers for spatio-temporal feature extraction of MI-electroencephalography in four motor rhythms. The Channel Self-Attention method was introduced for feature selection based on both global and local information, so as to build a model called Filter Bank Sinc-convolutional Network with Channel Self-Attention for high performance MI-decoding. Also, we proposed a data augmentation method based on multivariate empirical mode decomposition to improve the generalization capability of the model.Main results.We performed an intra-subject evaluation experiment on unseen data of three open MI datasets. The proposed method achieved mean accuracy of 78.20% (4-class scenario) on BCI Competition IV IIa, 87.34% (2-class scenario) on BCI Competition IV IIb, and 72.03% (2-class scenario) on Open Brain Machine Interface (OpenBMI) dataset, which are significantly higher than those of compared deep learning-based methods by at least 3.05% (p= 0.0469), 3.18% (p= 0.0371), and 2.27% (p= 0.0024) respectively.Significance.This work provides a new option for deep learning-based MI decoding, which can be employed for building BCI systems for motor rehabilitation.
Collapse
Affiliation(s)
- Jiaming Chen
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Dan Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Weibo Yi
- Beijing Machine and Equipment Institute, Beijing, People's Republic of China
| | - Meng Xu
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Xiyue Tan
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| |
Collapse
|
36
|
Wang J, Bi L, Feleke AG, Fei W. MRCPs-and-ERS/D-Oscillations-Driven Deep Learning Models for Decoding Unimanual and Bimanual Movements. IEEE Trans Neural Syst Rehabil Eng 2023; 31:1384-1393. [PMID: 37027527 DOI: 10.1109/tnsre.2023.3245617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
Abstract
Motor brain-computer interface (BCI) can intend to restore or compensate for central nervous system functionality. In the motor-BCI, motor execution (ME), which relies on patients' residual or intact movement functions, is a more intuitive and natural paradigm. Based on the ME paradigm, we can decode voluntary hand movement intentions from electroencephalography (EEG) signals. Numerous studies have investigated EEG-based unimanual movement decoding. Moreover, some studies have explored bimanual movement decoding since bimanual coordination is important in daily-life assistance and bilateral neurorehabilitation therapy. However, the multi-class classification of the unimanual and bimanual movements shows weak performance. To address this problem, in this work, we propose a neurophysiological signatures-driven deep learning model utilizing the movement-related cortical potentials (MRCPs) and event-related synchronization/ desynchronization (ERS/D) oscillations for the first time, inspired by the finding that brain signals encode motor-related information with both evoked potentials and oscillation components in ME. The proposed model consists of a feature representation module, an attention-based channel-weighting module, and a shallow convolutional neural network module. Results show that our proposed model has superior performance to the baseline methods. Six-class classification accuracies of unimanual and bimanual movements achieved 80.3%. Besides, each feature module of our model contributes to the performance. This work is the first to fuse the MRCPs and ERS/D oscillations of ME in deep learning to enhance the multi-class unimanual and bimanual movements' decoding performance. This work can facilitate the neural decoding of unimanual and bimanual movements for neurorehabilitation and assistance.
Collapse
|
37
|
Jaipriya D, Sriharipriya KC. A comparative analysis of masking empirical mode decomposition and a neural network with feed-forward and back propagation along with masking empirical mode decomposition to improve the classification performance for a reliable brain-computer interface. Front Comput Neurosci 2022; 16:1010770. [PMID: 36405787 PMCID: PMC9672820 DOI: 10.3389/fncom.2022.1010770] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 10/03/2022] [Indexed: 02/25/2024] Open
Abstract
In general, extraction and classification are used in various fields like image processing, pattern recognition, signal processing, and so on. Extracting effective characteristics from raw electroencephalogram (EEG) signals is a crucial role of the brain-computer interface for motor imagery. Recently, there has been a great deal of focus on motor imagery in the EEG signals since they encode a person's intent to do an action. Researchers have been using MI signals to assist paralyzed people and even move them on their own with certain equipment, like wheelchairs. As a result, proper decoding is an important step required for the interconnection of the brain and the computer. EEG decoding is a challenging process because of poor SNR, complexity, and other reasons. However, choosing an appropriate method to extract the features to improve the performance of motor imagery recognition is still a research hotspot. To extract the features of the EEG signal in the classification task, this paper proposes a Masking Empirical Mode Decomposition (MEMD) based Feed Forward Back Propagation Neural Network (MEMD-FFBPNN). The dataset consists of EEG signals which are first normalized using the minimax method and given as input to the MEMD to extract the features and then given to the FFBPNN to classify the tasks. The accuracy of the proposed method MEMD-FFBPNN has been measured using the confusion matrix, mean square error and which has been recorded up to 99.9%. Thus, the proposed method gives better accuracy than the other conventional methods.
Collapse
Affiliation(s)
| | - K. C. Sriharipriya
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, India
| |
Collapse
|
38
|
Image-Based Learning Using Gradient Class Activation Maps for Enhanced Physiological Interpretability of Motor Imagery Skills. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12031695] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Brain activity stimulated by the motor imagery paradigm (MI) is measured by Electroencephalography (EEG), which has several advantages to be implemented with the widely used Brain–Computer Interfaces (BCIs) technology. However, the substantial inter/intra variability of recorded data significantly influences individual skills on the achieved performance. This study explores the ability to distinguish between MI tasks and the interpretability of the brain’s ability to produce elicited mental responses with improved accuracy. We develop a Deep and Wide Convolutional Neuronal Network fed by a set of topoplots extracted from the multichannel EEG data. Further, we perform a visualization technique based on gradient-based class activation maps (namely, GradCam++) at different intervals along the MI paradigm timeline to account for intra-subject variability in neural responses over time. We also cluster the dynamic spatial representation of the extracted maps across the subject set to come to a deeper understanding of MI-BCI coordination skills. According to the results obtained from the evaluated GigaScience Database of motor-evoked potentials, the developed approach enhances the physiological explanation of motor imagery in aspects such as neural synchronization between rhythms, brain lateralization, and the ability to predict the MI onset responses and their evolution during training sessions.
Collapse
|
39
|
Lee DY, Jeong JH, Lee BH, Lee SW. Motor Imagery Classification Using Inter-Task Transfer Learning via A Channel-Wise Variational Autoencoder-based Convolutional Neural Network. IEEE Trans Neural Syst Rehabil Eng 2022; 30:226-237. [PMID: 35041605 DOI: 10.1109/tnsre.2022.3143836] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Highly sophisticated control based on a brain-computer interface (BCI) requires decoding kinematic information from brain signals. The forearm is a region of the upper limb that is often used in everyday life, but intuitive movements within the same limb have rarely been investigated in previous BCI studies. In this study, we focused on various forearm movement decoding from electroencephalography (EEG) signals using a small number of samples. Ten healthy participants took part in an experiment and performed motor execution (ME) and motor imagery (MI) of the intuitive movement tasks (Dataset I). We propose a convolutional neural network using a channel-wise variational autoencoder (CVNet) based on inter-task transfer learning. We approached that training the reconstructed ME-EEG signals together will also achieve more sufficient classification performance with only a small amount of MI-EEG signals. The proposed CVNet was validated on our own Dataset I and a public dataset, BNCI Horizon 2020 (Dataset II). The classification accuracies of various movements are confirmed to be 0.83 (±0.04) and 0.69 (±0.04) for Dataset I and II, respectively. The results show that the proposed method exhibits performance increases of approximately 0.09~0.27 and 0.08~0.24 compared with the conventional models for Dataset I and II, respectively. The outcomes suggest that the training model for decoding imagined movements can be performed using data from ME and a small number of data samples from MI. Hence, it is presented the feasibility of BCI learning strategies that can sufficiently learn deep learning with a few amount of calibration dataset and time only, with stable performance.
Collapse
|
40
|
Wang X, Cavigelli L, Schneider T, Benini L. Sub-100 μW Multispectral Riemannian Classification for EEG-Based Brain-Machine Interfaces. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2021; 15:1149-1160. [PMID: 34932486 DOI: 10.1109/tbcas.2021.3137290] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Motor imagery (MI) brain-machine interfaces (BMIs) enable us to control machines by merely thinking of performing a motor action. Practical use cases require a wearable solution where the classification of the brain signals is done locally near the sensor using machine learning models embedded on energy-efficient microcontroller units (MCUs), for assured privacy, user comfort, and long-term usage. In this work, we provide practical insights on the accuracy-cost trade-off for embedded BMI solutions. Our multispectral Riemannian classifier reaches 75.1% accuracy on a 4-class MI task. The accuracy is further improved by tuning different types of classifiers to each subject, achieving 76.4%. We further scale down the model by quantizing it to mixed-precision representations with a minimal accuracy loss of 1% and 1.4%, respectively, which is still up to 4.1% more accurate than the state-of-the-art embedded convolutional neural network. We implement the model on a low-power MCU within an energy budget of merely 198 μJ and taking only 16.9 ms per classification. Classifying samples continuously, overlapping the 3.5 s samples by 50% to avoid missing user inputs allows for operation at just 85 μW. Compared to related works in embedded MI-BMIs, our solution sets the new state-of-the-art in terms of accuracy-energy trade-off for near-sensor classification.
Collapse
|
41
|
|
42
|
Altaheri H, Muhammad G, Alsulaiman M, Amin SU, Altuwaijri GA, Abdul W, Bencherif MA, Faisal M. Deep learning techniques for classification of electroencephalogram (EEG) motor imagery (MI) signals: a review. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06352-5] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
|
43
|
Singh A, Hussain AA, Lal S, Guesgen HW. A Comprehensive Review on Critical Issues and Possible Solutions of Motor Imagery Based Electroencephalography Brain-Computer Interface. SENSORS 2021; 21:s21062173. [PMID: 33804611 PMCID: PMC8003721 DOI: 10.3390/s21062173] [Citation(s) in RCA: 36] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/28/2020] [Revised: 03/15/2021] [Accepted: 03/16/2021] [Indexed: 01/16/2023]
Abstract
Motor imagery (MI) based brain–computer interface (BCI) aims to provide a means of communication through the utilization of neural activity generated due to kinesthetic imagination of limbs. Every year, a significant number of publications that are related to new improvements, challenges, and breakthrough in MI-BCI are made. This paper provides a comprehensive review of the electroencephalogram (EEG) based MI-BCI system. It describes the current state of the art in different stages of the MI-BCI (data acquisition, MI training, preprocessing, feature extraction, channel and feature selection, and classification) pipeline. Although MI-BCI research has been going for many years, this technology is mostly confined to controlled lab environments. We discuss recent developments and critical algorithmic issues in MI-based BCI for commercial deployment.
Collapse
|