1
|
Li Y, Chen B, Yoshimura N, Koike Y, Yamashita O. Sparse Bayesian correntropy learning for robust muscle activity reconstruction from noisy brain recordings. Neural Netw 2025; 182:106899. [PMID: 39571386 DOI: 10.1016/j.neunet.2024.106899] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 10/15/2024] [Accepted: 11/07/2024] [Indexed: 12/17/2024]
Abstract
Sparse Bayesian learning has promoted many effective frameworks of brain activity decoding for the brain-computer interface, including the direct reconstruction of muscle activity using brain recordings. However, existing sparse Bayesian learning algorithms mainly use Gaussian distribution as error assumption in the reconstruction task, which is not necessarily the truth in the real-world application. On the other hand, brain recording is known to be highly noisy and contains many non-Gaussian noises, which could lead to large performance degradation for sparse Bayesian learning algorithms. The goal of this paper is to propose a novel robust implementation of sparse Bayesian learning so that robustness and sparseness can be realized simultaneously. Motivated by the exceptional robustness of maximum correntropy criterion (MCC), we proposed integrating MCC to the sparse Bayesian learning regime. To be specific, we derived the explicit error assumption inherent in the MCC, and then leveraged it for the likelihood function. Meanwhile, we utilized the automatic relevance determination technique as the sparse prior distribution. To fully evaluate the proposed method, a synthetic example and a real-world muscle activity reconstruction task with two different brain modalities were leveraged. Experimental results showed, our proposed sparse Bayesian correntropy learning framework significantly improves the robustness for the noisy regression tasks. Our proposed algorithm could realize higher correlation coefficients and lower root mean squared errors for the real-world muscle activity reconstruction scenario. Sparse Bayesian correntropy learning provides a powerful approach for brain activity decoding which will promote the development of brain-computer interface technology.
Collapse
Affiliation(s)
- Yuanhao Li
- Center for Advanced Intelligence Project, RIKEN, Tokyo, 103-0027, Japan; Department of Computational Brain Imaging, Advanced Telecommunication Research Institute International, Kyoto, 619-0237, Japan.
| | - Badong Chen
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Natsue Yoshimura
- School of Computing, Institute of Science Tokyo, Yokohama, 226-8501, Japan
| | - Yasuharu Koike
- Institute of Integrated Research, Institute of Science Tokyo, Yokohama, 226-8501, Japan
| | - Okito Yamashita
- Center for Advanced Intelligence Project, RIKEN, Tokyo, 103-0027, Japan; Department of Computational Brain Imaging, Advanced Telecommunication Research Institute International, Kyoto, 619-0237, Japan
| |
Collapse
|
2
|
Yuan P, You X, Chen H, Wang Y, Peng Q, Zou B. Sparse Additive Machine With the Correntropy-Induced Loss. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:1989-2003. [PMID: 37289610 DOI: 10.1109/tnnls.2023.3280349] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Sparse additive machines (SAMs) have shown competitive performance on variable selection and classification in high-dimensional data due to their representation flexibility and interpretability. However, the existing methods often employ the unbounded or nonsmooth functions as the surrogates of 0-1 classification loss, which may encounter the degraded performance for data with outliers. To alleviate this problem, we propose a robust classification method, named SAM with the correntropy-induced loss (CSAM), by integrating the correntropy-induced loss (C-loss), the data-dependent hypothesis space, and the weighted -norm regularizer ( ) into additive machines. In theory, the generalization error bound is estimated via a novel error decomposition and the concentration estimation techniques, which shows that the convergence rate can be achieved under proper parameter conditions. In addition, the theoretical guarantee on variable selection consistency is analyzed. Experimental evaluations on both synthetic and real-world datasets consistently validate the effectiveness and robustness of the proposed approach.
Collapse
|
3
|
Chen B, Zheng Y, Ren P. Error Loss Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:5256-5268. [PMID: 36099217 DOI: 10.1109/tnnls.2022.3202989] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
A novel model called error loss network (ELN) is proposed to build an error loss function for supervised learning. The ELN is similar in structure to a radial basis function (RBF) neural network, but its input is an error sample and output is a loss corresponding to that error sample. That means the nonlinear input-output mapper of the ELN creates an error loss function. The proposed ELN provides a unified model for a large class of error loss functions, which includes some information-theoretic learning (ITL) loss functions as special cases. The activation function, weight parameters, and network size of the ELN can be predetermined or learned from the error samples. On this basis, we propose a new machine learning paradigm where the learning process is divided into two stages: first, learning a loss function using an ELN; second, using the learned loss function to continue to perform the learning. Experimental results are presented to demonstrate the desirable performance of the new method.
Collapse
|
4
|
Li Y, Chen B, Wang G, Yoshimura N, Koike Y. Partial maximum correntropy regression for robust electrocorticography decoding. Front Neurosci 2023; 17:1213035. [PMID: 37457015 PMCID: PMC10347400 DOI: 10.3389/fnins.2023.1213035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Accepted: 06/15/2023] [Indexed: 07/18/2023] Open
Abstract
The Partial Least Square Regression (PLSR) method has shown admirable competence for predicting continuous variables from inter-correlated electrocorticography signals in the brain-computer interface. However, PLSR is essentially formulated with the least square criterion, thus, being considerably prone to the performance deterioration caused by the brain recording noises. To address this problem, this study aims to propose a new robust variant for PLSR. To this end, the maximum correntropy criterion (MCC) is utilized to propose a new robust implementation of PLSR, called Partial Maximum Correntropy Regression (PMCR). The half-quadratic optimization is utilized to calculate the robust projectors for the dimensionality reduction, and the regression coefficients are optimized by a fixed-point optimization method. The proposed PMCR is evaluated with a synthetic example and a public electrocorticography dataset under three performance indicators. For the synthetic example, PMCR realized better prediction results compared with the other existing methods. PMCR could also abstract valid information with a limited number of decomposition factors in a noisy regression scenario. For the electrocorticography dataset, PMCR achieved superior decoding performance in most cases, and also realized the minimal neurophysiological pattern deterioration with the interference of the noises. The experimental results demonstrate that, the proposed PMCR could outperform the existing methods in a noisy, inter-correlated, and high-dimensional decoding task. PMCR could alleviate the performance degradation caused by the adverse noises and ameliorate the electrocorticography decoding robustness for the brain-computer interface.
Collapse
Affiliation(s)
- Yuanhao Li
- Institute of Innovative Research, Tokyo Institute of Technology, Yokohama, Japan
| | - Badong Chen
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Gang Wang
- Key Laboratory of Biomedical Information Engineering of Ministry of Education, Xi'an Jiaotong University, Xi'an, China
| | - Natsue Yoshimura
- School of Computing, Tokyo Institute of Technology, Yokohama, Japan
| | - Yasuharu Koike
- Institute of Innovative Research, Tokyo Institute of Technology, Yokohama, Japan
| |
Collapse
|
5
|
Fang X, Tan Y, Zhang F, Duan S, Wang L. Transient Response and Firing Behaviors of Memristive Neuron Circuit. Front Neurosci 2022; 16:922086. [PMID: 35812218 PMCID: PMC9257141 DOI: 10.3389/fnins.2022.922086] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2022] [Accepted: 05/30/2022] [Indexed: 11/13/2022] Open
Abstract
The signal transmission mechanism of the Resistor-Capacitor (RC) circuit is similar to the intracellular and extracellular signal propagating mechanism of the neuron. Thus, the RC circuit can be utilized as the circuit model of the neuron cell membrane. However, resistors are electronic components with the fixed-resistance and have no memory properties. A memristor is a promising neuro-morphological electronic device with nonvolatile, switching, and nonlinear characteristics. First of all, we consider replacing the resistor in the RC neuron circuit with a memristor, which is named the Memristor-Capacitor (MC) circuit, then the MC neuron model is constructed. We compare the charging and discharging processes between the RC and MC neuron circuits. Secondly, two models are compared under the different external stimuli. Finally, the synchronous and asynchronous activities of the RC and MC neuron circuits are performed. Extensive experimental results suggest that the charging and discharging speed of the MC neuron circuit is faster than that of the RC neuron circuit. Given sufficient time and proper external stimuli, the RC and MC neuron circuits can produce the action potentials. The synchronous and asynchronous phenomena in the two neuron circuits reproduce nonlinear dynamic behaviors of the biological neurons.
Collapse
Affiliation(s)
- Xiaoyan Fang
- College of Artificial Intelligence, Southwest University, Chongqing, China
| | - Yao Tan
- Department of Big Data and Machine Learning, Chongqing University of Technology, Chongqing, China
| | - Fengqing Zhang
- College of Artificial Intelligence, Southwest University, Chongqing, China
| | - Shukai Duan
- College of Artificial Intelligence, Southwest University, Chongqing, China
| | - Lidan Wang
- College of Artificial Intelligence, Southwest University, Chongqing, China
- *Correspondence: Lidan Wang
| |
Collapse
|
6
|
Yang S, Tan J, Chen B. Robust Spike-Based Continual Meta-Learning Improved by Restricted Minimum Error Entropy Criterion. ENTROPY 2022; 24:e24040455. [PMID: 35455118 PMCID: PMC9031894 DOI: 10.3390/e24040455] [Citation(s) in RCA: 44] [Impact Index Per Article: 14.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Revised: 03/19/2022] [Accepted: 03/23/2022] [Indexed: 02/04/2023]
Abstract
The spiking neural network (SNN) is regarded as a promising candidate to deal with the great challenges presented by current machine learning techniques, including the high energy consumption induced by deep neural networks. However, there is still a great gap between SNNs and the online meta-learning performance of artificial neural networks. Importantly, existing spike-based online meta-learning models do not target the robust learning based on spatio-temporal dynamics and superior machine learning theory. In this invited article, we propose a novel spike-based framework with minimum error entropy, called MeMEE, using the entropy theory to establish the gradient-based online meta-learning scheme in a recurrent SNN architecture. We examine the performance based on various types of tasks, including autonomous navigation and the working memory test. The experimental results show that the proposed MeMEE model can effectively improve the accuracy and the robustness of the spike-based meta-learning performance. More importantly, the proposed MeMEE model emphasizes the application of the modern information theoretic learning approach on the state-of-the-art spike-based learning algorithms. Therefore, in this invited paper, we provide new perspectives for further integration of advanced information theory in machine learning to improve the learning performance of SNNs, which could be of great merit to applied developments with spike-based neuromorphic systems.
Collapse
Affiliation(s)
- Shuangming Yang
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China; (S.Y.); (J.T.)
| | - Jiangtong Tan
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China; (S.Y.); (J.T.)
| | - Badong Chen
- Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an 710049, China
- Correspondence:
| |
Collapse
|