1
|
Li Y, Chen B, Yoshimura N, Koike Y, Yamashita O. Sparse Bayesian correntropy learning for robust muscle activity reconstruction from noisy brain recordings. Neural Netw 2025; 182:106899. [PMID: 39571386 DOI: 10.1016/j.neunet.2024.106899] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 10/15/2024] [Accepted: 11/07/2024] [Indexed: 12/17/2024]
Abstract
Sparse Bayesian learning has promoted many effective frameworks of brain activity decoding for the brain-computer interface, including the direct reconstruction of muscle activity using brain recordings. However, existing sparse Bayesian learning algorithms mainly use Gaussian distribution as error assumption in the reconstruction task, which is not necessarily the truth in the real-world application. On the other hand, brain recording is known to be highly noisy and contains many non-Gaussian noises, which could lead to large performance degradation for sparse Bayesian learning algorithms. The goal of this paper is to propose a novel robust implementation of sparse Bayesian learning so that robustness and sparseness can be realized simultaneously. Motivated by the exceptional robustness of maximum correntropy criterion (MCC), we proposed integrating MCC to the sparse Bayesian learning regime. To be specific, we derived the explicit error assumption inherent in the MCC, and then leveraged it for the likelihood function. Meanwhile, we utilized the automatic relevance determination technique as the sparse prior distribution. To fully evaluate the proposed method, a synthetic example and a real-world muscle activity reconstruction task with two different brain modalities were leveraged. Experimental results showed, our proposed sparse Bayesian correntropy learning framework significantly improves the robustness for the noisy regression tasks. Our proposed algorithm could realize higher correlation coefficients and lower root mean squared errors for the real-world muscle activity reconstruction scenario. Sparse Bayesian correntropy learning provides a powerful approach for brain activity decoding which will promote the development of brain-computer interface technology.
Collapse
Affiliation(s)
- Yuanhao Li
- Center for Advanced Intelligence Project, RIKEN, Tokyo, 103-0027, Japan; Department of Computational Brain Imaging, Advanced Telecommunication Research Institute International, Kyoto, 619-0237, Japan.
| | - Badong Chen
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, 710049, China
| | - Natsue Yoshimura
- School of Computing, Institute of Science Tokyo, Yokohama, 226-8501, Japan
| | - Yasuharu Koike
- Institute of Integrated Research, Institute of Science Tokyo, Yokohama, 226-8501, Japan
| | - Okito Yamashita
- Center for Advanced Intelligence Project, RIKEN, Tokyo, 103-0027, Japan; Department of Computational Brain Imaging, Advanced Telecommunication Research Institute International, Kyoto, 619-0237, Japan
| |
Collapse
|
2
|
Zheng X, Zhang L, Yan L, Zhao L. A robust semi-supervised regressor with correntropy-induced manifold regularization and adaptive graph. Neural Netw 2025; 182:106902. [PMID: 39577044 DOI: 10.1016/j.neunet.2024.106902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 08/20/2024] [Accepted: 11/07/2024] [Indexed: 11/24/2024]
Abstract
For semi-supervised regression tasks, most existing methods ignore the impact of noise. However, the data inevitably contain noise. Therefore, this study proposes a novel correntropy-induced semi-supervised regression (CSSR) method that mitigates the adverse effects of noise. To implement the robustness of CSSR, a novel correntropy-induced manifold regularization (CMR) and a correntropy-induced adaptive graph (CAG) are designed. Specifically, CMR is inspired by the principles of correntropy and aims to learn a graph representation, whereas CAG inherits the robust characteristics of the correntropy metric and adaptively constructs an adjacency matrix. Finally, by incorporating CMR, CAG, and the correntropy-induced loss seamlessly, CSSR demonstrates the ability to deliver promising joint performance. The final solution of CSSR is achieved through an iterative process. Moreover, we validated the convergence of CSSR through a combination of theoretical analyses and empirical experiments. The experimental evaluation encompassed three synthetic, 15 benchmark, and two image datasets. The findings demonstrate that CSSR surpasses similar methods in the realm of semi-supervised regression tasks, demonstrating its effectiveness and robustness.
Collapse
Affiliation(s)
- Xiaohan Zheng
- School of Computer Science and Technology, Soochow University, 215006 Suzhou, China.
| | - Li Zhang
- School of Computer Science and Technology, Soochow University, 215006 Suzhou, China.
| | - Leilei Yan
- School of Computer Science and Technology, Soochow University, 215006 Suzhou, China.
| | - Lei Zhao
- School of Computer Science and Technology, Soochow University, 215006 Suzhou, China.
| |
Collapse
|
3
|
Yuan P, You X, Chen H, Wang Y, Peng Q, Zou B. Sparse Additive Machine With the Correntropy-Induced Loss. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:1989-2003. [PMID: 37289610 DOI: 10.1109/tnnls.2023.3280349] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Sparse additive machines (SAMs) have shown competitive performance on variable selection and classification in high-dimensional data due to their representation flexibility and interpretability. However, the existing methods often employ the unbounded or nonsmooth functions as the surrogates of 0-1 classification loss, which may encounter the degraded performance for data with outliers. To alleviate this problem, we propose a robust classification method, named SAM with the correntropy-induced loss (CSAM), by integrating the correntropy-induced loss (C-loss), the data-dependent hypothesis space, and the weighted -norm regularizer ( ) into additive machines. In theory, the generalization error bound is estimated via a novel error decomposition and the concentration estimation techniques, which shows that the convergence rate can be achieved under proper parameter conditions. In addition, the theoretical guarantee on variable selection consistency is analyzed. Experimental evaluations on both synthetic and real-world datasets consistently validate the effectiveness and robustness of the proposed approach.
Collapse
|
4
|
Li F, Yang H. A novel bounded loss framework for support vector machines. Neural Netw 2024; 178:106476. [PMID: 38959596 DOI: 10.1016/j.neunet.2024.106476] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 05/29/2024] [Accepted: 06/17/2024] [Indexed: 07/05/2024]
Abstract
This paper introduces a novel bounded loss framework for SVM and SVR. Specifically, using the Pinball loss as an illustration, we devise a novel bounded exponential quantile loss (Leq-loss) for both support vector machine classification and regression tasks. For Leq-loss, it not only enhances the robustness of SVM and SVR against outliers but also improves the robustness of SVM to resampling from a different perspective. Furthermore, EQSVM and EQSVR were constructed based on Leq-loss, and the influence functions and breakdown point lower bounds of their estimators are derived. It is proved that the influence functions are bounded, and the breakdown point lower bounds can reach the highest asymptotic breakdown point of 1/2. Additionally, we demonstrated the robustness of EQSVM to resampling and derived its generalization error bound based on Rademacher complexity. Due to the Leq-loss being non-convex, we can use the concave-convex procedure (CCCP) technique to transform the problem into a series of convex optimization problems and use the ClipDCD algorithm to solve these convex optimization problems. Numerous experiments have been conducted to confirm the effectiveness of the proposed EQSVM and EQSVR.
Collapse
Affiliation(s)
- Feihong Li
- College of Mathematics and Statistics, Chongqing University, Chongqing, 401331, China.
| | - Hu Yang
- College of Mathematics and Statistics, Chongqing University, Chongqing, 401331, China.
| |
Collapse
|
5
|
Ding Y, Zhou H, Zou Q, Yuan L. Identification of drug-side effect association via correntropy-loss based matrix factorization with neural tangent kernel. Methods 2023; 219:73-81. [PMID: 37783242 DOI: 10.1016/j.ymeth.2023.09.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 09/18/2023] [Accepted: 09/20/2023] [Indexed: 10/04/2023] Open
Abstract
Adverse drug reactions include side effects, allergic reactions, and secondary infections. Severe adverse reactions can cause cancer, deformity, or mutation. The monitoring of drug side effects is an important support for post marketing safety supervision of drugs, and an important basis for revising drug instructions. Its purpose is to timely detect and control drug safety risks. Traditional methods are time-consuming. To accelerate the discovery of side effects, we propose a machine learning based method, called correntropy-loss based matrix factorization with neural tangent kernel (CLMF-NTK), to solve the prediction of drug side effects. Our method and other computational methods are tested on three benchmark datasets, and the results show that our method achieves the best predictive performance.
Collapse
Affiliation(s)
- Yijie Ding
- Key Laboratory of Computational Science and Application of Hainan Province, Hainan Normal University, Haikou 571158, China; Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou 324000, China; School of Electronic and Information Engineering, Suzhou University of Science and Technology, Suzhou 215009, China
| | - Hongmei Zhou
- Beidahuang Industry Group General Hospital, Harbin 150001, China
| | - Quan Zou
- Yangtze Delta Region Institute (Quzhou), University of Electronic Science and Technology of China, Quzhou 324000, China.
| | - Lei Yuan
- Department of Hepatobiliary Surgery, Quzhou People's Hospital, 100# Minjiang Main Road, Quzhou 324000, China.
| |
Collapse
|
6
|
Capped Asymmetric Elastic Net Support Vector Machine for Robust Binary Classification. INT J INTELL SYST 2023. [DOI: 10.1155/2023/2201330] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/25/2023]
Abstract
Recently, there are lots of literature on improving the robustness of SVM by constructing nonconvex functions, but they seldom theoretically study the robust property of the constructed functions. In this paper, based on our recent work, we present a novel capped asymmetric elastic net (CaEN) loss and equip it with the SVM as CaENSVM. We derive the influence function of the estimators of the CaENSVM to theoretically explain the robustness of the proposed method. Our results can be easily extended to other similar nonconvex loss functions. We further show that the influence function of the CaENSVM is bounded, so that the robustness of the CaENSVM can be theoretically explained. Other theoretical analysis demonstrates that the CaENSVM satisfies the Bayes rule and the corresponding generalization error bound based on Rademacher complexity guarantees its good generalization capability. Since CaEN loss is concave, we implement an efficient DC procedure based on the stochastic gradient descent algorithm (Pegasos) to solve the optimization problem. A host of experiments are conducted to verify the effectiveness of our proposed CaENSVM model.
Collapse
|
7
|
Chen B, Xie Y, Wang X, Yuan Z, Ren P, Qin J. Multikernel Correntropy for Robust Learning. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:13500-13511. [PMID: 34550898 DOI: 10.1109/tcyb.2021.3110732] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
As a novel similarity measure that is defined as the expectation of a kernel function between two random variables, correntropy has been successfully applied in robust machine learning and signal processing to combat large outliers. The kernel function in correntropy is usually a zero-mean Gaussian kernel. In a recent work, the concept of mixture correntropy (MC) was proposed to improve the learning performance, where the kernel function is a mixture Gaussian kernel, namely, a linear combination of several zero-mean Gaussian kernels with different widths. In both correntropy and MC, the center of the kernel function is, however, always located at zero. In the present work, to further improve the learning performance, we propose the concept of multikernel correntropy (MKC), in which each component of the mixture Gaussian kernel can be centered at a different location. The properties of the MKC are investigated and an efficient approach is proposed to determine the free parameters in MKC. Experimental results show that the learning algorithms under the maximum MKC criterion (MMKCC) can outperform those under the original maximum correntropy criterion (MCC) and the maximum MC criterion (MMCC).
Collapse
|
8
|
Li Y, Chen B, Yoshimura N, Koike Y. Restricted Minimum Error Entropy Criterion for Robust Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:6599-6612. [PMID: 34077373 DOI: 10.1109/tnnls.2021.3082571] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The minimum error entropy (MEE) criterion is a powerful approach for non-Gaussian signal processing and robust machine learning. However, the instantiation of MEE on robust classification is a rather vacancy in the literature. The original MEE purely focuses on minimizing Renyi's quadratic entropy of the prediction errors, which could exhibit inferior capability in noisy classification tasks. To this end, we analyze the optimal error distribution with adverse outliers and introduce a specific codebook for restriction, which optimizes the error distribution toward the optimal case. Half-quadratic-based optimization and convergence analysis of the proposed learning criterion, called restricted MEE (RMEE), are provided. The experimental results considering logistic regression and extreme learning machine on synthetic data and UCI datasets, respectively, are presented to demonstrate the superior robustness of RMEE. Furthermore, we evaluate RMEE on a noisy electroencephalogram dataset, so as to strengthen its practical impact.
Collapse
|
9
|
Qi K, Yang H. A novel robust nonparallel support vector classifier based on one optimization problem. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07814-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
10
|
|
11
|
Qi K, Yang H. Joint rescaled asymmetric least squared nonparallel support vector machine with a stochastic quasi-Newton based algorithm. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03183-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
12
|
Ren Q, Yang L. A robust projection twin support vector machine with a generalized correntropy-based loss. APPL INTELL 2022. [DOI: 10.1007/s10489-021-02480-6] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
13
|
Xiong K, Iu HHC, Wang S. Kernel Correntropy Conjugate Gradient Algorithms Based on Half-Quadratic Optimization. IEEE TRANSACTIONS ON CYBERNETICS 2021; 51:5497-5510. [PMID: 31945006 DOI: 10.1109/tcyb.2019.2959834] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
As a nonlinear similarity measure defined in the kernel space, the correntropic loss (C-Loss) can address the stability issues of second-order similarity measures thanks to its ability to extract high-order statistics of data. However, the kernel adaptive filter (KAF) based on the C-Loss uses the stochastic gradient descent (SGD) method to update its weights and, thus, suffers from poor performance and a slow convergence rate. To address these issues, the conjugate gradient (CG)-based correntropy algorithm is developed by solving the combination of half-quadratic (HQ) optimization and weighted least-squares (LS) problems, generating a novel robust kernel correntropy CG (KCCG) algorithm. The proposed KCCG with less computational complexity achieves comparable performance to the kernel recursive maximum correntropy (KRMC) algorithm. To further curb the growth of the network in KCCG, the random Fourier features KCCG (RFFKCCG) algorithm is proposed by transforming the original input data into a fixed-dimensional random Fourier features space (RFFS). Since only one current error information is used in the loss function of RFFKCCG, it can provide a more efficient filter structure than the other KAFs with sparsification. The Monte Carlo simulations conducted in the prediction of synthetic and real-world chaotic time series and the regression for large-scale datasets validate the superiorities of the proposed algorithms in terms of robustness, filtering accuracy, and complexity.
Collapse
|
14
|
CTSVM: A robust twin support vector machine with correntropy-induced loss function for binary classification problems. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2021.01.006] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
15
|
Chen S, Yang J, Wei Y, Luo L, Lu GF, Gong C. δ-Norm-Based Robust Regression With Applications to Image Analysis. IEEE TRANSACTIONS ON CYBERNETICS 2021; 51:3371-3383. [PMID: 30872251 DOI: 10.1109/tcyb.2019.2901248] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Up to now, various matrix norms (e.g., l1 -norm, l2 -norm, l2,1 -norm, etc.) have been widely leveraged to form the loss function of different regression models, and have played an important role in image analysis. However, the previous regression models adopting the existing norms are sensitive to outliers and, thus, often bring about unsatisfactory results on the heavily corrupted images. This is because their adopted norms for measuring the data residual can hardly suppress the negative influence of noisy data, which will probably mislead the regression process. To address this issue, this paper proposes a novel δ (delta)-norm to count the nonzero blocks around an element in a vector or matrix, which weakens the impacts of outliers and also takes the structure property of examples into account. After that, we present the δ -norm-based robust regression (DRR) in which the data examples are mapped to the kernel space and measured by the proposed δ -norm. By exploring an explicit kernel function, we show that DRR has a closed-form solution, which suggests that DRR can be efficiently solved. To further handle the influences from mixed noise, DRR is extended to a multiscale version. The experimental results on image classification and background modeling datasets validate the superiority of the proposed approach to the existing state-of-the-art robust regression models.
Collapse
|
16
|
Kernel-based regression via a novel robust loss function and iteratively reweighted least squares. Knowl Inf Syst 2021. [DOI: 10.1007/s10115-021-01554-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
17
|
Ren LR, Gao YL, Liu JX, Shang J, Zheng CH. Correntropy induced loss based sparse robust graph regularized extreme learning machine for cancer classification. BMC Bioinformatics 2020; 21:445. [PMID: 33028187 PMCID: PMC7542897 DOI: 10.1186/s12859-020-03790-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2019] [Accepted: 09/30/2020] [Indexed: 01/17/2023] Open
Abstract
Background As a machine learning method with high performance and excellent generalization ability, extreme learning machine (ELM) is gaining popularity in various studies. Various ELM-based methods for different fields have been proposed. However, the robustness to noise and outliers is always the main problem affecting the performance of ELM. Results In this paper, an integrated method named correntropy induced loss based sparse robust graph regularized extreme learning machine (CSRGELM) is proposed. The introduction of correntropy induced loss improves the robustness of ELM and weakens the negative effects of noise and outliers. By using the L2,1-norm to constrain the output weight matrix, we tend to obtain a sparse output weight matrix to construct a simpler single hidden layer feedforward neural network model. By introducing the graph regularization to preserve the local structural information of the data, the classification performance of the new method is further improved. Besides, we design an iterative optimization method based on the idea of half quadratic optimization to solve the non-convex problem of CSRGELM. Conclusions The classification results on the benchmark dataset show that CSRGELM can obtain better classification results compared with other methods. More importantly, we also apply the new method to the classification problems of cancer samples and get a good classification effect.
Collapse
Affiliation(s)
- Liang-Rui Ren
- School of Computer Science, Qufu Normal University, Rizhao, 276826, China
| | - Ying-Lian Gao
- Qufu Normal University Library, Qufu Normal University, Rizhao, 276826, China
| | - Jin-Xing Liu
- School of Computer Science, Qufu Normal University, Rizhao, 276826, China.
| | - Junliang Shang
- School of Computer Science, Qufu Normal University, Rizhao, 276826, China
| | - Chun-Hou Zheng
- School of Computer Science, Qufu Normal University, Rizhao, 276826, China.,College of Computer Science and Technology, Anhui University, Hefei, 230601, China
| |
Collapse
|
18
|
|
19
|
Robust semi-supervised support vector machines with Laplace kernel-induced correntropy loss functions. APPL INTELL 2020. [DOI: 10.1007/s10489-020-01865-3] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
20
|
Yang L, Dong H. Robust support vector machine with generalized quantile loss for classification and regression. Appl Soft Comput 2019. [DOI: 10.1016/j.asoc.2019.105483] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|