1
|
Wong HT, Mai J, Wang Z, Leung CS. Generalized M-sparse algorithms for constructing fault tolerant RBF networks. Neural Netw 2024; 180:106633. [PMID: 39208461 DOI: 10.1016/j.neunet.2024.106633] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2023] [Revised: 11/02/2023] [Accepted: 08/12/2024] [Indexed: 09/04/2024]
Abstract
In the construction process of radial basis function (RBF) networks, two common crucial issues arise: the selection of RBF centers and the effective utilization of the given source without encountering the overfitting problem. Another important issue is the fault tolerant capability. That is, when noise or faults exist in a trained network, it is crucial that the network's performance does not undergo significant deterioration or decrease. However, without employing a fault tolerant procedure, a trained RBF network may exhibit significantly poor performance. Unfortunately, most existing algorithms are unable to simultaneously address all of the aforementioned issues. This paper proposes fault tolerant training algorithms that can simultaneously select RBF nodes and train RBF output weights. Additionally, our algorithms can directly control the number of RBF nodes in an explicit manner, eliminating the need for a time-consuming procedure to tune the regularization parameter and achieve the target RBF network size. Based on simulation results, our algorithms demonstrate improved test set performance when more RBF nodes are used, effectively utilizing the given source without encountering the overfitting problem. This paper first defines a fault tolerant objective function, which includes a term to suppress the effects of weight faults and weight noise. This term also prevents the issue of overfitting, resulting in better test set performance when more RBF nodes are utilized. With the defined objective function, the training process is designed to solve a generalized M-sparse problem by incorporating an ℓ0-norm constraint. The ℓ0-norm constraint allows us to directly and explicitly control the number of RBF nodes. To address the generalized M-sparse problem, we introduce the noise-resistant iterative hard thresholding (NR-IHT) algorithm. The convergence properties of the NR-IHT algorithm are subsequently discussed theoretically. To further enhance performance, we incorporate the momentum concept into the NR-IHT algorithm, referring to the modified version as "NR-IHT-Mom". Simulation results show that both the NR-IHT algorithm and the NR-IHT-Mom algorithm outperform several state-of-the-art comparison algorithms.
Collapse
Affiliation(s)
- Hiu-Tung Wong
- Center for Intelligent Multidimensional Data Analysis, Hong Kong Science Park, Shatin, Hong Kong Special Administrative Region of China; Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong Special Administrative Region of China.
| | - Jiajie Mai
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong Special Administrative Region of China.
| | - Zhenni Wang
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong Special Administrative Region of China.
| | - Chi-Sing Leung
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong Special Administrative Region of China.
| |
Collapse
|
2
|
Wang J, Gao R, Zheng H, Zhu H, Shi CJR. SSGCNet: A Sparse Spectra Graph Convolutional Network for Epileptic EEG Signal Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:12157-12171. [PMID: 37030729 DOI: 10.1109/tnnls.2023.3252569] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
In this article, we propose a sparse spectra graph convolutional network (SSGCNet) for epileptic electroencephalogram (EEG) signal classification. The goal is to develop a lightweighted deep learning model while retaining a high level of classification accuracy. To do so, we propose a weighted neighborhood field graph (WNFG) to represent EEG signals. The WNFG reduces redundant edges between graph nodes and has lower graph generation time and memory usage than the baseline solution. The sequential graph convolutional network is further developed from a WNFG by combining sparse weight pruning and the alternating direction method of multipliers (ADMM). Compared with the state-of-the-art method, our method has the same classification accuracy on the Bonn public dataset and the spikes and slow waves (SSW) clinical real dataset when the connection rate is ten times smaller.
Collapse
|
3
|
Zhao Y, Si D, Pei J, Yang X. Geodesic Basis Function Neural Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:8386-8400. [PMID: 37015442 DOI: 10.1109/tnnls.2022.3227296] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
In the learning of existing radial basis function neural networks-based methods, it is difficult to propagate errors back. This leads to an inconsistency between the learning and recognition task. This article proposes a geodesic basis function neural network with subclass extension learning (GBFNN-ScE). The geodesic basis function (GBF), which is defined here for the first time, uses the geodetic distance in the manifold as a measure to obtain the response of the sample with respect to the local center. To learn network parameters by back-propagating errors for the purpose of correct classification, a specific GBF based on a pruned gamma encoding cosine function is constructed. This function has a concise and explicit expression on the hyperspherical manifold, which is conducive to the realization of error back propagation. In the preprocessing layer, a sample unitization method with no loss of information, nonnegative unit hyperspherical crown (NUHC) mapping, is proposed. The sample can be mapped to the support set of the GBF. To alleviate the problem that one-hot encoding is not effective enough in the differential expression of data labels within a class, a subclass extension (ScE) learning strategy is proposed. The ScE learning strategy can help the learned network be more robust. For the working of GBFNN-ScE, the original sample is projected onto the support set of GBF through the NUHC mapping. Then the mapped samples are sent to the nonlinear computation units composed of GBFs in the hidden layer. Finally, the response obtained in the hidden layer is weighted by the learned weight to obtain the network output. This article theoretically proves that the separability of the data with ScE learning is stronger. The experimental results show that the proposed GBFNN-ScE has a better performance in recognition tasks than existing methods. The ablation experiments show that the ideas of the GBFNN-ScE contribute to the algorithm performance.
Collapse
|
4
|
Shi ZL, Li XP, Leung CS, So HC. Cardinality Constrained Portfolio Optimization via Alternating Direction Method of Multipliers. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:2901-2909. [PMID: 35895648 DOI: 10.1109/tnnls.2022.3192065] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Inspired by sparse learning, the Markowitz mean-variance model with a sparse regularization term is popularly used in sparse portfolio optimization. However, in penalty-based portfolio optimization algorithms, the cardinality level of the resultant portfolio relies on the choice of the regularization parameter. This brief formulates the mean-variance model as a cardinality ( l0 -norm) constrained nonconvex optimization problem, in which we can explicitly specify the number of assets in the portfolio. We then use the alternating direction method of multipliers (ADMMs) concept to develop an algorithm to solve the constrained nonconvex problem. Unlike some existing algorithms, the proposed algorithm can explicitly control the portfolio cardinality. In addition, the dynamic behavior of the proposed algorithm is derived. Numerical results on four real-world datasets demonstrate the superiority of our approach over several state-of-the-art algorithms.
Collapse
|
5
|
Liu Z, Leung CS, So HC. Formal Convergence Analysis on Deterministic ℓ1-Regularization based Mini-Batch Learning for RBF Networks. Neurocomputing 2023. [DOI: 10.1016/j.neucom.2023.02.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
|
6
|
Lai X, Cao J, Lin Z. An Accelerated Maximally Split ADMM for a Class of Generalized Ridge Regression. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:958-972. [PMID: 34437070 DOI: 10.1109/tnnls.2021.3104840] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Ridge regression (RR) has been commonly used in machine learning, but is facing computational challenges in big data applications. To meet the challenges, this article develops a highly parallel new algorithm, i.e., an accelerated maximally split alternating direction method of multipliers (A-MS-ADMM), for a class of generalized RR (GRR) that allows different regularization factors for different regression coefficients. Linear convergence of the new algorithm along with its convergence ratio is established. Optimal parameters of the algorithm for the GRR with a particular set of regularization factors are derived, and a selection scheme of the algorithm parameters for the GRR with general regularization factors is also discussed. The new algorithm is then applied in the training of single-layer feedforward neural networks. Experimental results on performance validation on real-world benchmark datasets for regression and classification and comparisons with existing methods demonstrate the fast convergence, low computational complexity, and high parallelism of the new algorithm.
Collapse
|
7
|
Zhao Y, Zheng S, Pei J, Yang X. Multiple Discriminant Preserving Support Subspace RBFNNs with Graph Similarity Learning. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.11.038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
8
|
|
9
|
Abstract
To improve the network performance of radial basis function (RBF) and back-propagation (BP) networks on complex nonlinear problems, an integrated neural network model with pre-RBF kernels is proposed. The proposed method is based on the framework of a single optimized BP network and an RBF network. By integrating and connecting the RBF kernel mapping layer and BP neural network, the local features of a sample set can be effectively extracted to improve separability; subsequently, the connected BP network can be used to perform learning and classification in the kernel space. Experiments on an artificial dataset and three benchmark datasets show that the proposed model combines the advantages of RBF and BP networks, as well as improves the performances of the two networks. Finally, the effectiveness of the proposed method is verified.
Collapse
Affiliation(s)
- Hui Wen
- Institute of Electromechanical and Information Engineering, Putian University, Putian, Fujian, China
| | - Tao Yan
- Institute of Electromechanical and Information Engineering, Putian University, Putian, Fujian, China
| | - Zhiqiang Liu
- Institute of Electromechanical and Information Engineering, Putian University, Putian, Fujian, China
| | - Deli Chen
- Institute of Electromechanical and Information Engineering, Putian University, Putian, Fujian, China
| |
Collapse
|
10
|
Lai X, Cao J, Huang X, Wang T, Lin Z. A Maximally Split and Relaxed ADMM for Regularized Extreme Learning Machines. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:1899-1913. [PMID: 31398134 DOI: 10.1109/tnnls.2019.2927385] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
One of the salient features of the extreme learning machine (ELM) is its fast learning speed. However, in a big data environment, the ELM still suffers from an overly heavy computational load due to the high dimensionality and the large amount of data. Using the alternating direction method of multipliers (ADMM), a convex model fitting problem can be split into a set of concurrently executable subproblems, each with just a subset of model coefficients. By maximally splitting across the coefficients and incorporating a novel relaxation technique, a maximally split and relaxed ADMM (MS-RADMM), along with a scalarwise implementation, is developed for the regularized ELM (RELM). The convergence conditions and the convergence rate of the MS-RADMM are established, which exhibits linear convergence with a smaller convergence ratio than the unrelaxed maximally split ADMM. The optimal parameter values of the MS-RADMM are obtained and a fast parameter selection scheme is provided. Experiments on ten benchmark classification data sets are conducted, the results of which demonstrate the fast convergence and parallelism of the MS-RADMM. Complexity comparisons with the matrix-inversion-based method in terms of the numbers of multiplication and addition operations, the computation time and the number of memory cells are provided for performance evaluation of the MS-RADMM.
Collapse
|
11
|
Dong J, Zhao Y, Liu C. Constrained PSO Based Center Selection for RBF Networks Under Concurrent Fault Situation. Neural Process Lett 2020. [DOI: 10.1007/s11063-020-10202-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
12
|
Zhao Y, Pei J, Chen H. Multi-layer radial basis function neural network based on multi-scale kernel learning. Appl Soft Comput 2019. [DOI: 10.1016/j.asoc.2019.105541] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
13
|
Dong J, Zhao Y, Liu C, Han ZF, Leung CS. Orthogonal least squares based center selection for fault-tolerant RBF networks. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.02.039] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|