1
|
Khan A, Ali A, Islam N, Manzoor S, Zeb H, Azeem M, Ihtesham S. Robust Extreme Learning Machine Using New Activation and Loss Functions Based on M-Estimation for Regression and Classification. SCIENTIFIC PROGRAMMING 2022; 2022:1-10. [DOI: 10.1155/2022/6446080] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/26/2023]
Abstract
This paper provides an analysis of the combining effect of novel activation function and loss function based on M-estimation in application to extreme learning machine (ELM), a feed-forward neural network. Due to the computational efficiency and classification/prediction accuracy of ELM and its variants, they have been widely exploited in the development of new technologies and applications. However, in real applications, the performance of classical ELMs deteriorates in the presence of outliers, thus, negatively impacting the precision and accuracy of the system. To further enhance the performance of ELM and its variants, we proposed novel activation functions based on the psi function of M and redescend the M-estimation method along with the smooth
2-norm weight-loss functions to reduce the negative impact of the outliers. The proposed psi functions of several M and redescending M-estimation methods are more flexible to make more distinct features space. For the first time, the idea of the psi function as an activation function in the neural network is introduced in the literature to ensure accurate prediction. In addition, new robust
2 norm-loss functions based on M and redescending M-estimation are proposed to deal with outliers efficiently in ELM. To evaluate the performance of the proposed methodology against other state-of-the-art techniques, experiments have been performed in diverse environments, which show promising improvements in application to regression and classification problems.
Collapse
Affiliation(s)
- Adnan Khan
- Department of Statistics Islamia College University Peshawar, Peshawar, Pakistan
| | - Amjad Ali
- Department of Statistics Islamia College University Peshawar, Peshawar, Pakistan
| | - Naveed Islam
- Department of Computer Science Islamia College University Peshawar, Peshawar, Pakistan
| | - Sadaf Manzoor
- Department of Statistics Islamia College University Peshawar, Peshawar, Pakistan
| | - Hassan Zeb
- Department of Statistics Islamia College University Peshawar, Peshawar, Pakistan
| | - Muhammad Azeem
- Department of Statistics, University of Malakand, Peshawar, Pakistan
| | - Shumaila Ihtesham
- Department of Statistics Islamia College University Peshawar, Peshawar, Pakistan
| |
Collapse
|
2
|
Multi-Scale LBP Texture Feature Learning Network for Remote Sensing Interpretation of Land Desertification. REMOTE SENSING 2022. [DOI: 10.3390/rs14143486] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Land desertification is a major challenge to global sustainable development. Therefore, the timely and accurate monitoring of the land desertification status can provide scientific decision support for desertification control. The existing automatic interpretation methods are affected by factors such as “same spectrum different matter”, “different spectrum same object”, staggered distribution of desertification areas, and wide ranges of ground objects. We propose an automatic interpretation method for the remote sensing of land desertification that incorporates multi-scale local binary pattern (MSLBP) and spectral features based on the above issues. First, a multi-scale convolutional LBP feature extraction network is designed to obtain the spatial texture features of remote sensing images and fuse them with spectral features to enhance the feature representation capability of the model. Then, considering the continuity of the distribution of the same kind of ground objects in local space, we designed an adaptive median filtering method to process the probability map of the extreme learning machine (ELM) classifier output to improve the classification accuracy. Four typical datasets were developed using GF-1 multispectral imagery with the Horqin Left Wing Rear Banner as the study area. Experimental results on four datasets show that the proposed method solves the problem of ill classification and omission in classifying the remote sensing images of desertification, effectively suppresses the effects of “homospectrum” and “heterospectrum”, and significantly improves the accuracy of the remote sensing interpretation of land desertification.
Collapse
|
3
|
Fuzzy rule dropout with dynamic compensation for wide learning algorithm of TSK fuzzy classifier. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.109410] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
4
|
Prototype Regularized Manifold Regularization Technique for Semi-Supervised Online Extreme Learning Machine. SENSORS 2022; 22:s22093113. [PMID: 35590801 PMCID: PMC9101820 DOI: 10.3390/s22093113] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Revised: 04/09/2022] [Accepted: 04/11/2022] [Indexed: 12/04/2022]
Abstract
Data streaming applications such as the Internet of Things (IoT) require processing or predicting from sequential data from various sensors. However, most of the data are unlabeled, making applying fully supervised learning algorithms impossible. The online manifold regularization approach allows sequential learning from partially labeled data, which is useful for sequential learning in environments with scarcely labeled data. Unfortunately, the manifold regularization technique does not work out of the box as it requires determining the radial basis function (RBF) kernel width parameter. The RBF kernel width parameter directly impacts the performance as it is used to inform the model to which class each piece of data most likely belongs. The width parameter is often determined off-line via hyperparameter search, where a vast amount of labeled data is required. Therefore, it limits its utility in applications where it is difficult to collect a great deal of labeled data, such as data stream mining. To address this issue, we proposed eliminating the RBF kernel from the manifold regularization technique altogether by combining the manifold regularization technique with a prototype learning method, which uses a finite set of prototypes to approximate the entire data set. Compared to other manifold regularization approaches, this approach instead queries the prototype-based learner to find the most similar samples for each sample instead of relying on the RBF kernel. Thus, it no longer necessitates the RBF kernel, which improves its practicality. The proposed approach can learn faster and achieve a higher classification performance than other manifold regularization techniques based on experiments on benchmark data sets. Results showed that the proposed approach can perform well even without using the RBF kernel, which improves the practicality of manifold regularization techniques for semi-supervised learning.
Collapse
|
5
|
|
6
|
Zheng Y, Chen B, Wang S, Wang W, Qin W. Mixture Correntropy-Based Kernel Extreme Learning Machines. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:811-825. [PMID: 33079685 DOI: 10.1109/tnnls.2020.3029198] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Kernel-based extreme learning machine (KELM), as a natural extension of ELM to kernel learning, has achieved outstanding performance in addressing various regression and classification problems. Compared with the basic ELM, KELM has a better generalization ability owing to no needs of the number of hidden nodes given beforehand and random projection mechanism. Since KELM is derived under the minimum mean square error (MMSE) criterion for the Gaussian assumption of noise, its performance may deteriorate under the non-Gaussian cases, seriously. To improve the robustness of KELM, this article proposes a mixture correntropy-based KELM (MC-KELM), which adopts the recently proposed maximum mixture correntropy criterion as the optimization criterion, instead of using the MMSE criterion. In addition, an online sequential version of MC-KELM (MCOS-KELM) is developed to deal with the case that the data arrive sequentially (one-by-one or chunk-by-chunk). Experimental results on regression and classification data sets are reported to validate the performance superiorities of the new methods.
Collapse
|
7
|
Wu D, Li T, Wan Q. A hybrid deep kernel incremental extreme learning machine based on improved coyote and beetle swarm optimization methods. COMPLEX INTELL SYST 2021. [DOI: 10.1007/s40747-021-00486-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
AbstractThe iteration times and learning efficiency of kernel incremental extreme learning machines are always affected by the redundant nodes. A hybrid deep kernel incremental extreme learning machine (DKIELM) based on the improved coyote and beetle swarm optimization methods was proposed in this paper. A hybrid intelligent optimization algorithm based on the improved coyote optimization algorithm (ICOA) and improved beetle swarm optimization algorithm (IBSOA) was proposed to optimize the parameters and determine the number of effectively hidden layer neurons for the proposed DKIELM. A Gaussian global best-growing operator was adopted to replace the original growing operator in the intelligent optimization algorithm to improve COA searching efficiency and convergence. In the meantime, IBSOA was designed based on tent mapping inverse learning and dynamic mutation strategies to avoid falling into a local optimum. The experimental results demonstrated the feasibility and effectiveness of the proposed DKIELM with encouraging performances compared with other ELMs.
Collapse
|
8
|
|
9
|
Kiranyaz S, Malik J, Abdallah HB, Ince T, Iosifidis A, Gabbouj M. Exploiting heterogeneity in operational neural networks by synaptic plasticity. Neural Comput Appl 2021. [DOI: 10.1007/s00521-020-05543-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
AbstractThe recently proposed network model, Operational Neural Networks (ONNs), can generalize the conventional Convolutional Neural Networks (CNNs) that are homogenous only with a linear neuron model. As a heterogenous network model, ONNs are based on a generalized neuron model that can encapsulate any set of non-linear operators to boost diversity and to learn highly complex and multi-modal functions or spaces with minimal network complexity and training data. However, the default search method to find optimal operators in ONNs, the so-called Greedy Iterative Search (GIS) method, usually takes several training sessions to find a single operator set per layer. This is not only computationally demanding, also the network heterogeneity is limited since the same set of operators will then be used for all neurons in each layer. To address this deficiency and exploit a superior level of heterogeneity, in this study the focus is drawn on searching the best-possible operator set(s) for the hidden neurons of the network based on the “Synaptic Plasticity” paradigm that poses the essential learning theory in biological neurons. During training, each operator set in the library can be evaluated by their synaptic plasticity level, ranked from the worst to the best, and an “elite” ONN can then be configured using the top-ranked operator sets found at each hidden layer. Experimental results over highly challenging problems demonstrate that the elite ONNs even with few neurons and layers can achieve a superior learning performance than GIS-based ONNs and as a result, the performance gap over the CNNs further widens.
Collapse
|
10
|
Liu L, Kuang Z, Chen Y, Xue JH, Yang W, Zhang W. IncDet: In Defense of Elastic Weight Consolidation for Incremental Object Detection. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2021; 32:2306-2319. [PMID: 32598286 DOI: 10.1109/tnnls.2020.3002583] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Elastic weight consolidation (EWC) has been successfully applied for general incremental learning to overcome the catastrophic forgetting issue. It adaptively constrains each parameter of the new model not to deviate much from its counterpart in the old model during fine-tuning on new class data sets, according to its importance weight for old tasks. However, the previous study demonstrates that it still suffers from catastrophic forgetting when directly used in object detection. In this article, we show EWC is effective for incremental object detection if with critical adaptations. First, we conduct controlled experiments to identify two core issues why EWC fails if trivially applied to incremental detection: 1) the absence of old class annotations in new class images makes EWC misclassify objects of old classes in these images as background and 2) the quadratic regularization loss in EWC easily leads to gradient explosion when balancing old and new classes. Then, based on the abovementioned findings, we propose the corresponding solutions to tackle these issues: 1) utilize pseudobounding box annotations of old classes on new data sets to compensate for the absence of old class annotations and 2) adopt a novel Huber regularization instead of the original quadratic loss to prevent from unstable training. Finally, we propose a general EWC-based incremental object detection framework and implement it under both Fast R-CNN and Faster R-CNN, showing its flexibility and versatility. In terms of either the final performance or the performance drop with respect to the upper bound of joint training on all seen classes, evaluations on the PASCAL VOC and COCO data sets show that our method achieves a new state of the art.
Collapse
|
11
|
Yu H, Chen C, Yang H. Two-Stage Game Strategy for Multiclass Imbalanced Data Online Prediction. Neural Process Lett 2020. [DOI: 10.1007/s11063-020-10358-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
12
|
Yang Y, Wu QMJ, Feng X, Akilan T. Recomputation of the Dense Layers for Performance Improvement of DCNN. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2020; 42:2912-2925. [PMID: 31107643 DOI: 10.1109/tpami.2019.2917685] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Gradient descent optimization of learning has become a paradigm for training deep convolutional neural networks (DCNN). However, utilizing other learning strategies in the training process of the DCNN has rarely been explored by the deep learning (DL) community. This serves as the motivation to introduce a non-iterative learning strategy to retrain neurons at the top dense or fully connected (FC) layers of DCNN, resulting in, higher performance. The proposed method exploits the Moore-Penrose Inverse to pull back the current residual error to each FC layer, generating well-generalized features. Further, the weights of each FC layers are recomputed according to the Moore-Penrose Inverse. We evaluate the proposed approach on six most widely accepted object recognition benchmark datasets: Scene-15, CIFAR-10, CIFAR-100, SUN-397, Places365, and ImageNet. The experimental results show that the proposed method obtains improvements over 30 state-of-the-art methods. Interestingly, it also indicates that any DCNN with the proposed method can provide better performance than the same network with its original Backpropagation (BP)-based training.
Collapse
|
13
|
Learning local discriminative representations via extreme learning machine for machine fault diagnosis. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.05.021] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
14
|
A Hybrid Method Based on Extreme Learning Machine and Self Organizing Map for Pattern Classification. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2020; 2020:2918276. [PMID: 32908471 PMCID: PMC7468594 DOI: 10.1155/2020/2918276] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/24/2020] [Revised: 06/29/2020] [Accepted: 07/14/2020] [Indexed: 11/18/2022]
Abstract
Extreme learning machine is a fast learning algorithm for single hidden layer feedforward neural network. However, an improper number of hidden neurons and random parameters have a great effect on the performance of the extreme learning machine. In order to select a suitable number of hidden neurons, this paper proposes a novel hybrid learning based on a two-step process. First, the parameters of hidden layer are adjusted by a self-organized learning algorithm. Next, the weights matrix of the output layer is determined using the Moore–Penrose inverse method. Nine classification datasets are considered to demonstrate the efficiency of the proposed approach compared with original extreme learning machine, Tikhonov regularization optimally pruned extreme learning machine, and backpropagation algorithms. The results show that the proposed method is fast and produces better accuracy and generalization performances.
Collapse
|
15
|
Ji H, Wu G, Wang G. Accelerating ELM training over data streams. INT J MACH LEARN CYB 2020. [DOI: 10.1007/s13042-020-01158-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
16
|
|
17
|
Song H, Qin AK, Salim FD. Evolutionary model construction for electricity consumption prediction. Neural Comput Appl 2020. [DOI: 10.1007/s00521-019-04310-w] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
18
|
Chu Y, Lin H, Yang L, Diao Y, Zhang D, Zhang S, Fan X, Shen C, Xu B, Yan D. Discriminative globality-locality preserving extreme learning machine for image classification. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2019.09.013] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
19
|
Abstract
AbstractFeed-forward, fully connected artificial neural networks or the so-called multi-layer perceptrons are well-known universal approximators. However, their learning performance varies significantly depending on the function or the solution space that they attempt to approximate. This is mainly because of their homogenous configuration based solely on the linear neuron model. Therefore, while they learn very well those problems with a monotonous, relatively simple and linearly separable solution space, they may entirely fail to do so when the solution space is highly nonlinear and complex. Sharing the same linear neuron model with two additional constraints (local connections and weight sharing), this is also true for the conventional convolutional neural networks (CNNs) and it is, therefore, not surprising that in many challenging problems only the deep CNNs with a massive complexity and depth can achieve the required diversity and the learning performance. In order to address this drawback and also to accomplish a more generalized model over the convolutional neurons, this study proposes a novel network model, called operational neural networks (ONNs), which can be heterogeneous and encapsulate neurons with any set of operators to boost diversity and to learn highly complex and multi-modal functions or spaces with minimal network complexity and training data. Finally, the training method to back-propagate the error through the operational layers of ONNs is formulated. Experimental results over highly challenging problems demonstrate the superior learning capabilities of ONNs even with few neurons and hidden layers.
Collapse
|
20
|
Zhou J, Jiang Z, Wang S. Laplacian least learning machine with dynamic updating for imbalanced classification. Appl Soft Comput 2020. [DOI: 10.1016/j.asoc.2019.106028] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
21
|
Naik SM, Jagannath RPK, Kuppili V. An automatic estimation of the ridge parameter for extreme learning machine. CHAOS (WOODBURY, N.Y.) 2020; 30:013106. [PMID: 32013505 DOI: 10.1063/1.5097747] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/28/2019] [Accepted: 12/09/2019] [Indexed: 06/10/2023]
Abstract
Extreme learning machine (ELM) is an emerging learning method with a single-hidden layer feed-forward neural network that involves obtaining a solution to the system of linear equations. Unlike traditional gradient-based back-propagating neural networks, ELM is computationally efficient with fast training speed and good generalization capability. However, most of the time when applied to real-time problems, the linear system becomes ill-posed in the structure and needs the inclusion of a ridge parameter to obtain a reliable solution, and hence, the selection of the ridge parameter (C) is a crucial task. The ridge parameter is chosen heuristically from a predefined set. The generalized cross-validation is a widely used technique for the automatic estimation of the same, which is computationally expensive as it involves inversion of large matrices. The focus of the proposed work is on pragmatic aspects of the time-efficient automatic estimation of ridge parameter that result in a better generalization performance. In this work, methods are proposed that use the L-curve and U-curve techniques to automatically estimate the ridge parameter, and these methods are effective in the estimation of the ridge parameter even for systems with larger data. Through extensive numerical results, it is shown that the proposed methods outperform the existing ones in terms of accuracy, precision, sensitivity, specificity, F1-score, and computational time on various benchmark binary as well as multiclass classification data sets. Finally, the proposed methods are statistically analyzed using the nonparametric Friedman ranking test, which is also proving the effectiveness of the proposed method as it is providing a better rank for the same over existing methods.
Collapse
Affiliation(s)
- Shraddha M Naik
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, Goa 403401, India
| | - Ravi Prasad K Jagannath
- Department of Applied Sciences, National Institute of Technology Goa, Ponda, Goa 403401, India
| | - Venkatanareshbabu Kuppili
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, Goa 403401, India
| |
Collapse
|
22
|
|
23
|
Spectral-Spatial Hyperspectral Image Classification with Superpixel Pattern and Extreme Learning Machine. REMOTE SENSING 2019. [DOI: 10.3390/rs11171983] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Spectral-spatial classification of hyperspectral images (HSIs) has recently attracted great attention in the research domain of remote sensing. It is well-known that, in remote sensing applications, spectral features are the fundamental information and spatial patterns provide the complementary information. With both spectral features and spatial patterns, hyperspectral image (HSI) applications can be fully explored and the classification performance can be greatly improved. In reality, spatial patterns can be extracted to represent a line, a clustering of points or image texture, which denote the local or global spatial characteristic of HSIs. In this paper, we propose a spectral-spatial HSI classification model based on superpixel pattern (SP) and kernel based extreme learning machine (KELM), called SP-KELM, to identify the land covers of pixels in HSIs. In the proposed SP-KELM model, superpixel pattern features are extracted by an advanced principal component analysis (PCA), which is based on superpixel segmentation in HSIs and used to denote spatial information. The KELM method is then employed to be a classifier in the proposed spectral-spatial model with both the original spectral features and the extracted spatial pattern features. Experimental results on three publicly available HSI datasets verify the effectiveness of the proposed SP-KELM model, with the performance improvement of 10% over the spectral approaches.
Collapse
|
24
|
Individual-Specific Classification of Mental Workload Levels Via an Ensemble Heterogeneous Extreme Learning Machine for EEG Modeling. Symmetry (Basel) 2019. [DOI: 10.3390/sym11070944] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
In a human–machine cooperation system, assessing the mental workload (MW) of the human operator is quite crucial to maintaining safe operation conditions. Among various MW indicators, electroencephalography (EEG) signals are particularly attractive because of their high temporal resolution and sensitivity to the occupation of working memory. However, the individual difference of the EEG feature distribution may impair the machine-learning based MW classifier. In this paper, we employed a fast-training neural network, extreme learning machine (ELM), as the basis to build an individual-specific classifier ensemble to recognize binary MW. To improve the diversity of the classification committee, heterogeneous member classifiers were adopted by fusing multiple ELMs and Bayesian models. Specifically, a deep network structure was applied in each weak model aiming at finding informative EEG feature representations. The structure of hyper-parameters of the proposed heterogeneous ensemble ELM (HE-ELM) was then identified and then its performance was compared against several competitive MW classifiers. We found that the HE-ELM model was superior for improving the individual-specific accuracy of MW assessments.
Collapse
|
25
|
Jin L, Huang Z, Chen L, Liu M, Li Y, Chou Y, Yi C. Modified single-output Chebyshev-polynomial feedforward neural network aided with subset method for classification of breast cancer. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.03.046] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
|
26
|
|
27
|
Rodriguez N, Barba L, Alvarez P, Cabrera-Guerrero G. Stationary Wavelet-Fourier Entropy and Kernel Extreme Learning for Bearing Multi-Fault Diagnosis. ENTROPY 2019; 21:e21060540. [PMID: 33267254 PMCID: PMC7515029 DOI: 10.3390/e21060540] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/08/2019] [Revised: 05/23/2019] [Accepted: 05/26/2019] [Indexed: 11/16/2022]
Abstract
Bearing fault diagnosis methods play an important role in rotating machine health monitoring. In recent years, various intelligent fault diagnosis methods have been proposed, which are mainly based on the features extraction method combined with either shallow or deep learning methods. During the last few years, Shannon entropy features have been widely used in machine health monitoring, improving the accuracy of the bearing fault diagnosis process. Therefore, in this paper, we consider the combination of multi-scale stationary wavelet packet analysis with the Fourier amplitude spectrum to obtain a new discriminative Shannon entropy feature that we call stationary wavelet packet Fourier entropy (SWPFE). Features extracted by our SWPFE method are then passed onto a shallow kernel extreme learning machine (KELM) classifier to diagnose bearing failure types with different severities. The proposed method was applied on two experimental vibration signal databases of a rolling element bearing and compared to two recently proposed methods called stationary wavelet packet permutation entropy (SWPPE) and stationary wavelet packet dispersion entropy (SWPPE). Based on our results, we can say that the proposed method is able to achieve better accuracy levels than both the SWPPE and SWPDE methods using fewer failure features. Further, as our method does not require any hyperparameter calibration step, it is less dependent on user experience/expertise.
Collapse
Affiliation(s)
- Nibaldo Rodriguez
- Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2374631, Chile
- Correspondence: ; Tel.: +56-32-227-3761
| | - Lida Barba
- Facultad de Ingeniería, Universidad Nacional de Chimborazo, Chimborazo 060108, Ecuador
| | - Pablo Alvarez
- Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2374631, Chile
| | - Guillermo Cabrera-Guerrero
- Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2374631, Chile
| |
Collapse
|
28
|
|
29
|
Combining Multi-Scale Wavelet Entropy and Kernelized Classification for Bearing Multi-Fault Diagnosis. ENTROPY 2019; 21:e21020152. [PMID: 33266868 PMCID: PMC7514634 DOI: 10.3390/e21020152] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/09/2019] [Revised: 01/26/2019] [Accepted: 01/26/2019] [Indexed: 11/20/2022]
Abstract
Discriminative feature extraction and rolling element bearing failure diagnostics are very important to ensure the reliability of rotating machines. Therefore, in this paper, we propose multi-scale wavelet Shannon entropy as a discriminative fault feature to improve the diagnosis accuracy of bearing fault under variable work conditions. To compute the multi-scale wavelet entropy, we consider integrating stationary wavelet packet transform with both dispersion (SWPDE) and permutation (SWPPE) entropies. The multi-scale entropy features extracted by our proposed methods are then passed on to the kernel extreme learning machine (KELM) classifier to diagnose bearing failure types with different severities. In the end, both the SWPDE–KELM and the SWPPE–KELM methods are evaluated on two bearing vibration signal databases. We compare these two feature extraction methods to a recently proposed method called stationary wavelet packet singular value entropy (SWPSVE). Based on our results, we can say that the diagnosis accuracy obtained by the SWPDE–KELM method is slightly better than the SWPPE–KELM method and they both significantly outperform the SWPSVE–KELM method.
Collapse
|
30
|
Zhang Y, Wu J, Cai Z, Du B, Yu PS. An unsupervised parameter learning model for RVFL neural network. Neural Netw 2019; 112:85-97. [PMID: 30771727 DOI: 10.1016/j.neunet.2019.01.007] [Citation(s) in RCA: 44] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2018] [Revised: 12/30/2018] [Accepted: 01/18/2019] [Indexed: 01/18/2023]
Abstract
With the direct input-output connections, a random vector functional link (RVFL) network is a simple and effective learning algorithm for single-hidden layer feedforward neural networks (SLFNs). RVFL is a universal approximator for continuous functions on compact sets with fast learning property. Owing to its simplicity and effectiveness, RVFL has attracted significant interest in numerous real-world applications. In reality, the performance of RVFL is often challenged by randomly assigned network parameters. In this paper, we propose a novel unsupervised network parameter learning method for RVFL, named sparse pre-trained random vector functional link (SP-RVFL for short) network. The proposed SP-RVFL uses a sparse autoencoder with ℓ1-norm regularization to adaptively learn superior network parameters for specific learning tasks. By doing so, the learned network parameters in SP-RVFL are embedded with the valuable information of input data, which alleviate the randomly generated parameter issue and improve the algorithmic performance. Experiments and comparisons on 16 diverse benchmarks from different domains confirm the effectiveness of the proposed SP-RVFL. The corresponding results also demonstrate that RVFL outperforms extreme learning machine (ELM).
Collapse
Affiliation(s)
- Yongshan Zhang
- School of Computer Science, China University of Geosciences, Wuhan 430074, China.
| | - Jia Wu
- Department of Computing, Faculty of Science and Engineering, Macquarie University, Sydney NSW 2109, Australia.
| | - Zhihua Cai
- School of Computer Science, China University of Geosciences, Wuhan 430074, China.
| | - Bo Du
- School of Computer Science, Wuhan University, Wuhan, 430072, China.
| | - Philip S Yu
- Department of Computer Science, University of Illinois at Chicago, Chicago, IL 60607, USA.
| |
Collapse
|
31
|
|
32
|
|
33
|
Kale A, Sonavane S. F-WSS $$^{++}$$ + + : incremental wrapper subset selection algorithm for fuzzy extreme learning machine. INT J MACH LEARN CYB 2018. [DOI: 10.1007/s13042-018-0859-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
|
34
|
Fixed-Size Extreme Learning Machines Through Simulated Annealing. Neural Process Lett 2018. [DOI: 10.1007/s11063-017-9700-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
35
|
Zhang L, Zhang D. Evolutionary Cost-Sensitive Extreme Learning Machine. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2017; 28:3045-3060. [PMID: 27740499 DOI: 10.1109/tnnls.2016.2607757] [Citation(s) in RCA: 48] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Conventional extreme learning machines (ELMs) solve a Moore-Penrose generalized inverse of hidden layer activated matrix and analytically determine the output weights to achieve generalized performance, by assuming the same loss from different types of misclassification. The assumption may not hold in cost-sensitive recognition tasks, such as face recognition-based access control system, where misclassifying a stranger as a family member may result in more serious disaster than misclassifying a family member as a stranger. Though recent cost-sensitive learning can reduce the total loss with a given cost matrix that quantifies how severe one type of mistake against another, in many realistic cases, the cost matrix is unknown to users. Motivated by these concerns, this paper proposes an evolutionary cost-sensitive ELM, with the following merits: 1) to the best of our knowledge, it is the first proposal of ELM in evolutionary cost-sensitive classification scenario; 2) it well addresses the open issue of how to define the cost matrix in cost-sensitive learning tasks; and 3) an evolutionary backtracking search algorithm is induced for adaptive cost matrix optimization. Experiments in a variety of cost-sensitive tasks well demonstrate the effectiveness of the proposed approaches, with about 5%-10% improvements.
Collapse
|
36
|
Salaken SM, Khosravi A, Nguyen T, Nahavandi S. Extreme learning machine based transfer learning algorithms: A survey. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2017.06.037] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
37
|
|
38
|
Short-Term Electricity-Load Forecasting Using a TSK-Based Extreme Learning Machine with Knowledge Representation. ENERGIES 2017. [DOI: 10.3390/en10101613] [Citation(s) in RCA: 29] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
39
|
Stationary Wavelet Singular Entropy and Kernel Extreme Learning for Bearing Multi-Fault Diagnosis. ENTROPY 2017. [DOI: 10.3390/e19100541] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
40
|
Yin Y, Zhao Y, Zhang B, Li C, Guo S. Enhancing ELM by Markov Boundary based feature selection. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.09.119] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
41
|
Discriminative extreme learning machine with supervised sparsity preserving for image classification. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.05.113] [Citation(s) in RCA: 65] [Impact Index Per Article: 8.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
42
|
Sun Y, Chen Y, Yuan Y, Wang G. Dynamic adjustment of hidden layer structure for convex incremental extreme learning machine. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.07.072] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
43
|
Zhao YP, Li ZQ, Xi PP, Liang D, Sun L, Chen TH. Gram–Schmidt process based incremental extreme learning machine. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2017.01.049] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
44
|
Han F, Zhao MR, Zhang JM, Ling QH. An improved incremental constructive single-hidden-layer feedforward networks for extreme learning machine based on particle swarm optimization. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.09.092] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
45
|
You ZH, Zhou M, Luo X, Li S. Highly Efficient Framework for Predicting Interactions Between Proteins. IEEE TRANSACTIONS ON CYBERNETICS 2017; 47:731-743. [PMID: 28113829 DOI: 10.1109/tcyb.2016.2524994] [Citation(s) in RCA: 36] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Protein-protein interactions (PPIs) play a central role in many biological processes. Although a large amount of human PPI data has been generated by high-throughput experimental techniques, they are very limited compared to the estimated 130 000 protein interactions in humans. Hence, automatic methods for human PPI-detection are highly desired. This work proposes a novel framework, i.e., Low-rank approximation-kernel Extreme Learning Machine (LELM), for detecting human PPI from a protein's primary sequences automatically. It has three main steps: 1) mapping each protein sequence into a matrix built on all kinds of adjacent amino acids; 2) applying the low-rank approximation model to the obtained matrix to solve its lowest rank representation, which reflects its true subspace structures; and 3) utilizing a powerful kernel extreme learning machine to predict the probability for PPI based on this lowest rank representation. Experimental results on a large-scale human PPI dataset demonstrate that the proposed LELM has significant advantages in accuracy and efficiency over the state-of-art approaches. Hence, this work establishes a new and effective way for the automatic detection of PPI.
Collapse
|
46
|
Pratama M, Zhang G, Er MJ, Anavatti S. An Incremental Type-2 Meta-Cognitive Extreme Learning Machine. IEEE TRANSACTIONS ON CYBERNETICS 2017; 47:339-353. [PMID: 26812744 DOI: 10.1109/tcyb.2016.2514537] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Existing extreme learning algorithm have not taken into account four issues: 1) complexity; 2) uncertainty; 3) concept drift; and 4) high dimensionality. A novel incremental type-2 meta-cognitive extreme learning machine (ELM) called evolving type-2 ELM (eT2ELM) is proposed to cope with the four issues in this paper. The eT2ELM presents three main pillars of human meta-cognition: 1) what-to-learn; 2) how-to-learn; and 3) when-to-learn. The what-to-learn component selects important training samples for model updates by virtue of the online certainty-based active learning method, which renders eT2ELM as a semi-supervised classifier. The how-to-learn element develops a synergy between extreme learning theory and the evolving concept, whereby the hidden nodes can be generated and pruned automatically from data streams with no tuning of hidden nodes. The when-to-learn constituent makes use of the standard sample reserved strategy. A generalized interval type-2 fuzzy neural network is also put forward as a cognitive component, in which a hidden node is built upon the interval type-2 multivariate Gaussian function while exploiting a subset of Chebyshev series in the output node. The efficacy of the proposed eT2ELM is numerically validated in 12 data streams containing various concept drifts. The numerical results are confirmed by thorough statistical tests, where the eT2ELM demonstrates the most encouraging numerical results in delivering reliable prediction, while sustaining low complexity.
Collapse
|
47
|
Iosifidis A, Tefas A, Pitas I. Approximate kernel extreme learning machine for large scale data classification. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.09.023] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
48
|
Yang Y, Wu QMJ. Extreme Learning Machine With Subnetwork Hidden Nodes for Regression and Classification. IEEE TRANSACTIONS ON CYBERNETICS 2016; 46:2885-2898. [PMID: 26552104 DOI: 10.1109/tcyb.2015.2492468] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
As demonstrated earlier, the learning effectiveness and learning speed of single-hidden-layer feedforward neural networks are in general far slower than required, which has been a major bottleneck for many applications. Huang et al. proposed extreme learning machine (ELM) which improves the training speed by hundreds of times as compared to its predecessor learning techniques. This paper offers an ELM-based learning method that can grow subnetwork hidden nodes by pulling back residual network error to the hidden layer. Furthermore, the proposed method provides a similar or better generalization performance with remarkably fewer hidden nodes as compared to other ELM methods employing huge number of hidden nodes. Thus, the learning speed of the proposed technique is hundred times faster compared to other ELMs as well as to back propagation and support vector machines. The experimental validations for all methods are carried out on 32 data sets.
Collapse
|
49
|
Ling QH, Song YQ, Han F, Yang D, Huang DS. An Improved Ensemble of Random Vector Functional Link Networks Based on Particle Swarm Optimization with Double Optimization Strategy. PLoS One 2016; 11:e0165803. [PMID: 27835638 PMCID: PMC5106042 DOI: 10.1371/journal.pone.0165803] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2016] [Accepted: 10/18/2016] [Indexed: 11/19/2022] Open
Abstract
For ensemble learning, how to select and combine the candidate classifiers are two key issues which influence the performance of the ensemble system dramatically. Random vector functional link networks (RVFL) without direct input-to-output links is one of suitable base-classifiers for ensemble systems because of its fast learning speed, simple structure and good generalization performance. In this paper, to obtain a more compact ensemble system with improved convergence performance, an improved ensemble of RVFL based on attractive and repulsive particle swarm optimization (ARPSO) with double optimization strategy is proposed. In the proposed method, ARPSO is applied to select and combine the candidate RVFL. As for using ARPSO to select the optimal base RVFL, ARPSO considers both the convergence accuracy on the validation data and the diversity of the candidate ensemble system to build the RVFL ensembles. In the process of combining RVFL, the ensemble weights corresponding to the base RVFL are initialized by the minimum norm least-square method and then further optimized by ARPSO. Finally, a few redundant RVFL is pruned, and thus the more compact ensemble of RVFL is obtained. Moreover, in this paper, theoretical analysis and justification on how to prune the base classifiers on classification problem is presented, and a simple and practically feasible strategy for pruning redundant base classifiers on both classification and regression problems is proposed. Since the double optimization is performed on the basis of the single optimization, the ensemble of RVFL built by the proposed method outperforms that built by some single optimization methods. Experiment results on function approximation and classification problems verify that the proposed method could improve its convergence accuracy as well as reduce the complexity of the ensemble system.
Collapse
Affiliation(s)
- Qing-Hua Ling
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, China
- School of Computer Science and Engineering, Jiangsu University of Science and Technology, Zhenjiang, China
- * E-mail:
| | - Yu-Qing Song
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, China
| | - Fei Han
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, China
| | - Dan Yang
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, China
| | - De-Shuang Huang
- School of Electronics and Information Engineering, Tongji University, Shanghai, China
| |
Collapse
|
50
|
Prieto A, Prieto B, Ortigosa EM, Ros E, Pelayo F, Ortega J, Rojas I. Neural networks: An overview of early research, current frameworks and new challenges. Neurocomputing 2016. [DOI: 10.1016/j.neucom.2016.06.014] [Citation(s) in RCA: 161] [Impact Index Per Article: 17.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/24/2023]
|