1
|
Salehi A, Khedmati M. Hybrid clustering strategies for effective oversampling and undersampling in multiclass classification. Sci Rep 2025; 15:3460. [PMID: 39870706 PMCID: PMC11772689 DOI: 10.1038/s41598-024-84786-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Accepted: 12/27/2024] [Indexed: 01/29/2025] Open
Abstract
Multiclass imbalance is a challenging problem in real-world datasets, where certain classes may have a low number of samples because they correspond to rare occurrences. To address the challenge of multiclass imbalance, this paper introduces a novel hybrid cluster-based oversampling and undersampling (HCBOU) technique. By clustering and separating classes into majority and minority categories, this algorithm retains the most information during undersampling while generating efficient data in the minority class. The classification is carried out using one-vs-one and one-vs-all decomposition schemes. Extensive experimentation was carried out on 30 datasets to evaluate the proposed algorithm's performance. The results were subsequently compared with those of several state-of-the-art algorithms. Based on the results, the proposed algorithm outperforms the competing algorithms under different scenarios. Finally, The HCBOU algorithm demonstrated robust performance across varying class imbalance levels, highlighting its effectiveness in handling imbalanced datasets.
Collapse
Affiliation(s)
- Amirreza Salehi
- Department of Industrial Engineering, Sharif University of Technology, Tehran, Iran
| | - Majid Khedmati
- Department of Industrial Engineering, Sharif University of Technology, Azadi Ave., Tehran, 1458889694, Iran.
| |
Collapse
|
2
|
Nguyen TV, Wang M, Maisto D. Editorial: Addressing large scale computing challenges in neuroscience: current advances and future directions. Front Neuroinform 2024; 18:1534396. [PMID: 39741921 PMCID: PMC11685192 DOI: 10.3389/fninf.2024.1534396] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Accepted: 12/03/2024] [Indexed: 01/03/2025] Open
Affiliation(s)
- Tam V. Nguyen
- Department of Computer Science, University of Dayton, Dayton, OH, United States
| | - Min Wang
- School of Information Technology and Systems, University of Canberra, Canberra, ACT, Australia
| | - Domenico Maisto
- Institute of Cognitive Sciences and Technologies, National Research Council of Italy, Rome, Italy
| |
Collapse
|
3
|
Peng X, Wang FY, Li L. MixGradient: A gradient-based re-weighting scheme with mixup for imbalanced data streams. Neural Netw 2023; 161:525-534. [PMID: 36805267 DOI: 10.1016/j.neunet.2023.02.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Revised: 11/22/2022] [Accepted: 02/09/2023] [Indexed: 02/15/2023]
Abstract
A challenge for contemporary deep neural networks in real-world problems is learning from an imbalanced data stream, where data tends to be received chunk by chunk over time, and the prior class distribution is severely imbalanced. Although many sophisticated algorithms have been derived, most of them overlook the importance of gradient information. From this perspective, the difficulty of learning from imbalanced data streams lies in the fact that the gradient estimated on an uneven class distribution is not informative enough to reflect the critical pattern of each class. To this end, we propose to assign higher weights on the training samples whose gradients are close to the gradient of corresponding typical samples, thus highlighting the important samples in minority classes and suppressing the noisy samples in majority classes. Such an idea can be combined with Mixup, which exploits the interpolation information of data to further compensate for the information of sample space that the typical samples do not provide and expand the role of the proposed re-weighting scheme. Experiments on artificially induced long-tailed CIFAR data streams and long-tailed MiniPlaces data stream show that the resulting method, termed MixGradient, boosts the generalization performance of DNNs under different imbalance ratios and achieves up to 10% accuracy improvement.
Collapse
Affiliation(s)
- Xinyu Peng
- Department of Automation, Tsinghua University, Beijing, 100084, China.
| | - Fei-Yue Wang
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, 100080, China.
| | - Li Li
- Department of Automation, Tsinghua University, Beijing, 100084, China; National Laboratory for Information Science and Technology, Tsinghua University, Beijing, 100084, China.
| |
Collapse
|
4
|
Liu X, Zhou Y, Meng W, Luo Q. Functional extreme learning machine for regression and classification. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:3768-3792. [PMID: 36899604 DOI: 10.3934/mbe.2023177] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Although Extreme Learning Machine (ELM) can learn thousands of times faster than traditional slow gradient algorithms for training neural networks, ELM fitting accuracy is limited. This paper develops Functional Extreme Learning Machine (FELM), which is a novel regression and classifier. It takes functional neurons as the basic computing units and uses functional equation-solving theory to guide the modeling process of functional extreme learning machines. The functional neuron function of FELM is not fixed, and its learning process refers to the process of estimating or adjusting the coefficients. It follows the spirit of extreme learning and solves the generalized inverse of the hidden layer neuron output matrix through the principle of minimum error, without iterating to obtain the optimal hidden layer coefficients. To verify the performance of the proposed FELM, it is compared with ELM, OP-ELM, SVM and LSSVM on several synthetic datasets, XOR problem, benchmark regression and classification datasets. The experimental results show that although the proposed FELM has the same learning speed as ELM, its generalization performance and stability are better than ELM.
Collapse
Affiliation(s)
- Xianli Liu
- College of Artificial Intelligence, Guangxi University for Nationalities, Nanning 530006, China
| | - Yongquan Zhou
- College of Artificial Intelligence, Guangxi University for Nationalities, Nanning 530006, China
- Xiangsihu College of Gunagxi University for Nationalities, Nanning, Guangxi 532100, China
- Guangxi Key Laboratories of Hybrid Computation and IC Design Analysis, Nanning 530006, China
| | - Weiping Meng
- College of Artificial Intelligence, Guangxi University for Nationalities, Nanning 530006, China
| | - Qifang Luo
- College of Artificial Intelligence, Guangxi University for Nationalities, Nanning 530006, China
- Guangxi Key Laboratories of Hybrid Computation and IC Design Analysis, Nanning 530006, China
| |
Collapse
|
5
|
Han M, Li A, Gao Z, Mu D, Liu S. A survey of multi-class imbalanced data classification methods. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-221902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
In reality, the data generated in many fields are often imbalanced, such as fraud detection, network intrusion detection and disease diagnosis. The class with fewer instances in the data is called the minority class, and the minority class in some applications contains the significant information. So far, many classification methods and strategies for binary imbalanced data have been proposed, but there are still many problems and challenges in multi-class imbalanced data that need to be solved urgently. The classification methods for multi-class imbalanced data are analyzed and summarized in terms of data preprocessing methods and algorithm-level classification methods, and the performance of the algorithms using the same dataset is compared separately. In the data preprocessing methods, the methods of oversampling, under-sampling, hybrid sampling and feature selection are mainly introduced. Algorithm-level classification methods are comprehensively introduced in four aspects: ensemble learning, neural network, support vector machine and multi-class decomposition technique. At the same time, all data preprocessing methods and algorithm-level classification methods are analyzed in detail in terms of the techniques used, comparison algorithms, pros and cons, respectively. Moreover, the evaluation metrics commonly used for multi-class imbalanced data classification methods are described comprehensively. Finally, the future directions of multi-class imbalanced data classification are given.
Collapse
Affiliation(s)
- Meng Han
- School of Computer Science and Engineering, North Minzu University, Yinchuan, China
| | - Ang Li
- School of Computer Science and Engineering, North Minzu University, Yinchuan, China
| | - Zhihui Gao
- School of Computer Science and Engineering, North Minzu University, Yinchuan, China
| | - Dongliang Mu
- School of Computer Science and Engineering, North Minzu University, Yinchuan, China
| | - Shujuan Liu
- School of Computer Science and Engineering, North Minzu University, Yinchuan, China
| |
Collapse
|
6
|
A multiple classifiers time-serial ensemble pruning algorithm based on the mechanism of forward supplement. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03855-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
7
|
Active learning with extreme learning machine for online imbalanced multiclass classification. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.107385] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
8
|
SMOTE-Based Weighted Deep Rotation Forest for the Imbalanced Hyperspectral Data Classification. REMOTE SENSING 2021. [DOI: 10.3390/rs13030464] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/25/2022]
Abstract
Conventional classification algorithms have shown great success in balanced hyperspectral data classification. However, the imbalanced class distribution is a fundamental problem of hyperspectral data, and it is regarded as one of the great challenges in classification tasks. To solve this problem, a non-ANN based deep learning, namely SMOTE-Based Weighted Deep Rotation Forest (SMOTE-WDRoF) is proposed in this paper. First, the neighboring pixels of instances are introduced as the spatial information and balanced datasets are created by using the SMOTE algorithm. Second, these datasets are fed into the WDRoF model that consists of the rotation forest and the multi-level cascaded random forests. Specifically, the rotation forest is used to generate rotation feature vectors, which are input into the subsequent cascade forest. Furthermore, the output probability of each level and the original data are stacked as the dataset of the next level. And the sample weights are automatically adjusted according to the dynamic weight function constructed by the classification results of each level. Compared with the traditional deep learning approaches, the proposed method consumes much less training time. The experimental results on four public hyperspectral data demonstrate that the proposed method can get better performance than support vector machine, random forest, rotation forest, SMOTE combined rotation forest, convolutional neural network, and rotation-based deep forest in multiclass imbalance learning.
Collapse
|
9
|
Lai X, Cao J, Huang X, Wang T, Lin Z. A Maximally Split and Relaxed ADMM for Regularized Extreme Learning Machines. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:1899-1913. [PMID: 31398134 DOI: 10.1109/tnnls.2019.2927385] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
One of the salient features of the extreme learning machine (ELM) is its fast learning speed. However, in a big data environment, the ELM still suffers from an overly heavy computational load due to the high dimensionality and the large amount of data. Using the alternating direction method of multipliers (ADMM), a convex model fitting problem can be split into a set of concurrently executable subproblems, each with just a subset of model coefficients. By maximally splitting across the coefficients and incorporating a novel relaxation technique, a maximally split and relaxed ADMM (MS-RADMM), along with a scalarwise implementation, is developed for the regularized ELM (RELM). The convergence conditions and the convergence rate of the MS-RADMM are established, which exhibits linear convergence with a smaller convergence ratio than the unrelaxed maximally split ADMM. The optimal parameter values of the MS-RADMM are obtained and a fast parameter selection scheme is provided. Experiments on ten benchmark classification data sets are conducted, the results of which demonstrate the fast convergence and parallelism of the MS-RADMM. Complexity comparisons with the matrix-inversion-based method in terms of the numbers of multiplication and addition operations, the computation time and the number of memory cells are provided for performance evaluation of the MS-RADMM.
Collapse
|
10
|
|
11
|
Shukla S, Raghuwanshi BS. Online sequential class-specific extreme learning machine for binary imbalanced learning. Neural Netw 2019; 119:235-248. [DOI: 10.1016/j.neunet.2019.08.018] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2019] [Revised: 07/03/2019] [Accepted: 08/15/2019] [Indexed: 12/25/2022]
|
12
|
Zhang L, Ray H, Priestley J, Tan S. A descriptive study of variable discretization and cost-sensitive logistic regression on imbalanced credit data. J Appl Stat 2019; 47:568-581. [PMID: 35706966 PMCID: PMC9041569 DOI: 10.1080/02664763.2019.1643829] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
Abstract
Training classification models on imbalanced data tends to result in bias towards the majority class. In this paper, we demonstrate how variable discretization and cost-sensitive logistic regression help mitigate this bias on an imbalanced credit scoring dataset, and further show the application of the variable discretization technique on the data from other domains, demonstrating its potential as a generic technique for classifying imbalanced data beyond credit socring. The performance measurements include ROC curves, Area under ROC Curve (AUC), Type I Error, Type II Error, accuracy, and F1 score. The results show that proper variable discretization and cost-sensitive logistic regression with the best class weights can reduce the model bias and/or variance. From the perspective of the algorithm, cost-sensitive logistic regression is beneficial for increasing the value of predictors even if they are not in their optimized forms while maintaining monotonicity. From the perspective of predictors, the variable discretization performs better than cost-sensitive logistic regression, provides more reasonable coefficient estimates for predictors which have nonlinear relationships against their empirical logit, and is robust to penalty weights on misclassifications of events and non-events determined by their apriori proportions.
Collapse
Affiliation(s)
- Lili Zhang
- Analytics and Data Science Ph.D. Program, Kennesaw State University, Kennesaw, Georgia, USA
| | - Herman Ray
- Analytics and Data Science Institute, Kennesaw State University, Kennesaw, Georgia, USA
| | - Jennifer Priestley
- Analytics and Data Science Institute, Kennesaw State University, Kennesaw, Georgia, USA
| | - Soon Tan
- Ermas Consulting Inc., Alpharetta, Georgia, USA
| |
Collapse
|
13
|
Alaba PA, Popoola SI, Olatomiwa L, Akanle MB, Ohunakin OS, Adetiba E, Alex OD, Atayero AA, Wan Daud WMA. Towards a more efficient and cost-sensitive extreme learning machine: A state-of-the-art review of recent trend. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.03.086] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
14
|
Yu H, Yang X, Zheng S, Sun C. Active Learning From Imbalanced Data: A Solution of Online Weighted Extreme Learning Machine. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2019; 30:1088-1103. [PMID: 30137013 DOI: 10.1109/tnnls.2018.2855446] [Citation(s) in RCA: 35] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
It is well known that active learning can simultaneously improve the quality of the classification model and decrease the complexity of training instances. However, several previous studies have indicated that the performance of active learning is easily disrupted by an imbalanced data distribution. Some existing imbalanced active learning approaches also suffer from either low performance or high time consumption. To address these problems, this paper describes an efficient solution based on the extreme learning machine (ELM) classification model, called active online-weighted ELM (AOW-ELM). The main contributions of this paper include: 1) the reasons why active learning can be disrupted by an imbalanced instance distribution and its influencing factors are discussed in detail; 2) the hierarchical clustering technique is adopted to select initially labeled instances in order to avoid the missed cluster effect and cold start phenomenon as much as possible; 3) the weighted ELM (WELM) is selected as the base classifier to guarantee the impartiality of instance selection in the procedure of active learning, and an efficient online updated mode of WELM is deduced in theory; and 4) an early stopping criterion that is similar to but more flexible than the margin exhaustion criterion is presented. The experimental results on 32 binary-class data sets with different imbalance ratios demonstrate that the proposed AOW-ELM algorithm is more effective and efficient than several state-of-the-art active learning algorithms that are specifically designed for the class imbalance scenario.
Collapse
|
15
|
Liu Z, Loo CK, Seera M. Meta-cognitive Recurrent Recursive Kernel OS-ELM for concept drift handling. Appl Soft Comput 2019. [DOI: 10.1016/j.asoc.2018.11.006] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
16
|
Class Imbalance Ensemble Learning Based on the Margin Theory. APPLIED SCIENCES-BASEL 2018. [DOI: 10.3390/app8050815] [Citation(s) in RCA: 83] [Impact Index Per Article: 11.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
|