1
|
Liu H, Zhou Y, Liu B, Zhao J, Yao R, Shao Z. Incremental learning with neural networks for computer vision: a survey. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10294-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
2
|
Agarwal S, Rattani A, Chowdary CR. A-iLearn: An adaptive incremental learning model for spoof fingerprint detection. MACHINE LEARNING WITH APPLICATIONS 2022. [DOI: 10.1016/j.mlwa.2021.100210] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022] Open
|
3
|
Wu Z, Gao P, Cui L, Chen J. An Incremental Learning Method Based on Dynamic Ensemble RVM for Intrusion Detection. IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT 2022. [DOI: 10.1109/tnsm.2021.3102388] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
|
4
|
Geng C, Huang SJ, Chen S. Recent Advances in Open Set Recognition: A Survey. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2021; 43:3614-3631. [PMID: 32191881 DOI: 10.1109/tpami.2020.2981604] [Citation(s) in RCA: 89] [Impact Index Per Article: 22.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/21/2023]
Abstract
In real-world recognition/classification tasks, limited by various objective factors, it is usually difficult to collect training samples to exhaust all classes when training a recognizer or classifier. A more realistic scenario is open set recognition (OSR), where incomplete knowledge of the world exists at training time, and unknown classes can be submitted to an algorithm during testing, requiring the classifiers to not only accurately classify the seen classes, but also effectively deal with unseen ones. This paper provides a comprehensive survey of existing open set recognition techniques covering various aspects ranging from related definitions, representations of models, datasets, evaluation criteria, and algorithm comparisons. Furthermore, we briefly analyze the relationships between OSR and its related tasks including zero-shot, one-shot (few-shot) recognition/learning techniques, classification with reject option, and so forth. Additionally, we also review the open world recognition which can be seen as a natural extension of OSR. Importantly, we highlight the limitations of existing approaches and point out some promising subsequent research directions in this field.
Collapse
|
5
|
Abstract
In this paper, we present a novel incremental learning technique to solve the catastrophic forgetting problem observed in the CNN architectures. We used a progressive deep neural network to incrementally learn new classes while keeping the performance of the network unchanged on old classes. The incremental training requires us to train the network only for new classes and fine-tune the final fully connected layer, without needing to train the entire network again, which significantly reduces the training time. We evaluate the proposed architecture extensively on image classification task using Fashion MNIST, CIFAR-100 and ImageNet-1000 datasets. Experimental results show that the proposed network architecture not only alleviates catastrophic forgetting but can also leverages prior knowledge via lateral connections to previously learned classes and their features. In addition, the proposed scheme is easily scalable and does not require structural changes on the network trained on the old task, which are highly required properties in embedded systems.
Collapse
|
6
|
Din SU, Shao J, Kumar J, Mawuli CB, Mahmud SMH, Zhang W, Yang Q. Data stream classification with novel class detection: a review, comparison and challenges. Knowl Inf Syst 2021. [DOI: 10.1007/s10115-021-01582-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
7
|
Hammami Z, Sayed-Mouchaweh M, Mouelhi W, Ben Said L. Neural networks for online learning of non-stationary data streams: a review and application for smart grids flexibility improvement. Artif Intell Rev 2020. [DOI: 10.1007/s10462-020-09844-3] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
8
|
EnsPKDE&IncLKDE: a hybrid time series prediction algorithm integrating dynamic ensemble pruning, incremental learning, and kernel density estimation. APPL INTELL 2020. [DOI: 10.1007/s10489-020-01802-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
9
|
Pławiak P, Abdar M, Rajendra Acharya U. Application of new deep genetic cascade ensemble of SVM classifiers to predict the Australian credit scoring. Appl Soft Comput 2019. [DOI: 10.1016/j.asoc.2019.105740] [Citation(s) in RCA: 83] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
10
|
Abstract
Abstract
Mining and analysing streaming data is crucial for many applications, and this area of research has gained extensive attention over the past decade. However, there are several inherent problems that continue to challenge the hardware and the state-of-the art algorithmic solutions. Examples of such problems include the unbound size, varying speed and unknown data characteristics of arriving instances from a data stream. The aim of this research is to portray key challenges faced by algorithmic solutions for stream mining, particularly focusing on the prevalent issue of concept drift. A comprehensive discussion of concept drift and its inherent data challenges in the context of stream mining is presented, as is a critical, in-depth review of relevant literature. Current issues with the evaluative procedure for concept drift detectors is also explored, highlighting problems such as a lack of established base datasets and the impact of temporal dependence on concept drift detection. By exposing gaps in the current literature, this study suggests recommendations for future research which should aid in the progression of stream mining and concept drift detection algorithms.
Collapse
|
11
|
Yu H, Webb GI. Adaptive online extreme learning machine by regulating forgetting factor by concept drift map. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2018.11.098] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
12
|
Yu Z, Wang D, Zhao Z, Chen CLP, You J, Wong HS, Zhang J. Hybrid Incremental Ensemble Learning for Noisy Real-World Data Classification. IEEE TRANSACTIONS ON CYBERNETICS 2019; 49:403-416. [PMID: 29990215 DOI: 10.1109/tcyb.2017.2774266] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Traditional ensemble learning approaches explore the feature space and the sample space, respectively, which will prevent them to construct more powerful learning models for noisy real-world dataset classification. The random subspace method only search for the selection of features. Meanwhile, the bagging approach only search for the selection of samples. To overcome these limitations, we propose the hybrid incremental ensemble learning (HIEL) approach which takes into consideration the feature space and the sample space simultaneously to handle noisy dataset. Specifically, HIEL first adopts the bagging technique and linear discriminant analysis to remove noisy attributes, and generates a set of bootstraps and the corresponding ensemble members in the subspaces. Then, the classifiers are selected incrementally based on a classifier-specific criterion function and an ensemble criterion function. The corresponding weights for the classifiers are assigned during the same process. Finally, the final label is summarized by a weighted voting scheme, which serves as the final result of the classification. We also explore various classifier-specific criterion functions based on different newly proposed similarity measures, which will alleviate the effect of noisy samples on the distance functions. In addition, the computational cost of HIEL is analyzed theoretically. A set of nonparametric tests are adopted to compare HIEL and other algorithms over several datasets. The experiment results show that HIEL performs well on the noisy datasets. HIEL outperforms most of the compared classifier ensemble methods on 14 out of 24 noisy real-world UCI and KEEL datasets.
Collapse
|
13
|
Li C, Wei F, Dong W, Wang X, Liu Q, Zhang X. Dynamic Structure Embedded Online Multiple-Output Regression for Streaming Data. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2019; 41:323-336. [PMID: 29994559 DOI: 10.1109/tpami.2018.2794446] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Online multiple-output regression is an important machine learning technique for modeling, predicting, and compressing multi-dimensional correlated data streams. In this paper, we propose a novel online multiple-output regression method, called MORES, for streaming data. MORES can dynamically learn the structure of the regression coefficients to facilitate the model's continuous refinement. Considering that limited expressive ability of regression models often leading to residual errors being dependent, MORES intends to dynamically learn and leverage the structure of the residual errors to improve the prediction accuracy. Moreover, we introduce three modified covariance matrices to extract necessary information from all the seen data for training, and set different weights on samples so as to track the data streams' evolving characteristics. Furthermore, an efficient algorithm is designed to optimize the proposed objective function, and an efficient online eigenvalue decomposition algorithm is developed for the modified covariance matrix. Finally, we analyze the convergence of MORES in certain ideal condition. Experiments on two synthetic datasets and three real-world datasets validate the effectiveness and efficiency of MORES. In addition, MORES can process at least 2,000 instances per second (including training and testing) on the three real-world datasets, more than 12 times faster than the state-of-the-art online learning algorithm.
Collapse
|
14
|
Bang SH, Ak R, Narayanan A, Lee YT, Cho H. A Survey on Knowledge Transfer for Manufacturing Data Analytics. COMPUT IND 2019; 104:10.1016/j.compind.2018.07.001. [PMID: 39440000 PMCID: PMC11495017 DOI: 10.1016/j.compind.2018.07.001] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Abstract
Data analytics techniques have been used for numerous manufacturing applications in various areas. A common assumption of data analytics models is that the environment that generates data is stationary, that is, the feature (or label) space or distribution of the data does not change over time. However, in the real world, this assumption is not valid especially for manufacturing. In non-stationary environments, the accuracy of the model decreases over time, so the model must be retrained periodically and adapted to the corresponding environment(s). Knowledge transfer for data analytics is an approach that trains a model with knowledge extracted from data or model(s). Knowledge transfer can be used when adapting to a new environment, while reducing or eliminating degradation in the accuracy of the model. This paper surveys knowledge transfer methods that have been widely used in various applications, and investigates the applicability of these methods for manufacturing problems. The surveyed knowledge transfer methods are analyzed from three viewpoints: types of non-stationary environments, availability of labeled data, and sources of knowledge. In addition, we categorize events that cause non-stationary environments in manufacturing, and present a mechanism to enable practitioners to select the appropriate methods for their manufacturing data analytics applications among the surveyed knowledge transfer methods. The mechanism includes the steps 1) to detect changes in data properties, 2) to define source and target, and 3) to select available knowledge transfer methods. By providing comprehensive information, this paper will support researchers to adopt knowledge transfer in manufacturing.
Collapse
Affiliation(s)
- Seung Hwan Bang
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
- Department of Industrial and Management Engineering, Pohang University of Science and Technology (POSTECH), Pohang, Gyeongbuk, 37673, Republic of Korea
| | - Ronay Ak
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
| | - Anantha Narayanan
- Department of Mechanical Engineering, University of Maryland, College Park, MD, 20742, USA
| | - Y Tina Lee
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
| | - Hyunbo Cho
- Department of Industrial and Management Engineering, Pohang University of Science and Technology (POSTECH), Pohang, Gyeongbuk, 37673, Republic of Korea
| |
Collapse
|
15
|
Abstract
Deep learning-based methods have reached state of the art performances, relying on a large quantity of available data and computational power. Such methods still remain highly inappropriate when facing a major open machine learning problem, which consists of learning incrementally new classes and examples over time. Combining the outstanding performances of Deep Neural Networks (DNNs) with the flexibility of incremental learning techniques is a promising venue of research. In this contribution, we introduce Transfer Incremental Learning using Data Augmentation (TILDA). TILDA is based on pre-trained DNNs as feature extractors, robust selection of feature vectors in subspaces using a nearest-class-mean based technique, majority votes and data augmentation at both the training and the prediction stages. Experiments on challenging vision datasets demonstrate the ability of the proposed method for low complexity incremental learning, while achieving significantly better accuracy than existing incremental counterparts.
Collapse
|
16
|
Beheshti M, Ashapure A, Rahnemoonfar M, Faichney J. Fluorescence microscopy image segmentation based on graph and fuzzy methods: A comparison with ensemble method. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2018. [DOI: 10.3233/jifs-17466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Affiliation(s)
- Maedeh Beheshti
- School of Information and Communication Technology, Griffith University, Australia
| | - Akash Ashapure
- College of Science and Engineering, Texas A&M University-Corpus Christi, USA
| | - Maryam Rahnemoonfar
- College of Science and Engineering, Texas A&M University-Corpus Christi, USA
| | - Jolon Faichney
- School of Information and Communication Technology, Griffith University, Australia
| |
Collapse
|
17
|
Sultan Zia M, Hussain M, Arfan Jaffar M. Incremental Learning-Based Facial Expression Classification System Using a Novel Multinomial Classifier. INT J PATTERN RECOGN 2017. [DOI: 10.1142/s0218001418560049] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Facial expressions recognition is a crucial task in pattern recognition and it becomes even crucial when cross-cultural emotions are encountered. Various studies in the past have shown that all the facial expressions are not innate and universal, but many of them are learned and culture-dependent. Extreme facial expression recognition methods employ different datasets for training and later use it for testing and demostrate high accuracy in recognition. Their performances degrade drastically when expression images are taken from different cultures. Moreover, there are many existing facial expression patterns which cannot be generated and used as training data in single training session. A facial expression recognition system can maintain its high accuracy and robustness globally and for a longer period if the system possesses the ability to learn incrementally. We also propose a novel classification algorithm for multinomial classification problems. It is an efficient classifier and can be a good choice for base classifier in real-time applications. We propose a facial expression recognition system that can learn incrementally. We use Local Binary Pattern (LBP) features to represent the expression space. The performance of the system is tested on static images from six different databases containing expressions from various cultures. The experiments using the incremental learning classification demonstrate promising results.
Collapse
Affiliation(s)
- M. Sultan Zia
- COMSATS, Institute of Information Technology, Sahiwal, Pakistan
| | - Majid Hussain
- COMSATS, Institute of Information Technology, Sahiwal, Pakistan
- Al Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - M. Arfan Jaffar
- Al Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| |
Collapse
|
18
|
Ngo Ho AK, Eglin V, Ragot N, Ramel JY. A multi-one-class dynamic classifier for adaptive digitization of document streams. INT J DOC ANAL RECOG 2017. [DOI: 10.1007/s10032-017-0286-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
19
|
Bolón-Canedo V, Sánchez-Maroño N, Alonso-Betanzos A. Feature selection for high-dimensional data. PROGRESS IN ARTIFICIAL INTELLIGENCE 2016. [DOI: 10.1007/s13748-015-0080-y] [Citation(s) in RCA: 31] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
20
|
Logistic Regression Learning Model for Handling Concept Drift with Unbalanced Data in Credit Card Fraud Detection System. ADVANCES IN INTELLIGENT SYSTEMS AND COMPUTING 2016. [DOI: 10.1007/978-81-322-2523-2_66] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|
21
|
Martínez-Rego D, Fernández-Francos D, Fontenla-Romero O, Alonso-Betanzos A. Stream change detection via passive-aggressive classification and Bernoulli CUSUM. Inf Sci (N Y) 2015. [DOI: 10.1016/j.ins.2015.01.022] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
|
22
|
Ahmad A, Brown G. Random Ordinality Ensembles: Ensemble methods for multi-valued categorical data. Inf Sci (N Y) 2015. [DOI: 10.1016/j.ins.2014.10.064] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
23
|
Yu Z, Li L, Liu J, Han G. Hybrid adaptive classifier ensemble. IEEE TRANSACTIONS ON CYBERNETICS 2015; 45:177-190. [PMID: 24860045 DOI: 10.1109/tcyb.2014.2322195] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Traditional random subspace-based classifier ensemble approaches (RSCE) have several limitations, such as viewing the same importance for the base classifiers trained in different subspaces, not considering how to find the optimal random subspace set. In this paper, we design a general hybrid adaptive ensemble learning framework (HAEL), and apply it to address the limitations of RSCE. As compared with RSCE, HAEL consists of two adaptive processes, i.e., base classifier competition and classifier ensemble interaction, so as to adjust the weights of the base classifiers in each ensemble and to explore the optimal random subspace set simultaneously. The experiments on the real-world datasets from the KEEL dataset repository for the classification task and the cancer gene expression profiles show that: 1) HAEL works well on both the real-world KEEL datasets and the cancer gene expression profiles and 2) it outperforms most of the state-of-the-art classifier ensemble approaches on 28 out of 36 KEEL datasets and 6 out of 6 cancer datasets.
Collapse
|
24
|
Riccardi A, Fernández-Navarro F, Carloni S. Cost-sensitive AdaBoost algorithm for ordinal regression based on extreme learning machine. IEEE TRANSACTIONS ON CYBERNETICS 2014; 44:1898-1909. [PMID: 25222730 DOI: 10.1109/tcyb.2014.2299291] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
In this paper, the well known stagewise additive modeling using a multiclass exponential (SAMME) boosting algorithm is extended to address problems where there exists a natural order in the targets using a cost-sensitive approach. The proposed ensemble model uses an extreme learning machine (ELM) model as a base classifier (with the Gaussian kernel and the additional regularization parameter). The closed form of the derived weighted least squares problem is provided, and it is employed to estimate analytically the parameters connecting the hidden layer to the output layer at each iteration of the boosting algorithm. Compared to the state-of-the-art boosting algorithms, in particular those using ELM as base classifier, the suggested technique does not require the generation of a new training dataset at each iteration. The adoption of the weighted least squares formulation of the problem has been presented as an unbiased and alternative approach to the already existing ELM boosting techniques. Moreover, the addition of a cost model for weighting the patterns, according to the order of the targets, enables the classifier to tackle ordinal regression problems further. The proposed method has been validated by an experimental study by comparing it with already existing ensemble methods and ELM techniques for ordinal regression, showing competitive results.
Collapse
|
25
|
CHIU CHIENYUAN, VERMA BRIJESH. RELATIONSHIP BETWEEN DATA SIZE, ACCURACY, DIVERSITY AND CLUSTERS IN NEURAL NETWORK ENSEMBLES. INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS 2014. [DOI: 10.1142/s1469026813400051] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
This paper presents an approach for analyzing relationships between data size, cluster, accuracy and diversity in neural network ensembles. The main objective of this research is to find out the influence of data size such as number of patterns, number of inputs and number of classes on various parameters such as clusters, accuracy and diversity of a neural network ensemble. The proposed approach is based on splitting data sets into different groups using the data size, clustering data and conducting training and testing of neural network ensembles. The test data is same for all groups and used to test all trained ensembles. The experiments have been conducted on 15 UCI machine learning benchmark datasets and results are presented in this paper.
Collapse
Affiliation(s)
- CHIEN-YUAN CHIU
- Central Queensland University, Rockhampton, Qld 4702, Australia
| | - BRIJESH VERMA
- Central Queensland University, Rockhampton, Qld 4702, Australia
| |
Collapse
|
26
|
Gomes JB, Gaber MM, Sousa PAC, Menasalvas E. Mining recurring concepts in a dynamic feature space. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2014; 25:95-110. [PMID: 24806647 DOI: 10.1109/tnnls.2013.2271915] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Most data stream classification techniques assume that the underlying feature space is static. However, in real-world applications the set of features and their relevance to the target concept may change over time. In addition, when the underlying concepts reappear, reusing previously learnt models can enhance the learning process in terms of accuracy and processing time at the expense of manageable memory consumption. In this paper, we propose mining recurring concepts in a dynamic feature space (MReC-DFS), a data stream classification system to address the challenges of learning recurring concepts in a dynamic feature space while simultaneously reducing the memory cost associated with storing past models. MReC-DFS is able to detect and adapt to concept changes using the performance of the learning process and contextual information. To handle recurring concepts, stored models are combined in a dynamically weighted ensemble. Incremental feature selection is performed to reduce the combined feature space. This contribution allows MReC-DFS to store only the features most relevant to the learnt concepts, which in turn increases the memory efficiency of the technique. In addition, an incremental feature selection method is proposed that dynamically determines the threshold between relevant and irrelevant features. Experimental results demonstrating the high accuracy of MReC-DFS compared with state-of-the-art techniques on a variety of real datasets are presented. The results also show the superior memory efficiency of MReC-DFS.
Collapse
|
27
|
Pratama M, Anavatti SG, Angelov PP, Lughofer E. PANFIS: a novel incremental learning machine. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2014; 25:55-68. [PMID: 24806644 DOI: 10.1109/tnnls.2013.2271933] [Citation(s) in RCA: 81] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Most of the dynamics in real-world systems are compiled by shifts and drifts, which are uneasy to be overcome by omnipresent neuro-fuzzy systems. Nonetheless, learning in nonstationary environment entails a system owning high degree of flexibility capable of assembling its rule base autonomously according to the degree of nonlinearity contained in the system. In practice, the rule growing and pruning are carried out merely benefiting from a small snapshot of the complete training data to truncate the computational load and memory demand to the low level. An exposure of a novel algorithm, namely parsimonious network based on fuzzy inference system (PANFIS), is to this end presented herein. PANFIS can commence its learning process from scratch with an empty rule base. The fuzzy rules can be stitched up and expelled by virtue of statistical contributions of the fuzzy rules and injected datum afterward. Identical fuzzy sets may be alluded and blended to be one fuzzy set as a pursuit of a transparent rule base escalating human's interpretability. The learning and modeling performances of the proposed PANFIS are numerically validated using several benchmark problems from real-world or synthetic datasets. The validation includes comparisons with state-of-the-art evolving neuro-fuzzy methods and showcases that our new method can compete and in some cases even outperform these approaches in terms of predictive fidelity and model complexity.
Collapse
|
28
|
Coop R, Mishtal A, Arel I. Ensemble learning in fixed expansion layer networks for mitigating catastrophic forgetting. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2013; 24:1623-1634. [PMID: 24808599 DOI: 10.1109/tnnls.2013.2264952] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Catastrophic forgetting is a well-studied attribute of most parameterized supervised learning systems. A variation of this phenomenon, in the context of feedforward neural networks, arises when nonstationary inputs lead to loss of previously learned mappings. The majority of the schemes proposed in the literature for mitigating catastrophic forgetting were not data driven and did not scale well. We introduce the fixed expansion layer (FEL) feedforward neural network, which embeds a sparsely encoding hidden layer to help mitigate forgetting of prior learned representations. In addition, we investigate a novel framework for training ensembles of FEL networks, based on exploiting an information-theoretic measure of diversity between FEL learners, to further control undesired plasticity. The proposed methodology is demonstrated on a basic classification task, clearly emphasizing its advantages over existing techniques. The architecture proposed can be enhanced to address a range of computational intelligence tasks, such as regression problems and system control.
Collapse
|
29
|
Woźniak M, Kasprzak A, Cal P. Weighted Aging Classifier Ensemble for the Incremental Drifted Data Streams. ACTA ACUST UNITED AC 2013. [DOI: 10.1007/978-3-642-40769-7_50] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/17/2023]
|
30
|
He H, Cao Y. SSC: a classifier combination method based on signal strength. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2012; 23:1100-1117. [PMID: 24807136 DOI: 10.1109/tnnls.2012.2198227] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
We propose a new classifier combination method, the signal strength-based combining (SSC) approach, to combine the outputs of multiple classifiers to support the decision-making process in classification tasks. As ensemble learning methods have attracted growing attention from both academia and industry recently, it is critical to understand the fundamental issues of the combining rule. Motivated by the signal strength concept, our proposed SSC algorithm can effectively integrate the individual vote from different classifiers in an ensemble learning system. Comparative studies of our method with nine major existing combining rules, namely, geometric average rule, arithmetic average rule, median value rule, majority voting rule, Borda count, max and min rule, weighted average, and weighted majority voting rules, is presented. Furthermore, we also discuss the relationship of the proposed method with respect to margin-based classifiers, including the boosting method (AdaBoost.M1 and AdaBoost.M2) and support vector machines by margin analysis. Detailed analyses of margin distribution graphs are presented to discuss the characteristics of the proposed method. Simulation results for various real-world datasets illustrate the effectiveness of the proposed method.
Collapse
|
31
|
Haibo He, Sheng Chen, Kang Li, Xin Xu. Incremental Learning From Stream Data. ACTA ACUST UNITED AC 2011; 22:1901-14. [DOI: 10.1109/tnn.2011.2171713] [Citation(s) in RCA: 92] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
32
|
|
33
|
Tortajada S, Fuster-Garcia E, Vicente J, Wesseling P, Howe FA, Julià-Sapé M, Candiota AP, Monleón D, Moreno-Torres A, Pujol J, Griffiths JR, Wright A, Peet AC, Martínez-Bisbal MC, Celda B, Arús C, Robles M, García-Gómez JM. Incremental Gaussian Discriminant Analysis based on Graybill and Deal weighted combination of estimators for brain tumour diagnosis. J Biomed Inform 2011; 44:677-87. [PMID: 21377545 DOI: 10.1016/j.jbi.2011.02.009] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2010] [Revised: 02/17/2011] [Accepted: 02/23/2011] [Indexed: 01/13/2023]
Abstract
In the last decade, machine learning (ML) techniques have been used for developing classifiers for automatic brain tumour diagnosis. However, the development of these ML models rely on a unique training set and learning stops once this set has been processed. Training these classifiers requires a representative amount of data, but the gathering, preprocess, and validation of samples is expensive and time-consuming. Therefore, for a classical, non-incremental approach to ML, it is necessary to wait long enough to collect all the required data. In contrast, an incremental learning approach may allow us to build an initial classifier with a smaller number of samples and update it incrementally when new data are collected. In this study, an incremental learning algorithm for Gaussian Discriminant Analysis (iGDA) based on the Graybill and Deal weighted combination of estimators is introduced. Each time a new set of data becomes available, a new estimation is carried out and a combination with a previous estimation is performed. iGDA does not require access to the previously used data and is able to include new classes that were not in the original analysis, thus allowing the customization of the models to the distribution of data at a particular clinical center. An evaluation using five benchmark databases has been used to evaluate the behaviour of the iGDA algorithm in terms of stability-plasticity, class inclusion and order effect. Finally, the iGDA algorithm has been applied to automatic brain tumour classification with magnetic resonance spectroscopy, and compared with two state-of-the-art incremental algorithms. The empirical results obtained show the ability of the algorithm to learn in an incremental fashion, improving the performance of the models when new information is available, and converging in the course of time. Furthermore, the algorithm shows a negligible instance and concept order effect, avoiding the bias that such effects could introduce.
Collapse
Affiliation(s)
- Salvador Tortajada
- IBIME, Instituto de Aplicaciones de las Tecnologías de la Información y de las Comunicaciones Avanzadas (ITACA), Universitat Politècnica de València, València, Spain.
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |
Collapse
|
34
|
Rahman A, Verma B. Novel Layered Clustering-Based Approach for Generating Ensemble of Classifiers. ACTA ACUST UNITED AC 2011; 22:781-92. [DOI: 10.1109/tnn.2011.2118765] [Citation(s) in RCA: 70] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
35
|
|
36
|
Grinblat GL, Uzal LC, Ceccatto HA, Granitto PM. Solving Nonstationary Classification Problems With Coupled Support Vector Machines. ACTA ACUST UNITED AC 2011; 22:37-51. [DOI: 10.1109/tnn.2010.2083684] [Citation(s) in RCA: 34] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
37
|
Towards incremental learning of nonstationary imbalanced data stream: a multiple selectively recursive approach. EVOLVING SYSTEMS 2010. [DOI: 10.1007/s12530-010-9021-y] [Citation(s) in RCA: 88] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
38
|
Abstract
In recent years, learning from imbalanced data has attracted growing attention from both academia and industry due to the explosive growth of applications that use and produce imbalanced data. However, because of the complex characteristics of imbalanced data, many real-world solutions struggle to provide robust efficiency in learning-based applications. In an effort to address this problem, this paper presents Ranked Minority Oversampling in Boosting (RAMOBoost), which is a RAMO technique based on the idea of adaptive synthetic data generation in an ensemble learning system. Briefly, RAMOBoost adaptively ranks minority class instances at each learning iteration according to a sampling probability distribution that is based on the underlying data distribution, and can adaptively shift the decision boundary toward difficult-to-learn minority and majority class instances by using a hypothesis assessment procedure. Simulation analysis on 19 real-world datasets assessed over various metrics-including overall accuracy, precision, recall, F-measure, G-mean, and receiver operation characteristic analysis-is used to illustrate the effectiveness of this method.
Collapse
Affiliation(s)
- Sheng Chen
- Department of Electrical and Computer Engineering, Stevens Institute of Technology, Hoboken, NJ 07030, USA.
| | | | | |
Collapse
|
39
|
Incremental Learning of New Classes in Unbalanced Datasets: Learn + + .UDNC. MULTIPLE CLASSIFIER SYSTEMS 2010. [DOI: 10.1007/978-3-642-12127-2_4] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
|