51
|
Feng L, Zhao C, Chen CP, Li Y, Zhou M, Qiao H, Fu C. BNGBS: An efficient network boosting system with triple incremental learning capabilities for more nodes, samples, and classes. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.06.100] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
52
|
Yu H, Lu J, Zhang G. Online Topology Learning by a Gaussian Membership-Based Self-Organizing Incremental Neural Network. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:3947-3961. [PMID: 31725398 DOI: 10.1109/tnnls.2019.2947658] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/05/2023]
Abstract
In order to extract useful information from data streams, incremental learning has been introduced in more and more data mining algorithms. For instance, a self-organizing incremental neural network (SOINN) has been proposed to extract a topological structure that consists of one or more neural networks to closely reflect the data distribution of data streams. However, SOINN has the tradeoff between deleting previously learned nodes and inserting new nodes, i.e., the stability-plasticity dilemma. Therefore, it is not guaranteed that the topological structure obtained by the SOINN will closely represent data distribution. For solving the stability-plasticity dilemma, we propose a Gaussian membership-based SOINN (Gm-SOINN). Unlike other SOINN-based methods that allow only one node to be identified as a "winner" (the nearest node), the Gm-SOINN uses a Gaussian membership to indicate to which degree the node is a winner. Hence, the Gm-SOINN avoids the topological structure that cannot represent the data distribution because previously learned nodes overly deleted or noisy nodes inserted. In addition, an evolving Gaussian mixture model is integrated into the Gm-SOINN to estimate the density distribution of nodes, thereby avoiding the wrong connection between two nodes. Experiments involving both artificial and real-world data sets indicate that our proposed Gm-SOINN achieves better performance than other topology learning methods.
Collapse
|
53
|
Chu Z, Yu J, Hamdulla A. LPG-model: A novel model for throughput prediction in stream processing, using a light gradient boosting machine, incremental principal component analysis, and deep gated recurrent unit network. Inf Sci (N Y) 2020. [DOI: 10.1016/j.ins.2020.05.042] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
54
|
Zhao Z, Cristian A, Rosen G. Keeping up with the genomes: efficient learning of our increasing knowledge of the tree of life. BMC Bioinformatics 2020; 21:412. [PMID: 32957925 PMCID: PMC7507296 DOI: 10.1186/s12859-020-03744-7] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2020] [Accepted: 09/08/2020] [Indexed: 11/26/2022] Open
Abstract
BACKGROUND It is a computational challenge for current metagenomic classifiers to keep up with the pace of training data generated from genome sequencing projects, such as the exponentially-growing NCBI RefSeq bacterial genome database. When new reference sequences are added to training data, statically trained classifiers must be rerun on all data, resulting in a highly inefficient process. The rich literature of "incremental learning" addresses the need to update an existing classifier to accommodate new data without sacrificing much accuracy compared to retraining the classifier with all data. RESULTS We demonstrate how classification improves over time by incrementally training a classifier on progressive RefSeq snapshots and testing it on: (a) all known current genomes (as a ground truth set) and (b) a real experimental metagenomic gut sample. We demonstrate that as a classifier model's knowledge of genomes grows, classification accuracy increases. The proof-of-concept naïve Bayes implementation, when updated yearly, now runs in 1/4th of the non-incremental time with no accuracy loss. CONCLUSIONS It is evident that classification improves by having the most current knowledge at its disposal. Therefore, it is of utmost importance to make classifiers computationally tractable to keep up with the data deluge. The incremental learning classifier can be efficiently updated without the cost of reprocessing nor the access to the existing database and therefore save storage as well as computation resources.
Collapse
Affiliation(s)
- Zhengqiao Zhao
- Ecological and Evolutionary Signal-process and Informatics (EESI) Lab, Department of Electrical and Computer Engineering, Drexel University, Market Street, Philadelphia, US
| | - Alexandru Cristian
- Department of Computer Science, Drexel University, Market Street, Philadelphia, US
| | - Gail Rosen
- Ecological and Evolutionary Signal-process and Informatics (EESI) Lab, Department of Electrical and Computer Engineering, Drexel University, Market Street, Philadelphia, US
| |
Collapse
|
55
|
EnsPKDE&IncLKDE: a hybrid time series prediction algorithm integrating dynamic ensemble pruning, incremental learning, and kernel density estimation. APPL INTELL 2020. [DOI: 10.1007/s10489-020-01802-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
56
|
Khuat TT, Ruta D, Gabrys B. Hyperbox-based machine learning algorithms: a comprehensive survey. Soft comput 2020. [DOI: 10.1007/s00500-020-05226-7] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/22/2023]
|
57
|
Mello AR, Stemmer MR, Koerich AL. Incremental and decremental fuzzy bounded twin support vector machine. Inf Sci (N Y) 2020. [DOI: 10.1016/j.ins.2020.03.038] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
58
|
Costa J, Silva C, Antunes M, Ribeiro B. Boosting dynamic ensemble’s performance in Twitter. Neural Comput Appl 2020. [DOI: 10.1007/s00521-019-04599-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
59
|
A Novel Drift Detection Algorithm Based on Features’ Importance Analysis in a Data Streams Environment. JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH 2020. [DOI: 10.2478/jaiscr-2020-0019] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022] Open
Abstract
Abstract
The training set consists of many features that influence the classifier in different degrees. Choosing the most important features and rejecting those that do not carry relevant information is of great importance to the operating of the learned model. In the case of data streams, the importance of the features may additionally change over time. Such changes affect the performance of the classifier but can also be an important indicator of occurring concept-drift. In this work, we propose a new algorithm for data streams classification, called Random Forest with Features Importance (RFFI), which uses the measure of features importance as a drift detector. The RFFT algorithm implements solutions inspired by the Random Forest algorithm to the data stream scenarios. The proposed algorithm combines the ability of ensemble methods for handling slow changes in a data stream with a new method for detecting concept drift occurrence. The work contains an experimental analysis of the proposed algorithm, carried out on synthetic and real data.
Collapse
|
60
|
GeoBoost: An Incremental Deep Learning Approach toward Global Mapping of Buildings from VHR Remote Sensing Images. REMOTE SENSING 2020. [DOI: 10.3390/rs12111794] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Modern convolutional neural networks (CNNs) are often trained on pre-set data sets with a fixed size. As for the large-scale applications of satellite images, for example, global or regional mappings, these images are collected incrementally by multiple stages in general. In other words, the sizes of training datasets might be increased for the tasks of mapping rather than be fixed beforehand. In this paper, we present a novel algorithm, called GeoBoost, for the incremental-learning tasks of semantic segmentation via convolutional neural networks. Specifically, the GeoBoost algorithm is trained in an end-to-end manner on the newly available data, and it does not decrease the performance of previously trained models. The effectiveness of the GeoBoost algorithm is verified on the large-scale data set of DREAM-B. This method avoids the need for training on the enlarged data set from scratch and would become more effective along with more available data.
Collapse
|
61
|
Wang Z, Li HX, Chen C. Incremental Reinforcement Learning in Continuous Spaces via Policy Relaxation and Importance Weighting. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:1870-1883. [PMID: 31395556 DOI: 10.1109/tnnls.2019.2927320] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
In this paper, a systematic incremental learning method is presented for reinforcement learning in continuous spaces where the learning environment is dynamic. The goal is to adjust the previously learned policy in the original environment to a new one incrementally whenever the environment changes. To improve the adaptability to the ever-changing environment, we propose a two-step solution incorporated with the incremental learning procedure: policy relaxation and importance weighting. First, the behavior policy is relaxed to a random one in the initial learning episodes to encourage a proper exploration in the new environment. It alleviates the conflict between the new information and the existing knowledge for a better adaptation in the long term. Second, it is observed that episodes receiving higher returns are more in line with the new environment, and hence contain more new information. During parameter updating, we assign higher importance weights to the learning episodes that contain more new information, thus encouraging the previous optimal policy to be faster adapted to a new one that fits in the new environment. Empirical studies on continuous controlling tasks with varying configurations verify that the proposed method achieves a significantly faster adaptation to various dynamic environments than the baselines.
Collapse
|
62
|
Wang L, Wu C. Dynamic imbalanced business credit evaluation based on Learn++ with sliding time window and weight sampling and FCM with multiple kernels. Inf Sci (N Y) 2020. [DOI: 10.1016/j.ins.2020.02.011] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
63
|
Infinite Lattice Learner: an ensemble for incremental learning. Soft comput 2020. [DOI: 10.1007/s00500-019-04330-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
64
|
Lobo JL, Oregi I, Bifet A, Del Ser J. Exploiting the stimuli encoding scheme of evolving Spiking Neural Networks for stream learning. Neural Netw 2020; 123:118-133. [DOI: 10.1016/j.neunet.2019.11.021] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2019] [Revised: 09/05/2019] [Accepted: 11/25/2019] [Indexed: 10/25/2022]
|
65
|
|
66
|
Zhang Z, Zhao Y, Liao X, Shi W, Li K, Zou Q, Peng S. Deep learning in omics: a survey and guideline. Brief Funct Genomics 2020; 18:41-57. [PMID: 30265280 DOI: 10.1093/bfgp/ely030] [Citation(s) in RCA: 85] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2018] [Revised: 07/31/2018] [Accepted: 08/30/2018] [Indexed: 01/17/2023] Open
Abstract
Omics, such as genomics, transcriptome and proteomics, has been affected by the era of big data. A huge amount of high dimensional and complex structured data has made it no longer applicable for conventional machine learning algorithms. Fortunately, deep learning technology can contribute toward resolving these challenges. There is evidence that deep learning can handle omics data well and resolve omics problems. This survey aims to provide an entry-level guideline for researchers, to understand and use deep learning in order to solve omics problems. We first introduce several deep learning models and then discuss several research areas which have combined omics and deep learning in recent years. In addition, we summarize the general steps involved in using deep learning which have not yet been systematically discussed in the existent literature on this topic. Finally, we compare the features and performance of current mainstream open source deep learning frameworks and present the opportunities and challenges involved in deep learning. This survey will be a good starting point and guideline for omics researchers to understand deep learning.
Collapse
Affiliation(s)
- Zhiqiang Zhang
- School of Computer Science, National University of Defense Technology, Changsha, China
| | - Yi Zhao
- Institute of Computing Technology,Chinese Academy of Sciences, Beijing, China
| | - Xiangke Liao
- School of Computer Science, National University of Defense Technology, Changsha, China
| | - Wenqiang Shi
- School of Computer Science, National University of Defense Technology, Changsha, China
| | - Kenli Li
- College of Computer Science and Electronic Engineering & National Supercomputer Centre in Changsha, Hunan University, Changsha, China
| | - Quan Zou
- School of Computer Science and Technology, Tianjin University, Tianjin, China
| | - Shaoliang Peng
- School of Computer Science, National University of Defense Technology, Changsha, China.,College of Computer Science and Electronic Engineering & National Supercomputer Centre in Changsha, Hunan University, Changsha, China
| |
Collapse
|
67
|
Deng CH, Wang XJ, Gu J, Wang W. The Online Soft Computing Models of key variables based on the Boundary Forest method. Soft comput 2019. [DOI: 10.1007/s00500-019-04584-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
68
|
Incremental Learning to Personalize Human Activity Recognition Models: The Importance of Human AI Collaboration. SENSORS 2019; 19:s19235151. [PMID: 31775243 PMCID: PMC6928956 DOI: 10.3390/s19235151] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2019] [Revised: 11/15/2019] [Accepted: 11/19/2019] [Indexed: 11/24/2022]
Abstract
This study presents incremental learning based methods to personalize human activity recognition models. Initially, a user-independent model is used in the recognition process. When a new user starts to use the human activity recognition application, personal streaming data can be gathered. Of course, this data does not have labels. However, there are three different ways to obtain this data: non-supervised, semi-supervised, and supervised. The non-supervised approach relies purely on predicted labels, the supervised approach uses only human intelligence to label the data, and the proposed method for semi-supervised learning is a combination of these two: It uses artificial intelligence (AI) in most cases to label the data but in uncertain cases it relies on human intelligence. After labels are obtained, the personalization process continues by using the streaming data and these labels to update the incremental learning based model, which in this case is Learn++. Learn++ is an ensemble method that can use any classifier as a base classifier, and this study compares three base classifiers: linear discriminant analysis (LDA), quadratic discriminant analysis (QDA), and classification and regression tree (CART). Moreover, three datasets are used in the experiment to show how well the presented method generalizes on different datasets. The results show that personalized models are much more accurate than user-independent models. On average, the recognition rates are: 87.0% using the user-independent model, 89.1% using the non-supervised personalization approach, 94.0% using the semi-supervised personalization approach, and 96.5% using the supervised personalization approach. This means that by relying on predicted labels with high confidence, and asking the user to label only uncertain observations (6.6% of the observations when using LDA, 7.7% when using QDA, and 18.3% using CART), almost as low error rates can be achieved as by using the supervised approach, in which labeling is fully based on human intelligence.
Collapse
|
69
|
Besrour A, Ksantini R. Incremental Subclass Support Vector Machine. INT J ARTIF INTELL T 2019. [DOI: 10.1142/s0218213019500209] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Support Vector Machine (SVM) is a very competitive linear classifier based on convex optimization problem, were support vectors fully describe decision boundary. Hence, SVM is sensitive to data spread and does not take into account the existence of class subclasses, nor minimizes data dispersion for classification performance improvement. Thus, Kernel subclass SVM (KSSVM) was proposed to handle multimodal data and to minimize data dispersion. Nevertheless, KSSVM has difficulties in classifying sequentially obtained data and handling large scale datasets, since it is based on batch learning. For this reason, we propose a novel incremental KSSVM (iKSSVM) which handles dynamic and large data in a proper manner. The iKSSVM is still based on convex optimization problem and minimizes data dispersion within and between data subclasses incrementally, in order to improve discriminative power and classification performance. An extensive comparative evaluation of the iKSSVM to batch KSSVM, as well as, other contemporary incremental classifiers, on real world datasets, has shown clearly its superiority in terms of classification accuracy.
Collapse
Affiliation(s)
- Amine Besrour
- Higher School of Communications of Tunis (Sup’Com), MEDIATRON Lab, Carthage University, Tunis, Tunisia, CP 2080, Tunisia
| | - Riadh Ksantini
- Digital Security Research Lab, Higher School of Communication of Tunis (Sup’Com), University of Carthage, Tunisia, Tunisia
- University of Windsor, 401, Sunset Avenue, Windsor, ON, Canada
| |
Collapse
|
70
|
Vaičiulytė J, Sakalauskas L. Recursive parameter estimation algorithm of the Dirichlet hidden Markov model. J STAT COMPUT SIM 2019. [DOI: 10.1080/00949655.2019.1679144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Affiliation(s)
- Jūratė Vaičiulytė
- Institute of Data Science and Digital Technologies, Vilnius University, Vilnius, Lithuania
| | - Leonidas Sakalauskas
- Department of Informatics and Statistics, Klaipėda University, Klaipėda, Lithuania
| |
Collapse
|
71
|
Abstract
Abstract
Mining and analysing streaming data is crucial for many applications, and this area of research has gained extensive attention over the past decade. However, there are several inherent problems that continue to challenge the hardware and the state-of-the art algorithmic solutions. Examples of such problems include the unbound size, varying speed and unknown data characteristics of arriving instances from a data stream. The aim of this research is to portray key challenges faced by algorithmic solutions for stream mining, particularly focusing on the prevalent issue of concept drift. A comprehensive discussion of concept drift and its inherent data challenges in the context of stream mining is presented, as is a critical, in-depth review of relevant literature. Current issues with the evaluative procedure for concept drift detectors is also explored, highlighting problems such as a lack of established base datasets and the impact of temporal dependence on concept drift detection. By exposing gaps in the current literature, this study suggests recommendations for future research which should aid in the progression of stream mining and concept drift detection algorithms.
Collapse
|
72
|
Abderrahmane N, Lemaire E, Miramond B. Design Space Exploration of Hardware Spiking Neurons for Embedded Artificial Intelligence. Neural Netw 2019; 121:366-386. [PMID: 31593842 DOI: 10.1016/j.neunet.2019.09.024] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2019] [Revised: 09/16/2019] [Accepted: 09/17/2019] [Indexed: 11/30/2022]
Abstract
Machine learning is yielding unprecedented interest in research and industry, due to recent success in many applied contexts such as image classification and object recognition. However, the deployment of these systems requires huge computing capabilities, thus making them unsuitable for embedded systems. To deal with this limitation, many researchers are investigating brain-inspired computing, which would be a perfect alternative to the conventional Von Neumann architecture based computers (CPU/GPU) that meet the requirements for computing performance, but not for energy-efficiency. Therefore, neuromorphic hardware circuits that are adaptable for both parallel and distributed computations need to be designed. In this paper, we focus on Spiking Neural Networks (SNNs) with a comprehensive study of neural coding methods and hardware exploration. In this context, we propose a framework for neuromorphic hardware design space exploration, which allows to define a suitable architecture based on application-specific constraints and starting from a wide variety of possible architectural choices. For this framework, we have developed a behavioral level simulator for neuromorphic hardware architectural exploration named NAXT. Moreover, we propose modified versions of the standard Rate Coding technique to make trade-offs with the Time Coding paradigm, which is characterized by the low number of spikes propagating in the network. Thus, we are able to reduce the number of spikes while keeping the same neuron's model, which results in an SNN with fewer events to process. By doing so, we seek to reduce the amount of power consumed by the hardware. Furthermore, we present three neuromorphic hardware architectures in order to quantitatively study the implementation of SNNs. One of these architectures integrates a novel hybrid structure: a highly-parallel computation core for most solicited layers, and time-multiplexed computation units for deeper layers. These architectures are derived from a novel funnel-like Design Space Exploration framework for neuromorphic hardware.
Collapse
Affiliation(s)
| | - Edgar Lemaire
- Université Côte d'Azur, CNRS, LEAT, France; Thales Research Technology / STI Group / LCHP, Palaiseau, France.
| | | |
Collapse
|
73
|
Junsawang P, Phimoltares S, Lursinsap C. Streaming chunk incremental learning for class-wise data stream classification with fast learning speed and low structural complexity. PLoS One 2019; 14:e0220624. [PMID: 31498787 PMCID: PMC6733468 DOI: 10.1371/journal.pone.0220624] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2018] [Accepted: 07/20/2019] [Indexed: 11/18/2022] Open
Abstract
Due to the fast speed of data generation and collection from advanced equipment, the amount of data obviously overflows the limit of available memory space and causes difficulties achieving high learning accuracy. Several methods based on discard-after-learn concept have been proposed. Some methods were designed to cope with a single incoming datum but some were designed for a chunk of incoming data. Although the results of these approaches are rather impressive, most of them are based on temporally adding more neurons to learn new incoming data without any neuron merging process which can obviously increase the computational time and space complexities. Only online versatile elliptic basis function (VEBF) introduced neuron merging to reduce the space-time complexity of learning only a single incoming datum. This paper proposed a method for further enhancing the capability of discard-after-learn concept for streaming data-chunk environment in terms of low computational time and neural space complexities. A set of recursive functions for computing the relevant parameters of a new neuron, based on statistical confidence interval, was introduced. The newly proposed method, named streaming chunk incremental learning (SCIL), increases the plasticity and the adaptabilty of the network structure according to the distribution of incoming data and their classes. When being compared to the others in incremental-like manner, based on 11 benchmarked data sets of 150 to 581,012 samples with attributes ranging from 4 to 1,558 formed as streaming data, the proposed SCIL gave better accuracy and time in most data sets.
Collapse
Affiliation(s)
- Prem Junsawang
- Department of Mathematics and Computer Science, Faculty of Science, Chulalongkorn University, Bangkok, Thailand
| | - Suphakant Phimoltares
- Department of Mathematics and Computer Science, Faculty of Science, Chulalongkorn University, Bangkok, Thailand
- * E-mail:
| | - Chidchanok Lursinsap
- Department of Mathematics and Computer Science, Faculty of Science, Chulalongkorn University, Bangkok, Thailand
| |
Collapse
|
74
|
Incremental learning of concept drift in Multiple Instance Learning for industrial visual inspection. COMPUT IND 2019. [DOI: 10.1016/j.compind.2019.04.006] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
75
|
Continual lifelong learning with neural networks: A review. Neural Netw 2019; 113:54-71. [DOI: 10.1016/j.neunet.2019.01.012] [Citation(s) in RCA: 322] [Impact Index Per Article: 53.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2018] [Revised: 01/18/2019] [Accepted: 01/22/2019] [Indexed: 10/27/2022]
|
76
|
Yu H, Webb GI. Adaptive online extreme learning machine by regulating forgetting factor by concept drift map. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2018.11.098] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
77
|
Incremental supervised learning: algorithms and applications in pattern recognition. EVOLUTIONARY INTELLIGENCE 2019. [DOI: 10.1007/s12065-019-00203-y] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
78
|
Bang SH, Ak R, Narayanan A, Lee YT, Cho H. A Survey on Knowledge Transfer for Manufacturing Data Analytics. COMPUT IND 2019; 104:10.1016/j.compind.2018.07.001. [PMID: 39440000 PMCID: PMC11495017 DOI: 10.1016/j.compind.2018.07.001] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Abstract
Data analytics techniques have been used for numerous manufacturing applications in various areas. A common assumption of data analytics models is that the environment that generates data is stationary, that is, the feature (or label) space or distribution of the data does not change over time. However, in the real world, this assumption is not valid especially for manufacturing. In non-stationary environments, the accuracy of the model decreases over time, so the model must be retrained periodically and adapted to the corresponding environment(s). Knowledge transfer for data analytics is an approach that trains a model with knowledge extracted from data or model(s). Knowledge transfer can be used when adapting to a new environment, while reducing or eliminating degradation in the accuracy of the model. This paper surveys knowledge transfer methods that have been widely used in various applications, and investigates the applicability of these methods for manufacturing problems. The surveyed knowledge transfer methods are analyzed from three viewpoints: types of non-stationary environments, availability of labeled data, and sources of knowledge. In addition, we categorize events that cause non-stationary environments in manufacturing, and present a mechanism to enable practitioners to select the appropriate methods for their manufacturing data analytics applications among the surveyed knowledge transfer methods. The mechanism includes the steps 1) to detect changes in data properties, 2) to define source and target, and 3) to select available knowledge transfer methods. By providing comprehensive information, this paper will support researchers to adopt knowledge transfer in manufacturing.
Collapse
Affiliation(s)
- Seung Hwan Bang
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
- Department of Industrial and Management Engineering, Pohang University of Science and Technology (POSTECH), Pohang, Gyeongbuk, 37673, Republic of Korea
| | - Ronay Ak
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
| | - Anantha Narayanan
- Department of Mechanical Engineering, University of Maryland, College Park, MD, 20742, USA
| | - Y Tina Lee
- Systems Integration Division, Engineering Laboratory, National Institute of Standards and Technology (NIST), Gaithersburg, MD, 20899, USA
| | - Hyunbo Cho
- Department of Industrial and Management Engineering, Pohang University of Science and Technology (POSTECH), Pohang, Gyeongbuk, 37673, Republic of Korea
| |
Collapse
|
79
|
Abstract
Deep learning-based methods have reached state of the art performances, relying on a large quantity of available data and computational power. Such methods still remain highly inappropriate when facing a major open machine learning problem, which consists of learning incrementally new classes and examples over time. Combining the outstanding performances of Deep Neural Networks (DNNs) with the flexibility of incremental learning techniques is a promising venue of research. In this contribution, we introduce Transfer Incremental Learning using Data Augmentation (TILDA). TILDA is based on pre-trained DNNs as feature extractors, robust selection of feature vectors in subspaces using a nearest-class-mean based technique, majority votes and data augmentation at both the training and the prediction stages. Experiments on challenging vision datasets demonstrate the ability of the proposed method for low complexity incremental learning, while achieving significantly better accuracy than existing incremental counterparts.
Collapse
|
80
|
Incremental Learning for Classification of Unstructured Data Using Extreme Learning Machine. ALGORITHMS 2018. [DOI: 10.3390/a11100158] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Unstructured data are irregular information with no predefined data model. Streaming data which constantly arrives over time is unstructured, and classifying these data is a tedious task as they lack class labels and get accumulated over time. As the data keeps growing, it becomes difficult to train and create a model from scratch each time. Incremental learning, a self-adaptive algorithm uses the previously learned model information, then learns and accommodates new information from the newly arrived data providing a new model, which avoids the retraining. The incrementally learned knowledge helps to classify the unstructured data. In this paper, we propose a framework CUIL (Classification of Unstructured data using Incremental Learning) which clusters the metadata, assigns a label for each cluster and then creates a model using Extreme Learning Machine (ELM), a feed-forward neural network, incrementally for each batch of data arrived. The proposed framework trains the batches separately, reducing the memory resources, training time significantly and is tested with metadata created for the standard image datasets like MNIST, STL-10, CIFAR-10, Caltech101, and Caltech256. Based on the tabulated results, our proposed work proves to show greater accuracy and efficiency.
Collapse
|
81
|
Sun Y, Tang K, Zhu Z, Yao X. Concept Drift Adaptation by Exploiting Historical Knowledge. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4822-4832. [PMID: 29993956 DOI: 10.1109/tnnls.2017.2775225] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Incremental learning with concept drift has often been tackled by ensemble methods, where models built in the past can be retrained to attain new models for the current data. Two design questions need to be addressed in developing ensemble methods for incremental learning with concept drift, i.e., which historical (i.e., previously trained) models should be preserved and how to utilize them. A novel ensemble learning method, namely, Diversity and Transfer-based Ensemble Learning (DTEL), is proposed in this paper. Given newly arrived data, DTEL uses each preserved historical model as an initial model and further trains it with the new data via transfer learning. Furthermore, DTEL preserves a diverse set of historical models, rather than a set of historical models that are merely accurate in terms of classification accuracy. Empirical studies on 15 synthetic data streams and 5 real-world data streams (all with concept drifts) demonstrate that DTEL can handle concept drift more effectively than 4 other state-of-the-art methods.
Collapse
|
82
|
Wang S, Minku LL, Yao X. A Systematic Study of Online Class Imbalance Learning With Concept Drift. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4802-4821. [PMID: 29993955 DOI: 10.1109/tnnls.2017.2771290] [Citation(s) in RCA: 49] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
As an emerging research topic, online class imbalance learning often combines the challenges of both class imbalance and concept drift. It deals with data streams having very skewed class distributions, where concept drift may occur. It has recently received increased research attention; however, very little work addresses the combined problem where both class imbalance and concept drift coexist. As the first systematic study of handling concept drift in class-imbalanced data streams, this paper first provides a comprehensive review of current research progress in this field, including current research focuses and open challenges. Then, an in-depth experimental study is performed, with the goal of understanding how to best overcome concept drift in online learning with class imbalance.
Collapse
|
83
|
Incremental learning method for cyber intelligence, surveillance, and reconnaissance in closed military network using converged IT techniques. Soft comput 2018. [DOI: 10.1007/s00500-018-3433-1] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
|
84
|
Affiliation(s)
| | - Bing Liu
- University of Illinois at Chicago
| |
Collapse
|
85
|
Gupta S, Sanyal S. INNAMP: An incremental neural network architecture with monitor perceptron. AI COMMUN 2018. [DOI: 10.3233/aic-180767] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Affiliation(s)
- Sharad Gupta
- Information Technology Department, Indian Institute of Information Technology, Allahabad, Deoghat, Jhalwa, Allahabad, India. E-mail:
| | - Sudip Sanyal
- Computer Science and Engineering, BML Munjal University, Gurgaon, Haryana, India. E-mail:
| |
Collapse
|
86
|
Jaworski M, Duda P, Rutkowski L, Jaworski M, Duda P, Rutkowski L, Rutkowski L, Duda P, Jaworski M. New Splitting Criteria for Decision Trees in Stationary Data Streams. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:2516-2529. [PMID: 28500013 DOI: 10.1109/tnnls.2017.2698204] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
The most popular tools for stream data mining are based on decision trees. In previous 15 years, all designed methods, headed by the very fast decision tree algorithm, relayed on Hoeffding's inequality and hundreds of researchers followed this scheme. Recently, we have demonstrated that although the Hoeffding decision trees are an effective tool for dealing with stream data, they are a purely heuristic procedure; for example, classical decision trees such as ID3 or CART cannot be adopted to data stream mining using Hoeffding's inequality. Therefore, there is an urgent need to develop new algorithms, which are both mathematically justified and characterized by good performance. In this paper, we address this problem by developing a family of new splitting criteria for classification in stationary data streams and investigating their probabilistic properties. The new criteria, derived using appropriate statistical tools, are based on the misclassification error and the Gini index impurity measures. The general division of splitting criteria into two types is proposed. Attributes chosen based on type- splitting criteria guarantee, with high probability, the highest expected value of split measure. Type- criteria ensure that the chosen attribute is the same, with high probability, as it would be chosen based on the whole infinite data stream. Moreover, in this paper, two hybrid splitting criteria are proposed, which are the combinations of single criteria based on the misclassification error and Gini index.
Collapse
|
87
|
|
88
|
Sultan Zia M, Hussain M, Arfan Jaffar M. Incremental Learning-Based Facial Expression Classification System Using a Novel Multinomial Classifier. INT J PATTERN RECOGN 2017. [DOI: 10.1142/s0218001418560049] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Facial expressions recognition is a crucial task in pattern recognition and it becomes even crucial when cross-cultural emotions are encountered. Various studies in the past have shown that all the facial expressions are not innate and universal, but many of them are learned and culture-dependent. Extreme facial expression recognition methods employ different datasets for training and later use it for testing and demostrate high accuracy in recognition. Their performances degrade drastically when expression images are taken from different cultures. Moreover, there are many existing facial expression patterns which cannot be generated and used as training data in single training session. A facial expression recognition system can maintain its high accuracy and robustness globally and for a longer period if the system possesses the ability to learn incrementally. We also propose a novel classification algorithm for multinomial classification problems. It is an efficient classifier and can be a good choice for base classifier in real-time applications. We propose a facial expression recognition system that can learn incrementally. We use Local Binary Pattern (LBP) features to represent the expression space. The performance of the system is tested on static images from six different databases containing expressions from various cultures. The experiments using the incremental learning classification demonstrate promising results.
Collapse
Affiliation(s)
- M. Sultan Zia
- COMSATS, Institute of Information Technology, Sahiwal, Pakistan
| | - Majid Hussain
- COMSATS, Institute of Information Technology, Sahiwal, Pakistan
- Al Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - M. Arfan Jaffar
- Al Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| |
Collapse
|
89
|
Reducing the complexity of an adaptive radial basis function network with a histogram algorithm. Neural Comput Appl 2017. [DOI: 10.1007/s00521-016-2350-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
90
|
Xiao J, Xiang Z, Wang D, Xiao Z. Nonparametric kernel smoother on topology learning neural networks for incremental and ensemble regression. Neural Comput Appl 2017. [DOI: 10.1007/s00521-017-3218-y] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
91
|
Affiliation(s)
- Yue Dong
- Department of Computer Science; McGill University; Montreal, QC Canada
| | | |
Collapse
|
92
|
|
93
|
|
94
|
Ngo Ho AK, Eglin V, Ragot N, Ramel JY. A multi-one-class dynamic classifier for adaptive digitization of document streams. INT J DOC ANAL RECOG 2017. [DOI: 10.1007/s10032-017-0286-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
95
|
Lee SW, Lee CY, Kwak DH, Ha JW, Kim J, Zhang BT. Dual-memory neural networks for modeling cognitive activities of humans via wearable sensors. Neural Netw 2017; 92:17-28. [PMID: 28318904 DOI: 10.1016/j.neunet.2017.02.008] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2016] [Revised: 02/13/2017] [Accepted: 02/14/2017] [Indexed: 10/20/2022]
Abstract
Wearable devices, such as smart glasses and watches, allow for continuous recording of everyday life in a real world over an extended period of time or lifelong. This possibility helps better understand the cognitive behavior of humans in real life as well as build human-aware intelligent agents for practical purposes. However, modeling the human cognitive activity from wearable-sensor data stream is challenging because learning new information often results in loss of previously acquired information, causing a problem known as catastrophic forgetting. Here we propose a deep-learning neural network architecture that resolves the catastrophic forgetting problem. Based on the neurocognitive theory of the complementary learning systems of the neocortex and hippocampus, we introduce a dual memory architecture (DMA) that, on one hand, slowly acquires the structured knowledge representations and, on the other hand, rapidly learns the specifics of individual experiences. The DMA system learns continuously through incremental feature adaptation and weight transfer. We evaluate the performance on two real-life datasets, the CIFAR-10 image-stream dataset and the 46-day Lifelog dataset collected from Google Glass, showing that the proposed model outperforms other online learning methods.
Collapse
Affiliation(s)
- Sang-Woo Lee
- School of Computer Science and Engineering, Seoul National University, Seoul 08826, South Korea
| | - Chung-Yeon Lee
- School of Computer Science and Engineering, Seoul National University, Seoul 08826, South Korea
| | - Dong-Hyun Kwak
- Interdisciplinary Program in Neuroscience, Seoul National University, Seoul 08826, South Korea
| | - Jung-Woo Ha
- NAVER LABS, NAVER Corp., Bundang 13561, South Korea
| | - Jeonghee Kim
- NAVER LABS, NAVER Corp., Bundang 13561, South Korea
| | - Byoung-Tak Zhang
- School of Computer Science and Engineering, Seoul National University, Seoul 08826, South Korea; Interdisciplinary Program in Neuroscience, Seoul National University, Seoul 08826, South Korea; Surromind Robotics, 1 Gwanak-ro Gwanak-gu, Seoul 08826, South Korea.
| |
Collapse
|
96
|
|
97
|
Abstract
According to consciousness involvement, human’s learning can be roughly classified into explicit learning and implicit learning. Contrasting strongly to explicit learning with clear targets and rules, such as our school study of mathematics, learning is implicit when we acquire new information without intending to do so. Research from psychology indicates that implicit learning is ubiquitous in our daily life. Moreover, implicit learning plays an important role in human visual perception. But in the past 60 years, most of the well-known machine-learning models aimed to simulate explicit learning while the work of modeling implicit learning was relatively limited, especially for computer vision applications. This article proposes a novel unsupervised computational model for implicit visual learning by exploring dissipative system, which provides a unifying macroscopic theory to connect biology with physics. We test the proposed Dissipative Implicit Learning Model (DILM) on various datasets. The experiments show that DILM not only provides a good match to human behavior but also improves the explicit machine-learning performance obviously on image classification tasks.
Collapse
Affiliation(s)
- Yan Liu
- The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Yang Liu
- The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Shenghua Zhong
- The Hong Kong Polytechnic University, Hong Kong SAR, China
| | - Songtao Wu
- The Hong Kong Polytechnic University, Hong Kong SAR, China
| |
Collapse
|
98
|
Quitadamo LR, Cavrini F, Sbernini L, Riillo F, Bianchi L, Seri S, Saggio G. Support vector machines to detect physiological patterns for EEG and EMG-based human-computer interaction: a review. J Neural Eng 2017; 14:011001. [PMID: 28068295 DOI: 10.1088/1741-2552/14/1/011001] [Citation(s) in RCA: 47] [Impact Index Per Article: 5.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
Support vector machines (SVMs) are widely used classifiers for detecting physiological patterns in human-computer interaction (HCI). Their success is due to their versatility, robustness and large availability of free dedicated toolboxes. Frequently in the literature, insufficient details about the SVM implementation and/or parameters selection are reported, making it impossible to reproduce study analysis and results. In order to perform an optimized classification and report a proper description of the results, it is necessary to have a comprehensive critical overview of the applications of SVM. The aim of this paper is to provide a review of the usage of SVM in the determination of brain and muscle patterns for HCI, by focusing on electroencephalography (EEG) and electromyography (EMG) techniques. In particular, an overview of the basic principles of SVM theory is outlined, together with a description of several relevant literature implementations. Furthermore, details concerning reviewed papers are listed in tables and statistics of SVM use in the literature are presented. Suitability of SVM for HCI is discussed and critical comparisons with other classifiers are reported.
Collapse
Affiliation(s)
- L R Quitadamo
- Department of Electronic Engineering, University of Rome Tor Vergata, Rome, Italy. School of Life and Health Sciences, Aston Brain Center, Aston University, Birmingham, UK
| | | | | | | | | | | | | |
Collapse
|
99
|
|
100
|
Feng Z, Wang M, Yang S, Jiao L. Incremental Semi-Supervised classification of data streams via self-representative selection. Appl Soft Comput 2016. [DOI: 10.1016/j.asoc.2016.02.023] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|