1
|
Wind/Storage Power Scheduling Based on Time–Sequence Rolling Optimization. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2022. [DOI: 10.1007/s13369-022-07220-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
2
|
Evaluating the Effectiveness of Machine Learning and Deep Learning Models Combined Time-Series Satellite Data for Multiple Crop Types Classification over a Large-Scale Region. REMOTE SENSING 2022. [DOI: 10.3390/rs14102341] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
Abstract
Accurate extraction of crop cultivated area and spatial distribution is essential for food security. Crop classification methods based on machine learning and deep learning and remotely sensed time-series data are widely utilized to detect crop planting area. However, few studies assess the effectiveness of machine learning and deep learning algorithm integrated time-series satellite data for identifying multiple crop type classification over a large-scale region. Hence, this study aims to evaluate the effectiveness of machine learning and deep learning models in crop classification and provide a framework for large-scale multiple crop type classification based on time-series of satellite data. The time-series of the Normalized Difference Vegetation Index (NDVI), Enhanced Vegetation Index (EVI), and NaE (combined NDVI and EVI) were adopted as input features, and four widely used machine learning models, including Support Vector Machine (SVM), Random Forest (RF), K-Nearest Neighbor (KNN), and their integration (Stacking) were compared to examine the performance of multiple crop types (i.e., wheat, corn, wheat-corn, early rice, and early rice-late rice) classification in the North China Plain. The performance of two types of deep learning classifiers: the One-dimensional Convolutional Neural Network (Conv1D) and Long Short-Term Memory Networks (LSTM), were also tested. The results showed that the NaE feature performed best among three input features, and the Stacking model produced the highest accuracy (77.12%) compared to other algorithms.
Collapse
|
3
|
Selecting the Suitable Resampling Strategy for Imbalanced Data Classification Regarding Dataset Properties. An Approach Based on Association Models. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11188546] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
Abstract
In many application domains such as medicine, information retrieval, cybersecurity, social media, etc., datasets used for inducing classification models often have an unequal distribution of the instances of each class. This situation, known as imbalanced data classification, causes low predictive performance for the minority class examples. Thus, the prediction model is unreliable although the overall model accuracy can be acceptable. Oversampling and undersampling techniques are well-known strategies to deal with this problem by balancing the number of examples of each class. However, their effectiveness depends on several factors mainly related to data intrinsic characteristics, such as imbalance ratio, dataset size and dimensionality, overlapping between classes or borderline examples. In this work, the impact of these factors is analyzed through a comprehensive comparative study involving 40 datasets from different application areas. The objective is to obtain models for automatic selection of the best resampling strategy for any dataset based on its characteristics. These models allow us to check several factors simultaneously considering a wide range of values since they are induced from very varied datasets that cover a broad spectrum of conditions. This differs from most studies that focus on the individual analysis of the characteristics or cover a small range of values. In addition, the study encompasses both basic and advanced resampling strategies that are evaluated by means of eight different performance metrics, including new measures specifically designed for imbalanced data classification. The general nature of the proposal allows the choice of the most appropriate method regardless of the domain, avoiding the search for special purpose techniques that could be valid for the target data.
Collapse
|
4
|
Sucholutsky I, Schonlau M. Optimal 1-NN prototypes for pathological geometries. PeerJ Comput Sci 2021; 7:e464. [PMID: 33954242 PMCID: PMC8049135 DOI: 10.7717/peerj-cs.464] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2020] [Accepted: 03/11/2021] [Indexed: 06/12/2023]
Abstract
Using prototype methods to reduce the size of training datasets can drastically reduce the computational cost of classification with instance-based learning algorithms like the k-Nearest Neighbour classifier. The number and distribution of prototypes required for the classifier to match its original performance is intimately related to the geometry of the training data. As a result, it is often difficult to find the optimal prototypes for a given dataset, and heuristic algorithms are used instead. However, we consider a particularly challenging setting where commonly used heuristic algorithms fail to find suitable prototypes and show that the optimal number of prototypes can instead be found analytically. We also propose an algorithm for finding nearly-optimal prototypes in this setting, and use it to empirically validate the theoretical results. Finally, we show that a parametric prototype generation method that normally cannot solve this pathological setting can actually find optimal prototypes when combined with the results of our theoretical analysis.
Collapse
|
5
|
EEkNN: k-Nearest Neighbor Classifier with an Evidential Editing Procedure for Training Samples. ELECTRONICS 2019. [DOI: 10.3390/electronics8050592] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The k-nearest neighbor (kNN) rule is one of the most popular classification algorithms applied in many fields because it is very simple to understand and easy to design. However, one of the major problems encountered in using the kNN rule is that all of the training samples are considered equally important in the assignment of the class label to the query pattern. In this paper, an evidential editing version of the kNN rule is developed within the framework of belief function theory. The proposal is composed of two procedures. An evidential editing procedure is first proposed to reassign the original training samples with new labels represented by an evidential membership structure, which provides a general representation model regarding the class membership of the training samples. After editing, a classification procedure specifically designed for evidently edited training samples is developed in the belief function framework to handle the more general situation in which the edited training samples are assigned dependent evidential labels. Three synthetic datasets and six real datasets collected from various fields were used to evaluate the performance of the proposed method. The reported results show that the proposal achieves better performance than other considered kNN-based methods, especially for datasets with high imprecision ratios.
Collapse
|
6
|
|
7
|
An improved fast edit approach for two-string approximated mean computation applied to OCR. Pattern Recognit Lett 2013. [DOI: 10.1016/j.patrec.2012.11.019] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
8
|
|
9
|
TOUSSAINT GODFRIED. GEOMETRIC PROXIMITY GRAPHS FOR IMPROVING NEAREST NEIGHBOR METHODS IN INSTANCE-BASED LEARNING AND DATA MINING. ACTA ACUST UNITED AC 2012. [DOI: 10.1142/s0218195905001622] [Citation(s) in RCA: 50] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
In the typical nonparametric approach to classification in instance-based learning and data mining, random data (the training set of patterns) are collected and used to design a decision rule (classifier). One of the most well known such rules is the k-nearest-neighbor decision rule (also known as lazy learning) in which an unknown pattern is classified into the majority class among its k nearest neighbors in the training set. Several questions related to this rule have received considerable attention over the years. Such questions include the following. How can the storage of the training set be reduced without degrading the performance of the decision rule? How should the reduced training set be selected to represent the different classes? How large should k be? How should the value of k be chosen? Should all k neighbors be equally weighted when used to decide the class of an unknown pattern? If not, how should the weights be chosen? Should all the features (attributes) we weighted equally and if not how should the feature weights be chosen? What distance metric should be used? How can the rule be made robust to overlapping classes or noise present in the training data? How can the rule be made invariant to scaling of the measurements? How can the nearest neighbors of a new point be computed efficiently? What is the smallest neural network that can implement nearest neighbor decision rules? Geometric proximity graphs such as Voronoi diagrams and their many relatives provide elegant solutions to these problems, as well as other related data mining problems such as outlier detection. After a non-exhaustive review of some of the classical canonical approaches to these problems, the methods that use proximity graphs are discussed, some new observations are made, and open problems are listed.
Collapse
Affiliation(s)
- GODFRIED TOUSSAINT
- School of Computer Science, McGill University, 3480 University, St., McConnell Eng. Building, Room 318, Montréal, Québec H3A 2A7, Canada
| |
Collapse
|
10
|
|
11
|
Paredes R, Vidal E. Learning weighted metrics to minimize nearest-neighbor classification error. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2006; 28:1100-10. [PMID: 16792099 DOI: 10.1109/tpami.2006.145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
In order to optimize the accuracy of the Nearest-Neighbor classification rule, a weighted distance is proposed, along with algorithms to automatically learn the corresponding weights. These weights may be specific for each class and feature, for each individual prototype, or for both. The learning algorithms are derived by (approximately) minimizing the Leaving-One-Out classification error of the given training set. The proposed approach is assessed through a series of experiments with UCI/STATLOG corpora, as well as with a more specific task of text classification which entails very sparse data representation and huge dimensionality. In all these experiments, the proposed approach shows a uniformly good behavior, with results comparable to or better than state-of-the-art results published with the same data so far.
Collapse
Affiliation(s)
- Roberto Paredes
- Departamento de Sistemas Informáticos y Computación, Instituto Tecnológico de Informática, Universidad Politiécnica de Valencia, Spain.
| | | |
Collapse
|
12
|
|
13
|
Correcting the Training Data. ACTA ACUST UNITED AC 2003. [DOI: 10.1007/978-1-4613-0231-5_1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register]
|
14
|
Paredes R, Vidal E. A class-dependent weighted dissimilarity measure for nearest neighbor classification problems. Pattern Recognit Lett 2000. [DOI: 10.1016/s0167-8655(00)00064-7] [Citation(s) in RCA: 52] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
15
|
Decontamination of Training Samples for Supervised Pattern Recognition Methods. ADVANCES IN PATTERN RECOGNITION 2000. [DOI: 10.1007/3-540-44522-6_64] [Citation(s) in RCA: 42] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
16
|
Kuyel T, Geisler W, Ghosh J. Fast image classification using a sequence of visual fixations. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS. PART B, CYBERNETICS : A PUBLICATION OF THE IEEE SYSTEMS, MAN, AND CYBERNETICS SOCIETY 1999; 29:304-308. [PMID: 18252304 DOI: 10.1109/3477.752805] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Based on human retinal sampling distributions and eye movements, a sequential resolution image preprocessor is developed. Combined with a nearest neighbor classifier, this preprocessor provides an efficient image classification method, the sequential resolution nearest neighbor (SRNN) classifier. The human eye has a typical fixation sequence that exploits the nonuniform sampling distribution of its retina. If the retinal resolution is not sufficient to identify an object, the eye moves in such a way that the projection of the object falls onto a retinal region with a higher sampling density. Similarly, the SRNN classifier uses a sequence of increasing resolutions until a final class decision is made. Experimental results on texture segmentation show that the preprocessor used in the SRNN classifier is considerably faster than traditional multiresolution algorithms which use all the available resolution levels to analyze the input data.
Collapse
Affiliation(s)
- T Kuyel
- Texas Instrum. Inc., Dallas, TX
| | | | | |
Collapse
|
17
|
Ferri F, Albert J, Vidal E. Considerations about sample-size sensitivity of a family of edited nearest-neighbor rules. ACTA ACUST UNITED AC 1999; 29:667-72. [DOI: 10.1109/3477.790454] [Citation(s) in RCA: 44] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
18
|
Romer C, Kandel A. Comments on "Constraints on belief functions imposed by fuzzy random variables": some technical remarks on Romer/Kandel. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS. PART B, CYBERNETICS : A PUBLICATION OF THE IEEE SYSTEMS, MAN, AND CYBERNETICS SOCIETY 1999; 29:672. [PMID: 18252347 DOI: 10.1109/3477.790455] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
First, we would like to thank V. Kratschmer for his validation of our results in the paper regarding the belief measure by using a topological approach. Though assertions (1) and (3) are presented in a weakened fashion, our results still remain valid, as he claims. It is true that assertion (2) has been proved by us only for Borel sets B, which have at most countable components. We were not able to prove the same result for Borel sets with uncountable components (such as the irrational numbers, for example) using our line of reasoning. We therefore applaud the proof presented by V. Kratschmer for the more general Borel sets using an interesting use of some topological properties induced by the Hansdorff metric defined on the space of closed intervals of the real numbers. This certainly makes our original approach to fuzzy data analysis combining fuzzy sets theory and Dempster-Shafer even more useful.
Collapse
Affiliation(s)
- C Romer
- Dept. of Comput. Sci. & Eng., Univ. of South Florida, Tampa, FL
| | | |
Collapse
|
19
|
TURNEY PETER. Theoretical analyses of cross-validation error and voting in instance-based learning. J EXP THEOR ARTIF IN 1994. [DOI: 10.1080/09528139408953793] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
20
|
Loizou G, Maybank SJ. The nearest neighbor and the bayes error rates. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1987; 9:254-262. [PMID: 21869395 DOI: 10.1109/tpami.1987.4767899] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
The (k, l) nearest neighbor method of pattern classification is compared to the Bayes method. If the two acceptance rates are equal then the asymptotic error rates satisfy the inequalities Ek,l + 1 ¿ E*(¿) ¿ Ek,l dE*(¿), where d is a function of k, l, and the number of pattern classes, and ¿ is the reject threshold for the Bayes method. An explicit expression for d is given which is optimal in the sense that for some probability distributions Ek,l and dE* (¿) are equal.
Collapse
Affiliation(s)
- G Loizou
- Department of Computer Science, Birkbeck College, University of London, Malet Street, London WC1E 7HX, England
| | | |
Collapse
|
21
|
|
22
|
Fukunaga K, Mantock JM. A nonparametric two-dimensional display for classification. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1982; 4:427-436. [PMID: 21869059 DOI: 10.1109/tpami.1982.4767276] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
A two-dimensional display whose coordinates are related to the distance to the kth-nearest neighbor of each class is presented. Applications of the display to minimum error, minimum cost, minimax, and Neyman-Pearson type classifier designs are given. The display is shown to present risk information in a manner that easily allows the specification of reject regions. Two methods of error estimation using the display, an error counting technique and a risk averaging method, are detailed. It is shown that the classifiers that result are generalizations of the standard k-NN majority vote classifier. As a result of the properties of the display, classifiers can be readily evaluated and modified. In addition, a condensing algorithm that preserves the nearest neighbor error count of any preclassified data set is described. The display is used to graphically illustrate the distance relationships that are central to the algorithm.
Collapse
Affiliation(s)
- K Fukunaga
- Department of Electrical Engineering, Purdue University, West Lafayette, IN 47907
| | | |
Collapse
|
23
|
Dasarathy BV. Nosing around the neighborhood: a new system structure and classification rule for recognition in partially exposed environments. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1980; 2:67-71. [PMID: 22499625 DOI: 10.1109/tpami.1980.4766972] [Citation(s) in RCA: 24] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
The scope of the classical k-NN classification techniques is enlarged under this study to cover partially exposed environments. The modified classification system structure required for successful operation in environments, wherein all the inherent pattern classes are not exposed to the system prior to deployment, is developed and illustrated with the aid of a specific classification rule-the neighborhood census rule (NCR). Admittedly, alternative rules can be visualized to fit this modified structure. However, this study concentrates on the use of NCR to bring out the underlying philosophy and develops optimum thresholds for admittance of unknown samples into the set of presently known classes. These thresholds are learned from the available training samples of these classes. This learning represents a new dimensionality of the learning system structure in that estimates of the domains of the known classes are developed in addition to learning of the discrimination among these classes. This facilitates identification of samples belonging to the classes previously unexposed to the recognition system. Experimental results are also presented in support of the proposed concepts and methodology for operation in partially exposed environments.
Collapse
|
24
|
|
25
|
|
26
|
|