1
|
Ma J, Yu G. Distribution-free Bayesian regularized learning framework for semi-supervised learning. Neural Netw 2024; 174:106262. [PMID: 38547803 DOI: 10.1016/j.neunet.2024.106262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 02/09/2024] [Accepted: 03/19/2024] [Indexed: 04/14/2024]
Abstract
In machine learning it is often necessary to assume or know the distribution of the data, however it is difficult to do so in practical applications. Aiming to this problem, this work, we propose a novel distribution-free Bayesian regularized learning framework for semi-supervised learning, which is called Hessian regularized twin minimax probability extreme learning machine (HRTMPELM). In this framework, we attempt to construct two non-parallel hyperplanes by introducing the high separation probability assumption, such that each hyperplane separates samples from one class with maximum probability while moving away from samples from the other class. Subsidiently, the framework can be utilized to construct reasonable semi-supervised classifiers by using the information of the inherent geometric distribution of the samples through the Hessian regularization term. Additionally, the proposed framework controls the misclassification error of samples by minimizing the upper limit of the worst-case misclassification probability, and improves the generalization performance of the model by introducing the idea of regularization to avoid the occurrence of ill-posedness and overfitting problems. More importantly, the framework has no hyperparameters, making the learning process very simplified and efficient. Finally, a simple and reliable algorithm with globally optimal solutions via multivariate Chebyshev inequalities is designed for solving the proposed learning framework. Experiments on multiple datasets demonstrate the reliability and effectiveness of the proposed learning framework compared to other methods. Especially, we applied the framework to Ningxia wolfberry quality detection, which greatly enriches and facilitates the application of machine learning algorithms in the agricultural field.
Collapse
Affiliation(s)
- Jun Ma
- School of Mathematics and Information Sciences, North Minzu University, Yinchuan Ningxia 750021, PR China.
| | - Guolin Yu
- School of Mathematics and Information Sciences, North Minzu University, Yinchuan Ningxia 750021, PR China.
| |
Collapse
|
2
|
Alsharari F, Saber Y, Alohali H, Alqahtani MH, Ebodey M, Elmasry T, Alsharif J, Soliman AF, Smarandache F, Sikander F. On stratified single-valued soft topogenous structures. Heliyon 2024; 10:e27926. [PMID: 39670082 PMCID: PMC11636833 DOI: 10.1016/j.heliyon.2024.e27926] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2023] [Revised: 03/03/2024] [Accepted: 03/08/2024] [Indexed: 12/14/2024] Open
Abstract
This paper presents novel concepts including stratified single-valued neutrosophic soft topogenous (stratified svns-topogenous), stratified single-valued neutrosophic soft filter (stratified svns-filter), stratified single-valued neutrosophic soft quasi uniformity (stratified svnsq-uniformity) and stratified single-valued neutrosophic soft quasi proximity (stratified svnsq-proximity). Additionally, we present the idea of single-valued neutrosophic soft topogenous structures, formed by integrating svns-topogenous with svns-filter, and discuss their properties. Furthermore, we explore the connections between these single-valued neutrosophic soft topological structures and their corresponding stratifications.
Collapse
Affiliation(s)
- Fahad Alsharari
- Department of Mathematics, College of Science, Jouf University, Sakaka 72311, Saudi Arabia
| | - Yaser Saber
- Department of Mathematics, College of Science Al-Zulfi, Majmaah University, P. O. Box 66, Al-Majmaah 11952, Saudi Arabia
- Department of Mathematics, Faculty of Science, Al-Azhar University, Assiut, 71524, Egypt
| | - Hanan Alohali
- Department of Mathematics, College of Science, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
| | - Mesfer H. Alqahtani
- Department of Mathematics, University College of Umluj, University of Tabuk, Tabuk 48322, Saudi Arabia
| | - Mubarak Ebodey
- Department of Business Administration, Faculty of Science and Humanities at Hotat Sudair, Majmaah University, 11952, Riyadh, Saudi Arabia
| | - Tawfik Elmasry
- Department of Business Administration, Faculty of Science and Humanities at Hotat Sudair, Majmaah University, 11952, Riyadh, Saudi Arabia
| | - Jafar Alsharif
- Department of Business Administration, Faculty of Science and Humanities at Hotat Sudair, Majmaah University, 11952, Riyadh, Saudi Arabia
| | - Amal F. Soliman
- Department of Mathematics, College of Science and Humanities in Alkharj, Prince Sattam Bin Abdulaziz University, Alkharj, Saudi Arabia
- Department of Basic Science, Benha Faculty of Engineering, Benha University, Banha, Egypt
| | | | - Fahad Sikander
- Department of Basics Sciences, College of Science and Theoretical studies, Saudi Electronic University, Jeddah 23442, Saudi Arabia
| |
Collapse
|
3
|
Wang L, Cui G, Cai X. Fuzzy clustering optimal k selection method based on multi-objective optimization. Soft comput 2023. [DOI: 10.1007/s00500-022-07727-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
|
4
|
|
5
|
Liu L, Li P, Chu M, Cai H. Stochastic gradient support vector machine with local structural information for pattern recognition. INT J MACH LEARN CYB 2021. [DOI: 10.1007/s13042-021-01303-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
6
|
Kernel-Free Quadratic Surface Minimax Probability Machine for a Binary Classification Problem. Symmetry (Basel) 2021. [DOI: 10.3390/sym13081378] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
In this paper, we propose a novel binary classification method called the kernel-free quadratic surface minimax probability machine (QSMPM), that makes use of the kernel-free techniques of the quadratic surface support vector machine (QSSVM) and inherits the advantage of the minimax probability machine (MPM) without any parameters. Specifically, it attempts to find a quadratic hypersurface that separates two classes of samples with maximum probability. However, the optimization problem derived directly was too difficult to solve. Therefore, a nonlinear transformation was introduced to change the quadratic function involved into a linear function. Through such processing, our optimization problem finally became a second-order cone programming problem, which was solved efficiently by an alternate iteration method. It should be pointed out that our method is both kernel-free and parameter-free, making it easy to use. In addition, the quadratic hypersurface obtained by our method was allowed to be any general form of quadratic hypersurface. It has better interpretability than the methods with the kernel function. Finally, in order to demonstrate the geometric interpretation of our QSMPM, five artificial datasets were implemented, including showing the ability to obtain a linear separating hyperplane. Furthermore, numerical experiments on benchmark datasets confirmed that the proposed method had better accuracy and less CPU time than corresponding methods.
Collapse
|
7
|
Tao X, Chen W, Li X, Zhang X, Li Y, Guo J. The ensemble of density-sensitive SVDD classifier based on maximum soft margin for imbalanced datasets. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.106897] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
8
|
Gu B, Shan Y, Quan X, Zheng G. Accelerating Sequential Minimal Optimization via Stochastic Subgradient Descent. IEEE TRANSACTIONS ON CYBERNETICS 2021; 51:2215-2223. [PMID: 30736010 DOI: 10.1109/tcyb.2019.2893289] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
Sequential minimal optimization (SMO) is one of the most popular methods for solving a variety of support vector machines (SVMs). The shrinking and caching techniques are commonly used to accelerate SMO. An interesting phenomenon of SMO is that most of the computational time is wasted on the first half of iterations for building a good solution closing to the optimal. However, as we all know, the stochastic subgradient descent (SSGD) method is extremely fast for building a good solution. In this paper, we propose a generalized framework of accelerating SMO through SSGD for a variety of SVMs of binary classification, regression, ordinal regression, and so on. We also provide a deep insight about why SSGD can accelerate SMO. Experimental results on a variety of datasets and learning applications confirm that our method can effectively speed up SMO.
Collapse
|
9
|
Yang L, Wen Y, Zhang M, Wang X. Twin minimax probability machine for pattern classification. Neural Netw 2020; 131:201-214. [PMID: 32801059 DOI: 10.1016/j.neunet.2020.07.030] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2019] [Revised: 06/01/2020] [Accepted: 07/23/2020] [Indexed: 11/29/2022]
Abstract
We propose a new distribution-free Bayes optimal classifier, called the twin minimax probability machine (TWMPM), which combines the benefits of both minimax probability machine(MPM) and twin support vector machine (TWSVM). TWMPM tries to construct two nonparallel hyperplanes such that each hyperplane separates one class samples with maximal probability, and is distant from the other class samples simultaneously. Moreover, the proposed TWMPM can control the misclassification error of samples in a worst-case setting by minimizing the upper bound on misclassification probability. An efficient algorithm for TWMPM is first proposed, which transforms TWMPM into concave fractional programming by applying multivariate Chebyshev inequality. Then the proposed TWMPM is reformulated as a pair of convex quadric programming (QP) by proper mathematical transformations. This guarantees TWMPM to have global optimal solution and be solved simply and effectively. In addition, we develop also an iterative algorithm for the proposed TWMPM. By comparing the two proposed algorithms theoretically, it is easy to know that the convex quadric programming algorithm is with lower computation burden than iterative algorithm for the TWMPM. A linear TWMPM version is first built, and then we show how to exploit mercer kernel to obtain nonlinear TWMPM version. The computation complexity for QP algorithm of TWMPM is in the same order as the traditional twin support vector machine (TWSVM). Experiments are carried out on three databases: UCI benchmark database, a practical application database and an artificial database. With low computation complexity and fewer parameters, experiments show the feasibility and effectiveness of the proposed TWMPM and its QP algorithm.
Collapse
Affiliation(s)
- Liming Yang
- College of Science, China Agricultural University, Beijing, 100083, China.
| | - Yakun Wen
- College of electronic information and engineering, China Agricultural University, China
| | - Min Zhang
- College of Science, China Agricultural University, Beijing, 100083, China
| | - Xue Wang
- College of electronic information and engineering, China Agricultural University, China
| |
Collapse
|
10
|
Wu J, Sheng VS, Zhang J, Li H, Dadakova T, Swisher CL, Cui Z, Zhao P. Multi-Label Active Learning Algorithms for Image Classification: Overview and Future Promise. ACM COMPUTING SURVEYS 2020; 53:28. [PMID: 34421185 PMCID: PMC8376181 DOI: 10.1145/3379504] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/01/2017] [Accepted: 12/01/2019] [Indexed: 05/13/2023]
Abstract
Image classification is a key task in image understanding, and multi-label image classification has become a popular topic in recent years. However, the success of multi-label image classification is closely related to the way of constructing a training set. As active learning aims to construct an effective training set through iteratively selecting the most informative examples to query labels from annotators, it was introduced into multi-label image classification. Accordingly, multi-label active learning is becoming an important research direction. In this work, we first review existing multi-label active learning algorithms for image classification. These algorithms can be categorized into two top groups from two aspects respectively: sampling and annotation. The most important component of multi-label active learning is to design an effective sampling strategy that actively selects the examples with the highest informativeness from an unlabeled data pool, according to various information measures. Thus, different informativeness measures are emphasized in this survey. Furthermore, this work also makes a deep investigation on existing challenging issues and future promises in multi-label active learning with a focus on four core aspects: example dimension, label dimension, annotation, and application extension.
Collapse
Affiliation(s)
- Jian Wu
- Soochow University, China and Human Longevity, Inc., USA
| | | | - Jing Zhang
- Nanjing University of Science and Technology, China
| | - Hua Li
- Washington University in St. Louis, USA
| | | | | | - Zhiming Cui
- Suzhou University of Science and Technology, China
| | | |
Collapse
|
11
|
Ma J, Shen J. RETRACTED: A novel twin minimax probability machine for classification and regression. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2020.105703] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
|
12
|
Wang Y, Pan Z, Pan Y. A Training Data Set Cleaning Method by Classification Ability Ranking for the k -Nearest Neighbor Classifier. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:1544-1556. [PMID: 31265416 DOI: 10.1109/tnnls.2019.2920864] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
The k -nearest neighbor (KNN) rule is a successful technique in pattern classification due to its simplicity and effectiveness. As a supervised classifier, KNN classification performance usually suffers from low-quality samples in the training data set. Thus, training data set cleaning (TDC) methods are needed for enhancing the classification accuracy by cleaning out noisy, or even wrong, samples in the original training data set. In this paper, we propose a classification ability ranking (CAR)-based TDC method to improve the performance of a KNN classifier, namely CAR-based TDC method. The proposed classification ability function ranks a training sample in terms of its contribution to correctly classify other training samples as a KNN through the leave-one-out (LV1) strategy in the cleaning stage. The training sample that likely misclassifies the other samples during the KNN classifications according to the LV1 strategy is considered to have lower classification ability and will be cleaned out from the original training data set. Extensive experiments, based on ten real-world data sets, show that the proposed CAR-based TDC method can significantly reduce the classification error rates of KNN-based classifiers, while reducing computational complexity thanks to a smaller cleaned training data set.
Collapse
|
13
|
Ma J, Yang L, Wen Y, Sun Q. Twin minimax probability extreme learning machine for pattern recognition. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2019.06.014] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
14
|
Investigation of High-Efficiency Iterative ILU Preconditioner Algorithm for Partial-Differential Equation Systems. Symmetry (Basel) 2019. [DOI: 10.3390/sym11121461] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
In this paper, we investigate an iterative incomplete lower and upper (ILU) factorization preconditioner for partial-differential equation systems. We discretize the partial-differential equations into linear equation systems. An iterative scheme of linear systems is used. The ILU preconditioners of linear systems are performed on the different computation nodes of multi-central processing unit (CPU) cores. Firstly, the preconditioner of general tridiagonal matrix equations is tested on supercomputers. Then, the effects of partial-differential equation systems on the speedup of parallel multiprocessors are examined. The numerical results estimate that the parallel efficiency is higher than in other algorithms.
Collapse
|
15
|
A quantum-behaved particle swarm optimization algorithm with the flexible single-/multi-population strategy and multi-stage perturbation strategy based on the characteristics of objective function. Soft comput 2019. [DOI: 10.1007/s00500-019-04328-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
16
|
|
17
|
Xu ZC, Xiao X, Qiu WR, Wang P, Fang XZ. iAI-DSAE: A Computational Method for Adenosine to Inosine Editing Site Prediction. LETT ORG CHEM 2019. [DOI: 10.2174/1570178615666181016112546] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022]
Abstract
As an important post-transcriptional modification, adenosine-to-inosine RNA editing generally occurs in both coding and noncoding RNA transcripts in which adenosines are converted to inosines. Accordingly, the diversification of the transcriptome can be resulted in by this modification. It is significant to accurately identify adenosine-to-inosine editing sites for further understanding their biological functions. Currently, the adenosine-to-inosine editing sites would be determined by experimental methods, unfortunately, it may be costly and time consuming. Furthermore, there are only a few existing computational prediction models in this field. Therefore, the work in this study is starting to develop other computational methods to address these problems. Given an uncharacterized RNA sequence that contains many adenosine resides, can we identify which one of them can be converted to inosine, and which one cannot? To deal with this problem, a novel predictor called iAI-DSAE is proposed in the current study. In fact, there are two key issues to address: one is ‘what feature extraction methods should be adopted to formulate the given sample sequence?’ The other is ‘what classification algorithms should be used to construct the classification model?’ For the former, a 540-dimensional feature vector is extracted to formulate the sample sequence by dinucleotide-based auto-cross covariance, pseudo dinucleotide composition, and nucleotide density methods. For the latter, we use the present more popular method i.e. deep spare autoencoder to construct the classification model. Generally, ACC and MCC are considered as the two of the most important performance indicators of a predictor. In this study, in comparison with those of predictor PAI, they are up 2.46% and 4.14%, respectively. The two other indicators, Sn and Sp, rise at certain degree also. This indicates that our predictor can be as an important complementary tool to identify adenosine-toinosine RNA editing sites. For the convenience of most experimental scientists, an easy-to-use web-server for identifying adenosine-to-inosine editing sites has been established at: http://www.jci-bioinfo.cn/iAI-DSAE, by which users can easily obtain their desired results without the need to go through the complicated mathematical equations involved. It is important to identify adenosine-to-inosine editing sites in RNA sequences for the intensive study on RNA function and the development of new medicine. In current study, a novel predictor, called iAI-DSAE, was proposed by using three feature extraction methods including dinucleotidebased auto-cross covariance, pseudo dinucleotide composition and nucleotide density. The jackknife test results of the iAI-DSAE predictor based on deep spare auto-encoder model show that our predictor is more stable and reliable. It has not escaped our notice that the methods proposed in the current paper can be used to solve many other problems in genome analysis.
Collapse
Affiliation(s)
- Zhao-Chun Xu
- Computer Department, Jing-De-Zhen Ceramic Institute, Jing-De-Zhen 333403, China
| | - Xuan Xiao
- Computer Department, Jing-De-Zhen Ceramic Institute, Jing-De-Zhen 333403, China
| | - Wang-Ren Qiu
- Computer Department, Jing-De-Zhen Ceramic Institute, Jing-De-Zhen 333403, China
| | - Peng Wang
- Computer Department, Jing-De-Zhen Ceramic Institute, Jing-De-Zhen 333403, China
| | - Xin-Zhu Fang
- Computer Department, Jing-De-Zhen Ceramic Institute, Jing-De-Zhen 333403, China
| |
Collapse
|
18
|
Bonyadi MR, Tieng QM, Reutens DC. Optimization of Distributions Differences for Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2019; 30:511-523. [PMID: 29994733 DOI: 10.1109/tnnls.2018.2844723] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
In this paper, we introduce a new classification algorithm called the optimization of distribution differences (ODD). The algorithm aims to find a transformation from the feature space to a new space where the instances in the same class are as close as possible to one another, whereas the gravity centers of these classes are as far as possible from one another. This aim is formulated as a multiobjective optimization problem that is solved by a hybrid of an evolutionary strategy and the quasi-Newton method. The choice of the transformation function is flexible and could be any continuous space function. We experiment with a linear and a nonlinear transformation in this paper. We show that the algorithm can outperform eight other classification methods, namely naive Bayes, support vector machines, linear discriminant analysis, multilayer perceptrons, decision trees, and k -nearest neighbors, and two recently proposed classification methods, in 12 standard classification data sets. Our results show that the method is less sensitive to the imbalanced number of instances compared with these methods. We also show that ODD maintains its performance better than other classification methods in these data sets and hence offers a better generalization ability.
Collapse
|
19
|
Torkzadeh S, Soltanizadeh H, Orouji AA. Multi-constraint QoS routing using a customized lightweight evolutionary strategy. Soft comput 2019. [DOI: 10.1007/s00500-018-3018-z] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
20
|
A discriminated similarity matrix construction based on sparse subspace clustering algorithm for hyperspectral imagery. COGN SYST RES 2019. [DOI: 10.1016/j.cogsys.2018.01.003] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
21
|
|
22
|
Chen Y, Tian X, Liu H, Chen Z, Yang B, Liao W, Zhang P, He R, Yang M. Parallel ILU preconditioners in GPU computation. Soft comput 2018. [DOI: 10.1007/s00500-017-2764-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
23
|
|
24
|
Guo W, Si C, Xue Y, Mao Y, Wang L, Wu Q. A Grouping Particle Swarm Optimizer with Personal-Best-Position Guidance for Large Scale Optimization. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2018; 15:1904-1915. [PMID: 28489542 DOI: 10.1109/tcbb.2017.2701367] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Particle Swarm Optimization (PSO) is a popular algorithm which is widely investigated and well implemented in many areas. However, the canonical PSO does not perform well in population diversity maintenance so that usually leads to a premature convergence or local optima. To address this issue, we propose a variant of PSO named Grouping PSO with Personal-Best-Position ( Pbest) Guidance (GPSO-PG) which maintains the population diversity by preserving the diversity of exemplars. On one hand, we adopt uniform random allocation strategy to assign particles into different groups and in each group the losers will learn from the winner. On the other hand, we employ personal historical best position of each particle in social learning rather than the current global best particle. In this way, the exemplars diversity increases and the effect from the global best particle is eliminated. We test the proposed algorithm to the benchmarks in CEC 2008 and CEC 2010, which concern the large scale optimization problems (LSOPs). By comparing several current peer algorithms, GPSO-PG exhibits a competitive performance to maintain population diversity and obtains a satisfactory performance to the problems.
Collapse
|
25
|
Learning from crowds with active learning and self-healing. Neural Comput Appl 2018. [DOI: 10.1007/s00521-017-2878-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
26
|
|
27
|
Shen X, Shen F, Liu L, Yuan YH, Liu W, Sun QS. Multiview Discrete Hashing for Scalable Multimedia Search. ACM T INTEL SYST TEC 2018. [DOI: 10.1145/3178119] [Citation(s) in RCA: 22] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
Abstract
Hashing techniques have recently gained increasing research interest in multimedia studies. Most existing hashing methods only employ single features for hash code learning. Multiview data with each view corresponding to a type of feature generally provides more comprehensive information. How to efficiently integrate multiple views for learning compact hash codes still remains challenging. In this article, we propose a novel unsupervised hashing method, dubbed multiview discrete hashing (MvDH), by effectively exploring multiview data. Specifically, MvDH performs matrix factorization to generate the hash codes as the latent representations shared by multiple views, during which spectral clustering is performed simultaneously. The joint learning of hash codes and cluster labels enables that MvDH can generate more discriminative hash codes, which are optimal for classification. An efficient alternating algorithm is developed to solve the proposed optimization problem with guaranteed convergence and low computational complexity. The binary codes are optimized via the discrete cyclic coordinate descent (DCC) method to reduce the quantization errors. Extensive experimental results on three large-scale benchmark datasets demonstrate the superiority of the proposed method over several state-of-the-art methods in terms of both accuracy and scalability.
Collapse
Affiliation(s)
| | - Fumin Shen
- University of Electronic Science and Technology of China, Chengdu, China
| | - Li Liu
- Northumbria University, UK
| | | | - Weiwei Liu
- The University of New South Wales, Sydney, NSW, Australia
| | - Quan-Sen Sun
- Nanjing University of Science and Technology, Nanjing, China
| |
Collapse
|
28
|
Shen X, Liu W, Tsang IW, Sun QS, Ong YS. Multilabel Prediction via Cross-View Search. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4324-4338. [PMID: 29990175 DOI: 10.1109/tnnls.2017.2763967] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Embedding methods have shown promising performance in multilabel prediction, as they are able to discover the label dependence. However, most methods ignore the correlations between the input and output, such that their learned embeddings are not well aligned, which leads to degradation in prediction performance. This paper presents a formulation for multilabel learning, from the perspective of cross-view learning, that explores the correlations between the input and the output. The proposed method, called Co-Embedding (CoE), jointly learns a semantic common subspace and view-specific mappings within one framework. The semantic similarity structure among the embeddings is further preserved, ensuring that close embeddings share similar labels. Additionally, CoE conducts multilabel prediction through the cross-view $k$ nearest neighborhood ( $k$ NN) search among the learned embeddings, which significantly reduces computational costs compared with conventional decoding schemes. A hashing-based model, i.e., Co-Hashing (CoH), is further proposed. CoH is based on CoE, and imposes the binary constraint on continuous latent embeddings. CoH aims to generate compact binary representations to improve the prediction efficiency by benefiting from the efficient $k$ NN search of multiple labels in the Hamming space. Extensive experiments on various real-world data sets demonstrate the superiority of the proposed methods over the state of the arts in terms of both prediction accuracy and efficiency.
Collapse
|
29
|
Wang Y, Kim K, Lee B, Youn HY. Word clustering based on POS feature for efficient twitter sentiment analysis. HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES 2018. [DOI: 10.1186/s13673-018-0140-y] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
AbstractWith rapid growth of social networking service on Internet, huge amount of information are continuously generated in real time. As a result, sentiment analysis of online reviews and messages has become a popular research issue [1]. In this paper a novel modified Chi Square-based feature clustering and weighting scheme is proposed for the sentiment analysis of twitter message. Along with the part of speech tagging, the discriminability and dependency of the words in the tagged training dataset are taken into account in the clustering and weighting process. The multinomial Naïve Bayes model is also employed to handle redundant features, and the influence of emotional words is raised for maximizing the accuracy. Computer simulation with Sentiment 140 workload shows that the proposed scheme significantly outperforms four existing representative sentiment analysis schemes in terms of the accuracy regardless of the size of training and test data.
Collapse
|
30
|
He G, Zhao W, Xia X, Peng R, Wu X. An ensemble of shapelet-based classifiers on inter-class and intra-class imbalanced multivariate time series at the early stage. Soft comput 2018. [DOI: 10.1007/s00500-018-3261-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
31
|
|
32
|
Learning effective binary descriptors for micro-expression recognition transferred by macro-information. Pattern Recognit Lett 2018. [DOI: 10.1016/j.patrec.2017.07.010] [Citation(s) in RCA: 32] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
33
|
|
34
|
A hybrid gravitational search algorithm with swarm intelligence and deep convolutional feature for object tracking optimization. Appl Soft Comput 2018. [DOI: 10.1016/j.asoc.2018.02.037] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
35
|
|
36
|
|
37
|
Tao J, Zhou D, Zhu B. Multi-source adaptation embedding with feature selection by exploiting correlation information. Knowl Based Syst 2018. [DOI: 10.1016/j.knosys.2017.12.016] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
38
|
Ma T, Hao Y, Suo X, Xue Y, Cao J. A weighted collaboration network generalization method for privacy protection in C-DBLP. INTELL DATA ANAL 2018. [DOI: 10.3233/ida-163482] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Affiliation(s)
- Tinghuai Ma
- CICAEET, Jiangsu Engineering Center of Network Monitoring, Nanjing University of Information Science & Technology, Nanjing 210044, Jiangsu, China
| | - Yu Hao
- School of Computer & Software, Nanjing University of Information Science & Technology, Jiangsu, Nanjing 210044, Jiangsu, China
| | - Xiafei Suo
- School of Computer & Software, Nanjing University of Information Science & Technology, Jiangsu, Nanjing 210044, Jiangsu, China
| | - Yu Xue
- School of Computer & Software, Nanjing University of Information Science & Technology, Jiangsu, Nanjing 210044, Jiangsu, China
| | - Jie Cao
- School of Economics & Management, Nanjing University of Information Science & Technology, Nanjing 210044, Jiangsu, China
| |
Collapse
|
39
|
Wang F, Yang H, Yang Y. Swarming movement of dynamical multi-agent systems with sampling control and time delays. Soft comput 2018. [DOI: 10.1007/s00500-018-3035-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
40
|
Shang R, Wang W, Stolkin R, Jiao L. Non-Negative Spectral Learning and Sparse Regression-Based Dual-Graph Regularized Feature Selection. IEEE TRANSACTIONS ON CYBERNETICS 2018; 48:793-806. [PMID: 28287996 DOI: 10.1109/tcyb.2017.2657007] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Feature selection is an important approach for reducing the dimension of high-dimensional data. In recent years, many feature selection algorithms have been proposed, but most of them only exploit information from the data space. They often neglect useful information contained in the feature space, and do not make full use of the characteristics of the data. To overcome this problem, this paper proposes a new unsupervised feature selection algorithm, called non-negative spectral learning and sparse regression-based dual-graph regularized feature selection (NSSRD). NSSRD is based on the feature selection framework of joint embedding learning and sparse regression, but extends this framework by introducing the feature graph. By using low dimensional embedding learning in both data space and feature space, NSSRD simultaneously exploits the geometric information of both spaces. Second, the algorithm uses non-negative constraints to constrain the low-dimensional embedding matrix of both feature space and data space, ensuring that the elements in the matrix are non-negative. Third, NSSRD unifies the embedding matrix of the feature space and the sparse transformation matrix. To ensure the sparsity of the feature array, the sparse transformation matrix is constrained using the -norm. Thus feature selection can obtain accurate discriminative information from these matrices. Finally, NSSRD uses an iterative and alternative updating rule to optimize the objective function, enabling it to select the representative features more quickly and efficiently. This paper explains the objective function, the iterative updating rules and a proof of convergence. Experimental results show that NSSRD is significantly more effective than several other feature selection algorithms from the literature, on a variety of test data.
Collapse
|
41
|
Kumar DA, Meher SK, Kumari KP. Fusion of progressive granular neural networks for pattern classification. Soft comput 2018. [DOI: 10.1007/s00500-018-3052-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
42
|
Three-way decisions approach to multiple attribute group decision making with linguistic information-based decision-theoretic rough fuzzy set. Int J Approx Reason 2018. [DOI: 10.1016/j.ijar.2017.11.015] [Citation(s) in RCA: 101] [Impact Index Per Article: 14.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
43
|
Zhang W, Hu H, Hu H, Fang J. Semantic distance between vague concepts in a framework of modeling with words. Soft comput 2018. [DOI: 10.1007/s00500-017-2992-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
44
|
Mishra S, Panda M. Bat Algorithm for Multilevel Colour Image Segmentation Using Entropy-Based Thresholding. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2018. [DOI: 10.1007/s13369-017-3017-x] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
45
|
|
46
|
|
47
|
Chen SB, Zhang Y, Ding CH, Zhou ZL, Luo B. A discriminative multi-class feature selection method via weighted l2,1-norm and Extended Elastic Net. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2017.09.055] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
48
|
Dynamic Gesture Recognition with a Terahertz Radar Based on Range Profile Sequences and Doppler Signatures. SENSORS 2017; 18:s18010010. [PMID: 29267249 PMCID: PMC5795606 DOI: 10.3390/s18010010] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/03/2017] [Revised: 11/15/2017] [Accepted: 12/14/2017] [Indexed: 11/18/2022]
Abstract
The frequency of terahertz radar ranges from 0.1 THz to 10 THz, which is higher than that of microwaves. Multi-modal signals, including high-resolution range profile (HRRP) and Doppler signatures, can be acquired by the terahertz radar system. These two kinds of information are commonly used in automatic target recognition; however, dynamic gesture recognition is rarely discussed in the terahertz regime. In this paper, a dynamic gesture recognition system using a terahertz radar is proposed, based on multi-modal signals. The HRRP sequences and Doppler signatures were first achieved from the radar echoes. Considering the electromagnetic scattering characteristics, a feature extraction model is designed using location parameter estimation of scattering centers. Dynamic Time Warping (DTW) extended to multi-modal signals is used to accomplish the classifications. Ten types of gesture signals, collected from a terahertz radar, are applied to validate the analysis and the recognition system. The results of the experiment indicate that the recognition rate reaches more than 91%. This research verifies the potential applications of dynamic gesture recognition using a terahertz radar.
Collapse
|
49
|
Unsupervised feature selection based on self-representation sparse regression and local similarity preserving. INT J MACH LEARN CYB 2017. [DOI: 10.1007/s13042-017-0760-y] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
50
|
Dhingra K, Yadav SK. Spam analysis of big reviews dataset using Fuzzy Ranking Evaluation Algorithm and Hadoop. INT J MACH LEARN CYB 2017. [DOI: 10.1007/s13042-017-0768-3] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|