1
|
Qi F, Guo J, Li J, Liao Y, Liao W, Cai H, Chen J. Multi-kernel clustering with tensor fusion on Grassmann manifold for high-dimensional genomic data. Methods 2024; 231:215-225. [PMID: 39396747 DOI: 10.1016/j.ymeth.2024.09.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 09/21/2024] [Accepted: 09/26/2024] [Indexed: 10/15/2024] Open
Abstract
The high dimensionality and noise challenges in genomic data make it difficult for traditional clustering methods. Existing multi-kernel clustering methods aim to improve the quality of the affinity matrix by learning a set of base kernels, thereby enhancing clustering performance. However, directly learning from the original base kernels presents challenges in handling errors and redundancies when dealing with high-dimensional data, and there is still a lack of feasible multi-kernel fusion strategies. To address these issues, we propose a Multi-Kernel Clustering method with Tensor fusion on Grassmann manifolds, called MKCTM. Specifically, we maximize the clustering consensus among base kernels by imposing tensor low-rank constraints to eliminate noise and redundancy. Unlike traditional kernel fusion approaches, our method fuses learned base kernels on the Grassmann manifold, resulting in a final consensus matrix for clustering. We integrate tensor learning and fusion processes into a unified optimization model and propose an effective iterative optimization algorithm for solving it. Experimental results on ten datasets, comparing against 12 popular baseline clustering methods, confirm the superiority of our approach. Our code is available at https://github.com/foureverfei/MKCTM.git.
Collapse
Affiliation(s)
- Fei Qi
- Data Science and Information Engineering, Guizhou Minzu University, Guiyang, 550025, Guizhou, China; Computer Science and Engineering, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Jin Guo
- Big Data and Information Engineering, Guiyang Institute of Humanities and Technology, Guiyang, 550025, Guizhou, China
| | - Junyu Li
- Computer Science and Engineering, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Yi Liao
- Computer Science and Engineering, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Wenxiong Liao
- Computer Science and Engineering, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Hongmin Cai
- Computer Science and Engineering, South China University of Technology, Guangzhou, 510006, Guangdong, China
| | - Jiazhou Chen
- Computer Science and Technology, Guangdong University of Technology, Guangzhou, 510006, Guangdong, China.
| |
Collapse
|
2
|
Wu Y, Shen XJ, Abhadiomhen SE, Yang Y, Gu JN. Kernel ensemble support vector machine with integrated loss in shared parameters space. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 82:18077-18096. [DOI: 10.1007/s11042-022-14226-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/20/2021] [Revised: 02/18/2022] [Accepted: 11/04/2022] [Indexed: 12/04/2024]
|
3
|
Rahimzadeh Arashloo S. One-Class Classification Using ℓp-Norm Multiple Kernel Fisher Null Approach. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2023; 32:1843-1856. [PMID: 37028349 DOI: 10.1109/tip.2023.3255102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
We address the one-class classification (OCC) problem and advocate a one-class MKL (multiple kernel learning) approach for this purpose. To this aim, based on the Fisher null-space OCC principle, we present a multiple kernel learning algorithm where an $\ell _{p}$ -norm regularisation ( $p \geq 1$ ) is considered for kernel weight learning. We cast the proposed one-class MKL problem as a min-max saddle point Lagrangian optimisation task and propose an efficient approach to optimise it. An extension of the proposed approach is also considered where several related one-class MKL tasks are learned concurrently by constraining them to share common weights for kernels. An extensive evaluation of the proposed MKL approach on a range of data sets from different application domains confirms its merits against the baseline and several other algorithms.
Collapse
|
4
|
Gupta A, Das S. Transfer Clustering Using a Multiple Kernel Metric Learned Under Multi-Instance Weak Supervision. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE 2022. [DOI: 10.1109/tetci.2021.3110526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Affiliation(s)
- Avisek Gupta
- Electronics and Communication Sciences Unit, Indian Statistical Institute, Kolkata, India
| | - Swagatam Das
- Electronics and Communication Sciences Unit, Indian Statistical Institute, Kolkata, India
| |
Collapse
|
5
|
Distance metric learning with local multiple kernel embedding. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-021-01487-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
6
|
Alioscha-Perez M, Oveneke MC, Sahli H. SVRG-MKL: A Fast and Scalable Multiple Kernel Learning Solution for Features Combination in Multi-Class Classification Problems. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2020; 31:1710-1723. [PMID: 31283489 DOI: 10.1109/tnnls.2019.2922123] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
In this paper, we present a novel strategy to combine a set of compact descriptors to leverage an associated recognition task. We formulate the problem from a multiple kernel learning (MKL) perspective and solve it following a stochastic variance reduced gradient (SVRG) approach to address its scalability, currently an open issue. MKL models are ideal candidates to jointly learn the optimal combination of features along with its associated predictor. However, they are unable to scale beyond a dozen thousand of samples due to high computational and memory requirements, which severely limits their applicability. We propose SVRG-MKL, an MKL solution with inherent scalability properties that can optimally combine multiple descriptors involving millions of samples. Our solution takes place directly in the primal to avoid Gram matrices computation and memory allocation, whereas the optimization is performed with a proposed algorithm of linear complexity and hence computationally efficient. Our proposition builds upon recent progress in SVRG with the distinction that each kernel is treated differently during optimization, which results in a faster convergence than applying off-the-shelf SVRG into MKL. Extensive experimental validation conducted on several benchmarking data sets confirms a higher accuracy and a significant speedup of our solution. Our technique can be extended to other MKL problems, including visual search and transfer learning, as well as other formulations, such as group-sensitive (GMKL) and localized MKL (LMKL) in convex settings.
Collapse
|
7
|
|