1
|
Zhang Y, Shangguan C, Zhang X, Ma J, He J, Jia M, Chen N. Computer-Aided Diagnosis of Complications After Liver Transplantation Based on Transfer Learning. Interdiscip Sci 2024; 16:123-140. [PMID: 37875773 DOI: 10.1007/s12539-023-00588-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2023] [Revised: 09/20/2023] [Accepted: 09/22/2023] [Indexed: 10/26/2023]
Abstract
Liver transplantation is one of the most effective treatments for acute liver failure, cirrhosis, and even liver cancer. The prediction of postoperative complications is of great significance for liver transplantation. However, the existing prediction methods based on traditional machine learning are often unavailable or unreliable due to the insufficient amount of real liver transplantation data. Therefore, we propose a new framework to increase the accuracy of computer-aided diagnosis of complications after liver transplantation with transfer learning, which can handle small-scale but high-dimensional data problems. Furthermore, since data samples are often high dimensional in the real world, capturing key features that influence postoperative complications can help make the correct diagnosis for patients. So, we also introduce the SHapley Additive exPlanation (SHAP) method into our framework for exploring the key features of postoperative complications. We used data obtained from 425 patients with 456 features in our experiments. Experimental results show that our approach outperforms all compared baseline methods in predicting postoperative complications. In our work, the average precision, the mean recall, and the mean F1 score reach 91.22%, 91.70%, and 91.18%, respectively.
Collapse
Affiliation(s)
- Ying Zhang
- School of Control and Computer Engineering, North China Electric Power University, Beijing, 102206, China.
| | - Chenyuan Shangguan
- School of Control and Computer Engineering, North China Electric Power University, Beijing, 102206, China
| | - Xuena Zhang
- Department of Anesthesiology Medicine, Beijing Chaoyang Hospital, Capital Medical University, Beijing, 100069, China
| | - Jialin Ma
- Tianjin Zhuoman Technology Co., Ltd., Tianjin, 300000, China
| | - Jiyuan He
- School of Control and Computer Engineering, North China Electric Power University, Beijing, 102206, China
| | - Meng Jia
- School of Control and Computer Engineering, North China Electric Power University, Beijing, 102206, China
| | - Na Chen
- Hebei Vocational College of Rail Transportation, Shijiazhuang, 050051, China
| |
Collapse
|
2
|
Xue X, Zhang K, Tan KC, Feng L, Wang J, Chen G, Zhao X, Zhang L, Yao J. Affine Transformation-Enhanced Multifactorial Optimization for Heterogeneous Problems. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:6217-6231. [PMID: 33320820 DOI: 10.1109/tcyb.2020.3036393] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Evolutionary multitasking (EMT) is a newly emerging research topic in the community of evolutionary computation, which aims to improve the convergence characteristic across multiple distinct optimization tasks simultaneously by triggering knowledge transfer among them. Unfortunately, most of the existing EMT algorithms are only capable of boosting the optimization performance for homogeneous problems which explicitly share the same (or similar) fitness landscapes. Seldom efforts have been devoted to generalize the EMT for solving heterogeneous problems. A few preliminary studies employ domain adaptation techniques to enhance the transferability between two distinct tasks. However, almost all of these methods encounter a severe issue which is the so-called degradation of intertask mapping. Keeping this in mind, a novel rank loss function for acquiring a superior intertask mapping is proposed in this article. In particular, with an evolutionary-path-based representation model for optimization instance, an analytical solution of affine transformation for bridging the gap between two distinct problems is mathematically derived from the proposed rank loss function. It is worth mentioning that the proposed mapping-based transferability enhancement technique can be seamlessly embedded into an EMT paradigm. Finally, the efficacy of our proposed method against several state-of-the-art EMTs is verified experimentally on a number of synthetic multitasking and many-tasking benchmark problems, as well as a practical case study.
Collapse
|
3
|
Wang H, Zhou Z. Rough margin-based ν-twin support tensor machine in pattern recognition. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-200573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
In Rough margin-based ν-Twin Support Vector Machine (Rν-TSVM) algorithm, the rough theory is introduced. Rν-TSVM gives different penalties to the corresponding misclassified samples according to their positions, so it avoids the overfitting problem to some extent. While the input data is a tensor, Rν-TSVM cannot handle it directly and may not utilize the data information effectively. Therefore, we propose a novel classifier based on tensor data, termed as Rough margin-based ν-Twin Support Tensor Machine (Rν-TSTM). Similar to Rν-TSVM, Rν-TSTM constructs rough lower margin, rough upper margin and rough boundary in tensor space. Rν-TSTM not only retains the superiority of Rν-TSVM, but also has its unique advantages. Firstly, the data topology is retained more efficiently by the direct use of tensor representation. Secondly, it has better classification performance compared to other classification algorithms. Thirdly, it can avoid overfitting problem to a great extent. Lastly, it is more suitable for high dimensional and small sample size problem. To solve the corresponding optimization problem in Rν-TSTM, we adopt the alternating iteration method in which the parameters corresponding to the hyperplanes are estimated by solving a series of Rν-TSVM optimization problem. The efficiency and superiority of the proposed method are demonstrated by computational experiments.
Collapse
Affiliation(s)
- Huiru Wang
- College of Science, Beijing Forestry University, Haidian, Beijing, China
| | - Zhijian Zhou
- College of Science, China Agricultural University, Haidian, Beijing, China
| |
Collapse
|
4
|
Verma M, Shukla K. Convergence analysis of accelerated proximal extra-gradient method with applications. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.01.049] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
5
|
|
6
|
Kim J, Bukhari W, Lee M. Feature Analysis of Unsupervised Learning for Multi-task Classification Using Convolutional Neural Network. Neural Process Lett 2017. [DOI: 10.1007/s11063-017-9724-1] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
7
|
Jiang M, Huang W, Huang Z, Yen GG. Integration of Global and Local Metrics for Domain Adaptation Learning Via Dimensionality Reduction. IEEE TRANSACTIONS ON CYBERNETICS 2017; 47:38-51. [PMID: 26672056 DOI: 10.1109/tcyb.2015.2502483] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Domain adaptation learning (DAL) investigates how to perform a task across different domains. In this paper, we present a kernelized local-global approach to solve domain adaptation problems. The basic idea of the proposed method is to consider the global and local information regarding the domains (e.g., maximum mean discrepancy and intraclass distance) and to convert the domain adaptation problem into a bi-object optimization problem via the kernel method. A solution for the optimization problem will help us identify a latent space in which the distributions of the different domains will be close to each other in the global sense, and the local properties of the labeled source samples will be preserved. Therefore, classic classification algorithms can be used to recognize unlabeled target domain data, which has a significant difference on the source samples. Based on the analysis, we validate the proposed algorithm using four different sources of data: synthetic, textual, object, and facial image. The experimental results indicate that the proposed method provides a reasonable means to improve DAL algorithms.
Collapse
|
8
|
Weighted multifeature hyperspectral image classification via kernel joint sparse representation. Neurocomputing 2016. [DOI: 10.1016/j.neucom.2015.07.114] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
9
|
|
10
|
He X, Mourot G, Maquin D, Ragot J, Beauseroy P, Smolarz A, Grall-Maës E. Multi-task learning with one-class SVM. Neurocomputing 2014. [DOI: 10.1016/j.neucom.2013.12.022] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|
11
|
|
12
|
Roy A, Mackin PD, Mukhopadhyay S. Methods for pattern selection, class-specific feature selection and classification for automated learning. Neural Netw 2013; 41:113-29. [DOI: 10.1016/j.neunet.2012.12.007] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2012] [Revised: 10/15/2012] [Accepted: 12/17/2012] [Indexed: 11/28/2022]
|
13
|
Han F, Zhu JS. Improved Particle Swarm Optimization Combined with Backpropagation for Feedforward Neural Networks. INT J INTELL SYST 2012. [DOI: 10.1002/int.21569] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Affiliation(s)
- Fei Han
- School of Computer Science and Telecommunication Engineering; Jiangsu University; Zhenjiang, Jiangsu; People's Republic of China
| | - Jian-Sheng Zhu
- School of Computer Science and Telecommunication Engineering; Jiangsu University; Zhenjiang, Jiangsu; People's Republic of China
| |
Collapse
|
14
|
Yuan XT, Liu X, Yan S. Visual classification with multitask joint sparse representation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2012; 21:4349-4360. [PMID: 22736645 DOI: 10.1109/tip.2012.2205006] [Citation(s) in RCA: 58] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
We address the problem of visual classification with multiple features and/or multiple instances. Motivated by the recent success of multitask joint covariate selection, we formulate this problem as a multitask joint sparse representation model to combine the strength of multiple features and/or instances for recognition. A joint sparsity-inducing norm is utilized to enforce class-level joint sparsity patterns among the multiple representation vectors. The proposed model can be efficiently optimized by a proximal gradient method. Furthermore, we extend our method to the setup where features are described in kernel matrices. We then investigate into two applications of our method to visual classification: 1) fusing multiple kernel features for object categorization and 2) robust face recognition in video with an ensemble of query images. Extensive experiments on challenging real-world data sets demonstrate that the proposed method is competitive to the state-of-the-art methods in respective applications.
Collapse
Affiliation(s)
- Xiao-Tong Yuan
- Department of Statistics, Rutgers University, Newark, NJ 08854, USA.
| | | | | |
Collapse
|
15
|
Jiang H, Zhang B. Dynamical memory control based on projection technique for online regression. Soft comput 2012. [DOI: 10.1007/s00500-012-0929-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
16
|
|
17
|
Performance evaluation of multilayer perceptrons for discriminating and quantifying multiple kinds of odors with an electronic nose. Neural Netw 2012; 33:204-15. [PMID: 22717447 DOI: 10.1016/j.neunet.2012.05.009] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2012] [Revised: 05/14/2012] [Accepted: 05/23/2012] [Indexed: 11/21/2022]
Abstract
This paper studies several types and arrangements of perceptron modules to discriminate and quantify multiple odors with an electronic nose. We evaluate the following types of multilayer perceptron. (A) A single multi-output (SMO) perceptron both for discrimination and for quantification. (B) An SMO perceptron for discrimination followed by multiple multi-output (MMO) perceptrons for quantification. (C) An SMO perceptron for discrimination followed by multiple single-output (MSO) perceptrons for quantification. (D) MSO perceptrons for discrimination followed by MSO perceptrons for quantification, called the MSO-MSO perceptron model, under the following conditions: (D1) using a simple one-against-all (OAA) decomposition method; (D2) adopting a simple OAA decomposition method and virtual balance step; and (D3) employing a local OAA decomposition method, virtual balance step and local generalization strategy all together. The experimental results for 12 kinds of volatile organic compounds at 85 concentration levels in the training set and 155 concentration levels in the test set show that the MSO-MSO perceptron model with the D3 learning procedure is the most effective of those tested for discrimination and quantification of many kinds of odors.
Collapse
|
18
|
Rakotomamonjy A, Flamary R, Gasso G, Canu S. lp-lq penalty for sparse linear and sparse multiple kernel multitask learning. ACTA ACUST UNITED AC 2012; 22:1307-20. [PMID: 21813358 DOI: 10.1109/tnn.2011.2157521] [Citation(s) in RCA: 60] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Recently, there has been much interest around multitask learning (MTL) problem with the constraints that tasks should share a common sparsity profile. Such a problem can be addressed through a regularization framework where the regularizer induces a joint-sparsity pattern between task decision functions. We follow this principled framework and focus on l(p)-l(q) (with 0 ≤ p ≤ 1 and 1 ≤ q ≤ 2) mixed norms as sparsity-inducing penalties. Our motivation for addressing such a larger class of penalty is to adapt the penalty to a problem at hand leading thus to better performances and better sparsity pattern. For solving the problem in the general multiple kernel case, we first derive a variational formulation of the l(1)-l(q) penalty which helps us in proposing an alternate optimization algorithm. Although very simple, the latter algorithm provably converges to the global minimum of the l(1)-l(q) penalized problem. For the linear case, we extend existing works considering accelerated proximal gradient to this penalty. Our contribution in this context is to provide an efficient scheme for computing the l(1)-l(q) proximal operator. Then, for the more general case, when , we solve the resulting nonconvex problem through a majorization-minimization approach. The resulting algorithm is an iterative scheme which, at each iteration, solves a weighted l(1)-l(q) sparse MTL problem. Empirical evidences from toy dataset and real-word datasets dealing with brain-computer interface single-trial electroencephalogram classification and protein subcellular localization show the benefit of the proposed approaches and algorithms.
Collapse
|
19
|
Duan L, Xu D, Tsang IWH. Domain adaptation from multiple sources: a domain-dependent regularization approach. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2012; 23:504-518. [PMID: 24808555 DOI: 10.1109/tnnls.2011.2178556] [Citation(s) in RCA: 110] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
In this paper, we propose a new framework called domain adaptation machine (DAM) for the multiple source domain adaption problem. Under this framework, we learn a robust decision function (referred to as target classifier) for label prediction of instances from the target domain by leveraging a set of base classifiers which are prelearned by using labeled instances either from the source domains or from the source domains and the target domain. With the base classifiers, we propose a new domain-dependent regularizer based on smoothness assumption, which enforces that the target classifier shares similar decision values with the relevant base classifiers on the unlabeled instances from the target domain. This newly proposed regularizer can be readily incorporated into many kernel methods (e.g., support vector machines (SVM), support vector regression, and least-squares SVM (LS-SVM)). For domain adaptation, we also develop two new domain adaptation methods referred to as FastDAM and UniverDAM. In FastDAM, we introduce our proposed domain-dependent regularizer into LS-SVM as well as employ a sparsity regularizer to learn a sparse target classifier with the support vectors only from the target domain, which thus makes the label prediction on any test instance very fast. In UniverDAM, we additionally make use of the instances from the source domains as Universum to further enhance the generalization ability of the target classifier. We evaluate our two methods on the challenging TRECIVD 2005 dataset for the large-scale video concept detection task as well as on the 20 newsgroups and email spam datasets for document retrieval. Comprehensive experiments demonstrate that FastDAM and UniverDAM outperform the existing multiple source domain adaptation methods for the two applications.
Collapse
|
20
|
Jeong S, Lee M. Adaptive object recognition model using incremental feature representation and hierarchical classification. Neural Netw 2012; 25:130-40. [DOI: 10.1016/j.neunet.2011.06.020] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2011] [Revised: 06/09/2011] [Accepted: 06/27/2011] [Indexed: 11/17/2022]
|
21
|
Nishikawa H, Ozawa S. Radial Basis Function Network for Multitask Pattern Recognition. Neural Process Lett 2011. [DOI: 10.1007/s11063-011-9178-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
22
|
Pan SJ, Tsang IW, Kwok JT, Yang Q. Domain Adaptation via Transfer Component Analysis. ACTA ACUST UNITED AC 2011; 22:199-210. [DOI: 10.1109/tnn.2010.2091281] [Citation(s) in RCA: 2119] [Impact Index Per Article: 151.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
23
|
Vassilieva E, Pinto G, Acacio de Barros J, Suppes P. Learning Pattern Recognition Through Quasi-Synchronization of Phase Oscillators. ACTA ACUST UNITED AC 2011; 22:84-95. [DOI: 10.1109/tnn.2010.2086476] [Citation(s) in RCA: 36] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
24
|
Hisada M, Ozawa S, Zhang K, Kasabov N. Incremental linear discriminant analysis for evolving feature spaces in multitask pattern recognition problems. EVOLVING SYSTEMS 2010. [DOI: 10.1007/s12530-010-9000-3] [Citation(s) in RCA: 24] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|