1
|
Rocchetta R, Mey A, Oliehoek FA. A Survey on Scenario Theory, Complexity, and Compression-Based Learning and Generalization. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:16985-16999. [PMID: 37703153 DOI: 10.1109/tnnls.2023.3308828] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/15/2023]
Abstract
This work investigates formal generalization error bounds that apply to support vector machines (SVMs) in realizable and agnostic learning problems. We focus on recently observed parallels between probably approximately correct (PAC)-learning bounds, such as compression and complexity-based bounds, and novel error guarantees derived within scenario theory. Scenario theory provides nonasymptotic and distributional-free error bounds for models trained by solving data-driven decision-making problems. Relevant theorems and assumptions are reviewed and discussed. We propose a numerical comparison of the tightness and effectiveness of theoretical error bounds for support vector classifiers trained on several randomized experiments from 13 real-life problems. This analysis allows for a fair comparison of different approaches from both conceptual and experimental standpoints. Based on the numerical results, we argue that the error guarantees derived from scenario theory are often tighter for realizable problems and always yield informative results, i.e., probability bounds tighter than a vacuous [0, 1] interval. This work promotes scenario theory as an alternative tool for model selection, structural-risk minimization, and generalization error analysis of SVMs. In this way, we hope to bring the communities of scenario and statistical learning theory closer, so that they can benefit from each other's insights.
Collapse
|
2
|
Wang H, Zhu J, Zhang S. Safe screening rules for multi-view support vector machines. Neural Netw 2023; 166:326-343. [PMID: 37541164 DOI: 10.1016/j.neunet.2023.07.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2022] [Revised: 07/11/2023] [Accepted: 07/12/2023] [Indexed: 08/06/2023]
Abstract
Multi-view learning aims to make use of the advantages of different views to complement each other and fully mines the potential information in the data. However, the complexity of multi-view learning algorithm is much higher than that of single view learning algorithm. Based on the optimality conditions of two classical multi-view models: SVM-2K and multi-view twin support vector machine (MvTwSVM), this paper analyzes the corresponding relationship between dual variables and samples, and derives their safe screening rules for the first time, termed as SSR-SVM-2K and SSR-MvTwSVM. It can assign or delete four groups of different dual variables in advance before solving the optimization problem, so as to greatly reduce the scale of the optimization problem and improve the solution speed. More importantly, the safe screening criterion is "safe", that is, the solution of the reduced optimization problem is the same as that of the original problem before screening. In addition, we further give a sequence screening rule to speed up the parameter optimization process, and analyze its properties, including the similarities and differences of safe screening rules between multi-view SVMs and single-view SVMs, the computational complexity, and the relationship between the parameter interval and screening rate. Numerical experiments verify the effectiveness of the proposed methods.
Collapse
Affiliation(s)
- Huiru Wang
- Department of Mathematics, College of Science, Beijing Forestry University, No. 35 Qinghua East Road, 100083 Haidian, Beijing, China.
| | - Jiayi Zhu
- School of Computer Science and Engineering and Guangdong Province Key Laboratory of Computational Science, Sun Yat-Sen University, Guangzhou, Guangdong 510006, China
| | - Siyuan Zhang
- College of Information and Electrical Engineering, China Agricultural University, No. 17 Qinghua East Road, 100083 Haidian, Beijing, China
| |
Collapse
|
3
|
Tang L, Tian Y, Wang X, Pardalos PM. A simple and reliable instance selection for fast training support vector machine: Valid Border Recognition. Neural Netw 2023; 166:379-395. [PMID: 37549607 DOI: 10.1016/j.neunet.2023.07.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2023] [Revised: 07/11/2023] [Accepted: 07/12/2023] [Indexed: 08/09/2023]
Abstract
Support vector machines (SVMs) are powerful statistical learning tools, but their application to large datasets can cause time-consuming training complexity. To address this issue, various instance selection (IS) approaches have been proposed, which choose a small fraction of critical instances and screen out others before training. However, existing methods have not been able to balance accuracy and efficiency well. Some methods miss critical instances, while others use complicated selection schemes that require even more execution time than training with all original instances, thus violating the initial intention of IS. In this work, we present a newly developed IS method called Valid Border Recognition (VBR). VBR selects the closest heterogeneous neighbors as valid border instances and incorporates this process into the creation of a reduced Gaussian kernel matrix, thus minimizing the execution time. To improve reliability, we propose a strengthened version of VBR (SVBR). Based on VBR, SVBR gradually adds farther heterogeneous neighbors as complements until the Lagrange multipliers of already selected instances become stable. In numerical experiments, the effectiveness of our proposed methods is verified on benchmark and synthetic datasets in terms of accuracy, execution time and inference time.
Collapse
Affiliation(s)
- Long Tang
- School of Artificial Intelligence, Nanjing University of Information Science & Technology, Nanjing, 210044, China; Research Institute of Talent Big Data, Nanjing University of Information Science & Technology, Nanjing, 210044, China.
| | - Yingjie Tian
- Research Center on Fictitious Economy and Data Science, Chinese Academy of Sciences, Beijing 100190, China
| | - Xiaowei Wang
- Research Institute of Talent Big Data, Nanjing University of Information Science & Technology, Nanjing, 210044, China
| | - Panos M Pardalos
- Center for Applied Optimization, Department of Industrial and Systems Engineering, University of Florida, Gainesville, 32611, USA
| |
Collapse
|
4
|
Wang H, Zhu J, Feng F. Elastic net twin support vector machine and its safe screening rules. Inf Sci (N Y) 2023. [DOI: 10.1016/j.ins.2023.03.131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/29/2023]
|
5
|
Qi K, Yang H. Elastic Net Nonparallel Hyperplane Support Vector Machine and Its Geometrical Rationality. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:7199-7209. [PMID: 34097622 DOI: 10.1109/tnnls.2021.3084404] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Twin support vector machine (TWSVM), which constructs two nonparallel classifying hyperplanes, is widely applied to various fields. However, TWSVM solves two quadratic programming problems (QPPs) separately such that the final classifiers lack consistency and enough prediction accuracy. Moreover, by reason of only considering the 1-norm penalty for slack variables, TWSVM is not well defined in the geometrical view. In this article, we propose a novel elastic net nonparallel hyperplane support vector machine (ENNHSVM), which adopts elastic net penalty for slack variables and constructs two nonparallel separating hyperplanes simultaneously. We further discuss the properties of ENNHSVM theoretically and derive the violation tolerance upper bound to better demonstrate the relative violations of training samples in the same class. In particular, we design a safe screening rule for ENNHSVM to speed up the calculations. We finally compare the performance of ENNHSVM on both synthetic datasets and benchmark datasets with the Lagrangian SVM, the twin parametric-margin SVM, the elastic net SVM, the TWSVM, and the nonparallel hyperplane SVM.
Collapse
|
6
|
Pan X, Xu Y. A Safe Feature Elimination Rule for L 1-Regularized Logistic Regression. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2022; 44:4544-4554. [PMID: 33822720 DOI: 10.1109/tpami.2021.3071138] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The L1-regularized logistic regression (L1-LR) is popular for classification problems. To accelerate its training speed for high-dimensional data, techniques named safe screening rules have been proposed recently. They can safely delete the inactive features in data so as to greatly reduce the training cost of L1-LR. The screening power of these rules is determined by their corresponding safe regions, which is also the core technique of safe screening rules. In this paper, we introduce a new safe feature elimination rule (SFER) for L1-LR. Compared to existing safe rules, the safe region of SFER is improved in two aspects: (1) a smaller sphere region is constructed by using the strong convexity of dual L1-LR twice; (2) multiple half-spaces, which correspond to the potential active constraints, are added for further contraction. Both improvements can enhance the screening ability of SFER. As for the complexity of SFER, an iterative filtering framework is given by decomposing the safe region into multiple "domes". In this way, SFER admits a closed form solution and the identified features will not be scanned repeatedly. Experiments on ten benchmark data sets demonstrate that SFER gives superior performance than existing methods on training efficiency.
Collapse
|
7
|
Fan Y, Zhao J. Safe sample screening rules for multicategory angle-based support vector machines. Comput Stat Data Anal 2022. [DOI: 10.1016/j.csda.2022.107508] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/03/2022]
|
8
|
Zhou K, Zhang Q, Li J. TSVMPath: Fast Regularization Parameter Tuning Algorithm for Twin Support Vector Machine. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-10870-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
9
|
Pang X, Zhao J, Xu Y. A novel ramp loss-based multi-task twin support vector machine with multi-parameter safe acceleration. Neural Netw 2022; 150:194-212. [DOI: 10.1016/j.neunet.2022.03.006] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2021] [Revised: 02/13/2022] [Accepted: 03/03/2022] [Indexed: 10/18/2022]
|
10
|
Xiao X, Xu Y, Zhang Y, Zhong P. A novel self-weighted Lasso and its safe screening rule. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03316-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
11
|
Xie F, Xu Y, Ma M, Pang X. A safe acceleration method for multi-task twin support vector machine. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-021-01481-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
12
|
|
13
|
Zhao J, Xu Y, Xu C, Wang T. A two-stage safe screening method for non-convex support vector machine with ramp loss. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.107250] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
14
|
Xie F, Pang X, Xu Y. Pinball loss-based multi-task twin support vector machine and its safe acceleration method. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06173-6] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
15
|
Multi-parameter safe screening rule for hinge-optimal margin distribution machine. APPL INTELL 2021. [DOI: 10.1007/s10489-020-02024-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
16
|
Wu W, Xu Y, Pang X. A hybrid acceleration strategy for nonparallel support vector machine. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2020.08.067] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
17
|
|
18
|
Wang X, Yang Y, Xu Y, Chen Q, Wang H, Gao H. Predicting hypoglycemic drugs of type 2 diabetes based on weighted rank support vector machine. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2020.105868] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
|
19
|
Multi-variable estimation-based safe screening rule for small sphere and large margin support vector machine. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2019.105223] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
20
|
Shang R, Xu K, Shang F, Jiao L. Sparse and low-redundant subspace learning-based dual-graph regularized robust feature selection. Knowl Based Syst 2020. [DOI: 10.1016/j.knosys.2019.07.001] [Citation(s) in RCA: 30] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
21
|
Abhishek, Verma S. Optimal manifold neighborhood and kernel width for robust non-linear dimensionality reduction. Knowl Based Syst 2019. [DOI: 10.1016/j.knosys.2019.104953] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
22
|
|
23
|
|
24
|
Pan X, Xu Y. A Novel and Safe Two-Stage Screening Method for Support Vector Machine. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2019; 30:2263-2274. [PMID: 30507539 DOI: 10.1109/tnnls.2018.2879800] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
To make support vector machine (SVM) applicable to large-scale data sets, safe screening rules are developed recently. The main idea is to reduce the scale of SVM by safely discarding the redundant training samples. Among existing safe screening rules, the dual screening method with variational inequalities (DVI) and the dynamic screening rule (DSR) based on duality gap are two representative strategies. DVI is efficient, while its safety may be affected by inaccurate solving algorithms. DSR is guaranteed to be safe; however, accurate feasible solutions are required for good efficiency. Based on the above-mentioned studies, in this paper, a two-stage screening (TSS) rule, which fully exploits the advantages of the above-mentioned two approaches and improves their shortcomings, is proposed. First, DVI is applied prior to training for sample screening. It reduces the scale of SVM and, meanwhile, produces a better initial feasible solution for DSR. Then, by embedding DSR into the solving algorithm, the solver becomes more accurate, and the safety of DVI can be strengthened. In the end, for safety guarantee, a postchecking step is added to search the wrongly identified samples and retrain them. To theoretically analyze the safety of DVI, an upper bound of the deviation in DVI is estimated, and a Safe-DVI is given based on it. To ensure the efficiency of TSS, the superiority of DVI over initial DSR is verified. In addition, kernel version of TSS is also given for nonlinear SVM. Numerical experiments on synthetic data sets and 12 real-world data sets verify the efficiency and safety of this TSS.
Collapse
|
25
|
An improved non-parallel Universum support vector machine and its safe sample screening rule. Knowl Based Syst 2019. [DOI: 10.1016/j.knosys.2019.01.031] [Citation(s) in RCA: 47] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022]
|
26
|
Wu W, Xu Y. Accelerating improved twin support vector machine with safe screening rule. INT J MACH LEARN CYB 2019. [DOI: 10.1007/s13042-019-00946-1] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
27
|
Pan X, Xu Y. A safe reinforced feature screening strategy for lasso based on feasible solutions. Inf Sci (N Y) 2019. [DOI: 10.1016/j.ins.2018.10.031] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
28
|
Xu Y, Tian Y, Pan X, Wang H. E-ENDPP: a safe feature selection rule for speeding up Elastic Net. APPL INTELL 2019. [DOI: 10.1007/s10489-018-1295-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
|
29
|
|
30
|
Yang Z, Pan X, Xu Y. Piecewise linear solution path for pinball twin support vector machine. Knowl Based Syst 2018. [DOI: 10.1016/j.knosys.2018.07.022] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
|
31
|
An accelerator for support vector machines based on the local geometrical information and data partition. INT J MACH LEARN CYB 2018. [DOI: 10.1007/s13042-018-0877-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
32
|
|