1
|
Sugasawa S, Kobayashi G. Robust fitting of mixture models using weighted complete estimating equations. Comput Stat Data Anal 2022. [DOI: 10.1016/j.csda.2022.107526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/03/2022]
|
2
|
Sepahdar A, Madadi M, Balakrishnan N, Jamalizadeh A. Parsimonious mixture‐of‐experts based on mean mixture of multivariate normal distributions. Stat (Int Stat Inst) 2022. [DOI: 10.1002/sta4.421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Afsaneh Sepahdar
- Department of Statistics, Faculty of Mathematics and Computer Shahid Bahonar University of Kerman Kerman 76169‐14111 Iran
| | - Mohsen Madadi
- Department of Statistics, Faculty of Mathematics and Computer Shahid Bahonar University of Kerman Kerman 76169‐14111 Iran
| | | | - Ahad Jamalizadeh
- Department of Statistics, Faculty of Mathematics and Computer Shahid Bahonar University of Kerman Kerman 76169‐14111 Iran
| |
Collapse
|
3
|
Wang X, Feng Z. Component selection for exponential power mixture models. J Appl Stat 2021; 50:291-314. [PMID: 36698546 PMCID: PMC9870023 DOI: 10.1080/02664763.2021.1990225] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
Exponential Power (EP) family is a much flexible distribution family including Gaussian family as a sub-family. In this article, we study component selection and estimation for EP mixture models and regressions. The assumption on zero component mean in [X. Cao, Q. Zhao, D. Meng, Y. Chen, and Z. Xu, Robust low-rank matrix factorization under general mixture noise distributions, IEEE. Trans. Image. Process. 25 (2016), pp. 4677-4690.] is relaxed. To select components and estimate parameters simultaneously, we propose a penalized likelihood method, which can shrink mixing proportions to zero to achieve components selection. Modified EM algorithms are proposed, and the consistency of estimated component number is obtained. Simulation studies show the advantages of the proposed methods on accuracies of component number selection, parameter estimation, and density estimation. Analysis of value at risk of SHIBOR and a climate change data are given as illustration.
Collapse
Affiliation(s)
- Xinyi Wang
- The Wang Yanan Institute for Studies in Economics, Xiamen University, Xiamen, People's Republic of China
| | - Zhenghui Feng
- The Wang Yanan Institute for Studies in Economics, Xiamen University, Xiamen, People's Republic of China,MOE Key Laboratory of Econometrics, Department of Statistics, School of Economics, Xiamen University, Xiamen, People's Republic of China,Zhenghui Feng The Wang Yanan Institute for Studies in Economics, Xiamen University, Xiamen 361005, People's Republic of China MOE Key Laboratory of Econometrics, Department of Statistics, School of Economics, Xiamen University, Xiamen361005, People's Republic of China
| |
Collapse
|
4
|
Nguyen HD, Nguyen T, Chamroukhi F, McLachlan GJ. Approximations of conditional probability density functions in Lebesgue spaces via mixture of experts models. JOURNAL OF STATISTICAL DISTRIBUTIONS AND APPLICATIONS 2021. [DOI: 10.1186/s40488-021-00125-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
AbstractMixture of experts (MoE) models are widely applied for conditional probability density estimation problems. We demonstrate the richness of the class of MoE models by proving denseness results in Lebesgue spaces, when inputs and outputs variables are both compactly supported. We further prove an almost uniform convergence result when the input is univariate. Auxiliary lemmas are proved regarding the richness of the soft-max gating function class, and their relationships to the class of Gaussian gating functions.
Collapse
|
5
|
Mixture of linear experts model for censored data: A novel approach with scale-mixture of normal distributions. Comput Stat Data Anal 2021. [DOI: 10.1016/j.csda.2021.107182] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
6
|
Sugasawa S. Grouped Heterogeneous Mixture Modeling for Clustered Data. J Am Stat Assoc 2021. [DOI: 10.1080/01621459.2020.1777136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Affiliation(s)
- Shonosuke Sugasawa
- Center for Spatial Information Science, The University of Tokyo, Kashiwa, Japan
| |
Collapse
|
7
|
Nguyen HD, Chamroukhi F, Forbes F. Approximation results regarding the multiple-output Gaussian gated mixture of linear experts model. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.08.014] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
8
|
Yin J, Wu L, Lu H, Dai L. New estimation in mixture of experts models using the Pearson type VII distribution. COMMUN STAT-SIMUL C 2019. [DOI: 10.1080/03610918.2018.1485943] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Affiliation(s)
- Junhui Yin
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Liucang Wu
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Hanchi Lu
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Lin Dai
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| |
Collapse
|
9
|
Dai L, Yin J, Xie Z, Wu L. Robust variable selection in finite mixture of regression models using the t distribution. COMMUN STAT-THEOR M 2018. [DOI: 10.1080/03610926.2018.1513143] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Affiliation(s)
- Lin Dai
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Junhui Yin
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Zhengfen Xie
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| | - Liucang Wu
- Faculty of Science, Kunming University of Science and Technology, Kunming, P. R. China
| |
Collapse
|
10
|
Lloyd-Jones LR, Nguyen HD, McLachlan GJ. A globally convergent algorithm for lasso-penalized mixture of linear regression models. Comput Stat Data Anal 2018. [DOI: 10.1016/j.csda.2017.09.003] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
11
|
|
12
|
Nguyen HD, Lloyd-Jones LR, McLachlan GJ. A Universal Approximation Theorem for Mixture-of-Experts Models. Neural Comput 2016; 28:2585-2593. [DOI: 10.1162/neco_a_00892] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
The mixture-of-experts (MoE) model is a popular neural network architecture for nonlinear regression and classification. The class of MoE mean functions is known to be uniformly convergent to any unknown target function, assuming that the target function is from a Sobolev space that is sufficiently differentiable and that the domain of estimation is a compact unit hypercube. We provide an alternative result, which shows that the class of MoE mean functions is dense in the class of all continuous functions over arbitrary compact domains of estimation. Our result can be viewed as a universal approximation theorem for MoE models. The theorem we present allows MoE users to be confident in applying such models for estimation when data arise from nonlinear and nondifferentiable generative processes.
Collapse
Affiliation(s)
- Hien D. Nguyen
- School of Mathematics and Physics, University of Queensland, Brisbane, Queensland 4072, Australia
| | - Luke R. Lloyd-Jones
- Centre for Neurogenetics and Statistical Genetics, Queensland Brain Institute, University of Queensland, Brisbane, Queensland 4072, Australia
| | - Geoffrey J. McLachlan
- School of Mathematics and Physics, University of Queensland, Brisbane, Queensland 4072, Australia
| |
Collapse
|
13
|
|
14
|
Nguyen HD, McLachlan GJ. Maximum likelihood estimation of triangular and polygonal distributions. Comput Stat Data Anal 2016. [DOI: 10.1016/j.csda.2016.04.003] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
15
|
Chamroukhi F. Robust mixture of experts modeling using the t distribution. Neural Netw 2016; 79:20-36. [PMID: 27093693 DOI: 10.1016/j.neunet.2016.03.002] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2015] [Revised: 01/23/2016] [Accepted: 03/11/2016] [Indexed: 11/30/2022]
Abstract
Mixture of Experts (MoE) is a popular framework for modeling heterogeneity in data for regression, classification, and clustering. For regression and cluster analyses of continuous data, MoE usually uses normal experts following the Gaussian distribution. However, for a set of data containing a group or groups of observations with heavy tails or atypical observations, the use of normal experts is unsuitable and can unduly affect the fit of the MoE model. We introduce a robust MoE modeling using the t distribution. The proposed t MoE (TMoE) deals with these issues regarding heavy-tailed and noisy data. We develop a dedicated expectation-maximization (EM) algorithm to estimate the parameters of the proposed model by monotonically maximizing the observed data log-likelihood. We describe how the presented model can be used in prediction and in model-based clustering of regression data. The proposed model is validated on numerical experiments carried out on simulated data, which show the effectiveness and the robustness of the proposed model in terms of modeling non-linear regression functions as well as in model-based clustering. Then, it is applied to the real-world data of tone perception for musical data analysis, and the one of temperature anomalies for the analysis of climate change data. The obtained results show the usefulness of the TMoE model for practical applications.
Collapse
Affiliation(s)
- F Chamroukhi
- Université de Toulon, CNRS, LSIS, UMR 7296, 83957 La Garde, France; Aix Marseille Université, CNRS, ENSAM, LSIS, UMR 7296, 13397 Marseille, France; Laboratoire Paul Painlevé (LPP), UMR CNRS 8524, Université Lille 1, 59650 Villeneuve d'Ascq, France.
| |
Collapse
|
16
|
|
17
|
Hinde J, Ingrassia S, Lin TI, McNicholas P. The Third Special Issue on Advances in Mixture Models. Comput Stat Data Anal 2016. [DOI: 10.1016/j.csda.2015.08.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|