Originally introduced as a neural network for ensemble learning, mixture of experts (MoE) has recently become a fundamental building block of highly successful modern deep neural …
T Manole, N Ho - International Conference on Machine …, 2022 - proceedings.mlr.press
We revisit the classical problem of deriving convergence rates for the maximum likelihood estimator (MLE) in finite mixture models. The Wasserstein distance has become a standard …
N Ho, CY Yang, MI Jordan - Journal of Machine Learning Research, 2022 - jmlr.org
We provide a theoretical treatment of over-specified Gaussian mixtures of experts with covariate-free gating networks. We establish the convergence rates of the maximum …
We study the maximum likelihood estimation (MLE) in the multivariate deviated model where the data are generated from the density function $(1-\lambda^{\ast}) h_ {0}(x)+\lambda^{\ast} …
Y Zhang, Y Yang - Journal of the Royal Statistical Society Series …, 2024 - academic.oup.com
This article considers Bayesian model selection via mean-field (MF) variational approximation. Towards this goal, we study the non-asymptotic properties of MF inference …
We analyze the posterior contraction rates of parameters in Bayesian models via the Langevin diffusion process, in particular by controlling moments of the stochastic process …
D Do, L Do, XL Nguyen - Electronic Journal of Statistics, 2025 - projecteuclid.org
Mixtures of regression are useful for regression learning with respect to an uncertain and heterogeneous response variable of interest. In addition to being a rich predictive model for …