We consider the problem of testing for a difference in means between clusters of observations identified via k-means clustering. In this setting, classical hypothesis tests lead …
T Manole, A Khalili - The Annals of Statistics, 2021 - projecteuclid.org
Estimating the number of components in finite mixture models via the Group-Sort-Fuse procedure Page 1 The Annals of Statistics 2021, Vol. 49, No. 6, 3043–3069 https://doi.org/10.1214/21-AOS2072 …
Top-K sparse softmax gating mixture of experts has been widely used for scaling up massive deep-learning architectures without increasing the computational cost. Despite its popularity …
Originally introduced as a neural network for ensemble learning, mixture of experts (MoE) has recently become a fundamental building block of highly successful modern deep neural …
T Manole, N Ho - International Conference on Machine …, 2022 - proceedings.mlr.press
We revisit the classical problem of deriving convergence rates for the maximum likelihood estimator (MLE) in finite mixture models. The Wasserstein distance has become a standard …
Scientists and engineers are often interested in learning the number of subpopulations (or components) present in a data set. A common suggestion is to use a finite mixture model …
Continuously learning to solve unseen tasks with limited experience has been extensively pursued in meta-learning and continual learning, but with restricted assumptions such as …
A Guha, N Ho, XL Nguyen - International Conference on …, 2023 - proceedings.mlr.press
Dirichlet Process mixture models (DPMM) in combination with Gaussian kernels have been an important modeling tool for numerous data domains arising from biological, physical, and …