Selective inference for k-means clustering

YT Chen, DM Witten - Journal of Machine Learning Research, 2023 - jmlr.org
We consider the problem of testing for a difference in means between clusters of
observations identified via k-means clustering. In this setting, classical hypothesis tests lead …

Data fission: splitting a single data point

J Leiner, B Duan, L Wasserman… - Journal of the American …, 2023 - Taylor & Francis
Suppose we observe a random vector X from some distribution in a known family with
unknown parameters. We ask the following question: when is it possible to split X into two …

More powerful selective inference for the graph fused lasso

Y Chen, S Jewell, D Witten - Journal of Computational and …, 2023 - Taylor & Francis
The graph fused lasso—which includes as a special case the one-dimensional fused lasso—
is widely used to reconstruct signals that are piecewise constant on a graph, meaning that …

Exact selective inference with randomization

S Panigrahi, K Fry, J Taylor - Biometrika, 2024 - academic.oup.com
We introduce a pivot for exact selective inference with randomization. Not only does our
pivot lead to exact inference in Gaussian regression models, but it is also available in closed …

Generalized data thinning using sufficient statistics

A Dharamshi, A Neufeld, K Motwani… - Journal of the …, 2024 - Taylor & Francis
Our goal is to develop a general strategy to decompose a random variable X into multiple
independent random variables, without sacrificing any information about unknown …

Valid inference after causal discovery

P Gradu, T Zrnic, Y Wang, MI Jordan - Journal of the American …, 2024 - Taylor & Francis
Causal discovery and causal effect estimation are two fundamental tasks in causal
inference. While many methods have been developed for each task individually, statistical …

Post-selection inference via algorithmic stability

T Zrnic, MI Jordan - The Annals of Statistics, 2023 - projecteuclid.org
Post-selection inference via algorithmic stability Page 1 The Annals of Statistics 2023, Vol. 51,
No. 4, 1666–1691 https://doi.org/10.1214/23-AOS2303 © Institute of Mathematical Statistics …

Selective inference using randomized group lasso estimators for general models

Y Huang, S Pirenne, S Panigrahi… - arXiv preprint arXiv …, 2023 - arxiv.org
Selective inference methods are developed for group lasso estimators for use with a wide
class of distributions and loss functions. The method includes the use of exponential family …

Reducing the complexity of high-dimensional environmental data: An analytical framework using LASSO with considerations of confounding for statistical inference

S Frndak, G Yu, Y Oulhote, EI Queirolo, G Barg… - International journal of …, 2023 - Elsevier
Purpose Frameworks for selecting exposures in high-dimensional environmental datasets,
while considering confounding, are lacking. We present a two-step approach for exposure …

Data thinning for convolution-closed distributions

A Neufeld, A Dharamshi, LL Gao, D Witten - Journal of Machine Learning …, 2024 - jmlr.org
We propose data thinning, an approach for splitting an observation into two or more
independent parts that sum to the original observation, and that follow the same distribution …