Fetchsgd: Communication-efficient federated learning with sketching

D Rothchild, A Panda, E Ullah, N Ivkin… - International …, 2020 - proceedings.mlr.press
Existing approaches to federated learning suffer from a communication bottleneck as well as
convergence issues due to sparse client participation. In this paper we introduce a novel …

New frameworks for offline and streaming coreset constructions

V Braverman, D Feldman, H Lang, A Statman… - arXiv preprint arXiv …, 2016 - arxiv.org
A coreset for a set of points is a small subset of weighted points that approximately
preserves important properties of the original set. Specifically, if $ P $ is a set of points, $ Q …

Sharper Bounds for Sensitivity Sampling

D Woodruff, T Yasuda - International Conference on …, 2023 - proceedings.mlr.press
In large scale machine learning, random sampling is a popular way to approximate datasets
by a small representative subset of examples. In particular, sensitivity sampling is an …

Tight bounds for adversarially robust streams and sliding windows via difference estimators

DP Woodruff, S Zhou - 2021 IEEE 62nd Annual Symposium on …, 2022 - ieeexplore.ieee.org
In the adversarially robust streaming model, a stream of elements is presented to an
algorithm and is allowed to depend on the output of the algorithm at earlier times during the …

Adversarial robustness of streaming algorithms through importance sampling

V Braverman, A Hassidim, Y Matias… - Advances in …, 2021 - proceedings.neurips.cc
Robustness against adversarial attacks has recently been at the forefront of algorithmic
design for machine learning tasks. In the adversarial streaming model, an adversary gives …

Online lewis weight sampling

DP Woodruff, T Yasuda - Proceedings of the 2023 Annual ACM-SIAM …, 2023 - SIAM
The seminal work of Cohen and Peng [CP15](STOC 2015) introduced Lewis weight
sampling to the theoretical computer science community, which yields fast row sampling …

New subset selection algorithms for low rank approximation: Offline and online

DP Woodruff, T Yasuda - Proceedings of the 55th Annual ACM …, 2023 - dl.acm.org
Subset selection for the rank k approximation of an n× d matrix A offers improvements in the
interpretability of matrices, as well as a variety of computational savings. This problem is well …

Streaming Euclidean k-median and k-means with o (log n) Space

V Cohen-Addad, DP Woodruff… - 2023 IEEE 64th Annual …, 2023 - ieeexplore.ieee.org
We consider the classic Euclidean k-median and k-means objective on data streams, where
the goal is to provide a (1+ε)-approximation to the optimal k-median or k-means solution …

Coresets for classification–simplified and strengthened

T Mai, C Musco, A Rao - Advances in Neural Information …, 2021 - proceedings.neurips.cc
We give relative error coresets for training linear classifiers with a broad class of loss
functions, including the logistic loss and hinge loss. Our construction achieves …

Near-Optimal -Clustering in the Sliding Window Model

D Woodruff, P Zhong, S Zhou - Advances in Neural …, 2024 - proceedings.neurips.cc
Clustering is an important technique for identifying structural information in large-scale data
analysis, where the underlying dataset may be too large to store. In many applications …