Gaussian processes and kernel methods: A review on connections and equivalences

M Kanagawa, P Hennig, D Sejdinovic… - arXiv preprint arXiv …, 2018 - arxiv.org
This paper is an attempt to bridge the conceptual gaps between researchers working on the
two widely used approaches based on positive definite kernels: Bayesian learning or …

Explaining neural scaling laws

Y Bahri, E Dyer, J Kaplan, J Lee, U Sharma - Proceedings of the National …, 2024 - pnas.org
The population loss of trained deep neural networks often follows precise power-law scaling
relations with either the size of the training dataset or the number of parameters in the …

Quasi-oracle estimation of heterogeneous treatment effects

X Nie, S Wager - Biometrika, 2021 - academic.oup.com
Flexible estimation of heterogeneous treatment effects lies at the heart of many statistical
applications, such as personalized medicine and optimal resource allocation. In this article …

Survey on cyberspace security

H Zhang, W Han, X Lai, D Lin, J Ma, JH Li - Science China Information …, 2015 - Springer
Along with the rapid development and wide application of information technology, human
society has entered the information era. In this era, people live and work in cyberspace …

Learning curves of generic features maps for realistic datasets with a teacher-student model

B Loureiro, C Gerbelot, H Cui, S Goldt… - Advances in …, 2021 - proceedings.neurips.cc
Teacher-student models provide a framework in which the typical-case performance of high-
dimensional supervised learning can be described in closed form. The assumptions of …

The inductive bias of quantum kernels

J Kübler, S Buchholz… - Advances in Neural …, 2021 - proceedings.neurips.cc
It has been hypothesized that quantum computers may lend themselves well to applications
in machine learning. In the present work, we analyze function classes defined via quantum …

Generalization properties of learning with random features

A Rudi, L Rosasco - Advances in neural information …, 2017 - proceedings.neurips.cc
We study the generalization properties of ridge regression with random features in the
statistical learning framework. We show for the first time that $ O (1/\sqrt {n}) $ learning …

Less is more: Nyström computational regularization

A Rudi, R Camoriano… - Advances in neural …, 2015 - proceedings.neurips.cc
We study Nyström type subsampling approaches to large scale kernel methods, and prove
learning bounds in the statistical learning setting, where random sampling and high …

On the equivalence between kernel quadrature rules and random feature expansions

F Bach - Journal of machine learning research, 2017 - jmlr.org
We show that kernel-based quadrature rules for computing integrals can be seen as a
special case of random feature expansions for positive definite kernels, for a particular …

[PDF][PDF] Divide and conquer kernel ridge regression: A distributed algorithm with minimax optimal rates

Y Zhang, J Duchi, M Wainwright - The Journal of Machine Learning …, 2015 - jmlr.org
We study a decomposition-based scalable approach to kernel ridge regression, and show
that it achieves minimax optimal convergence rates under relatively mild conditions. The …