A Basteri, D Trevisan - Machine Learning, 2024 - Springer
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound from above the quadratic Wasserstein distance between its output …
R Eldan, D Mikulincer… - Conference on Learning …, 2021 - proceedings.mlr.press
We study the extent to which wide neural networks may be approximated by Gaussian processes, when initialized with random weights. It is a well-established fact that as the …
We carry out an information-theoretical analysis of a two-layer neural network trained from input-output pairs generated by a teacher network with matching architecture, in …
SQ Zhang, F Wang, FL Fan - IEEE Transactions on Neural …, 2022 - ieeexplore.ieee.org
Recent years have witnessed an increasing interest in the correspondence between infinitely wide networks and Gaussian processes. Despite the effectiveness and elegance of …
A Bordino, S Favaro, S Fortini - Proceedings of Machine …, 2024 - iris.unibocconi.it
There is a recent and growing literature on large-width asymptotic and non-asymptotic properties of deep Gaussian neural networks (NNs), namely NNs with weights initialized as …
We consider the optimisation of large and shallow neural networks via gradient flow, where the output of each hidden node is scaled by some positive parameter. We focus on the case …
D Trevisan - arXiv preprint arXiv:2312.11737, 2023 - arxiv.org
We establish novel rates for the Gaussian approximation of random deep neural networks with Gaussian parameters (weights and biases) and Lipschitz activation functions, in the …
State-of-the-art neural network training methods depend on the gradient of the network function. Therefore, they cannot be applied to networks whose activation functions do not …
Infinitely wide or deep neural networks (NNs) with independent and identically distributed (iid) parameters have been shown to be equivalent to Gaussian processes. Because of the …