Identifiability of deep generative models without auxiliary information

B Kivva, G Rajendran, P Ravikumar… - Advances in Neural …, 2022 - proceedings.neurips.cc
We prove identifiability of a broad class of deep latent variable models that (a) have
universal approximation capabilities and (b) are the decoders of variational autoencoders …

Abide by the law and follow the flow: Conservation laws for gradient flows

S Marcotte, R Gribonval… - Advances in neural …, 2024 - proceedings.neurips.cc
Understanding the geometric properties of gradient descent dynamics is a key ingredient in
deciphering the recent success of very large machine learning models. A striking …

Parameter identifiability of a deep feedforward ReLU neural network

J Bona-Pellissier, F Bachoc, F Malgouyres - Machine Learning, 2023 - Springer
The possibility for one to recover the parameters—weights and biases—of a neural network
thanks to the knowledge of its function on a subset of the input space can be, depending on …

Does a sparse ReLU network training problem always admit an optimum?

T LE, R Gribonval, E Riccietti - Advances in Neural …, 2023 - proceedings.neurips.cc
Given a training set, a loss function, and a neural network architecture, it is often taken for
granted that optimal network parameters exist, and a common practice is to apply available …

Local identifiability of deep reLU neural networks: the theory

J Bona-Pellissier, F Malgouyres… - Advances in neural …, 2022 - proceedings.neurips.cc
Is a sample rich enough to determine, at least locally, the parameters of a neural network?
To answer this question, we introduce a new local parameterization of a given deep ReLU …

A path-norm toolkit for modern networks: consequences, promises and challenges

A Gonon, N Brisebarre, E Riccietti… - arXiv preprint arXiv …, 2023 - arxiv.org
This work introduces the first toolkit around path-norms that is fully able to encompass
general DAG ReLU networks with biases, skip connections and any operation based on the …

Functional equivalence and path connectivity of reducible hyperbolic tangent networks

M Farrugia-Roberts - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Understanding the learning process of artificial neural networks requires clarifying the
structure of the parameter space within which learning takes place. A neural network …

[HTML][HTML] NPSFF-Net: Enhanced Building Segmentation in Remote Sensing Images via Novel Pseudo-Siamese Feature Fusion

N Guo, M Jiang, X Hu, Z Su, W Zhang, R Li, J Luo - Remote Sensing, 2024 - mdpi.com
Building segmentation has extensive research value and application prospects in high-
resolution remote sensing image (HRSI) processing. However, complex architectural …

[PDF][PDF] Complexity of Feed-Forward Neural Networks from the Perspective of Functional Equivalence.

G Shen - arXiv preprint arXiv, 2023 - researchgate.net
In this paper, we investigate the complexity of feed-forward neural networks by examining
the concept of functional equivalence, which suggests that different network …

Geometry-induced Implicit Regularization in Deep ReLU Neural Networks

J Bona-Pellissier - arXiv preprint arXiv:2402.08269, 2024 - arxiv.org
It is well known that neural networks with many more parameters than training examples do
not overfit. Implicit regularization phenomena, which are still not well understood, occur …