Y Zhou, T Pang, K Liu… - Advances in Neural …, 2024 - proceedings.neurips.cc
Regularization in modern machine learning is crucial, and it can take various forms in algorithmic design: training set, model family, error function, regularization terms, and …
Selecting suitable architecture parameters and training hyperparameters is essential for enhancing machine learning (ML) model performance. Several recent empirical studies …
H Guo, J Jin, B Liu - Applied Sciences, 2023 - mdpi.com
Averaging neural network weights sampled by a backbone stochastic gradient descent (SGD) is a simple-yet-effective approach to assist the backbone SGD in finding better …
Ensembling has a long history in statistical data analysis, with many impactful applications. However, in many modern machine learning settings, the benefits of ensembling are less …
C Cianfarani, AN Bhagoji, V Sehwag… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Representation learning,\textit {ie} the generation of representations useful for downstream applications, is a task of fundamental importance that underlies much of the …
J Park, I Pelakh, S Wojtowytsch - Advances in Neural …, 2023 - proceedings.neurips.cc
We investigate how shallow ReLU networks interpolate between known regions. Our analysis shows that empirical risk minimizers converge to a minimum norm interpolant as …
Selecting suitable architecture parameters and training hyperparameters is essential for enhancing machine learning (ML) model performance. Several recent empirical studies …
Y Zhou, Y Yang, A Chang… - … on Machine Learning, 2023 - proceedings.mlr.press
Recent work has highlighted the complex influence training hyperparameters, eg, the number of training epochs, can have on the prunability of machine learning models …
Deep learning sometimes appears to work in unexpected ways. In pursuit of a deeper understanding of its surprising behaviors, we investigate the utility of a simple yet accurate …