A comprehensive survey on regularization strategies in machine learning

Y Tian, Y Zhang - Information Fusion, 2022 - Elsevier
In machine learning, the model is not as complicated as possible. Good generalization
ability means that the model not only performs well on the training data set, but also can …

Normalization techniques in training dnns: Methodology, analysis and application

L Huang, J Qin, Y Zhou, F Zhu, L Liu… - IEEE transactions on …, 2023 - ieeexplore.ieee.org
Normalization techniques are essential for accelerating the training and improving the
generalization of deep neural networks (DNNs), and have successfully been used in various …

Micro-batch training with batch-channel normalization and weight standardization

S Qiao, H Wang, C Liu, W Shen, A Yuille - arXiv preprint arXiv:1903.10520, 2019 - arxiv.org
Batch Normalization (BN) has become an out-of-box technique to improve deep network
training. However, its effectiveness is limited for micro-batch training, ie, each GPU typically …

Representative batch normalization with feature calibration

SH Gao, Q Han, D Li, MM Cheng… - Proceedings of the …, 2021 - openaccess.thecvf.com
Abstract Batch Normalization (BatchNorm) has become the default component in modern
neural networks to stabilize training. In BatchNorm, centering and scaling operations, along …

Revisiting internal covariate shift for batch normalization

M Awais, MTB Iqbal, SH Bae - IEEE Transactions on Neural …, 2020 - ieeexplore.ieee.org
Despite the success of batch normalization (BatchNorm) and a plethora of its variants, the
exact reasons for its success are still shady. The original BatchNorm article explained it as a …

Switchable normalization for learning-to-normalize deep representation

P Luo, R Zhang, J Ren, Z Peng… - IEEE transactions on …, 2019 - ieeexplore.ieee.org
We address a learning-to-normalize problem by proposing Switchable Normalization (SN),
which learns to select different normalizers for different normalization layers of a deep neural …

Beyond batchnorm: Towards a unified understanding of normalization in deep learning

ES Lubana, R Dick, H Tanaka - Advances in Neural …, 2021 - proceedings.neurips.cc
Inspired by BatchNorm, there has been an explosion of normalization layers in deep
learning. Recent works have identified a multitude of beneficial properties in BatchNorm to …

Differentiable dynamic quantization with mixed precision and adaptive resolution

Z Zhang, W Shao, J Gu, X Wang… - … Conference on Machine …, 2021 - proceedings.mlr.press
Abstract Model quantization is challenging due to many tedious hyper-parameters such as
precision (bitwidth), dynamic range (minimum and maximum discrete values) and stepsize …

Differentiable learning-to-group channels via groupable convolutional neural networks

Z Zhang, J Li, W Shao, Z Peng… - Proceedings of the …, 2019 - openaccess.thecvf.com
Group convolution, which divides the channels of ConvNets into groups, has achieved
impressive improvement over the regular convolution operation. However, existing models …

Proxy-normalizing activations to match batch normalization while removing batch dependence

A Labatie, D Masters… - Advances in Neural …, 2021 - proceedings.neurips.cc
We investigate the reasons for the performance degradation incurred with batch-
independent normalization. We find that the prototypical techniques of layer normalization …