Lsq+: Improving low-bit quantization through learnable offsets and better initialization

Y Bhalgat, J Lee, M Nagel… - Proceedings of the …, 2020 - openaccess.thecvf.com
Unlike ReLU, newer activation functions (like Swish, H-swish, Mish) that are frequently
employed in popular efficient architectures can also result in negative activation values, with …

Overcoming oscillations in quantization-aware training

M Nagel, M Fournarakis… - International …, 2022 - proceedings.mlr.press
When training neural networks with simulated quantization, we observe that quantized
weights can, rather unexpectedly, oscillate between two grid-points. The importance of this …

Eq-net: Elastic quantization neural networks

K Xu, L Han, Y Tian, S Yang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Current model quantization methods have shown their promising capability in reducing
storage space and computation complexity. However, due to the diversity of quantization …

Learnable companding quantization for accurate low-bit neural networks

K Yamamoto - Proceedings of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Quantizing deep neural networks is an effective method for reducing memory consumption
and improving inference speed, and is thus useful for implementation in resource …

Adabits: Neural network quantization with adaptive bit-widths

Q Jin, L Yang, Z Liao - … of the IEEE/CVF Conference on …, 2020 - openaccess.thecvf.com
Deep neural networks with adaptive configurations have gained increasing attention due to
the instant and flexible deployment of these models on platforms with different resource …

Explicit loss-error-aware quantization for low-bit deep neural networks

A Zhou, A Yao, K Wang, Y Chen - Proceedings of the IEEE …, 2018 - openaccess.thecvf.com
Benefiting from tens of millions of hierarchically stacked learnable parameters, Deep Neural
Networks (DNNs) have demonstrated overwhelming accuracy on a variety of artificial …

Hawq: Hessian aware quantization of neural networks with mixed-precision

Z Dong, Z Yao, A Gholami… - Proceedings of the …, 2019 - openaccess.thecvf.com
Abstract Model size and inference speed/power have become a major challenge in the
deployment of neural networks for many applications. A promising approach to address …

Post-training piecewise linear quantization for deep neural networks

J Fang, A Shafiee, H Abdel-Aziz, D Thorsley… - Computer Vision–ECCV …, 2020 - Springer
Quantization plays an important role in the energy-efficient deployment of deep neural
networks on resource-limited devices. Post-training quantization is highly desirable since it …

Data-free quantization through weight equalization and bias correction

M Nagel, M Baalen, T Blankevoort… - Proceedings of the …, 2019 - openaccess.thecvf.com
We introduce a data-free quantization method for deep neural networks that does not
require fine-tuning or hyperparameter selection. It achieves near-original model …

Dsconv: Efficient convolution operator

MG Nascimento, R Fawcett… - Proceedings of the …, 2019 - openaccess.thecvf.com
Quantization is a popular way of increasing the speed and lowering the memory usage of
Convolution Neural Networks (CNNs). When labelled training data is available, network …