When training neural networks with simulated quantization, we observe that quantized weights can, rather unexpectedly, oscillate between two grid-points. The importance of this …
K Xu, L Han, Y Tian, S Yang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Current model quantization methods have shown their promising capability in reducing storage space and computation complexity. However, due to the diversity of quantization …
K Yamamoto - Proceedings of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Quantizing deep neural networks is an effective method for reducing memory consumption and improving inference speed, and is thus useful for implementation in resource …
Q Jin, L Yang, Z Liao - … of the IEEE/CVF Conference on …, 2020 - openaccess.thecvf.com
Deep neural networks with adaptive configurations have gained increasing attention due to the instant and flexible deployment of these models on platforms with different resource …
Benefiting from tens of millions of hierarchically stacked learnable parameters, Deep Neural Networks (DNNs) have demonstrated overwhelming accuracy on a variety of artificial …
Abstract Model size and inference speed/power have become a major challenge in the deployment of neural networks for many applications. A promising approach to address …
Quantization plays an important role in the energy-efficient deployment of deep neural networks on resource-limited devices. Post-training quantization is highly desirable since it …
We introduce a data-free quantization method for deep neural networks that does not require fine-tuning or hyperparameter selection. It achieves near-original model …
MG Nascimento, R Fawcett… - Proceedings of the …, 2019 - openaccess.thecvf.com
Quantization is a popular way of increasing the speed and lowering the memory usage of Convolution Neural Networks (CNNs). When labelled training data is available, network …