A comprehensive review of binary neural network

C Yuan, SS Agaian - Artificial Intelligence Review, 2023 - Springer
Deep learning (DL) has recently changed the development of intelligent systems and is
widely adopted in many real-life applications. Despite their various benefits and potentials …

A survey of quantization methods for efficient neural network inference

A Gholami, S Kim, Z Dong, Z Yao… - Low-Power Computer …, 2022 - taylorfrancis.com
This chapter provides approaches to the problem of quantizing the numerical values in deep
Neural Network computations, covering the advantages/disadvantages of current methods …

Adabin: Improving binary neural networks with adaptive binary sets

Z Tu, X Chen, P Ren, Y Wang - European conference on computer vision, 2022 - Springer
This paper studies the Binary Neural Networks (BNNs) in which weights and activations are
both binarized into 1-bit values, thus greatly reducing the memory usage and computational …

A comprehensive survey on model quantization for deep neural networks

B Rokh, A Azarpeyvand, A Khanteymoori - arXiv preprint arXiv …, 2022 - arxiv.org
Recent advances in machine learning by deep neural networks are significant. But using
these networks has been accompanied by a huge number of parameters for storage and …

Intraq: Learning synthetic images with intra-class heterogeneity for zero-shot network quantization

Y Zhong, M Lin, G Nan, J Liu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Learning to synthesize data has emerged as a promising direction in zero-shot quantization
(ZSQ), which represents neural networks by low-bit integer without accessing any of the real …

Recu: Reviving the dead weights in binary neural networks

Z Xu, M Lin, J Liu, J Chen, L Shao… - Proceedings of the …, 2021 - openaccess.thecvf.com
Binary neural networks (BNNs) have received increasing attention due to their superior
reductions of computation and memory. Most existing works focus on either lessening the …

Data-free knowledge distillation for image super-resolution

Y Zhang, H Chen, X Chen, Y Deng… - Proceedings of the …, 2021 - openaccess.thecvf.com
Convolutional network compression methods require training data for achieving acceptable
results, but training data is routinely unavailable due to some privacy and transmission …

Hard sample matters a lot in zero-shot quantization

H Li, X Wu, F Lv, D Liao, TH Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
Zero-shot quantization (ZSQ) is promising for compressing and accelerating deep neural
networks when the data for training full-precision models are inaccessible. In ZSQ, network …

Pruning networks with cross-layer ranking & k-reciprocal nearest filters

M Lin, L Cao, Y Zhang, L Shao… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
This article focuses on filter-level network pruning. A novel pruning method, termed CLR-
RNF, is proposed. We first reveal a “long-tail” pruning problem in magnitude-based weight …

Bibench: Benchmarking and analyzing network binarization

H Qin, M Zhang, Y Ding, A Li, Z Cai… - International …, 2023 - proceedings.mlr.press
Network binarization emerges as one of the most promising compression approaches
offering extraordinary computation and memory savings by minimizing the bit-width …