A survey on deep neural network compression: Challenges, overview, and solutions

R Mishra, HP Gupta, T Dutta - arXiv preprint arXiv:2010.03954, 2020 - arxiv.org
Deep Neural Network (DNN) has gained unprecedented performance due to its automated
feature extraction capability. This high order performance leads to significant incorporation …

Transforming large-size to lightweight deep neural networks for IoT applications

R Mishra, H Gupta - ACM Computing Surveys, 2023 - dl.acm.org
Deep Neural Networks (DNNs) have gained unprecedented popularity due to their high-
order performance and automated feature extraction capability. This has encouraged …

Opq: Compressing deep neural networks with one-shot pruning-quantization

P Hu, X Peng, H Zhu, MMS Aly, J Lin - Proceedings of the AAAI …, 2021 - ojs.aaai.org
Abstract As Deep Neural Networks (DNNs) usually are overparameterized and have millions
of weight parameters, it is challenging to deploy these large DNN models on resource …

DeepSZ: A novel framework to compress deep neural networks by using error-bounded lossy compression

S Jin, S Di, X Liang, J Tian, D Tao… - Proceedings of the 28th …, 2019 - dl.acm.org
Today's deep neural networks (DNNs) are becoming deeper and wider because of
increasing demand on the analysis quality and more and more complex applications to …

Edp: An efficient decomposition and pruning scheme for convolutional neural network compression

X Ruan, Y Liu, C Yuan, B Li, W Hu, Y Li… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Model compression methods have become popular in recent years, which aim to alleviate
the heavy load of deep neural networks (DNNs) in real-world applications. However, most of …

Literature review of deep network compression

A Alqahtani, X Xie, MW Jones - Informatics, 2021 - mdpi.com
Deep networks often possess a vast number of parameters, and their significant redundancy
in parameterization has become a widely-recognized property. This presents significant …

Automatic neural network compression by sparsity-quantization joint learning: A constrained optimization-based approach

H Yang, S Gui, Y Zhu, J Liu - Proceedings of the IEEE/CVF …, 2020 - openaccess.thecvf.com
Abstract Deep Neural Networks (DNNs) are applied in a wide range of usecases. There is
an increased demand for deploying DNNs on devices that do not have abundant resources …

Soft weight-sharing for neural network compression

K Ullrich, E Meeds, M Welling - arXiv preprint arXiv:1702.04008, 2017 - arxiv.org
The success of deep learning in numerous application domains created the de-sire to run
and train them on mobile devices. This however, conflicts with their computationally, memory …

An overview of neural network compression

JO Neill - arXiv preprint arXiv:2006.03669, 2020 - arxiv.org
Overparameterized networks trained to convergence have shown impressive performance
in domains such as computer vision and natural language processing. Pushing state of the …

[PDF][PDF] Optimization based Layer-wise Magnitude-based Pruning for DNN Compression.

G Li, C Qian, C Jiang, X Lu, K Tang - IJCAI, 2018 - ijcai.org
Layer-wise magnitude-based pruning (LMP) is a very popular method for deep neural
network (DNN) compression. However, tuning the layerspecific thresholds is a difficult task …