PermDNN: Efficient compressed DNN architecture with permuted diagonal matrices

C Deng, S Liao, Y Xie, KK Parhi… - 2018 51st Annual …, 2018 - ieeexplore.ieee.org
Deep neural network (DNN) has emerged as the most important and popular artificial
intelligent (AI) technique. The growth of model size poses a key energy efficiency challenge …

Opq: Compressing deep neural networks with one-shot pruning-quantization

P Hu, X Peng, H Zhu, MMS Aly, J Lin - Proceedings of the AAAI …, 2021 - ojs.aaai.org
Abstract As Deep Neural Networks (DNNs) usually are overparameterized and have millions
of weight parameters, it is challenging to deploy these large DNN models on resource …

DeepSZ: A novel framework to compress deep neural networks by using error-bounded lossy compression

S Jin, S Di, X Liang, J Tian, D Tao… - Proceedings of the 28th …, 2019 - dl.acm.org
Today's deep neural networks (DNNs) are becoming deeper and wider because of
increasing demand on the analysis quality and more and more complex applications to …

Lightweight deep learning: An overview

CH Wang, KY Huang, Y Yao, JC Chen… - IEEE consumer …, 2022 - ieeexplore.ieee.org
With the recent success of the deep neural networks (DNNs) in the field of artificial
intelligence, the urge of deploying DNNs has drawn tremendous attention because it can …

A survey on deep neural network compression: Challenges, overview, and solutions

R Mishra, HP Gupta, T Dutta - arXiv preprint arXiv:2010.03954, 2020 - arxiv.org
Deep Neural Network (DNN) has gained unprecedented performance due to its automated
feature extraction capability. This high order performance leads to significant incorporation …

CirCNN: accelerating and compressing deep neural networks using block-circulant weight matrices

C Ding, S Liao, Y Wang, Z Li, N Liu, Y Zhuo… - Proceedings of the 50th …, 2017 - dl.acm.org
Large-scale deep neural networks (DNNs) are both compute and memory intensive. As the
size of DNNs continues to grow, it is critical to improve the energy efficiency and …

Non-structured DNN weight pruning—Is it beneficial in any platform?

X Ma, S Lin, S Ye, Z He, L Zhang… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Large deep neural network (DNN) models pose the key challenge to energy efficiency due
to the significantly higher energy consumption of off-chip DRAM accesses than arithmetic or …

TIE: Energy-efficient tensor train-based inference engine for deep neural network

C Deng, F Sun, X Qian, J Lin, Z Wang… - Proceedings of the 46th …, 2019 - dl.acm.org
In the era of artificial intelligence (AI), deep neural networks (DNNs) have emerged as the
most important and powerful AI technique. However, large DNN models are both storage …

Automatic neural network compression by sparsity-quantization joint learning: A constrained optimization-based approach

H Yang, S Gui, Y Zhu, J Liu - Proceedings of the IEEE/CVF …, 2020 - openaccess.thecvf.com
Abstract Deep Neural Networks (DNNs) are applied in a wide range of usecases. There is
an increased demand for deploying DNNs on devices that do not have abundant resources …

Discrete model compression with resource constraint for deep neural networks

S Gao, F Huang, J Pei, H Huang - Proceedings of the IEEE …, 2020 - openaccess.thecvf.com
In this paper, we target to address the problem of compression and acceleration of
Convolutional Neural Networks (CNNs). Specifically, we propose a novel structural pruning …