Non-structured DNN weight pruning—Is it beneficial in any platform?

X Ma, S Lin, S Ye, Z He, L Zhang… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Large deep neural network (DNN) models pose the key challenge to energy efficiency due
to the significantly higher energy consumption of off-chip DRAM accesses than arithmetic or …

Structadmm: Achieving ultrahigh efficiency in structured pruning for dnns

T Zhang, S Ye, X Feng, X Ma, K Zhang… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Weight pruning methods of deep neural networks (DNNs) have been demonstrated to
achieve a good model pruning rate without loss of accuracy, thereby alleviating the …

[PDF][PDF] Adam-admm: A unified, systematic framework of structured weight pruning for dnns

T Zhang, K Zhang, S Ye, J Li, J Tang… - arXiv preprint arXiv …, 2018 - yeshaokai.github.io
Weight pruning methods of deep neural networks (DNNs) have been demonstrated to
achieve a good model pruning ratio without loss of accuracy, thereby alleviating the …

Admm-nn: An algorithm-hardware co-design framework of dnns using alternating direction methods of multipliers

A Ren, T Zhang, S Ye, J Li, W Xu, X Qian, X Lin… - Proceedings of the …, 2019 - dl.acm.org
Model compression is an important technique to facilitate efficient embedded and hardware
implementations of deep neural networks (DNNs), a number of prior works are dedicated to …

Opq: Compressing deep neural networks with one-shot pruning-quantization

P Hu, X Peng, H Zhu, MMS Aly, J Lin - Proceedings of the AAAI …, 2021 - ojs.aaai.org
Abstract As Deep Neural Networks (DNNs) usually are overparameterized and have millions
of weight parameters, it is challenging to deploy these large DNN models on resource …

CirCNN: accelerating and compressing deep neural networks using block-circulant weight matrices

C Ding, S Liao, Y Wang, Z Li, N Liu, Y Zhuo… - Proceedings of the 50th …, 2017 - dl.acm.org
Large-scale deep neural networks (DNNs) are both compute and memory intensive. As the
size of DNNs continues to grow, it is critical to improve the energy efficiency and …

Lightweight deep learning: An overview

CH Wang, KY Huang, Y Yao, JC Chen… - IEEE consumer …, 2022 - ieeexplore.ieee.org
With the recent success of the deep neural networks (DNNs) in the field of artificial
intelligence, the urge of deploying DNNs has drawn tremendous attention because it can …

A systematic dnn weight pruning framework using alternating direction method of multipliers

T Zhang, S Ye, K Zhang, J Tang… - Proceedings of the …, 2018 - openaccess.thecvf.com
Weight pruning methods for deep neural networks (DNNs) have been investigated recently,
but prior work in this area is mainly heuristic, iterative pruning, thereby lacking guarantees …

A unified framework of DNN weight pruning and weight clustering/quantization using ADMM

S Ye, T Zhang, K Zhang, J Li, J Xie, Y Liang… - arXiv preprint arXiv …, 2018 - arxiv.org
Many model compression techniques of Deep Neural Networks (DNNs) have been
investigated, including weight pruning, weight clustering and quantization, etc. Weight …

Global sparse momentum sgd for pruning very deep neural networks

X Ding, X Zhou, Y Guo, J Han… - Advances in Neural …, 2019 - proceedings.neurips.cc
Abstract Deep Neural Network (DNN) is powerful but computationally expensive and
memory intensive, thus impeding its practical usage on resource-constrained front-end …