Recent advances on neural network pruning at initialization

H Wang, C Qin, Y Bai, Y Zhang, Y Fu - arXiv preprint arXiv:2103.06460, 2021 - arxiv.org
Neural network pruning typically removes connections or neurons from a pretrained
converged model; while a new pruning paradigm, pruning at initialization (PaI), attempts to …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Pruning neural networks without any data by iteratively conserving synaptic flow

H Tanaka, D Kunin, DL Yamins… - Advances in neural …, 2020 - proceedings.neurips.cc
Pruning the parameters of deep neural networks has generated intense interest due to
potential savings in time, memory and energy both during training and at test time. Recent …

Picking winning tickets before training by preserving gradient flow

C Wang, G Zhang, R Grosse - arXiv preprint arXiv:2002.07376, 2020 - arxiv.org
Overparameterization has been shown to benefit both the optimization and generalization of
neural networks, but large networks are resource hungry at both training and test time …

Woodfisher: Efficient second-order approximation for neural network compression

SP Singh, D Alistarh - Advances in Neural Information …, 2020 - proceedings.neurips.cc
Second-order information, in the form of Hessian-or Inverse-Hessian-vector products, is a
fundamental tool for solving optimization problems. Recently, there has been significant …

Effective sparsification of neural networks with global sparsity constraint

X Zhou, W Zhang, H Xu… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Weight pruning is an effective technique to reduce the model size and inference time for
deep neural networks in real world deployments. However, since magnitudes and relative …

Eigendamage: Structured pruning in the kronecker-factored eigenbasis

C Wang, R Grosse, S Fidler… - … conference on machine …, 2019 - proceedings.mlr.press
Reducing the test time resource requirements of a neural network while preserving test
accuracy is crucial for running inference on resource-constrained devices. To achieve this …

Efficient neural network training via forward and backward propagation sparsification

X Zhou, W Zhang, Z Chen, S Diao… - Advances in neural …, 2021 - proceedings.neurips.cc
Sparse training is a natural idea to accelerate the training speed of deep neural networks
and save the memory usage, especially since large modern neural networks are …

[Retracted] DeepCompNet: A Novel Neural Net Model Compression Architecture

M Mary Shanthi Rani, P Chitra… - Computational …, 2022 - Wiley Online Library
The emergence of powerful deep learning architectures has resulted in breakthrough
innovations in several fields such as healthcare, precision farming, banking, education, and …

From TensorFlow graphs to LUTs and wires: Automated sparse and physically aware CNN hardware generation

M Hall, V Betz - 2020 International Conference on Field …, 2020 - ieeexplore.ieee.org
We present algorithms and an architectural methodology to enable zero skipping while
increasing frequency in per-layer customized data flow Convolutional Neural Network …