A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Recent advances on neural network pruning at initialization

H Wang, C Qin, Y Bai, Y Zhang, Y Fu - arXiv preprint arXiv:2103.06460, 2021 - arxiv.org
Neural network pruning typically removes connections or neurons from a pretrained
converged model; while a new pruning paradigm, pruning at initialization (PaI), attempts to …

Spvit: Enabling faster vision transformers via latency-aware soft token pruning

Z Kong, P Dong, X Ma, X Meng, W Niu, M Sun… - European conference on …, 2022 - Springer
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Validating the lottery ticket hypothesis with inertial manifold theory

Z Zhang, J Jin, Z Zhang, Y Zhou… - Advances in neural …, 2021 - proceedings.neurips.cc
Despite achieving remarkable efficiency, traditional network pruning techniques often follow
manually-crafted heuristics to generate pruned sparse networks. Such heuristic pruning …

F8net: Fixed-point 8-bit only multiplication for network quantization

Q Jin, J Ren, R Zhuang, S Hanumante, Z Li… - arXiv preprint arXiv …, 2022 - arxiv.org
Neural network quantization is a promising compression technique to reduce memory
footprint and save energy consumption, potentially leading to real-time inference. However …

Most activation functions can win the lottery without excessive depth

R Burkholz - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
The strong lottery ticket hypothesis has highlighted the potential for training deep neural
networks by pruning, which has inspired interesting practical and theoretical insights into …

Convolutional and residual networks provably contain lottery tickets

R Burkholz - International Conference on Machine Learning, 2022 - proceedings.mlr.press
Abstract The Lottery Ticket Hypothesis continues to have a profound practical impact on the
quest for small scale deep neural networks that solve modern deep learning tasks at …

Effective model sparsification by scheduled grow-and-prune methods

X Ma, M Qin, F Sun, Z Hou, K Yuan, Y Xu… - arXiv preprint arXiv …, 2021 - arxiv.org
Deep neural networks (DNNs) are effective in solving many real-world problems. Larger
DNN models usually exhibit better quality (eg, accuracy) but their excessive computation …

Plant'n'Seek: Can You Find the Winning Ticket?

J Fischer, R Burkholz - arXiv preprint arXiv:2111.11153, 2021 - arxiv.org
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that
aim to reduce the computational costs associated with deep learning during training and …