Pruning vs quantization: which is better?

A Kuzmin, M Nagel, M Van Baalen… - Advances in neural …, 2024 - proceedings.neurips.cc
Neural network pruning and quantization techniques are almost as old as neural networks
themselves. However, to date, only ad-hoc comparisons between the two have been …

Deep learning paradigm and its bias for coronary artery wall segmentation in intravascular ultrasound scans: a closer look

V Kumari, N Kumar, S Kumar K, A Kumar… - Journal of …, 2023 - mdpi.com
Background and Motivation: Coronary artery disease (CAD) has the highest mortality rate;
therefore, its diagnosis is vital. Intravascular ultrasound (IVUS) is a high-resolution imaging …

Compresso: Structured pruning with collaborative prompting learns compact large language models

S Guo, J Xu, LL Zhang, M Yang - arXiv preprint arXiv:2310.05015, 2023 - arxiv.org
Despite the remarkable success of Large Language Models (LLMs), the massive size poses
significant deployment challenges, particularly on resource-constrained hardware. While …

Prune Efficiently by Soft Pruning

P Agarwal, M Mathew, KR Patel… - Proceedings of the …, 2024 - openaccess.thecvf.com
Embedded systems are power sensitive and have limited memory hence inferencing large
networks on such systems is difficult. Pruning techniques have been instrumental in …

Magnitude attention-based dynamic pruning

J Back, N Ahn, J Kim - arXiv preprint arXiv:2306.05056, 2023 - arxiv.org
Existing pruning methods utilize the importance of each weight based on specified criteria
only when searching for a sparse structure but do not utilize it during training. In this work …

A systematic DNN weight pruning framework based on symmetric accelerated stochastic ADMM

M Yuan, J Bai, F Jiang, L Du - Neurocomputing, 2024 - Elsevier
Weight pruning is widely employed in compressing Deep Neural Networks (DNNs) because
of the increasing computation and storage requirement. However, related work failed to …

LRNAS: Differentiable Searching for Adversarially Robust Lightweight Neural Architecture

Y Feng, Z Lv, H Chen, S Gao, F An… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
The adversarial robustness is critical to deep neural networks (DNNs) in deployment.
However, the improvement of adversarial robustness often requires compromising with the …

FedNISP: Neuron Importance Scope Propagation pruning for communication efficient federated learning

G Kumar, D Toshniwal - Computers and Electrical Engineering, 2024 - Elsevier
Federated learning (FL) has gained significant attention in academia and industry due to its
privacy-preserving nature. FL is a decentralized approach that allows clients to collaborate …

Pruning on-the-fly: A recoverable pruning method without fine-tuning

D Liu, X Liu - arXiv preprint arXiv:2212.12651, 2022 - arxiv.org
Most existing pruning works are resource-intensive, requiring retraining or fine-tuning of the
pruned models for accuracy. We propose a retraining-free pruning method based on …

COOL: Conservation of Output Links for Pruning Algorithms in Network Intrusion Detection

TN Dao, HJ Lee - IEEE Internet of Things Journal, 2023 - ieeexplore.ieee.org
To reduce network intrusion detection latency in a high volume of data traffic, on-device
detection with neuron pruning has been widely adopted by eliminating ineffective …