J Wu, D Zhu, L Fang, Y Deng… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Network pruning is one of the chief means for improving the computational efficiency of Deep Neural Networks (DNNs). Pruning-based methods generally discard network kernels …
H Liu, S Elkerdawy, N Ray… - Proceedings of the …, 2021 - openaccess.thecvf.com
Neural network quantization has achieved a high compression rate using a fixed low bit- width representation of weights and activations while maintaining the accuracy of the high …
The deployment of Deep Neural Network (DNN)-based networks on resource-constrained devices remains a significant challenge due to their high computational and parameter …
H Tang, Y Lu, Q Xuan - ICASSP 2023-2023 IEEE International …, 2023 - ieeexplore.ieee.org
Despite the popularization of deep neural networks (DNNs) in many fields, it is still challenging to deploy state-of-the-art models to resource-constrained devices due to high …
X Luo, D Liu, H Kong, S Huai… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
The prohibitive complexity of convolutional neural networks (CNNs) has triggered an increasing demand for network simplification. To this end, one natural solution is to remove …
Abstract Deep Convolutional Neural Networks (CNNs), continue to demonstrate remarkable performance across various tasks. However, their computational demands and energy …
J Tmamna, EB Ayed, MB Ayed - … 2021, Sanur, Bali, Indonesia, December 8 …, 2021 - Springer
Pruning has recently become ever-important research to compress deep neural networks. Previous pruning methods focus on removing filters, channels, or weights to reduce the …
Y Lu, P Zhang, J Wang, L Ma, X Yang… - arXiv preprint arXiv …, 2024 - arxiv.org
Deep learning has revolutionized computing in many real-world applications, arguably due to its remarkable performance and extreme convenience as an end-to-end solution …