作者
Haocheng Zhao, Limin Yu, Runwei Guan, Liye Jia, Junqing Zhang, Yutao Yue
发表日期
2023/12/15
研讨会论文
2023 International Conference on Machine Learning and Applications (ICMLA)
页码范围
1-8
出版商
IEEE
简介
In the current era of multi-modal and large models gradually revealing their potential, neural network pruning has emerged as a crucial means of model compression. It is widely recognized that models tend to be over-parameterized, and pruning enables the removal of unimportant weights, leading to improved inference speed while preserving accuracy. From early methods such as gradient-based, and magnitude-based pruning to modern algorithms like iterative magnitude pruning, lottery ticket hypothesis, and pruning at initialization, researchers have strived to increase the compression ratio of model parameters while maintaining high accuracy. Currently, mainstream algorithms focus on the global pruning of neural networks using various scoring functions, followed by different pruning strategies to enhance the accuracy of sparse model. Recent studies have shown that random pruning with varying layer-wise …
学术搜索中的文章
H Zhao, L Yu, R Guan, L Jia, J Zhang, Y Yue - 2023 International Conference on Machine Learning …, 2023