作者
Li Li, Zhu Li, Yue Li, Birendra Kathariya, Shuvra Bhattacharyya
发表日期
2019/3/26
研讨会论文
2019 Data Compression Conference (DCC)
页码范围
590-590
出版商
IEEE
简介
In this paper, based on the Hessian approximation, an incremental pruning method is proposed to compress the deep neural network. The proposed method starts from the idea of using the Hessian to measure the "importance" of each weight in a deep neural network, and it mainly has the following key contributions. First, we propose to use the second moment in Adam optimizer as a measure of the "importance" of each weight to avoid calculating the Hessian matrix. Second, an incremental method is proposed to prune the neural network step by step. The incremental method can adjust the remaining non-zero weights of the whole network after each pruning to help boost the performance of the pruned network. Last but not least, the proposed method applies an automatically-generated global threshold for all the weights among all the layers, which achieves the inter-layer bit allocation automatically. Such a …
引用总数
学术搜索中的文章
L Li, Z Li, Y Li, B Kathariya, S Bhattacharyya - 2019 Data Compression Conference (DCC), 2019