Perforatedcnns: Acceleration through elimination of redundant convolutions

M Figurnov, A Ibraimova… - Advances in neural …, 2016 - proceedings.neurips.cc
We propose a novel approach to reduce the computational cost of evaluation of
convolutional neural networks, a factor that has hindered their deployment in low-power …

Optimizing the deep neural networks by layer‐wise refined pruning and the acceleration on FPGA

H Li, X Yue, Z Wang, Z Chai, W Wang… - Computational …, 2022 - Wiley Online Library
To accelerate the practical applications of artificial intelligence, this paper proposes a high
efficient layer‐wise refined pruning method for deep neural networks at the software level …

PipeCNN: An OpenCL-based FPGA accelerator for large-scale convolution neuron networks

D Wang, J An, K Xu - arXiv preprint arXiv:1611.02450, 2016 - arxiv.org
Convolutional neural networks (CNNs) have been widely employed in many applications
such as image classification, video analysis and speech recognition. Being compute …

Recent advances in convolutional neural network acceleration

Q Zhang, M Zhang, T Chen, Z Sun, Y Ma, B Yu - Neurocomputing, 2019 - Elsevier
In recent years, convolutional neural networks (CNNs) have shown great performance in
various fields such as image classification, pattern recognition, and multi-media …

Dynet: Dynamic convolution for accelerating convolutional neural networks

Y Zhang, J Zhang, Q Wang, Z Zhong - arXiv preprint arXiv:2004.10694, 2020 - arxiv.org
Convolution operator is the core of convolutional neural networks (CNNs) and occupies the
most computation cost. To make CNNs more efficient, many methods have been proposed …

Gate decorator: Global filter pruning method for accelerating deep convolutional neural networks

Z You, K Yan, J Ye, M Ma… - Advances in neural …, 2019 - proceedings.neurips.cc
Filter pruning is one of the most effective ways to accelerate and compress convolutional
neural networks (CNNs). In this work, we propose a global filter pruning algorithm called …

Full-stack optimization for accelerating cnns using powers-of-two weights with fpga validation

B McDanel, SQ Zhang, HT Kung, X Dong - Proceedings of the ACM …, 2019 - dl.acm.org
We present a full-stack optimization framework for accelerating inference of CNNs
(Convolutional Neural Networks) and validate the approach with a field-programmable gate …

Structured probabilistic pruning for convolutional neural network acceleration

H Wang, Q Zhang, Y Wang, H Hu - arXiv preprint arXiv:1709.06994, 2017 - arxiv.org
In this paper, we propose a novel progressive parameter pruning method for Convolutional
Neural Network acceleration, named Structured Probabilistic Pruning (SPP), which …

Channel pruning based on mean gradient for accelerating convolutional neural networks

C Liu, H Wu - Signal Processing, 2019 - Elsevier
Abstract Convolutional Neural Networks (CNNs) are getting deeper and wider to improve
their performance and thus increase their computational complexity. We apply channel …

Channel pruning for accelerating very deep neural networks

Y He, X Zhang, J Sun - Proceedings of the IEEE …, 2017 - openaccess.thecvf.com
In this paper, we introduce a new channel pruning method to accelerate very deep
convolutional neural networks. Given a trained CNN model, we propose an iterative two …