An efficient and flexible accelerator design for sparse convolutional neural networks

X Xie, J Lin, Z Wang, J Wei - … on Circuits and Systems I: Regular …, 2021 - ieeexplore.ieee.org
Designing hardware accelerators for convolutional neural networks (CNNs) has recently
attracted tremendous attention. Plenty of existing accelerators are built for dense CNNs or …

OMNI: A framework for integrating hardware and software optimizations for sparse CNNs

Y Liang, L Lu, J Xie - … on Computer-Aided Design of Integrated …, 2020 - ieeexplore.ieee.org
Convolution neural networks (CNNs) as one of today's main flavor of deep learning
techniques dominate in various image recognition tasks. As the model size of modern CNNs …

ESCALATE: Boosting the efficiency of sparse CNN accelerator with kernel decomposition

S Li, E Hanson, X Qian, HH Li, Y Chen - MICRO-54: 54th Annual IEEE …, 2021 - dl.acm.org
The ever-growing parameter size and computation cost of Convolutional Neural Network
(CNN) models hinder their deployment onto resource-constrained platforms. Network …

SqueezeFlow: A sparse CNN accelerator exploiting concise convolution rules

J Li, S Jiang, S Gong, J Wu, J Yan… - IEEE Transactions on …, 2019 - ieeexplore.ieee.org
Convolutional Neural Networks (CNNs) have been widely used in machine learning tasks.
While delivering state-of-the-art accuracy, CNNs are known as both compute-and memory …

Sparcnet: A hardware accelerator for efficient deployment of sparse convolutional networks

A Page, A Jafari, C Shea, T Mohsenin - ACM Journal on Emerging …, 2017 - dl.acm.org
Deep neural networks have been shown to outperform prior state-of-the-art solutions that
often relied heavily on hand-engineered feature extraction techniques coupled with simple …

GoSPA: An energy-efficient high-performance globally optimized sparse convolutional neural network accelerator

C Deng, Y Sui, S Liao, X Qian… - 2021 ACM/IEEE 48th …, 2021 - ieeexplore.ieee.org
The co-existence of activation sparsity and model sparsity in convolutional neural network
(CNN) models makes sparsity-aware CNN hardware designs very attractive. The existing …

An efficient hardware accelerator for structured sparse convolutional neural networks on FPGAs

C Zhu, K Huang, S Yang, Z Zhu… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Deep convolutional neural networks (CNNs) have achieved state-of-the-art performance in a
wide range of applications. However, deeper CNN models, which are usually computation …

WinoNN: Optimizing FPGA-based convolutional neural network accelerators using sparse Winograd algorithm

X Wang, C Wang, J Cao, L Gong… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
In recent years, a variety of accelerators on FPGAs have been proposed to speed up the
convolutional neural network (CNN) in many domain-specific application fields. Besides …

SWM: A high-performance sparse-winograd matrix multiplication CNN accelerator

D Wu, X Fan, W Cao, L Wang - IEEE Transactions on Very …, 2021 - ieeexplore.ieee.org
Many convolutional neural network (CNN) accelerators are proposed to exploit the sparsity
of the networks recently to enjoy the benefits of both computation and memory reduction …

SCNN: An accelerator for compressed-sparse convolutional neural networks

A Parashar, M Rhu, A Mukkara, A Puglielli… - ACM SIGARCH …, 2017 - dl.acm.org
Convolutional Neural Networks (CNNs) have emerged as a fundamental technology for
machine learning. High performance and extreme energy efficiency are critical for …