作者
Eric Qin, Geonhwa Jeong, William Won, Sheng-Chun Kao, Hyoukjun Kwon, Sudarshan Srinivasan, Dipankar Das, Gordon E Moon, Sivasankaran Rajamanickam, Tushar Krishna
发表日期
2021/5/17
研讨会论文
2021 IEEE International Parallel and Distributed Processing Symposium (IPDPS)
页码范围
1014-1024
出版商
IEEE
简介
Sparsity, which occurs in both scientific applications and Deep Learning (DL) models, has been a key target of optimization within recent ASIC accelerators due to the potential memory and compute savings. These applications use data stored in a variety of compression formats. We demonstrate that both the compactness of different compression formats and the compute efficiency of the algorithms enabled by them vary across tensor dimensions and amount of sparsity. Since DL and scientific workloads span across all sparsity regions, there can be numerous format combinations for optimizing memory and compute efficiency. Unfortunately, many proposed accelerators operate on one or two fixed format combinations. This work proposes hardware extensions to accelerators for supporting numerous format combinations seamlessly and demonstrates ~ 4 x speedup over performing format conversions in software.
引用总数
20212022202320242293
学术搜索中的文章
E Qin, G Jeong, W Won, SC Kao, H Kwon, S Srinivasan… - 2021 IEEE International Parallel and Distributed …, 2021