A systematic survey of general sparse matrix-matrix multiplication

J Gao, W Ji, F Chang, S Han, B Wei, Z Liu… - ACM Computing …, 2023 - dl.acm.org
General Sparse Matrix-Matrix Multiplication (SpGEMM) has attracted much attention from
researchers in graph analyzing, scientific computing, and deep learning. Many optimization …

Flexagon: A multi-dataflow sparse-sparse matrix multiplication accelerator for efficient dnn processing

F Muñoz-Martínez, R Garg, M Pellauer… - Proceedings of the 28th …, 2023 - dl.acm.org
Sparsity is a growing trend in modern DNN models. Existing Sparse-Sparse Matrix
Multiplication (SpMSpM) accelerators are tailored to a particular SpMSpM dataflow (ie, Inner …

Spade: A flexible and scalable accelerator for spmm and sddmm

G Gerogiannis, S Yesil, D Lenadora, D Cao… - Proceedings of the 50th …, 2023 - dl.acm.org
The widespread use of Sparse Matrix Dense Matrix Multiplication (SpMM) and Sampled
Dense Matrix Dense Matrix Multiplication (SDDMM) kernels makes them candidates for …

Exploring architecture, dataflow, and sparsity for gcn accelerators: A holistic framework

L Yin, J Wang, H Zheng - Proceedings of the Great Lakes Symposium on …, 2023 - dl.acm.org
Recent years have seen an increasing number of Graph Convolutional Network (GCN)
models employed in various real-world applications. However, designing efficient …

Training recipe for n: M structured sparsity with decaying pruning mask

SC Kao, A Yazdanbakhsh, S Subramanian… - arXiv preprint arXiv …, 2022 - arxiv.org
Sparsity has become one of the promising methods to compress and accelerate Deep
Neural Networks (DNNs). Among different categories of sparsity, structured sparsity has …

Progressive Gradient Flow for Robust N: M Sparsity Training in Transformers

AR Bambhaniya, A Yazdanbakhsh… - arXiv preprint arXiv …, 2024 - arxiv.org
N: M Structured sparsity has garnered significant interest as a result of relatively modest
overhead and improved efficiency. Additionally, this form of sparsity holds considerable …

Copernicus: Characterizing the performance implications of compression formats used in sparse workloads

B Asgari, R Hadidi, J Dierberger… - 2021 IEEE …, 2021 - ieeexplore.ieee.org
Sparse matrices are the key ingredients of several application domains, from scientific
computing to machine learning. The primary challenge with sparse matrices has been …

SAGA: Sparsity-Agnostic Graph Convolutional Network Acceleration with Near-Optimal Workload Balance

S Gandham, L Yin, H Zheng… - 2023 IEEE/ACM …, 2023 - ieeexplore.ieee.org
Graph Convolutional Networks (GCNs) have shown much promise in resolving
sophisticated scientific problems with non-Euclidean data, such as traffic prediction, disease …

Cerberus: Triple mode acceleration of sparse matrix and vector multiplication

S Hwang, D Baek, J Park, J Huh - ACM Transactions on Architecture and …, 2024 - dl.acm.org
The multiplication of sparse matrix and vector (SpMV) is one of the most widely used kernels
in high-performance computing as well as machine learning acceleration for sparse neural …

IAP-SpTV: An input-aware adaptive pipeline SpTV via GCN on CPU-GPU

H Wang, W Yang, R Hu, R Ouyang, K Li, K Li - Journal of Parallel and …, 2023 - Elsevier
Sparse tensor-times-vector (SpTV) is the core computation of tensor decomposition.
Optimizing the computational performance of SpTV on CPU-GPU becomes a challenge due …