Sparcl: Sparse continual learning on the edge

Z Wang, Z Zhan, Y Gong, G Yuan… - Advances in …, 2022 - proceedings.neurips.cc
Existing work in continual learning (CL) focuses on mitigating catastrophic forgetting, ie,
model performance deterioration on past tasks when learning a new task. However, the …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Compute-efficient deep learning: Algorithmic trends and opportunities

BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …

Interspace pruning: Using adaptive filter representations to improve training of sparse cnns

P Wimmer, J Mehnert… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Unstructured pruning is well suited to reduce the memory footprint of convolutional neural
networks (CNNs), both at training and inference time. CNNs contain parameters arranged in …

Effective model sparsification by scheduled grow-and-prune methods

X Ma, M Qin, F Sun, Z Hou, K Yuan, Y Xu… - arXiv preprint arXiv …, 2021 - arxiv.org
Deep neural networks (DNNs) are effective in solving many real-world problems. Larger
DNN models usually exhibit better quality (eg, accuracy) but their excessive computation …

Dimensionality reduced training by pruning and freezing parts of a deep neural network: a survey

P Wimmer, J Mehnert, AP Condurache - Artificial Intelligence Review, 2023 - Springer
State-of-the-art deep learning models have a parameter count that reaches into the billions.
Training, storing and transferring such models is energy and time consuming, thus costly. A …

Layer freezing & data sieving: missing pieces of a generic framework for sparse training

G Yuan, Y Li, S Li, Z Kong, S Tulyakov… - Advances in …, 2022 - proceedings.neurips.cc
Recently, sparse training has emerged as a promising paradigm for efficient deep learning
on edge devices. The current research mainly devotes the efforts to reducing training costs …

COPS: Controlled pruning before training starts

W Paul, M Jens, C Alexandru - 2021 International Joint …, 2021 - ieeexplore.ieee.org
State-of-the-art deep neural network (DNN) pruning techniques, applied one-shot before
training starts, evaluate sparse architectures with the help of a single criterion-called pruning …

ELSA: Partial Weight Freezing for Overhead-Free Sparse Network Deployment

P Halvachi, A Peste, D Alistarh, CH Lampert - arXiv preprint arXiv …, 2023 - arxiv.org
We present ELSA, a practical solution for creating deep networks that can easily be
deployed at different levels of sparsity. The core idea is to embed one or more sparse …

CRUS: A hardware-efficient algorithm mitigating highly nonlinear weight update in CIM crossbar arrays for artificial neural networks

J Lee, J Hwang, Y Cho, MK Park… - IEEE Journal on …, 2022 - ieeexplore.ieee.org
Mitigating the nonlinear weight update of synaptic devices is one of the main challenges in
designing compute-in-memory (CIM) crossbar arrays for artificial neural networks (ANNs) …