Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Head2toe: Utilizing intermediate representations for better transfer learning

U Evci, V Dumoulin, H Larochelle… - … on Machine Learning, 2022 - proceedings.mlr.press
Transfer-learning methods aim to improve performance in a data-scarce target domain using
a model pretrained on a data-rich source domain. A cost-efficient strategy, linear probing …

Sparse training via boosting pruning plasticity with neuroregeneration

S Liu, T Chen, X Chen, Z Atashgahi… - Advances in …, 2021 - proceedings.neurips.cc
Works on lottery ticket hypothesis (LTH) and single-shot network pruning (SNIP) have raised
a lot of attention currently on post-training pruning (iterative magnitude pruning), and before …

Do we actually need dense over-parameterization? in-time over-parameterization in sparse training

S Liu, L Yin, DC Mocanu… - … on Machine Learning, 2021 - proceedings.mlr.press
In this paper, we introduce a new perspective on training deep neural networks capable of
state-of-the-art performance without the need for the expensive over-parameterization by …

Dynamic sparse training for deep reinforcement learning

G Sokar, E Mocanu, DC Mocanu, M Pechenizkiy… - arXiv preprint arXiv …, 2021 - arxiv.org
Deep reinforcement learning (DRL) agents are trained through trial-and-error interactions
with the environment. This leads to a long training time for dense neural networks to achieve …

Ten lessons we have learned in the new" sparseland": A short handbook for sparse neural network researchers

S Liu, Z Wang - arXiv preprint arXiv:2302.02596, 2023 - arxiv.org
This article does not propose any novel algorithm or new hardware for sparsity. Instead, it
aims to serve the" common good" for the increasingly prosperous Sparse Neural Network …

Sparse training theory for scalable and efficient agents

DC Mocanu, E Mocanu, T Pinto, S Curci… - arXiv preprint arXiv …, 2021 - arxiv.org
A fundamental task for artificial intelligence is learning. Deep Neural Networks have proven
to cope perfectly with all learning paradigms, ie supervised, unsupervised, and …

[HTML][HTML] Quick and robust feature selection: the strength of energy-efficient sparse training for autoencoders

Z Atashgahi, G Sokar, T van der Lee, E Mocanu… - Machine Learning, 2022 - Springer
Major complications arise from the recent increase in the amount of high-dimensional data,
including high computational costs and memory requirements. Feature selection, which …

Supervised Feature Selection via Ensemble Gradient Information from Sparse Neural Networks

K Liu, Z Atashgahi, G Sokar… - International …, 2024 - proceedings.mlr.press
Feature selection algorithms aim to select a subset of informative features from a dataset to
reduce the data dimensionality, consequently saving resource consumption and improving …

Don't be so dense: Sparse-to-sparse gan training without sacrificing performance

S Liu, Y Tian, T Chen, L Shen - International Journal of Computer Vision, 2023 - Springer
This paper does not describe a novel method. Instead, it studies an incremental, yet must-
know baseline given the recent progress in sparse neural network training and Generative …