Spanning training progress: Temporal dual-depth scoring (tdds) for enhanced dataset pruning

X Zhang, J Du, Y Li, W Xie… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Dataset pruning aims to construct a coreset capable of achieving performance comparable
to the original full dataset. Most existing dataset pruning methods rely on snapshot-based …

HyperSparse Neural Networks: Shifting Exploration to Exploitation through Adaptive Regularization

P Glandorf, T Kaiser… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Sparse neural networks are a key factor in developing resource-efficient machine learning
applications. We propose the novel and powerful sparse learning method Adaptive …

Can we achieve robustness from data alone?

N Tsilivis, J Su, J Kempe - arXiv preprint arXiv:2207.11727, 2022 - arxiv.org
We introduce a meta-learning algorithm for adversarially robust classification. The proposed
method tries to be as model agnostic as possible and optimizes a dataset prior to its …

Robust active distillation

C Baykal, K Trinh, F Iliopoulos, G Menghani… - arXiv preprint arXiv …, 2022 - arxiv.org
Distilling knowledge from a large teacher model to a lightweight one is a widely successful
approach for generating compact, powerful models in the semi-supervised learning setting …

Simultaneous linear connectivity of neural networks modulo permutation

E Sharma, D Kwok, T Denton, DM Roy… - … Conference on Machine …, 2024 - Springer
Neural networks typically exhibit permutation symmetry, as reordering neurons in each layer
does not change the underlying function they compute. These symmetries contribute to the …

DeepMem: ML Models as storage channels and their (mis-) applications

MAA Mamun, QM Alam, E Shaigani, P Zaree… - arXiv preprint arXiv …, 2023 - arxiv.org
Machine learning (ML) models are overparameterized to support generality and avoid
overfitting. Prior works have shown that these additional parameters can be used for both …

Data pruning for efficient model pruning in neural machine translation

A Azeemi, I Qazi, A Raza - Findings of the Association for …, 2023 - aclanthology.org
Abstract Model pruning methods reduce memory requirements and inference time of large-
scale pre-trained language models after deployment. However, the actual pruning …

When Layers Play the Lottery, all Tickets Win at Initialization

A Jordao, G de Araújo… - Proceedings of the …, 2023 - openaccess.thecvf.com
Pruning is a standard technique for reducing the computational cost of deep networks. Many
advances in pruning leverage concepts from the Lottery Ticket Hypothesis (LTH). LTH …

On the special role of class-selective neurons in early training

O Ranadive, N Thakurdesai, AS Morcos… - arXiv preprint arXiv …, 2023 - arxiv.org
It is commonly observed that deep networks trained for classification exhibit class-selective
neurons in their early and intermediate layers. Intriguingly, recent studies have shown that …

Effective Subset Selection Through The Lens of Neural Network Pruning

N Bar, R Giryes - arXiv preprint arXiv:2406.01086, 2024 - arxiv.org
Having large amounts of annotated data significantly impacts the effectiveness of deep
neural networks. However, the annotation task can be very expensive in some domains …