Task-specific skill localization in fine-tuned language models

A Panigrahi, N Saunshi, H Zhao… - … on Machine Learning, 2023 - proceedings.mlr.press
Pre-trained language models can be fine-tuned to solve diverse NLP tasks, including in few-
shot settings. Thus fine-tuning allows the model to quickly pick up task-specific" skills," but …

Rankfeat: Rank-1 feature removal for out-of-distribution detection

Y Song, N Sebe, W Wang - Advances in Neural Information …, 2022 - proceedings.neurips.cc
The task of out-of-distribution (OOD) detection is crucial for deploying machine learning
models in real-world settings. In this paper, we observe that the singular value distributions …

Advancing model pruning via bi-level optimization

Y Zhang, Y Yao, P Ram, P Zhao… - Advances in …, 2022 - proceedings.neurips.cc
The deployment constraints in practical applications necessitate the pruning of large-scale
deep learning models, ie, promoting their weight sparsity. As illustrated by the Lottery Ticket …

Model sparsity can simplify machine unlearning

J Liu, P Ram, Y Yao, G Liu, Y Liu… - Advances in Neural …, 2024 - proceedings.neurips.cc
In response to recent data regulation requirements, machine unlearning (MU) has emerged
as a critical process to remove the influence of specific examples from a given model …

Compute-efficient deep learning: Algorithmic trends and opportunities

BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …

Improving robustness of vision transformers by reducing sensitivity to patch corruptions

Y Guo, D Stutz, B Schiele - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Despite their success, vision transformers still remain vulnerable to image corruptions, such
as noise or blur. Indeed, we find that the vulnerability mainly stems from the unstable self …

Prime: A few primitives can boost robustness to common corruptions

A Modas, R Rade, G Ortiz-Jiménez… - … on Computer Vision, 2022 - Springer
Despite their impressive performance on image classification tasks, deep networks have a
hard time generalizing to unforeseen corruptions of their data. To fix this vulnerability, prior …

Defending against image corruptions through adversarial augmentations

DA Calian, F Stimberg, O Wiles, SA Rebuffi… - arXiv preprint arXiv …, 2021 - arxiv.org
Modern neural networks excel at image classification, yet they remain vulnerable to common
image corruptions such as blur, speckle noise or fog. Recent methods that focus on this …

Benchmark generation framework with customizable distortions for image classifier robustness

S Sarkar, AR Babu, S Mousavi… - Proceedings of the …, 2024 - openaccess.thecvf.com
We present a novel framework for generating adversarial benchmarks to evaluate the
robustness of image classification models. The RLAB framework allows users to customize …

Dimensionality reduced training by pruning and freezing parts of a deep neural network: a survey

P Wimmer, J Mehnert, AP Condurache - Artificial Intelligence Review, 2023 - Springer
State-of-the-art deep learning models have a parameter count that reaches into the billions.
Training, storing and transferring such models is energy and time consuming, thus costly. A …