A primer on zeroth-order optimization in signal processing and machine learning: Principals, recent advances, and applications

S Liu, PY Chen, B Kailkhura, G Zhang… - IEEE Signal …, 2020 - ieeexplore.ieee.org
Zeroth-order (ZO) optimization is a subset of gradient-free optimization that emerges in many
signal processing and machine learning (ML) applications. It is used for solving optimization …

Fine-tuning language models with just forward passes

S Malladi, T Gao, E Nichani… - Advances in …, 2023 - proceedings.neurips.cc
Fine-tuning language models (LMs) has yielded success on diverse downstream tasks, but
as LMs grow in size, backpropagation requires a prohibitively large amount of memory …

Hopskipjumpattack: A query-efficient decision-based attack

J Chen, MI Jordan… - 2020 ieee symposium on …, 2020 - ieeexplore.ieee.org
The goal of a decision-based adversarial attack on a trained model is to generate
adversarial examples based solely on observing output labels returned by the targeted …

Derivative-free optimization methods

J Larson, M Menickelly, SM Wild - Acta Numerica, 2019 - cambridge.org
In many optimization problems arising from scientific, engineering and artificial intelligence
applications, objective and constraint functions are available only as the output of a black …

Blackvip: Black-box visual prompting for robust transfer learning

C Oh, H Hwang, H Lee, YT Lim… - Proceedings of the …, 2023 - openaccess.thecvf.com
With the surge of large-scale pre-trained models (PTMs), fine-tuning these models to
numerous downstream tasks becomes a crucial problem. Consequently, parameter efficient …

Gradient-free methods for deterministic and stochastic nonsmooth nonconvex optimization

T Lin, Z Zheng, M Jordan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Nonsmooth nonconvex optimization problems broadly emerge in machine learning and
business decision making, whereas two core challenges impede the development of …

A guide through the zoo of biased SGD

Y Demidovich, G Malinovsky… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Stochastic Gradient Descent (SGD) is arguably the most important single algorithm
in modern machine learning. Although SGD with unbiased gradient estimators has been …

A theoretical and empirical comparison of gradient approximations in derivative-free optimization

AS Berahas, L Cao, K Choromanski… - Foundations of …, 2022 - Springer
In this paper, we analyze several methods for approximating gradients of noisy functions
using only function values. These methods include finite differences, linear interpolation …

Transfer learning without knowing: Reprogramming black-box machine learning models with scarce data and limited resources

YY Tsai, PY Chen, TY Ho - International Conference on …, 2020 - proceedings.mlr.press
Current transfer learning methods are mainly based on finetuning a pretrained model with
target-domain data. Motivated by the techniques from adversarial machine learning (ML) …

Zo-adamm: Zeroth-order adaptive momentum method for black-box optimization

X Chen, S Liu, K Xu, X Li, X Lin… - Advances in neural …, 2019 - proceedings.neurips.cc
The adaptive momentum method (AdaMM), which uses past gradients to update descent
directions and learning rates simultaneously, has become one of the most popular first-order …