A zeroth-order block coordinate descent algorithm for huge-scale black-box optimization

HQ Cai, Y Lou, D McKenzie… - … Conference on Machine …, 2021 - proceedings.mlr.press
We consider the zeroth-order optimization problem in the huge-scale setting, where the
dimension of the problem is so large that performing even basic vector operations on the …

Zeroth-order optimization meets human feedback: Provable learning via ranking oracles

Z Tang, D Rybin, TH Chang - arxiv preprint arxiv:2303.03751, 2023 - arxiv.org
In this study, we delve into an emerging optimization challenge involving a black-box
objective function that can only be gauged via a ranking oracle-a situation frequently …

Prompt-tuning decision transformer with preference ranking

S Hu, L Shen, Y Zhang, D Tao - arxiv preprint arxiv:2305.09648, 2023 - arxiv.org
Prompt-tuning has emerged as a promising method for adapting pre-trained models to
downstream tasks or aligning with human preferences. Prompt learning is widely used in …

A Hamilton–Jacobi-based proximal operator

S Osher, H Heaton, S Wu Fung - Proceedings of the …, 2023 - National Acad Sciences
First-order optimization algorithms are widely used today. Two standard building blocks in
these algorithms are proximal operators (proximals) and gradients. Although gradients can …

Zeroth-order regularized optimization (zoro): Approximately sparse gradients and adaptive sampling

HQ Cai, D McKenzie, W Yin, Z Zhang - SIAM Journal on Optimization, 2022 - SIAM
We consider the problem of minimizing a high-dimensional objective function, which may
include a regularization term, using only (possibly noisy) evaluations of the function. Such …

Stochastic zeroth-order gradient and Hessian estimators: variance reduction and refined bias bounds

Y Feng, T Wang - Information and Inference: A Journal of the …, 2023 - academic.oup.com
We study stochastic zeroth-order gradient and Hessian estimators for real-valued functions
in. We show that, via taking finite difference along random orthogonal directions, the …

Global solutions to nonconvex problems by evolution of Hamilton-Jacobi PDEs

H Heaton, S Wu Fung, S Osher - Communications on Applied …, 2024 - Springer
Computing tasks may often be posed as optimization problems. The objective functions for
real-world scenarios are often nonconvex and/or nondifferentiable. State-of-the-art methods …

Curvature-aware derivative-free optimization

B Kim, HQ Cai, D McKenzie, W Yin - arxiv preprint arxiv:2109.13391, 2021 - arxiv.org
The paper discusses derivative-free optimization (DFO), which involves minimizing a
function without access to gradients or directional derivatives, only function evaluations …

Stochastic zeroth order descent with structured directions

M Rando, C Molinari, S Villa, L Rosasco - Computational Optimization and …, 2024 - Springer
We introduce and analyze Structured Stochastic Zeroth order Descent (S-SZD), a finite
difference approach that approximates a stochastic gradient on a set of l≤ d orthogonal …

Sequential stochastic blackbox optimization with zeroth-order gradient estimators

C Audet, J Bigeon, R Couderc… - arxiv preprint arxiv …, 2023 - arxiv.org
This work considers stochastic optimization problems in which the objective function values
can only be computed by a blackbox corrupted by some random noise following an …