Stochastic model-based minimization of weakly convex functions
We consider a family of algorithms that successively sample and minimize simple stochastic
models of the objective function. We show that under reasonable conditions on …
models of the objective function. We show that under reasonable conditions on …
Weakly-convex–concave min–max optimization: provable algorithms and applications in machine learning
Min–max problems have broad applications in machine learning, including learning with
non-decomposable loss and learning with robustness to data distribution. Convex–concave …
non-decomposable loss and learning with robustness to data distribution. Convex–concave …
Stochastic methods for composite and weakly convex optimization problems
We consider minimization of stochastic functionals that are compositions of a (potentially)
nonsmooth convex function h and smooth function c and, more generally, stochastic weakly …
nonsmooth convex function h and smooth function c and, more generally, stochastic weakly …
Biased stochastic first-order methods for conditional stochastic optimization and applications in meta learning
Conditional stochastic optimization covers a variety of applications ranging from invariant
learning and causal inference to meta-learning. However, constructing unbiased gradient …
learning and causal inference to meta-learning. However, constructing unbiased gradient …
The importance of better models in stochastic optimization
Standard stochastic optimization methods are brittle, sensitive to stepsize choice and other
algorithmic parameters, and they exhibit instability outside of well-behaved families of …
algorithmic parameters, and they exhibit instability outside of well-behaved families of …
Subgradient methods for sharp weakly convex functions
Subgradient methods converge linearly on a convex function that grows sharply away from
its solution set. In this work, we show that the same is true for sharp functions that are only …
its solution set. In this work, we show that the same is true for sharp functions that are only …
Weakly convex optimization over Stiefel manifold using Riemannian subgradient-type methods
We consider a class of nonsmooth optimization problems over the Stiefel manifold, in which
the objective function is weakly convex in the ambient Euclidean space. Such problems are …
the objective function is weakly convex in the ambient Euclidean space. Such problems are …
Stochastic subgradient method converges at the rate on weakly convex functions
We prove that the proximal stochastic subgradient method, applied to a weakly convex
problem, drives the gradient of the Moreau envelope to zero at the rate $ O (k^{-1/4}) $. As a …
problem, drives the gradient of the Moreau envelope to zero at the rate $ O (k^{-1/4}) $. As a …
Convergence of a stochastic gradient method with momentum for non-smooth non-convex optimization
V Mai, M Johansson - International conference on machine …, 2020 - proceedings.mlr.press
Stochastic gradient methods with momentum are widely used in applications and at the core
of optimization subroutines in many popular machine learning libraries. However, their …
of optimization subroutines in many popular machine learning libraries. However, their …
Stochastic bias-reduced gradient methods
We develop a new primitive for stochastic optimization: a low-bias, low-cost estimator of the
minimizer $ x_\star $ of any Lipschitz strongly-convex function $ f $. In particular, we use a …
minimizer $ x_\star $ of any Lipschitz strongly-convex function $ f $. In particular, we use a …