Stochastic policy gradient methods: Improved sample complexity for fisher-non-degenerate policies
Recently, the impressive empirical success of policy gradient (PG) methods has catalyzed
the development of their theoretical foundations. Despite the huge efforts directed at the …
the development of their theoretical foundations. Despite the huge efforts directed at the …
Momentum and stochastic momentum for stochastic gradient, newton, proximal point and subspace descent methods
In this paper we study several classes of stochastic optimization algorithms enriched with
heavy ball momentum. Among the methods studied are: stochastic gradient descent …
heavy ball momentum. Among the methods studied are: stochastic gradient descent …
Momentum improves normalized sgd
We provide an improved analysis of normalized SGD showing that adding momentum
provably removes the need for large batch sizes on non-convex objectives. Then, we …
provably removes the need for large batch sizes on non-convex objectives. Then, we …
High-probability bounds for non-convex stochastic optimization with heavy tails
We consider non-convex stochastic optimization using first-order algorithms for which the
gradient estimates may have heavy tails. We show that a combination of gradient clip** …
gradient estimates may have heavy tails. We show that a combination of gradient clip** …
The marginal value of momentum for small learning rate sgd
Momentum is known to accelerate the convergence of gradient descent in strongly convex
settings without stochastic gradient noise. In stochastic optimization, such as training neural …
settings without stochastic gradient noise. In stochastic optimization, such as training neural …
Momentum ensures convergence of signsgd under weaker assumptions
Abstract Sign Stochastic Gradient Descent (signSGD) is a communication-efficient stochastic
algorithm that only uses the sign information of the stochastic gradient to update the model's …
algorithm that only uses the sign information of the stochastic gradient to update the model's …
Meta learning on a sequence of imbalanced domains with difficulty awareness
Recognizing new objects by learning from a few labeled examples in an evolving
environment is crucial to obtain excellent generalization ability for real-world machine …
environment is crucial to obtain excellent generalization ability for real-world machine …
Online optimization over riemannian manifolds
Online optimization has witnessed a massive surge of research attention in recent years. In
this paper, we propose online gradient descent and online bandit algorithms over …
this paper, we propose online gradient descent and online bandit algorithms over …
Riemannian optimistic algorithms
In this paper, we consider Riemannian online convex optimization with dynamic regret. First,
we propose two novel algorithms, namely the Riemannian Online Optimistic Gradient …
we propose two novel algorithms, namely the Riemannian Online Optimistic Gradient …
Root-sgd: Sharp nonasymptotics and asymptotic efficiency in a single algorithm
We study the problem of solving strongly convex and smooth unconstrained optimization
problems using stochastic first-order algorithms. We devise a novel algorithm, referred to …
problems using stochastic first-order algorithms. We devise a novel algorithm, referred to …