A survey of distributed optimization
In distributed optimization of multi-agent systems, agents cooperate to minimize a global
function which is a sum of local objective functions. Motivated by applications including …
function which is a sum of local objective functions. Motivated by applications including …
A review of nonlinear FFT-based computational homogenization methods
M Schneider - Acta Mechanica, 2021 - Springer
Since their inception, computational homogenization methods based on the fast Fourier
transform (FFT) have grown in popularity, establishing themselves as a powerful tool …
transform (FFT) have grown in popularity, establishing themselves as a powerful tool …
Lookahead optimizer: k steps forward, 1 step back
The vast majority of successful deep neural networks are trained using variants of stochastic
gradient descent (SGD) algorithms. Recent attempts to improve SGD can be broadly …
gradient descent (SGD) algorithms. Recent attempts to improve SGD can be broadly …
Train faster, generalize better: Stability of stochastic gradient descent
We show that parametric models trained by a stochastic gradient method (SGM) with few
iterations have vanishing generalization error. We prove our results by arguing that SGM is …
iterations have vanishing generalization error. We prove our results by arguing that SGM is …
An introduction to continuous optimization for imaging
A large number of imaging problems reduce to the optimization of a cost function, with
typical structural properties. The aim of this paper is to describe the state of the art in …
typical structural properties. The aim of this paper is to describe the state of the art in …
A differential equation for modeling Nesterov's accelerated gradient method: Theory and insights
We derive a second-order ordinary differential equation (ODE) which is the limit of
Nesterov's accelerated gradient method. This ODE exhibits approximate equivalence to …
Nesterov's accelerated gradient method. This ODE exhibits approximate equivalence to …
A variational perspective on accelerated methods in optimization
Accelerated gradient methods play a central role in optimization, achieving optimal rates in
many settings. Although many generalizations and extensions of Nesterov's original …
many settings. Although many generalizations and extensions of Nesterov's original …
Understanding the acceleration phenomenon via high-resolution differential equations
Gradient-based optimization algorithms can be studied from the perspective of limiting
ordinary differential equations (ODEs). Motivated by the fact that existing ODEs do not …
ordinary differential equations (ODEs). Motivated by the fact that existing ODEs do not …
Federated accelerated stochastic gradient descent
Abstract We propose Federated Accelerated Stochastic Gradient Descent (FedAc), a
principled acceleration of Federated Averaging (FedAvg, also known as Local SGD) for …
principled acceleration of Federated Averaging (FedAvg, also known as Local SGD) for …
The limit points of (optimistic) gradient descent in min-max optimization
Motivated by applications in Optimization, Game Theory, and the training of Generative
Adversarial Networks, the convergence properties of first order methods in min-max …
Adversarial Networks, the convergence properties of first order methods in min-max …