Implicit learning dynamics in stackelberg games: Equilibria characterization, convergence analysis, and empirical study

T Fiez, B Chasnov, L Ratliff - International Conference on …, 2020 - proceedings.mlr.press
Contemporary work on learning in continuous games has commonly overlooked the
hierarchical decision-making structure present in machine learning problems formulated as …

Global convergence to local minmax equilibrium in classes of nonconvex zero-sum games

T Fiez, L Ratliff, E Mazumdar… - Advances in Neural …, 2021 - proceedings.neurips.cc
We study gradient descent-ascent learning dynamics with timescale separation ($\tau $-
GDA) in unconstrained continuous action zero-sum games where the minimizing player …

[PDF][PDF] Local convergence analysis of gradient descent ascent with finite timescale separation

T Fiez, LJ Ratliff - Proceedings of the International Conference on …, 2021 - par.nsf.gov
We study the role that a finite timescale separation parameter τ has on gradient descent-
ascent in non-convex, non-concave zero-sum games where the learning rate of player 1 is …

Policy-gradient algorithms have no guarantees of convergence in linear quadratic games

E Mazumdar, LJ Ratliff, MI Jordan, SS Sastry - arxiv preprint arxiv …, 2019 - arxiv.org
We show by counterexample that policy-gradient algorithms have no guarantees of even
local convergence to Nash equilibria in continuous action and state space multi-agent …

Solving min-max optimization with hidden structure via gradient descent ascent

EV Vlatakis-Gkaragkounis, L Flokas… - Advances in Neural …, 2021 - proceedings.neurips.cc
Many recent AI architectures are inspired by zero-sum games, however, the behavior of their
dynamics is still not well understood. Inspired by this, we study standard gradient descent …

Gradient descent-ascent provably converges to strict local minmax equilibria with a finite timescale separation

T Fiez, L Ratliff - arxiv preprint arxiv:2009.14820, 2020 - arxiv.org
We study the role that a finite timescale separation parameter $\tau $ has on gradient
descent-ascent in two-player non-convex, non-concave zero-sum games where the learning …

Generalized natural gradient flows in hidden convex-concave games and gans

A Mladenovic, I Sakos, G Gidel… - … Conference on Learning …, 2021 - openreview.net
Game-theoretic formulations in machine learning have recently risen in prominence,
whereby entire modeling paradigms are best captured as zero-sum games. Despite their …

Limiting behaviors of nonconvex-nonconcave minimax optimization via continuous-time systems

B Grimmer, H Lu, P Worah… - … on Algorithmic Learning …, 2022 - proceedings.mlr.press
Unlike nonconvex optimization, where gradient descent is guaranteed to converge to a local
optimizer, algorithms for nonconvex-nonconcave minimax optimization can have …

A note on large deviations for interacting particle dynamics for finding mixed equilibria in zero-sum games

V Nilsson, P Nyquist - arxiv preprint arxiv:2206.15177, 2022 - arxiv.org
Finding equilibria points in continuous minimax games has become a key problem within
machine learning, in part due to its connection to the training of generative adversarial …

[BOEK][B] Beyond Worst-Case Analysis of Optimization in the Era of Machine Learning

EV Vlatakis-Gkaragkounis - 2022 - search.proquest.com
Worst-case analysis (WCA) has been the dominant tool for understanding the performance
of the lion share of algorithmic arsenal of theoretical computer science. While WCA has …