When can we learn general-sum Markov games with a large number of players sample-efficiently?

Z Song, S Mei, Y Bai - arxiv preprint arxiv:2110.04184, 2021 - arxiv.org
Multi-agent reinforcement learning has made substantial empirical progresses in solving
games with a large number of players. However, theoretically, the best known sample …

The complexity of markov equilibrium in stochastic games

C Daskalakis, N Golowich… - The Thirty Sixth Annual …, 2023 - proceedings.mlr.press
We show that computing approximate stationary Markov coarse correlated equilibria (CCE)
in general-sum stochastic games is PPAD-hard, even when there are two players, the game …

Independent policy gradient for large-scale markov potential games: Sharper rates, function approximation, and game-agnostic convergence

D Ding, CY Wei, K Zhang… - … Conference on Machine …, 2022 - proceedings.mlr.press
We examine global non-asymptotic convergence properties of policy gradient methods for
multi-agent reinforcement learning (RL) problems in Markov potential games (MPGs). To …

Model-based multi-agent rl in zero-sum markov games with near-optimal sample complexity

K Zhang, SM Kakade, T Basar, LF Yang - Journal of Machine Learning …, 2023 - jmlr.org
Model-based reinforcement learning (RL), which finds an optimal policy after establishing an
empirical model, has long been recognized as one of the cornerstones of RL. It is especially …

On improving model-free algorithms for decentralized multi-agent reinforcement learning

W Mao, L Yang, K Zhang… - … Conference on Machine …, 2022 - proceedings.mlr.press
Multi-agent reinforcement learning (MARL) algorithms often suffer from an exponential
sample complexity dependence on the number of agents, a phenomenon known as the …

Breaking the curse of multiagency: Provably efficient decentralized multi-agent rl with function approximation

Y Wang, Q Liu, Y Bai, C ** - The Thirty Sixth Annual …, 2023 - proceedings.mlr.press
A unique challenge in Multi-Agent Reinforcement Learning (MARL) is the\emph {curse of
multiagency}, where the description length of the game as well as the complexity of many …

Breaking the curse of multiagents in a large state space: Rl in markov games with independent linear function approximation

Q Cui, K Zhang, S Du - The Thirty Sixth Annual Conference …, 2023 - proceedings.mlr.press
We propose a new model,\emph {independent linear Markov game}, for multi-agent
reinforcement learning with a large state space and a large number of agents. This is a class …

On last-iterate convergence beyond zero-sum games

I Anagnostides, I Panageas, G Farina… - International …, 2022 - proceedings.mlr.press
Most existing results about last-iterate convergence of learning dynamics are limited to two-
player zero-sum games, and only apply under rigid assumptions about what dynamics the …

Independent natural policy gradient always converges in markov potential games

R Fox, SM Mcaleer, W Overman… - International …, 2022 - proceedings.mlr.press
Natural policy gradient has emerged as one of the most successful algorithms for computing
optimal policies in challenging Reinforcement Learning (RL) tasks, yet, very little was known …

Revisiting some common practices in cooperative multi-agent reinforcement learning

W Fu, C Yu, Z Xu, J Yang, Y Wu - arxiv preprint arxiv:2206.07505, 2022 - arxiv.org
Many advances in cooperative multi-agent reinforcement learning (MARL) are based on two
common design principles: value decomposition and parameter sharing. A typical MARL …