Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Federated reinforcement learning: Linear speedup under markovian sampling
Since reinforcement learning algorithms are notoriously data-intensive, the task of sampling
observations from the environment is usually split across multiple agents. However …
observations from the environment is usually split across multiple agents. However …
The blessing of heterogeneity in federated q-learning: Linear speedup and beyond
In this paper, we consider federated Q-learning, which aims to learn an optimal Q-function
by periodically aggregating local Q-estimates trained on local data alone. Focusing on …
by periodically aggregating local Q-estimates trained on local data alone. Focusing on …
The sample-communication complexity trade-off in federated q-learning
We consider the problem of Federated Q-learning, where $ M $ agents aim to collaboratively
learn the optimal Q-function of an unknown infinite horizon Markov Decision Process with …
learn the optimal Q-function of an unknown infinite horizon Markov Decision Process with …
Distributed momentum-based Frank-Wolfe algorithm for stochastic optimization
This paper considers distributed stochastic optimization, in which a number of agents
cooperate to optimize a global objective function through local computations and information …
cooperate to optimize a global objective function through local computations and information …
Sample and communication-efficient decentralized actor-critic algorithms with finite-time analysis
Actor-critic (AC) algorithms have been widely used in decentralized multi-agent systems to
learn the optimal joint control policy. However, existing decentralized AC algorithms either …
learn the optimal joint control policy. However, existing decentralized AC algorithms either …
Federated q-learning: Linear regret speedup with low communication cost
In this paper, we consider federated reinforcement learning for tabular episodic Markov
Decision Processes (MDP) where, under the coordination of a central server, multiple …
Decision Processes (MDP) where, under the coordination of a central server, multiple …
Taming communication and sample complexities in decentralized policy evaluation for cooperative multi-agent reinforcement learning
Cooperative multi-agent reinforcement learning (MARL) has received increasing attention in
recent years and has found many scientific and engineering applications. However, a key …
recent years and has found many scientific and engineering applications. However, a key …
Federated Q-learning with reference-advantage decomposition: almost optimal regret and logarithmic communication cost
In this paper, we consider model-free federated reinforcement learning for tabular episodic
Markov decision processes. Under the coordination of a central server, multiple agents …
Markov decision processes. Under the coordination of a central server, multiple agents …
Central limit theorem for two-timescale stochastic approximation with markovian noise: Theory and applications
Two-timescale stochastic approximation (TTSA) is among the most general frameworks for
iterative stochastic algorithms. This includes well-known stochastic optimization methods …
iterative stochastic algorithms. This includes well-known stochastic optimization methods …
Distributed TD (0) with almost no communication
We provide a new non-asymptotic analysis of distributed temporal difference learning with
linear function approximation. Our approach relies on “one-shot averaging,” where N agents …
linear function approximation. Our approach relies on “one-shot averaging,” where N agents …