Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Lower bounds and optimal algorithms for personalized federated learning
In this work, we consider the optimization formulation of personalized federated learning
recently introduced by Hanzely & Richtarik (2020) which was shown to give an alternative …
recently introduced by Hanzely & Richtarik (2020) which was shown to give an alternative …
Batched multi-armed bandits problem
In this paper, we study the multi-armed bandit problem in the batched setting where the
employed policy must split data into a small number of batches. While the minimax regret for …
employed policy must split data into a small number of batches. While the minimax regret for …
Provably efficient q-learning with low switching cost
We take initial steps in studying PAC-MDP algorithms with limited adaptivity, that is,
algorithms that change its exploration policy as infrequently as possible during regret …
algorithms that change its exploration policy as infrequently as possible during regret …
An algorithm with optimal dimension-dependence for zero-order nonsmooth nonconvex stochastic optimization
We study the complexity of producing $(\delta,\epsilon) $-stationary points of Lipschitz
objectives which are possibly neither smooth nor convex, using only noisy function …
objectives which are possibly neither smooth nor convex, using only noisy function …
Graph oracle models, lower bounds, and gaps for parallel stochastic optimization
We suggest a general oracle-based framework that captures parallel stochastic optimization
in different parallelization settings described by a dependency graph, and derive generic …
in different parallelization settings described by a dependency graph, and derive generic …
The min-max complexity of distributed stochastic convex optimization with intermittent communication
We resolve the min-max complexity of distributed stochastic convex optimization (up to a log
factor) in the intermittent communication setting, where $ M $ machines work in parallel over …
factor) in the intermittent communication setting, where $ M $ machines work in parallel over …
Linear bandits with limited adaptivity and learning distributional optimal design
Motivated by practical needs such as large-scale learning, we study the impact of adaptivity
constraints to linear contextual bandits, a central problem in online learning and decision …
constraints to linear contextual bandits, a central problem in online learning and decision …
Experimenting in equilibrium
Classical approaches to experimental design assume that intervening on one unit does not
affect other units. There are many important settings, however, where this noninterference …
affect other units. There are many important settings, however, where this noninterference …
Complexity of highly parallel non-smooth convex optimization
A landmark result of non-smooth convex optimization is that gradient descent is an optimal
algorithm whenever the number of computed gradients is smaller than the dimension $ d …
algorithm whenever the number of computed gradients is smaller than the dimension $ d …
Policy finetuning in reinforcement learning via design of experiments using offline data
In some applications of reinforcement learning, a dataset of pre-collected experience is
already availablebut it is also possible to acquire some additional online data to help …
already availablebut it is also possible to acquire some additional online data to help …