Lower bounds and optimal algorithms for personalized federated learning

F Hanzely, S Hanzely, S Horváth… - Advances in Neural …, 2020 - proceedings.neurips.cc
In this work, we consider the optimization formulation of personalized federated learning
recently introduced by Hanzely & Richtarik (2020) which was shown to give an alternative …

Batched multi-armed bandits problem

Z Gao, Y Han, Z Ren, Z Zhou - Advances in Neural …, 2019 - proceedings.neurips.cc
In this paper, we study the multi-armed bandit problem in the batched setting where the
employed policy must split data into a small number of batches. While the minimax regret for …

Provably efficient q-learning with low switching cost

Y Bai, T **e, N Jiang, YX Wang - Advances in Neural …, 2019 - proceedings.neurips.cc
We take initial steps in studying PAC-MDP algorithms with limited adaptivity, that is,
algorithms that change its exploration policy as infrequently as possible during regret …

An algorithm with optimal dimension-dependence for zero-order nonsmooth nonconvex stochastic optimization

G Kornowski, O Shamir - Journal of Machine Learning Research, 2024 - jmlr.org
We study the complexity of producing $(\delta,\epsilon) $-stationary points of Lipschitz
objectives which are possibly neither smooth nor convex, using only noisy function …

Graph oracle models, lower bounds, and gaps for parallel stochastic optimization

BE Woodworth, J Wang, A Smith… - Advances in neural …, 2018 - proceedings.neurips.cc
We suggest a general oracle-based framework that captures parallel stochastic optimization
in different parallelization settings described by a dependency graph, and derive generic …

The min-max complexity of distributed stochastic convex optimization with intermittent communication

BE Woodworth, B Bullins, O Shamir… - … on Learning Theory, 2021 - proceedings.mlr.press
We resolve the min-max complexity of distributed stochastic convex optimization (up to a log
factor) in the intermittent communication setting, where $ M $ machines work in parallel over …

Linear bandits with limited adaptivity and learning distributional optimal design

Y Ruan, J Yang, Y Zhou - Proceedings of the 53rd Annual ACM SIGACT …, 2021 - dl.acm.org
Motivated by practical needs such as large-scale learning, we study the impact of adaptivity
constraints to linear contextual bandits, a central problem in online learning and decision …

Experimenting in equilibrium

S Wager, K Xu - Management Science, 2021 - pubsonline.informs.org
Classical approaches to experimental design assume that intervening on one unit does not
affect other units. There are many important settings, however, where this noninterference …

Complexity of highly parallel non-smooth convex optimization

S Bubeck, Q Jiang, YT Lee, Y Li… - Advances in neural …, 2019 - proceedings.neurips.cc
A landmark result of non-smooth convex optimization is that gradient descent is an optimal
algorithm whenever the number of computed gradients is smaller than the dimension $ d …

Policy finetuning in reinforcement learning via design of experiments using offline data

R Zhang, A Zanette - Advances in Neural Information …, 2023 - proceedings.neurips.cc
In some applications of reinforcement learning, a dataset of pre-collected experience is
already availablebut it is also possible to acquire some additional online data to help …