Towards a theory of non-log-concave sampling: first-order stationarity guarantees for Langevin Monte Carlo

K Balasubramanian, S Chewi… - … on Learning Theory, 2022 - proceedings.mlr.press
For the task of sampling from a density $\pi\propto\exp (-V) $ on $\R^ d $, where $ V $ is
possibly non-convex but $ L $-gradient Lipschitz, we prove that averaged Langevin Monte …

Randomized exploration in cooperative multi-agent reinforcement learning

HL Hsu, W Wang, M Pajic, P Xu - Advances in Neural …, 2025 - proceedings.neurips.cc
We present the first study on provably efficient randomized exploration in cooperative multi-
agent reinforcement learning (MARL). We propose a unified algorithm framework for …

Langevin monte carlo for contextual bandits

P Xu, H Zheng, EV Mazumdar… - International …, 2022 - proceedings.mlr.press
We study the efficiency of Thompson sampling for contextual bandits. Existing Thompson
sampling-based algorithms need to construct a Laplace approximation (ie, a Gaussian …

Reverse transition kernel: A flexible framework to accelerate diffusion inference

X Huang, D Zou, H Dong, Y Ma… - Advances in Neural …, 2025 - proceedings.neurips.cc
To generate data from trained diffusion models, most inference algorithms, such as DDPM,
DDIM, and other variants, rely on discretizing the reverse SDEs or their equivalent ODEs. In …

Provable and practical: Efficient exploration in reinforcement learning via langevin monte carlo

H Ishfaq, Q Lan, P Xu, AR Mahmood, D Precup… - arxiv preprint arxiv …, 2023 - arxiv.org
We present a scalable and effective exploration strategy based on Thompson sampling for
reinforcement learning (RL). One of the key shortcomings of existing Thompson sampling …

Differentiable annealed importance sampling and the perils of gradient noise

G Zhang, K Hsu, J Li, C Finn… - Advances in Neural …, 2021 - proceedings.neurips.cc
Annealed importance sampling (AIS) and related algorithms are highly effective tools for
marginal likelihood estimation, but are not fully differentiable due to the use of Metropolis …

Faster sampling without isoperimetry via diffusion-based Monte Carlo

X Huang, D Zou, H Dong, YA Ma… - The Thirty Seventh …, 2024 - proceedings.mlr.press
To sample from a general target distribution $ p_*\propto e^{-f_*} $ beyond the isoperimetric
condition, Huang et al.(2023) proposed to perform sampling through reverse diffusion …

Provably fast finite particle variants of SVGD via virtual particle stochastic approximation

A Das, D Nagaraj - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Abstract Stein Variational Gradient Descent (SVGD) is a popular particle-based variational
inference algorithm with impressive empirical performance across various domains …

Improved convergence rate of stochastic gradient langevin dynamics with variance reduction and its application to optimization

Y Kinoshita, T Suzuki - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Abstract The stochastic gradient Langevin Dynamics is one of the most fundamental
algorithms to solve sampling problems and non-convex optimization appearing in several …

Fisher information lower bounds for sampling

S Chewi, P Gerber, H Lee, C Lu - … Conference on Algorithmic …, 2023 - proceedings.mlr.press
We prove two lower bounds for the complexity of non-log-concave sampling within the
framework of Balasubramanian et al.(2022), who introduced the use of Fisher information …