Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Towards a theory of non-log-concave sampling: first-order stationarity guarantees for Langevin Monte Carlo
For the task of sampling from a density $\pi\propto\exp (-V) $ on $\R^ d $, where $ V $ is
possibly non-convex but $ L $-gradient Lipschitz, we prove that averaged Langevin Monte …
possibly non-convex but $ L $-gradient Lipschitz, we prove that averaged Langevin Monte …
Randomized exploration in cooperative multi-agent reinforcement learning
We present the first study on provably efficient randomized exploration in cooperative multi-
agent reinforcement learning (MARL). We propose a unified algorithm framework for …
agent reinforcement learning (MARL). We propose a unified algorithm framework for …
Langevin monte carlo for contextual bandits
We study the efficiency of Thompson sampling for contextual bandits. Existing Thompson
sampling-based algorithms need to construct a Laplace approximation (ie, a Gaussian …
sampling-based algorithms need to construct a Laplace approximation (ie, a Gaussian …
Reverse transition kernel: A flexible framework to accelerate diffusion inference
To generate data from trained diffusion models, most inference algorithms, such as DDPM,
DDIM, and other variants, rely on discretizing the reverse SDEs or their equivalent ODEs. In …
DDIM, and other variants, rely on discretizing the reverse SDEs or their equivalent ODEs. In …
Provable and practical: Efficient exploration in reinforcement learning via langevin monte carlo
We present a scalable and effective exploration strategy based on Thompson sampling for
reinforcement learning (RL). One of the key shortcomings of existing Thompson sampling …
reinforcement learning (RL). One of the key shortcomings of existing Thompson sampling …
Differentiable annealed importance sampling and the perils of gradient noise
Annealed importance sampling (AIS) and related algorithms are highly effective tools for
marginal likelihood estimation, but are not fully differentiable due to the use of Metropolis …
marginal likelihood estimation, but are not fully differentiable due to the use of Metropolis …
Faster sampling without isoperimetry via diffusion-based Monte Carlo
To sample from a general target distribution $ p_*\propto e^{-f_*} $ beyond the isoperimetric
condition, Huang et al.(2023) proposed to perform sampling through reverse diffusion …
condition, Huang et al.(2023) proposed to perform sampling through reverse diffusion …
Provably fast finite particle variants of SVGD via virtual particle stochastic approximation
Abstract Stein Variational Gradient Descent (SVGD) is a popular particle-based variational
inference algorithm with impressive empirical performance across various domains …
inference algorithm with impressive empirical performance across various domains …
Improved convergence rate of stochastic gradient langevin dynamics with variance reduction and its application to optimization
Abstract The stochastic gradient Langevin Dynamics is one of the most fundamental
algorithms to solve sampling problems and non-convex optimization appearing in several …
algorithms to solve sampling problems and non-convex optimization appearing in several …
Fisher information lower bounds for sampling
We prove two lower bounds for the complexity of non-log-concave sampling within the
framework of Balasubramanian et al.(2022), who introduced the use of Fisher information …
framework of Balasubramanian et al.(2022), who introduced the use of Fisher information …