Amortizing intractable inference in diffusion models for vision, language, and control

S Venkatraman, M Jain, L Scimeca, M Kim… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion models have emerged as effective distribution estimators in vision, language, and
reinforcement learning, but their use as priors in downstream tasks poses an intractable …

Beyond ELBOs: A large-scale evaluation of variational methods for sampling

D Blessing, X Jia, J Esslinger, F Vargas… - arxiv preprint arxiv …, 2024 - arxiv.org
Monte Carlo methods, Variational Inference, and their combinations play a pivotal role in
sampling from intractable probability distributions. However, current studies lack a unified …

Steering masked discrete diffusion models via discrete denoising posterior prediction

J Rector-Brooks, M Hasan, Z Peng, Z Quinn… - arxiv preprint arxiv …, 2024 - arxiv.org
Generative modeling of discrete data underlies important applications spanning text-based
agents like ChatGPT to the design of the very building blocks of life in protein sequences …

Sequential controlled langevin diffusions

J Chen, L Richter, J Berner, D Blessing… - arxiv preprint arxiv …, 2024 - arxiv.org
An effective approach for sampling from unnormalized densities is based on the idea of
gradually transporting samples from an easy prior to the complicated target distribution. Two …

From discrete-time policies to continuous-time diffusion samplers: Asymptotic equivalences and faster training

J Berner, L Richter, M Sendera, J Rector-Brooks… - arxiv preprint arxiv …, 2025 - arxiv.org
We study the problem of training neural stochastic differential equations, or diffusion models,
to sample from a Boltzmann distribution without access to target samples. Existing methods …

[PDF][PDF] Flow of reasoning: Efficient training of llm policy with divergent thinking

F Yu, L Jiang, H Kang, S Hao, L Qin - arxiv preprint arxiv …, 2024 - academia.edu
Divergent thinking, the cognitive process of generating diverse solutions, is a hallmark of
human creativity and problem-solving. For machines, sampling diverse solution trajectories …

Pessimistic backward policy for gflownets

H Jang, Y Jang, M Kim, J Park, S Ahn - arxiv preprint arxiv:2405.16012, 2024 - arxiv.org
This paper studies Generative Flow Networks (GFlowNets), which learn to sample objects
proportionally to a given reward function through the trajectory of state transitions. In this …

Can a Bayesian Oracle Prevent Harm from an Agent?

Y Bengio, MK Cohen, N Malkin, M MacDermott… - arxiv preprint arxiv …, 2024 - arxiv.org
Is there a way to design powerful AI systems based on machine learning methods that would
satisfy probabilistic safety guarantees? With the long-term goal of obtaining a probabilistic …

Adaptive teachers for amortized samplers

M Kim, S Choi, T Yun, E Bengio, L Feng… - arxiv preprint arxiv …, 2024 - arxiv.org
Amortized inference is the task of training a parametric model, such as a neural network, to
approximate a distribution with a given unnormalized density where exact sampling is …

Streaming Bayes GFlowNets

T Silva, DA de Souza… - Advances in Neural …, 2025 - proceedings.neurips.cc
Bayes' rule naturally allows for inference refinement in a streaming fashion, without the need
to recompute posteriors from scratch whenever new data arrives. In principle, Bayesian …