Dynamical variational autoencoders: A comprehensive review

L Girin, S Leglaive, X Bie, J Diard, T Hueber… - arxiv preprint arxiv …, 2020 - arxiv.org
Variational autoencoders (VAEs) are powerful deep generative models widely used to
represent high-dimensional complex data through a low-dimensional latent space learned …

A survey of unsupervised generative models for exploratory data analysis and representation learning

M Abukmeil, S Ferrari, A Genovese, V Piuri… - Acm computing surveys …, 2021 - dl.acm.org
For more than a century, the methods for data representation and the exploration of the
intrinsic structures of data have developed remarkably and consist of supervised and …

Diffusion-based generation, optimization, and planning in 3d scenes

S Huang, Z Wang, P Li, B Jia, T Liu… - Proceedings of the …, 2023 - openaccess.thecvf.com
We introduce SceneDiffuser, a conditional generative model for 3D scene understanding.
SceneDiffuser provides a unified model for solving scene-conditioned generation …

Versatile diffusion: Text, images and variations all in one diffusion model

X Xu, Z Wang, G Zhang, K Wang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recent advances in diffusion models have set an impressive milestone in many generation
tasks, and trending works such as DALL-E2, Imagen, and Stable Diffusion have attracted …

Cogview: Mastering text-to-image generation via transformers

M Ding, Z Yang, W Hong, W Zheng… - Advances in neural …, 2021 - proceedings.neurips.cc
Text-to-Image generation in the general domain has long been an open problem, which
requires both a powerful generative model and cross-modal understanding. We propose …

Implicit generation and modeling with energy based models

Y Du, I Mordatch - Advances in Neural Information …, 2019 - proceedings.neurips.cc
Energy based models (EBMs) are appealing due to their generality and simplicity in
likelihood modeling, but have been traditionally difficult to train. We present techniques to …

Cyclical annealing schedule: A simple approach to mitigating kl vanishing

H Fu, C Li, X Liu, J Gao, A Celikyilmaz… - arxiv preprint arxiv …, 2019 - arxiv.org
Variational autoencoders (VAEs) with an auto-regressive decoder have been applied for
many natural language processing (NLP) tasks. The VAE objective consists of two terms,(i) …

Learning generative vision transformer with energy-based latent space for saliency prediction

J Zhang, J **e, N Barnes, P Li - Advances in Neural …, 2021 - proceedings.neurips.cc
Vision transformer networks have shown superiority in many computer vision tasks. In this
paper, we take a step further by proposing a novel generative vision transformer with latent …

Dlow: Diversifying latent flows for diverse human motion prediction

Y Yuan, K Kitani - Computer Vision–ECCV 2020: 16th European …, 2020 - Springer
Deep generative models are often used for human motion prediction as they are able to
model multi-modal data distributions and characterize diverse human behavior. While much …

Uncertainty inspired RGB-D saliency detection

J Zhang, DP Fan, Y Dai, S Anwar… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
We propose the first stochastic framework to employ uncertainty for RGB-D saliency
detection by learning from the data labeling process. Existing RGB-D saliency detection …