Text data augmentation for deep learning
Abstract Natural Language Processing (NLP) is one of the most captivating applications of
Deep Learning. In this survey, we consider how the Data Augmentation training strategy can …
Deep Learning. In this survey, we consider how the Data Augmentation training strategy can …
Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models
Deep generative models are a class of techniques that train deep neural networks to model
the distribution of training samples. Research has fragmented into various interconnected …
the distribution of training samples. Research has fragmented into various interconnected …
Sdxl: Improving latent diffusion models for high-resolution image synthesis
We present SDXL, a latent diffusion model for text-to-image synthesis. Compared to
previous versions of Stable Diffusion, SDXL leverages a three times larger UNet backbone …
previous versions of Stable Diffusion, SDXL leverages a three times larger UNet backbone …
Variational diffusion models
Diffusion-based generative models have demonstrated a capacity for perceptually
impressive synthesis, but can they also be great likelihood-based models? We answer this …
impressive synthesis, but can they also be great likelihood-based models? We answer this …
Analog bits: Generating discrete data using diffusion models with self-conditioning
We present Bit Diffusion: a simple and generic approach for generating discrete data with
continuous state and continuous time diffusion models. The main idea behind our approach …
continuous state and continuous time diffusion models. The main idea behind our approach …
Score-based generative modeling in latent space
Score-based generative models (SGMs) have recently demonstrated impressive results in
terms of both sample quality and distribution coverage. However, they are usually applied …
terms of both sample quality and distribution coverage. However, they are usually applied …
Graph contrastive learning automated
Self-supervised learning on graph-structured data has drawn recent interest for learning
generalizable, transferable and robust representations from unlabeled graphs. Among …
generalizable, transferable and robust representations from unlabeled graphs. Among …
Maximum likelihood training of score-based diffusion models
Score-based diffusion models synthesize samples by reversing a stochastic process that
diffuses data to noise, and are trained by minimizing a weighted combination of score …
diffuses data to noise, and are trained by minimizing a weighted combination of score …
Score-based generative modeling with critically-damped langevin diffusion
Score-based generative models (SGMs) have demonstrated remarkable synthesis quality.
SGMs rely on a diffusion process that gradually perturbs the data towards a tractable …
SGMs rely on a diffusion process that gradually perturbs the data towards a tractable …
Efficient training of language models to fill in the middle
We show that autoregressive language models can learn to infill text after we apply a
straightforward transformation to the dataset, which simply moves a span of text from the …
straightforward transformation to the dataset, which simply moves a span of text from the …