Text data augmentation for deep learning

C Shorten, TM Khoshgoftaar, B Furht - Journal of big Data, 2021 - Springer
Abstract Natural Language Processing (NLP) is one of the most captivating applications of
Deep Learning. In this survey, we consider how the Data Augmentation training strategy can …

Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models

S Bond-Taylor, A Leach, Y Long… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Deep generative models are a class of techniques that train deep neural networks to model
the distribution of training samples. Research has fragmented into various interconnected …

Sdxl: Improving latent diffusion models for high-resolution image synthesis

D Podell, Z English, K Lacey, A Blattmann… - arxiv preprint arxiv …, 2023 - arxiv.org
We present SDXL, a latent diffusion model for text-to-image synthesis. Compared to
previous versions of Stable Diffusion, SDXL leverages a three times larger UNet backbone …

Variational diffusion models

D Kingma, T Salimans, B Poole… - Advances in neural …, 2021 - proceedings.neurips.cc
Diffusion-based generative models have demonstrated a capacity for perceptually
impressive synthesis, but can they also be great likelihood-based models? We answer this …

Analog bits: Generating discrete data using diffusion models with self-conditioning

T Chen, R Zhang, G Hinton - arxiv preprint arxiv:2208.04202, 2022 - arxiv.org
We present Bit Diffusion: a simple and generic approach for generating discrete data with
continuous state and continuous time diffusion models. The main idea behind our approach …

Score-based generative modeling in latent space

A Vahdat, K Kreis, J Kautz - Advances in neural information …, 2021 - proceedings.neurips.cc
Score-based generative models (SGMs) have recently demonstrated impressive results in
terms of both sample quality and distribution coverage. However, they are usually applied …

Graph contrastive learning automated

Y You, T Chen, Y Shen, Z Wang - … Conference on Machine …, 2021 - proceedings.mlr.press
Self-supervised learning on graph-structured data has drawn recent interest for learning
generalizable, transferable and robust representations from unlabeled graphs. Among …

Maximum likelihood training of score-based diffusion models

Y Song, C Durkan, I Murray… - Advances in neural …, 2021 - proceedings.neurips.cc
Score-based diffusion models synthesize samples by reversing a stochastic process that
diffuses data to noise, and are trained by minimizing a weighted combination of score …

Score-based generative modeling with critically-damped langevin diffusion

T Dockhorn, A Vahdat, K Kreis - arxiv preprint arxiv:2112.07068, 2021 - arxiv.org
Score-based generative models (SGMs) have demonstrated remarkable synthesis quality.
SGMs rely on a diffusion process that gradually perturbs the data towards a tractable …

Efficient training of language models to fill in the middle

M Bavarian, H Jun, N Tezak, J Schulman… - arxiv preprint arxiv …, 2022 - arxiv.org
We show that autoregressive language models can learn to infill text after we apply a
straightforward transformation to the dataset, which simply moves a span of text from the …