Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models

S Bond-Taylor, A Leach, Y Long… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Deep generative models are a class of techniques that train deep neural networks to model
the distribution of training samples. Research has fragmented into various interconnected …

Controllable protein design with language models

N Ferruz, B Höcker - Nature Machine Intelligence, 2022 - nature.com
The twenty-first century is presenting humankind with unprecedented environmental and
medical challenges. The ability to design novel proteins tailored for specific purposes would …

Structured denoising diffusion models in discrete state-spaces

J Austin, DD Johnson, J Ho, D Tarlow… - Advances in neural …, 2021 - proceedings.neurips.cc
Denoising diffusion probabilistic models (DDPMs)[Ho et al. 2021] have shown impressive
results on image and waveform generation in continuous state spaces. Here, we introduce …

Show-o: One single transformer to unify multimodal understanding and generation

J **e, W Mao, Z Bai, DJ Zhang, W Wang, KQ Lin… - arxiv preprint arxiv …, 2024 - arxiv.org
We present a unified transformer, ie, Show-o, that unifies multimodal understanding and
generation. Unlike fully autoregressive models, Show-o unifies autoregressive and …

Restoring and attributing ancient texts using deep neural networks

Y Assael, T Sommerschield, B Shillingford, M Bordbar… - Nature, 2022 - nature.com
Ancient history relies on disciplines such as epigraphy—the study of inscribed texts known
as inscriptions—for evidence of the thought, language, society and history of past …

Realtoxicityprompts: Evaluating neural toxic degeneration in language models

S Gehman, S Gururangan, M Sap, Y Choi… - arxiv preprint arxiv …, 2020 - arxiv.org
Pretrained neural language models (LMs) are prone to generating racist, sexist, or otherwise
toxic language which hinders their safe deployment. We investigate the extent to which …

It's not just size that matters: Small language models are also few-shot learners

T Schick, H Schütze - arxiv preprint arxiv:2009.07118, 2020 - arxiv.org
When scaled to hundreds of billions of parameters, pretrained language models such as
GPT-3 (Brown et al., 2020) achieve remarkable few-shot performance. However, enormous …

Self-diagnosis and self-debiasing: A proposal for reducing corpus-based bias in nlp

T Schick, S Udupa, H Schütze - Transactions of the Association for …, 2021 - direct.mit.edu
Abstract⚠ This paper contains prompts and model outputs that are offensive in nature. When
trained on large, unfiltered crawls from the Internet, language models pick up and reproduce …

Framework for a foreign language teaching software for children utilizing AR, voicebots and ChatGPT (large language models)

O Topsakal, E Topsakal - The Journal of Cognitive Systems, 2022 - dergipark.org.tr
The cognitive capabilities of children develop during the early years of their life. Research
shows that learning a foreign language helps develop cognitive skills. Moreover, learning a …

What artificial neural networks can tell us about human language acquisition

A Warstadt, SR Bowman - Algebraic structures in natural …, 2022 - taylorfrancis.com
Rapid progress in machine learning for natural language processing has the potential to
transform debates about how humans learn language. However, the learning environments …