A survey of controllable text generation using transformer-based pre-trained language models

H Zhang, H Song, S Li, M Zhou, D Song - ACM Computing Surveys, 2023 - dl.acm.org
Controllable Text Generation (CTG) is an emerging area in the field of natural language
generation (NLG). It is regarded as crucial for the development of advanced text generation …

Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

Lamda: Language models for dialog applications

R Thoppilan, D De Freitas, J Hall, N Shazeer… - arxiv preprint arxiv …, 2022 - arxiv.org
We present LaMDA: Language Models for Dialog Applications. LaMDA is a family of
Transformer-based neural language models specialized for dialog, which have up to 137B …

A survey of natural language generation

C Dong, Y Li, H Gong, M Chen, J Li, Y Shen… - ACM Computing …, 2022 - dl.acm.org
This article offers a comprehensive review of the research on Natural Language Generation
(NLG) over the past two decades, especially in relation to data-to-text generation and text-to …

Recipes for building an open-domain chatbot

S Roller, E Dinan, N Goyal, D Ju, M Williamson… - arxiv preprint arxiv …, 2020 - arxiv.org
Building open-domain chatbots is a challenging area for machine learning research. While
prior work has shown that scaling neural models in the number of parameters and the size of …

Transformers: State-of-the-art natural language processing

T Wolf, L Debut, V Sanh, J Chaumond… - Proceedings of the …, 2020 - aclanthology.org
Recent progress in natural language processing has been driven by advances in both
model architecture and model pretraining. Transformer architectures have facilitated …

Dialogpt: Large-scale generative pre-training for conversational response generation

Y Zhang, S Sun, M Galley, YC Chen, C Brockett… - arxiv preprint arxiv …, 2019 - arxiv.org
We present a large, tunable neural conversational response generation model, DialoGPT
(dialogue generative pre-trained transformer). Trained on 147M conversation-like …

[PDF][PDF] Language models are unsupervised multitask learners

A Radford, J Wu, R Child, D Luan… - OpenAI …, 2019 - storage.prod.researchhub.com
Natural language processing tasks, such as question answering, machine translation,
reading comprehension, and summarization, are typically approached with supervised …

Multi-task pre-training for plug-and-play task-oriented dialogue system

Y Su, L Shu, E Mansimov, A Gupta, D Cai… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained language models have been recently shown to benefit task-oriented dialogue
(TOD) systems. Despite their success, existing methods often formulate this task as a …

TOD-BERT: Pre-trained natural language understanding for task-oriented dialogue

CS Wu, S Hoi, R Socher, C **ong - arxiv preprint arxiv:2004.06871, 2020 - arxiv.org
The underlying difference of linguistic patterns between general text and task-oriented
dialogue makes existing pre-trained language models less useful in practice. In this work …