Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing

P Liu, W Yuan, J Fu, Z Jiang, H Hayashi… - ACM Computing …, 2023 - dl.acm.org
This article surveys and organizes research works in a new paradigm in natural language
processing, which we dub “prompt-based learning.” Unlike traditional supervised learning …

Neural machine translation: A review

F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …

[PDF][PDF] Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension

M Lewis - arxiv preprint arxiv:1910.13461, 2019 - fq.pkwyx.com
We present BART, a denoising autoencoder for pretraining sequence-to-sequence models.
BART is trained by (1) corrupting text with an arbitrary noising function, and (2) learning a …

Gedi: Generative discriminator guided sequence generation

B Krause, AD Gotmare, B McCann, NS Keskar… - arxiv preprint arxiv …, 2020 - arxiv.org
While large-scale language models (LMs) are able to imitate the distribution of natural
language well enough to generate realistic text, it is difficult to control which regions of the …

Prophetnet: Predicting future n-gram for sequence-to-sequence pre-training

W Qi, Y Yan, Y Gong, D Liu, N Duan, J Chen… - arxiv preprint arxiv …, 2020 - arxiv.org
This paper presents a new sequence-to-sequence pre-training model called ProphetNet,
which introduces a novel self-supervised objective named future n-gram prediction and the …

Neural text generation with unlikelihood training

S Welleck, I Kulikov, S Roller, E Dinan, K Cho… - arxiv preprint arxiv …, 2019 - arxiv.org
Neural text generation is a key tool in natural language applications, but it is well known
there are major problems at its core. In particular, standard likelihood training and decoding …

Don't give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization

S Narayan, SB Cohen, M Lapata - arxiv preprint arxiv:1808.08745, 2018 - arxiv.org
We introduce extreme summarization, a new single-document summarization task which
does not favor extractive strategies and calls for an abstractive modeling approach. The idea …

Wizard of wikipedia: Knowledge-powered conversational agents

E Dinan, S Roller, K Shuster, A Fan, M Auli… - arxiv preprint arxiv …, 2018 - arxiv.org
In open-domain dialogue intelligent agents should exhibit the use of knowledge, however
there are few convincing demonstrations of this to date. The most popular sequence to …

GSum: A general framework for guided neural abstractive summarization

ZY Dou, P Liu, H Hayashi, Z Jiang, G Neubig - arxiv preprint arxiv …, 2020 - arxiv.org
Neural abstractive summarization models are flexible and can produce coherent summaries,
but they are sometimes unfaithful and can be difficult to control. While previous studies …

ELI5: Long form question answering

A Fan, Y Jernite, E Perez, D Grangier, J Weston… - arxiv preprint arxiv …, 2019 - arxiv.org
We introduce the first large-scale corpus for long-form question answering, a task requiring
elaborate and in-depth answers to open-ended questions. The dataset comprises 270K …