Machine-generated text: A comprehensive survey of threat models and detection methods

EN Crothers, N Japkowicz, HL Viktor - IEEE Access, 2023 - ieeexplore.ieee.org
Machine-generated text is increasingly difficult to distinguish from text authored by humans.
Powerful open-source models are freely available, and user-friendly tools that democratize …

Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

Dart: Open-domain structured data record to text generation

L Nan, D Radev, R Zhang, A Rau, A Sivaprasad… - arxiv preprint arxiv …, 2020 - arxiv.org
We present DART, an open domain structured DAta Record to Text generation dataset with
over 82k instances (DARTs). Data-to-Text annotations can be a costly process, especially …

Structural adapters in pretrained language models for amr-to-text generation

LFR Ribeiro, Y Zhang, I Gurevych - arxiv preprint arxiv:2103.09120, 2021 - arxiv.org
Pretrained language models (PLM) have recently advanced graph-to-text generation, where
the input graph is linearized into a sequence and fed into the PLM to obtain its …

Evaluating semantic accuracy of data-to-text generation with natural language inference

O Dušek, Z Kasner - arxiv preprint arxiv:2011.10819, 2020 - arxiv.org
A major challenge in evaluating data-to-text (D2T) generation is measuring the semantic
accuracy of the generated text, ie checking if the output text contains all and only facts …

Sticking to the facts: Confident decoding for faithful data-to-text generation

R Tian, S Narayan, T Sellam, AP Parikh - arxiv preprint arxiv:1910.08684, 2019 - arxiv.org
We address the issue of hallucination in data-to-text generation, ie, reducing the generation
of text that is unsupported by the source. We conjecture that hallucination can be caused by …

Neural pipeline for zero-shot data-to-text generation

Z Kasner, O Dušek - arxiv preprint arxiv:2203.16279, 2022 - arxiv.org
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data
representation and repeating training data noise. We examine how to avoid finetuning …

Evaluating generative models for graph-to-text generation

S Yuan, M Färber - arxiv preprint arxiv:2307.14712, 2023 - arxiv.org
Large language models (LLMs) have been widely employed for graph-to-text generation
tasks. However, the process of finetuning LLMs requires significant training resources and …

Control prefixes for parameter-efficient text generation

J Clive, K Cao, M Rei - arxiv preprint arxiv:2110.08329, 2021 - arxiv.org
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language
model to a downstream application. However, it uses the same dataset-level tuned prompt …

Logic-consistency text generation from semantic parses

C Shu, Y Zhang, X Dong, P Shi, T Yu… - arxiv preprint arxiv …, 2021 - arxiv.org
Text generation from semantic parses is to generate textual descriptions for formal
representation inputs such as logic forms and SQL queries. This is challenging due to two …