A contrastive framework for neural text generation

Y Su, T Lan, Y Wang, D Yogatama… - Advances in Neural …, 2022 - proceedings.neurips.cc
Text generation is of great importance to many natural language processing applications.
However, maximization-based decoding methods (eg, beam search) of neural language …

Multi-task pre-training for plug-and-play task-oriented dialogue system

Y Su, L Shu, E Mansimov, A Gupta, D Cai… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained language models have been recently shown to benefit task-oriented dialogue
(TOD) systems. Despite their success, existing methods often formulate this task as a …

A survey on retrieval-augmented text generation

H Li, Y Su, D Cai, Y Wang, L Liu - arxiv preprint arxiv:2202.01110, 2022 - arxiv.org
Recently, retrieval-augmented text generation attracted increasing attention of the
computational linguistics community. Compared with conventional generation models …

Language models can see: Plugging visual controls in text generation

Y Su, T Lan, Y Liu, F Liu, D Yogatama, Y Wang… - arxiv preprint arxiv …, 2022 - arxiv.org
Generative language models (LMs) such as GPT-2/3 can be prompted to generate text with
remarkable quality. While they are designed for text-prompted generation, it remains an …

A survey on neural data-to-text generation

Y Lin, T Ruan, J Liu, H Wang - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Data-to-text Generation (D2T) aims to generate textual natural language statements that can
fluently and precisely describe the structured data such as graphs, tables, and meaning …

Plan-then-generate: Controlled data-to-text generation via planning

Y Su, D Vandyke, S Wang, Y Fang, N Collier - arxiv preprint arxiv …, 2021 - arxiv.org
Recent developments in neural networks have led to the advance in data-to-text generation.
However, the lack of ability of neural models to control the structure of generated output can …

Retrieving multimodal information for augmented generation: A survey

R Zhao, H Chen, W Wang, F Jiao, XL Do, C Qin… - arxiv preprint arxiv …, 2023 - arxiv.org
As Large Language Models (LLMs) become popular, there emerged an important trend of
using multimodality to augment the LLMs' generation ability, which enables LLMs to better …

TaCL: Improving BERT pre-training with token-aware contrastive learning

Y Su, F Liu, Z Meng, T Lan, L Shu, E Shareghi… - arxiv preprint arxiv …, 2021 - arxiv.org
Masked language models (MLMs) such as BERT and RoBERTa have revolutionized the
field of Natural Language Understanding in the past few years. However, existing pre …

Bioreader: a retrieval-enhanced text-to-text transformer for biomedical literature

G Frisoni, M Mizutani, G Moro… - Proceedings of the 2022 …, 2022 - aclanthology.org
The latest batch of research has equipped language models with the ability to attend over
relevant and factual information from non-parametric external sources, drawing a …

Neural pipeline for zero-shot data-to-text generation

Z Kasner, O Dušek - arxiv preprint arxiv:2203.16279, 2022 - arxiv.org
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data
representation and repeating training data noise. We examine how to avoid finetuning …