Decoding methods in neural language generation: a survey

S Zarrieß, H Voigt, S Schüz - Information, 2021 - mdpi.com
Neural encoder-decoder models for language generation can be trained to predict words
directly from linguistic or non-linguistic inputs. When generating with these so-called end-to …

If beam search is the answer, what was the question?

C Meister, T Vieira, R Cotterell - arxiv preprint arxiv:2010.02650, 2020 - arxiv.org
Quite surprisingly, exact maximum a posteriori (MAP) decoding of neural language
generators frequently leads to low-quality results. Rather, most state-of-the-art results on …

Learning to reason deductively: Math word problem solving as complex relation extraction

Z Jie, J Li, W Lu - arxiv preprint arxiv:2203.10316, 2022 - arxiv.org
Solving math word problems requires deductive reasoning over the quantities in the text.
Various recent research efforts mostly relied on sequence-to-sequence or sequence-to-tree …

Transformer neural network for protein-specific de novo drug generation as a machine translation problem

D Grechishnikova - Scientific reports, 2021 - nature.com
Drug discovery for a protein target is a very laborious, long and costly process. Machine
learning approaches and, in particular, deep generative networks can substantially reduce …

On decoding strategies for neural text generators

G Wiher, C Meister, R Cotterell - Transactions of the Association for …, 2022 - direct.mit.edu
When generating text from probabilistic models, the chosen decoding strategy has a
profound effect on the resulting text. Yet the properties elicited by various decoding …

Is MAP decoding all you need? the inadequacy of the mode in neural machine translation

B Eikema, W Aziz - arxiv preprint arxiv:2005.10283, 2020 - arxiv.org
Recent studies have revealed a number of pathologies of neural machine translation (NMT)
systems. Hypotheses explaining these mostly suggest there is something fundamentally …

Understanding and improving sequence-to-sequence pretraining for neural machine translation

W Wang, W Jiao, Y Hao, X Wang, S Shi, Z Tu… - arxiv preprint arxiv …, 2022 - arxiv.org
In this paper, we present a substantial step in better understanding the SOTA sequence-to-
sequence (Seq2Seq) pretraining for neural machine translation~(NMT). We focus on …

Machine translation decoding beyond beam search

R Leblond, JB Alayrac, L Sifre, M Pislar… - arxiv preprint arxiv …, 2021 - arxiv.org
Beam search is the go-to method for decoding auto-regressive machine translation models.
While it yields consistent improvements in terms of BLEU, it is only concerned with finding …

Language model evaluation beyond perplexity

C Meister, R Cotterell - arxiv preprint arxiv:2106.00085, 2021 - arxiv.org
We propose an alternate approach to quantifying how well language models learn natural
language: we ask how well they match the statistical tendencies of natural language. To …

Plate: Visually-grounded planning with transformers in procedural tasks

J Sun, DA Huang, B Lu, YH Liu… - IEEE Robotics and …, 2022 - ieeexplore.ieee.org
In this work, we study the problem of how to leverage instructional videos to facilitate the
understanding of human decision-making processes, focusing on training a model with the …