SummaReranker: A multi-task mixture-of-experts re-ranking framework for abstractive summarization
Sequence-to-sequence neural networks have recently achieved great success in abstractive
summarization, especially through fine-tuning large pre-trained language models on the …
summarization, especially through fine-tuning large pre-trained language models on the …
Learning to break the loop: Analyzing and mitigating repetitions for neural text generation
While large-scale neural language models, such as GPT2 and BART, have achieved
impressive results on various text generation tasks, they tend to get stuck in undesirable …
impressive results on various text generation tasks, they tend to get stuck in undesirable …
Repetition in repetition out: Towards understanding neural text degeneration from the data perspective
There are a number of diverging hypotheses about the neural text degeneration problem, ie,
generating repetitive and dull loops, which makes this problem both interesting and …
generating repetitive and dull loops, which makes this problem both interesting and …
Understanding in-context learning from repetitions
This paper explores the elusive mechanism underpinning in-context learning in Large
Language Models (LLMs). Our work provides a novel perspective by examining in-context …
Language Models (LLMs). Our work provides a novel perspective by examining in-context …
Nearest neighbor knowledge distillation for neural machine translation
k-nearest-neighbor machine translation (NN-MT), proposed by Khandelwal et al.(2021), has
achieved many state-of-the-art results in machine translation tasks. Although effective, NN …
achieved many state-of-the-art results in machine translation tasks. Although effective, NN …
R2D2: Robust data-to-text with replacement detection
Unfaithful text generation is a common problem for text generation systems. In the case of
Data-to-Text (D2T) systems, the factuality of the generated text is particularly crucial for any …
Data-to-Text (D2T) systems, the factuality of the generated text is particularly crucial for any …
From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Modern language models rely on the transformer architecture and attention mechanism to
perform language understanding and text generation. In this work, we study learning a 1 …
perform language understanding and text generation. In this work, we study learning a 1 …
[PDF][PDF] InferDPT: Privacy-preserving inference for black-box large language model
Large language models (LLMs), represented by ChatGPT, have greatly simplified text
generation tasks. However, they have also raised concerns about privacy risks such as data …
generation tasks. However, they have also raised concerns about privacy risks such as data …
Exploring automatic text simplification of german narrative documents
In this paper, we apply transformer-based Natural Language Generation (NLG) techniques
to the problem of text simplification. Currently, there are only a few German datasets …
to the problem of text simplification. Currently, there are only a few German datasets …
Decoupled non-parametric knowledge distillation for end-to-end speech translation
H Zhang, N Si, Y Chen, W Zhang… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
Existing techniques often attempt to make knowledge transfer from a powerful machine
translation (MT) to speech translation (ST) model with some elaborate techniques, which …
translation (MT) to speech translation (ST) model with some elaborate techniques, which …