Survey of hallucination in natural language generation
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …
the development of sequence-to-sequence deep learning technologies such as Transformer …
Repairing the cracked foundation: A survey of obstacles in evaluation practices for generated text
Abstract Evaluation practices in natural language generation (NLG) have many known flaws,
but improved evaluation approaches are rarely widely adopted. This issue has become …
but improved evaluation approaches are rarely widely adopted. This issue has become …
Glm-130b: An open bilingual pre-trained model
We introduce GLM-130B, a bilingual (English and Chinese) pre-trained language model
with 130 billion parameters. It is an attempt to open-source a 100B-scale model at least as …
with 130 billion parameters. It is an attempt to open-source a 100B-scale model at least as …
Palm: Scaling language modeling with pathways
Large language models have been shown to achieve remarkable performance across a
variety of natural language tasks using few-shot learning, which drastically reduces the …
variety of natural language tasks using few-shot learning, which drastically reduces the …
Finetuned language models are zero-shot learners
This paper explores a simple method for improving the zero-shot learning abilities of
language models. We show that instruction tuning--finetuning language models on a …
language models. We show that instruction tuning--finetuning language models on a …
Adapting large language models via reading comprehension
We explore how continued pre-training on domain-specific corpora influences large
language models, revealing that training on the raw corpora endows the model with domain …
language models, revealing that training on the raw corpora endows the model with domain …
Spot: Better frozen model adaptation through soft prompt transfer
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …
Ext5: Towards extreme multi-task scaling for transfer learning
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …
processing (NLP), few works have systematically studied the effect of scaling up the number …
Preventing verbatim memorization in language models gives a false sense of privacy
Studying data memorization in neural language models helps us understand the risks (eg, to
privacy or copyright) associated with models regurgitating training data and aids in the …
privacy or copyright) associated with models regurgitating training data and aids in the …
Unified demonstration retriever for in-context learning
In-context learning is a new learning paradigm where a language model conditions on a few
input-output pairs (demonstrations) and a test input, and directly outputs the prediction. It has …
input-output pairs (demonstrations) and a test input, and directly outputs the prediction. It has …