Learning to summarize with human feedback
As language models become more powerful, training and evaluation are increasingly
bottlenecked by the data and metrics used for a particular task. For example, summarization …
bottlenecked by the data and metrics used for a particular task. For example, summarization …
Symbolic chain-of-thought distillation: Small models can also" think" step-by-step
Chain-of-thought prompting (eg," Let's think step-by-step") primes large language models to
verbalize rationalization for their predictions. While chain-of-thought can lead to dramatic …
verbalize rationalization for their predictions. While chain-of-thought can lead to dramatic …
Improving code generation by training with natural language feedback
The potential for pre-trained large language models (LLMs) to use natural language
feedback at inference time has been an exciting recent development. We build upon this …
feedback at inference time has been an exciting recent development. We build upon this …
Lirex: Augmenting language inference with relevant explanations
Natural language explanations (NLEs) are a special form of data annotation in which
annotators identify rationales (most significant text tokens) when assigning labels to data …
annotators identify rationales (most significant text tokens) when assigning labels to data …
Towards interpretable natural language understanding with explanations as latent variables
Recently generating natural language explanations has shown very promising results in not
only offering interpretable explanations but also providing additional information and …
only offering interpretable explanations but also providing additional information and …
Relationship-embedded representation learning for grounding referring expressions
Grounding referring expressions in images aims to locate the object instance in an image
described by a referring expression. It involves a joint understanding of natural language …
described by a referring expression. It involves a joint understanding of natural language …
Training language models with language feedback
Pretrained language models often do not perform tasks in ways that are in line with our
preferences, eg, generating offensive text or factually incorrect summaries. Recent work …
preferences, eg, generating offensive text or factually incorrect summaries. Recent work …
Training language models with language feedback
Pretrained language models often do not perform tasks in ways that are in line with our
preferences, eg, generating offensive text or factually incorrect summaries. Recent work …
preferences, eg, generating offensive text or factually incorrect summaries. Recent work …
What if you said that differently?: How Explanation Formats Affect Human Feedback Efficacy and User Perception
Eliciting feedback from end users of NLP models can be beneficial for improving models.
However, how should we present model responses to users so they are most amenable to …
However, how should we present model responses to users so they are most amenable to …
Unpaired image captioning by image-level weakly-supervised visual concept recognition
The goal of unpaired image captioning (UIC) is to describe images without using image-
caption pairs in the training phase. Although challenging, we expect the task can be …
caption pairs in the training phase. Although challenging, we expect the task can be …