Ultrafeedback: Boosting language models with high-quality feedback
Reinforcement learning from human feedback (RLHF) has become a pivot technique in
aligning large language models (LLMs) with human preferences. In RLHF practice …
aligning large language models (LLMs) with human preferences. In RLHF practice …
Aligning large language models with human: A survey
Large Language Models (LLMs) trained on extensive textual corpora have emerged as
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …
Prometheus: Inducing fine-grained evaluation capability in language models
Recently, GPT-4 has become the de facto evaluator for long-form text generated by large
language models (LLMs). However, for practitioners and researchers with large and custom …
language models (LLMs). However, for practitioners and researchers with large and custom …
Generative judge for evaluating alignment
The rapid development of Large Language Models (LLMs) has substantially expanded the
range of tasks they can address. In the field of Natural Language Processing (NLP) …
range of tasks they can address. In the field of Natural Language Processing (NLP) …
Training language models to self-correct via reinforcement learning
Self-correction is a highly desirable capability of large language models (LLMs), yet it has
consistently been found to be largely ineffective in modern LLMs. Current methods for …
consistently been found to be largely ineffective in modern LLMs. Current methods for …
Flask: Fine-grained language model evaluation based on alignment skill sets
Evaluation of Large Language Models (LLMs) is challenging because aligning to human
values requires the composition of multiple skills and the required set of skills varies …
values requires the composition of multiple skills and the required set of skills varies …
Shepherd: A critic for language model generation
As large language models improve, there is increasing interest in techniques that leverage
these models' capabilities to refine their own outputs. In this work, we introduce Shepherd, a …
these models' capabilities to refine their own outputs. In this work, we introduce Shepherd, a …
Internal consistency and self-feedback in large language models: A survey
Large language models (LLMs) often exhibit deficient reasoning or generate hallucinations.
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
A survey on knowledge distillation of large language models
This survey presents an in-depth exploration of knowledge distillation (KD) techniques
within the realm of Large Language Models (LLMs), spotlighting the pivotal role of KD in …
within the realm of Large Language Models (LLMs), spotlighting the pivotal role of KD in …
Confidence matters: Revisiting intrinsic self-correction capabilities of large language models
The recent success of Large Language Models (LLMs) has catalyzed an increasing interest
in their self-correction capabilities. This paper presents a comprehensive investigation into …
in their self-correction capabilities. This paper presents a comprehensive investigation into …