A survey on data selection for language models
A major factor in the recent success of large language models is the use of enormous and
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
Foundational challenges in assuring alignment and safety of large language models
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …
language models (LLMs). These challenges are organized into three different categories …
How johnny can persuade llms to jailbreak them: Rethinking persuasion to challenge ai safety by humanizing llms
Most traditional AI safety research has approached AI models as machines and centered on
algorithm-focused attacks developed by security experts. As large language models (LLMs) …
algorithm-focused attacks developed by security experts. As large language models (LLMs) …
Olmo: Accelerating the science of language models
Language models (LMs) have become ubiquitous in both NLP research and in commercial
product offerings. As their commercial importance has surged, the most powerful models …
product offerings. As their commercial importance has surged, the most powerful models …
Bigcodebench: Benchmarking code generation with diverse function calls and complex instructions
Task automation has been greatly empowered by the recent advances in Large Language
Models (LLMs) via Python code, where the tasks ranging from software engineering …
Models (LLMs) via Python code, where the tasks ranging from software engineering …
No" zero-shot" without exponential data: Pretraining concept frequency determines multimodal model performance
Web-crawled pretraining datasets underlie the impressive" zero-shot" evaluation
performance of multimodal models, such as CLIP for classification and Stable-Diffusion for …
performance of multimodal models, such as CLIP for classification and Stable-Diffusion for …
An archival perspective on pretraining data
Alongside an explosion in research and development related to large language models,
there has been a concomitant rise in the creation of pretraining datasets—massive …
there has been a concomitant rise in the creation of pretraining datasets—massive …
Open problems in technical ai governance
AI progress is creating a growing range of risks and opportunities, but it is often unclear how
they should be navigated. In many cases, the barriers and uncertainties faced are at least …
they should be navigated. In many cases, the barriers and uncertainties faced are at least …
The bias amplification paradox in text-to-image generation
Bias amplification is a phenomenon in which models increase imbalances present in the
training data. In this paper, we study bias amplification in the text-to-image domain using …
training data. In this paper, we study bias amplification in the text-to-image domain using …
Evaluating copyright takedown methods for language models
Language models (LMs) derive their capabilities from extensive training on diverse data,
including potentially copyrighted material. These models can memorize and generate …
including potentially copyrighted material. These models can memorize and generate …