3d common corruptions and data augmentation
We introduce a set of image transformations that can be used as corruptions to evaluate the
robustness of models as well as data augmentation mechanisms for training neural …
robustness of models as well as data augmentation mechanisms for training neural …
A whac-a-mole dilemma: Shortcuts come in multiples where mitigating one amplifies others
Abstract Machine learning models have been found to learn shortcuts---unintended decision
rules that are unable to generalize---undermining models' reliability. Previous works address …
rules that are unable to generalize---undermining models' reliability. Previous works address …
Datamodels: Predicting predictions from training data
We present a conceptual framework, datamodeling, for analyzing the behavior of a model
class in terms of the training data. For any fixed" target" example $ x $, training set $ S $, and …
class in terms of the training data. For any fixed" target" example $ x $, training set $ S $, and …
Salient imagenet: How to discover spurious features in deep learning?
Deep neural networks can be unreliable in the real world especially when they heavily use
{\it spurious} features for their predictions. Focusing on image classifications, we define {\it …
{\it spurious} features for their predictions. Focusing on image classifications, we define {\it …
Adaptive testing of computer vision models
Vision models often fail systematically on groups of data that share common semantic
characteristics (eg, rare objects or unusual scenes), but identifying these failure modes is a …
characteristics (eg, rare objects or unusual scenes), but identifying these failure modes is a …
Artificial intelligence foundation and pre-trained models: Fundamentals, applications, opportunities, and social impacts
A Kolides, A Nawaz, A Rathor, D Beeman… - … Modelling Practice and …, 2023 - Elsevier
With the emergence of foundation models (FMs) that are trained on large amounts of data at
scale and adaptable to a wide range of downstream applications, AI is experiencing a …
scale and adaptable to a wide range of downstream applications, AI is experiencing a …
Red teaming deep neural networks with feature synthesis tools
Interpretable AI tools are often motivated by the goal of understanding model behavior in out-
of-distribution (OOD) contexts. Despite the attention this area of study receives, there are …
of-distribution (OOD) contexts. Despite the attention this area of study receives, there are …
Modeldiff: A framework for comparing learning algorithms
We study the problem of (learning) algorithm comparison, where the goal is to find
differences between models trained with two different learning algorithms. We begin by …
differences between models trained with two different learning algorithms. We begin by …
Diagnosing and rectifying vision models using language
Recent multi-modal contrastive learning models have demonstrated the ability to learn an
embedding space suitable for building strong vision classifiers, by leveraging the rich …
embedding space suitable for building strong vision classifiers, by leveraging the rich …
Dataset interfaces: Diagnosing model failures using controllable counterfactual generation
Distribution shift is a major source of failure for machine learning models. However,
evaluating model reliability under distribution shift can be challenging, especially since it …
evaluating model reliability under distribution shift can be challenging, especially since it …