Ethical machine learning in healthcare
The use of machine learning (ML) in healthcare raises numerous ethical concerns,
especially as models can amplify existing health inequities. Here, we outline ethical …
especially as models can amplify existing health inequities. Here, we outline ethical …
On the opportunities and risks of foundation models
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …
Towards out-of-distribution generalization: A survey
Traditional machine learning paradigms are based on the assumption that both training and
test data follow the same statistical pattern, which is mathematically referred to as …
test data follow the same statistical pattern, which is mathematically referred to as …
Just train twice: Improving group robustness without training group information
Standard training via empirical risk minimization (ERM) can produce models that achieve
low error on average but high error on minority groups, especially in the presence of …
low error on average but high error on minority groups, especially in the presence of …
AI for radiographic COVID-19 detection selects shortcuts over signal
Artificial intelligence (AI) researchers and radiologists have recently reported AI systems that
accurately detect COVID-19 in chest radiographs. However, the robustness of these systems …
accurately detect COVID-19 in chest radiographs. However, the robustness of these systems …
On feature learning in the presence of spurious correlations
Deep classifiers are known to rely on spurious features—patterns which are correlated with
the target on the training data but not inherently relevant to the learning problem, such as the …
the target on the training data but not inherently relevant to the learning problem, such as the …
Fishr: Invariant gradient variances for out-of-distribution generalization
Learning robust models that generalize well under changes in the data distribution is critical
for real-world applications. To this end, there has been a growing surge of interest to learn …
for real-world applications. To this end, there has been a growing surge of interest to learn …
Improving out-of-distribution robustness via selective augmentation
Abstract Machine learning algorithms typically assume that training and test examples are
drawn from the same distribution. However, distribution shift is a common problem in real …
drawn from the same distribution. However, distribution shift is a common problem in real …
Gradient starvation: A learning proclivity in neural networks
We identify and formalize a fundamental gradient descent phenomenon resulting in a
learning proclivity in over-parameterized neural networks. Gradient Starvation arises when …
learning proclivity in over-parameterized neural networks. Gradient Starvation arises when …
Wilds: A benchmark of in-the-wild distribution shifts
Distribution shifts—where the training distribution differs from the test distribution—can
substantially degrade the accuracy of machine learning (ML) systems deployed in the wild …
substantially degrade the accuracy of machine learning (ML) systems deployed in the wild …