Shortcut learning in deep neural networks

R Geirhos, JH Jacobsen, C Michaelis… - Nature Machine …, 2020 - nature.com
Deep learning has triggered the current rise of artificial intelligence and is the workhorse of
today's machine intelligence. Numerous success stories have rapidly spread all over …

On feature learning in the presence of spurious correlations

P Izmailov, P Kirichenko, N Gruver… - Advances in Neural …, 2022 - proceedings.neurips.cc
Deep classifiers are known to rely on spurious features—patterns which are correlated with
the target on the training data but not inherently relevant to the learning problem, such as the …

Dataset distillation with infinitely wide convolutional networks

T Nguyen, R Novak, L **ao… - Advances in Neural …, 2021 - proceedings.neurips.cc
The effectiveness of machine learning algorithms arises from being able to extract useful
features from large amounts of data. As model and dataset sizes increase, dataset …

Gradient starvation: A learning proclivity in neural networks

M Pezeshki, O Kaba, Y Bengio… - Advances in …, 2021 - proceedings.neurips.cc
We identify and formalize a fundamental gradient descent phenomenon resulting in a
learning proclivity in over-parameterized neural networks. Gradient Starvation arises when …

Trustworthy ai: A computational perspective

H Liu, Y Wang, W Fan, X Liu, Y Li, S Jain, Y Liu… - ACM Transactions on …, 2022 - dl.acm.org
In the past few decades, artificial intelligence (AI) technology has experienced swift
developments, changing everyone's daily life and profoundly altering the course of human …

[PDF][PDF] Machine psychology: Investigating emergent capabilities and behavior in large language models using psychological methods

T Hagendorff - arxiv preprint arxiv:2303.13988, 2023 - cybershafarat.com
Large language models (LLMs) are currently at the forefront of intertwining AI systems with
human communication and everyday life. Due to rapid technological advances and their …

Do wide and deep networks learn the same things? uncovering how neural network representations vary with width and depth

T Nguyen, M Raghu, S Kornblith - arxiv preprint arxiv:2010.15327, 2020 - arxiv.org
A key factor in the success of deep neural networks is the ability to scale models to improve
performance by varying the architecture depth and width. This simple property of neural …

The pitfalls of simplicity bias in neural networks

H Shah, K Tamuly, A Raghunathan… - Advances in …, 2020 - proceedings.neurips.cc
Several works have proposed Simplicity Bias (SB)---the tendency of standard training
procedures such as Stochastic Gradient Descent (SGD) to find simple models---to justify why …

Getting aligned on representational alignment

I Sucholutsky, L Muttenthaler, A Weller, A Peng… - arxiv preprint arxiv …, 2023 - arxiv.org
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …

Intriguing properties of contrastive losses

T Chen, C Luo, L Li - Advances in Neural Information …, 2021 - proceedings.neurips.cc
We study three intriguing properties of contrastive learning. First, we generalize the standard
contrastive loss to a broader family of losses, and we find that various instantiations of the …