Shortcut learning in deep neural networks
Deep learning has triggered the current rise of artificial intelligence and is the workhorse of
today's machine intelligence. Numerous success stories have rapidly spread all over …
today's machine intelligence. Numerous success stories have rapidly spread all over …
On feature learning in the presence of spurious correlations
Deep classifiers are known to rely on spurious features—patterns which are correlated with
the target on the training data but not inherently relevant to the learning problem, such as the …
the target on the training data but not inherently relevant to the learning problem, such as the …
Dataset distillation with infinitely wide convolutional networks
The effectiveness of machine learning algorithms arises from being able to extract useful
features from large amounts of data. As model and dataset sizes increase, dataset …
features from large amounts of data. As model and dataset sizes increase, dataset …
Gradient starvation: A learning proclivity in neural networks
We identify and formalize a fundamental gradient descent phenomenon resulting in a
learning proclivity in over-parameterized neural networks. Gradient Starvation arises when …
learning proclivity in over-parameterized neural networks. Gradient Starvation arises when …
Trustworthy ai: A computational perspective
In the past few decades, artificial intelligence (AI) technology has experienced swift
developments, changing everyone's daily life and profoundly altering the course of human …
developments, changing everyone's daily life and profoundly altering the course of human …
[PDF][PDF] Machine psychology: Investigating emergent capabilities and behavior in large language models using psychological methods
T Hagendorff - arxiv preprint arxiv:2303.13988, 2023 - cybershafarat.com
Large language models (LLMs) are currently at the forefront of intertwining AI systems with
human communication and everyday life. Due to rapid technological advances and their …
human communication and everyday life. Due to rapid technological advances and their …
Do wide and deep networks learn the same things? uncovering how neural network representations vary with width and depth
A key factor in the success of deep neural networks is the ability to scale models to improve
performance by varying the architecture depth and width. This simple property of neural …
performance by varying the architecture depth and width. This simple property of neural …
The pitfalls of simplicity bias in neural networks
Several works have proposed Simplicity Bias (SB)---the tendency of standard training
procedures such as Stochastic Gradient Descent (SGD) to find simple models---to justify why …
procedures such as Stochastic Gradient Descent (SGD) to find simple models---to justify why …
Getting aligned on representational alignment
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
Intriguing properties of contrastive losses
We study three intriguing properties of contrastive learning. First, we generalize the standard
contrastive loss to a broader family of losses, and we find that various instantiations of the …
contrastive loss to a broader family of losses, and we find that various instantiations of the …