Modeling similarity and psychological space
Similarity and categorization are fundamental processes in human cognition that help
complex organisms make sense of the cacophony of information in their environment. These …
complex organisms make sense of the cacophony of information in their environment. These …
[HTML][HTML] Analyzing biological and artificial neural networks: challenges with opportunities for synergy?
Highlights•Artificial and biological neural networks can be analyzed using similar
methods.•Neural analysis has revealed similarities between the representations in artificial …
methods.•Neural analysis has revealed similarities between the representations in artificial …
Memorization without overfitting: Analyzing the training dynamics of large language models
Despite their wide adoption, the underlying training and memorization dynamics of very
large language models is not well understood. We empirically study exact memorization in …
large language models is not well understood. We empirically study exact memorization in …
Do vision transformers see like convolutional neural networks?
Convolutional neural networks (CNNs) have so far been the de-facto model for visual data.
Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or …
Recent work has shown that (Vision) Transformer models (ViT) can achieve comparable or …
A toy model of universality: Reverse engineering how networks learn group operations
Universality is a key hypothesis in mechanistic interpretability–that different models learn
similar features and circuits when trained on similar tasks. In this work, we study the …
similar features and circuits when trained on similar tasks. In this work, we study the …
Layer-wise analysis of a self-supervised speech representation model
Recently proposed self-supervised learning approaches have been successful for pre-
training speech representation models. The utility of these learned representations has been …
training speech representation models. The utility of these learned representations has been …
Learning from failure: De-biasing classifier from biased classifier
Neural networks often learn to make predictions that overly rely on spurious corre-lation
existing in the dataset, which causes the model to be biased. While previous work tackles …
existing in the dataset, which causes the model to be biased. While previous work tackles …
Toward transparent ai: A survey on interpreting the inner structures of deep neural networks
The last decade of machine learning has seen drastic increases in scale and capabilities.
Deep neural networks (DNNs) are increasingly being deployed in the real world. However …
Deep neural networks (DNNs) are increasingly being deployed in the real world. However …
Rapid learning or feature reuse? towards understanding the effectiveness of maml
Similarity of neural network representations revisited
Recent work has sought to understand the behavior of neural networks by comparing
representations between layers and between different trained models. We examine methods …
representations between layers and between different trained models. We examine methods …