The fast committor machine: Interpretable prediction with kernels
D Aristoff, M Johnson, G Simpson… - The Journal of chemical …, 2024 - pubs.aip.org
In the study of stochastic systems, the committor function describes the probability that a
system starting from an initial configuration x will reach a set B before a set A. This paper …
system starting from an initial configuration x will reach a set B before a set A. This paper …
Wide neural networks trained with weight decay provably exhibit neural collapse
Deep neural networks (DNNs) at convergence consistently represent the training data in the
last layer via a highly symmetric geometric structure referred to as neural collapse. This …
last layer via a highly symmetric geometric structure referred to as neural collapse. This …
Emergence in non-neural models: grokking modular arithmetic via average gradient outer product
Neural networks trained to solve modular arithmetic tasks exhibit grokking, a phenomenon
where the test accuracy starts improving long after the model achieves 100% training …
where the test accuracy starts improving long after the model achieves 100% training …
Neural collapse vs. low-rank bias: Is deep neural collapse really optimal?
Deep neural networks (DNNs) exhibit a surprising structure in their final layer known as
neural collapse (NC), and a growing body of works has currently investigated the …
neural collapse (NC), and a growing body of works has currently investigated the …
Theoretical characterisation of the Gauss-Newton conditioning in Neural Networks
The Gauss-Newton (GN) matrix plays an important role in machine learning, most evident in
its use as a preconditioning matrix for a wide family of popular adaptive methods to speed …
its use as a preconditioning matrix for a wide family of popular adaptive methods to speed …
Neural Collapse versus Low-rank Bias: Is Deep Neural Collapse Really Optimal?
Deep neural networks (DNNs) exhibit a surprising structure in their final layer known as
neural collapse (NC), and a growing body of works has currently investigated the …
neural collapse (NC), and a growing body of works has currently investigated the …
Compressible Dynamics in Deep Overparameterized Low-Rank Learning & Adaptation
While overparameterization in machine learning models offers great benefits in terms of
optimization and generalization, it also leads to increased computational requirements as …
optimization and generalization, it also leads to increased computational requirements as …
The Persistence of Neural Collapse Despite Low-Rank Bias: An Analytic Perspective Through Unconstrained Features
C Garrod, JP Keating - arxiv preprint arxiv:2410.23169, 2024 - arxiv.org
Modern deep neural networks have been observed to exhibit a simple structure in their final
layer features and weights, commonly referred to as neural collapse. This phenomenon has …
layer features and weights, commonly referred to as neural collapse. This phenomenon has …
Neural Collapse Beyond the Unconstrainted Features Model: Landscape, Dynamics, and Generalization in the Mean-Field Regime
D Wu, M Mondelli - arxiv preprint arxiv:2501.19104, 2025 - arxiv.org
Neural Collapse is a phenomenon where the last-layer representations of a well-trained
neural network converge to a highly structured geometry. In this paper, we focus on its first …
neural network converge to a highly structured geometry. In this paper, we focus on its first …
Neural Collapse Inspired Feature Alignment for Out-of-Distribution Generalization
The spurious correlation between the background features of the image and its label arises
due to that the samples labeled with the same class in the training set often co-occurs with a …
due to that the samples labeled with the same class in the training set often co-occurs with a …