Self-consistent dynamical field theory of kernel evolution in wide neural networks
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …
through a self-consistent dynamical field theory. We construct a collection of deterministic …
Dynamics of finite width kernel and prediction fluctuations in mean field neural networks
We analyze the dynamics of finite width effects in wide but finite feature learning neural
networks. Starting from a dynamical mean field theory description of infinite width deep …
networks. Starting from a dynamical mean field theory description of infinite width deep …
The tunnel effect: Building data representations in deep neural networks
Deep neural networks are widely known for their remarkable effectiveness across various
tasks, with the consensus that deeper networks implicitly learn more complex data …
tasks, with the consensus that deeper networks implicitly learn more complex data …
Width and depth limits commute in residual networks
We show that taking the width and depth to infinity in a deep neural network with skip
connections, when branches are scaled by $1/\sqrt {depth} $, result in the same covariance …
connections, when branches are scaled by $1/\sqrt {depth} $, result in the same covariance …
The influence of learning rule on representation dynamics in wide neural networks
It is unclear how changing the learning rule of a deep neural network alters its learning
dynamics and representations. To gain insight into the relationship between learned …
dynamics and representations. To gain insight into the relationship between learned …
Leave-one-out distinguishability in machine learning
We introduce a new analytical framework to quantify the changes in a machine learning
algorithm's output distribution following the inclusion of a few data points in its training set, a …
algorithm's output distribution following the inclusion of a few data points in its training set, a …
Neural (tangent kernel) collapse
This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures
the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) …
the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) …
On the infinite-depth limit of finite-width neural networks
S Hayou - Transactions on Machine Learning Research, 2022 - openreview.net
In this paper, we study the infinite-depth limit of finite-width residual neural networks with
random Gaussian weights. With proper scaling, we show that by fixing the width and taking …
random Gaussian weights. With proper scaling, we show that by fixing the width and taking …
A spring-block theory of feature learning in deep neural networks
Feature-learning deep nets progressively collapse data to a regular low-dimensional
geometry. How this phenomenon emerges from collective action of nonlinearity, noise …
geometry. How this phenomenon emerges from collective action of nonlinearity, noise …
Self-consistent dynamical field theory of kernel evolution in wide neural networks
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …
through a self-consistent dynamical field theory. We construct a collection of deterministic …