Self-consistent dynamical field theory of kernel evolution in wide neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …

Dynamics of finite width kernel and prediction fluctuations in mean field neural networks

B Bordelon, C Pehlevan - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We analyze the dynamics of finite width effects in wide but finite feature learning neural
networks. Starting from a dynamical mean field theory description of infinite width deep …

The tunnel effect: Building data representations in deep neural networks

W Masarczyk, M Ostaszewski, E Imani… - Advances in …, 2024 - proceedings.neurips.cc
Deep neural networks are widely known for their remarkable effectiveness across various
tasks, with the consensus that deeper networks implicitly learn more complex data …

Width and depth limits commute in residual networks

S Hayou, G Yang - International Conference on Machine …, 2023 - proceedings.mlr.press
We show that taking the width and depth to infinity in a deep neural network with skip
connections, when branches are scaled by $1/\sqrt {depth} $, result in the same covariance …

The influence of learning rule on representation dynamics in wide neural networks

B Bordelon, C Pehlevan - The Eleventh International Conference on …, 2022 - openreview.net
It is unclear how changing the learning rule of a deep neural network alters its learning
dynamics and representations. To gain insight into the relationship between learned …

Leave-one-out distinguishability in machine learning

J Ye, A Borovykh, S Hayou, R Shokri - arxiv preprint arxiv:2309.17310, 2023 - arxiv.org
We introduce a new analytical framework to quantify the changes in a machine learning
algorithm's output distribution following the inclusion of a few data points in its training set, a …

Neural (tangent kernel) collapse

M Seleznova, D Weitzner, R Giryes… - Advances in …, 2024 - proceedings.neurips.cc
This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures
the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) …

On the infinite-depth limit of finite-width neural networks

S Hayou - Transactions on Machine Learning Research, 2022 - openreview.net
In this paper, we study the infinite-depth limit of finite-width residual neural networks with
random Gaussian weights. With proper scaling, we show that by fixing the width and taking …

A spring-block theory of feature learning in deep neural networks

C Shi, L Pan, I Dokmanić - arxiv preprint arxiv:2407.19353, 2024 - arxiv.org
Feature-learning deep nets progressively collapse data to a regular low-dimensional
geometry. How this phenomenon emerges from collective action of nonlinearity, noise …

Self-consistent dynamical field theory of kernel evolution in wide neural networks

B Bordelon, C Pehlevan - Journal of Statistical Mechanics: Theory …, 2023 - iopscience.iop.org
We analyze feature learning in infinite-width neural networks trained with gradient flow
through a self-consistent dynamical field theory. We construct a collection of deterministic …