Local kernel renormalization as a mechanism for feature learning in overparametrized convolutional neural networks

R Aiudi, R Pacelli, P Baglioni, A Vezzani… - Nature …, 2025 - nature.com
Empirical evidence shows that fully-connected neural networks in the infinite-width limit (lazy
training) eventually outperform their finite-width counterparts in most computer vision tasks; …

A spring-block theory of feature learning in deep neural networks

C Shi, L Pan, I Dokmanić - arxiv preprint arxiv:2407.19353, 2024 - arxiv.org
Feature-learning deep nets progressively collapse data to a regular low-dimensional
geometry. How this phenomenon emerges from collective action of nonlinearity, noise …

Adaptive kernel predictors from feature-learning infinite limits of neural networks

C Lauditi, B Bordelon, C Pehlevan - arxiv preprint arxiv:2502.07998, 2025 - arxiv.org
Previous influential work showed that infinite width limits of neural networks in the lazy
training regime are described by kernel machines. Here, we show that neural networks …

Deep Linear Network Training Dynamics from Random Initialization: Data, Width, Depth, and Hyperparameter Transfer

B Bordelon, C Pehlevan - arxiv preprint arxiv:2502.02531, 2025 - arxiv.org
We theoretically characterize gradient descent dynamics in deep linear networks trained at
large width from random initialization and on large quantities of random data. Our theory …

Feature learning in finite-width Bayesian deep linear networks with multiple outputs and convolutional layers

F Bassetti, M Gherardi, A Ingrosso, M Pastore… - arxiv preprint arxiv …, 2024 - arxiv.org
Deep linear networks have been extensively studied, as they provide simplified models of
deep learning. However, little is known in the case of finite-width architectures with multiple …

Proportional infinite-width infinite-depth limit for deep linear neural networks

F Bassetti, L Ladelli, P Rotondo - arxiv preprint arxiv:2411.15267, 2024 - arxiv.org
We study the distributional properties of linear neural networks with random parameters in
the context of large networks, where the number of layers diverges in proportion to the …

[PDF][PDF] Confronting Large Fluctuations in Numerical Stochastic Perturbation Theory

P Baglioni - 2024 - repository.unipr.it
Perturbation theory is universally recognized as a fundamental tool in modern theoretical
physics. In the functional integral formalism, perturbation theory provides a method for …

Kernel Shape Renormalization In Bayesian Shallow Networks: a Gaussian Process Perspective

R Pacelli, L Giambagli… - 2024 IEEE Workshop on …, 2024 - ieeexplore.ieee.org
The Bayesian approach has proven to be a valuable tool for analytical inspection of neural
networks. Recent theoretical advances have led to the development of an effective statistical …