Representations and generalization in artificial and brain neural networks

Q Li, B Sorscher, H Sompolinsky - Proceedings of the National Academy of …, 2024 - pnas.org
Humans and animals excel at generalizing from limited data, a capability yet to be fully
replicated in artificial intelligence. This perspective investigates generalization in biological …

Task arithmetic in the tangent space: Improved editing of pre-trained models

G Ortiz-Jimenez, A Favero… - Advances in Neural …, 2023 - proceedings.neurips.cc
Task arithmetic has recently emerged as a cost-effective and scalable approach to edit pre-
trained models directly in weight space: By adding the fine-tuned weights of different tasks …

Structure-free graph condensation: From large-scale graphs to condensed graph-free data

X Zheng, M Zhang, C Chen… - Advances in …, 2023 - proceedings.neurips.cc
Graph condensation, which reduces the size of a large-scale graph by synthesizing a small-
scale condensed graph as its substitution, has immediate benefits for various graph learning …

Efficient dataset distillation using random feature approximation

N Loo, R Hasani, A Amini… - Advances in Neural …, 2022 - proceedings.neurips.cc
Dataset distillation compresses large datasets into smaller synthetic coresets which retain
performance with the aim of reducing the storage and computational burden of processing …

A kernel-based view of language model fine-tuning

S Malladi, A Wettig, D Yu, D Chen… - … on Machine Learning, 2023 - proceedings.mlr.press
It has become standard to solve NLP tasks by fine-tuning pre-trained language models
(LMs), especially in low-data settings. There is minimal theoretical understanding of …

A simple linear algebra identity to optimize large-scale neural network quantum states

R Rende, LL Viteritti, L Bardone, F Becca… - Communications …, 2024 - nature.com
Neural-network architectures have been increasingly used to represent quantum many-body
wave functions. These networks require a large number of variational parameters and are …

More than a toy: Random matrix models predict how real-world neural representations generalize

A Wei, W Hu, J Steinhardt - International conference on …, 2022 - proceedings.mlr.press
Of theories for why large-scale machine learning models generalize despite being vastly
overparameterized, which of their assumptions are needed to capture the qualitative …

Neural tangent kernel: A survey

E Golikov, E Pokonechnyy, V Korviakov - arxiv preprint arxiv:2208.13614, 2022 - arxiv.org
A seminal work [Jacot et al., 2018] demonstrated that training a neural network under
specific parameterization is equivalent to performing a particular kernel method as width …

[HTML][HTML] Deep networks for system identification: a survey

G Pillonetto, A Aravkin, D Gedon, L Ljung, AH Ribeiro… - Automatica, 2025 - Elsevier
Deep learning is a topic of considerable current interest. The availability of massive data
collections and powerful software resources has led to an impressive amount of results in …

Kecor: Kernel coding rate maximization for active 3d object detection

Y Luo, Z Chen, Z Fang, Z Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Achieving a reliable LiDAR-based object detector in autonomous driving is paramount, but
its success hinges on obtaining large amounts of precise 3D annotations. Active learning …