Ai alignment: A comprehensive survey

J Ji, T Qiu, B Chen, B Zhang, H Lou, K Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
AI alignment aims to make AI systems behave in line with human intentions and values. As
AI systems grow more capable, so do risks from misalignment. To provide a comprehensive …

Deep model fusion: A survey

W Li, Y Peng, M Zhang, L Ding, H Hu… - arxiv preprint arxiv …, 2023 - arxiv.org
Deep model fusion/merging is an emerging technique that merges the parameters or
predictions of multiple deep learning models into a single one. It combines the abilities of …

Repair: Renormalizing permuted activations for interpolation repair

K Jordan, H Sedghi, O Saukh, R Entezari… - arxiv preprint arxiv …, 2022 - arxiv.org
In this paper we look into the conjecture of Entezari et al.(2021) which states that if the
permutation invariance of neural networks is taken into account, then there is likely no loss …

Mechanistic mode connectivity

ES Lubana, EJ Bigelow, RP Dick… - International …, 2023 - proceedings.mlr.press
We study neural network loss landscapes through the lens of mode connectivity, the
observation that minimizers of neural networks retrieved via training on a dataset are …

Class incremental learning with multi-teacher distillation

H Wen, L Pan, Y Dai, H Qiu, L Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Distillation strategies are currently the primary approaches for mitigating forgetting in class
incremental learning (CIL). Existing methods generally inherit previous knowledge from a …

Proving linear mode connectivity of neural networks via optimal transport

D Ferbach, B Goujaud, G Gidel… - International …, 2024 - proceedings.mlr.press
The energy landscape of high-dimensional non-convex optimization problems is crucial to
understanding the effectiveness of modern deep neural network architectures. Recent works …

The empirical impact of neural parameter symmetries, or lack thereof

D Lim, TM Putterman, R Walters, H Maron… - arxiv preprint arxiv …, 2024 - arxiv.org
Many algorithms and observed phenomena in deep learning appear to be affected by
parameter symmetries--transformations of neural network parameters that do not change the …

Topological obstruction to the training of shallow ReLU neural networks

M Nurisso, P Leroy, F Vaccarino - Advances in Neural …, 2025 - proceedings.neurips.cc
Studying the interplay between the geometry of the loss landscape and the optimization
trajectories of simple neural networks is a fundamental step for understanding their behavior …

Learning through atypical phase transitions in overparameterized neural networks

C Baldassi, C Lauditi, EM Malatesta, R Pacelli… - Physical Review E, 2022 - APS
Current deep neural networks are highly overparameterized (up to billions of connection
weights) and nonlinear. Yet they can fit data almost perfectly through variants of gradient …

Symmetries, flat minima, and the conserved quantities of gradient flow

B Zhao, I Ganev, R Walters, R Yu… - arxiv preprint arxiv …, 2022 - arxiv.org
Empirical studies of the loss landscape of deep networks have revealed that many local
minima are connected through low-loss valleys. Yet, little is known about the theoretical …