Neural collapse: A review on modelling principles and generalization

V Kothapalli - arxiv preprint arxiv:2206.04041, 2022 - arxiv.org
Deep classifier neural networks enter the terminal phase of training (TPT) when training
error reaches zero and tend to exhibit intriguing Neural Collapse (NC) properties. Neural …

On the optimization landscape of neural collapse under mse loss: Global optimality with unconstrained features

J Zhou, X Li, T Ding, C You, Q Qu… - … on Machine Learning, 2022 - proceedings.mlr.press
When training deep neural networks for classification tasks, an intriguing empirical
phenomenon has been widely observed in the last-layer classifiers and features, where (i) …

A geometric analysis of neural collapse with unconstrained features

Z Zhu, T Ding, J Zhou, X Li, C You… - Advances in Neural …, 2021 - proceedings.neurips.cc
We provide the first global optimization landscape analysis of Neural Collapse--an intriguing
empirical phenomenon that arises in the last-layer classifiers and features of neural …

Understanding imbalanced semantic segmentation through neural collapse

Z Zhong, J Cui, Y Yang, X Wu, X Qi… - Proceedings of the …, 2023 - openaccess.thecvf.com
A recent study has shown a phenomenon called neural collapse in that the within-class
means of features and the classifier weight vectors converge to the vertices of a simplex …

Inducing neural collapse in imbalanced learning: Do we really need a learnable classifier at the end of deep neural network?

Y Yang, S Chen, X Li, L **e, Z Lin… - Advances in neural …, 2022 - proceedings.neurips.cc
Modern deep neural networks for classification usually jointly learn a backbone for
representation and a linear classifier to output the logit of each class. A recent study has …

Extended unconstrained features model for exploring deep neural collapse

T Tirer, J Bruna - International Conference on Machine …, 2022 - proceedings.mlr.press
The modern strategy for training deep neural networks for classification tasks includes
optimizing the network's weights even after the training error vanishes to further push the …

Are all losses created equal: A neural collapse perspective

J Zhou, C You, X Li, K Liu, S Liu… - Advances in Neural …, 2022 - proceedings.neurips.cc
While cross entropy (CE) is the most commonly used loss function to train deep neural
networks for classification tasks, many alternative losses have been developed to obtain …

Imbalance trouble: Revisiting neural-collapse geometry

C Thrampoulidis, GR Kini… - Advances in Neural …, 2022 - proceedings.neurips.cc
Neural Collapse refers to the remarkable structural properties characterizing the geometry of
class embeddings and classifier weights, found by deep nets when trained beyond zero …

Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training

C Fang, H He, Q Long, WJ Su - Proceedings of the National …, 2021 - National Acad Sciences
In this paper, we introduce the Layer-Peeled Model, a nonconvex, yet analytically tractable,
optimization program, in a quest to better understand deep neural networks that are trained …

Neural collapse under mse loss: Proximity to and dynamics on the central path

XY Han, V Papyan, DL Donoho - arxiv preprint arxiv:2106.02073, 2021 - arxiv.org
The recently discovered Neural Collapse (NC) phenomenon occurs pervasively in today's
deep net training paradigm of driving cross-entropy (CE) loss towards zero. During NC, last …