Neural collapse: A review on modelling principles and generalization
V Kothapalli - arxiv preprint arxiv:2206.04041, 2022 - arxiv.org
Deep classifier neural networks enter the terminal phase of training (TPT) when training
error reaches zero and tend to exhibit intriguing Neural Collapse (NC) properties. Neural …
error reaches zero and tend to exhibit intriguing Neural Collapse (NC) properties. Neural …
Recent advances of continual learning in computer vision: An overview
In contrast to batch learning where all training data is available at once, continual learning
represents a family of methods that accumulate knowledge and learn continuously with data …
represents a family of methods that accumulate knowledge and learn continuously with data …
A geometric analysis of neural collapse with unconstrained features
We provide the first global optimization landscape analysis of Neural Collapse--an intriguing
empirical phenomenon that arises in the last-layer classifiers and features of neural …
empirical phenomenon that arises in the last-layer classifiers and features of neural …
Neural collapse with normalized features: A geometric analysis over the riemannian manifold
When training overparameterized deep networks for classification tasks, it has been widely
observed that the learned features exhibit a so-called" neural collapse'" phenomenon. More …
observed that the learned features exhibit a so-called" neural collapse'" phenomenon. More …
Derivative-informed projected neural networks for high-dimensional parametric maps governed by PDEs
Many-query problems–arising from, eg, uncertainty quantification, Bayesian inversion,
Bayesian optimal experimental design, and optimization under uncertainty–require …
Bayesian optimal experimental design, and optimization under uncertainty–require …
Incremental learning via rate reduction
Current deep learning architectures suffer from catastrophic forgetting, a failure to retain
knowledge of previously learned classes when incrementally trained on new classes. The …
knowledge of previously learned classes when incrementally trained on new classes. The …
Optimization inspired multi-branch equilibrium models
Works have shown the strong connections between some implicit models and optimization
problems. However, explorations on such relationships are limited. Most works pay attention …
problems. However, explorations on such relationships are limited. Most works pay attention …
[HTML][HTML] Terahertz nanoscopy: Advances, challenges, and the road ahead
Exploring nanoscale material properties through light-matter interactions is essential to
unveil new phenomena and manipulate materials at the atomic level, paving the way for …
unveil new phenomena and manipulate materials at the atomic level, paving the way for …
Designing Universally-Approximating Deep Neural Networks: A First-Order Optimization Approach
Universal approximation capability, also referred to as universality, is an important property
of deep neural networks, endowing them with the potency to accurately represent the …
of deep neural networks, endowing them with the potency to accurately represent the …
Variance of the gradient also matters: Privacy leakage from gradients
Distributed machine learning (DML) enables model training on a large corpus of
decentralized data from users and only collects local models or gradients for global …
decentralized data from users and only collects local models or gradients for global …