Neural collapse: A review on modelling principles and generalization

V Kothapalli - arxiv preprint arxiv:2206.04041, 2022 - arxiv.org
Deep classifier neural networks enter the terminal phase of training (TPT) when training
error reaches zero and tend to exhibit intriguing Neural Collapse (NC) properties. Neural …

Recent advances of continual learning in computer vision: An overview

H Qu, H Rahmani, L Xu, B Williams, J Liu - arxiv preprint arxiv …, 2021 - arxiv.org
In contrast to batch learning where all training data is available at once, continual learning
represents a family of methods that accumulate knowledge and learn continuously with data …

A geometric analysis of neural collapse with unconstrained features

Z Zhu, T Ding, J Zhou, X Li, C You… - Advances in Neural …, 2021 - proceedings.neurips.cc
We provide the first global optimization landscape analysis of Neural Collapse--an intriguing
empirical phenomenon that arises in the last-layer classifiers and features of neural …

Neural collapse with normalized features: A geometric analysis over the riemannian manifold

C Yaras, P Wang, Z Zhu… - Advances in neural …, 2022 - proceedings.neurips.cc
When training overparameterized deep networks for classification tasks, it has been widely
observed that the learned features exhibit a so-called" neural collapse'" phenomenon. More …

Derivative-informed projected neural networks for high-dimensional parametric maps governed by PDEs

T O'Leary-Roseberry, U Villa, P Chen… - Computer Methods in …, 2022 - Elsevier
Many-query problems–arising from, eg, uncertainty quantification, Bayesian inversion,
Bayesian optimal experimental design, and optimization under uncertainty–require …

Incremental learning via rate reduction

Z Wu, C Baek, C You, Y Ma - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Current deep learning architectures suffer from catastrophic forgetting, a failure to retain
knowledge of previously learned classes when incrementally trained on new classes. The …

Optimization inspired multi-branch equilibrium models

M Li, Y Wang, X **e, Z Lin - International Conference on Learning …, 2022 - openreview.net
Works have shown the strong connections between some implicit models and optimization
problems. However, explorations on such relationships are limited. Most works pay attention …

[HTML][HTML] Terahertz nanoscopy: Advances, challenges, and the road ahead

X Guo, K Bertling, BC Donose, M Brünig… - Applied Physics …, 2024 - pubs.aip.org
Exploring nanoscale material properties through light-matter interactions is essential to
unveil new phenomena and manipulate materials at the atomic level, paving the way for …

Designing Universally-Approximating Deep Neural Networks: A First-Order Optimization Approach

Z Wu, M **ao, C Fang, Z Lin - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Universal approximation capability, also referred to as universality, is an important property
of deep neural networks, endowing them with the potency to accurately represent the …

Variance of the gradient also matters: Privacy leakage from gradients

Y Wang, J Deng, D Guo, C Wang… - … Joint Conference on …, 2022 - ieeexplore.ieee.org
Distributed machine learning (DML) enables model training on a large corpus of
decentralized data from users and only collects local models or gradients for global …