AUC maximization in the era of big data and AI: A survey

T Yang, Y Ying - ACM computing surveys, 2022 - dl.acm.org
Area under the ROC curve, aka AUC, is a measure of choice for assessing the performance
of a classifier for imbalanced data. AUC maximization refers to a learning paradigm that …

Adan: Adaptive nesterov momentum algorithm for faster optimizing deep models

X **e, P Zhou, H Li, Z Lin, S Yan - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
In deep learning, different kinds of deep networks typically need different optimizers, which
have to be chosen after multiple trials, making the training process inefficient. To relieve this …

Convergence of adam under relaxed assumptions

H Li, A Rakhlin, A Jadbabaie - Advances in Neural …, 2023 - proceedings.neurips.cc
In this paper, we provide a rigorous proof of convergence of the Adaptive Moment Estimate
(Adam) algorithm for a wide class of optimization objectives. Despite the popularity and …

Adam can converge without any modification on update rules

Y Zhang, C Chen, N Shi, R Sun… - Advances in neural …, 2022 - proceedings.neurips.cc
Ever since\citet {reddi2019convergence} pointed out the divergence issue of Adam, many
new variants have been designed to obtain convergence. However, vanilla Adam remains …

Provably faster algorithms for bilevel optimization

J Yang, K Ji, Y Liang - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Bilevel optimization has been widely applied in many important machine learning
applications such as hyperparameter optimization and meta-learning. Recently, several …

A framework for bilevel optimization that enables stochastic and global variance reduction algorithms

M Dagréou, P Ablin, S Vaiter… - Advances in Neural …, 2022 - proceedings.neurips.cc
Bilevel optimization, the problem of minimizing a value function which involves the arg-
minimum of another function, appears in many areas of machine learning. In a large scale …

Fednest: Federated bilevel, minimax, and compositional optimization

DA Tarzanagh, M Li… - … on Machine Learning, 2022 - proceedings.mlr.press
Standard federated optimization methods successfully apply to stochastic problems with
single-level structure. However, many contemporary ML problems-including adversarial …

Closing the gap between the upper bound and lower bound of Adam's iteration complexity

B Wang, J Fu, H Zhang, N Zheng… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Recently, Arjevani et al.[1] establish a lower bound of iteration complexity for the
first-order optimization under an $ L $-smooth condition and a bounded noise variance …

A fully single loop algorithm for bilevel optimization without hessian inverse

J Li, B Gu, H Huang - Proceedings of the AAAI Conference on Artificial …, 2022 - ojs.aaai.org
In this paper, we propose a novel Hessian inverse free Fully Single Loop Algorithm (FSLA)
for bilevel optimization problems. Classic algorithms for bilevel optimization admit a double …

The power of adaptivity in sgd: Self-tuning step sizes with unbounded gradients and affine variance

M Faw, I Tziotis, C Caramanis… - … on Learning Theory, 2022 - proceedings.mlr.press
We study convergence rates of AdaGrad-Norm as an exemplar of adaptive stochastic
gradient methods (SGD), where the step sizes change based on observed stochastic …