Nonconvex optimization meets low-rank matrix factorization: An overview

Y Chi, YM Lu, Y Chen - IEEE Transactions on Signal …, 2019 - ieeexplore.ieee.org
Substantial progress has been made recently on develo** provably accurate and efficient
algorithms for low-rank matrix factorization via nonconvex optimization. While conventional …

From inverse optimal control to inverse reinforcement learning: A historical review

N Ab Azar, A Shahmansoorian, M Davoudi - Annual Reviews in Control, 2020 - Elsevier
Inverse optimal control (IOC) is a powerful theory that addresses the inverse problems in
control systems, robotics, Machine Learning (ML) and optimization taking into account the …

A survey of optimization methods from a machine learning perspective

S Sun, Z Cao, H Zhu, J Zhao - IEEE transactions on cybernetics, 2019 - ieeexplore.ieee.org
Machine learning develops rapidly, which has made many theoretical breakthroughs and is
widely applied in various fields. Optimization, as an important part of machine learning, has …

Interpretable hyperspectral artificial intelligence: When nonconvex modeling meets hyperspectral remote sensing

D Hong, W He, N Yokoya, J Yao, L Gao… - … and Remote Sensing …, 2021 - ieeexplore.ieee.org
Hyperspectral (HS) imaging, also known as image spectrometry, is a landmark technique in
geoscience and remote sensing (RS). In the past decade, enormous efforts have been made …

Parametric deep energy approach for elasticity accounting for strain gradient effects

VM Nguyen-Thanh, C Anitescu, N Alajlan… - Computer Methods in …, 2021 - Elsevier
In this work, we present a Parametric Deep Energy Method (P-DEM) for elasticity problems
accounting for strain gradient effects. The approach is based on physics-informed neural …

Spider: Near-optimal non-convex optimization via stochastic path-integrated differential estimator

C Fang, CJ Li, Z Lin, T Zhang - Advances in neural …, 2018 - proceedings.neurips.cc
In this paper, we propose a new technique named\textit {Stochastic Path-Integrated
Differential EstimatoR}(SPIDER), which can be used to track many deterministic quantities of …

A sufficient condition for convergences of adam and rmsprop

F Zou, L Shen, Z Jie, W Zhang… - Proceedings of the IEEE …, 2019 - openaccess.thecvf.com
Adam and RMSProp are two of the most influential adaptive stochastic algorithms for
training deep neural networks, which have been pointed out to be divergent even in the …

EF21: A new, simpler, theoretically better, and practically faster error feedback

P Richtárik, I Sokolov… - Advances in Neural …, 2021 - proceedings.neurips.cc
Error feedback (EF), also known as error compensation, is an immensely popular
convergence stabilization mechanism in the context of distributed training of supervised …

Robustness to unbounded smoothness of generalized signsgd

M Crawshaw, M Liu, F Orabona… - Advances in neural …, 2022 - proceedings.neurips.cc
Traditional analyses in non-convex optimization typically rely on the smoothness
assumption, namely requiring the gradients to be Lipschitz. However, recent evidence …

Videoprism: A foundational visual encoder for video understanding

L Zhao, NB Gundavarapu, L Yuan, H Zhou… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce VideoPrism, a general-purpose video encoder that tackles diverse video
understanding tasks with a single frozen model. We pretrain VideoPrism on a …