Complete dictionary recovery over the sphere I: Overview and the geometric picture
We consider the problem of recovering a complete (ie, square and invertible) matrix A 0,
from Y∈ R n× p with Y= A 0 X 0, provided X 0 is sufficiently sparse. This recovery problem is …
from Y∈ R n× p with Y= A 0 X 0, provided X 0 is sufficiently sparse. This recovery problem is …
Recent scalability improvements for semidefinite programming with applications in machine learning, control, and robotics
Historically, scalability has been a major challenge for the successful application of
semidefinite programming in fields such as machine learning, control, and robotics. In this …
semidefinite programming in fields such as machine learning, control, and robotics. In this …
[BUCH][B] An introduction to optimization on smooth manifolds
N Boumal - 2023 - books.google.com
Optimization on Riemannian manifolds-the result of smooth geometry and optimization
merging into one elegant modern framework-spans many areas of science and engineering …
merging into one elegant modern framework-spans many areas of science and engineering …
Learning single-index models with shallow neural networks
Single-index models are a class of functions given by an unknown univariate``link''function
applied to an unknown one-dimensional projection of the input. These models are …
applied to an unknown one-dimensional projection of the input. These models are …
What is local optimality in nonconvex-nonconcave minimax optimization?
Minimax optimization has found extensive applications in modern machine learning, in
settings such as generative adversarial networks (GANs), adversarial training and multi …
settings such as generative adversarial networks (GANs), adversarial training and multi …
A geometric analysis of phase retrieval
Can we recover a complex signal from its Fourier magnitudes? More generally, given a set
of m measurements, y_k=\left| a _k^* x\right| yk= ak∗ x for k= 1, ..., mk= 1,…, m, is it possible …
of m measurements, y_k=\left| a _k^* x\right| yk= ak∗ x for k= 1, ..., mk= 1,…, m, is it possible …
On the power of over-parametrization in neural networks with quadratic activation
We provide new theoretical insights on why over-parametrization is effective in learning
neural networks. For a $ k $ hidden node shallow network with quadratic activation and $ n …
neural networks. For a $ k $ hidden node shallow network with quadratic activation and $ n …
Accelerated gradient descent escapes saddle points faster than gradient descent
Nesterov's accelerated gradient descent (AGD), an instance of the general family of
“momentum methods,” provably achieves faster convergence rate than gradient descent …
“momentum methods,” provably achieves faster convergence rate than gradient descent …
SE-Sync: A certifiably correct algorithm for synchronization over the special Euclidean group
Many important geometric estimation problems naturally take the form of synchronization
over the special Euclidean group: estimate the values of a set of unknown group elements x …
over the special Euclidean group: estimate the values of a set of unknown group elements x …
Global rates of convergence for nonconvex optimization on manifolds
We consider the minimization of a cost function f on a manifold using Riemannian gradient
descent and Riemannian trust regions (RTR). We focus on satisfying necessary optimality …
descent and Riemannian trust regions (RTR). We focus on satisfying necessary optimality …