Hidden convexity of wasserstein GANs: Interpretable generative models with closed-form solutions
Generative Adversarial Networks (GANs) are commonly used for modeling complex
distributions of data. Both the generators and discriminators of GANs are often modeled by …
distributions of data. Both the generators and discriminators of GANs are often modeled by …
Globally optimal training of neural networks with threshold activation functions
Threshold activation functions are highly preferable in neural networks due to their efficiency
in hardware implementations. Moreover, their mode of operation is more interpretable and …
in hardware implementations. Moreover, their mode of operation is more interpretable and …
Optimal neural network approximation of wasserstein gradient direction via convex optimization
The calculation of the direction of the Wasserstein gradient is vital for addressing problems
related to posterior sampling and scientific computing. To approximate the Wasserstein …
related to posterior sampling and scientific computing. To approximate the Wasserstein …
Parallel deep neural networks have zero duality gap
Training deep neural networks is a challenging non-convex optimization problem. Recent
work has proven that the strong duality holds (which means zero duality gap) for regularized …
work has proven that the strong duality holds (which means zero duality gap) for regularized …
The convex geometry of backpropagation: Neural network gradient flows converge to extreme points of the dual convex program
We study non-convex subgradient flows for training two-layer ReLU neural networks from a
convex geometry and duality perspective. We characterize the implicit bias of unregularized …
convex geometry and duality perspective. We characterize the implicit bias of unregularized …
A decomposition augmented lagrangian method for low-rank semidefinite programming
We develop a decomposition method based on the augmented Lagrangian framework to
solve a broad family of semidefinite programming problems, possibly with nonlinear …
solve a broad family of semidefinite programming problems, possibly with nonlinear …
Training quantized neural networks to global optimality via semidefinite programming
Neural networks (NNs) have been extremely successful across many tasks in machine
learning. Quantization of NN weights has become an important topic due to its impact on …
learning. Quantization of NN weights has become an important topic due to its impact on …
Overparameterized relu neural networks learn the simplest models: Neural isometry and exact recovery
The practice of deep learning has shown that neural networks generalize remarkably well
even with an extreme number of learned parameters. This appears to contradict traditional …
even with an extreme number of learned parameters. This appears to contradict traditional …
Neural Fisher discriminant analysis: Optimal neural network embeddings in polynomial time
Abstract Fisher's Linear Discriminant Analysis (FLDA) is a statistical analysis method that
linearly embeds data points to a lower dimensional space to maximize a discrimination …
linearly embeds data points to a lower dimensional space to maximize a discrimination …
System identification and control using quadratic neural networks
This letter proposes convex formulations of system identification and control for nonlinear
systems using two layer quadratic neural networks. The results in this letter cast system …
systems using two layer quadratic neural networks. The results in this letter cast system …