Hidden convexity of wasserstein GANs: Interpretable generative models with closed-form solutions

A Sahiner, T Ergen, B Ozturkler, B Bartan… - arxiv preprint arxiv …, 2021 - arxiv.org
Generative Adversarial Networks (GANs) are commonly used for modeling complex
distributions of data. Both the generators and discriminators of GANs are often modeled by …

Globally optimal training of neural networks with threshold activation functions

T Ergen, HI Gulluk, J Lacotte, M Pilanci - arxiv preprint arxiv:2303.03382, 2023 - arxiv.org
Threshold activation functions are highly preferable in neural networks due to their efficiency
in hardware implementations. Moreover, their mode of operation is more interpretable and …

Optimal neural network approximation of wasserstein gradient direction via convex optimization

Y Wang, P Chen, M Pilanci, W Li - SIAM Journal on Mathematics of Data …, 2024 - SIAM
The calculation of the direction of the Wasserstein gradient is vital for addressing problems
related to posterior sampling and scientific computing. To approximate the Wasserstein …

Parallel deep neural networks have zero duality gap

Y Wang, T Ergen, M Pilanci - arxiv preprint arxiv:2110.06482, 2021 - arxiv.org
Training deep neural networks is a challenging non-convex optimization problem. Recent
work has proven that the strong duality holds (which means zero duality gap) for regularized …

The convex geometry of backpropagation: Neural network gradient flows converge to extreme points of the dual convex program

Y Wang, M Pilanci - arxiv preprint arxiv:2110.06488, 2021 - arxiv.org
We study non-convex subgradient flows for training two-layer ReLU neural networks from a
convex geometry and duality perspective. We characterize the implicit bias of unregularized …

A decomposition augmented lagrangian method for low-rank semidefinite programming

Y Wang, K Deng, H Liu, Z Wen - SIAM Journal on Optimization, 2023 - SIAM
We develop a decomposition method based on the augmented Lagrangian framework to
solve a broad family of semidefinite programming problems, possibly with nonlinear …

Training quantized neural networks to global optimality via semidefinite programming

B Bartan, M Pilanci - International Conference on Machine …, 2021 - proceedings.mlr.press
Neural networks (NNs) have been extremely successful across many tasks in machine
learning. Quantization of NN weights has become an important topic due to its impact on …

Overparameterized relu neural networks learn the simplest models: Neural isometry and exact recovery

Y Wang, Y Hua, E Candés, M Pilanci - arxiv preprint arxiv:2209.15265, 2022 - arxiv.org
The practice of deep learning has shown that neural networks generalize remarkably well
even with an extreme number of learned parameters. This appears to contradict traditional …

Neural Fisher discriminant analysis: Optimal neural network embeddings in polynomial time

B Bartan, M Pilanci - International Conference on Machine …, 2022 - proceedings.mlr.press
Abstract Fisher's Linear Discriminant Analysis (FLDA) is a statistical analysis method that
linearly embeds data points to a lower dimensional space to maximize a discrimination …

System identification and control using quadratic neural networks

L Rodrigues, S Givigi - IEEE Control Systems Letters, 2023 - ieeexplore.ieee.org
This letter proposes convex formulations of system identification and control for nonlinear
systems using two layer quadratic neural networks. The results in this letter cast system …