Implicit regularization in nonconvex statistical estimation: Gradient descent converges linearly for phase retrieval and matrix completion

C Ma, K Wang, Y Chi, Y Chen - International Conference on …, 2018 - proceedings.mlr.press
Recent years have seen a flurry of activities in designing provably efficient nonconvex
optimization procedures for solving statistical estimation problems. For various problems like …

Gradient descent with random initialization: Fast global convergence for nonconvex phase retrieval

Y Chen, Y Chi, J Fan, C Ma - Mathematical Programming, 2019 - Springer
This paper considers the problem of solving systems of quadratic equations, namely,
recovering an object of interest x^ ♮ ∈ R^ nx♮∈ R n from m quadratic equations/samples …

Solving random quadratic systems of equations is nearly as easy as solving linear systems

Y Chen, E Candes - Advances in Neural Information …, 2015 - proceedings.neurips.cc
This paper is concerned with finding a solution x to a quadratic system of equations yi=|< ai,
x>|^ 2, i= 1, 2,..., m. We prove that it is possible to solve unstructured quadratic systems in n …

Solving random quadratic systems of equations is nearly as easy as solving linear systems

Y Chen, EJ Candès - Communications on pure and applied …, 2017 - Wiley Online Library
We consider the fundamental problem of solving quadratic systems of equations in, and is
unknown. We propose a novel method, which starts with an initial guess computed by …

Demystifying softmax gating function in Gaussian mixture of experts

H Nguyen, TT Nguyen, N Ho - Advances in Neural …, 2023 - proceedings.neurips.cc
Understanding the parameter estimation of softmax gating Gaussian mixture of experts has
remained a long-standing open problem in the literature. It is mainly due to three …

Exact and stable covariance estimation from quadratic sampling via convex programming

Y Chen, Y Chi, AJ Goldsmith - IEEE Transactions on …, 2015 - ieeexplore.ieee.org
Statistical inference and information processing of high-dimensional data often require an
efficient and accurate estimation of their second-order statistics. With rapidly changing data …

Reducibility and statistical-computational gaps from secret leakage

M Brennan, G Bresler - Conference on Learning Theory, 2020 - proceedings.mlr.press
Inference problems with conjectured statistical-computational gaps are ubiquitous
throughout modern statistics, computer science, statistical physics and discrete probability …

List-decodable linear regression

S Karmalkar, A Klivans… - Advances in neural …, 2019 - proceedings.neurips.cc
List-decodable Linear Regression Page 1 List-decodeable Linear Regression Sushrut
Karmalkar University of Texas at Austin sushrutk@cs.utexas.edu Adam R. Klivans University of …

List decodable learning via sum of squares

P Raghavendra, M Yau - Proceedings of the Fourteenth Annual ACM-SIAM …, 2020 - SIAM
In the list-decodable learning setup, an overwhelming majority (say a 1–β-fraction) of the
input data consists of outliers and the goal of an algorithm is to output a small list of …

Learning mixtures of linear regressions with nearly optimal complexity

Y Li, Y Liang - Conference On Learning Theory, 2018 - proceedings.mlr.press
Abstract Mixtures of Linear Regressions (MLR) is an important mixture model with many
applications. In this model, each observation is generated from one of the several unknown …