[HTML][HTML] Cleaning large correlation matrices: tools from random matrix theory
J Bun, JP Bouchaud, M Potters - Physics Reports, 2017 - Elsevier
This review covers recent results concerning the estimation of large covariance matrices
using tools from Random Matrix Theory (RMT). We introduce several RMT methods and …
using tools from Random Matrix Theory (RMT). We introduce several RMT methods and …
[HTML][HTML] Surprises in high-dimensional ridgeless least squares interpolation
Interpolators—estimators that achieve zero training error—have attracted growing attention
in machine learning, mainly because state-of-the art neural networks appear to be models of …
in machine learning, mainly because state-of-the art neural networks appear to be models of …
[HTML][HTML] Complex diffusion-weighted image estimation via matrix recovery under general noise models
We propose a patch-based singular value shrinkage method for diffusion magnetic
resonance image estimation targeted at low signal to noise ratio and accelerated …
resonance image estimation targeted at low signal to noise ratio and accelerated …
High-dimensional asymptotics of prediction: Ridge regression and classification
E Dobriban, S Wager - The Annals of Statistics, 2018 - JSTOR
We provide a unified analysis of the predictive risk of ridge regression and regularized
discriminant analysis in a dense random effects model. We work in a high-dimensional …
discriminant analysis in a dense random effects model. We work in a high-dimensional …
A continuous-time view of early stop** for least squares regression
We study the statistical properties of the iterates generated by gradient descent, applied to
the fundamental problem of least squares regression. We take a continuous-time view, ie …
the fundamental problem of least squares regression. We take a continuous-time view, ie …
Estimation of the number of spiked eigenvalues in a covariance matrix by bulk eigenvalue matching analysis
ZT Ke, Y Ma, X Lin - Journal of the American Statistical Association, 2023 - Taylor & Francis
The spiked covariance model has gained increasing popularity in high-dimensional data
analysis. A fundamental problem is determination of the number of spiked eigenvalues, K …
analysis. A fundamental problem is determination of the number of spiked eigenvalues, K …
Deterministic parallel analysis: an improved method for selecting factors and principal components
E Dobriban, AB Owen - Journal of the Royal Statistical Society …, 2019 - academic.oup.com
Factor analysis and principal component analysis are used in many application areas. The
first step, choosing the number of components, remains a serious challenge. Our work …
first step, choosing the number of components, remains a serious challenge. Our work …
Nonparametric Bayesian posterior contraction rates for discretely observed scalar diffusions
Nonparametric Bayesian posterior contraction rates for discretely observed scalar diffusions
Page 1 The Annals of Statistics 2017, Vol. 45, No. 4, 1664–1693 DOI: 10.1214/16-AOS1504 © …
Page 1 The Annals of Statistics 2017, Vol. 45, No. 4, 1664–1693 DOI: 10.1214/16-AOS1504 © …
How to reduce dimension with PCA and random projections?
In our “big data” age, the size and complexity of data is steadily increasing. Methods for
dimension reduction are ever more popular and useful. Two distinct types of dimension …
dimension reduction are ever more popular and useful. Two distinct types of dimension …
Asymptotics of the sketched pseudoinverse
We take a random matrix theory approach to random sketching and show an asymptotic first-
order equivalence of the regularized sketched pseudoinverse of a positive semidefinite …
order equivalence of the regularized sketched pseudoinverse of a positive semidefinite …