Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions
Low-rank matrix approximations, such as the truncated singular value decomposition and
the rank-revealing QR decomposition, play a central role in data analysis and scientific …
the rank-revealing QR decomposition, play a central role in data analysis and scientific …
Literature survey on low rank approximation of matrices
N Kishore Kumar, J Schneider - Linear and Multilinear Algebra, 2017 - Taylor & Francis
Low rank approximation of matrices has been well studied in literature. Singular value
decomposition, QR decomposition with column pivoting, rank revealing QR factorization …
decomposition, QR decomposition with column pivoting, rank revealing QR factorization …
Sketching as a tool for numerical linear algebra
DP Woodruff - … and Trends® in Theoretical Computer Science, 2014 - nowpublishers.com
This survey highlights the recent advances in algorithms for numerical linear algebra that
have come from the technique of linear sketching, whereby given a matrix, one first …
have come from the technique of linear sketching, whereby given a matrix, one first …
Randomized numerical linear algebra: A perspective on the field with an eye to software
Randomized numerical linear algebra-RandNLA, for short-concerns the use of
randomization as a resource to develop improved algorithms for large-scale linear algebra …
randomization as a resource to develop improved algorithms for large-scale linear algebra …
Low-rank approximation and regression in input sparsity time
We design a new distribution over m× n matrices S so that, for any fixed n× d matrix A of rank
r, with probability at least 9/10,∥ SAx∥ 2=(1±ε)∥ Ax∥ 2 simultaneously for all x∈ R d …
r, with probability at least 9/10,∥ SAx∥ 2=(1±ε)∥ Ax∥ 2 simultaneously for all x∈ R d …
Core-sets: Updated survey
D Feldman - Sampling techniques for supervised or unsupervised …, 2020 - Springer
In optimization or machine learning problems we are given a set of items, usually points in
some metric space, and the goal is to minimize or maximize an objective function over some …
some metric space, and the goal is to minimize or maximize an objective function over some …
Turning Big Data Into Tiny Data: Constant-Size Coresets for -Means, PCA, and Projective Clustering
We develop and analyze a method to reduce the size of a very large set of data points in a
high-dimensional Euclidean space R^d to a small set of weighted points such that the result …
high-dimensional Euclidean space R^d to a small set of weighted points such that the result …
Dimensionality reduction for k-means clustering and low rank approximation
We show how to approximate a data matrix A with a much smaller sketch~ A that can be
used to solve a general class of constrained k-rank approximation problems to within (1+ ε) …
used to solve a general class of constrained k-rank approximation problems to within (1+ ε) …
A unified framework for approximating and clustering data
Given a set F of n positive functions over a ground set X, we consider the problem of
computing x* that minimizes the expression∑ f∈ Ff (x), over x∈ X. A typical application is …
computing x* that minimizes the expression∑ f∈ Ff (x), over x∈ X. A typical application is …
Improved approximation algorithms for large matrices via random projections
T Sarlos - 2006 47th annual IEEE symposium on foundations of …, 2006 - ieeexplore.ieee.org
Several results appeared that show significant reduction in time for matrix multiplication,
singular value decomposition as well as linear (lscr 2) regression, all based on data …
singular value decomposition as well as linear (lscr 2) regression, all based on data …