Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions

N Halko, PG Martinsson, JA Tropp - SIAM review, 2011 - SIAM
Low-rank matrix approximations, such as the truncated singular value decomposition and
the rank-revealing QR decomposition, play a central role in data analysis and scientific …

The shapley value in machine learning

B Rozemberczki, L Watson, P Bayer, HT Yang… - arxiv preprint arxiv …, 2022 - arxiv.org
Over the last few years, the Shapley value, a solution concept from cooperative game theory,
has found numerous applications in machine learning. In this paper, we first discuss …

Sketching as a tool for numerical linear algebra

DP Woodruff - … and Trends® in Theoretical Computer Science, 2014 - nowpublishers.com
This survey highlights the recent advances in algorithms for numerical linear algebra that
have come from the technique of linear sketching, whereby given a matrix, one first …

Towards efficient data valuation based on the shapley value

R Jia, D Dao, B Wang, FA Hubis… - The 22nd …, 2019 - proceedings.mlr.press
Abstract {\em “How much is my data worth?”} is an increasingly common question posed by
organizations and individuals alike. An answer to this question could allow, for instance …

Low-rank approximation and regression in input sparsity time

KL Clarkson, DP Woodruff - Journal of the ACM (JACM), 2017 - dl.acm.org
We design a new distribution over m× n matrices S so that, for any fixed n× d matrix A of rank
r, with probability at least 9/10,∥ SAx∥ 2=(1±ε)∥ Ax∥ 2 simultaneously for all x∈ R d …

[PDF][PDF] Finding structure with randomness: Stochastic algorithms for constructing approximate matrix decompositions

N Halko, PG Martinsson… - arxiv preprint …, 2009 - machinelearningbigdata.pbworks …
Low-rank matrix approximations, such as the truncated singular value decomposition and
the rank-revealing QR decomposition, play a central role in data analysis and scientific …

A unified framework for approximating and clustering data

D Feldman, M Langberg - Proceedings of the forty-third annual ACM …, 2011 - dl.acm.org
Given a set F of n positive functions over a ground set X, we consider the problem of
computing x* that minimizes the expression∑ f∈ Ff (x), over x∈ X. A typical application is …

OSNAP: Faster numerical linear algebra algorithms via sparser subspace embeddings

J Nelson, HL Nguyên - 2013 ieee 54th annual symposium on …, 2013 - ieeexplore.ieee.org
An oblivious subspace embedding (OSE) given some parameters ε, d is a distribution D over
matrices Π∈ R m× n such that for any linear subspace W⊆ R n with dim (W)= d, P Π~ D (∀ …

Granular ball sampling for noisy label classification or imbalanced classification

S **a, S Zheng, G Wang, X Gao… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
This article presents a general sampling method, called granular-ball sampling (GBS), for
classification problems by introducing the idea of granular computing. The GBS method …

Solving empirical risk minimization in the current matrix multiplication time

YT Lee, Z Song, Q Zhang - Conference on Learning Theory, 2019 - proceedings.mlr.press
Many convex problems in machine learning and computer science share the same
form:\begin {align*}\min_ {x}\sum_ {i} f_i (A_i x+ b_i),\end {align*} where $ f_i $ are convex …