Mixed precision algorithms in numerical linear algebra

NJ Higham, T Mary - Acta Numerica, 2022 - cambridge.org
Today's floating-point arithmetic landscape is broader than ever. While scientific computing
has traditionally used single precision and double precision floating-point arithmetics, half …

Simulating low precision floating-point arithmetic

NJ Higham, S Pranesh - SIAM Journal on Scientific Computing, 2019 - SIAM
The half-precision (fp16) floating-point format, defined in the 2008 revision of the IEEE
standard for floating-point arithmetic, and a more recently proposed half-precision format …

A class of fast and accurate summation algorithms

P Blanchard, NJ Higham, T Mary - SIAM journal on scientific computing, 2020 - SIAM
The need to sum floating-point numbers is ubiquitous in scientific computing. Standard
recursive summation of n summands, often implemented in a blocked form, has a backward …

Mixed precision iterative refinement with sparse approximate inverse preconditioning

E Carson, N Khan - SIAM Journal on Scientific Computing, 2023 - SIAM
With the commercial availability of mixed precision hardware, mixed precision GMRES-
based iterative refinement schemes have emerged as popular approaches for solving …

Random matrices generating large growth in LU factorization with pivoting

DJ Higham, NJ Higham, S Pranesh - SIAM Journal on Matrix Analysis and …, 2021 - SIAM
We identify a class of random, dense, n*n matrices for which LU factorization with any form
of pivoting produces a growth factor typically of size at least n/(4\logn) for large n. The …

Solving block low-rank linear systems by LU factorization is numerically stable

NJ Higham, T Mary - IMA Journal of Numerical Analysis, 2022 - academic.oup.com
Block low-rank (BLR) matrices possess a blockwise low-rank property that can be exploited
to reduce the complexity of numerical linear algebra algorithms. The impact of these low …

Randomized approximate class-specific kernel spectral regression analysis for large-scale face verification

K Li, G Wu - Machine Learning, 2022 - Springer
Kernel methods are known to be effective to analyse complex objects by implicitly
embedding them into some feature space. The approximate class-specific kernel spectral …

Iterative preconditioned methods in Krylov spaces: trends of the 21st century

VP Il'in - Computational Mathematics and Mathematical Physics, 2021 - Springer
A analytic review of major problems and new mathematical and technological discoveries in
methods for solving SLAEs is given. This stage of mathematical modeling is a bottleneck …

Randomized algorithms for large-scale dictionary learning

G Wu, J Yang - Neural Networks, 2024 - Elsevier
Dictionary learning is an important sparse representation algorithm which has been widely
used in machine learning and artificial intelligence. However, for massive data in the big …

A robust algebraic multilevel domain decomposition preconditioner for sparse symmetric positive definite matrices

H Al Daas, P Jolivet - SIAM Journal on Scientific Computing, 2022 - SIAM
Domain decomposition (DD) methods are widely used as preconditioner techniques. Their
effectiveness relies on the choice of a locally constructed coarse space. Thus far, this …