Mixed precision algorithms in numerical linear algebra
Today's floating-point arithmetic landscape is broader than ever. While scientific computing
has traditionally used single precision and double precision floating-point arithmetics, half …
has traditionally used single precision and double precision floating-point arithmetics, half …
Simulating low precision floating-point arithmetic
NJ Higham, S Pranesh - SIAM Journal on Scientific Computing, 2019 - SIAM
The half-precision (fp16) floating-point format, defined in the 2008 revision of the IEEE
standard for floating-point arithmetic, and a more recently proposed half-precision format …
standard for floating-point arithmetic, and a more recently proposed half-precision format …
A class of fast and accurate summation algorithms
The need to sum floating-point numbers is ubiquitous in scientific computing. Standard
recursive summation of n summands, often implemented in a blocked form, has a backward …
recursive summation of n summands, often implemented in a blocked form, has a backward …
Mixed precision iterative refinement with sparse approximate inverse preconditioning
With the commercial availability of mixed precision hardware, mixed precision GMRES-
based iterative refinement schemes have emerged as popular approaches for solving …
based iterative refinement schemes have emerged as popular approaches for solving …
Random matrices generating large growth in LU factorization with pivoting
We identify a class of random, dense, n*n matrices for which LU factorization with any form
of pivoting produces a growth factor typically of size at least n/(4\logn) for large n. The …
of pivoting produces a growth factor typically of size at least n/(4\logn) for large n. The …
Solving block low-rank linear systems by LU factorization is numerically stable
Block low-rank (BLR) matrices possess a blockwise low-rank property that can be exploited
to reduce the complexity of numerical linear algebra algorithms. The impact of these low …
to reduce the complexity of numerical linear algebra algorithms. The impact of these low …
Randomized approximate class-specific kernel spectral regression analysis for large-scale face verification
K Li, G Wu - Machine Learning, 2022 - Springer
Kernel methods are known to be effective to analyse complex objects by implicitly
embedding them into some feature space. The approximate class-specific kernel spectral …
embedding them into some feature space. The approximate class-specific kernel spectral …
Iterative preconditioned methods in Krylov spaces: trends of the 21st century
VP Il'in - Computational Mathematics and Mathematical Physics, 2021 - Springer
A analytic review of major problems and new mathematical and technological discoveries in
methods for solving SLAEs is given. This stage of mathematical modeling is a bottleneck …
methods for solving SLAEs is given. This stage of mathematical modeling is a bottleneck …
Randomized algorithms for large-scale dictionary learning
G Wu, J Yang - Neural Networks, 2024 - Elsevier
Dictionary learning is an important sparse representation algorithm which has been widely
used in machine learning and artificial intelligence. However, for massive data in the big …
used in machine learning and artificial intelligence. However, for massive data in the big …
A robust algebraic multilevel domain decomposition preconditioner for sparse symmetric positive definite matrices
H Al Daas, P Jolivet - SIAM Journal on Scientific Computing, 2022 - SIAM
Domain decomposition (DD) methods are widely used as preconditioner techniques. Their
effectiveness relies on the choice of a locally constructed coarse space. Thus far, this …
effectiveness relies on the choice of a locally constructed coarse space. Thus far, this …