Sparsetir: Composable abstractions for sparse compilation in deep learning

Z Ye, R Lai, J Shao, T Chen, L Ceze - Proceedings of the 28th ACM …, 2023 - dl.acm.org
Sparse tensors are rapidly becoming critical components of modern deep learning
workloads. However, develo** high-performance sparse operators can be difficult and …

Sparse supernodal solver using block low-rank compression: Design, performance and analysis

G Pichon, E Darve, M Faverge, P Ramet… - Journal of computational …, 2018 - Elsevier
This paper presents two approaches using a Block Low-Rank (BLR) compression technique
to reduce the memory footprint and/or the time-to-solution of the sparse supernodal solver …

A comparison of two effective methods for reordering columns within supernodes

MO Karsavuran, EG Ng, BW Peyton - arxiv preprint arxiv:2501.08395, 2025 - arxiv.org
In some recent papers, researchers have found two very good methods for reordering
columns within supernodes in sparse Cholesky factors; these reorderings can be very useful …

Improving predication efficiency through compaction/restoration of simd instructions

A Barredo, JM Cebrian, M Moretó… - … symposium on high …, 2020 - ieeexplore.ieee.org
Vector processors offer a wide range of unexplored opportunities to improve performance
and energy efficiency. However, despite its potential, vector code generation and execution …

Sparsity analysis and optimization for state-space-based simulation of power electronic systems

Z Yu, B Shi, S Jia, H Xu, Y **ao… - CSEE Journal of Power …, 2024 - ieeexplore.ieee.org
Computer-aided analysis tool is playing an important role in the design of power electronics
converters. However, with the increase of system scale and complexity, the existing …

Sparse supernodal solver using block low-rank compression

G Pichon, E Darve, M Faverge… - 2017 IEEE …, 2017 - ieeexplore.ieee.org
This paper presents two approaches using a Block Low-Rank (BLR) compression technique
to reduce the memory footprint and/or the time-to-solution of the sparse supernodal solver …

Some new techniques to use in serial sparse Cholesky factorization algorithms

MO Karsavuran, EG Ng, BW Peyton… - arxiv preprint arxiv …, 2024 - arxiv.org
We present a new variant of serial right-looking supernodal sparse Cholesky factorization
(RL). Our comparison of RL with the multifrontal method confirms that RL is simpler, slightly …

Optimizing partitioned CSR-based SpGEMM on the Sunway TaihuLight

Y Chen, G **ao, W Yang - Neural Computing and Applications, 2020 - Springer
General sparse matrix-sparse matrix (SpGEMM) multiplication is one of the basic kernels in
a great many applications. Several works focus on various optimizations for SpGEMM. To …

Improved sparsity techniques for solving network equations in transient stability simulations

T **ao, J Wang, Y Gao, D Gan - IEEE Transactions on Power …, 2018 - ieeexplore.ieee.org
When solving network algebraic equations during power system transient stability
simulations, the nonzero elements in the independent vector and the elements needed in …

Blocking Sparse Matrices to Leverage Dense-Specific Multiplication

PS Labini, M Bernaschi, W Nutt… - 2022 IEEE/ACM …, 2022 - ieeexplore.ieee.org
Research to accelerate matrix multiplication, pushed by the growing computational
demands of deep learning, has sprouted many efficient architectural solutions, such as …