Tensor networks for dimensionality reduction and large-scale optimization: Part 1 low-rank tensor decompositions

A Cichocki, N Lee, I Oseledets, AH Phan… - … and Trends® in …, 2016 - nowpublishers.com
Modern applications in engineering and data science are increasingly based on
multidimensional data of exceedingly high volume, variety, and structural richness …

Low-rank tensor networks for dimensionality reduction and large-scale optimization problems: Perspectives and challenges part 1

A Cichocki, N Lee, IV Oseledets, AH Phan… - arxiv preprint arxiv …, 2016 - arxiv.org
Machine learning and data mining algorithms are becoming increasingly important in
analyzing large volume, multi-relational and multi--modal datasets, which are often …

Strassen's algorithm reloaded

J Huang, TM Smith, GM Henry… - SC'16: Proceedings of …, 2016 - ieeexplore.ieee.org
We dispel with “street wisdom” regarding the practical implementation of Strassen's
algorithm for matrix-matrix multiplication (DGEMM). Conventional wisdom: it is only practical …

Matrix multiplication, a little faster

E Karstadt, O Schwartz - Journal of the ACM (JACM), 2020 - dl.acm.org
Strassen's algorithm (1969) was the first sub-cubic matrix multiplication algorithm. Winograd
(1971) improved the leading coefficient of its complexity from 6 to 7. There have been many …

Strassen's algorithm reloaded on GPUs

J Huang, CD Yu, RA Geijn - ACM Transactions on Mathematical …, 2020 - dl.acm.org
Conventional Graphics Processing Unit (GPU) implementations of Strassen's algorithm
(Strassen) rely on the existing high-performance matrix multiplication (gemm), trading space …

Investigating Bayesian Optimization for rail network optimization

B Hickish, DI Fletcher, RF Harrison - International Journal of Rail …, 2020 - Taylor & Francis
Optimizing the operation of rail networks using simulations is an on-going task where
heuristic methods such as Genetic Algorithms have been applied. However, these …

Pebbling game and alternative basis for high performance matrix multiplication

O Schwartz, N Vaknin - SIAM Journal on Scientific Computing, 2023 - SIAM
Matrix multiplication is one of the most extensively used kernels in scientific computing.
Although subcubic algorithms exist, most high performance implementations are based on …

Error analysis and improving the accuracy of Winograd convolution for deep neural networks

B Barabasz, A Anderson, KM Soodhalter… - ACM Transactions on …, 2020 - dl.acm.org
Popular deep neural networks (DNNs) spend the majority of their execution time computing
convolutions. The Winograd family of algorithms can greatly reduce the number of arithmetic …

Generating families of practical fast matrix multiplication algorithms

J Huang, L Rice, DA Matthews… - 2017 IEEE …, 2017 - ieeexplore.ieee.org
Matrix multiplication (GEMM) is a core operation to numerous scientific applications.
Traditional implementations of Strassen-like fast matrix multiplication (FMM) algorithms often …

Design and Performance Analysis of 6T SRAM cell in 22nm CMOS and FinFET technology Nodes

SR Sanjana, R Banu, P Shubham - … Conference on Recent …, 2017 - ieeexplore.ieee.org
In modern day VLSI system design memories are the vital blocks and they need to be
thoroughly investigated with respect to area, power and performance before their fabrication …