Fast multiplication of random dense matrices with fixed sparse matrices
October 24, 2023 ยท Declared Dead ยท ๐ arXiv.org
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Tianyu Liang, Riley Murray, Aydฤฑn Buluรง, James Demmel
arXiv ID
2310.15419
Category
cs.CE: Computational Engineering
Cross-listed
cs.DC,
cs.DS
Citations
0
Venue
arXiv.org
Last Checked
2 months ago
Abstract
This work focuses on accelerating the multiplication of a dense random matrix with a (fixed) sparse matrix, which is frequently used in sketching algorithms. We develop a novel scheme that takes advantage of blocking and recomputation (on-the-fly random number generation) to accelerate this operation. The techniques we propose decrease memory movement, thereby increasing the algorithm's parallel scalability in shared memory architectures. On the Intel Frontera architecture, our algorithm can achieve 2x speedups over libraries such as Eigen and Intel MKL on some examples. In addition, with 32 threads, we can obtain a parallel efficiency of up to approximately 45%. We also present a theoretical analysis for the memory movement lower bound of our algorithm, showing that under mild assumptions, it's possible to beat the data movement lower bound of general matrix-matrix multiply (GEMM) by a factor of $\sqrt M$, where $M$ is the cache size. Finally, we incorporate our sketching algorithm into a randomized least squares solver. For extremely over-determined sparse input matrices, we show that our results are competitive with SuiteSparse; in some cases, we obtain a speedup of 10x over SuiteSparse.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computational Engineering
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
A Probabilistic Graphical Model Foundation for Enabling Predictive Digital Twins at Scale
R.I.P.
๐ป
Ghosted
Temporal Attention augmented Bilinear Network for Financial Time-Series Data Analysis
R.I.P.
๐ป
Ghosted
Linked Component Analysis from Matrices to High Order Tensors: Applications to Biomedical Data
R.I.P.
๐ป
Ghosted
Deep Dynamical Modeling and Control of Unsteady Fluid Flows
R.I.P.
๐ป
Ghosted
Design and Optimization of Conforming Lattice Structures
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted