Published in

Elsevier, Parallel Computing: Systems & Applications, 5-6(40), p. 47-58

DOI: 10.1016/j.parco.2014.03.012

Links

Tools

Export citation

Search in Google Scholar

Sparse Matrix Multiplication: The Distributed Block-Compressed Sparse Row Library

Journal article published in 2014 by Urban Borštnik, Joost VandeVondele ORCID, Valéry Weber, Jrg Hutter
This paper is available in a repository.
This paper is available in a repository.

Full text: Download

Green circle
Preprint: archiving allowed
Red circle
Postprint: archiving forbidden
Red circle
Published version: archiving forbidden
Data provided by SHERPA/RoMEO

Abstract

Efficient parallel multiplication of sparse matrices is key to enabling many large-scale calculations. This article presents the DBCSR (Distributed Block Compressed Sparse Row) library for scalable sparse matrix-matrix multiplication and its use in the CP2K program for linear-scaling quantum-chemical calculations. The library combines several approaches to implement sparse matrix multiplication in a way that performs well and is demonstrably scalable. Parallel communication has well-defined limits. Data volume decreases with O(1/P) with increasing process counts P and every process communicates with at most O(P) others. Local sparse matrix multiplication is handled efficiently using a combination of techniques: blocking elements together in an application-relevant way, an autotuning library for small matrix multiplications, cache-oblivious recursive multiplication, and multithreading. Additionally, on-the-fly filtering not only increases sparsity but also avoids performing calculations that fall below the filtering threshold. We demonstrate and analyze the performance of the DBCSR library and its various scaling behaviors.