Accelerating sparse matrix-vector multiplication on GPUs using bit-representation-optimized schemes

  • Authors:
  • Wai Teng Tang;Wen Jun Tan;Rajarshi Ray;Yi Wen Wong;Weiguang Chen;Shyh-hao Kuo;Rick Siow Mong Goh;Stephen John Turner;Weng-Fai Wong

  • Affiliations:
  • Nanyang Technological University, Singapore;Nanyang Technological University, Singapore;National University of Singapore, Singapore;National University of Singapore, Singapore;National University of Singapore, Singapore;Institute of High Performance Computing, Singapore;Institute of High Performance Computing, Singapore;Nanyang Technological University, Singapore;National University of Singapore, Singapore

  • Venue:
  • SC '13 Proceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

The sparse matrix-vector (SpMV) multiplication routine is an important building block used in many iterative algorithms for solving scientific and engineering problems. One of the main challenges of SpMV is its memory-boundedness. Although compression has been proposed previously to improve SpMV performance on CPUs, its use has not been demonstrated on the GPU because of the serial nature of many compression and decompression schemes. In this paper, we introduce a family of bit-representation-optimized (BRO) compression schemes for representing sparse matrices on GPUs. The proposed schemes, BRO-ELL, BRO-COO, and BRO-HYB, perform compression on index data and help to speed up SpMV on GPUs through reduction of memory traffic. Furthermore, we formulate a BRO-aware matrix reordering scheme as a data clustering problem and use it to increase compression ratios. With the proposed schemes, experiments show that average speedups of 1.5x compared to ELLPACK and HYB can be achieved for SpMV on GPUs.