A New Parallel Matrix Multiplication Algorithm on Distributed-Memory Concurrent Computers

  • Authors:
  • Jaeyoung Choi

  • Affiliations:
  • -

  • Venue:
  • HPC-ASIA '97 Proceedings of the High-Performance Computing on the Information Superhighway, HPC-Asia '97
  • Year:
  • 1997

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a new parallel matrix multiplication algorithm on distributed memory concurrent computers, which is fast and scalable, and whose performance is independent of data distribution on processors, and call it DIMMA (Distribution-Independent Matrix Multiplication Algorithm). The algorithm is based on two new ideas; it uses a modified pipelined communication scheme to overlap computation and communication effectively, and exploits the LCM block concept to obtain the maximum performance of the sequential BLAS routine in each processor even when the block size is very small as well as very large. The algorithm is implemented and compared with SUMMA on the Intel Paragon computer.