Reducing Connection Memory Requirements of MPI for InfiniBand Clusters: A Message Coalescing Approach

  • Authors:
  • Matthew J. Koop;Terry Jones;Dhabaleswar K. Panda

  • Affiliations:
  • Ohio State University;Lawrence Livermore National Laboratory;Ohio State University

  • Venue:
  • CCGRID '07 Proceedings of the Seventh IEEE International Symposium on Cluster Computing and the Grid
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Clusters in the area of high-performance computing have been growing in size at a considerable rate. In these clusters, the dominate programming model is the Message Passing Interface (MPI), so the MPI library has a key role in resource usage and performance. To obtain maximal performance, many clusters deploy a high-speed interconnect between compute nodes. One such interconnect, InfiniBand, has been gaining in popularity due to its various features including Remote Data Memory Access (RDMA), and high-performance. As a result, it is being deployed in a significant number of clusters and has been chosen as the standard interconnect for capacity clusters within the DOE Tri-Labs. As these clusters grow in size, care must be taken to ensure the resource usage does not increase too significantly with scale. In particular, the MPI library resource usage should not grow at a rate which will exhaust the node memory or starve user applications.