Neighborhood communication paradigm to increase scalability in large-scale dynamic scientific applications

  • Authors:
  • Aleksandr Ovcharenko;Daniel Ibanez;Fabien Delalondre;Onkar Sahni;Kenneth E. Jansen;Christopher D. Carothers;Mark S. Shephard

  • Affiliations:
  • Scientific Computation Research Center (SCOREC), Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA;Scientific Computation Research Center (SCOREC), Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA;Scientific Computation Research Center (SCOREC), Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA;Scientific Computation Research Center (SCOREC), Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA;Department of Aerospace Engineering Sciences, University of Colorado at Boulder, Boulder, CO 80309, USA;Department of Computer Science, Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA;Scientific Computation Research Center (SCOREC), Rensselaer Polytechnic Institute, 110 8th Street, Troy, NY 12180, USA

  • Venue:
  • Parallel Computing
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper introduces a general-purpose communication package built on top of MPI which is aimed at improving inter-processor communications independently of the supercomputer architecture being considered. The package is developed to support parallel applications that rely on computation characterized by large number of messages of various sizes, often small, that are focused within processor neighborhoods. In some cases, such as solvers having static mesh partitions, the number and size of messages are known a priori. However, in other cases such as mesh adaptation, the messages evolve and vary in number and size and include the dynamic movement of partition objects. The current package provides a utility for dynamic applications based on two key attributes that are: (i) explicit consideration of the neighborhood communication pattern to avoid many-to-many calls and also to reduce the number of collective calls to a minimum, and (ii) use of non-blocking MPI functions along with message packing to manage message flow control and reduce the number and time of communication calls. The test application demonstrated is parallel unstructured mesh adaptation. Results on IBM Blue Gene/P and Cray XE6 computers show that the use of neighborhood-based communication control leads to scalable results when executing generally imbalanced mesh adaptation runs.