OpenMP parallelism for fluid and fluid-particulate systems

  • Authors:
  • Amit Amritkar;Danesh Tafti;Rui Liu;Rick Kufrin;Barbara Chapman

  • Affiliations:
  • Virginia Tech, 114 I Randolph Hall, Department of Mechanical Engineering, Blacksburg, VA 24061, United States;Virginia Tech, 114 I Randolph Hall, Department of Mechanical Engineering, Blacksburg, VA 24061, United States;National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign, 1205 W. Clark St., Room 1008, Urbana, IL 61801, United States;National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign, 1205 W. Clark St., Room 1008, Urbana, IL 61801, United States;Department of Computer Science, University of Houston, Houston, TX 77204-3010, United States

  • Venue:
  • Parallel Computing
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

In order to exploit the flexibility of OpenMP in parallelizing large scale multi-physics applications where different modes of parallelism are needed for efficient computation, it is first necessary to be able to scale OpenMP codes as well as MPI on large core counts. In this research we have implemented fine grained OpenMP parallelism for a large CFD code GenIDLEST and investigated the performance from 1 to 256 cores using a variety of performance optimization and measurement tools. It is shown through weak and strong scaling studies that OpenMP performance can be made to match that of MPI on the SGI Altix systems for up to 256 cores. Data placement and locality were established to be key components in obtaining good scalability with OpenMP. It is also shown that a hybrid implementation on a dual core system gives the same performance as standalone MPI or OpenMP. Finally, it is shown that in irregular multi-physics applications which do not adhere solely to the SPMD (Single Process, Multiple Data) mode of computation, as encountered in tightly coupled fluid-particulate systems, the flexibility of OpenMP can have a big performance advantage over MPI.