Using MPI: portable parallel programming with the message-passing interface
Using MPI: portable parallel programming with the message-passing interface
Multigrid strategies for viscous flow solvers on anisotropic unstructured meshes
Journal of Computational Physics
Parallel performance investigations of an unstructured mesh Navier-Stokes solver
Parallel performance investigations of an unstructured mesh Navier-Stokes solver
Performance of a new CFD flow solver using a hybrid programming paradigm
Journal of Parallel and Distributed Computing
An Application-Based Performance Characterization of the Columbia Supercluster
SC '05 Proceedings of the 2005 ACM/IEEE conference on Supercomputing
Digital Flight: The Last CFD Aeronautical Grand Challenge
Journal of Scientific Computing
Robust scalability analysis and SPM case studies
The Journal of Supercomputing
A massively parallel fractional step solver for incompressible flows
Journal of Computational Physics
Hi-index | 0.00 |
This paper focuses on the parallel performance of two high-performance aerodynamic simulation packages on the newly installed NASA Columbia supercomputer. These packages include both a high-fidelity, unstructured, Reynolds-averaged Navier-Stokes solver, and a fully-automated inviscid flow package for cut-cell Cartesian grids. The complementary combination of these two simulation codes enables high-fidelity characterization of aerospace vehicle design performance over the entire flight envelope through extensive parametric analysis and detailed simulation of critical regions of the flight envelope. Both packages are industrial-level codes designed for complex geometry and incorporate customized multigrid solution algorithms. The performance of these codes on Columbia is examined using both MPI and OpenMP and using both the NUMAlink and InfiniBand interconnect fabrics. Numerical results demonstrate good scalability on up to 2016 CPUs using the NUMAlink4 interconnect, with measured computational rates in the vicinity of 3 TFLOP/s, while InfiniBand showed some performance degradation at high CPU counts, particularly with multigrid. Nonetheless, the results are encouraging enough to indicate that larger test cases using combined MPI/OpenMP communication should scale well on even more processors.