HPCVIEW: A Tool for Top-down Analysis of Node Performance
The Journal of Supercomputing
Toward Scalable Performance Visualization with Jumpshot
International Journal of High Performance Computing Applications
PerfExplorer: A Performance Data Mining Framework For Large-Scale Parallel Computing
SC '05 Proceedings of the 2005 ACM/IEEE conference on Supercomputing
A Flexible and Dynamic Infrastructure for MPI Tool Interoperability
ICPP '06 Proceedings of the 2006 International Conference on Parallel Processing
Parallel clustering algorithms for structured AMR
Journal of Parallel and Distributed Computing
PNMPI tools: a whole lot greater than the sum of their parts
Proceedings of the 2007 ACM/IEEE conference on Supercomputing
Open | SpeedShop: An open source infrastructure for parallel performance analysis
Scientific Programming - Large-Scale Programming Tools and Environments
Diagnosing performance bottlenecks in emerging petascale applications
Proceedings of the Conference on High Performance Computing Networking, Storage and Analysis
Scalable fine-grained call path tracing
Proceedings of the international conference on Supercomputing
Interpreting Performance Data across Intuitive Domains
ICPP '11 Proceedings of the 2011 International Conference on Parallel Processing
Hi-index | 0.00 |
Performance analysis of parallel scientific codes is becoming increasingly difficult due to the rapidly growing complexity of applications and architectures. Existing tools fall short in providing intuitive views that facilitate the process of performance debugging and tuning. In this paper, we extend recent ideas of projecting and visualizing performance data for faster, more intuitive analysis of applications. We collect detailed per-level and per-phase measurements for a dynamically load-balanced, structured AMR library and project per-core data collected in the hardware domain on to the application's communication topology. We show how our projections and visualizations lead to a rapid diagnosis of and mitigation strategy for a previously elusive scaling bottleneck in the library that is hard to detect using conventional tools. Our new insights have resulted in a 22% performance improvement for a 65,536-core run of the AMR library on an IBM Blue Gene/P system.