The integration of application and system based metrics in a parallel program performance tool
PPOPP '91 Proceedings of the third ACM SIGPLAN symposium on Principles and practice of parallel programming
The visualization of parallel systems: an overview
Journal of Parallel and Distributed Computing - Special issue on tools and methods for visualization of parallel systems and computations
Journal of Parallel and Distributed Computing - Special issue on tools and methods for visualization of parallel systems and computations
Analytical performance prediction on multicomputers
Proceedings of the 1993 ACM/IEEE conference on Supercomputing
Normalized performance indices for message passing parallel programs
ICS '94 Proceedings of the 8th international conference on Supercomputing
SIGMETRICS '94 Proceedings of the 1994 ACM SIGMETRICS conference on Measurement and modeling of computer systems
Software—Practice & Experience
Automated performance prediction of message-passing parallel programs
Supercomputing '95 Proceedings of the 1995 ACM/IEEE conference on Supercomputing
Analysis and optimization of software pipeline performance on MIMD parallel computers
Journal of Parallel and Distributed Computing
Parallel performance prediction using lost cycles analysis
Proceedings of the 1994 ACM/IEEE conference on Supercomputing
Visualizing the Performance of Parallel Programs
IEEE Software
Automated Modeling of Message-Passing Programs
MASCOTS '94 Proceedings of the Second International Workshop on Modeling, Analysis, and Simulation On Computer and Telecommunication Systems
Scalability Analysis Tools for SPMD Message-Passing Parallel Programs
MASCOTS '94 Proceedings of the Second International Workshop on Modeling, Analysis, and Simulation On Computer and Telecommunication Systems
Automated performance prediction of message-passing parallel programs
Supercomputing '95 Proceedings of the 1995 ACM/IEEE conference on Supercomputing
Trace-Based Load Characterization for Generating Performance Software Models
IEEE Transactions on Software Engineering
Evaluating the Scalability of Distributed Systems
IEEE Transactions on Parallel and Distributed Systems
Comparison of two code scalability tests
Information Processing Letters
Performance Prediction: A Case Study Using a Scalable Shared-Virtual-Memory Machine
IEEE Parallel & Distributed Technology: Systems & Technology
Performance Prediction for Complex Parallel Applications
IPPS '97 Proceedings of the 11th International Symposium on Parallel Processing
Event-Based Study of the Effect of Execution Environments on Parallel Program Performance
MASCOTS '96 Proceedings of the 4th International Workshop on Modeling, Analysis, and Simulation of Computer and Telecommunications Systems
Algorithm-system scalability of heterogeneous computing
Journal of Parallel and Distributed Computing
Hi-index | 0.00 |
Most performance tools for parallel programs are geared toward small-scale executions of small programs. Scalability analysis tools can help programmers predict the efficiency of scaled-up versions of such programs. Scalability analysis characterizes a program's asymptotic speedup as a function of problem size and the number of processors. It often requires modeling the program, which is a complex and tedious task. Scalability analysis toolkits that automate modeling can ease the programmer's burden.The authors introduce a methodology and a toolkit for automated scalability analysis of message-passing programs. The toolkit, called MK, generates an augmented parse tree, which is an abstraction of a program's actual parse tree. MK extracts the APT's skeleton from the original parse tree and annotates it with symbolic information about loop bounds and corresponding complexity expressions. The toolkit augments this annotated parse tree with computation phase graphs, which contain communication pattern information.MK then traverses the APT's nodes to perform two types of scalability analysis: complexity analysis or abstract interpretation. Complexity analysis, which involves algorithmic analysis of computation and communication costs, predicts a program's general performance. Abstract interpretation, which simulates a program's communication but abstracts out computation, pinpoints causes of poor performance.The authors demonstrate their approach by automatically characterizing the scalability of several scientific applications that run on Intel's iPSC/860 and Paragon supercomputers. Complexity analyses of a pipeline tridiagonal solver and a 2D atmospheric simulation predict times that are close to the actual execution times. The pipeline solver simulation has a maximum error of less than 20%; the atmospheric simulation has a mean error of 9.14%. Abstract interpretations of a transpose-based tridiagonal solver and a pipeline tridiagonal solver produce similar results. The transposed-based solver simulation's error is within 8%, and the pipeline simulation's error is approximately 15%.The authors plan to extend MK to handle more complicated analyses, and to extend the analyses to other platforms and architectures.