Astrophysical N-body simulations using hierarchical tree data structures
Proceedings of the 1992 ACM/IEEE conference on Supercomputing
Using MPI: portable parallel programming with the message-passing interface
Using MPI: portable parallel programming with the message-passing interface
Highly portable and efficient implementations of parallel adaptive N-body methods
SC '97 Proceedings of the 1997 ACM/IEEE conference on Supercomputing
Designing and Building Parallel Programs: Concepts and Tools for Parallel Software Engineering
Designing and Building Parallel Programs: Concepts and Tools for Parallel Software Engineering
Hiding Communication Latency in Reconfigurable Message-Passing Environments
IPPS '99/SPDP '99 Proceedings of the 13th International Symposium on Parallel Processing and the 10th Symposium on Parallel and Distributed Processing
Experimental Evaluation of QSM, a Simple Shared-Memory Model
IPPS '99/SPDP '99 Proceedings of the 13th International Symposium on Parallel Processing and the 10th Symposium on Parallel and Distributed Processing
The Semantics of Blocking and Nonblocking Send and Receive Primitives
Proceedings of the 8th International Symposium on Parallel Processing
Interprocedural Communication Optimizations for Message Passing Architectures
FRONTIERS '99 Proceedings of the The 7th Symposium on the Frontiers of Massively Parallel Computation
Optimizing Message Aggregation for Parallel Simulation on High Performance Clusters
MASCOTS '99 Proceedings of the 7th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems
GPS: a parallel programming tool based on process groups
SCCC '97 Proceedings of the 17th International Conference of the Chilean Computer Science Society
Parallel Implementation of 3D FMA using MPI
MPIDC '96 Proceedings of the Second MPI Developers Conference
Identifying the Capability of Overlapping Computation with Communication
PACT '96 Proceedings of the 1996 Conference on Parallel Architectures and Compilation Techniques
Hi-index | 0.00 |
Overlap of computations and communications is an effective mechanism to improve the performance of parallel/distributed applications significantly. This overlap can be achieved efficiently by using data partitioning and properly scheduling the data transfer. Various asynchronous communication primitives, that are provided by most message passing tools (e.g. PVM, MPI), can be used to implement the required overlapping mechanism. In this paper, we present a design model, the Distributed Software Design Model (DSDM) and show how it can be applied to optimize parallel/distributed applications. We show through several examples, the Master-Slave Merge Sorting Application and the astrophysical N-Body Problem, how the DSDM can be used to develop efficient and optimized implementations of parallel and distributed algorithms.