Communications of the ACM - Special issue on parallelism
Communications of the ACM
Process decomposition through locality of reference
PLDI '89 Proceedings of the ACM SIGPLAN 1989 Conference on Programming language design and implementation
Strand: new concepts in parallel programming
Strand: new concepts in parallel programming
Compiling Fortran D for MIMD distributed-memory machines
Communications of the ACM
The high performance Fortran handbook
The high performance Fortran handbook
A parallel programming environment supporting multiple data-parallel modules
International Journal of Parallel Programming
Using MPI: portable parallel programming with the message-passing interface
Using MPI: portable parallel programming with the message-passing interface
Digital Picture Processing
Task Parallelism in a High Performance Fortran Framework
IEEE Parallel & Distributed Technology: Systems & Technology
Automatic Extraction of Functional Parallelism from Ordinary Programs
IEEE Transactions on Parallel and Distributed Systems
Optimal Processor Assignment for a Class of Pipelined Computations
IEEE Transactions on Parallel and Distributed Systems
FRONTIERS '95 Proceedings of the Fifth Symposium on the Frontiers of Massively Parallel Computation (Frontiers'95)
AN OVERVIEW OF THE OPUS LANGUAGE AND RUNTIME SYSTEM
AN OVERVIEW OF THE OPUS LANGUAGE AND RUNTIME SYSTEM
A new model for integrated nested task and data parallel programming
PPOPP '97 Proceedings of the sixth ACM SIGPLAN symposium on Principles and practice of parallel programming
A task- and data-parallel programming language based on shared objects
ACM Transactions on Programming Languages and Systems (TOPLAS)
Thread fork/join techniques for multi-level parallelism exploitation in NUMA multiprocessors
ICS '99 Proceedings of the 13th international conference on Supercomputing
Developing and evaluating abstractions for distributed supercomputing
Cluster Computing
Approaches for Integrating Task and Data Parallelism
IEEE Concurrency
Interoperability of Data Parallel Runtime Libraries
IPPS '97 Proceedings of the 11th International Symposium on Parallel Processing
Interoperability of Data Parallel Runtime Libraries
IPPS '97 Proceedings of the 11th International Symposium on Parallel Processing
OpenMP Extensions for Thread Groups and Their Run-Time Support
LCPC '00 Proceedings of the 13th International Workshop on Languages and Compilers for Parallel Computing-Revised Papers
Exploiting Advanced Task Parallelism in High Performance Fortran via a Task Library
Euro-Par '99 Proceedings of the 5th International Euro-Par Conference on Parallel Processing
Task and data parallelism in P3L
Patterns and skeletons for parallel and distributed computing
Airshed Pollution Modeling: A Case Study in Application Development in an HPF Environment
IPPS '98 Proceedings of the 12th. International Parallel Processing Symposium on International Parallel Processing Symposium
Sourcebook of parallel computing
Tlib-a library to support programming with hierarchical multi-processor tasks
Journal of Parallel and Distributed Computing
OpenGR: A directive-based grid programming environment
Parallel Computing - OpenMp
Opus: A Coordination Language for Multidisciplinary Applications
Scientific Programming
Programming support and scheduling for communicating parallel tasks
Journal of Parallel and Distributed Computing
Hi-index | 0.00 |
High Performance Fortran (HPF) does not allow efficient expression of mixed task/data-parallel computations or the coupling of separately compiled data-parallel modules. In this paper, we show how a coordination library implementing the Message Passing Interface (MPI) can be used to represent these common parallel program structures. This library allows data-parallel tasks to exchange distributed data structures using calls to simple communication functions. We present microbenchmark results that characterize the performance of this library and that quantify the impact of optimizations that allow reuse of communication schedules in common situations. In addition, results from two-dimensional FFT, convolution, and multiblock programs demonstrate that the HPF/MPI library can provide performance superior to that of pure HPF. We conclude that this synergistic combination of two parallel programming standards represents a useful approach to task parallelism in a data-parallel framework, increasing the range of problems addressable in HPF without requiring complex compiler technology.