Compiling Fortran D for MIMD distributed-memory machines
Communications of the ACM
The high performance Fortran handbook
The high performance Fortran handbook
Latency and bandwidth considerations in parallel robotics image processing
Proceedings of the 1993 ACM/IEEE conference on Supercomputing
Generating communication for array statements: design, implementation, and evaluation
Journal of Parallel and Distributed Computing - Special issue on data parallel algorithms and programming
Using MPI: portable parallel programming with the message-passing interface
Using MPI: portable parallel programming with the message-passing interface
Optimal mapping of sequences of data parallel tasks
PPOPP '95 Proceedings of the fifth ACM SIGPLAN symposium on Principles and practice of parallel programming
Decoupling synchronization and data transfer in message passing systems of parallel computers
ICS '95 Proceedings of the 9th international conference on Supercomputing
Optimal latency-throughput tradeoffs for data parallel pipelines
Proceedings of the eighth annual ACM symposium on Parallel algorithms and architectures
Double standards: bringing task parallelism to HPF via the message passing interface
Supercomputing '96 Proceedings of the 1996 ACM/IEEE conference on Supercomputing
Communication and memory requirements as the basis for mapping task and data parallel programs
Proceedings of the 1994 ACM/IEEE conference on Supercomputing
Task Parallelism in a High Performance Fortran Framework
IEEE Parallel & Distributed Technology: Systems & Technology
IEEE Transactions on Pattern Analysis and Machine Intelligence
IEEE Transactions on Parallel and Distributed Systems
Do&Merge: Integrating Parallel Loops and Reductions
Proceedings of the 6th International Workshop on Languages and Compilers for Parallel Computing
A SOFTWARE ARCHITECTURE FOR MULTIDISCIPLINARY APPICATIONS: INTEGRATING TASK AND DATA PARALLELISM
A SOFTWARE ARCHITECTURE FOR MULTIDISCIPLINARY APPICATIONS: INTEGRATING TASK AND DATA PARALLELISM
AN OVERVIEW OF THE OPUS LANGUAGE AND RUNTIME SYSTEM
AN OVERVIEW OF THE OPUS LANGUAGE AND RUNTIME SYSTEM
Parallel breadth-first BDD construction
PPOPP '97 Proceedings of the sixth ACM SIGPLAN symposium on Principles and practice of parallel programming
A task- and data-parallel programming language based on shared objects
ACM Transactions on Programming Languages and Systems (TOPLAS)
A coordination language for mixed task and and data parallel programs
Proceedings of the 1999 ACM symposium on Applied computing
Building programs in the network of tasks model
SAC '00 Proceedings of the 2000 ACM symposium on Applied computing - Volume 1
ORT: a communication library for orthogonal processor groups
Proceedings of the 2001 ACM/IEEE conference on Supercomputing
A border-based coordination language for integrating task and data parallelism
Journal of Parallel and Distributed Computing
A resource query interface for network-aware applications
Cluster Computing
A data and task parallel image processing environment
Parallel Computing - Parallel computing in image and video processing
Approaches for Integrating Task and Data Parallelism
IEEE Concurrency
Integrating Task and Data Parallelism by Means of Coordination Patterns
HIPS '01 Proceedings of the 6th International Workshop on High-Level Parallel Programming Models and Supportive Environments
Network-Aware Parallel Computing with Remos
LCPC '98 Proceedings of the 11th International Workshop on Languages and Compilers for Parallel Computing
Exploiting Advanced Task Parallelism in High Performance Fortran via a Task Library
Euro-Par '99 Proceedings of the 5th International Euro-Par Conference on Parallel Processing
Library support for hierarchical multi-processor tasks
Proceedings of the 2002 ACM/IEEE conference on Supercomputing
Task and data parallelism in P3L
Patterns and skeletons for parallel and distributed computing
Adaptive Distributed Applications on Heterogeneous Networks
HCW '99 Proceedings of the Eighth Heterogeneous Computing Workshop
Domain interaction patterns to coordinate HPF tasks
Parallel Computing
Airshed Pollution Modeling: A Case Study in Application Development in an HPF Environment
IPPS '98 Proceedings of the 12th. International Parallel Processing Symposium on International Parallel Processing Symposium
Exploiting processor groups to extend scalability of the GA shared memory programming model
Proceedings of the 2nd conference on Computing frontiers
Tlib-a library to support programming with hierarchical multi-processor tasks
Journal of Parallel and Distributed Computing
Experiences with optimizing two stream-based applications for cluster execution
Journal of Parallel and Distributed Computing
OpenGR: A directive-based grid programming environment
Parallel Computing - OpenMp
Irregular computations in Fortran - expression and implementation strategies
Scientific Programming
Communicating Multiprocessor-Tasks
Languages and Compilers for Parallel Computing
OoOJava: an out-of-order approach to parallel programming
HotPar'10 Proceedings of the 2nd USENIX conference on Hot topics in parallelism
OoOJava: software out-of-order execution
Proceedings of the 16th ACM symposium on Principles and practice of parallel programming
Finding, expressing and managing parallelism in programs executed on clusters of workstations
Computer Communications
Programming support and scheduling for communicating parallel tasks
Journal of Parallel and Distributed Computing
Combined scheduling and mapping for scalable computing with parallel tasks
Scientific Programming - Biological Knowledge Discovery and Data Mining
Hi-index | 0.00 |
High Performance Fortran (HPF) has emerged as a standard language fordata parallel computing. However, a wide variety of scientific applications are best programmed by a combination of task and data parallelism. Therefore, a good model of task parallelism is important for continued success of HPF for parallel programming. This paper presents a task parallelism model that is simple, elegant, and relatively easy to implement in an HPF environment. Task parallelism is exploited by mechanisms for dividing processors into subgroups and mapping computations and data onto processor subgroups. This model of task parallelism has been implemented in the Fx compiler at Carnegie Mellon University. The paper addresses the main issues in compiling integrated task and data parallel programs and reports on the use of this model for programming various flat and nested task structures. Performance results are presented for a set of programs spanning signal processing, image processing, computer vision and environment modeling. A variant of this task model is a new approved extension of HPF and this paper offers insight into the power of expression and ease of implementation of this extension.