A bridging model for parallel computation
Communications of the ACM
LogP: towards a realistic model of parallel computation
PPOPP '93 Proceedings of the fourth ACM SIGPLAN symposium on Principles and practice of parallel programming
Synchronization and communication in the T3E multiprocessor
Proceedings of the seventh international conference on Architectural support for programming languages and operating systems
A quantitative comparison of parallel computation models
ACM Transactions on Computer Systems (TOCS)
Efficient Parallel Graph Algorithms For Coarse Grained Multicomputers and BSP
ICALP '97 Proceedings of the 24th International Colloquium on Automata, Languages and Programming
Parallelism in random access machines
STOC '78 Proceedings of the tenth annual ACM symposium on Theory of computing
MPI: A Message-Passing Interface Standard
MPI: A Message-Passing Interface Standard
A Language for the Complexity Analysis of Parallel Programs
PPAM '01 Proceedings of the th International Conference on Parallel Processing and Applied Mathematics-Revised Papers
Performance Prediction of Oblivious BSP Programs
Euro-Par '01 Proceedings of the 7th International Euro-Par Conference Manchester on Parallel Processing
Predicting the performance of parallel programs
Parallel Computing
A tool for performance modeling of parallel programs
Scientific Programming
Model oriented profiling of parallel programs
EUROMICRO-PDP'02 Proceedings of the 10th Euromicro conference on Parallel, distributed and network-based processing
Hi-index | 0.00 |
The BSP cost model provides a general framework to design efficient and portable data-parallel algorithms. Execution costs of BSP programs are predicted combining a limited number of program and machine dependent parameters. BSP programs can be written using several programming tools. In this work we explore the predictability of Bulk Synchronous programs implemented with the Message Passing Interface. Two classic computational geometry problems: the Convex Hull (CH) and the Lower Envelope (LE) are considered as cases of study. Efficient BSP algorithms have been implemented using MPI and executed on three different parallel architectures: a Fujitsu AP1000 (distributed memory), a CRAY T3E (distributed shared memory) and a cluster of PCs (Backus). The paper compares the degree of predictability on these architectures analysing the main sources of error.