Effects of synchronization barriers on multiprocessor performance
Parallel Computing
Efficient distributed event-driven simulations of multiple-loop networks
Communications of the ACM
The cost of conservative synchronization in parallel discrete event simulations
Journal of the ACM (JACM)
Global Virtual Time and distributed synchronization
PADS '95 Proceedings of the ninth workshop on Parallel and distributed simulation
Optimistic simulation of parallel architectures using program executables
PADS '96 Proceedings of the tenth workshop on Parallel and distributed simulation
Analysis of bounded time warp and comparison with YAWNS
ACM Transactions on Modeling and Computer Simulation (TOMACS)
The impact of adding aggressiveness to a non-aggressive windowing protocol
WSC '93 Proceedings of the 25th conference on Winter simulation
Hi-index | 0.00 |
A number of optimistic synchronization schemes for parallel simulation rely upon a global synchronization. The problem is to determine when every processor has completed all its work, and there are no messages in transit in the system that will cause more work. Most previous solutions to the problem have used distributed termination algorithms, which are inherently serial; other parallel mechanisms may be inefficient. In this paper we describe an efficient parallel algorithm derived from a common “barrier” synchronization algorithm used in parallel processing. The algorithm's principal attraction is speed, and generality—it is designed to be used in contexts more general than parallel discrete-event simulation. To establish our claim to speed, we compare our algorithm's performance with the standard barrier algorithm, and find that its additional costs are not excessive. Our experiments are conducted using up to 256 processors on the Intel Touchstone Delta.