Algorithms for scalable synchronization on shared-memory multiprocessors
ACM Transactions on Computer Systems (TOCS)
Atomic snapshots of shared memory
Journal of the ACM (JACM)
Direct bulk-synchronous parallel algorithms
Journal of Parallel and Distributed Computing
A Fair Fast Scalable Rea,der-Writer Lock
ICPP '93 Proceedings of the 1993 International Conference on Parallel Processing - Volume 02
Time-optimal, space-efficient single-scanner snapshots & multi-scanner snapshots using CAS
Proceedings of the twenty-sixth annual ACM symposium on Principles of distributed computing
PRO: a model for the design and analysis of efficient and scalable parallel algorithms
Nordic Journal of Computing
Out-of-Core Wavefront Computations with Reduced Synchronization
PDP '08 Proceedings of the 16th Euromicro Conference on Parallel, Distributed and Network-Based Processing (PDP 2008)
A framework for scalable greedy coloring on distributed-memory parallel computers
Journal of Parallel and Distributed Computing
Atomic shared register access by asynchronous hardware
SFCS '86 Proceedings of the 27th Annual Symposium on Foundations of Computer Science
Graph coloring on coarse grained multicomputers
Discrete Applied Mathematics
Lock-Free parallel algorithms: an experimental study
HiPC'04 Proceedings of the 11th international conference on High Performance Computing
Modeling and experimental validation of the data handover API
GPC'11 Proceedings of the 6th international conference on Advances in grid and pervasive computing
Relaxed synchronization with ordered read-write locks
Euro-Par'11 Proceedings of the 2011 international conference on Parallel Processing
Hi-index | 0.00 |
We introduce the framework of ordered read-write locks, ORWL, that are characterized by two main features: a strict FIFO policy for access, and the attribution of access to lock-handles instead of processes or threads. These two properties together allow applications to have a controlled pro-active access to resources and thereby to achieve a high degree of asynchronicity between different tasks of the same application. For the case of iterative computations with many parallel tasks which access their resources in a cyclic pattern, we provide a generic technique to implement them by means of ORWL. We show that the possible execution patterns for such a system correspond to a combinatorial lattice structure and that this lattice is finite if and only if the configuration contains a potential deadlock. In addition, we provide efficient algorithms: one that allows for a deadlock-free initialization of such a system and another one for the detection of deadlocks in an already initialized system.