Consensus in the presence of partial synchrony
Journal of the ACM (JACM)
Combining tentative and definite executions for very fast dependable parallel computing
STOC '91 Proceedings of the twenty-third annual ACM symposium on Theory of computing
Time-optimal message-efficient work performance in the presence of faults
PODC '94 Proceedings of the thirteenth annual ACM symposium on Principles of distributed computing
Sharing memory robustly in message-passing systems
Journal of the ACM (JACM)
Parallel algorithms with processor failures and delays
Journal of Algorithms
Constructions of permutation arrays for certain scheduling cost measures
Random Structures & Algorithms
Algorithms for the Certified Write-All Problem
SIAM Journal on Computing
Performing Work Efficiently in the Presence of Faults
SIAM Journal on Computing
The do-all problem in broadcast networks
Proceedings of the twentieth annual ACM symposium on Principles of distributed computing
Fault-Tolerant Parallel Computation
Fault-Tolerant Parallel Computation
Bounding Work and Communication in Robust Cooperative Computation
DISC '02 Proceedings of the 16th International Conference on Distributed Computing
RAMBO: A Reconfigurable Atomic Memory Service for Dynamic Networks
DISC '02 Proceedings of the 16th International Conference on Distributed Computing
Resolving message complexity of Byzantine Agreement and beyond
FOCS '95 Proceedings of the 36th Annual Symposium on Foundations of Computer Science
Performing tasks on synchronous restartable message-passing processors
Distributed Computing
Emulating shared-memory Do-All algorithms in asynchronous message-passing systems
Journal of Parallel and Distributed Computing
Hi-index | 0.00 |
This paper considers the problem of performing tasks in asynchronous distributed settings. This problem, called Do-All, has been substantially studied in synchronous models, but there is a dearth of efficient algorithms for asynchronous message-passing processors. Do-All can be trivially solved without any communication by an algorithm where each processor performs all tasks. Assuming p processors and t tasks, this requires work @Q(p.t). Thus, it is important to develop subquadratic solutions (when p and t are comparable) by trading computation for communication. Following the observation that it is not possible to obtain subquadratic work when the message delay d is substantial, e.g., d=@Q(t), this work pursues a message-delay-sensitive approach. Here, the upper bounds on work and communication are given as functions of p, t, and d, the upper bound on message delays, however, algorithms have no knowledge of d and they cannot rely on the existence of an upper bound on d. This paper presents two families of asynchronous algorithms achieving, for the first time, subquadratic work as long as d=o(t). The first family uses as its basis a shared-memory algorithm without having to emulate atomic registers assumed by that algorithm. These deterministic algorithms have work O(tp^@e+pd@?t/d@?^@e) for any @e0. The second family uses specific permutations of tasks, with certain combinatorial properties, to sequence the work of the processors. These randomized (deterministic) algorithms have expected (worst-case) work O(tlogp+pdlog(2+t/d)). Another important contribution in this work is the first delay-sensitive lower bound for this problem that helps explain the behavior of our algorithms: any randomized (deterministic) algorithm has expected (worst-case) work of @W(t+pdlog"d"+"1t).