ACM Transactions on Programming Languages and Systems (TOPLAS)
Sharing memory robustly in message-passing systems
Journal of the ACM (JACM)
Distributed Algorithms
DISC '02 Proceedings of the 16th International Conference on Distributed Computing
Robust emulation of shared memory using dynamic quorum-acknowledged broadcasts
FTCS '97 Proceedings of the 27th International Symposium on Fault-Tolerant Computing (FTCS '97)
A Decentralized Algorithm for Erasure-Coded Virtual Disks
DSN '04 Proceedings of the 2004 International Conference on Dependable Systems and Networks
Efficient Byzantine-Tolerant Erasure-Coded Storage
DSN '04 Proceedings of the 2004 International Conference on Dependable Systems and Networks
Reliable and total order broadcast in the crash-recovery model
Journal of Parallel and Distributed Computing
Using Erasure Codes Efficiently for Storage in a Distributed System
DSN '05 Proceedings of the 2005 International Conference on Dependable Systems and Networks
Optimal Resilience for Erasure-Coded Byzantine Distributed Storage
DSN '06 Proceedings of the International Conference on Dependable Systems and Networks
Verifying distributed erasure-coded data
Proceedings of the twenty-sixth annual ACM symposium on Principles of distributed computing
Low-overhead byzantine fault-tolerant storage
Proceedings of twenty-first ACM SIGOPS symposium on Operating systems principles
The collective memory of amnesic processes
ACM Transactions on Algorithms (TALG)
Consistency and fault tolerance for erasure-coded distributed storage systems
Proceedings of the fifth international workshop on Data-Intensive Distributed Computing Date
Hi-index | 0.00 |
We study erasure-coded atomic register implementations in an asynchronous crash-recovery model. Erasure coding provides a cheap and space-efficient way to tolerate failures in a distributed system. This paper presents ORCAS, Optimistic eRasure-Coded Atomic Storage, which consists of two separate implementations, ORCAS-A and ORCAS-B. In terms of storage space used, ORCAS-A is more efficient in systems where we expect large number of concurrent writes, whereas, ORCAS-B is more suitable if not many writes are invoked concurrently. Compared to replication based implementations, both ORCAS implementations significantly save on the storage space. The implementations are optimistic in the sense that the used storage is lower in synchronous periods, which are considered common in practice, as compared to asynchronous periods. Indirectly, we show that tolerating asynchronous periods does not increase storage overhead during synchronous periods.