Failure correction techniques for large disk arrays
ASPLOS III Proceedings of the third international conference on Architectural support for programming languages and operating systems
A project on high performance I/0 subsystems
ACM SIGARCH Computer Architecture News
Large-scale sorting in parallel memories (extended abstract)
SPAA '91 Proceedings of the third annual ACM symposium on Parallel algorithms and architectures
Deterministic distribution sort in shared and distributed memory multiprocessors
SPAA '93 Proceedings of the fifth annual ACM symposium on Parallel algorithms and architectures
EVENODD: an optimal scheme for tolerating double disk failures in RAID architectures
ISCA '94 Proceedings of the 21st annual international symposium on Computer architecture
EVENODD: An Efficient Scheme for Tolerating Double Disk Failures in RAID Architectures
IEEE Transactions on Computers - Special issue on fault-tolerant computing
Greed sort: optimal deterministic sorting on parallel disks
Journal of the ACM (JACM)
Fast concurrent access to parallel disks
SODA '00 Proceedings of the eleventh annual ACM-SIAM symposium on Discrete algorithms
Asynchronous scheduling of redundant disk arrays
Proceedings of the twelfth annual ACM symposium on Parallel algorithms and architectures
Timely and fault-tolerant data access from broadcast disks: a pinwheel-based approach
CIKM '96 Proceedings of the workshop on Databases: active and real-time
Efficient Placement of Parity and Data to Tolerate Two Disk Failures in Disk Array Systems
IEEE Transactions on Parallel and Distributed Systems
Asynchronous Scheduling of Redundant Disk Arrays
IEEE Transactions on Computers
Hi-index | 0.00 |
The ever increasing need for I/O bandwidth will be met with ever larger arrays of disks. These arrays require redundancy to protect against data loss. This paper examines alternative choices for encodings, or codes, that reliably store information in disk arrays. Codes are selected to maximize mean time to data loss or minimize disks containing redundant data, but are all constrained to minimize performance penalties associated with updating information or recovering from catastrophic disk failures. We show codes that give highly reliable data storage with low redundant data overhead for arrays of 1000 information disks.