Combinatorics: set systems, hypergraphs, families of vectors, and combinatorial probability
Combinatorics: set systems, hypergraphs, families of vectors, and combinatorial probability
Multi-disk management algorithms
SIGMETRICS '87 Proceedings of the 1987 ACM SIGMETRICS conference on Measurement and modeling of computer systems
A case for redundant arrays of inexpensive disks (RAID)
SIGMOD '88 Proceedings of the 1988 ACM SIGMOD international conference on Management of data
Simulation and the Monte Carlo Method
Simulation and the Monte Carlo Method
Proceedings of the Second International Conference on Data Engineering
Coding Techniques for Handling Failures in Large Disk Arrays
Coding Techniques for Handling Failures in Large Disk Arrays
A project on high performance I/0 subsystems
ACM SIGARCH Computer Architecture News
Beyond striping: the bridge multiprocessor file system
ACM SIGARCH Computer Architecture News
A study parallel disk organizations
ACM SIGARCH Computer Architecture News
An Evaluation of Multiple-Disk I/O Systems
IEEE Transactions on Computers
Dynamic file allocation in disk arrays
SIGMOD '91 Proceedings of the 1991 ACM SIGMOD international conference on Management of data
Parity logging overcoming the small write problem in redundant disk arrays
ISCA '93 Proceedings of the 20th annual international symposium on computer architecture
Improved parity-declustered layouts for disk arrays
SPAA '94 Proceedings of the sixth annual ACM symposium on Parallel algorithms and architectures
The TickerTAIP parallel RAID architecture
ACM Transactions on Computer Systems (TOCS)
Crosshatch disk array for improved reliability and performance
ISCA '94 Proceedings of the 21st annual international symposium on Computer architecture
Evaluating approximately balanced parity-declustered data layouts for disk arrays
Proceedings of the fourth workshop on I/O in parallel and distributed systems: part of the federated computing research conference
Flexible usage of parity storage space in disk arrays
Proceedings of the eighth annual ACM symposium on Parallel algorithms and architectures
Tolerating multiple failures in RAID architectures with optimal storage and uniform declustering
Proceedings of the 24th annual international symposium on Computer architecture
Node-covering, Error-correcting Codes and Multiprocessors with Very High Average Fault Tolerance
IEEE Transactions on Computers
Declustered disk array architectures with optimal and near-optimal parallelism
Proceedings of the 25th annual international symposium on Computer architecture
IEEE Transactions on Parallel and Distributed Systems
Reliability and performance of hierarchical RAID with multiple controllers
Proceedings of the twentieth annual ACM symposium on Principles of distributed computing
Data partitioning and load balancing in parallel disk systems
The VLDB Journal — The International Journal on Very Large Data Bases
Algorithm-Based Diskless Checkpointing for Fault-Tolerant Matrix Operations
FTCS '95 Proceedings of the Twenty-Fifth International Symposium on Fault-Tolerant Computing
Node Covering, Error Correcting Codes and Multiprocessors with Very High Average Fault Tolerance
FTCS '95 Proceedings of the Twenty-Fifth International Symposium on Fault-Tolerant Computing
Physical Database Design: the database professional's guide to exploiting indexes, views, storage, and more
WEAVER codes: highly fault tolerant erasure codes for storage systems
FAST'05 Proceedings of the 4th conference on USENIX Conference on File and Storage Technologies - Volume 4
Trust-Based Design and Check of FPGA Circuits Using Two-Level Randomized ECC Structures
ACM Transactions on Reconfigurable Technology and Systems (TRETS)
DiskReduce: RAID for data-intensive scalable computing
Proceedings of the 4th Annual Workshop on Petascale Data Storage
Hi-index | 0.01 |
The ever increasing need for I/O bandwidth will be met with ever larger arrays of disks. These arrays require redundancy to protect against data loss. This paper examines alternative choices for encodings, or codes, that reliably store information in disk arrays. Codes are selected to maximize mean time to data loss or minimize disks containing redundant data, but are all constrained to minimize performance penalties associated with updating information or recovering from catastrophic disk failures. We also codes that give highly reliable data storage with low redundant data overhead for arrays of 1000 information disks.