Grid Computing: Making the Global Infrastructure a Reality
Grid Computing: Making the Global Infrastructure a Reality
Automatic Fingerprint Recognition Systems
Automatic Fingerprint Recognition Systems
Face recognition: A literature survey
ACM Computing Surveys (CSUR)
Disk failures in the real world: what does an MTTF of 1,000,000 hours mean to you?
FAST '07 Proceedings of the 5th USENIX conference on File and Storage Technologies
Failure trends in a large disk drive population
FAST '07 Proceedings of the 5th USENIX conference on File and Storage Technologies
ICB '09 Proceedings of the Third International Conference on Advances in Biometrics
IEEE Transactions on Circuits and Systems for Video Technology
Scripting distributed scientific workflows using Weaver
Concurrency and Computation: Practice & Experience
Hi-index | 0.00 |
Quality of data plays a very important role in any scientific research. In this paper we present some of the challenges that we face in managing and maintaining data quality for a terabyte scale biometrics repository. We have developed a step by step model to capture, ingest, validate, and prepare data for biometrics research. During these processes, there are many hidden errors which can be introduced into the data. Those errors can affect the overall quality of data, and thus can skew the results of biometrics research. We discuss necessary steps we have taken to reduce and eliminate the errors. Steps such as data replication, automated data validation, and logging metadata changes are both necessary and crucial to improve the quality and reliability of our data.