A novel lossless data compression scheme based on the error correcting Hamming codes

  • Authors:
  • Hussein Al-Bahadili

  • Affiliations:
  • Department of Computer Information Systems, Faculty of Information Systems and Technology, The Arab Academy for Banking and Financial Sciences, P.O. Box 13190, Amman 11942, Jordan

  • Venue:
  • Computers & Mathematics with Applications
  • Year:
  • 2008

Quantified Score

Hi-index 0.09

Visualization

Abstract

This paper introduces a novel lossless binary data compression scheme that is based on the error correcting Hamming codes, namely the HCDC scheme. In this scheme, the binary sequence to be compressed is divided into blocks of n bits length. To utilize the Hamming codes, the block is considered as a Hamming codeword that consists of p parity bits and d data bits (n=d+p). Then each block is tested to find if it is a valid or a non-valid Hamming codeword. For a valid block, only the d data bits preceded by 1 are written to the compressed file, while for a non-valid block all n bits preceded by 0 are written to the compressed file. These additional 1 and 0 bits are used to distinguish the valid and the non-valid blocks during the decompression process. An analytical formula is derived for computing the compression ratio as a function of block size, and fraction of valid data blocks in the sequence. The performance of the HCDC scheme is analyzed, and the results obtained are presented in tables and graphs. Finally, conclusions and recommendations for future works are pointed out.