Gaussian codes and Shannon bounds for multiple descriptions

  • Authors:
  • R. Zamir

  • Affiliations:
  • Dept. of Electr. Eng.-Syst., Tel Aviv Univ.

  • Venue:
  • IEEE Transactions on Information Theory
  • Year:
  • 1999

Quantified Score

Hi-index 754.96

Visualization

Abstract

A pair of well-known inequalities, due to Shannon, upper/lower bound the rate-distortion function of a real source by the rate-distortion function of the Gaussian source with the same variance/entropy. We extend these bounds to multiple descriptions, a problem for which a general “single-letter” solution is not known. We show that the set DX(R1, R2) of achievable marginal (d1, d2) and central (d0) mean-squared errors in decoding X from two descriptions at rates R1 and R2 satisfies D*(σx2, R1, R2)⊆D X(R1, R2)⊆D*(Px, R1, R2) where σx2 and Px are the variance and the entropy-power of X, respectively, and D*(σ2, R1, R2) is the multiple description distortion region for a Gaussian source with variance σ2 found by Ozarow (1980). We further show that like in the single description case, a Gaussian random code achieves the outer bound in the limit as d1, d2→0, thus the outer bound is asymptotically tight at high resolution conditions