On inequalities between mutual information and variation

  • Authors:
  • V. V. Prelov

  • Affiliations:
  • Kharkevich Institute for Information Transmission Problems, RAS, Moscow, Russia

  • Venue:
  • Problems of Information Transmission
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

We continue studying the relationship between mutual information and variational distance started in Pinsker's paper [1], where an upper bound for the mutual information via variational distance was obtained. We present a simple lower bound, which in some cases is optimal or asymptotically optimal. A uniform upper bound for the mutual information via variational distance is also derived for random variables with a finite number of values. For such random variables, the asymptotic behaviour of the maximum of mutual information is also investigated in the cases where the variational distance tends either to zero or to its maximum value.