A Chebyshev Inequality For Multivariate Normal Distribution

  • Authors:
  • Davaadorjin Monhor

  • Affiliations:
  • Faculty of Geoinformatics, University of West Hungary, H-8002 Szekesfehervar, Hungary, E-mail: monhor@ella.hu

  • Venue:
  • Probability in the Engineering and Informational Sciences
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Chebyshev inequality estimates the probability for exceeding the deviation of a random variable from its mathematical expectation in terms of the variance of the random variable. In modern probability theory, the Chebyshev inequality is the most frequently used tool for proving different convergence processes; for example, it plays a fundamental role in proofs of various forms of laws of large numbers. The mathematical expression of the bound on the probability in the Chebyshev inequality is very simple and can be modified easily for different kinds of sequence of random variables (e.g., for the case of sums of independent random variables). This fact lies behind these frequent applications. In this setting, the Chebyshev inequality has pure theoretical “applications” in probability theory and its role is to provide “a guarantee” of convergence but not to give a bound on concrete probability content. In the present article we consider the Chebyshev inequality as a probability bound that is essential for the translation from its conventional theoretical applications to the practical setting if easy-to-compute multivariate generalizations are derived. Such an inequality for the random vectors having multivariate Normal distribution is proved. The new inequality gives a lower bound in terms of variances on the probability that the random vector in question falls into an Euclidean ball with center at mean vector. The need and importance of consideration of this kind of multivariate Chebyshev inequality stemmed from several problems in engineering and informational sciences (Hassibi and Boyd [9], Jeng [10], Jeng and Woods [11], Molina, Katseggelos, Mateos, Hermoso, and Segall [13]). Jeng [10] derived an inequality that gives an upper bound for the probability in question. The simultaneous application of the established multivariate Chebyshev inequality and Jeng's inequality is useful in practical problems by providing lower and upper bounds on the probability content. The inequality is attractive by its being easy to compute and its similarity to the original Chebyshev inequality, in contrast to well-known complicated multivariate Chebyshev inequalities. The present article also gives some insights into the very origin of the Chebyshev inequality, which makes the article self-contained.