Hostname: page-component-89b8bd64d-5bvrz Total loading time: 0 Render date: 2026-05-07T03:29:28.862Z Has data issue: false hasContentIssue false

A CHEBYSHEV INEQUALITY FOR MULTIVARIATE NORMAL DISTRIBUTION

Published online by Cambridge University Press:  27 February 2007

Davaadorjin Monhor
Affiliation:
Faculty of Geoinformatics, University of West Hungary, H-8002 Szekesfehervar, Hungary, E-mail: monhor@ella.hu

Abstract

Chebyshev inequality estimates the probability for exceeding the deviation of a random variable from its mathematical expectation in terms of the variance of the random variable. In modern probability theory, the Chebyshev inequality is the most frequently used tool for proving different convergence processes; for example, it plays a fundamental role in proofs of various forms of laws of large numbers. The mathematical expression of the bound on the probability in the Chebyshev inequality is very simple and can be modified easily for different kinds of sequence of random variables (e.g., for the case of sums of independent random variables). This fact lies behind these frequent applications. In this setting, the Chebyshev inequality has pure theoretical “applications” in probability theory and its role is to provide “a guarantee” of convergence but not to give a bound on concrete probability content.

In the present article we consider the Chebyshev inequality as a probability bound that is essential for the translation from its conventional theoretical applications to the practical setting if easy-to-compute multivariate generalizations are derived.

Such an inequality for the random vectors having multivariate Normal distribution is proved. The new inequality gives a lower bound in terms of variances on the probability that the random vector in question falls into an Euclidean ball with center at mean vector. The need and importance of consideration of this kind of multivariate Chebyshev inequality stemmed from several problems in engineering and informational sciences (Hassibi and Boyd [9], Jeng [10], Jeng and Woods [11], Molina, Katseggelos, Mateos, Hermoso, and Segall [13]). Jeng [10] derived an inequality that gives an upper bound for the probability in question. The simultaneous application of the established multivariate Chebyshev inequality and Jeng's inequality is useful in practical problems by providing lower and upper bounds on the probability content.

The inequality is attractive by its being easy to compute and its similarity to the original Chebyshev inequality, in contrast to well-known complicated multivariate Chebyshev inequalities. The present article also gives some insights into the very origin of the Chebyshev inequality, which makes the article self-contained.

Information

Type
Research Article
Copyright
© 2007 Cambridge University Press

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

Article purchase

Temporarily unavailable