

Study with the several resources on Docsity
Earn points by helping other students or get them with a premium plan
Prepare for your exams
Study with the several resources on Docsity
Earn points to download
Earn points by helping other students or get them with a premium plan
Community
Ask the community for help and clear up your study doubts
Discover the best universities in your country according to Docsity users
Free resources
Download our free guides on studying techniques, anxiety management strategies, and thesis advice from Docsity tutors
Study Material. This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the Table of Contents. The notes were prepared by Ewa Paszek and Marek Kimmel. The development of this course has been supported by NSF 0203396 grant. Chebyshev’s Inequality, Connexions Web site. http://cnx.org/content/m13502/1.3/, Oct 8, 2007. Chebyshev, Inequality, Theorem, Inequality, Random, variable, Equation, Standard, deviation, Significance, Empirical,
Typology: Study notes
1 / 2
This page cannot be seen from the preview
Don't miss anything!
Connexions module: m13502 1
This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License †
Abstract This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the Table of Contents. The notes were prepared by Ewa Paszek and Marek Kimmel. The development of this course has been supported by NSF 0203396 grant.
In this paragraph the Chebyshev's inequality is used to show, in another sense, that the sample mean, x, is a good statistic to use to estimate a population with mean μ ; the relative frequency of successes in n Bernoulli trials, y/n, is a good statistic for estimating p; and the empirical distribution function, Fn (x), can be used to estimate the theoretical distribution function F (x). The eect of the sample size n on these estimates is discussed. At the beginning, it is showed that the Chebyshev's inequality gives added signicance to the standard deviation in terms of bounding certain probabilities. The inequality is valid for all distributions for which the standard deviation exists. The proof is given for the discrete case, but it holds for the continuous case with integrals replacing summations.
Theorem 1: Chebyshev's Inequality If the random variable X has a mean μ and variance σ^2 , then for every k ≥ 1 ,
P (|X − μ| ≥ kσ) ≤
k^2
Proof: Let f (x) denote p.d.f. of X. Then
σ^2 = E
(X − μ)^2
x∈R
(x − μ)^2 f (x) =
x∈A
(x − μ)^2 f (x) +
x∈A'
(x − μ)^2 f (x) ,
where A = (x : |x − μ| ≥ kσ). The second term in the right-hand member of the equation is the sum of nonnegative numbers and thus is greater than or equal to zero, Hence
σ^2 ≥
x∈A
(x − μ)^2 f (x).
∗Version 1.3: Oct 8, 2007 3:19 pm GMT- †http://creativecommons.org/licenses/by/2.0/
http://cnx.org/content/m13502/1.3/
Connexions module: m13502 2
However, in A, |x − μ| ≥ kσ so
σ^2 ≥
x∈A
(kσ)^2 f (x) = k^2 σ^2
x∈A
f (x).
But the latter summation equals P (X ∈ A), and thus
σ^2 ≥ k^2 σ^2 P (X ∈ A) = k^2 σ^2 P (|X − μ| ≥ kσ).
That is, P (|X − μ| ≥ kσ) ≤
k^2
If = kσ, then P (|X − μ| ≥ ) ≤
σ^2 ^2
In words, Chebyshev's inequality states that the probability that X diers from its mean by at least k standard deviations is less than or equal to 1 k^2. It follows that the probability that X diers from its mean by less than k standard deviations is at least 1 k^2. That is,
P (|X − μ| < kσ) ≥ 1 −
k^2
From the corollary, it also follows that
P (|X − μ| < ) ≥ 1 −
σ^2 ^2
Thus Chebyshev's inequality can be used as a bound for certain probabilities. However, in many instances, the bound is not very close to the true probability.
Example 1 If it is known that X has a mean of 25 and a variance of 16, then, σ = 4 a lower bound for P (17 < X < 33) is given by
P (17 < X < 33) = P (|X − 25 | < 8) = P (|X − μ| < 2 σ) ≥ 1 −
and an upper bound for P (|X − 25 | ≥ 12) is found to be
P (|X − 25 | ≥ 12) = P (|X − μ| ≥ 3 σ) ≤
note: Note that the results of the last example hold for any distribution with mean 25 and standard deviation 4. But, even stronger, the probability that any random variable X diers from its mean by 3 or more standard deviations is at most 1/9 by letting k =3 in the theorem. Also the probability that any random variable X diers from its mean by less than 2 standard deviations is at least 3/ by letting k=2.
http://cnx.org/content/m13502/1.3/