Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Chebyshev's inequality, college study notes - Chebyshev's inequality, Study notes of Signals and Systems Theory

Study Material. This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the Table of Contents. The notes were prepared by Ewa Paszek and Marek Kimmel. The development of this course has been supported by NSF 0203396 grant. Chebyshev’s Inequality, Connexions Web site. http://cnx.org/content/m13502/1.3/, Oct 8, 2007. Chebyshev, Inequality, Theorem, Inequality, Random, variable, Equation, Standard, deviation, Significance, Empirical,

Typology: Study notes

2011/2012

Uploaded on 10/17/2012

wualter
wualter 🇺🇸

4.8

(95)

288 documents

1 / 2

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
Connexions module: m13502 1
Chebyshev's Inequality
Ewa Paszek
This work is produced by The Connexions Project and licensed under the
Creative Commons Attribution License
Abstract
This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the
Table of Contents. The notes were prepared by Ewa Paszek and Marek Kimmel. The development of
this course has been supported by NSF 0203396 grant.
1 Chebyshev's Inequality
In this paragraph the Chebyshev's inequality is used to show, in another sense, that the sample mean,
x
,
is a good statistic to use to estimate a population with mean
µ
; the relative frequency of successes in
n
Bernoulli trials,
y/n
, is a good statistic for estimating
p
; and the empirical distribution function,
Fn(x)
,
can be used to estimate the theoretical distribution function
F(x)
. The eect of the sample size
n
on these
estimates is discussed.
At the beginning, it is showed that the Chebyshev's inequality gives added signicance to the standard
deviation in terms of bounding certain probabilities. The inequality is valid for all distributions for which
the standard deviation exists. The proof is given for the discrete case, but it holds for the continuous case
with integrals replacing summations.
Theorem 1:
Chebyshev's Inequality
If the random variable
X
has a mean
µ
and variance
σ2
, then for every
k1
,
P(|Xµ| )1
k2.
Proof:
Let
f(x)
denote p.d.f. of
X
. Then
σ2=Eh(Xµ)2i=X
xR
(xµ)2f(x) = X
xA
(xµ)2f(x) + X
xA
'
(xµ)2f(x),
where
A= (x:|xµ| )
. The second term in the right-hand member of the equation is the
sum of nonnegative numbers and thus is greater than or equal to zero, Hence
σ2X
xA
(xµ)2f(x).
Version 1.3: Oct 8, 2007 3:19 pm GMT-5
http://creativecommons.org/licenses/by/2.0/
http://cnx.org/content/m13502/1.3/
pf2

Partial preview of the text

Download Chebyshev's inequality, college study notes - Chebyshev's inequality and more Study notes Signals and Systems Theory in PDF only on Docsity!

Connexions module: m13502 1

Chebyshev's Inequality

Ewa Paszek

This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License †

Abstract This course is a short series of lectures on Introductory Statistics. Topics covered are listed in the Table of Contents. The notes were prepared by Ewa Paszek and Marek Kimmel. The development of this course has been supported by NSF 0203396 grant.

1 Chebyshev's Inequality

In this paragraph the Chebyshev's inequality is used to show, in another sense, that the sample mean, x, is a good statistic to use to estimate a population with mean μ ; the relative frequency of successes in n Bernoulli trials, y/n, is a good statistic for estimating p; and the empirical distribution function, Fn (x), can be used to estimate the theoretical distribution function F (x). The eect of the sample size n on these estimates is discussed. At the beginning, it is showed that the Chebyshev's inequality gives added signicance to the standard deviation in terms of bounding certain probabilities. The inequality is valid for all distributions for which the standard deviation exists. The proof is given for the discrete case, but it holds for the continuous case with integrals replacing summations.

Theorem 1: Chebyshev's Inequality If the random variable X has a mean μ and variance σ^2 , then for every k ≥ 1 ,

P (|X − μ| ≥ kσ) ≤

k^2

Proof: Let f (x) denote p.d.f. of X. Then

σ^2 = E

[

(X − μ)^2

]

x∈R

(x − μ)^2 f (x) =

x∈A

(x − μ)^2 f (x) +

x∈A'

(x − μ)^2 f (x) ,

where A = (x : |x − μ| ≥ kσ). The second term in the right-hand member of the equation is the sum of nonnegative numbers and thus is greater than or equal to zero, Hence

σ^2 ≥

x∈A

(x − μ)^2 f (x).

∗Version 1.3: Oct 8, 2007 3:19 pm GMT- †http://creativecommons.org/licenses/by/2.0/

http://cnx.org/content/m13502/1.3/

Connexions module: m13502 2

However, in A, |x − μ| ≥ kσ so

σ^2 ≥

x∈A

(kσ)^2 f (x) = k^2 σ^2

x∈A

f (x).

But the latter summation equals P (X ∈ A), and thus

σ^2 ≥ k^2 σ^2 P (X ∈ A) = k^2 σ^2 P (|X − μ| ≥ kσ).

That is, P (|X − μ| ≥ kσ) ≤

k^2

COROLLARY

If  = kσ, then P (|X − μ| ≥ ) ≤

σ^2 ^2

In words, Chebyshev's inequality states that the probability that X diers from its mean by at least k standard deviations is less than or equal to 1 k^2. It follows that the probability that X diers from its mean by less than k standard deviations is at least 1 k^2. That is,

P (|X − μ| < kσ) ≥ 1 −

k^2

From the corollary, it also follows that

P (|X − μ| < ) ≥ 1 −

σ^2 ^2

Thus Chebyshev's inequality can be used as a bound for certain probabilities. However, in many instances, the bound is not very close to the true probability.

Example 1 If it is known that X has a mean of 25 and a variance of 16, then, σ = 4 a lower bound for P (17 < X < 33) is given by

P (17 < X < 33) = P (|X − 25 | < 8) = P (|X − μ| < 2 σ) ≥ 1 −

and an upper bound for P (|X − 25 | ≥ 12) is found to be

P (|X − 25 | ≥ 12) = P (|X − μ| ≥ 3 σ) ≤

note: Note that the results of the last example hold for any distribution with mean 25 and standard deviation 4. But, even stronger, the probability that any random variable X diers from its mean by 3 or more standard deviations is at most 1/9 by letting k =3 in the theorem. Also the probability that any random variable X diers from its mean by less than 2 standard deviations is at least 3/ by letting k=2.

http://cnx.org/content/m13502/1.3/