Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

The Negative Binomial Distribution - Final Exam Notes | MATH 329, Exams of Probability and Statistics

Material Type: Exam; Professor: Neal; Class: PROBAB/STAT I; Subject: Mathematics (Univ); University: Western Kentucky University; Term: Unknown 1989;

Typology: Exams

Pre 2010

Uploaded on 08/18/2009

koofers-user-8ja
koofers-user-8ja 🇺🇸

10 documents

1 / 3

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
Dr. Neal, WKU
MATH 329 The Negative Binomial
Distribution
A negative binomial random variable, denoted by X~nb(r,p), is a generalization of the
geometric random variable. Suppose we have probability p>0 of success on every
attempt and we make a sequence of independent attempts. Then
X
counts the number
of attempts needed to obtain the
r
th success, for some designated integer
r1
. For
r=1
, then X~geo(p). Because its takes at least
r
attempts to obtain
r
successes, the
range of X~nb(r,p) is {
r
,
r+1
,
r+2
, . . . }.
Probability Distribution Function
We again let q=1p be the probability of a failure on each attempt. To have
X
equal
k
, for some
kr
, we must have exactly
r
successes in
k
attempts with the last success
being on the
k
th (i.e., final) attempt. And we must choose
r1
of the preceding
k1
attempts to have the other successes. So there are k1
r1
ways to have a sequence of
k
attempts with exactly
r
successes and with the last success being on the last attempt.
And each such individual sequence occurs with probability p
rqkr
. Thus, the
probability of having the
r
th success on the
k
th attempt, for
kr
, is given by
P(X=k)=k1
r1
prqkr.
There is no closed-form formula for the cumulative probability P(Xk) or for
computing probabilities such as P(jXk). In each case, the individual probabilities
must be summed, or we must use a calculator/computer command:
P(Xk) = i=r
= i1
r1
prqir
i=r
k
and P(jXk) = i1
r1
prqir
i=j
k
.
Expected Value, Variance, and Standard Deviation
We can easily derive the expected value and variance of X~nb(r,p) by writing
X
as a
sum of independent geometric random variables. First, let X
1
~ geo(p) be the number
of attempts needed for the first success. Then for
2ir
, let Xi ~ geo(p) be the
additional number of attempts after the (i1)st success until the
i
th success.
Then the Xi are independent of each other (due to independent attempts) and the
total number of attempts needed for
r
successes is given by
X
1+X2+... +Xr=X~nb(r,p)
. Moreover, E[Xi]=1/ p and Var(Xi)=q/p
2
for
1ir
. By the linearity of expected value, we then have
E[X]=E[X1+X2+.. .+Xr]=E[X1]+E[X2]+... +E[Xr]
=1/ p+1/ p+... +1/ p=r
p.
pf3

Partial preview of the text

Download The Negative Binomial Distribution - Final Exam Notes | MATH 329 and more Exams Probability and Statistics in PDF only on Docsity!

MATH 329 The Negative Binomial

Distribution

A negative binomial random variable, denoted by X ~ nb ( r , p ), is a generalization of the geometric random variable. Suppose we have probability p > 0 of success on every attempt and we make a sequence of independent attempts. Then X counts the number of attempts needed to obtain the r th success, for some designated integer r ≥ 1. For r = 1 , then X ~ geo ( p ). Because its takes at least r attempts to obtain r successes, the range of X ~ nb ( r , p ) is { r , r + 1 , r + 2 ,... }.

Probability Distribution Function

We again let q = 1 − p be the probability of a failure on each attempt. To have X equal k , for some kr , we must have exactly r successes in k attempts with the last success being on the k th (i.e., final) attempt. And we must choose r − 1 of the preceding k − 1

attempts to have the other successes. So there are

k − 1 r − 1

 (^) ways to have a sequence of

k attempts with exactly r successes and with the last success being on the last attempt.

And each such individual sequence occurs with probability pr^ qkr^. Thus, the probability of having the r th success on the k th attempt, for kr , is given by

P ( X = k ) =

k − 1 r − 1

 (^) pr^ qk^ − r^.

There is no closed-form formula for the cumulative probability P ( Xk ) or for computing probabilities such as P ( jXk ). In each case, the individual probabilities must be summed, or we must use a calculator/computer command:

P ( Xk ) = i = r

k

∑ P (^ X^ =^ i )^ =^

i − 1 r − 1

 (^) pr^ qir i = r

k

∑ and^ P (^ j^ ≤^ X^ ≤^ k )^ =^

i − 1 r − 1

 (^) pr^ qir i = j

k

Expected Value, Variance, and Standard Deviation

We can easily derive the expected value and variance of X ~ nb ( r , p ) by writing X as a sum of independent geometric random variables. First, let X 1 ~ geo ( p ) be the number of attempts needed for the first success. Then for 2 ≤ ir , let Xi ~ geo ( p ) be the

additional number of attempts after the ( i − 1)st success until the i th success. Then the Xi are independent of each other (due to independent attempts) and the total number of attempts needed for r successes is given by

X 1 + X 2 +. .. + Xr = X ~ nb ( r , p ). Moreover, E [ Xi ] = 1/ p and Var ( Xi ) = q / p^2 for

1 ≤ ir. By the linearity of expected value, we then have

E [ X ] = E [ X 1 + X 2 + ... + Xr ] = E [ X 1 ] + E [ X 2 ] +... + E [ Xr ]

= 1/ p + 1/ p +. .. + 1/ p =

r p

Then by independence of the Xi , we have

Var ( X ) = Var ( X 1 + X 2 +... + Xr ) = Var ( X 1 ) + Var ( X 2 ) + ... + Var ( Xr )

= q / p^2 + q / p^2 + ... + q / p^2 = rq p^2

And for X ~ nb ( r , p ), the standard deviation is given by (^) X =

rq p

Mode

The most likely number of attempts needed to attain r successes, i.e., the mode , is always

given by the largest integer k such that k

rq p

, denoted by k =

rq p

. However, a

negative binomial distribution will be bi-modal when k =

rq p

is an integer. In this

case, this value of k and the previous integer k − 1 will be the two modes, except in the case of r = 1. When r = 1 , then k = 1 is the only mode. The expression for the mode can be derived with the same argument used for deriving the mode of a binomial distribution.

Example. Suppose we have a 40% chance of winning a bet and we play until we win 5 times.

(a) What is the most likely number of attempts needed for 5 wins and the what is the probability of needing exactly this number of attempts? (b) What are the average number and the standard deviation of the number of attempts needed to win 5 times? (c) What is the probability that it takes at least 10 attempts to achieve 5 wins?

Solution. Here X ~ nb (5,0.40). (a) The most likely number of attempts needed for 5

wins is k =

^

 =^

^

 =^ ^11 ^.^ So 10 and 11 are both the most likely number of attempts needed with

P ( X = 10) =

 0.4^5 0.6^5 ≈ 0.100329 and P ( X = 11) =

 0.4 5 0.6^6 ≈ 0..

(b) The average number of attempts needed is E [ X ] = 5/ 0.4 = 12.5 with a standard

deviation of (^) X =

5 × 0.

≈ 4.33. (c) The probability that it takes at least 10 attempts

is given by

P X (^ ≥ 10 ) = 1 − P (5 ≤ X ≤ 9)

= 1 −

 0.4^5 0.6 0 −

 0.4^5 0.6^1 −

 0.4 5 0.6^2 −

 0.4^5 0.6^3 −

 0.4^5 0.6^4