Chapter 5

advertisement
Chapter 5: Part A
The Exponential Distribution

Computing probability by conditioning (Chapter 3, P. 119)
o Discrete: P[ E ]   P[ E | Y  y]P{Y  y}
y

o Continuous: P[ X ] 

 P[ X | Y  y] fY ( y)dy

Computing expectation by conditioning: (Chapter 3, P. 106)
o E[X|Y] means E[X|Y = y]
o Discrete: E[ X ]   E[ X | Y  y]P{Y  y}
y

o Continuous: E[ X ] 





 E[ X | Y  y] fY ( y)dy

Computing variance by conditioning (Chapter 3, P. 118)
o Var(X | Y) means E[[X – E(X |Y)]2 | Y]
o Var(X) = E[Var(X | Y)] + Var(E[X |Y])
The needs for mathematical models: Proper mathematical models allow us to
approximate real world phenomena with adequate details.
Simplification: Not so much that the mathematical model fails to represent the real
world phenomena but enough that it renders the mathematics tractable.
The exponential distribution: It is easy to work with and often a good approximation
of the real world.
Definition and characteristics:
o Probability density function:
 e  x
x0
f ( x)  
 0
x0
o Cumulative distribution function (CDF):
 1  e  x
x0
F ( x)  
0
x0
o  (or ) is called the rate of the distribution. It has a unit of 1/time. This
parameter determines the rate of occurrence of events.
o Another parameter that determines the exponential distribution is the mean
time  (= reciprocal of the rate). It indicates the mean time between events.
o Expectation:
E[X] = 1 / 
E[X 2] = 2 / 2
o Moment generating function (see p. 64 for more information on moment
generating function): (t) =  / ( – t)
o Variance (see p. 46-47 for more information on variance):
Var(X) =E[X2] – (E[X])2 = (t)” – (E[X])2 = 1 / 2
o Properties:
 Memoryless: P{X > s + t | X > t} = P{X > s} for all s , t  0
t
t+s
P{ X  s  t , X  t}
 P{ X  s}
s
P{ X  t}
Since (X > s + t)  (X > t),
we have P{X > s + t , X > t} = P{X > s + t}
For exponential distribution: e-(s + t) = e- s e- t
 P{ X  s  t}  P{ X  t}P{ X  s}





e se t
 t
 e   s  P{ X  s}
e
Example: If Michael Jordan scores an average of 1 basket per 3.5
minutes, a) if he fails to score in the first half, on average how long
would it take for him to score his first basket in the second half? b)
what is the probability that he would not score for an entire game?
a) 3.5 minutes (memoryless)
b) P{X>48} = 1 – P(X<48} = 1 – F(48) = 1 – (1 – e-t) = 1.1E-6
The exponential distribution is the only distribution possessing the
memoryless property.
Remaining lifetime after t years:
1  F (s  t )
1  F (t )
For exp. distribution: P{lifetime > s + t | lifetime > t} = 1 – F(s)
Failure rate r(t): The conditional probability density that a t-year-old
item will fail.
f (t )
In general: r (t ) 
1  F (t )
In general: P{lifetime > s + t | lifetime > t} =



Exponential distribution: r(t) =  e-t / e-t = 
Problems: 1, 2, 4, and 5
For two exponential processes with 1and 2:
P{ X i  X j } 

i
i   j
i, j  1,2,
For N exponential processes with 1, 2, … N , The minimum :
1
1
1
 min 


1  i max

i


1 x
x


P{min( X 1 , X 2 ,, X N )  x}  e   i  e   i
P{ X i  min X j } 
j
i
 j


E[tdeparture]
Example 5.3
Example: A facility performs 2 services for each customer. The first
service can only be served by Clerk #1 and the second service can then
be served by either clerk #2 or #3. Assuming that a customer can go to
either clerk with equal probability if both clerk 2 and 3 are not busy
and the service times are exponentially distributed with mean service
times 13, 15, and 14 minutes per customer, respectively. When you
arrive at 12 pm there is one customer (Mr. Jones) ahead of you, being
served by clerk #1, determine your expected departure time.
= E[twait for service 1] + E[tservice 1] + E[twait for service 2] + E[tservice 2]
= E[tw1] + E[ts1] + E[tw2] + E[ts2]
E[tw1] = E[ts1] = 13, E[tw2] = 0 (why?)
Jones
1/2
1/2
#2
#3
2
1   2
1
1   2
3
1  3
1/2


1/2
1
1  3
1/2


1/2


Total Expectation:
E[ts2] =
1 1
1 2   2   3  1 1
1 3   2   3 
3 
2 




2 1  2
2 1  2  2  2 1  3
2 1  3  2 
= 14.496
E[tdeparture]

=
40.496 minutes after 12 pm
Problems: 5-6, 15, 16, 20, 21, 22, 29, 30, and 32
Chapter 5: Part B
The Poisson Process

Poisson R.V. (Chapter 2, P. 32-33): It can be used to approximate the binomial R. V.
when n >> p.
i
p(i) = P{X = i} = e-

i = 0, 1, 2, …
i!
Examples:
o 2.11 (p. 33)
P{X = 0} = e-3 = 0.05
What about the probability that accident would occur today?
P{X  1} = 1 – e-3 = 0.95 (same as the exponential distribution: the
probability that the first accident would occur sometime today)
Or we can do the following:
0.95
o 2.12 (p. 33): Use both binomial and poisson
n large
n small



Stochastic processes (Chapter 2, P. 79):
o A stochastic process {X(t), t  T} is a family of random variables, that is, for
each t  T, X(t) is a random variable.
o A stochastic process describes the evolution through time of some (physical)
process.
o The index t is often referred to as time and as a result X(t) is referred to as the
state of the process at time t.
o T can be discrete or continuous.
Counting processes
o Definitions:
 A stochastic process {N(t), t  0} is a counting process if N(t)
represents the total number of events that have occurred from time
zero up to time t. A counting process has the following properties:
- N(t)  0
- N(t) is integer valued
- N(s)  N(t) if s < t
- For s < t, N(t) – N(s) = total number of events in (s, t)
 Independent increments: Number of events occurring in disjoint time
intervals are independent.
 Stationary increments: Number of events occurring in any time
intervals depends only on the length of the interval.
Poisson process
o Definitions:
 N(0) = 0
 Independent increments
 Stationary increments (For homogeneous Poisson processes only).
 It follows the Poisson distribution with mean =  t:
P{N (t  s)  N ( s)  n}  e
 t (t )
n!
n
n  0,1,2,
o Properties:
  is called the rate of the process.
 Expectation:
E[X] =  t
 Variance:
Var(X) =  t
 Example: The 1999 European soccer champion Manchester United is
one of the best offensive teams in the history with an average of 2.5
goals per game (90 minutes). a) What is probability that it would be
shut out for an entire game? b) What is probability that it would score
more than 2 goals for a game? What is probability that it would score 2
goals in 90 seconds?
a) t = 1, P{n = 0} = e-2.5 = 0.082
b) 1 – P{n = 0} – P{n = 1} – P{n = 2}
= 1 – e-2.5 – e-2.5 (2.5) – e-2.5 (2.5)2/2 = 0.46


c) t = 1.5 minutes  (1 game / 90 minutes) = 1/60 game
P{n = 2} = e-2.5/60 (2.5/60)2/2 = 0.00083
(note: In the 1999 European Champions League final, Manchester
United was behind 1 goal only a few minutes into the game. They
fought hard but failed to score for the rest of the game. After the
official 90 minutes expired, they score 2 goals during the 90
seconds injury time added at the end of the game and won the
European Championship that year.)
Interarrival time:
Tn is the elapsed time between the (n – 1)th and the nth event. Tn , n = 1,
2, …, are iid (independent identically distributed) exponential random
variables having mean 1/. This is a direct consequence of the
independent and stationary increments property of a Poisson process.
n
Waiting time: Sn =
 Ti
n1
i 1
Probability that the nth event occurs before or after t:
Before t: P{Sn  t} = P{N(t)  n} =
After t: P{Sn  t} = P{N(t) < n} =

e
j n
n 1
e
j 0

  t ( t )
j
j!
  t ( t )
j
j!
Example:
P.336, 5-37 and 38 (s – minute)
 = 3 / minute, P{N=0} = e-s = e-3s






P{S2  t} = P{N(t) < 2} = P{N=0} + P{N=1} = e-3s (3s + 1)
Example:
P.337, 5-42
E[S4] = 4/
E[S4|N(1) = 2] = 1 + E[time for 2 more events] = 1 + 2/
E[N(4) – N(2)|N(1) = 3] = E[N(4) – N(2)] = 2
Problems: p. 340, 5-57 and 58
Given that n events have occurred in (0, t), the set S1, S2, …, Sn are
uniformly distributed in (0, t).
Given that Sn = t, the set S1, S2, …, Sn-1 are uniformly distributed in
(0, t).
Example: Parts arrive at a warehouse at a Poisson rate of  per day and
all parts are shipped out weekly on Monday. If the storage cost is C
cents/part/day, the average weekly storage cost is:
Average weekly cost = C  7  7 / 2
Problem: p. 340, 5-60, 64 (part a only)
o Type I and type II events:
 If a Poisson process {N(t), t  0} contains two types of events, one
with probability p and the other (1 – p), then they called type I or type
II event, respectively, given that the probability of each event is
independent of all other events.
 Let N1(t) and N2(t) denote respectively the number of the type I and
type II events occurring in [0, t].
 N1(t) + N2(t) = N(t)
{N1(t), t  0} and {N2(t), t  0} are both independent Poisson
processes with rate  p and  (1 – p).
 Example: If the number of cars crossing a bridge are assumed to be a
Poisson process a rate of 1 per minute and 5% of the cars are trucks.
a) Find the probability that 1 or more trucks will cross the bridge in 1
hour.
 = 1/min, P = 0.05, P = 3/ hours
P{N1} = 1 – P{N=0} = 1 – e-3 = 0.95
b) If 10 trucks had crossed the bridge during the last hour, find the
expected number of cars crossed the bridge in that hour.
(1 – P) = 0.95/min = 57 non-truck/hour  10 + 57 = 67 cars
1
2
PS n  S m  - The probability that n events occur in one Poisson process (1)

o
before m events occur in a second and independent Poisson process (2).


1
1  2

n  m 1
P S11  S12 
P

S 1n
2
 Sm


 1 

P S 12  S12  
 1  2 
 n  m  1 1 


  
k
 1  2 
k n 
k
2
 2 


 1  2 
n  m 1 k
o Nonhomogeneous Poisson process: The rate  is a function of time.
t
 = (t), let m(t )    ( ) d and  t = m(t + s) – m(t)
0
We have: P{N(t+s) – N(s) = n} =

e
  t ( t )
n
n = 0, 1, 2, …
n!
Example: Text P. 344, 5-78, Problem 5-77 (answer: e -11  11n / n!)
10
8
4
8 10 12 2
5
42 + 82 + (82 + 22/2) + (34 + 36/2) = 8 + 16 + 18 + 21 = 63
12
 4 
 8 
P{N (10 : 10)  N (9 : 50)  1}  ?  t   10   10   2
6
 60 
 60 
e
 t (t )
n!
n
e
1
 2 ( 2)
1!
 0.271
Download