Lecture 3

advertisement
Sept. 30, 2004 LEC #3
ECON 240A-1
Probability
L. Phillips
I. Introduction
Probability has its origins in curiosity about the laws governing gambles. During
the Renaissance, the Chevalier De Mere posed the following puzzle. Which is more
likely (1) rolling at least one six in four throws of a single die or (2) rolling at least one
double six in 24 throws of a pair of dice? De Mere asked his friend Blaise Pascal this
question. In turn Pascal enlisted the interest of Pierre De Fermat. Pascal and Fermat
worked out the theory of probability.
While gambling and probability are interesting topics in their own right,
probability is also a step towards making statistical inferences. The key is using a random
sample combined with probability models to estimate, for example, the fraction of voters
who will vote for a candidate. The pathway of understanding is a sequence of Bernoulli
trials leading to the binomial distribution, which for large numbers of observations can be
approximated by the normal distribution. These distributions can be used to estimate the
fraction that will vote yes and to calculate intervals within which the true fraction will lie
say 95 percent of the time.
II. Random Experiments
A key to understanding probability is to model certain activities such as flipping a
coin or throwing a die. The set of elementary outcomes, e.g. {heads, tails} or
symbolically {H, T}, is the sample space. Branching or tree diagrams illustrate these
random experiments and their elementary outcomes.
Sept. 30, 2004 LEC #3
ECON 240A-2
Probability
L. Phillips
-----------------------------------------------------------------------------------------------------------1
H
2
3
4
T
5
6
Figure 1: Flipping a Coin
Figure 2: Rolling a Die
Flipping a coin is an example of a Bernoulli trial, a random experiment with two
elementary outcomes, such as yes/no or heads/tails.
In a fair game, the probability of heads equals that of tails, for example, but the
laws of probability hold whether the game is fair or not. If the game is fair, the
probabilities of elementary outcomes for simple random experiments are intuitively
obvious, i.e. equally likely. In the case of flipping a coin, the probability of heads, P (H),
equals the probability of tails, P (T), equals one half. In the example of rolling a die, the
probability of rolling a one, P (1), equals the probability of rolling a six, P (6), equals one
sixth.
Probabilities of elementary outcomes are non-negative and the probabilities of all
the elementary outcomes sum to one.
Sept. 30, 2004 LEC #3
ECON 240A-3
Probability
L. Phillips
III. Events
Elementary outcomes can be grouped into sets that define an event. For example,
in the random experiment of throwing a die, the die can come up even, with the set of
elementary outcomes {2,4,6}, or the die can come up odd with the set of elementary
outcomes {1, 3, 5}. The probability of the event even is the sum of the probabilities of
the elementary outcomes in its set, so if the die is fair, P(even) = P(2) + P(4) + P(6) =
3/6.
Consider the random experiment of flipping a nickel and a quarter
simultaneously. There are four elementary outcomes: (1) nickel heads and quarter heads,
{NH, QH}, (2) nickel heads and quarter tails, {NH, QT}, (3) nickel tails and quarter
heads, {NT, QH}, and (4) nickel tails and quarter tails, {NT, QT}.
This is illustrated in Figure 3.
-----------------------------------------------------------------------------------------------------------NH, QH
NH, QT
NT, QH
NT, QT
Figure 3: Flipping a Nickel and a Quarter Simultaneously, Elementary Outcomes
-----------------------------------------------------------------------------------------------------------Once again, these elementary events can be combined into events, for example,
the event nickel heads: event {Nickel Heads} = {{NH, QH} {NH, QT}}.
Assuming the coins are fair, the probability of the event nickel heads, P (NH), is
the sum of the probabilities of the elementary events that compose it: P(NH) = P(NH,
QH) + P(NH, QT) = 2/4.
Sept. 30, 2004 LEC #3
ECON 240A-4
Probability
L. Phillips
Similarly, the event quarter tails involves two elementary outcomes:
event {QT} = {{NH, QT} {NT, QT}}, with probability,
P(QT) = P(NH, QT) + P(NT, QT) = 2/4.
These events, nickel heads and quarter tails can also be combined into events: (1)
nickel heads or quarter tails, {nickel headsquarter tails}, i.e. one event or the other takes
place, (2) nickel heads and quarter tails, {nickel headsquarter tails}, i.e. both events take
place. (3)The event not nickel heads, { NH }, called the event complementary to nickel
heads, is equivalent to the event nickel tails, {NT}.
IV. The Addition Rule
The event {nickel heads or quarter tails}, {NHQT}, includes three elementary
events:
{NHQT} = {NH,QH}, {NH,QT}, {NT,QT}. The probability of nickel heads or quarter
tails is the probability of the first row of Figure 3, plus the probability of the second
column of Figure 3, minus the overlap or intersection, i.e. the double counting of the
upper right hand corner. This leads to the addition rule: P(NHQT) = P(NH) + P(QT) –
P(NHQT) = 2/4 +2/4 – ¼ = ¾.
If two events, e.g. event A and event B are mutually exclusive, then P(AB) = 0,
so P() = P(A) + P(B). Since elementary event are mutually exclusive by definition,
events composed of combining them, for example nickel heads (above), have
probabilities which are just the sum of the elementary outcomes composing this event,
e.g. recall that P(NH) = P(NH, QH) + P(NH, QT) = 2/4.
The probability of an event, and the probability of not the event must sum to one,
for example, P( NH ) + P(NH) =1.
Sept. 30, 2004 LEC #3
ECON 240A-5
Probability
L. Phillips
V. Interpretations or Meanings of Probability
There are various perspectives on the meaning of probability. One meaning is
classical or a perspective based on gambling. The basic assumption is that the gamble or
random experiment is fair and that the elementary outcomes are equally likely.
But not all uncertainty in life is associated with gambles. Much of life itself is
uncertain. For example, will the Dow Jones Industrials end the fourth quarter of the year
2004 higher than it was at the end of the third quarter? One approach to such questions is
empirical. In the last ten years, what fraction of the quarters witnessed a rise in the Dow?
Another perspective on uncertainty in events is subjective or personal. For example,
the likelihood of success in your academic career and professional career could be
assessed empirically, based on the experiences of others who faced similar
circumstances. Alternatively, you could make a personal or subjective estimate of your
chances based on your own attributes and other information you consider relevant.
VI. Conditional Probability
Suppose that when flipping a nickel and a quarter simultaneously, the quarter
lands tails while the nickel is still spinning. What is the probability of the nickel landing
heads given the quarter landed tails, i.e. P (NH/QT)?
The answer is that this conditional probability of nickel heads given quarter tails
is equal to the probability of nickel heads and quarter tails divided by the probability of
quarter tails:
P(NH/QT) = P(NHQT)/P(QT) = 1/41/2 = ½.
This can be rearranged and stated as the multiplication rule:
P(NHQT) = P(NH/QT) P(QT).
Sept. 30, 2004 LEC #3
ECON 240A-6
Probability
L. Phillips
Another example is rolling a die. Suppose you are told that the number appearing
is even. What is the probability that the number is four? The conditional sample space has
been reduced since the odd elementary events are ruled out. So P(4/even) =1/3, or using
the multiplication rule:
P(4/even) = P(4even) /P(even) = 1/6/1/2 = 1/3.
Note that P(4/even) is not equal to P(4) = 1/6, where {4} is one of six elementary
outcomes.
VII. Independence
In the example above, whether the nickel lands heads or tails is not conditional on
the outcome for the quarter. Indeed, the outcome for the nickel is independent of the
outcome for the quarter. In this case, the conditional probability of the nickel landing
heads is just the probability of getting heads in the flip of a nickel, i.e.
P(NH/QT) = P (NH),
And so the multiplication rule can be stated, given independence, as:
P(NHQT) = P(NH) P(QT)
VIII. De Mere Again
With all of this probability apparatus at our disposal, we might turn to De Mere’s
question again. What is the probability of rolling at least one six in four throws of a single
die?
We could use a probability tree, as partially illustrated in Figure 4.
Sept. 30, 2004 LEC #3
ECON 240A-7
Probability
L. Phillips
1
2
1
3
2
4
3
5
4
6
5
6
Figure 4: Partial Probability Tree for Four Throws of the Die
-----------------------------------------------------------------------------------------------------If we had patience and a large enough piece of paper, we could enumerate all of the
elementary outcomes, such as {1,1,1,1}. But there will be 6*6*6*6 = 1296 elementary
outcomes. There must be an easier way.
We can use this conceptual framework of the tree or branching diagram. First, for
simplicity, consider two rolls of the die. Rolling at least one six could be one six on either
roll, or could be a six on each roll. The complement to this event is rolling zero sixes, no
six on the first roll and no six on the second roll. We can figure out the probability of this
event:
P( 6  6 ) = P( 6 ) * P ( 6 ) = (5/6)2 ,
since each roll of the die is a random experiment and hence the outcome on each roll is
independent of previous rolls of the die, we can use the special multiplication rule. It is a
simple extrapolation to four rolls of the die:
P( 6  6  6  6 ) = P( 6 ) * P ( 6 ) * P( 6 ) * P ( 6 ) = (5/6)4 .
Sept. 30, 2004 LEC #3
ECON 240A-8
Probability
L. Phillips
So the probability of rolling one or more sixes in four throws is:
1 – (5/6)4 = 1 – 625/1296 = 1 – 0.482 = 0.518.
In solving De Mere’s problem, we used the conceptual framework of the
probability tree, assumed a fair die, invoked random experiments to obtain independence,
and solved for the event complementary to one or more sixes. The same approach can be
used to solve for the probability of rolling at least one double six in 24 throws of a pair of
dice.
Independence is one key to the solution. We were able to invoke it because we conducted
random experiments. This is the motivation behind using random samples to make
inferences.
We will next pursue a sequence of Bernoulli events, for example, flipping a coin
two or more times, using the branching diagram and independence to derive the discrete
binomial distribution. We will see that for a large number of trials or flips, the binomial
can be approximated by the continuous normal or Gaussian distribution.
Sept. 30, 2004 LEC #3
ECON 240A-9
Probability
L. Phillips
Download