Frequently Asked Questions: Basic Probability 1. What does “correlation” mean? If two events are correlated, then they tend to change together. Knowing something about the one event tells you something about the other. If two events are independent, then the occurrence or nonoccurrence of the one tells you nothing about the other. Consider this example. I look out the window and tell you “I see a person coming out of the Wright building. What is your subjective probability that that person is female? Write it down on paper.” Then I tell you that the person’s last name is “Jones” and I invite you to revise your subjective probability. Since in your experience there has been no relationship between having a last name of “Jones” and being of the one or the other sex, your new subjective probability is unchanged. You consider the two events (being of the one sex or the other and having a last name of “Jones”) to be independent, that is, not correlated. Then I add, “and Jones is wearing a skirt, you may revise your subjective probability again, if you wish.” Now you probably would revise your subjective probability, upwards, since in your experience people who wear skirts are usually female. You think of being female and wearing a skirt as positively correlated (they tend to go together), not independent. Likewise, you think of wearing a skirt and being male as negatively correlated – knowing that the one event has happened makes the other event less likely. 2. How does one decide which approach to use when estimating a probability – empirical, rational, or subjective? It is both a matter of personal preference and the data that are available. Suppose I told you that I have a box of poker chips. Some are red, some are white, some are blue. I ask you to estimate the probability of drawing, on a single trial, a blue chip. You ask me if you can first experiment with the box of chips. I say “sorry, no.” You think about it (the rational approach). You make the assumption that the three types of chips are all equally likely to be drawn (there are equal numbers of them in the box, they are all the same size and shape, and they are randomly distributed within the box). You know (or assume) that drawing a blue chip and drawing a not blue chip are mutually exclusive events – a chip cannot be both blue and not blue, and the same is true of red versus not red and white versus not white. You also assume that the drawing blue, drawing red, and drawing white are mutually exhaustive events – there are no other events that can happen. Accordingly, you reason that P(blue) + P(red) + P (white) = 1 or 100%. Since you have assumed that the three events are equally likely, that is, P(blue) = P(red) = P(white), each of the probabilities must be 1/3. Now I offer you 3:1 odds against the drawn chip being blue (this is just another way of saying 1:3 odds in favor of the drawn chip being blue). You are thinking that is a good bet (better than fair) if betting on blue. If you are right, and the P(blue) is 1/3, then 2:1 odds would be fair. You take the bet. You put in $1, I put in $2. You draw a chip from the box. It is white. You lose a dollar. Hey, what do you expect, you probably would not hit blue on the first draw, bet again. We go through ten bets, and you get only one blue chip. You won $2 on that one bet but lost $9 on the other nine. That does not sound fair. I want you to keep on betting. You say “only if I can experiment with the box first.” Reluctantly, I say “OK.” You do a 1,000 trial experiment. On each trial you draw one chip, note its color, and then put it back in the box and shake the box well. In 1,000 trials you draw blue chips only 102 times. From that result, you estimate the probability of drawing a blue chip to be 102/1,000 = about 10%. This is the empirical approach. After you finish, I say “OK, lets bet.” You say “OK, but the bet won’t be fair unless the odds are 9:1 against getting a blue chip.” The subjective approach is whatever it is that humans do when making decisions that involve the likelihood of something happening or not. Although there are strong common threads here, there are also individual differences. Generally speaking, humans make a lot of errors in this domain. There is a large and interesting research literature on this topic, and I would love to talk about it, but, frankly, we just do not have the time to talk about that right now. 3. What is the relationship between the addition rule and mutual exclusion? Suppose that I ask you to bet on whether the next chip drawn from the box will be blue or red. From your experiment you know that the P(blue) = 10% and the P(red) = 30%. The addition rule is P(A B) = P(A) + P(B) if A and B are mutually exclusive. You did not see any chips that were both red and blue, so you estimate the P(blue or red) to be 10% + 30% = 40%. A fair bet would be 60:40 or 3:2 odds against the chip being blue or red. Now I bring out a new box of chips. I promise you that for this box P(blue) = 10% and P(red) = 30%. What is the probability that a randomly drawn chip will be blue or red? What I do not tell you is that 5% of the chips are red with blue polka dots, that is they are both blue and red. Now being blue and being red are not mutually exclusive. In this case the simplified addition rule (above) will not work, we need the more general addition rule, P(A B) = P(A) + P(B) - P(A B). P(blue or red) = 10% + 30% -5% = 35%, Look back at my handout “Basic Probability, Including Contingency Tables” for another example of the addition rule when events are not mutually exclusive, where we would have decided that 127% of female students are going to hell if we did not take into account that smoking and being sexually active are not mutually exclusive. Consider the diagram to the right. The area of the red circle on the left represents P(A). The area of the blue circle on the right represents P(B). The overlap between the two circles, with the purple dots, represents P(AB), the joint occurrence of A and B – events that are both A and B. A and B are not mutually exclusive. If we were to calculate P(A B) = P(A) + P(B) we would count the overlap area twice. To correct for this, we need to subtract out the overlap area once, that is, P(A B) = P(A) + P(B) P(A B).