Probability and Discrete Random Variable Probability What is Probability? • When we talk about probability, we are talking about a (mathematical) measure of how likely it is for some particular thing to happen • Probability deals with chance behavior – We study outcomes, or results of experiments – Each time we conduct an experiment, we may get a different result Why study Probability? • Descriptive statistics - describing and summarizing sample data, deals with data as it is • Probability - Modeling the probabilities for sample results to occur in situations in which the population is known • The combination of the two will let us do our inferential statistics techniques. Objectives 1. Learn the basic concepts of probability 2. Understand the rules of probabilities 3. Compute and interpret probabilities using the empirical method 4. Compute and interpret probabilities using the classical method 5. Compute the probabilities for the compound events. Sample Space & Outcomes • Some definitions – An experiment is a repeatable process where the results are uncertain – An outcome is one specific possible result – The set of all possible outcomes is the sample space denoted by a capital letter S • Example – Experiment … roll a fair 6 sided die – One of the outcomes … roll a “4” – The sample space … roll a “1” or “2” or “3” or “4” or “5” or “6”. So, S = {1, 2, 3, 4, 5, 6} (Include all outcomes in braces {…}.) Event • More definitions – An event is a collection of possible outcomes … we will use capital letters such as E for events – Outcomes are also sometimes called simple events … we will use lower case letters such as e for outcomes / simple events • Example (continued) – One of the events … E = {roll an even number} – E consists of the outcomes e2 = “roll a 2”, e4 = “roll a 4”, and e6 = “roll a 6” … we’ll write that as {2, 4, 6} Example Consider an experiment of rolling a die again. – There are 6 possible outcomes, e1 = “rolling a 1” which we’ll write as just {1}, e2 = “rolling a 2” or {2}, … – The sample space is the collection of those 6 outcomes. We write S = {1, 2, 3, 4, 5, 6} – One event of interest is E = “rolling an even number”. The event is indicated by E = {2, 4, 6} Probability of an Event • If E is an event, then we write P(E) as the probability of the event E happening • These probabilities must obey certain mathematical rules Probability Rule # 1 • Rule # 1 – the probability of any event must be greater than or equal to 0 and less than or equal to 1, i.e., 0 P( E ) 1 – It does not make sense to say that there is a -30% chance of rain – It does not make sense to say that there is a 140% chance of rain Note – probabilities can be written as decimals (0, 0.3, 1.0), or as percents (0%, 30%, 100%), or as fractions (3/10) Probability Rule # 2 • Rule #2 – the sum of the probabilities of all the outcomes must equal 1. P (e ) 1 i all outcomes – If we examine all possible outcomes, one of them must happen – It does not make sense to say that there are two possibilities, one occurring with probability 20% and the other with probability 50% (where did the other 30% go?) Example On the way to work Bob’s personal judgment is that he is four times more likely to get caught in a traffic jam (TJ) than have an easy commute (EC). What values should be assigned to P(TJ) and P(EC)? Solution: Given P (TJ ) 4 P( EC) Since P(TJ ) P( EC) 1 Which means 4 P ( EC) P ( EC) 1 5 P ( EC) 1 1 P ( EC) 5 1 4 P ( TJ ) 4 P ( EC) 4 5 5 Probability Rule (continued) • Probability models must satisfy both of these rules • There are some special types of events – If an event is impossible, then its probability must be equal to 0 (i.e. it can never happen) – If an event is a certainty, then its probability must be equal to 1 (i.e. it always happens) Unusual Events • A more sophisticated concept – An unusual event is one that has a low probability of occurring – This is not precise … how low is “low? • Typically, probabilities of 5% or less are considered low … events with probabilities of 5% or lower are considered unusual • However, this cutoff point can vary by the context of the problem How To Compute the Probability? The probability of an event may be obtained in three different ways: – Theoretically (a classical approach) – Empirically (an experimental approach) – Subjectively Compute Probability theoretically Equally Likely Outcomes • The classical method of calculating the probability applies to situations (or by assuming the situations) where all possible outcomes have the same probability which is called equally likely outcomes • Examples – Flipping a fair coin … two outcomes (heads and tails) … both equally likely – Rolling a fair die … six outcomes (1, 2, 3, 4, 5, and 6) … all equally likely – Choosing one student out of 250 in a simple random sample … 250 outcomes … all equally likely Equally Likely Outcomes • Because all the outcomes are equally likely, then each outcome occurs with probability 1/n where n is the number of outcomes • Examples – Flipping a fair coin … two outcomes (heads and tails) … each occurs with probability 1/2 – Rolling a fair die … six outcomes (1, 2, 3, 4, 5, and 6) … each occurs with probability 1/6 – Choosing one student out of 250 in a simple random sample … 250 outcomes … each occurs with probability 1/250 Theoretical Probability • The general formula is P (E ) Number of ways E can occur Number of possible outcomes • If we have an experiment where – There are n equally likely outcomes (i.e. N(S) = n) – The event E consists of m of them (i.e. N(E) = m) then m N (E ) P (E ) n N(S ) A More Complex Example Here we consider an example of select two subjects at random instead of just one subject: Three students (Katherine (K), Michael (M), and Dana (D)) want to go to a concert but there are only two tickets available. Two of the three students are selected at random. Question 1: What is the sample space of who goes? Solution: S = {(K,M),(K,D),(M,D)} Question 2: What is the probability that Katherine goes? Solution: Because 2 students are selected at random, each outcome in the sample space has equal chance to occur. Therefore, P( Katherine goes) = 2/3. Another Example A local automobile dealer classifies purchases by number of doors and transmission type. The table below gives the number of each classification. Manual Automatic Transmission Transmission 2-door 75 155 4-door 85 170 If one customer is selected at random, find the probability that: 1) The selected individual purchased a car with automatic transmission 2) The selected individual purchased a 2-door car Solutions Apply the formula P (E ) m N (E ) n N(S ) 1) P(Automatic Transmission) 155 170 325 65 75 85 155 170 485 97 2) P(2 - door ) 75 155 230 46 75 85 155 170 485 97 Compute Probability empirically Empirical Probability • If we do not know the probability of a certain event E, we can conduct a series of experiments to approximate it by P (E ) frequency of E number of trials of the experiment • This is called the empirical probability or experimental probability. It becomes a good approximation for P(E) if we have a large number of trials (the law of large numbers) Example We wish to determine what proportion of students at a certain school have type A blood – We perform an experiment (a simple random sample!) with 100 students – If 29 of those students have type A blood, then we would estimate that the proportion of students at this school with type A blood is 29% Example (continued) We wish to determine what proportion of students at a certain school have type AB blood – We perform an experiment (a simple random sample!) with 100 students – If 3 of those students have type AB blood, then we would estimate that the proportion of students at this school with type AB blood is 3% – This would be an unusual event Another Example Consider an experiment in which we roll two six-sided fair dice and record the number of 3s face up. The only possible outcomes are zero 3s, one 3, or two 3s. Here are the results after 100 rolls of these two dice, and also after 1000 rolls: 100 Rolls Outcome Frequency 0 80 1 19 2 1 1000 Rolls Outcome Frequency 0 690 1 282 2 28 Using a Histogram • We can express these results (from the 1000 rolls) in terms of relative frequencies and display the results using a histogram: 0.7 0.6 Relative Frequency 0.5 0.4 0.3 0.2 0.1 0.0 0 1 Three’s Face Up 2 Continuing the Experiment If we continue this experiment for several thousand more rolls, the relative frequency for each possible outcome will settle down and approach to a constant. This is so called the law of large numbers. Coin-Tossing Experiment Consider tossing a fair coin. Define the event H as the occurrence of a head. What is the probability of the event H, P(H)? • Theoretical approach – If we assume that the coin is fair, then there are two equally likely outcomes in a single toss of the coin. Intuitively, P(H) = 50%. • Empirical approach – If we do not know if the coin is fair or not. We then estimate the probability by tossing the coin many times and calculating the proportion of heads occurring. To show you the effect of applying large number of tosses on the accuracy of the estimation. What we actually do here is to toss the coin 10 times each time and repeated it 20 times. The results are shown in the next slide. We cumulate the total number of tosses over trials to compute the proportion of heads. We plot the proportions over trials in a graph as shown in the following slide. We observe that the proportion of heads tends to stabilize or settle down near 0.5 (50%). So, the proportion of heads over larger number of tosses is a better estimate of the true probability P(H). Experimental results of tossing a coin 10 times on each trial Trial 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Number of Relative Cumulative Heads Observed Frequency Relative Frequency 5 5/10 5/10 = 0.5000 4 4/10 9/20 = 0.4500 4 4/10 13/30 = 0.4333 5 5/10 18/40 = 0.4500 6 6/10 24/50 = 0.4800 7 7/10 28/60 = 0.4667 6 6/10 34/70 = 0.4857 4 4/10 38/80 = 0.4750 7 7/10 45/90 = 0.5000 3 3/10 48/100 = 0.4800 4 4/10 52/110 = 0.4727 6 6/10 58/120 = 0.4838 7 7/10 65/130 = 0.5000 4 4/10 69/140 = 0.4929 3 3/10 72/150 = 0.4800 7 7/10 79/160 = 0.4938 6 6/10 85/170 = 0.5000 3 3/10 88/180 = 0.4889 6 6/10 94/190 = 0.4947 4 4/10 98/200 = 0.4900 Cumulative Relative Frequency 0.6 Expected value = 1/2 0.55 0.5 0.45 0.4 0 5 10 15 20 25 Trial This stabilizing effect, or long-term average value, is often referred to as the law of large numbers. Law of Large Numbers If the number of times an experiment is repeated is increased, the ratio of the number of successful occurrences to the number of trials will tend to approach the theoretical probability of the outcome for an individual trial – Interpretation: The law of large numbers says: the larger the number of experimental trials, the closer the empirical probability is expected to be to the true probability P(A) Subjective Probability 1. Suppose the sample space elements are not equally likely and empirical probabilities cannot be used 2. Only method available for assigning probabilities may be personal judgment 3. These probability assignments are called subjective probabilities Summary • Probabilities describe the chances of events occurring … events consisting of outcomes in a sample space • Probabilities must obey certain rules such as always being greater than or equal to 0 and less then or equal to 1. • There are various ways to compute probabilities, including empirical method and classical method for experiments with equally likely outcomes. Compute Probabilities for Compound Events Venn Diagram • Venn Diagrams provide a useful way to visualize probabilities – The entire rectangle represents the sample space S – The circle represents an event E S E Example • In the Venn diagram below – The sample space is {0, 1, 2, 3, …, 9} – The event E is {0, 1, 2} – The event F is {8, 9} – The outcomes {3}, {4}, {5}, {6}, {7} are in neither event E nor event F Mutually Exclusive Events • Two events are disjoint if they do not have any outcomes in common • Another name for this is mutually exclusive • Two events are disjoint if it is impossible for both to happen at the same time • E and F below are disjoint Example The following table summarizes visitors to a local amusement park: Male Female Total All-Day Pass 1200 900 2100 Half-Day Pass 800 700 1500 Total 2000 1600 3600 One visitor from this group is selected at random: 1) Define the event A as “the visitor purchased an all-day pass” 2) Define the event B as “the visitor selected purchased a half-day pass” 3) Define the event C as “the visitor selected is female” Solutions 1) The events A and B are mutually exclusive 2) The events A and C are not mutually exclusive. The intersection of A and C can be seen in the table above or in the Venn diagram below: 3) A C 1200 900 700 800 Addition Rule for Disjoint Events • For disjoint events, the outcomes of (E or F) can be listed as the outcomes of E followed by the outcomes of F • There are no duplicates in this list • The Addition Rule for disjoint events is P(E or F) = P(E) + P(F) • Thus we can find P(E or F) if we know both P(E) and P(F) Addition Rule for More than Two Disjoint Events • This is also true for more than two disjoint events • If E, F, G, … are all disjoint (none of them have any outcomes in common), then P(E or F or G or …) = P(E) + P(F) + P(G) + … • The Venn diagram below is an example of this Example • In rolling a fair die, what is the chance of rolling a {2 or lower} or a {6} – The probability of {2 or lower} is 2/6 – The probability of {6} is 1/6 – The two events {1, 2} and {6} are disjoint • The total probability is 2/6 + 1/6 = 3/6 = 1/2 Note • • • • The addition rule only applies to events that are disjoint If the two events are not disjoint, then this rule must be modified Some outcomes will be double counted The Venn diagram below illustrates how the outcomes {1} and {3} are counted both in event E and event F Example In rolling a fair die, what is the chance of rolling a {2 or lower} or an even number? – The probability of {2 or lower} is 2/6 – The probability of {2, 4, 6} is 3/6 – The two events {1, 2} and {2, 4, 6} are not disjoint – The total probability is not 2/6 + 3/6 = 5/6 – The total probability is 4/6 because the event is {1, 2, 4, 6} Note: When we say A or B, we include outcomes either in A or in B or both. General Addition Rule • For the formula P(E) + P(F), all the outcomes that are in both events are counted twice • Thus, to compute P(E or F), these outcomes must be subtracted (once) • The General Addition Rule is P(E or F) = P(E) + P(F) – P(E and F) • This rule is true both for disjoint events and for not disjoint events. when E and F are disjoint, P(E and F) = 0 which leads to P(E or F) = P(E) + P(F) Example • When choosing a card at random out of a deck of 52 cards, what is the probability of choosing a queen or a heart? – E = “choosing a queen” – F = “choosing a heart” • E and F are not disjoint (it is possible to choose the queen of hearts), so we must use the General Addition Rule Solution • P(E) = P(queen) = 4/52 • P(F) = P(heart) = 13/52 • P(E and F) = P(queen of hearts) = 1/52, so P (queen or heart) P (queen) P (heart) P (queen and heart) 4 13 1 52 52 52 16 52 Another Example A manufacturer is testing the production of a new product on two assembly lines. A random sample of parts is selected and each part is inspected for defects. The results are summarized in the table below: Line 1 (1) Line 2 (2) Total Good (G) 70 80 150 Defective (D) 40 25 65 Total 110 105 215 Suppose a part is selected at random: 1) Find the probability the part is defective 2) Find the probability the part is produced on Line 1 3) Find the probability the part is good or produced on Line 2 Solutions 65 1) P( D) 215 (total defective divided by total number of parts) 110 2) P(1) 215 (total produced by Line 1 divided by total number of parts) 3) P(G or 2) n(G or 2) 175 n( S ) 215 (total good or produced on Line 2 divided by total parts) 175 P(G ) P(2) P (G and 2) 150 105 80 215 215 215 215 Complement Events • The set of all sample points in the sample space that do not belong to event E. The complement of event E is denoted by E or E c (read “E complement”). For example, The complement of the event “success” is “failure” The complement of the event “rain” is “no rain” The complement of the event “at least 3 patients recover” out of 5 patients is “2 or fewer recover” • The probability of the complement of an event E is 1 minus the probability of E P( E ) 1 P( E ) • This can be shown in one of two ways It’s obvious … if there is a 30% chance of rain, then there is a 70% chance of no rain. E = Rain, E = No rain E and E are two disjoint events that add up to the entire sample space Illustration The Complement Rule can also be illustrated using a Venn diagram Entire region The area of the region outside the circle represents Ec Notes: Complementary events are also mutually exclusive/disjoint. Mutually exclusive events are not necessarily complementary Example All employees at a certain company are classified as only one of the following: manager (A), service (B), sales (C), or staff(D). It is known that P(A) = 0.15, P(B) = 0.40, P(C) = 0.25, and P(D) = 0.20 Solution: P(A) 1 P(A) 1 0.15 0.85 P( A and B ) 0 (A and B are mutually exclusive ) P ( B or C ) P ( B) P ( C ) 0.40 0.25 0.65 P ( A or B or C ) P ( A ) P ( B) P ( C ) 0.15 + 0.40 + 0.25 = 0.80 Summary • Probabilities obey Additional Rules • For disjoint events, the Addition Rule is used for calculating “or” probabilities • For events that are not disjoint, the Addition Rule is not valid … instead the General Addition Rule is used for calculating “or” probabilities • The Complement Rule is used for calculating “not” probabilities Independence • Definition of independence: Events E and F are independent if the occurrence of E in a probability experiment does not affect the probability of event F • Other ways of saying the same thing – Knowing E does not give any additional information about F – Knowing F does not give any additional information about E – E and F are totally unrelated Examples of Independence – Flipping a coin and getting a “tail” (event E) and choosing a card and getting the “seven of clubs” (event F) – Choosing one student at random from University A (event E) and choosing another student at random from University B (event F) – Choosing a card and having it be a heart (event E) and having it be a jack (event F) Dependent Events • If the two events are not independent, then they are said to be dependent • Dependent does not mean that they completely rely on each other … it just means that they are not independent of each other • Dependent means that there is some kind of relationship between E and F – even if it is just a very small relationship Examples of Dependence – Whether Jack has brought an umbrella (event E) and whether his roommate Joe has brought an umbrella (event F) – Choosing a card and having it be a red card (event E) and having it be a heart (event F) – The number of people at a party (event E) and the noise level at the party (event F) Multiplication Rule Let A and B be two events defined in sample space S. If A and B are independent events, then: P(A and B) P(A) P( B) This formula can be expanded. If A, B, C, …, G are independent events, then P(A and B and C and ... and G) P(A) P( B) P(C) P(G) Example: Suppose the event A is “Allen gets a cold this winter,” B is “Bob gets a cold this winter,” and C is “Chris gets a cold this winter.” P(A) = 0.15, P(B) = 0.25, P(C) = 0.3, and all three events are independent. Find the probability that: 1. All three get colds this winter 2. Allen and Bob get a cold but Chris does not 3. None of the three gets a cold this winter Solutions 1) P(All three get colds this winter ) P(A and B and C) P(A ) P( B) P(C) = (0.15)(0.25)(0.30) = 0.0113 2) P ( Allen and Bob get a cold, but Chris does not) P ( A and B and C ) P ( A ) P ( B) P ( C ) = (0.15)(0.25)(0.70) = 0.0263 3) P ( None of the three gets a cold this winter) P ( A and B and C ) P ( A ) P ( B) P ( C) = (0.85)(0.75)(0.70) = 0.4463 Example A fair coin is tossed 5 times, and a head(H) or a tail (T) is recorded each time. What is the probability A = {at least one head in 5 tosses} Solution: Apply the complement rule: P(A) 1 P(A) 1 P(0 heads in 5 tosses ) 1 31 1 32 32 Note: P(0 heads in 5 tosses) = P( all tails in 5 tosses ) = P(1st toss is a head) P(2nd toss is a head)… p(5th toss is a head) 1 = 1 1 1 1 1 = (due to independence of tosses) 2 2 2 2 2 32 Example In a sample of 1200 residents, each person was asked if he or she favored building a new town playground. The responses are summarized in the table below: Age Less than 30 (Y) 30 to 50 (M) More than 50 (O) Total Favor (F) 250 600 100 950 Oppose 50 75 125 250 Total 300 675 225 1200 If one resident is selected at random, what is the probability the resident will: 1) Favor the new playground? 2) Be between 30 to 50 years old? 3) Are the events F and M independent? Solutions 1. P(F) 950 19 1200 2. 3. 24 675 9 P(M ) 1200 16 600 1 P(F & M ) 1200 2 Since P(F and M) is not equal to P(F) P(M), the two events F and M are not independent. Mutually Exclusive Events Are Not Independent • What’s the difference between disjoint events and independent events? • Disjoint events can never be independent – – – – Consider two events E and F that are disjoint Let’s say that event E has occurred Then we know that event F cannot have occurred Knowing information about event E has told us much information about event F – Thus E and F are not independent Summary • Compound Events are formed by combining several simple events: The probability that either event A or event B will occur: P(A or B) The probability that both events A and B will occur: P(A and B) • The “disjoint” concept corresponds to “or” and the Addition Rule … disjoint events and adding probabilities • The concept of independence corresponds to “and” and the Multiplication Rule … independent events and multiplying probabilities Discrete Probability Distributions Discrete Random Variables, Discrete Probability Distribution The Binomial Probability Distribution Learning Objectives 1. Distinguish between discrete and continuous random variables 2. Identify discrete probability distributions 3. Construct probability histograms 4. Compute and interpret the mean of a discrete random variable 5. Interpret the mean of a discrete random variable as an expected value 6. Compute the variance and standard deviation of a discrete random variable Discrete Random Variable Random Variables If the outcomes from an experiment are quantitative type (i.e. numbers), we denote the outcomes of this type with a variable using a capital letter such as X, Y , Z … That is, the variable contains all possible values (i.e. outcomes) from the experiment. Each possible value is denoted with lower case letter such as x, y, z… Since there is a probability/chance for each value of the variable to occur, we call the variable as a random variable. Examples of Random Variables • Tossing four coins and counting the number of heads – The number could be 0, 1, 2, 3, or 4 – The number could change when we toss another four coins • Measuring the heights of students - The heights could vary from student to student • Recording the number of computers sold per day by a local merchant with a random variable. Integer values ranging from zero to about 50 are possible values. Discrete Random Variables • A discrete random variable is a random variable that has either a finite or a countable number of values – A finite number of values such as {0, 1, 2, 3, and 4} – A countable number of values such as {1, 2, 3, …} • Discrete random variables are often “counts of …” Examples of Discrete Random Variables • The number of heads in tossing 3 coins There are four possible values – 0 heads, 1 head, 2 heads, and 3 heads – A finite number of possible values – a discrete random variable – This fits our general concept that discrete random variables are often “counts of …” • The number of pages in statistics textbooks – A countable number of possible values • The number of visitors to the White House in a day – A countable number of possible values Continuous Random Variables • A continuous random variable is a random variable that has an infinite, and more than countable, number of values – The values are any number in an interval • Continuous random variables are often “measurements of …” Examples of Continuous Random Variables • The possible temperature in Chicago at noon tomorrow, measured in degrees Fahrenheit – The possible values (assuming that we can measure temperature to great accuracy) are in an interval. So, 20 degrees can be recorded as 20.4 degrees or 20.41 degrees … etc. – The interval may be something like from -20 to 110 degrees. – This fits our general concept that continuous random variables are often “measurements of …” • The height of a college student – A value in an interval between 3 and 8 feet • The number of bytes of storage used on a 80 GB (80 billion bytes) hard drive – Although this is discrete, it is more reasonable to model it as a continuous random variable between 0 and 80 GB Discrete Probability Distribution Probability Distribution • Probability Distribution is a distribution of the probabilities associated with each of the values of a random variable. • The probability distribution is a theoretical distribution because the probabilities are theoretical probabilities; it is used to represent populations. Discrete Probability Distribution • The probability distribution of a discrete random variable X relates the values of X with their corresponding probabilities • A distribution could be – In the form of a table – In the form of a graph – In the form of a mathematical formula Probability Function • If X is a discrete random variable and x is a possible value for X, then we write P(x) as the probability that X is equal to x • Examples – In tossing one coin, if X is the number of heads, then P(0) = 0.5 and P(1) = 0.5 – In rolling one die, if X is the number rolled, then P(1) = 1/6 Properties of P(x) • Since P(x) form a probability distribution, they must satisfy the rules of probability – 0 ≤ P(x) ≤ 1 – Σ P(x) = 1 • In the second rule, the Σ sign means to add up the P(x)’s for all the possible x’s Probability Distribution Table • An example of a discrete probability distribution. All of the possible values x are listed in one column of a table, the corresponding probability P(x) for each value is listed on the next column as shown below: x 1 2 5 6 P(x) .2 .6 .1 .1 • All of the P(x) values need to be not only positive but also add up to 1 Not a Probability Distribution • An example that is not a probability distribution x 1 2 5 6 • Two things are wrong P(x) .2 .6 -.3 .1 – P(5) is negative – The P(x)’s do not add up to 1 Example The number of people staying in a randomly selected room at a local hotel is a random variable ranging in value from 0 to 4. The probability distribution is known and is given in various forms below: x P (x ) 0 2/15 1 4/15 2 5/15 3 3/15 4 1/15 Notes: This chart implies the only values x takes on are 0, 1, 2, 3, and 4 5 P( the randomvariablex equals 2) P(2) 15 Probability Histogram • A probability histogram is a histogram where – The horizontal axis corresponds to the possible values of X (i.e. the x’s) – The vertical axis corresponds to the probabilities for those values (i.e. the P(x)’s) • A probability histogram is very similar to a relative frequency histogram Probability Histogram • An example of a probability histogram • The histogram is drawn so that the height of the bar is the probability of that value Notes • The histogram of a probability distribution uses the area of each bar to represent its assigned probability • The width of each bar is 1 and the height of each bar is the assigned probability, so the area of each bar is also equal to the corresponding probability • The idea of area representing probability is important in the study of continuous random variables later Probability Function Sometimes the probability distribution for a random variable x is given by a functional expression. For example, 8 x P( x ) 15 for x 3, 4, 5, 6, 7 Find the probability associated with each value by using the probability function. Solution: 83 5 P(3) 15 15 86 2 P(6) 15 15 P(4) 84 4 15 15 P(7) P(5) 85 3 15 15 87 1 15 15 Note: P( x) follows the probability rules; each number is between 0 and 1 and sum of the probabilities is 1. Mean of Probability Distribution • Probability distribution is a population distribution, because the probability is regarded as an idealized relative frequency for an outcome to occur if the experiment is repeated large number of times (mostly infinite times if we can). • Since the mean of a population is denoted by a parameter m, the mean of a probability distribution is denoted by m as well • The mean of a probability distribution can be thought of in this way: – There are various possible values of a discrete random variable – The values that have the higher probabilities are the ones that occur more often – The values that occur more often should have a larger role in calculating the mean – The mean of the probability distribution is the weighted average of the values, weighted by the probabilities Mean of Discrete Probability Distribution • The mean of a discrete random variable is μ = Σ [ x • P(x) ] • In this formula – x are the possible values of X – P(x) is the probability that x occurs – Σ means to add up all of the products of these terms for all the possible values x Mean of Discrete Probability Distribution • Example of a calculation for the mean Multiply x P(x) x • P(x) 1 0.2 0.2 2 0.6 1.2 Multiply again 5 0.1 0.5 Multiply again 6 0.1 0.6 Multiply again • Add: 0.2 + 1.2 + 0.5 + 0.6 = 2.5 • The mean of this discrete random variable is 2.5 Mean of Discrete Probability Distribution • The calculation for this problem written out μ = Σ [ x • P(x) ] = [1• 0.2] + [2• 0.6] + [5• 0.1] + [6• 0.1] = 0.2 + 1.2 + 0.5 + 0.6 = 2.5 • The mean of this discrete random variable is 2.5 • The mean is an average value, so it does not have to be one of the possible values for X or an integer. Interpret the Mean of a Probability Distribution • The mean can also be thought of this way (as in the Law of Large Numbers) – If we repeat the experiment many times – If we record the result each time – If we calculate the mean of the results (this is just a mean of a group of numbers) – Then this mean of the results gets closer and closer to the mean of the random variable Expected Value • The expected value of a random variable is another term for its mean. That is, we often interpret the mean of a discrete random variable as an expected value • The term “expected value” illustrates the long term nature of the experiments as described on the previous slide – as we perform more and more experiments, the mean of the results of those experiments gets closer to the “expected value” of the random variable Variance and Standard Deviation of Discrete Probability Distribution Variance of a discrete random variable denoted by s2, is found by multiplying each possible value of the squared deviation from the mean, (x m)2, by its own probability and then adding all the products together: s 2 [( x m ) 2 P( x )] [ x 2 P( x )] { [ xP( x )]} 2 [ x 2 P( x )] m 2 Standard deviation of a discrete random variable is the positive square root of the variance: s s2 Short-cut Formula for Variance of Discrete Probability Distribution • The variance formula σ2 = Σ [ (x – μ)2 • P(x) ] can involve calculations with many decimals or fractions • An short-cut formula is σ2 = [ Σ x2 • P(x) ] – μ2 This formula is often easier to compute Example • To find the mean, variance and standard deviation of a probability distribution, you can extended the probability table as below: x2 P( x ) xP ( x ) x 2 P( x ) x 3 4 5 6 7 5/15 4/15 3/15 2/15 1/15 15/15 Totals 15/15 16/15 15/15 12/15 7/15 65/15 [ xP( x)] m [ xP ( x )] s 2 65 4.33 15 2 305 65 x P(x) m 1.56 15 15 2 2 s s 2 1.56 1.25 9 16 25 36 49 45/15 64/15 75/15 72/15 49/15 305/15 2 [ x P( x)] Calculate Mean and Variance of Discrete Probability Distribution • You can get the mean and variance of a discrete probability distribution from a TI graphing calculator. The procedure is the same as the one for getting the sample statistics from a frequency distribution (grouped data). • Enter all possible values of a variable in one list, say L1, and their corresponding probabilities in another list, say L2. Then, STAT CALC 1 ENTER L1, L2 ENTER • The sample mean x shown on the screen is actually the mean of the discrete probability distribution. The population standard deviation s x is the standard deviation s of the discrete probability distribution. ( Ignore other calculated statistics) m Summary • Discrete random variables are measures of outcomes that have discrete values • Discrete random variables are specified by their probability distributions which are regarded as population distributions • The mean of a discrete random variable is a parameter, can be interpreted as the long term average of repeated independent experiments • The variance of a discrete random variable is a parameter, measures its dispersion from its mean The Binomial Probability Distribution Learning Objectives 1. Determine whether a probability experiment is a binomial experiment 2. Compute probabilities of binomial experiments 3. Compute the mean and standard deviation of a binomial random variable 4. Construct binomial probability histograms Binomial Experiment • A binomial experiment has the following structure – The first trial of the experiment is performed … the result is either a success or a failure (Outcomes are classified into two categories, so it is term binomial experiment.) – The second trial is performed … the result is either a success or a failure. This result is independent of the first and the chance of success is the same as the first trial. – A third trial is performed … the result is either a success or a failure. The result is independent of the first two and the chance of success is the same – The process can go on and on. Example – A card is drawn from a deck. A “success” is for that card to be a heart … a “failure” is for any other suit – The card is then put back into the deck – A second card is drawn from the deck with the same definition of success. – The second card is put back into the deck – We continue for drawing 10 cards Binomial Experiment A binomial experiment is an experiment with the following characteristics – The experiment is performed a fixed number of times, each time called a trial – The trials are independently performed – Each trial has two possible outcomes, usually called a success (desired outcomes) and a failure (the rest of other outcomes) – The probability of success is the same for every trial Note: If an experiment contains more than 2 outcomes. We can always classify the outcome into two categories, success and failure. For instance, tossing a die, success is having a number less than 3, then failure is having a number not less than 3. Binomial Probability Distribution • Notation used for binomial distributions – The number of trials is represented by n – The probability of a success is represented by p – The total number of successes in n trials is the ransom variable X, the outcome observed • Because there cannot be a negative number of successes, and because there cannot be more than n successes (out of n attempts) 0≤X≤n Example • In our card drawing example – Each trial is the experiment of drawing one card – The experiment is performed 10 times, so n = 10 – The trials are independent because the drawn card is put back into the deck so that a card drawn before will not affect a card drawn next. – Each trial has two possible outcomes, a “success” of drawing a heart and a “failure” of drawing anything else – The probability of success is 0.25 (Since there are four suits in a deck where heart is one of the suits), the same for every trial, so p = 0.25 – X, the number of successes, is between 0 and 10 Notes • The word “success” does not mean that this is a good outcome or that we want this to be the outcome • A “success” in our card drawing experiment is to draw a heart • If we are counting hearts, then this is the outcome that we are measuring • There is no good or bad meaning to “success” Calculate Binomial Probability • We would like to calculate the probabilities of X, i.e. P(0), P(1), P(2), …, P(n) • Do a simpler example first – For n = 3 trials (e.g. toss a coin 3 time) – With p = .4 probability of success (e.g. Head is a success. The coin is loaded such that P(head) = 0.4) – Calculate P(2), the probability of 2 successes (i.e. 2 heads) Calculate Binomial Probability • For 3 trials, the possible ways of getting exactly 2 successes are –S S F –S F S –F S S • The probabilities for each (using the multiplication rule of independent events) are – 0.4 • 0.4 • 0.6 = 0.096 – 0.4 • 0.6 • 0.4 = 0.096 – 0.6 • 0.4 • 0.4 = 0.096 Calculate Binomial Probability • The total probability is P(2) = 0.096 + 0.096 + 0.096 = 0.288 • But there is a pattern – Each way had the same probability, … the probability of 2 success (0.4 times 0.4) times the probability of 1 failure (0.6 times 0.6), because each way contained 2 successes and 1 failure regardless of the order of the success and failure. • The probability for each case is (0.4)2 • (0.6)1 Counting formula • There are 3 possible sequences of success and failure – S S F could represent choosing a combination of 2 out of 3 … choosing the first and the second – S F S could represent choosing a second combination of 2 out of 3 … choosing the first and the third – F S S could represent choosing a third combination of 2 out of 3 • These are the 3 ways to choose 2 out of 3, which can be computed by a counting formula 3C2 or 3 ( read as 3 2 choose 2) Calculate Binomial Probability • Thus the total probability P(2) = 0.096 + 0.096 + 0.096 = 0.288 can also be written as P(2) = 3C2 • (0.4)2 • (0.6)1 • In other words, the probability is – The number of ways of choosing 2 out of 3, times – The probability of 2 successes, times – The probability of 1 failure General Formula for Binomial Probabilities • The general formula for the binomial probabilities is just this: • For P(x), the probability of x successes, the probability is – The number of ways of choosing x out of n, times – The probability of x successes, times – The probability of n-x failures • This formula is P(x) = nCx px (1 – p)n-x Calculate nCx • The formula nCx (called the binomial coefficient) for counting the number of ways to have x number of successes out of n trials can be computed as below: n! nC x = x!(n x )! Where n! is an abbreviation for n factorial: n! n (n 1)( n 2) (3)( 2)(1) • It is easier to compute it from the calculator For instance, to compute 3C2: 1. Enter 3 2. Click MATH PRB 3:nCr 3. Enter 2 Solution: 3C2 = 3 Example • A student guesses at random on a multiple choice quiz – There are n = 10 questions in total – There are 5 choices per question so that the probability of success (guess correctly) p = 1/5 = .2 (due to random guessing, each choice has equal chance) • What is the probability that the student gets 6 questions correct? Example Continued First, check if this is a binomial experiment – There are a finite number n = 10 of trials (10 questions, answer one question each trial) – Each trial has two outcomes (a correct guess and an incorrect guess) – The probability of success is independent from trial to trial (every question is done by a random guess) – The probability of success p = .2 is the same for each trial Example Continued • The probability of 6 correct guesses is P(x) = nCx px (1 – p)n-x = 10C6 (0.2)6 (0.8)4 = 210 • 0.000064 • 0.4096 = 0.005505 • This is less than a 1% chance • In fact, the chance of getting 6 or more correct (i.e. a passing score) is also less than 1% Example According to a recent study, 65% of all homes in a certain county have high levels of radon gas leaking into their basements. Four homes are selected at random and tested for radon. The random variable x is the number of homes with high levels of radon (out of the four). Solution: First check if it is a binomial experiment: 1. There are 4 repeated trials: n = 4. The trials are independent. 2. Each test for radon is a trial, and each test has two outcomes: radon or no radon 3. p = P(radon) = 0.65, q = P(no radon) = 0.35 4. x is the number of homes with high levels of radon out of 4 homes selected, possible values: 0, 1, 2, 3, 4 Yes! So apply the binomial probability distribution with n= 4, p=0.65 to compute the probabilities Example continued 4 P ( x ) (0.65 ) x (0.35 ) 4 x , for x 0, 1, 2, 3, 4 x 4 0 4 P ( 0 ) (0.65 ) (0. 35 ) (1)( 1)( 0 .0150 ) 0 .0150 0 4 1 3 P (1) (0. 65 ) (0. 35 ) ( 4 )( 0 .65 )( 0 .0429 ) 0 .1115 1 4 P ( 2 ) (0.65 ) 2 (0. 35 ) 2 ( 6 )( 0 .4225 )( 0 .1225 ) 0 .3105 2 4 P ( 3 ) (0. 65 ) 3 (0. 35 ) 1 ( 4 )( 0 .2746 )( 0 .35 ) 0 .3845 3 4 P ( 4 ) (0. 65 ) 4 (0. 35 ) 0 ( 1)( 0 .1785 )( 1) 0 .1785 4 Example In a certain automobile dealership, 70% of all customers purchase an extended warranty with their new car. For 15 customers selected at random: 1) Find the probability that exactly 12 will purchase an extended warranty 2) Find the probability at most 13 will purchase an extended warranty Solutions: • Let x be the number of customers who purchase an extended warranty. x is a binomial random variable. • The probability function associated with this experiment: 15 P( x ) (0.7) x (0.3)15 x , for x 0, 1, 2, ... ,15 x Solutions Continued 1) Probability exactly 12 purchase an extended warranty: 15 P(12) (0.7)12 (0.3) 3 0.1700 12 2) Probability at most 13 purchase an extended warranty: P( x 13) P(0) P(1) ... P(13) 1 P(14) P (15) 15 14 1 15 15 0 1 (0.7) (0. 3) (0.7) (0.3) 15 14 1 [0.0305 0.0047] 1 0.0352 0.9648 Note: Instead computing P (x 13) directly, It is easier to apply the complement rule to compute it’s probability by 1 – P (x >13) Calculate Binomial Probabilities • • • It is possible to use tables to look up these probabilities It is best to use a calculator routine or a software program to compute these probabilities. With TI graphing calculator, follow the steps below: Example 1: Compute an individual probability P(6) for a binomial random variable with n = 10, p = 0.4 using binompdf ( ) 1. Distr[2NDVARS] A:binompdf ENTER 2. Enter 10, 0.4, 6), then hit ENTER Solution: P(6) = 0.1115 • Example 2: Compute a cumulative probability for x = 0 to 5 using binomcdf( ), i.e., P (x 5) = P(0) + P(1) + P(2) + p(3) + P(4) + p(5) 1. Distr[2NDVARS] B:binomcdf ENTER 2. Enter 10, 0.4, 5), then hit ENTER Solution: P (x 5) = 0.8338 Note: Need to enter n, p, x in order to binompdf(n,p,x) ot binomcdf(n, p, x) Example • For the automobile dealership example considered on the previous slide. Apply a binomial probability with n = 15 and p = 0.7. Check the answers with a graphing calculator. Solution: P(12) = binompdf(15, 0.7, 12) = 0.1700 P ( x 13) = binomcdf(15, 0.7, 13) = 0.9647 • Also, find the cumulative probability x = 4 to 8 inclusively (means including 4 and 8). That is, P(4 x 8) = P(4) + P(5) + P(6) + P(7) + P(8) Solution: P(4 x 8) = binomcdf(15, 0.7, 8) – binomcdf(15, 0.7, 3) = 0.1311 Note: binomcdf(15, 0.7, 8) = P(0)+P(1)+P(2)+P(3)+P(4)+P(5)+P(6)+P(7)+P(8) binomcdf(15, 0.7, 3) = P(0)+P(1)+P(2)+P(3) So, the difference of these two cumulative probabilities covers the probabilities for x from 4 to 8. So, do not subtract binomcdf(15, 0.7, 4) from the binomcdf(15, 0.7, 8) since P(4) needs to be included. Mean of Binomial Probability Distribution • We would like to find the mean of a binomial distribution • You may apply the general formula for mean of a discrete probability distribution using μ = Σ [ x • P(x) ] to find the mean • It turns out that, for a binomial probability distribution, the mean can be quickly computed by μ=np • Example – There are 10 questions – The probability of success (correct guess) is 0.20 on each one – Then the expected number of correct answers would be 10 • 0.20 = 2 Variance and standard Deviation of a Binomial Probability Distribution • We would like to find the variance and standard deviation of a binomial distribution • You may also apply the general formula to compute σ2 = [ Σ x2 • P(x) ] – μ2 • It turns out that we can quickly get the variance and standard deviation by using the following formula: The variance is σ2 = n p (1 – p) The standard deviation is s np(1 - p) Example • For our random guessing on a quiz problem – n = 10 – p = 0.2 – x=6 • Therefore – The mean is np = 10 • 0.2 = 2 – The variance is np(1-p) = 10 • .2 • .8 = 0.16 – The standard deviation is 0.16 = 0.4 Shape of Binomial Distribution • With the formula for the binomial probabilities P(x), we can construct histograms for the binomial distribution • There are three different shapes for these histograms – When p < .5, the histogram is skewed right – When p = .5, the histogram is symmetric – When p > .5, the histogram is skewed left Right-skewed Binomial Distribution • For n = 10 and p = .2 (skewed right) – Mean = 2 – Standard deviation = .4 Symmetrical Binomial Distribution • For n = 10 and p = .5 (symmetric) – Mean = 5 – Standard deviation = .5 Left-skewed Binomial Distribution • For n = 10 and p = .8 (skewed left) – Mean = 8 – Standard deviation = .4 Notes • Despite binomial distributions being skewed, the histograms appear more and more bell shaped as n gets larger • This will be important! Summary • Binomial random variables model a series of independent trials, each of which can be a success or a failure, each of which has the same probability of success • The binomial random variable has mean equal to np and variance equal to np(1-p)