Transforming and Combining Random Variables Section 6.2 Reference Text: The Practice of Statistics, Fourth Edition. Starnes, Yates, Moore Objectives 1. Multiplying or Dividing by a constant 2. Adding or Subtracting by a constant 3. Putting it Together: Adding, Subtracting, Multiplying, or Dividing Linear transformation! 4. Mean of the Sum of Random Variables 5. Independent Random Variables 6. Variance of the Sum of Independent Random Variables 7. Mean of the Difference of Random Variables 8. Variance of the Differences of Random Variables Multiplying or Dividing by a Constant Adding or Subtracting by a Constant • If we took our values of our distinct random variables and added (which could be negative) (or subtracted) them by a constant (a)…. • Adds (or subtracts) measures of center and location (mean, median, quartiles, percentiles) by a. • Does not change the measure of spread • Does not change the shape of the distribution • Shall we look at Pete’s Jeep Tours? Putting It All Together: Linear Transformation • What happens if we transform a random variable by both adding or subtracting a constant and multiplying or dividing by a constant? • We could have gone directly from the number of passengers X on Pete’s Jeep Tours to the profit of: V = 150X -100 where we both subtracted 100 and multiplied by 150. • This is a linear transformation! In general can be written in the form of Y constants. = a + bX, where a and b are • Lets generalize on the next slide…. Effects of Linear Transformation on the mean and SD Mean of the Sum of Random Variables Independent Random Variables • If knowing whether any event involving X alone has occurred tells us nothing about the occurrence of any event involving Y alone, and vice versa, then X and Y are independent random variables. • But we already knew this! Just restating the idea of being independent! Independent Random Variables • Probability models often assume independence when the random variables describe outcomes that appear unrelated to each other. You should always ask whether the assumption of independence seems reasonable. • For instance, its reasonable to treat the random variables X = number of passengers on Pete’s trip and Y = number of passengers on Erin’s trip on a randomly chosen day as independent, since the siblings operate their trips in different parts of the country. Variance of the Sum of Independent Random Variables By the Way… • You might be wondering whether there’s a formula for computing the variance of the sum of two random variables that are not independent. There is, but its beyond the scope of this course. • Just remember, you can add variances only if the two random variables are independent, and that you can never add standard deviations. Mean of the Difference of Random Variables Variance of the Differences of Random Variables • Earlier, we saw that the variance of the sum of two independent random variables is the sum of their variances. Can you guess what the variance of the difference of two independent random variables will be? WRONG! THINK AGAIN! MUHAHAHA! Variance of the Differences of Random Variables Objectives 1. Multiplying or Dividing by a constant 2. Adding or Subtracting by a constant 3. Putting it Together: Adding, Subtracting, Multiplying, or Dividing Linear transformation! 4. Mean of the sum of random variables 5. Independent random variables 6. Variance of the sum of independent variables 7. Mean difference of Random Variables 8. Variance of the differences of Random Variables Homework Worksheet : I'm going to post the homework online- however the deal is: If I don’t post the homework online by Tuesday 11/25/14 then there is no homework over break. Deal?!