MOMENT GENERATING FUNCTION AND STATISTICAL DISTRIBUTIONS 1 MOMENT GENERATING FUNCTION The m.g.f. of random variable X is defined as e txf ( x )dx if X is cont. all x tX M X ( t ) E (e ) e txf ( x ) if X is discret e all x for t Є (-h,h) for some h>0. 2 Properties of m.g.f. • M(0)=E[1]=1 • If a r.v. X has m.g.f. M(t), then Y=aX+b has a m.g.f. ebtM(at) E(Xk ) M(k ) (0) where M(k ) is the k th derivative. • • M.g.f does not always exists (e.g. Cauchy distribution) 3 Example • Suppose that X has the following p.d.f. x f ( x ) xe for x 0 Find the m.g.f; expectation and variance. 4 CHARACTERISTIC FUNCTION The c.h.f. of random variable X is defined as e itx f ( x) dx if X is cont. all x itX X (t ) E (e ) itx e f ( x) if X is discrete all x for all real numbers t. i 2 1, i 1 C.h.f. always exists. 5 Uniqueness Theorem: 1. If two r.v.s have mg.f.s that exist and are equal, then they have the same distribution. 2. If two r,v,s have the same distribution, then they have the same m.g.f. (if they exist) Similar statements are true for c.h.f. 6 STATISTICAL DISTRIBUTIONS 7 SOME DISCRETE PROBABILITY DISTRIBUTIONS Degenerate, Uniform, Bernoulli, Binomial, Poisson, Negative Binomial, Geometric, Hypergeometric 8 DEGENERATE DISTRIBUTION • An rv X is degenerate at point k if 1, X k P X x 0, o.w. The cdf: 0, X k F x P X x 1, X k 9 UNIFORM DISTRIBUTION • A finite number of equally spaced values are equally likely to be observed. 1 P(X x ) ; x 1,2,...,N; N 1,2,... N • Example: throw a fair die. P(X=1)=…=P(X=6)=1/6 N 1 E(X) ; 2 ( N 1)( N 1) Var (X) 12 10 BERNOULLI DISTRIBUTION • A Bernoulli trial is an experiment with only two outcomes. An r.v. X has Bernoulli(p) distribution if 1 with probability p X ;0 p 1 0 with probability 1 p P(X x) p x (1 p)1 x for x 0,1; and 0 p 1 11 BINOMIAL DISTRIBUTION • Define an rv Y by Y = total number of successes in n Bernoulli trials. 1. There are n trials (n is finite and fixed). 2. Each trial can result in a success or a failure. 3. The probability p of success is the same for all the trials. 4. All the trials of the experiment are independent. Y X ~ Bin n, p where X ~ Ber p . n i 1 i i ~ Bin n , p . Then, independent Let X i i X ~ Bin n n k i 1 i 1 2 n , p . k 12 BINOMIAL DISTRIBUTION • Example: • There are black and white balls in a box. Select and record the color of the ball. Put it back and re-pick (sampling with replacement). • n: number of independent and identical trials • p: probability of success (e.g. probability of picking a black ball) • X: number of successes in n trials 13 BINOMIAL THEOREM • For any real numbers x and y and integer n>0 n i n i ( x y) x y i 0 i n n 14 BINOMIAL DISTRIBUTION • If Y~Bin(n,p), then E(Y) np Var(Y) np(1- p) MY (t ) [pet (1 p)]n 15 POISSON DISTRIBUTION • The number of occurrences in a given time interval can be modeled by the Poisson distribution. • e.g. waiting for bus, waiting for customers to arrive in a bank. • Another application is in spatial distributions. • e.g. modeling the distribution of bomb hits in an area or the distribution of fish in a lake. 16 POISSON DISTRIBUTION • If X~ Poi(λ), then E(X)= Var(X)=λ M X (t ) exp{[exp(t ) 1]} 17 Relationship between Binomial and Poisson X ~ Bin n, p with mgf M t pe 1 p t n X Let =np. lim M n t lim pe 1 p t X n n e 1 lim 1 M t e n t n n et 1 Y The mgf of Poisson() The limiting distribution of Binomial rv is the 18 Poisson distribution. NEGATIVE BINOMIAL DISTRIBUTION (PASCAL OR WAITING TIME DISTRIBUTION) • X: number of Bernoulli trials required to get a fixed number of failures before the r th success; or, alternatively, • Y: number of Bernoulli trials required to get a fixed number of successes, such as r successes. 19 NEGATIVE BINOMIAL DISTRIBUTION (PASCAL OR WAITING TIME DISTRIBUTION) X~NB(r,p) r x 1 r p (1 p) x ; P(X x) x x 0,1,...;0 p 1 MX (t ) pr [1 (1 p)e t ]r r(1 p) E(X) p Var (X) r(1 p) p2 20 NEGATIVE BINOMIAL DISTRIBUTION • An alternative form of the pdf: y 1 r p (1 p) y r ; y r, r 1,...; 0 p 1 P(Y y) r 1 Note: Y=X+r r r(1 p) E(Y) E(X) r Var (Y) Var (X) p p2 21 GEOMETRIC DISTRIBUTION • Distribution of the number of Bernoulli trials required to get the first success. • It is the special case of the Negative Binomial Distribution r=1. X~Geometric(p) P X x p 1 p , x 1,2, x 1 1 (1 p) E ( X) Var (X) p p2 22 GEOMETRIC DISTRIBUTION • Example: If probability is 0.001 that a light bulb will fail on any given day, then what is the probability that it will last at least 30 days? • Solution: P(X 30) x 1 30 0 . 001 ( 1 0 . 001 ) ( 0 . 999 ) 0.97 x 31 23 HYPERGEOMETRIC DISTRIBUTION • A box contains N marbles. Of these, M are red. Suppose that n marbles are drawn randomly from the box without replacement. The distribution of the number of red marbles, x is M N M X~Hypergeometric(N,M,n) x n x , x 0,1,..., n P X x N n It is dealing with finite population. 24 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS Uniform, Normal, Exponential, Gamma, Chi-Square, Beta Distributions 25 Uniform Distribution – A random variable X is said to be uniformly distributed if its density function is 1 f ( x) a x b. ba – The expected value and the variance are 2 ab (b a ) E(X) V( X ) 2 12 26 Uniform Distribution • Example 1 – The daily sale of gasoline is uniformly distributed between 2,000 and 5,000 gallons. Find the probability that sales are: – Between 2,500 and 3,000 gallons – More than 4,000 gallons – Exactly 2,500 gallons f(x) = 1/(5000-2000) = 1/3000 for x: [2000,5000] P(2500X3000) = (3000-2500)(1/3000) = .1667 1/3000 2000 2500 3000 5000 x 27 Uniform Distribution • Example 1 – The daily sale of gasoline is uniformly distributed between 2,000 and 5,000 gallons. Find the probability that sales are: – Between 2,500 and 3,500 gallons – More than 4,000 gallons – Exactly 2,500 gallons f(x) = 1/(5000-2000) = 1/3000 for x: [2000,5000] P(X4000) = (5000-4000)(1/3000) = .333 1/3000 2000 4000 5000 x 28 Uniform Distribution • Example 1 – The daily sale of gasoline is uniformly distributed between 2,000 and 5,000 gallons. Find the probability that sales are: – Between 2,500 and 3,500 gallons – More than 4,000 gallons – Exactly 2,500 gallons f(x) = 1/(5000-2000) = 1/3000 for x: [2000,5000] P(X=2500) = (2500-2500)(1/3000) = 0 1/3000 2000 2500 5000 x 29 Normal Distribution • This is the most popular continuous distribution. – Many distributions can be approximated by a normal distribution. – The normal distribution is the cornerstone distribution of statistical inference. 30 Normal Distribution • A random variable X with mean m and variance s2 is normally distributed if its probability density function is given by 2 x m (1 / 2) s e 1 f (x) x ; s 0 s 2 where 3.14159... and e 2.71828... 31 The Shape of the Normal Distribution The normal distribution is bell shaped, and symmetrical around m. 90 m Why symmetrical? Let m = 100. Suppose x = 110. f (110) 1 s 2 110100 (1/ 2) s e 2 1 s 2 10 (1/ 2) s e 110 Now suppose x = 90 2 f (90) 1 s 2 90100 (1 / 2) s e 2 1 s 2 10 (1 / 2) s e 2 The Effects of m and s How does the standard deviation affect the shape of f(x)? s= 2 s =3 s =4 How does the expected value affect the location of f(x)? m = 10 m = 11 m = 12 33 Finding Normal Probabilities • Two facts help calculate normal probabilities: – The normal distribution is symmetrical. – Any normal distribution can be transformed into a specific normal distribution called… “STANDARD NORMAL DISTRIBUTION” Example The amount of time it takes to assemble a computer is normally distributed, with a mean of 50 minutes and a standard deviation of 10 minutes. What is the probability that a computer is assembled in a time between 45 and 60 minutes? 34 STANDARD NORMAL DISTRIBUTION • NORMAL DISTRIBUTION WITH MEAN 0 AND VARIANCE 1. • IF X~N(m , s2), THEN Z X m ~ N (0,1) s NOTE: Z IS KNOWN AS Z SCORES. • “ ~ “ MEANS “DISTRIBUTED AS” 35 Finding Normal Probabilities • Solution – If X denotes the assembly time of a computer, we seek the probability P(45<X<60). – This probability can be calculated by creating a new normal variable the standard normal variable. Every normal variable with some m and s, can be transformed into this Z. X mx Z sx E(Z) = m = 0 Therefore, once probabilities for Z are calculated, probabilities of any normal variable can be found. V(Z) = s2 = 1 36 Standard normal probabilities Copied from Walck, C (2007) Handbook on Statistical Distributions for experimentalists 37 Standard normal table 1 38 Standard normal table 2 39 Standard normal table 3 40 Finding Normal Probabilities • Example - continued 45 - 50 X m 60 - 50 P(45<X<60) = P( < < ) s 10 10 = P(-0.5 < Z < 1) To complete the calculation we need to compute the probability under the standard normal distribution 41 Using the Standard Normal Table Standard normal probabilities have been calculated and are provided in a table . The tabulated probabilities correspond to the area between Z=0 and some Z = z0 >0 z 0.0 0.1 . . 1.0 . . 1.2 . . 0 0.0000 0.0398 . . 0.3413 . . 0.3849 . . 0.01 0.0040 0.0438 . . 0.3438 . . 0.3869 . . ……. ……. . . 0.05 0.0199 0.0596 . . 0.3531 . . 0.3944 . . P(0<Z<z0) Z=0 0.06 0.0239 0.0636 . . 0.3554 . . 0.3962 . . Z = z0 42 Finding Normal Probabilities • Example - continued 45 - 50 X m 60 - 50 P(45<X<60) = P( < < ) s 10 10 = P(-.5 < Z < 1) We need to find the shaded area z0 = -.5 z0 = 1 43 Finding Normal Probabilities • Example - continued 45 - 50 X m 60 - 50 P(45<X<60) = P( < < ) s 10 10 = P(-.5<Z<1) = P(-.5<Z<0)+ P(0<Z<1 P(0<Z<1) z 0.0 0.1 . . 1.0 . 0 0.0000 0.0398 . . 0.3413 . 0.1 0.0040 0.0438 . . 0.3438 . ……. 0.05 0.0199 0.0596 . .3413 . 0.3531 . z0 =-.5z=0 z0 = 1 0.06 0.0239 0.636 . . 0.3554 . 44 Finding Normal Probabilities • The symmetry of the normal distribution makes it possible to calculate probabilities for negative values of Z using the table as follows: -z0 0 +z0 P(-z0<Z<0) = P(0<Z<z0) 45 Finding Normal Probabilities • Example - continued 0 0.0000 0.0398 . . 0.1915 . 0.1 0.0040 0.0438 . . …. . ……. .3413 .1915 z 0.0 0.1 . . 0.5 . -.5 0.05 0.0199 0.0596 . . …. . 0.06 0.0239 0.636 . . …. . .5 46 Finding Normal Probabilities • Example - continued 0 0.0000 0.0398 . . 0.1915 . 0.1 0.0040 0.0438 . . …. . ……. .3413 .1915 .1915 .1915 z 0.0 0.1 . . 0.5 . -.5 0.05 0.0199 0.0596 . . …. . 0.06 0.0239 0.636 . . …. . .5 1.0 P(-.5<Z<1) = P(-.5<Z<0)+ P(0<Z<1) = .1915 + .3413 = .5328 47 Finding Normal Probabilities • Example - continued 0 0.5000 0.5398 . . 0.6915 . 0.1 0.5040 0.5438 . . …. . ……. .3413 z 0.0 0.1 . . 0.5 . 0.05 0.5199 0.5596 . . …. . 0.06 0.5239 0.5636 . . …. . P(Z<-0.5)=1-P(Z>-0.5)=1-0.6915=0.3085 By Symmetry P(Z<0.5) 48 Finding Normal Probabilities • Example – The rate of return (X) on an investment is normally distributed with a mean of 10% and standard deviation of (i) 5%, (ii) 10%. – What is the probability of losing money? 0% 10% 0 - 10 (i) P(X< 0 ) = P(Z< ) = P(Z< - 2) 5 X .4772 -2 =P(Z>2) = 0.5 - P(0<Z<2) = 0.5 - .4772 = .0228 0 2 Z 49 Finding Normal Probabilities • Example – The rate of return (X) on an investment is normally distributed with mean of 10% and standard deviation of (i) 5%, (ii) 10%. – What is the probability of losing money? X 0% 10% 0 - 10 (ii) P(X< 0 ) = P(Z< ) 10 .3413 -1 = P(Z< - 1) = P(Z>1) = 0.5 - P(0<Z<1) = 0.5 - .3413 = .1587 1 Z 50 AREAS UNDER THE STANDARD NORMAL DENSITY P(0<Z<1)=.3413 0 1 Z 51 AREAS UNDER THE STANDARD NORMAL DENSITY .3413 .4772 P(1<Z<2)=.4772-.3413=.1359 52 EXAMPLES • P( Z < 0.94 ) = 0.5 + P( 0 < Z < 0.94 ) = 0.5 + 0.3264 = 0.8264 0.8264 0 0.94 53 EXAMPLES • P( Z > 1.76 ) = 0.5 – P( 0 < Z < 1.76 ) = 0.5 – 0.4608 = 0.0392 0.0392 0 1.76 54 EXAMPLES • P( -1.56 < Z < 2.13 ) = = P( -1.56 < Z < 0 ) + P( 0 < Z < 2.13 ) Because of symmetry P(0 < Z < 1.56) = 0.4406 + 0.4834 = 0.9240 0.9240 -1.56 2.13 55 STANDARDIZATION FORMULA • If X~N(m , s2), then the standardized value Z of any ‘X-score’ associated with calculating probabilities for the X distribution is: Z X m s • The standardized value Z of any ‘X-score’ associated with calculating probabilities for the X distribution is: x m z.s (Converse Formula) 56 Finding Values of Z • Sometimes we need to find the value of Z for a given probability • We use the notation zA to express a Z value for which P(Z > zA) = A A zA 57 PERCENTILE • The pth percentile of a set of measurements is the value for which at most p% of the measurements are less than that value. • 80th percentile means P( Z < a ) = 0.80 • If Z ~ N(0,1) and A is any probability, then P( Z > zA) = A A zA 58 Finding Values of Z • Example – Determine z exceeded by 5% of the population – Determine z such that 5% of the population is below • Solution z.05 is defined as the z value for which the area on its right under the standard normal curve is .05. 0.45 0.05 0.05 -Z0.05 0 Z0.05 1.645 59 EXAMPLES • Let X be rate of return on a proposed investment. Mean is 0.30 and standard deviation is 0.1. a) P(X>.55)=? Standardization formula b) P(X<.22)=? c) P(.25<X<.35)=? d) 80th Percentile of X is? Converse Formula e) 30th Percentile of X is? 60 a)P(X 0.55) P b)P(X 0.22) P ANSWERS X - 0.3 0.55 - 0.3 =Z> = 2.5 = 0.5 - 0.4938 = 0.0062 0.1 0.1 X - 0.3 0.22 - 0.3 =Z = 0.8 = 0.5 - 0.2881 = 0.2119 0.1 0.1 c) P(0.25 X 0.35) P 0.25 0.3 0.5 X - 0.3 = Z 0.35 - 0.3 = 0.5 = 2.* (0.1915) 0.3830 0.1 0.1 0.1 d) 80th Percentile of X is x m s .z0.20 .3+(.85)*(.1)=.385 e) 30th Percentile of X is x m s .z0.70 .3+(-.53)*(.1)=.247 61 The Normal Approximation to the Binomial Distribution • The normal distribution provides a close approximation to the Binomial distribution when n (number of trials) is large and p (success probability) is close to 0.5. • The approximation is used only when np 5 and n(1-p) 5 62 The Normal Approximation to the Binomial Distribution • If the assumptions are satisfied, the Binomial random variable X can be approximated by normal distribution with mean m = np and s2 = np(1-p). • In probability calculations, the continuity correction improves the results. For example, if X is Binomial random variable, then P(X a) ≈ P(X<a+0.5) P(X a) ≈ P(X>a-0.5) 63 EXAMPLE • Let X ~ Binomial(25,0.6), and want to find P(X ≤ 13). • Exact binomial calculation: P(X 13) 13 25 x 25 x ( 0 . 6 ) ( 0 . 4 ) 0.267 x x 0 • Normal approximation (w/o correction): Y~N(15,2.45²) 13 15 P(X 13) P(Y 13) P( Z ) P( Z 0.82) 0.206 2.45 Normal approximation is good, but not great! 64 EXAMPLE, cont. Bars – Bin(25,0.6); line – N(15, 2.45²) Copied from Casella & Berger (1990) 65 EXAMPLE, cont. • Normal approximation (w correction): Y~N(15,2.45²) P(X 13) P(X 13.5) P(Y 13.5) P(Z 0.61) 0.271 Much better approximation to the exact value: 0.267 66 Exponential Distribution • The exponential distribution can be used to model – the length of time between telephone calls – the length of time between arrivals at a service station – the lifetime of electronic components. • When the number of occurrences of an event follows the Poisson distribution, the time between occurrences follows the exponential distribution. 67 Exponential Distribution A random variable is exponentially distributed if its probability density function is given by f 1 X x e x/ , x 0, 0 = e-x, f(x) x>=0. is a distribution parameter. is the distribution parameter >0). E(X) = V(X) = 2 The cumulative distribution function is F(x) =1e-x/, x0 68 Exponential distribution for 1 = .5, 1, 2 2.5 f(x) = 2e-2x 2 f(x) = 1e-1x 1.5 f(x) = .5e-.5x 1 0.5 0 0 1 2 3 4 5 2.5 2 1.5 1 P(a<X<b) = e-a/ e-b/ 0.5 0 a b 69 Exponential Distribution • Finding exponential probabilities is relatively easy: – P(X < a) = P(X ≤ a)=F(a)=1 – e –a/ – P(X > a) = e–a/ – P(a< X < b) = e – a/ – e – b/ 70 Exponential Distribution • Example The service rate at a supermarket checkout is 6 customers per hour. – If the service time is exponential, find the following probabilities: • A service is completed in 5 minutes, • A customer leaves the counter more than 10 minutes after arriving • A service is completed between 5 and 8 minutes. 71 Exponential Distribution • Solution – A service rate of 6 per hour = A service rate of .1 per minute (1 = .1/minute). – P(X < 5) = 1-e-.lx = 1 – e-.1(5) = .3935 – P(X >10) = e-.lx = e-.1(10) = .3679 – P(5 < X < 8) = e-.1(5) – e-.1(8) = .1572 72 Exponential Distribution • If pdf of lifetime of fluorescent lamp is exponential with mean 0.10, find the life for 95% reliability? The reliability function = R(t) = 1F(t) = e-t/ R(t) = 0.95 t = ? What is the mean time to failure? 73 GAMMA DISTRIBUTION • X~ Gamma(,) 1 f x x e 1 x/ , x 0, 0, 0 E X and Var X M t 1 t , t 2 1 74 GAMMA DISTRIBUTION • Gamma Function: x e dx 1 x 0 where is a positive integer. Properties: 1 , 0 n n 1! for any integer n 1 1 2 75 GAMMA DISTRIBUTION • Let X1,X2,…,Xn be independent rvs with Xi~Gamma(i, ). Then, X ~ Gamma , n i 1 i n i 1 i •Let X be an rv with X~Gamma(, ). Then, cX ~ Gamma , c where c is positive constant. • Let X1,X2,…,Xn be a random sample with Xi~Gamma(, ). Then, n X X ~ Gamma n , n n i 1 i 76 GAMMA DISTRIBUTION • Special cases: Suppose X~Gamma(α,β) – If α=1, then X~ Exponential(β) – If α=p/2, β=2, then X~ 2 (p) (will come back in a min.) – If Y=1/X, then Y ~ inverted gamma. 77 Integral tricks • Recall: h(t) is an odd function if h(-t)=-h(t); It is an even function if h(-t)=h(t). • Ex: h(x)=x exp{-x²/2} odd; h(x)= exp{-x²/2-x} neither odd nor even • odd function =0 • even function = 2* 0 even function 78 CHI-SQUARE DISTRIBUTION Chi-square with degrees of freedom • X~ 2()= Gamma(/2,2) 1 / 2 1 x / 2 f ( x) / 2 x e , x 0, 1,2,... 2 ( / 2) E X and Var X 2 M(t ) (1 2t ) p / 2 t 1/ 2 79 DEGREES OF FREEDOM • In statistics, the phrase degrees of freedom is used to describe the number of values in the final calculation of a statistic that are free to vary. • The number of independent pieces of information that go into the estimate of a parameter is called the degrees of freedom (df) . • How many components need to be known before the vector is fully determined? 80 CHI-SQUARE DISTRIBUTION • If rv X has Gamma(,) distribution, then Y=2X/ has Gamma(,2) distribution. If 2 2 is positive integer, then Y has 2 distribution. •Let X be an rv with X~N(0, 1). Then, X ~ 2 2 1 •Let X1,X2,…,Xn be a r.s. with Xi~N(0,1). Then, X ~ n i 1 2 2 i n 81 BETA DISTRIBUTION • The Beta family of distributions is a continuous family on (0,1) and often used to model proportions. 1 f x x 1 x , 0 x 1, 0, 0. B , where B , x 1 x dx 1 1 1 1 0 EX and Var X 1 82 2 CAUCHY DISTRIBUTION • It is a symmetric and bell-shaped distribution on (,) with pdf f (x) 1 s 1 ( 1 x s )2 ,s 0 Since E X , the mean does not exist. • The mgf does not exist. • measures the center of the distribution and it is the median. • If X and Y have N(0,1) distribution, then Z=X/Y has a Cauchy distribution with =0 and σ=1. 83 LOG-NORMAL DISTRIBUTION • An rv X is said to have the lognormal distribution, with parameters µ and s2, if Y=ln(X) has the N(µ, s2) distribution. •The lognormal distribution is used to model continuous random quantities when the distribution is believed to be skewed, such as certain income and lifetime variables. 84 STUDENT’S T DISTRIBUTION • This distribution will arise in the study of population mean when the underlying distribution is normal. • Let Z be a standard normal rv and let U be a chi-square distributed rv independent of Z, with degrees of freedom. Then, Z X ~ t U / When n, XN(0,1). 85 F DISTRIBUTION • Let U and V be independent rvs with chisquare distributions with 1 and 2 degrees of freedom. Then, U / X ~ F V / 1 1, 2 2 86 MULTIVARIATE DISTRIBUTIONS 87 EXTENDED HYPERGEOMETRIC DISTRIBUTION • Suppose that a collection consists of a finite number of items, N and that there are k+1 different types; M1 of type 1, M2 of type 2, and so on. Select n items at random without replacement, and let Xi be the number of items of type i that are selected. The vector X=(X1, X2,…,Xk) has an extended hypergeometric distribution and the joint pdf is M 1 M 2 M k M k 1 ... x x x x f x1 , x2 ,..., xk 1 2 k k 1 , xi {0,1,...,M i } N n k k i 1 i 1 where M k 1 N M i and x k 1 n xi . 88 MULTINOMIAL DISTRIBUTION • Let E1,E2,...,Ek,Ek+1 be k+1 mutually exclusive and exhaustive events which can occur on any trial of an experiment with P(Ei)=pi,i=1,2,…,k+1. On n independent trials of the experiment, let Xi be the number of occurrences of the event Ei. Then, the vector X=(X1, X2,…,Xk) has a multinomial distribution with joint pdf n! f x1 , x2 ,..., xk p1x1 p2x2 ...pkxk 11 , xi {0,1,...,n} x1! x2 !...xk 1! k k i 1 i 1 where x k 1 n xi and p k 1 1 pi . 89 BIVARIATE NORMAL DISTRIBUTION • A pair of continuous rvs X and Y is said to have a bivariate normal distribution if it has a joint pdf of the form 1 f x, y exp 2 2 1 2s1s 2 1 2 1 2 2 y m y m x m x m y y x 2 x s1 s1 s 2 s 2 x , y , s1 0, s 2 0, 1 1. 90 If BIVARIATE NORMAL DISTRIBUTION 2 2 X ,Y ~ BVN mx ,m y ,s1 ,s 2 , , then 2 2 X ~ N mx ,s1 and Y ~ N m y ,s 2 and is the correlation coefficient btw X and Y. 1. Conditional on X=x, s2 Y x ~ N m y (x mX ),s 22 1 2 s1 2. Conditional on Y=y, s1 X y ~ N m x ( y mY ),s12 1 2 s2 91