Formula Sheet for CPIT 701 (Advanced Probability and Statistics) PROBABILITY Probability of any event: 0 P (event) 1 Permutation: subsets of r elements from n different elements n! Prn = n (n − 1) (n − 2) (n − r + 1) = (n − r )! Permutation of similar objects: n1 is of one type, n2 is of second type, … among n = n1+n2+…+nr elements n! n1! n 2 ! n3 ! n r ! Combinations: subsets of size r from a set of n elements n n! nCr = Crn = = r r!(n − r )! Independent Events: P(AB) = P(A or B) = P(A) + P(B) – P(A and B) P(ABC) = P(A) + P(B) + P(C) – P(AB) – P(AC) - P(AB) = P(A and B) = P(A)P(B) P(A|B) = P(A) P(BC) + P(ABC) P(B) = P(BA) + P(BA’) = P(B|A)P(A)+P(B|A’)P(A’) P(B|A)=P(B) Dependent Events: For Mutually exclusive events AB=: P(A and B) = P(A) * P(B given A) P(AB) = P(A or B) = P(A) + P(B) P(AB)=P(A|B)P(B)=P(BA)=P(B|A)P(A) P(A and B and C) = P(A) * P(B | A) * P(C given A and B) Bayes’ Theorem P(AB) = P(AB) = P(A | B) P(B) = P(B | A) P(A) P(B | A) P( A) P( AB ) Conditional Probabilit y P( A | B) = P( A | B) = P(B) P(B | A) P( A) + P(B | A) P( A) A, B = any two events A’ = complement of A Markov’s Inequality Chebyshev’s Inequality If X is a non-negative random variable with mean , then If X is a random variable with a finite mean for any constant a > 0 and variance 2, then for any constant a > 0 P(X a) Mean of Central Tendency P( X - a) a 𝑁 1 𝑀𝑒𝑎𝑛 = ∑ 𝑥𝑖 𝑁 2 a2 𝑁 1 𝑀𝑒𝑎𝑛 𝑎𝑏𝑠𝑜𝑙𝑢𝑡𝑒 𝑑𝑒𝑣𝑖𝑎𝑡𝑖𝑜𝑛 = ∑|𝑥𝑖 − 𝑥̅ | 𝑁 𝑖=1 𝑖=1 Sample Variance (mean squared deviation from the mean) 𝑁 𝑠2 = 1 ∑(𝑥𝑖 − 𝑥̅ )2 𝑁−1 𝑖=1 𝐷𝑖𝑠𝑗𝑜𝑖𝑛𝑡 𝐴 ∩ 𝐵 = ∅ 𝑆𝑢𝑏𝑠𝑒𝑡 𝐴 ⊃ 𝐵 𝑖𝑓 𝐴 ∩ 𝐵 = 𝐴 𝑃𝑎𝑟𝑡𝑖𝑡𝑖𝑜𝑛 𝐴 ∪ 𝐵 ∪ 𝐶 ∪ 𝐷 = Ω 𝐶𝑜𝑚𝑝𝑙𝑖𝑚𝑒𝑛𝑡 𝐴𝑐 = 𝐴̅ = {𝑥 ∈ Ω | 𝑥 ∉ 𝐴} DeMorgan’s Laws Set Basics Intersection 𝐴 ∩ 𝐵 = {𝑥|𝑥 𝜖 𝐴 𝑎𝑛𝑑 𝑥 ∈ 𝐵} Union 𝐴 ∪ 𝐵 = {𝑥|𝑥 𝜖 𝐴 𝑜𝑟 𝑥 ∈ 𝐵} Useful Identities Probabilistic Law Ω𝑐 = ∅ 𝐴 ∪ A𝑐 = Ω (A𝒄 )𝒄 = 𝐴 𝐴 ∩ A𝑐 = ∅ 𝑛 𝑐 𝑛 𝑃(𝐸) = 𝑐 (⋃ 𝐸𝑖 ) = ⋂ E𝑖 𝑖=1 𝑛 𝑎𝑙𝑙 𝑜𝑢𝑡𝑐𝑜𝑚𝑒𝑠 𝑖𝑛 𝐸 𝑛 𝑖=1 𝑛 𝑐 ∑ p𝑖 𝑃 (⋃ 𝐸𝑖 ) = P(𝐸1 ) + P(𝐸2 ) + ⋯ + P(𝐸𝑛 ) (⋂ 𝐸𝑖 ) = ⋃ E𝑐𝑖 𝑖=1 𝑖=1 𝑖=1 Probability Identities Discrete Uniform Probability 𝑃(∅) = 0 𝑃(𝐸 𝑐 ) = 1 − P(E) 𝑃(𝐸) ≤ 1 𝑃(𝐸) = |𝐸| # 𝑒𝑙𝑒𝑚𝑒𝑛𝑡𝑠 𝑖𝑛 𝐸 = |Ω| 𝑁 If 𝐸 ⊂ 𝐹 𝑡ℎ𝑒𝑛 𝑃(𝐸) ≤ 𝑃(𝐹) 𝑃(𝐸 ∪ 𝐹) = P(E) + 𝑃(𝐹) − 𝑃(𝐸 ∩ 𝐹) Stage Counting Method Product Rule = 𝑛1 . 𝑛2 … 𝑛𝑟 For n sets Total number of possible subsets = 2𝑛 Counting without replacement (ordering without repitition)Permutation = n! K-Permutations Combinatorics Draw k from a set of n with replacement Order matters = 𝑛𝑘 Conditional Probability 𝑛𝑘 Order doesn’t matter = 𝑛! Draw k from a set of n without replacement 𝑛! Order matters = (𝑛−𝑘)! 𝑛 𝑛! = 𝑛𝑃𝑘 (𝑛 − 𝑘)! Combinations: Order does not matter Selecting k out of n = 𝑛𝑃𝑘 𝑛! 𝑛 ( )= = 𝑘 𝑘! 𝑘! (𝑛 − 𝑘)! 𝑃(𝐴 ∩ 𝐵) 𝑃(𝐵) 𝐵 = (𝐴1 ∩ 𝐵) ∪ (𝐴2 ∩ 𝐵) ∪ … (𝐴𝑛 ∩ 𝐵) 𝑃(𝐵) = 𝑃(𝐴1 ∩ 𝐵) + 𝑃(𝐴2 ∩ 𝐵) + ⋯ + 𝑃(𝐴𝑛 ∩ 𝐵) 𝑃(𝐴|𝐵) = = ∑ 𝑃(𝐴𝑖 ∩ 𝐵) 𝑃(𝐴𝑖 ∩ 𝐵) = 𝑃(𝐵|𝐴𝑖 )𝑃(𝐴𝑖 ) 𝑛! Order doesn’t matter = (𝑘 ) = 𝑘!(𝑛−𝑘)! 𝑃(𝐵) = ∑ 𝑃(𝐵|𝐴𝑖 )𝑃(𝐴𝑖 ) 𝑃(𝐵) = 𝑃(𝐵|𝐴)𝑃(𝐴) + 𝑃(𝐵|𝐴𝑐 )𝑃(𝐴𝑐 ) 𝑃(𝐵|𝐴)𝑃(𝐴) 𝑃(𝐴|𝐵) = 𝑃(𝐵) 𝑃(𝐵|𝐴)𝑃(𝐴) 𝑃(𝐴|𝐵) = 𝑃(𝐵|𝐴)𝑃(𝐴) + 𝑃(𝐵|𝐴𝑐 )𝑃(𝐴𝑐 ) Independent events 𝑃(𝐴|𝐵) = 𝑃(𝐴) 𝑃(𝐴 ∩ 𝐵) = 𝑃(𝐴)𝑃(𝐵) Discrete Random Variable Probability Mass Function (PMF) Discrete Random Variable Expectation Value 𝑃𝑋 (𝑥) = 𝑃(𝑋 = 𝑥) 𝐸[𝑋] = ∑ 𝑥 𝑝𝑋 (𝑥) ∑ 𝑝𝑋 (𝑘) = 1 𝑎𝑙𝑙 𝑥 Variance 𝑎𝑙𝑙 𝑘 𝑃(𝑎 ≤ 𝑋 ≤ 𝑏) = ∑ 𝑝𝑋 (𝑥) 𝑎𝑙𝑙 𝑎≤𝑋≤𝑏 Function of Random Variables 𝑝𝑋 (𝑥) 𝑎𝑛𝑑 𝑦 = 𝑔(𝑋) 𝑝𝑌 (𝑦) = ∑ 𝑥|𝑔(𝑥)=𝑦 𝑝𝑋 (𝑥) 𝑣𝑎𝑟[𝑋] = ∑ (𝑥 − 𝐸[𝑋])2 𝑝𝑋 (𝑥) 𝑎𝑙𝑙 𝑥 Bernoulli Random Variable 𝑝 𝑖𝑓 𝑘 = 1 𝑃𝑋 (𝑘) = { 1 − 𝑝 𝑖𝑓 𝑘 = 0 𝐸[𝑋] = 𝑝 2 𝑣𝑎𝑟[𝑋] = 𝑝(1 − 𝑝) Cumulative Distribution Function 𝐹𝑋 (𝑥) = 𝑃(𝑋 ≤ 𝑥) = ∑ 𝑝𝑋 (𝑘) 𝑎𝑙𝑙 𝑘≤𝑥 Binomial Distribution Geometric Distribution 𝑃𝑋 (𝑥) = (1 − 𝑝)𝑘−1 𝑝 𝐸[𝑋] = 1/𝑝 𝑣𝑎𝑟[𝑋] = (1 − 𝑝)/𝑝2 Poisson Distribution Uniform Distribution 𝑛 𝑃𝑋 (𝑥) = ( ) 𝑝𝑘 (1 − 𝑝)𝑛−𝑘 𝑘 𝐸[𝑋] = 𝑛𝑝 𝑣𝑎𝑟[𝑋] = 𝑛𝑝(1 − 𝑝) 𝑊ℎ𝑒𝑛 𝑛 ≫ 𝑘, 𝑐𝑜𝑛𝑠𝑖𝑑𝑒𝑟 𝐵𝑖𝑛𝑜𝑚𝑖𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑛𝑝 = 𝜆 = 𝑐𝑜𝑛𝑠𝑡𝑎𝑛𝑡 𝜆𝑘 𝑛 𝑃𝑋 (𝑥) = lim ( ) 𝑝𝑘 (1 − 𝑝)𝑛−𝑘 = 𝑒 −𝜆 𝑛→∞ 𝑘 𝑘! 𝐸[𝑋] = 𝜆 𝑣𝑎𝑟[𝑋] = 𝜆 Independence For all cases 𝑝𝑋 (𝑥) = 1/𝑁 𝑁 𝐸[𝑋] = ∑ 𝑥 𝑝𝑋 (𝑥) = ∑ 𝑎𝑙𝑙 𝑥 𝑣𝑎𝑟[𝑋] = ∑ (𝑥 − 𝑥𝑖 𝑁 𝑖=1 2 𝐸[𝑋]) 𝑝𝑋 (𝑥) 𝑎𝑙𝑙 𝑥 Continuous Random Variable 𝑏 𝑃(𝑎 ≤ 𝑋 ≤ 𝑏) = ∫ 𝑓𝑋 (𝑥)𝑑𝑥 𝐸[𝑋 + 𝑌] = 𝐸[𝑋] + 𝐸[𝑌] 𝐸[𝑎𝑋 + 𝑏] = 𝑎𝐸[𝑋] + 𝑏 𝑣𝑎𝑟[𝑎𝑋 + 𝑏] = 𝑎2 𝑣𝑎𝑟[𝑋] 𝑎 Probability Density Function (PDF) If X and Y are independent random variables 𝐸[𝑋𝑌] = 𝐸[𝑋]𝐸[𝑌] 𝑣𝑎𝑟[𝑋 + 𝑌] = 𝑣𝑎𝑟[𝑋] + 𝑣𝑎𝑟[𝑌] 𝑃(𝑥 ≤ 𝑋 ≤ 𝑥 + 𝛿) 𝛿⟶0 𝛿 Non-negativity 𝑓𝑋 (𝑥) ≥ 0 Normalization 𝑓𝑋 (𝑥) = lim ∞ ∫ 𝑓𝑋 (𝑥)𝑑𝑥 = 1 −∞ Probability at a point 𝑃(𝑋 = 𝑎) = 0 Continuous Random Variable Expectation Value Cumulative Distribution Function (CDF) ∞ 𝐹𝑋 (𝑏) = 𝑃(𝑋 ≤ 𝑏) = ∫ 𝑓𝑋 (𝑥)𝑑𝑥 ∞ −∞ 𝐸[𝑋] = ∫ 𝑥𝑓𝑋 (𝑥)𝑑𝑥 CDF is monotonically non-decreasing −∞ Variance ∞ 𝑣𝑎𝑟[𝑋] = ∫ (𝑥 − 𝐸[𝑋])2 𝑓𝑋 (𝑥)𝑑𝑥 −∞ Normal Distribution 𝑓𝑋 (𝑥) = 1 (𝑥−𝜇)2 − 𝑒 2𝜎2 Standard Normal Distribution 2 = 𝑁(𝜇, 𝜎 ) √2𝜋𝜎 1 𝑥−𝜇 𝐹𝑋 (𝑥) = [1 + 𝑒𝑟𝑓 ( )] 2 √2𝜎 𝐸[𝑋] = 𝜇 𝑣𝑎𝑟[𝑋] = 𝜎 2 Sample Mean 1 𝑋̅ = ∑ 𝑋𝑖 𝑛 𝑖=1,𝑛 Expectation of the sample mean 1 𝐸(𝑋̅) = ∑ 𝑋𝑖 = 𝐸(𝑋𝑖 ) = 𝜇 𝑛 𝑖=1,𝑛 Variance of the sample mean 𝑃(𝑎 ≤ 𝑋 ≤ 𝑏) = 𝐹𝑋 (𝑏) − 𝐹𝑋 (𝑎) 𝑑𝐹𝑋 (𝑥) 𝑓𝑋 (𝑥) = 𝑑𝑥 𝑥−𝜇 𝜎 𝐼𝑓 𝑋 = 𝑁(𝜇, 𝜎 2 ) 𝑡ℎ𝑒𝑛 𝑍 = 𝑁(0,1) 𝑧= Linear function of a normally distributed RV produces a normally distributed RV 𝐿𝑒𝑡 𝑌 = 𝛼𝑋 + 𝛽 𝐼𝑓 𝑋 = 𝑁(𝜇, 𝜎 2 ) 𝑡ℎ𝑒𝑛 𝑌 = 𝑁(𝛼𝜇 + 𝛽, (𝛼𝜎)2 ) RV with a distribution that approaches N(0,1) as 𝑛 ⟶ ∞ 𝑍= 𝑋̅ − 𝜇 𝜎/√𝑛 Student’s t Distribution (t is RV with DF = n-1) 𝑡= 𝑋̅ − 𝜇 𝑠/√𝑛 3 𝑣𝑎𝑟(𝑋̅) = 1 𝜎2 𝑣𝑎𝑟 ( ∑ 𝑋𝑖 ) = 𝑛2 𝑛 𝑖=1,𝑛 Standard deviation of the sample mean 𝜎𝑋̅ = 𝜎 √𝑛 Sampling distribution of the Variance (for sample variance S2), RV with a chi-square distribution and DF = n -1 (𝑛 − 1)𝑆 2 𝜎2 2 𝐸[𝜒 ] = 𝑛 − 1 𝑣𝑎𝑟[𝜒 2 ] = 2(𝑛 − 1) Sampling distribution of the Variance for iid random variables with 𝑿𝒊 ~𝑁(𝜇, 𝜎2 ) 𝜒2 = Signal to noise, 𝐸[𝑆 2 ] = 𝜎 2 2𝜎 4 𝑣𝑎𝑟[𝑆 2 ] = 𝑛−1 𝐸[𝑆 2 ] √𝑣𝑎𝑟[𝑆 2 ] 𝑛−1 =√ 2 Confidence interval for the variance (𝑛 − 1)𝑠 2 2 𝜒𝛼/2 < 𝜎2 < (𝑛 − 1)𝑠 2 2 𝜒1−𝛼/2 Comparing two sample variances, RV that follows F-distribution with n1-1 and n2-1 degrees of freedom 𝐹= Confidence Interval of the mean 𝑥̅ − 𝑧𝛼⁄2 𝑠 √𝑛 𝑠 < 𝜇 < 𝑥̅ + 𝑧𝛼⁄2 √𝑛 𝑆12 /𝜎12 𝑆22 /𝜎22 ̂𝑞̂ 𝑝 Standard error 𝑆𝐸(𝑝̂ ) = √ 𝑛 𝑋 𝑛 𝑤ℎ𝑒𝑟𝑒 𝑝 𝑖𝑠 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑠𝑢𝑐𝑒𝑐𝑠𝑠 𝑎𝑛𝑑 𝑞 𝑖𝑠 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑓𝑎𝑖𝑙𝑢𝑟𝑒 Margin of error = 𝑧𝛼⁄2 𝑆𝐸(𝑝̂ ) 95% CI (𝛼 = 0.05) 𝑝̂ = 𝑝̂ 𝑞̂ 𝑝̂ 𝑞̂ (𝑝̂ − 1.96√ , 𝑝̂ + 1.96√ ) 𝑛 𝑛 Confidence Interval of the mean 𝑥̅ − 𝑡𝛼⁄2 𝑠 √𝑛 < 𝜇 < 𝑥̅ + 𝑡𝛼⁄2 𝑠 √𝑛 For large sample, 95% confidence interval (𝑥̅ − 1.96 𝑠 √𝑛 , 𝑥̅ + 1.96 𝑠 √𝑛 ) Pooled Samples (𝑛1 − 1)𝑠12 + (𝑛2 − 1)𝑠22 𝑠𝑝2 = (𝑛1 − 1) + (𝑛2 − 1) For t-tests, DF = 𝑛1 + 𝑛2 − 2 ̅̅̅1 − ̅̅̅ 𝑣𝑎𝑟[𝑋 𝑋2 ] = √ Correlation coefficient 𝑆12 𝑛1 + Two independent Sample means ̅̅̅1 − ̅̅̅ 𝐸[𝑋 𝑋2 ] = 𝜇1 − 𝜇2 𝜎12 𝜎22 ̅̅̅ ̅̅̅ ] 𝑣𝑎𝑟[𝑋1 − 𝑋2 = + 𝑛1 𝑛2 ̅̅̅ ̅̅̅ 𝑋1 − 𝑋2 − (𝜇1 − 𝜇2 ) 𝑍= ~𝑁(0,1) 2 2 𝑆 𝑆 √ 1+ 2 𝑛1 𝑛2 Two Matched Samples 𝐷𝑖 = 𝑋𝑖 − 𝑌𝑖 For large samples 𝑍= 𝑆22 1 1 = 𝑆𝑝 √ + 𝑛2 𝑛1 𝑛2 ̅ − 𝜇𝐷 𝐷 𝑆𝐷 /√𝑛 ~𝑁(0,1) Covariance 4 Pearson’s Sample Correlation Coefficient 𝑛 𝑟= (𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 1 ∑ 𝑍𝑥𝑖 𝑍𝑦𝑖 = (𝑛 − 1)𝑆𝑥 𝑆𝑦 𝑛−1 𝑖=1 𝑐𝑜𝑣(𝑋, 𝑌) = 𝐸[(𝑋 − 𝐸[𝑋])(𝑌 − 𝐸[𝑌])] ∑𝑛𝑖=1(𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 𝑐𝑜𝑣(𝑋, 𝑌) = (𝑛 − 1) Correlation coefficient 𝑟= Covariance Properties 𝑐𝑜𝑣(𝑋, 𝑌) 𝑆𝑥 𝑆𝑦 Linear Model 𝑐𝑜𝑣(𝑋, 𝑌) = 𝐸[𝑋𝑌] − 𝐸[𝑋]𝐸[𝑌] With a and b as scalars 𝑌 = 𝑎𝑋 + 𝑏 Conditional expectation 𝑐𝑜𝑣(𝑋, 𝑎) = 0 𝑐𝑜𝑣(𝑋, 𝑋) = 𝜎𝑋2 𝑐𝑜𝑣(𝑋, 𝑌) = 𝑐𝑜𝑣(𝑌, 𝑋) 𝑐𝑜𝑣(𝑎𝑋, 𝑏𝑌) = 𝑎𝑏𝑐𝑜𝑠(𝑋, 𝑌) 𝑐𝑜𝑣(𝑋 + 𝑎, 𝑌 + 𝑏) = 𝑐𝑜𝑣(𝑋, 𝑌) 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 Law of iterated expectations 𝐸[𝐸[𝑌|𝑋]] = 𝐸[𝑌] Using expectation on both sides for 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 𝐸[𝑌] = 𝑎𝐸[𝑋] + 𝑏 𝑏 = 𝐸[𝑌] − 𝑎𝐸[𝑋] If X and Y are independent, then 𝑐𝑜𝑣(𝑋, 𝑌) = 0 Covariance 𝑐𝑜𝑣(𝑋, 𝑌) = 𝑎 𝑣𝑎𝑟(𝑋) 𝑐𝑜𝑣(𝑋, 𝑌) 𝑎= 𝑣𝑎𝑟(𝑋) Maximum Likelihood Estimator (MLE) Suppose iid Xi ~𝑁(𝜇, 𝜎2 ) MLE for the mean 𝜇 1 𝑃(𝑋𝑖 = 𝑥𝑖 |𝜇) = −(𝑥−𝜇)2 ⁄2𝜎 2 𝑒 √2𝜋𝜎 𝑛 1 𝑛 − ∑(𝑥−𝜇)2⁄2𝜎2 |𝜇) ∏ 𝑃(𝑋𝑖 = 𝑥𝑖 = ( ) 𝑒 √2𝜋𝜎 𝑖=1 𝑛 1 𝜇̂ = ∑ 𝑥𝑖 𝑛 𝑛 𝑖=1 Least Square Error (LSE) 𝑦𝑖 = 𝑎𝑥𝑖 + 𝑏 + 𝜖𝑖 Sum of the squares of the residuals (SSE) 𝑛 𝑆𝑆𝐸 = 𝑛 ∑ 𝜖𝑖2 𝑖=1 = ∑(𝑦𝑖 − 𝑎𝑥𝑖 − 𝑏)2 𝑖=1 Properties of the LSE Solution 𝑛 ∑ 𝜖𝑖 = 0 𝑖=1 Unbiased estimator for 𝑣𝑎𝑟(𝜖) is SSE/(n-2) 1 𝜎̂2 = ∑(𝑥𝑖 − 𝑥̅ )2 𝑛 𝑖=1 𝑛 1 ̂ 𝑦) = ∑(𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 𝑐𝑜𝑣(𝑥, 𝑛 𝑖=1 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 ∑𝑛𝑖=1(𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 𝑎= ∑𝑛𝑖=1(𝑥𝑖 − 𝑥̅ )2 𝑏 = 𝑦̅ − 𝑎𝑥̅ 5 Linear Model 𝑌 = 𝑎𝑋 + 𝑏 Conditional expectation REGRESSION MODELS Maximum Likelihood Estimator (MLE) Suppose iid Xi ~𝑁(𝜇, 𝜎2 ) 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 Law of iterated expectations 𝐸[𝐸[𝑌|𝑋]] = 𝐸[𝑌] Using expectation on both sides for 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 𝐸[𝑌] = 𝑎𝐸[𝑋] + 𝑏 𝑏 = 𝐸[𝑌] − 𝑎𝐸[𝑋] Covariance MLE for the mean 𝜇 1 𝑃(𝑋𝑖 = 𝑥𝑖 |𝜇) = 2 ⁄2𝜎2 √2𝜋𝜎 1 𝑛 − ∑(𝑥−𝜇)2 ⁄2𝜎2 ∏ 𝑃(𝑋𝑖 = 𝑥𝑖 |𝜇) = ( ) 𝑒 √2𝜋𝜎 𝑖=1 𝑛 𝑛 1 𝜇̂ = ∑ 𝑥𝑖 𝑛 𝑛 𝑐𝑜𝑣(𝑋, 𝑌) = 𝑎 𝑣𝑎𝑟(𝑋) 𝑐𝑜𝑣(𝑋, 𝑌) 𝑎= 𝑣𝑎𝑟(𝑋) 𝑒 −(𝑥−𝜇) 𝑖=1 1 𝜎̂2 = ∑(𝑥𝑖 − 𝑥̅ )2 𝑛 𝑖=1 𝑛 1 ̂ 𝑦) = ∑(𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 𝑐𝑜𝑣(𝑥, 𝑛 𝑖=1 𝐸[𝑌|𝑋] = 𝑎𝑋 + 𝑏 ∑𝑛𝑖=1(𝑥𝑖 − 𝑥̅ )(𝑦𝑖 − 𝑦̅) 𝑎= ∑𝑛𝑖=1(𝑥𝑖 − 𝑥̅ )2 𝑏 = 𝑦̅ − 𝑎𝑥̅ Least Square Error (LSE) Linear Regression 𝑦𝑖 = 𝑎𝑥𝑖 + 𝑏 + 𝜖𝑖 Sum of the squares of the residuals (SSE) 𝑛 𝑆𝑆𝐸 = 𝑛 ∑ 𝜖𝑖2 𝑖=1 = ∑(𝑦𝑖 − 𝑎𝑥𝑖 − 𝑏)2 𝑀𝑜𝑑𝑒𝑙: 𝑦𝑖 = 𝑎𝑥𝑖 + 𝑏 + 𝜖𝑖 𝑃𝑟𝑒𝑑𝑖𝑐𝑡𝑒𝑑 𝑉𝑎𝑙𝑢𝑒: 𝑦̂𝑖 = 𝑎𝑥𝑖 + 𝑏 𝑅𝑒𝑠𝑖𝑑𝑢𝑎𝑙: 𝜖𝑖 = 𝑦𝑖 − 𝑦̂𝑖 𝑖=1 Properties of the LSE Solution 𝑛 ∑ 𝜖𝑖 = 0 𝑖=1 Unbiased estimator for 𝑣𝑎𝑟(𝜖) is SSE/(n-2) Goodness of Fit 𝐺𝑜𝑜𝑑𝑛𝑒𝑠𝑠 𝑜𝑓 𝑓𝑖𝑡 𝑚𝑒𝑡𝑟𝑖𝑐: 𝑅 2 = 𝑟 2 𝑐𝑜𝑣(𝑋, 𝑌) 𝑟= √𝑣𝑎𝑟(𝑋)𝑣𝑎𝑟(𝑌) 𝑐𝑜𝑣 2 (𝑋, 𝑌) 𝑅2 = 𝑣𝑎𝑟(𝑋)𝑣𝑎𝑟(𝑌) 𝑣𝑎𝑟(𝜖) 𝑅2 = 1 − 𝑣𝑎𝑟(𝑌) Y = 0 + 1 X + Y = dependent variable (response) X = independent variable (predictor or explanatory) 0 = intercept (value of Y when X = 0) 1 = slope of the regression line = random error Uncertainty of Fit parameters 𝑦̂𝑖 = 𝑎𝑥𝑖 + 𝑏 Using t-distribution and DF = n - 2 𝑣𝑎𝑟(𝜖) 𝑛 𝑣𝑎𝑟(𝑋) 𝑣𝑎𝑟(𝜖) 𝑥̅ 2 𝑣𝑎𝑟(𝑏) = (1 + ) 𝑛 𝑣𝑎𝑟(𝑋) (𝑥𝑖 − 𝑥̅ )2 𝑣𝑎𝑟(𝜖) 𝑣𝑎𝑟(𝑦̂𝑖 ) = + (1 ) 𝑛 𝑣𝑎𝑟(𝑋) 𝑣𝑎𝑟(𝑎) = Yˆ = b0 + b1 X Yˆ = predicted value of Y b0 = estimate of 0 , based on sample results b1 = estimate of 1 , based on sample results 6 Error = (Actual value) – (Predicted value) e = Y − Yˆ X= Y= b1 = X = average (mean) of X values n Y = average (mean) of Y values n ( X − X )(Y − Y ) (X − X) 2 b0 = Y − b1 X Sum of Squares Total = SST = (Y − Y ) 2 Sum of Squares Error = SSE = e 2 = (Y − Yˆ ) 2 Correlation Coefficient = r = r 2 SSR SSE =1– SST SST SSE Mean Squared Error = s 2 = MSE = n − k −1 Generic Linear Model Y = 0 + 1 X + SST = SSR + SSE Sum of Squares Regression = SSR = (Yˆ − Y ) 2 Standard Error of Estimate = s = MSE SSR MSR = k k = number of independent variables in the model Hypothesis Test H 0 : 1 = 0 Coefficien t of Determination = r 2 = MSR MSE degrees of freedom for the numerator = df1 = k degrees of freedom for the denominator = df2 = n – k–1 F Statistic : F = Y = 0 + 1X1 + 2X2 + … + kXk + Y= dependent variable (response variable) Xi = ith independent variable (predictor or explanatory variable) 0 = intercept (value of Y when all Xi = 0) i = coefficient of the ith independent variable k= number of independent variables = random error SSE /( n − k − 1) Adjusted r 2 = 1 − SST /( n − 1) H 1 : 1 0 Reject if Fcalculated F , df1 , df2 df 1 = k df 2 = n − k − 1 p - value = P( F calculated test statistic) Reject if p - value Yˆ = b0 + b1 X 1 + b2 X 2 + ... + bk X k Yˆ = predicted value of Y b0 = sample intercept (an estimate of 0) bi = sample coefficient of the i th variable (an estimate of i) (∑𝑛𝑖=1 𝑦𝑖 )(∑𝑛𝑖=1 𝑥𝑖 ) 𝑆𝑥𝑦 𝑛 Least Square Estimates β̂0 = 𝑦̅ − β̂1 𝑥̅ 𝑎𝑛𝑑 β̂1 = = 2 𝑆𝑥𝑥 (∑𝑛 𝑥 ) ∑𝑛𝑖=1 𝑥𝑖2 − 𝑖=1 𝑖 𝑛 1 Where 𝑥̅ = (𝑛) ∑𝑛𝑖=1 𝑥𝑖 𝑎𝑛𝑑 𝑦̅ = (1/𝑛) ∑𝑛𝑖=1 𝑦𝑖 ∑𝑛𝑖=1 𝑦𝑖 𝑥𝑖 − 𝑛 𝐸𝑟𝑟𝑜𝑟 𝑠𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠, 𝑆𝑆𝐸 = 𝑆𝑆 𝐸 𝑈𝑛𝑏𝑖𝑎𝑠𝑒𝑑 𝑒𝑠𝑡𝑖𝑚𝑎𝑡𝑜𝑟, 𝜎̂ 2 = 𝑛−2 ∑ 𝑒𝑖2 𝑖=1 𝑛 = ∑(𝑦𝑖 − 𝑦̂𝑖 )2 𝑖=1 𝑆𝑆𝐸 = 𝑆𝑆𝑇 − β̂1 𝑆𝑥𝑦 SS 𝑇 = ∑𝑛𝑖=1(𝑦𝑖 − 𝑦̅)2 𝜎̂ 2 Estimated Standard error of the slope, se(β̂1 ) = √ 𝑆𝑥𝑥 7 1 𝑥̅ 2 Estimated Standard error of the intercept, se(β̂0 ) = √𝜎̂ 2 [ + ] 𝑛 𝑆𝑥𝑥 𝐻𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠 𝑇𝑒𝑠𝑡𝑖𝑛𝑔 (𝑡 − 𝑡𝑒𝑠𝑡) Null hypothesis: 𝐻0 : 𝛽1 = 𝛽1,0 𝐻1 : 𝛽1 ≠ 𝛽1,0 Test Statistic: : 𝑇0 : ̂1 −𝛽1,0 𝛽 Reject null hypothesis if |𝑡0 | > t α/2,n−2 ̂ 2 /𝑆𝑥𝑥 √𝜎 Null hypothesis: 𝐻0 : 𝛽0 = 𝛽0,0 𝐻1 : 𝛽0 ≠ 𝛽0,0 Test Statistic: : 𝑇0 : ̂0 −𝛽0,0 𝛽 ̅2 1 𝑥 ] 𝑛 𝑆𝑥𝑥 Reject null hypothesis if |𝑡0 | > t α/2,n−2 √𝜎 ̂ 2[ + 𝑛 𝑛 𝑛 2 Analysis of Variance Identity, ∑(𝑦𝑖 − 𝑦̅) = ∑(𝑦̂𝑖 − 𝑦̅) + ∑(𝑦𝑖 − 𝑦̂𝑖 )2 𝑖=1 2 𝑖=1 𝑖=1 𝑆𝑆𝑇 = 𝑆𝑆𝑅 + 𝑆𝑆𝐸 𝑆𝑆𝑅 /1 𝑀𝑆𝑅 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝑠𝑖𝑔𝑛𝑖𝑓𝑎𝑛𝑐𝑒 𝑜𝑓 𝑅𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛, 𝐹0 = = 𝑓𝑜𝑙𝑙𝑜𝑤𝑠 𝐹1,𝑛−2 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑆𝑆𝐸 /(𝑛 − 2) 𝑀𝑆𝐸 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠, 𝑖𝑓 𝑓0 > 𝑓𝛼,1,𝑛−2 2 ̂ ̂ 𝜎 𝜎 100(1−∝)% CI of slope β0 is β̂1 − 𝑡𝛼,𝑛−2 √ ≤ 𝛽1 ≤ β̂1 + 𝑡𝛼,𝑛−2 √ 2 𝑆𝑥𝑥 2 2 𝑆𝑥𝑥 𝑥̅ 2 2 1 1 𝑥̅ 100(1−∝)% CI of slope β1 is β̂0 − 𝑡𝛼,𝑛−2 √𝜎̂ 2 [𝑛 + 𝑆 ] ≤ 𝛽0 ≤ β̂0 + 𝑡𝛼,𝑛−2 √𝜎̂ 2 [𝑛 + 𝑆 ] 𝑥𝑥 2 𝑥𝑥 2 100(1−∝)% CI about the mean response x = x0 , say μ𝑌|𝑥0 , is 1 (𝑥0 − 𝑥̅ )2 1 (𝑥0 − 𝑥̅ )2 μ̂𝑌|𝑥0 − 𝑡𝛼,𝑛−2 √𝜎̂ 2 [ + ] ≤ μ𝑌|𝑥0 ≤ μ̂𝑌|𝑥0 + 𝑡𝛼,𝑛−2 √𝜎̂ 2 [ + ] 𝑛 𝑆𝑥𝑥 𝑛 𝑆𝑥𝑥 2 2 𝑤ℎ𝑒𝑟𝑒 μ̂𝑌|𝑥 = β̂0 + β̂1 𝑥0 𝑖𝑠 𝑐𝑜𝑚𝑝𝑢𝑡𝑒𝑑 𝑓𝑟𝑜𝑚 𝑡ℎ𝑒 𝑓𝑖𝑡𝑡𝑒𝑑 𝑟𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑚𝑜𝑑𝑒𝑙 0 100(1−∝)% PI on a future observation Y0 at the value x0 is 1 (𝑥0 − 𝑥̅ )2 1 (𝑥0 − 𝑥̅ )2 ŷ0 − 𝑡𝛼,𝑛−2 √𝜎̂ 2 [ + ] ≤ Y0 ≤ ŷ0 + 𝑡𝛼,𝑛−2 √𝜎̂ 2 [ + ] 𝑛 𝑆𝑥𝑥 𝑛 𝑆𝑥𝑥 2 2 𝑤ℎ𝑒𝑟𝑒 ŷ0 = β̂0 + β̂1 𝑥0 𝑖𝑠 𝑐𝑜𝑚𝑝𝑢𝑡𝑒𝑑 𝑓𝑟𝑜𝑚 𝑡ℎ𝑒 𝑟𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑚𝑜𝑑𝑒𝑙 𝑅𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 𝑒𝑖 = 𝑦𝑖 − ŷ𝑖 d𝑖 = 𝑒𝑖 /√𝜎̂ 2 𝑆𝑆 𝑆𝑆 𝑅 Coefficient of determination, R2 = =1− 𝐸 𝐶𝑜𝑟𝑟𝑒𝑙𝑎𝑡𝑖𝑜𝑛 𝐶𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡, 𝜌 = 𝑆𝑆𝑇 𝑇0 = 𝑅√𝑛 − 2 𝑆𝑆𝑇 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝑧𝑒𝑟𝑜 𝑐𝑜𝑟𝑟𝑒𝑙𝑎𝑡𝑖𝑜𝑛: 𝐻0 : 𝜌 = 0 𝜎𝑋𝑌 𝜎𝑋 𝜎𝑌 ℎ𝑎𝑠 𝑡 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑛 − 2 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 √1 − 𝑅 2 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠, 𝑖𝑓 |𝑡0 | > 𝑡𝛼,𝑛−2 2 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝑐𝑜𝑟𝑟𝑒𝑙𝑎𝑡𝑖𝑜𝑛: 𝐻0 : 𝜌 = 𝜌0 𝑍0 = (arctanh 𝑅 − arctanh 𝜌0 )(𝑛 − 3)1/2 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠, 𝑖𝑓 |𝑧0 | > 𝑧𝛼 2 100(1−∝)% CI for correlation coefficient is 𝑧𝛼 𝑧𝛼/2 2 tanh (arctanh 𝑟 − ) ≤ 𝜌 ≤ tanh (arctanh 𝑟 + ) 𝑤ℎ𝑒𝑟𝑒 tanh 𝑢 = (𝑒𝑢 −𝑒−𝑢 )/(𝑒𝑢 +𝑒−𝑢 ) √𝑛 − 3 √𝑛 − 3 1 𝐹𝑖𝑡𝑡𝑒𝑑 𝑙𝑜𝑔𝑖𝑠𝑡𝑖𝑐 𝑒𝑥𝑝𝑟𝑒𝑠𝑠𝑖𝑜𝑛, 𝑦̂ = 1 + 𝑒𝑥𝑝[−(𝛽0 + 𝛽1 𝑥)] Multiple Regression 8 𝑦1 𝑦2 𝑦=[⋮] 𝑦𝑛 𝑦 = 𝑋𝛽 + 𝜖 𝑦𝑖 = 𝛽0 + 𝛽1 𝑥𝑖1 + 𝛽2 𝑥𝑖2 + ⋯ + 𝛽𝑘 𝑥𝑖𝑘 + 𝜖𝑖 𝑖 = 1,2, … , 𝑛 𝜖1 𝛽1 1 𝑥11 𝑥12 … 𝑥1𝑘 … 𝑥 𝑥 𝑥 𝜖 𝛽 21 22 2𝑘 2 𝑋 = [1 𝛽 = [ 2] 𝜖=[⋮] ⋮ ⋮ ⋮ ] ⋮ 1 𝑥𝑛1 𝑥𝑛2 ⋮ … 𝑥𝑛𝑘 𝑛 𝜖𝑛 𝛽𝑛 𝑛 𝑛 𝑛 𝑛 𝛽̂0 + 𝛽̂1 ∑ 𝑥𝑖1 + 𝛽̂2 ∑ 𝑥𝑖2 + ⋯ + 𝛽̂𝑘 ∑ 𝑥𝑖𝑘 = ∑ 𝑦𝑖 𝑖=1 𝑛 𝑛 𝛽̂0 ∑ 𝑥𝑖1 + 𝑖=1 𝑛 2 𝛽̂1 ∑ 𝑥𝑖1 𝑖=1 𝑛 𝑖=1 𝑛 𝑖=1 𝑛 𝑖=1 𝑛 + 𝛽̂2 ∑ 𝑥𝑖1 𝑥𝑖2 + ⋯ + 𝛽̂𝑘 ∑ 𝑥𝑖1 𝑥𝑖𝑘 = ∑ 𝑥𝑖1 𝑦𝑖 𝑖=1 𝑖=1 𝑛 𝑖=1 𝑛 𝑛 2 𝛽̂0 ∑ 𝑥𝑖𝑘 + 𝛽̂1 ∑ 𝑥𝑖𝑘 𝑥𝑖1 + 𝛽̂2 ∑ 𝑥𝑖𝑘 𝑥𝑖2 + ⋯ + 𝛽̂𝑘 ∑ 𝑥𝑖𝑘 = ∑ 𝑥𝑖𝑘 𝑦𝑖 𝑖=1 𝑖=1 𝑛 𝑛 𝑛 𝑖=1 𝑁𝑜𝑟𝑚𝑎𝑙 𝐸𝑞𝑢𝑎𝑡𝑖𝑜𝑛𝑠: 𝑋 ′ 𝑋𝛽̂ = 𝑋′𝑌 𝑛 ∑ 𝑥𝑖1 ∑ 𝑥𝑖2 𝑖=1 𝑛 𝑖=1 𝑛 2 ∑ 𝑥𝑖1 ∑ 𝑥𝑖1 𝑥𝑖2 𝑖=1 𝑖=1 𝑖=1 𝑛 ∑ 𝑥𝑖𝑘 [ 𝑖=1 ⋯ ∑ 𝑥𝑖𝑘 𝑥𝑖1 ∑ 𝑥𝑖𝑘 𝑥𝑖2 … 𝑖=1 𝑖=1 𝑖=1 𝑛 ⋮ 2 ∑ 𝑥𝑖𝑘 𝑛 𝛽̂0 ⋯ ∑ 𝑥𝑖1 𝑥𝑖𝑘 ⋮ 𝑛 ∑ 𝑥𝑖𝑘 𝑖=1 𝑛 ⋮ 𝑛 ⋮ 𝑖=1 𝑛 ∑ 𝑥𝑖1 ⋮ 𝑖=1 ∑ 𝑦𝑖 𝑖=1 𝑛 ∑ 𝑥𝑖1 𝑦𝑖 𝛽̂1 = ⋮ 𝑖=1 [𝛽̂𝑘 ] 𝑛 ⋮ ∑ 𝑥𝑖𝑘 𝑦𝑖 ] [ 𝑖=1 ] 𝐿𝑒𝑎𝑠𝑡 𝑠𝑞𝑢𝑎𝑟𝑒 𝑒𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝛽: 𝛽̂ = (𝑋 ′ 𝑋)−1 𝑋′𝑌 ∑𝑛𝑖=1 𝑒𝑖2 𝑆𝑆𝐸 2 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑜𝑟 𝑜𝑓 𝑉𝑎𝑟𝑖𝑎𝑛𝑐𝑒, 𝜎̂ = = 𝑛−𝑝 𝑛−𝑝 𝑖=1 𝑛 𝐸𝑟𝑟𝑜𝑟 𝑠𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠, 𝑆𝑆𝐸 = 𝑛 2 ∑ 𝑒𝑖 = ∑(𝑦𝑖 𝑖=1 𝑖=1 ′ −1 − 𝑦̂𝑖 )2 = 𝑒′ 𝑒 𝐶𝑜𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒, 𝐶 = (𝑋 𝑋) ̂ 2 𝐶𝑖𝑗 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒𝑑 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑒𝑟𝑟𝑜𝑟 𝑜𝑓 𝛽̂𝑗 = 𝑠𝑒(𝛽̂𝑗 ) = √𝜎 Hypothesis of ANOVA Test Null hypothesis: 𝐻0 : 𝛽1 = 𝛽2 = ⋯ = 𝛽𝑘 = 0 𝐻1 : 𝛽𝑗 ≠ 0 𝑓𝑜𝑟 𝑎𝑡 𝑙𝑒𝑎𝑠𝑡 𝑜𝑛𝑒 𝑗 𝑆𝑆𝑅 /𝑘 𝑀𝑆𝑅 𝑇𝑒𝑠𝑡 𝑠𝑡𝑎𝑡𝑖𝑠𝑡𝑖𝑐 𝑓𝑜𝑟 𝐴𝑁𝑂𝑉𝐴, 𝐹0 = = 𝑆𝑆𝐸 /(𝑛 − 𝑝) 𝑀𝑆𝐸 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠, 𝑖𝑓 𝑓0 > 𝑓𝛼,𝑘,𝑛−𝑝 (∑𝑛𝑖=1 𝑦𝑖 )2 (∑𝑛𝑖=1 𝑦𝑖 )2 ′ ′ ̂ 𝑆𝑆𝐸 = 𝑦 𝑦 − − [ 𝛽𝑋 𝑦 − ] 𝑛 𝑛 (∑𝑛𝑖=1 𝑦𝑖 )2 𝑆𝑆𝑅 = 𝛽̂ 𝑋 ′ 𝑦 − 𝑛 𝑆𝑆𝑅 𝑆𝑆𝐸 2 Coefficient of determination, R = =1− 𝑆𝑆𝑇 𝑆𝑆𝑇 𝑆𝑆 /(𝑛 − 𝑝) 𝐸 Adjusted R2 , R2𝑎𝑑𝑗 = 𝑆𝑆𝑇 /(𝑛 − 1) Null hypothesis: 𝐻0 : 𝛽𝑗 = 𝛽𝑗,0 𝐻1 : 𝛽𝑗 ≠ 𝛽𝑗,0 9 Test Statistic: : 𝑇0 : ̂𝑗 −𝛽𝑗,0 𝛽 ̂ 2 𝐶𝑗𝑗 √𝜎 = ̂𝑗 −𝛽𝑗,0 𝛽 ̂𝑗 ) 𝑠𝑒(𝛽 Reject null hypothesis if |𝑡0 | > t α/2,n−p 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝑠𝑖𝑔𝑛𝑖𝑓𝑎𝑛𝑐𝑒 𝑜𝑓 𝑅𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝐻0 : 𝛽1 = 0 𝐻1 : 𝛽1 ≠ 0 𝑆𝑆𝑅 (𝛽1 |𝛽2 )/𝑟 𝐹0 = 𝑀𝑆𝐸 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠, 𝑖𝑓 𝑓0 > 𝑓𝛼,𝑟,𝑛−𝑝 . 𝑇ℎ𝑖𝑠 𝑐𝑜𝑛𝑐𝑙𝑢𝑑𝑒𝑠 𝑡ℎ𝑎𝑡 𝑎𝑡 𝑙𝑒𝑎𝑠𝑡 𝑜𝑛𝑒 𝑜𝑓 𝑝𝑎𝑟𝑎𝑚𝑒𝑡𝑒𝑟𝑠 𝑖𝑛 𝛽1 𝑖𝑠 𝑛𝑜𝑡 𝑧𝑒𝑟𝑜. 100(1−∝)% CI on the regression coefficient β𝑗 j = 0,1,2, … , k is β̂𝑗 − 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 𝐶𝑖𝑗 ≤ 𝛽𝑗 ≤ β̂𝑗 + 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 𝐶𝑖𝑗 2 2 100(1−∝)% CI about the mean response at point x01 , x02 , … , x0𝑘 , is μ̂𝑌|𝑥0 − 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 𝑥0′ (𝑋 ′ 𝑋)−1 𝑥0 ≤ μ𝑌|𝑥0 ≤ μ̂𝑌|𝑥0 + 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 𝑥0′ (𝑋 ′ 𝑋)−1 𝑥0 2 2 100(1−∝)% PI on a future observation Y0 at the value x01 , x02 , … , x0𝑘 is ŷ0 − 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 (1 + 𝑥0′ (𝑋 ′ 𝑋)−1 𝑥0 ) ≤ Y0 ≤ ŷ0 + 𝑡𝛼,𝑛−𝑝 √𝜎̂ 2 (1 + 𝑥0′ (𝑋 ′ 𝑋)−1 𝑥0 ) 2 2 𝑅𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 𝑒𝑖 = 𝑦𝑖 − ŷ𝑖 𝑒𝑖 d𝑖 = = 𝑒𝑖 /√𝜎̂ 2 √𝑀𝑆𝐸 𝑒𝑖 𝑆𝑡𝑢𝑑𝑒𝑛𝑡𝑖𝑧𝑒𝑑 𝑅𝑒𝑠𝑖𝑑𝑢𝑎𝑙𝑠 𝑟𝑖 = 𝑊ℎ𝑒𝑟𝑒 𝑖 = 1,2, … , 𝑛 √𝜎̂ 2 (1 − ℎ𝑖𝑖 ) 𝐻𝑎𝑡 𝑚𝑎𝑡𝑟𝑖𝑥: 𝐻 = 𝑋(𝑋 ′ 𝑋)−1 𝑋′ ℎ𝑖𝑖 = 𝑖𝑡ℎ 𝑑𝑖𝑎𝑔𝑜𝑛𝑎𝑙 𝑒𝑙𝑒𝑚𝑒𝑛𝑡 𝑜𝑓 𝑡ℎ𝑒 𝐻 𝑚𝑎𝑡𝑟𝑖𝑥 = 𝑥𝑖′ (𝑋 ′ 𝑋)−1 𝑥𝑖 ′ 𝑆𝑆𝐸 (𝑝) 𝐶𝑜𝑜𝑘 𝑠𝐷𝑖𝑠𝑡𝑎𝑛𝑐𝑒 𝐶𝑝 𝑆𝑡𝑎𝑡𝑖𝑠𝑡𝑖𝑐, 𝐶𝑝 = − 𝑛 + 2𝑝 ′ ′ ̂ ̂ ̂ ̂ 𝜎̂ 2 (𝛽(𝑖) − 𝛽 ) 𝑋 𝑋(𝛽(𝑖) − 𝛽 ) 𝐷𝑖 = 𝑤ℎ𝑒𝑟𝑒 𝑖 = 1,2, … , 𝑛 𝑃𝑟𝑒𝑑𝑖𝑐𝑡𝑖𝑜𝑛 𝐸𝑟𝑟𝑜𝑟 𝑆𝑢𝑚 𝑜𝑓 𝑆𝑞𝑢𝑎𝑟𝑒𝑠 𝑝𝜎̂ 2 𝑛 𝑛 2 𝑒𝑖 2 𝑟𝑖2 ℎ𝑖𝑖 𝑃𝑅𝐸𝑆𝑆 = ∑(𝑦𝑖 − 𝑦̂(𝑡) ) = ∑ ( ) 𝐷𝑖 = 𝑤ℎ𝑒𝑟𝑒 𝑖 = 1,2, … , 𝑛 1 − ℎ𝑖𝑖 𝑝 (1 − ℎ ) 𝑖𝑖 𝑖=1 𝑆𝑡𝑒𝑝𝑤𝑖𝑠𝑒 𝑅𝑒𝑔𝑟𝑒𝑠𝑠𝑖𝑜𝑛 𝑆𝑆𝑅 (𝛽𝑗 |𝛽1 , 𝛽0 ) 𝐹𝑗 = 𝑀𝑆𝐸 (𝑥𝑗 , 𝑥1 ) Single-Factor Experiments 𝑖=1 𝑉𝑎𝑟𝑖𝑎𝑛𝑐𝑒 𝐼𝑛𝑓𝑙𝑎𝑡𝑖𝑜𝑛 𝐹𝑎𝑐𝑡𝑜𝑟 (𝑉𝐼𝐹) 1 𝑉𝐼𝐹(𝛽𝑗 ) = 𝑊ℎ𝑒𝑟𝑒 𝑗 = 1,2, … , 𝑘 (1 − 𝑅𝑗2 ) 𝑎 𝑛 𝑎 2 𝑎 𝑆𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠 𝐼𝑑𝑒𝑛𝑡𝑖𝑡𝑦, ∑ ∑(𝑦𝑖𝑗 − 𝑦̅.. ) = 𝑎 ∑(𝑦𝑖. − 𝑦̅.. 𝑖=1 𝑗=1 )2 𝑖=1 𝑆𝑆𝑇 = 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 + 𝑆𝑆𝐸 𝑛 2 + ∑ ∑(𝑦𝑖𝑗 − 𝑦̅𝑖. ) 𝑖=1 𝑗=1 𝑎 𝐸(𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 ) = (𝑎 − 1)𝜎 + 𝑛 ∑ 𝜏𝑖2 2 𝑖=1 𝐸(𝑆𝑆𝐸 ) = 𝑎(𝑛 − 1)𝜎 2 𝑀𝑆𝐸 = 𝑆𝑆𝐸 /[𝑎(𝑛 − 1)] 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 /(𝑎 − 1) 𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 𝐹0 = = 𝑆𝑆𝐸 /[𝑎(𝑛 − 1)] 𝑀𝑆𝐸 𝑎 𝑛 2 𝑦.. 2 𝑆𝑆𝑇 = ∑ ∑ 𝑦𝑖𝑗 − 𝑁 𝑖=1 𝑗=1 𝑎 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 = ∑ 𝑖=1 𝑦𝑖.2 𝑦..2 − 𝑛 𝑁 𝑆𝑆𝐸 = 𝑆𝑆𝑇 − 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 10 𝑇= 𝑌̅𝑖. − 𝜇𝑖 ℎ𝑎𝑠 𝑡 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑎(𝑛 − 1) 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 √𝑀𝑆𝐸 /𝑛 100(1−∝)% CI about the mean of the i𝑡ℎ treatment μ𝑖 , is 𝑀𝑆𝐸 𝑀𝑆𝐸 𝑦̅𝑖. − 𝑡𝛼,𝑎(𝑛−1) √ ≤ μ𝑖 ≤ 𝑦̅𝑖. + 𝑡𝛼,𝑎(𝑛−1) √ 𝑛 𝑛 2 2 100(1−∝)% CI on the difference in two treatment means μ𝑖 − μ𝑗 , is 2𝑀𝑆𝐸 2𝑀𝑆𝐸 𝑦̅𝑖. − 𝑦̅𝑗. − 𝑡𝛼,𝑎(𝑛−1) √ ≤ μ𝑖 − μ𝑗 ≤ 𝑦̅𝑖. − 𝑦̅𝑗. + 𝑡𝛼,𝑎(𝑛−1) √ 𝑛 𝑛 2 2 𝐹𝑖𝑠ℎ𝑒𝑟 ′ 𝑠 𝐿𝑒𝑎𝑠𝑡 𝑆𝑖𝑔𝑛𝑖𝑓𝑖𝑐𝑎𝑛𝑡 𝐷𝑖𝑓𝑓𝑒𝑟𝑒𝑛𝑐𝑒 (𝐿𝑆𝐷)𝑚𝑒𝑡ℎ𝑜𝑑 𝑁𝑢𝑙𝑙 𝐻𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠 𝐻0 : μ𝑖 = μ𝑗 𝑤ℎ𝑒𝑟𝑒 𝑖 ≠ 𝑗 𝑦̅𝑖. − 𝑦̅𝑗. 𝑡0 = √2𝑀𝑆𝐸 𝑛 𝑅𝑒𝑗𝑒𝑐𝑡 𝑛𝑢𝑙𝑙 ℎ𝑦𝑝𝑜𝑡ℎ𝑒𝑠𝑖𝑠 𝑖𝑓 |𝑦̅𝑖. − 𝑦̅𝑗. | > 𝐿𝑆𝐷 2𝑀𝑆𝐸 𝐿𝑆𝐷 = 𝑡𝛼,𝑎(𝑛−1) √ 𝑛 2 𝐼𝑓 𝑠𝑎𝑚𝑝𝑙𝑒 𝑠𝑖𝑧𝑒𝑠 𝑎𝑟𝑒 𝑑𝑖𝑓𝑓𝑒𝑟𝑒𝑛𝑡, 𝐿𝑆𝐷 = 𝑡𝛼,𝑁−𝑎 √𝑀𝑆𝐸 ( 2 1 1 + ) 𝑛𝑖 𝑛𝑗 𝑃𝑜𝑤𝑒𝑟 𝑜𝑓 𝐴𝑁𝑂𝑉𝐴 𝑡𝑒𝑠𝑡, 1 − 𝛽 = 𝑃{𝑅𝑒𝑗𝑒𝑐𝑡 𝐻0 |𝐻0 𝑖𝑠 𝑓𝑎𝑙𝑠𝑒} = 𝑃{𝐹0 > 𝑓𝛼,𝑎−1,𝑎(𝑛−1) |𝐻0 𝑖𝑠 𝑓𝑎𝑙𝑠𝑒} 𝑛 ∑𝑎𝑖=1 𝜏𝑖2 𝜙2 = 𝑎𝜎 2 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 𝐸(𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 ) = 𝐸 ( ) = 𝜎 2 + 𝑛𝜎𝜏2 𝑎−1 𝑆𝑆𝐸 𝐸(𝑀𝑆𝐸 ) = 𝐸 [ ] = 𝜎2 𝑎(𝑛 − 1) 𝜎̂ 2 = 𝑀𝑆𝐸 𝑀𝑆 𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 − 𝑀𝑆𝐸 𝜎̂𝜏2 = 𝑛 Randomized Block Experiment 𝑎 𝑏 2 𝑆𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠 𝐼𝑑𝑒𝑛𝑡𝑖𝑡𝑦, ∑ ∑(𝑦𝑖𝑗 − 𝑦̅.. ) 𝑎 𝑖=1 𝑗=1 = 𝑏 ∑(𝑦̅𝑖. − 𝑦̅.. 𝑖=1 )2 𝑏 2 𝑎 𝑏 2 + 𝑎 ∑(𝑦̅.𝑗 − 𝑦̅.. ) + ∑ ∑(𝑦𝑖𝑗 − 𝑦̅.𝑗 − 𝑦̅𝑖. + 𝑦̅.. ) 𝑗=1 𝑖=1 𝑗=1 𝑆𝑆𝑇 = 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 + 𝑆𝑆𝐵𝑙𝑜𝑐𝑘𝑠 + 𝑆𝑆𝐸 𝑏 ∑𝑎𝑖=1 𝜏𝑖2 𝐸(𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 ) = 𝜎 2 + 𝑎−1 ∑𝑏𝑗=1 𝛽𝑗2 𝑎 𝐸(𝑀𝑆𝐵𝑙𝑜𝑐𝑘𝑠 ) = 𝜎 2 + 𝑏−1 𝐸(𝑀𝑆𝐸 ) = 𝜎 2 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 𝑀𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 = 𝑎−1 𝑆𝑆𝐵𝑙𝑜𝑐𝑘𝑠 𝑀𝑆𝐵𝑙𝑜𝑐𝑘𝑠 = 𝑏−1 11 𝑀𝑆𝐸 = 𝑆𝑆𝐸 (𝑎 − 1)(𝑏 − 1) 𝑎 𝑏 2 𝑆𝑆𝑇 = ∑ ∑ 𝑦𝑖𝑗 − 𝑖=1 𝑗=1 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 𝑎 1 𝑦..2 = ∑ 𝑦𝑖.2 − 𝑏 𝑎𝑏 𝑖=1 𝑏 𝑆𝑆𝐵𝑙𝑜𝑐𝑘𝑠 = 𝑦..2 𝑎𝑏 1 𝑦..2 ∑ 𝑦.𝑗2 − 𝑎 𝑎𝑏 𝑗=1 𝑆𝑆𝐸 = 𝑆𝑆𝑇 − 𝑆𝑆𝑇𝑟𝑒𝑎𝑡𝑚𝑒𝑛𝑡𝑠 − 𝑆𝑆𝐵𝑙𝑜𝑐𝑘𝑠 𝑒𝑖𝑗 = 𝑦𝑖𝑗 − 𝑦̂𝑖𝑗 𝑦̂𝑖𝑗 = 𝑦̅𝑖. + 𝑦̅.𝑗 − 𝑦̅.. Two Factors experiment 𝑏 𝑛 𝑦𝑖.. = ∑ ∑ 𝑦𝑖𝑗𝑘 𝑗=1 𝑘=1 𝑦𝑖.. 𝑦̅𝑖.. = 𝑤ℎ𝑒𝑟𝑒 𝑖 = 1,2, … , 𝑎 𝑏𝑛 𝑎 𝑛 𝑦.𝑗. = ∑ ∑ 𝑦𝑖𝑗𝑘 𝑖=1 𝑘=1 𝑦.𝑗. 𝑦̅.𝑗. = 𝑤ℎ𝑒𝑟𝑒 𝑗 = 1,2, … , 𝑏 𝑎𝑛 𝑛 𝑦𝑖𝑗. = ∑ 𝑦𝑖𝑗𝑘 𝑘=1 𝑦̅𝑖𝑗. = 𝑦𝑖𝑗. 𝑤ℎ𝑒𝑟𝑒 𝑖 = 1,2, … , 𝑎 𝑛 𝑎 𝑏 𝑛 𝑦... = ∑ ∑ ∑ 𝑦𝑖𝑗𝑘 𝑖=1 𝑗=1 𝑘=1 𝑦... 𝑦̅... = 𝑤ℎ𝑒𝑟𝑒 𝑗 = 1,2, … , 𝑏 𝑎𝑏𝑛 𝑎 𝑏 𝑛 2 𝑆𝑆𝑇 = ∑ ∑ ∑(𝑦𝑖𝑗𝑘 − 𝑦̅... ) 𝑖=1 𝑗=1 𝑘=1 𝑎 𝑏 𝑛 2 𝑆𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠 𝐼𝑑𝑒𝑛𝑡𝑖𝑡𝑦, ∑ ∑ ∑(𝑦𝑖𝑗𝑘 − 𝑦̅... ) 𝑖=1 𝑗=1 𝑘=1 𝑎 = 𝑏𝑛 ∑(𝑦̅𝑖.. − 𝑦̅... 𝑎 𝑖=1 𝑏 𝑛 )2 𝑏 2 𝑎 𝑏 2 + 𝑎𝑛 ∑(𝑦̅.𝑗. − 𝑦̅... ) + 𝑛 ∑ ∑(𝑦̅𝑖𝑗. − 𝑦̅𝑖.. − 𝑦̅.𝑗. + 𝑦̅... ) 𝑗=1 𝑖=1 𝑗=1 2 + ∑ ∑ ∑(𝑦𝑖𝑗𝑘 − 𝑦̅𝑖𝑗. ) 𝑖=1 𝑗=1 𝑘=1 𝑆𝑆𝑇 = 𝑆𝑆𝐴 + 𝑆𝑆𝐵 + 𝑆𝑆𝐴𝐵 + 𝑆𝑆𝐸 𝑆𝑆𝐴 𝑏𝑛 ∑𝑎𝑖=1 𝜏𝑖2 𝐸(𝑀𝑆𝐴 ) = 𝐸 ( ) = 𝜎2 + 𝑎−1 𝑎−1 ∑𝑏𝑗=1 𝛽𝑗2 𝑎𝑛 𝑆𝑆𝐵 2 ) 𝐸(𝑀𝑆𝐵 = 𝐸 ( )=𝜎 + 𝑏−1 𝑏−1 12 𝑛 ∑𝑎𝑖=1 ∑𝑏𝑗=1(𝜏𝛽)2𝑖𝑗 𝑆𝑆𝐴𝐵 ) = 𝜎2 + (𝑎 − 1)(𝑏 − 1) (𝑎 − 1)(𝑏 − 1) 𝑆𝑆𝐸 𝐸(𝑀𝑆𝐸 ) = 𝐸 ( ) = 𝜎2 𝑎𝑏(𝑛 − 1) 𝐸(𝑀𝑆𝐴𝐵 ) = 𝐸 ( 𝑀𝑆𝐴 𝑀𝑆𝐸 𝑀𝑆𝐵 𝐹 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝐹𝑎𝑐𝑡𝑜𝑟 𝐵, 𝐹0 = 𝑀𝑆𝐸 𝑀𝑆𝐴𝐵 𝐹 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝐴𝐵 𝐼𝑛𝑡𝑒𝑟𝑎𝑐𝑡𝑖𝑜𝑛, 𝐹0 = 𝑀𝑆𝐸 𝐹 𝑇𝑒𝑠𝑡 𝑓𝑜𝑟 𝐹𝑎𝑐𝑡𝑜𝑟 𝐴, 𝐹0 = 𝑎 𝑏 𝑛 2 𝑆𝑆𝑇 = ∑ ∑ ∑ 𝑦𝑖𝑗𝑘 − 𝑖=1 𝑗=1 𝑘=1 𝑎 𝑦𝑖..2 𝑦...2 𝑆𝑆𝐴 = ∑ − 𝑏𝑛 𝑎𝑏𝑛 𝑖=1 𝑏 2 𝑦.𝑗. 𝑦...2 𝑆𝑆𝐵 = ∑ − 𝑎𝑛 𝑎𝑏𝑛 𝑗=1 𝑎 𝑏 2 𝑦𝑖𝑗. 𝑦...2 𝑆𝑆𝐴𝐵 = ∑ ∑ 𝑖=1 𝑗=1 𝑛 − 𝑎𝑏𝑛 𝑦...2 𝑎𝑏𝑛 − 𝑆𝑆𝐴 − 𝑆𝑆𝐵 𝑆𝑆𝐸 = 𝑆𝑆𝑇 − 𝑆𝑆𝐴𝐵 − 𝑆𝑆𝐴 − 𝑆𝑆𝐵 𝑒𝑖𝑗𝑘 = 𝑦𝑖𝑗𝑘 − 𝑦̂𝑖𝑗. Three-Factor Fixed Effects Model Source Sum of Degrees of of Squares Freedom Variation A SSA a–1 Mean Square MSA B SSB b–1 MSB C SSC c–1 MSC AB SSAB (a – 1)(b – 1) MSAB AC SSAC (a – 1)(c – 1) MSAC BC SSBC (b – 1)(c – 1) MSBC ABC SSABC (a – 1)(b – 1)(c – 1) MSABC abc(n – 1) abcn – 1 MSE Error SSE Total SST k 2 Factorial Designs Expected Mean Squares F0 𝑏𝑐𝑛 ∑ 𝜏𝑖2 𝑎−1 ∑ 𝛽𝑗2 𝑎𝑐𝑛 2 𝜎 + 𝑏−1 ∑ 𝛾𝑘2 𝑎𝑏𝑛 2 𝜎 + 𝑐−1 ∑ ∑(𝜏𝛽)2𝑖𝑗 𝑐𝑛 2 𝜎 + (𝑎 − 1)(𝑏 − 1) 𝑏𝑛 ∑ ∑(𝜏𝛾)2𝑖𝑘 𝜎2 + (𝑎 − 1)(𝑐 − 1) 𝑎𝑛 ∑ ∑(𝛽𝛾)2𝑗𝑘 2 𝜎 + (𝑏 − 1)(𝑐 − 1) ∑ ∑ ∑(𝜏𝛽𝛾)2𝑖𝑗𝑘 𝑛 𝜎2 + (𝑎 − 1)(𝑏 − 1)(𝑐 − 1) 𝜎2 𝜎2 + 𝑀𝑆𝐴 𝑀𝑆𝐸 𝑀𝑆𝐵 𝑀𝑆𝐸 𝑀𝑆𝐶 𝑀𝑆𝐸 𝑀𝑆𝐴𝐵 𝑀𝑆𝐸 𝑀𝑆𝐴𝐶 𝑀𝑆𝐸 𝑀𝑆𝐵𝐶 𝑀𝑆𝐸 𝑀𝑆𝐴𝐵𝐶 𝑀𝑆𝐸 (l) Represents the treatment combination with both factors at the low level. 𝑀𝑎𝑖𝑛 𝐸𝑓𝑓𝑒𝑐𝑡 𝑜𝑓 𝐹𝑎𝑐𝑡𝑜𝑟 𝐴: 𝐴 = 𝑦̅𝐴+ − 𝑦̅𝐴− = 𝑎 + 𝑎𝑏 𝑏 + (𝑙) 1 [𝑎 + 𝑎𝑏 − 𝑏 − (𝑙)] − = 2𝑛 2𝑛 2𝑛 13 𝑏 + 𝑎𝑏 𝑎 + (𝑙) 1 [𝑏 + 𝑎𝑏 − 𝑎 − (𝑙)] − = 2𝑛 2𝑛 2𝑛 𝑎𝑏 + (𝑙) 𝑎 + 𝑏 1 [𝑎𝑏 + (𝑙) − 𝑎 − 𝑏] 𝐼𝑛𝑡𝑒𝑟𝑎𝑐𝑡𝑖𝑜𝑛 𝐸𝑓𝑓𝑒𝑐𝑡 𝐴𝐵: 𝐴𝐵 = − = 2𝑛 2𝑛 2𝑛 𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡 𝐶𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡𝑠 𝑎𝑟𝑒 𝑎𝑙𝑤𝑎𝑦𝑠 + 1 𝑜𝑟 − 1. 𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡𝐴 = 𝑎 + 𝑎𝑏 − 𝑏 − (𝑙) 𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡 𝐸𝑓𝑓𝑒𝑐𝑡 = 𝑛2𝑘−1 (𝐶𝑜𝑛𝑡𝑟𝑎𝑠𝑡)2 𝑆𝑢𝑚 𝑜𝑓 𝑠𝑞𝑢𝑎𝑟𝑒𝑠 𝑓𝑜𝑟 𝑎𝑛 𝑒𝑓𝑓𝑒𝑐𝑡, 𝑆𝑆 = 𝑛2𝑘 𝑌 = 𝛽0 + 𝛽1 𝑥1 + 𝜖 𝑒𝑓𝑓𝑒𝑐𝑡 𝑦̅+ − 𝑦̅− 𝐶𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡 𝑎𝑛𝑑 𝑒𝑓𝑓𝑒𝑐𝑡, 𝛽̂ = = 2 2 𝑀𝑎𝑖𝑛 𝐸𝑓𝑓𝑒𝑐𝑡 𝑜𝑓 𝐹𝑎𝑐𝑡𝑜𝑟 𝐵: 𝐵 = 𝑦̅𝐵+ − 𝑦̅𝐵− = 𝑆𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑒𝑟𝑟𝑜𝑟 𝑜𝑓 𝑎 𝑐𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡, 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑒𝑟𝑟𝑜𝑟 𝛽̂ = 𝑡 − 𝑠𝑡𝑎𝑡𝑖𝑠𝑡𝑖𝑐 𝑓𝑜𝑟 𝑎 𝑐𝑜𝑒𝑓𝑓𝑖𝑐𝑖𝑒𝑛𝑡, 𝑡 = 𝜎̂ 1 1 1 √ 𝑘−1 + 𝑘−1 = 𝜎̂√ 𝑘 2 𝑛2 𝑛2 𝑛2 𝛽̂ 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑒𝑟𝑟𝑜𝑟 𝛽̂ = (𝑦̅+ − 𝑦̅− )/2 1 𝑛2𝑘 𝜎̂√ 2k Factorial Designs for k 3 factors 1 [𝑎 + 𝑎𝑏 + 𝑎𝑐 + 𝑎𝑏𝑐 − (1) − 𝑏 − 𝑐 − 𝑏𝑐] 4𝑛 1 [𝑏 + 𝑎𝑏 + 𝑏𝑐 + 𝑎𝑏𝑐 − (1) − 𝑎 − 𝑐 − 𝑎𝑐] 𝐵 = 𝑦̅𝐵+ − 𝑦̅𝐵− = 4𝑛 1 [𝑐 + 𝑎𝑐 + 𝑏𝑐 + 𝑎𝑏𝑐 − (1) − 𝑎 − 𝑏 − 𝑎𝑏] 𝐶 = 𝑦̅𝐶+ − 𝑦̅𝐶− = 4𝑛 𝐴 = 𝑦̅𝐴+ − 𝑦̅𝐴− = 1 [𝑎𝑏𝑐 − 𝑏𝑐 + 𝑎𝑏 − 𝑏 − 𝑎𝑐 + 𝑐 − 𝑎 + (1)] 4𝑛 1 [(𝑎) − 𝑎 + 𝑏 − 𝑎𝑏 − 𝑐 + 𝑎𝑐 − 𝑏𝑐 + 𝑎𝑏𝑐] 𝐴𝐶 = 4𝑛 1 [(1) + 𝑎 − 𝑏 − 𝑎𝑏 − 𝑐 − 𝑎𝑐 + 𝑏𝑐 + 𝑎𝑏𝑐] 𝐵𝐶 = 4𝑛 1 [𝑎𝑏𝑐 − 𝑏𝑐 − 𝑎𝑐 + 𝑐 − 𝑎𝑏 + 𝑏 + 𝑎 − (1)] 𝐴𝐵𝐶 = 4𝑛 𝑌 = 𝛽0 + 𝛽1 𝑥1 + 𝛽2 𝑥2 + 𝛽12 𝑥1 𝑥2 + 𝜖 𝐴𝐵 = 2 2 𝑆𝑆𝐶𝑢𝑟𝑣𝑎𝑡𝑢𝑟𝑒 = 𝑛𝐹 𝑛𝐶 (𝑦̅𝐹 − 𝑦̅𝐶 ) = 𝑛𝐹 + 𝑛𝐶 𝑦̅𝐹 − 𝑦̅𝐶 1 1 √ + ( 𝑛𝐹 𝑛𝐶 ) 𝐹𝑖𝑟𝑠𝑡 𝑜𝑟𝑑𝑒𝑟 𝑚𝑜𝑑𝑒𝑙: 𝑌 = 𝛽0 + 𝛽1 𝑥1 + 𝛽2 𝑥2 + 𝛽𝑘 𝑥𝑘 + 𝜖 𝑘 𝑘 𝑆𝑒𝑐𝑜𝑛𝑑 𝑜𝑟𝑑𝑒𝑟 𝑚𝑜𝑑𝑒𝑙: 𝑌 = 𝛽0 + ∑ 𝛽𝑖 𝑥𝑖 + ∑ 𝛽𝑖𝑖 𝑥𝑖2 + ∑ ∑ 𝛽𝑖𝑗 𝑥𝑖 𝑥𝑗 + 𝜖 𝑖=1 𝑖=1 𝑖<𝑗 𝑘 𝑆𝑡𝑒𝑒𝑝𝑒𝑠𝑡 𝐴𝑠𝑐𝑒𝑛𝑡 𝑦̂ = 𝛽̂0 + ∑ 𝛽̂𝑖 𝑥𝑖 𝑘 𝑖=1 𝑘 𝐹𝑖𝑡𝑡𝑒𝑑 𝑆𝑒𝑐𝑜𝑛𝑑 𝑜𝑟𝑑𝑒𝑟 𝑚𝑜𝑑𝑒𝑙 𝑦̂ = 𝛽̂0 + ∑ 𝛽̂𝑖 𝑥𝑖 + ∑ 𝛽̂𝑖𝑖 𝑥𝑖2 + ∑ ∑ 𝛽̂𝑖𝑗 𝑥𝑖 𝑥𝑗 𝑖=1 𝑖=1 𝑖<𝑗 14 WAITING LINES AND QUEUING THEORY MODELS Single-Channel Model, Poisson Arrivals, Exponential Multichannel Model, Poisson Arrivals, Exponential Service Times (M/M/1) Service Times (M/M/m or M/M/s) m = number of channels open = mean number of arrivals per time period (arrival rate) = average arrival rate = mean number of customers or units served per time = average service rate at each channel period (service rate) The probability that there are zero customers in the The average number of customers or units in the system, system L 1 P0 = for m n n =m –1 1 1 m m L= + − n =0 n! m! m − The average time a customer spends in the system, W The average number of customers or units in the system 1 W= ( / )m − L= P + 2 0 (m – 1)!(m − ) The average number of customers in the queue, Lq 2 The average time a unit spends in the waiting line or Lq = being served, in the system ( − ) ( / ) m 1 L The average time a customer spends waiting in the W= P + = 2 0 (m – 1)!(m − ) queue, Wq The average number of customers or units in line Wq = waiting for service ( − ) The utilization factor for the system, (rho), the probability the service facility is being used = The percent idle time, P0, or the probability no one is in the system P0 = 1 − Pn>k The average number of customers or units in line waiting for service 1 L Wq = W − = q The average number of customers or units in line waiting for service (Utilization rate) The probability that the number of customers in the system is greater than k, Pn>k = Lq = L − = m k +1 𝜆 𝑛 𝐶 = ( ) = 𝜌𝑛 𝜇 𝑃0 = 1 − 𝜌 𝑃𝑛 = (1 − 𝜌)𝜌𝑛 Finite Population Model (M/M/1 with Finite Source) = mean arrival rate = mean service rate N = size of the population Probability that the system is empty Total service cost = (Number of channels) x (Cost per channel) Total service cost = mCs m = number of channels Cs = service cost (labor cost) of each channel Total waiting cost = (Total time spent waiting by all arrivals) x (Cost of waiting) 15 P0 = 1 N N! (N – n)! n n =0 Average length of the queue + Lq = N − (1 – P0 ) Average number of customers (units) in the system L = Lq + (1 – P0 ) = (Number of arrivals) x (Average wait per arrival)Cw = (W)Cw Total waiting cost (based on time in queue) = (Wq)Cw Total cost = Total service cost + Total waiting cost Total cost = mCs + WCw Total cost (based on time in queue) = mCs + WqCw Average waiting time in the queue Lq Wq = (N – L) Average time in the system 1 W = Wq + Probability of n units in the system n N! P for n = 0,1,..., N Pn = (N – n )! 0 Constant Service Time Model (M/D/1) Average length of the queue Lq = 2 2 ( − ) Average waiting time in the queue Wq = 2 ( − ) Little’s Flow Equations L = W (or W = L/) Lq = Wq (or Wq = Lq/) Average time in system = average time in queue + average time receiving service W = Wq + 1/ Average number of customers in the system L = Lq + Average time in the system 1 W = Wq + N(t) = Number of customers in queuing system at time t (t >= 0) Pn(t) = Probability of exactly n customers in queuing system at time t, given number at time 0. s = number of servers (parallel service channels) in queuing system n = mean arrival rate (expected number of arrivals per unit time) of new customers when n customers are in system = expected interarrival time n = mean service rate for overall system (expected number of customers completing service per unit time) when n customers are in system. Represents combined rate at which all busy servers (those serving customers) achieve service completions = expected service time Utilization factor = = /(s) Steady-state condition 𝑃𝑛 = 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑒𝑥𝑎𝑐𝑡𝑙𝑦 𝑛 𝑐𝑢𝑠𝑡𝑜𝑚𝑒𝑟𝑠𝑖𝑛 𝑞𝑢𝑒𝑢𝑖𝑛𝑔 𝑠𝑦𝑠𝑡𝑒𝑚 16 ∞ 𝐿 = 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑐𝑢𝑠𝑡𝑜𝑚𝑒𝑟𝑠 𝑖𝑛 𝑞𝑢𝑒𝑢𝑖𝑛𝑔 𝑠𝑦𝑠𝑡𝑒𝑚 = ∑ 𝑛 𝑃𝑛 𝑛=0 ∞ 𝐿𝑞 = 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑞𝑢𝑒𝑢𝑒 𝑙𝑒𝑛𝑔𝑡ℎ (𝑒𝑥𝑐𝑙𝑢𝑑𝑒𝑠 𝑐𝑢𝑠𝑡𝑜𝑚𝑒𝑟𝑠 𝑏𝑒𝑖𝑛𝑔 𝑠𝑒𝑟𝑣𝑒𝑑) = ∑(𝑛 − 𝑠) 𝑃𝑛 𝑛=𝑠 𝑊 = 𝐸(𝑤𝑎𝑖𝑡𝑖𝑛𝑔 𝑡𝑖𝑚𝑒 𝑖𝑛 𝑠𝑦𝑠𝑡𝑒𝑚 (𝑖𝑛𝑐𝑙𝑢𝑑𝑒𝑠 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑡𝑖𝑚𝑒)𝑓𝑜𝑟 𝑒𝑎𝑐ℎ 𝑖𝑛𝑑𝑖𝑣𝑖𝑑𝑢𝑎𝑙 𝑐𝑢𝑠𝑡𝑜𝑚𝑒𝑟) 𝑊𝑞 = 𝐸(𝑤𝑎𝑖𝑡𝑖𝑛𝑔 𝑡𝑖𝑚𝑒 𝑖𝑛 𝑞𝑢𝑒𝑢𝑒 (𝑒𝑥𝑐𝑙𝑢𝑑𝑒𝑠 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑡𝑖𝑚𝑒)𝑓𝑜𝑟 𝑒𝑎𝑐ℎ 𝑖𝑛𝑑𝑖𝑣𝑖𝑑𝑢𝑎𝑙 𝑐𝑢𝑠𝑡𝑜𝑚𝑒𝑟) 𝐿𝑖𝑡𝑡𝑙𝑒 ′ 𝑠 𝑓𝑜𝑟𝑚𝑢𝑙𝑎: 𝐿 = 𝜆𝑊 𝐿𝑞 = 𝜆𝑊𝑞 1 𝑊 = 𝑊𝑞 + 𝜇 Impact of Exponential distribution on Queuing Model 𝛼𝑒 −𝛼𝑡 𝑓𝑜𝑟 𝑡 ≥ 0 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙 𝐷𝑖𝑠𝑡𝑟𝑖𝑏𝑡𝑢𝑖𝑜𝑛′ 𝑠 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑑𝑒𝑛𝑠𝑖𝑡𝑦 𝑓𝑢𝑛𝑐𝑡𝑖𝑜𝑛, 𝑓𝑇 (𝑡) = { 0 𝑓𝑜𝑟 𝑡 < 0 −𝛼𝑡 −𝛼𝑡 𝐶𝑢𝑚𝑢𝑙𝑎𝑡𝑖𝑣𝑒 𝑃𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑖𝑒𝑠: 𝑃{𝑇 ≤ 𝑡} = 1 − 𝑒 ; 𝑃{𝑇 > 𝑡} = 𝑒 𝑃𝑟𝑜𝑝𝑒𝑟𝑡𝑦 1: 𝑓𝑇 (𝑡) 𝑖𝑠 𝑎 𝑠𝑡𝑟𝑖𝑐𝑡𝑙𝑦 𝑑𝑒𝑐𝑟𝑒𝑎𝑠𝑖𝑛𝑔 𝑓𝑢𝑛𝑐𝑡𝑖𝑜𝑛 𝑜𝑓 𝑡 (𝑡 ≥ 0); 𝑠𝑜 𝑃{0 ≤ 𝑇 ≤ ∆𝑡} > 𝑃{𝑡 ≤ 𝑇 ≤ 𝑡 + ∆𝑡} 𝑃𝑟𝑜𝑝𝑒𝑟𝑡𝑦 2: 𝐿𝑎𝑐𝑘 𝑜𝑓 𝑚𝑒𝑚𝑜𝑟𝑦; 𝑠𝑜 𝑃{𝑇 > 𝑡 + ∆𝑡 | 𝑇 > ∆𝑡} = 𝑃{𝑇 > 𝑡} 𝑃𝑟𝑜𝑝𝑒𝑟𝑡𝑦 3: 𝑀𝑖𝑛𝑖𝑚𝑢𝑚 𝑜𝑓 𝑠𝑒𝑣𝑒𝑟𝑎𝑙 𝑖𝑛𝑑𝑒𝑝𝑒𝑛𝑑𝑒𝑛𝑡 𝑒𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙 𝑟𝑎𝑛𝑑𝑜𝑚 𝑣𝑎𝑟𝑖𝑎𝑏𝑙𝑒𝑠 ℎ𝑎𝑠 𝑎𝑛 𝑒𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛. 𝑈 = 𝑚𝑖𝑛 {𝑇1 , 𝑇2 , … , 𝑇𝑛 } 𝑛 𝑃{𝑈 > 𝑇} = 𝑒𝑥𝑝 (− ∑ 𝛼𝑖 𝑡) 𝑛 𝑖=1 𝛼 = ∑ 𝛼𝑖 𝑖=1 𝑃𝑟𝑜𝑝𝑒𝑟𝑡𝑦 4: 𝑅𝑒𝑙𝑎𝑡𝑖𝑜𝑛𝑠ℎ𝑖𝑝 𝑡𝑜 𝑃𝑜𝑖𝑠𝑠𝑜𝑛 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑛 −𝛼𝑡 (𝛼𝑡) 𝑒 𝑃{𝑋(𝑡) = 𝑛} = , 𝑓𝑜𝑟 𝑛 = 0, 1, 2, … ; 𝐴𝑙𝑠𝑜 𝑚𝑒𝑎𝑛 𝐸{𝑋(𝑡)} = 𝛼𝑡; 𝑤ℎ𝑒𝑟𝑒 𝛼 𝑖𝑠 𝑚𝑒𝑎𝑛 𝑟𝑎𝑡𝑒 𝑛! 𝑋(𝑡) = 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑜𝑐𝑐𝑢𝑟𝑟𝑒𝑛𝑐𝑒𝑠 𝑏𝑦 𝑡𝑖𝑚𝑒 𝑡 (𝑡 ≥ 0) 𝑃𝑟𝑜𝑝𝑒𝑟𝑡𝑦 5: 𝐹𝑜𝑟 𝑎𝑙𝑙 𝑝𝑜𝑠𝑖𝑡𝑖𝑣𝑒 𝑣𝑎𝑙𝑢𝑒𝑠 𝑜𝑓 𝑡, 𝑃{𝑇 ≤ 𝑡 + ∆𝑡|𝑇 > 𝑡} ≈ 𝛼∆𝑡, 𝑓𝑜𝑟 𝑠𝑚𝑎𝑙𝑙 ∆𝑡 𝐵𝑖𝑟𝑡ℎ − 𝑎𝑛𝑑 − 𝐷𝑒𝑎𝑡ℎ 𝑃𝑟𝑜𝑐𝑒𝑠𝑠 𝜆𝑛−1 𝜆𝑛−2 ⋯ 𝜆0 𝐶𝑛 = , 𝑓𝑜𝑟 𝑛 = 1, 2, … 𝜇𝑛 𝜇𝑛−1 ⋯ 𝜇1 𝐹𝑜𝑟 𝑛 = 0, 𝐶𝑛 = 1 𝑃𝑛 = 𝐶𝑛 𝑃0 , 𝑓𝑜𝑟 𝑛 = 0, 1, 2, … ∞ ∑ 𝑃𝑛 = 1 𝑛=0 ∞ −1 𝑃0 = (∑ 𝐶𝑛 ) 𝑛=0 ∞ 𝐴𝑣𝑒𝑟𝑎𝑔𝑒 𝑎𝑟𝑟𝑖𝑣𝑎𝑙 𝑟𝑎𝑡𝑒 𝑜𝑣𝑒𝑟 𝑡ℎ𝑒 𝑙𝑜𝑛𝑔 𝑟𝑢𝑛, 𝜆̅ = ∑ 𝜆𝑛 𝑃𝑛 T1, T2, … be independent service-time random variables having an exponential distribution with parameter, and let Sn+1 = T1 + T2 + …+ Tn+1, for n = 0, 1, 2, … 𝑛=0 M/M/s, with s > 1 17 Sn+1 represents the conditional waiting time given n customers already in the system. Sn+1 is known to have an Erlang distribution. ∞ 𝑃{𝐸(𝑊) > 𝑡} = ∑ 𝑃𝑛 𝑃{𝑆𝑛+1 > 𝑡} = 𝑒 −𝜇(1−𝜌)𝑡 , 𝑓𝑜𝑟 𝑡 𝑛=0 ≥0 ∞ 𝜆 𝑛 𝜇 ( ) 𝐶𝑛 = 𝜆 𝑠 𝜇 ( ) { 𝑠! 𝜆 (𝑠𝜇) = 𝑠−1 𝜆 𝑛 𝜇 𝑠!𝑠𝑛−𝑠 ( ) 𝑓𝑜𝑟 𝑛 = 𝑠, 𝑠 + 1, … (𝜆/𝜇)𝑛 (𝜆/𝜇)𝑠 1 𝑃0 = 1⁄[∑ + ] 𝑛! 𝑠! 1 − 𝜆/(𝑠𝜇) 𝑛=0 𝑃{𝐸(𝑊𝑞 ) > 𝑡} = ∑ 𝑃𝑛 𝑃{𝑆𝑛 > 𝑡} = 𝜌𝑒 −𝜇(1−𝜌)𝑡 , 𝑓𝑜𝑟 𝑡 𝑛=0 ≥0 𝑃{𝐸(𝑊𝑞 ) > 𝑡|𝐸(𝑊𝑞 ) > 0} = 𝑒 −𝜇(1−𝜌)𝑡 , 𝑓𝑜𝑟 𝑡 ≥ 0 𝑓𝑜𝑟 𝑛 = 1, 2, … , 𝑠 𝑛! 𝑛−𝑠 𝑃𝑛 = 𝜆 𝑛 (𝜇 ) 𝑃 𝑖𝑓 0 ≤ 𝑛 ≤ 𝑠 𝑛! 0 𝜆 𝑛 (𝜇 ) 𝑖𝑓 𝑛 ≥ 𝑠 {𝑠! 𝑠 𝑛−𝑠 𝑃0 𝑃0 (𝜆/𝜇)𝑠 𝜌 𝐿𝑞 = 𝑠! (1 − 𝜌)2 𝐿𝑞 𝑊𝑞 = 𝜆 1 𝑊 = 𝑊𝑞 + 𝜇 𝜆 𝐿 = 𝐿𝑞 + 𝜇 𝑃{𝐸(𝑊) > 𝑡} = 𝑒 −𝜇𝑡 [1 𝜆 −𝜇𝑡(𝑠−1− ) 𝜇 𝑃0 (𝜆/𝜇)𝑠 1 − 𝑒 ( )] + 𝑠! (1 − 𝜌) 𝑠 − 1 − 𝜆/𝜇 𝑃{𝐸(𝑊𝑞 ) > 𝑡} = (1 − 𝑃{𝐸(𝑊𝑞 ) = 0})𝑒 −𝑠𝜇(1−𝜌)𝑡 𝑠−1 𝑃{𝐸(𝑊𝑞 ) = 0} = ∑ 𝑃𝑛 𝑛=0 Finite Queue Variation of M/M/s (M/M/s/K model) Finite Queue: Number of customers in the system cannot exceed a specified number, K. Queue capacity is K-s 𝜆 𝑓𝑜𝑟 𝑛 = 0, 1, 2, … , 𝐾 − 1 𝜆𝑛 = { 0 𝑓𝑜𝑟 𝑛 ≥ 𝐾 M/M/1/K (𝜆/𝜇)𝑛 = 𝜌𝑛 𝑓𝑜𝑟 𝑛 = 1, 2, … , 𝐾 𝐶𝑛 = { 0 𝑓𝑜𝑟 𝑛 > 𝐾 1−𝜌 𝑃0 = 1 − 𝜌𝐾+1 1−𝜌 𝑃𝑛 = 𝜌𝑛 , 𝑓𝑜𝑟 𝑛 = 0, 1, 2, … , 𝐾 𝐾+1 1−𝜌 𝜌 (𝐾 + 1)𝜌𝐾+1 𝐿= − 1−𝜌 1 − 𝜌𝐾+1 𝐿𝑞 = 𝐿 − (1 − 𝑃0 ) Finite Calling Population variation of M/M/s M/M/1 with finite population 𝑁! 𝜆 𝑛 ( ) 𝑓𝑜𝑟 𝑛 ≤ 𝑁 𝐶𝑛 = {(𝑁 − 𝑛)! 𝜇 0 𝑓𝑜𝑟 𝑛 > 𝑁 𝑁 𝑁! 𝜆 𝑛 ( ) ] 𝑃0 = 1⁄∑ [ (𝑁 − 𝑛 )! 𝜇 𝑛=0 𝑁! 𝜆 𝑛 ( ) 𝑃0 , 𝑖𝑓 𝑛 = 1, 2, … , 𝑁 𝑃𝑛 = (𝑁 − 𝑛)! 𝜇 𝑁 𝐿𝑞 = ∑(𝑛 − 1)𝑃0 = 𝑁 − 𝑛=1 𝜆+𝜇 (1 − 𝑃0 ) 𝜇 𝜇 𝐿 = 𝑁 − (1 − 𝑃0 ) 𝜆 𝐿 𝑊= 𝜆̅ 18 𝐿𝑞 𝜆̅ 𝐿 𝑊= 𝜆̅ 𝜆̅ = ∑ 𝜆𝑛 𝑃𝑛 = 𝜆(𝑁 − 𝐿) 𝜆̅ = ∑ 𝜆𝑛 𝑃𝑛 = 𝜆(1 − 𝑃𝐾 ) M/M/s with finite population and s > 1 𝑊𝑞 = ∞ 𝑊𝑞 = ∞ 𝑛=0 𝐶𝑛 = 𝑛=0 M/M/s/K 𝐶𝑛 (𝜆/𝜇)𝑛 𝑓𝑜𝑟 𝑛 = 1, 2, … , 𝑠 𝑛! 𝑛−𝑠 (𝜆/𝜇)𝑛 = (𝜆/𝜇)𝑠 𝜆 ( ) = 𝑠! 𝑠𝜇 𝑠! 𝑠 𝑛−𝑠 𝑓𝑜𝑟 𝑛 = 𝑠, 𝑠 + 1, … , 𝐾 { 0 𝑓𝑜𝑟 𝑛 > 𝐾 𝑛 𝜆 𝑠 𝐾 𝑠 (𝜆 ) (𝜇 ) 𝜆 𝑛−𝑠 𝜇 𝑃0 = 1⁄[∑ + ∑ ( ) ] 𝑛! 𝑠! 𝑠𝜇 𝑛=0 𝑛 𝑛=𝑠+1 𝜆 (𝜇 ) 𝑁! 𝜆 𝑛 ( ) (𝑁−𝑛)!𝑛! 𝜇 𝑓𝑜𝑟 𝑛 = 0, 1, 2, … , 𝑠 𝑁! 𝜆 𝑛 ( ) (𝑁−𝑛)!𝑠!𝑠𝑛−𝑠 𝜇 𝑓𝑜𝑟 𝑛 = 𝑠, 𝑠 + 1, … , 𝐾 0 𝑓𝑜𝑟 𝑛 > 𝐾 { 𝑁! 𝜆 𝑛 𝑁! 𝜆 𝑛 𝑁 𝑃0 = 1⁄[∑𝑠−1 𝑛=0 (𝑁−𝑛)!𝑛! (𝜇 ) + ∑𝑛=𝑠 (𝑁−𝑛)!𝑠!𝑠𝑛−𝑠 (𝜇 ) ] 𝑁! 𝜆 𝑛 ( ) 𝑃0 (𝑁 − 𝑛)! 𝑛! 𝜇 𝑃𝑛 = 𝑁! 𝜆 𝑛 ( ) 𝑃0 (𝑁 − 𝑛)! 𝑠! 𝑠 𝑛−𝑠 𝜇 0 𝑖𝑓 𝑛 > 𝑁 { 𝑖𝑓 0 ≤ 𝑛 ≤ 𝑠 𝑖𝑓 𝑠 ≤ 𝑛 ≤ 𝑁 𝑁 𝑃 𝑓𝑜𝑟 𝑛 = 1, 2, … , 𝑠 𝑛! 0 𝑃𝑛 = 𝜆 𝑛 (𝜇 ) 𝑓𝑜𝑟 𝑛 = 𝑠, 𝑠 + 1, … , 𝐾 𝑃 𝑠! 𝑠 𝑛−𝑠 0 {0 𝑓𝑜𝑟 𝑛 > 𝐾 𝑃0 (𝜆/𝜇)𝑠 𝜌 𝐿𝑞 = [1 − 𝜌𝐾−𝑠 − (𝐾 − 𝑠)𝜌𝐾−𝑠 (1−)] 𝑠! (1 − 𝜌)2 𝑠−1 𝐿𝑞 = ∑(𝑛 − 𝑠)𝑃𝑛 𝑛=𝑠 𝑠−1 𝐿𝑞 𝑊𝑞 = 𝜆̅ 𝐿 𝑊= 𝜆̅ M/G/1 Model 𝑃0 = 1 − 𝜌 𝑛=0 𝑃𝑜𝑙𝑙𝑎𝑐𝑧𝑒𝑘 − 𝐾ℎ𝑖𝑛𝑡𝑐ℎ𝑖𝑛𝑒 𝐹𝑜𝑟𝑚𝑢𝑙𝑎, 𝐿𝑞 = 𝑠−1 𝐿 = ∑ 𝑛𝑃𝑛 + 𝐿𝑞 + 𝑠 (1 − ∑ 𝑃𝑛 ) 𝑛=0 𝑠−1 𝐿 = ∑ 𝑛𝑃𝑛 + 𝐿𝑞 + 𝑠 (1 − ∑ 𝑃𝑛 ) 𝑛=0 𝐿𝑞 𝜆̅ ∞ 𝐿 𝑊= 𝜆̅ 𝐿𝑞 𝑊𝑞 = 𝜆̅ 𝑛=0 𝜆̅ = ∑ 𝜆𝑛 𝑃𝑛 = 𝜆(𝑁 − 𝐿) 𝑛=0 M/D/s Model 𝜆2 𝜎 2 + 𝜌2 2(1 − 𝜌) 𝐿 = 𝜌 + 𝐿𝑞 𝐿𝑞 𝑊𝑞 = 𝜆 1 𝑊 = 𝑊𝑞 + 𝜇 𝐹𝑜𝑟 𝑎𝑛𝑦 𝑓𝑖𝑥𝑒𝑑 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑡𝑖𝑚𝑒 1⁄𝜇 , 𝑛𝑜𝑡𝑖𝑐𝑒 𝑡ℎ𝑎𝑡 𝐿𝑞 , 𝐿, 𝑊𝑞 𝑎𝑛𝑑 𝑊 𝑎𝑙𝑙 𝑖𝑛𝑐𝑟𝑒𝑎𝑠𝑒 𝑎𝑠 𝜎 2 𝑖𝑠 𝑖𝑛𝑐𝑟𝑒𝑎𝑠𝑒𝑑. M/Ek/s Model 𝐸𝑟𝑙𝑎𝑛𝑔 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛′ 𝑠 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑑𝑒𝑛𝑠𝑖𝑡𝑦 𝑓𝑢𝑛𝑐𝑡𝑖𝑜𝑛 𝐹𝑜𝑟 𝑀/𝐷/1 𝑚𝑜𝑑𝑒𝑙, 𝐿𝑞 = 𝜌2 2(1 − 𝜌) Nonpreemptive Priorities Model 𝑆𝑡𝑒𝑎𝑑𝑦 − 𝑠𝑡𝑎𝑡𝑒 𝑒𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑤𝑒𝑖𝑔ℎ𝑡𝑖𝑛𝑔 𝑡𝑖𝑚𝑒 𝑖𝑛 𝑡ℎ𝑒 𝑠𝑦𝑠𝑡𝑒𝑚 (𝑖𝑛𝑐𝑙𝑢𝑑𝑖𝑛𝑔 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑡𝑖𝑚𝑒)𝑓𝑜𝑟 𝑎 19 (𝜇𝑘)𝑘 𝑘−1 −𝑘𝜇𝑡 𝑡 𝑒 , 𝑓𝑜𝑟 𝑡 ≥ 0 (𝑘 − 1)! 𝜇 𝑎𝑛𝑑 𝑘 𝑎𝑟𝑒 𝑠𝑡𝑟𝑖𝑐𝑡𝑙𝑦 𝑝𝑜𝑠𝑖𝑡𝑖𝑣𝑒 𝐴𝑁𝐷 𝑘 𝑚𝑢𝑠𝑡 𝑏𝑒 𝑎𝑛 𝑖𝑛𝑡𝑒𝑔𝑒𝑟 𝐸𝑥𝑐𝑙𝑢𝑑𝑖𝑛𝑔 𝑖𝑛𝑡𝑒𝑔𝑒𝑟 𝑟𝑒𝑠𝑡𝑟𝑖𝑐𝑡𝑖𝑜𝑛, 𝐸𝑟𝑙𝑎𝑛𝑔 𝑖𝑠 𝑠𝑎𝑚𝑒 𝑎𝑠 𝑔𝑎𝑚𝑚𝑎 1 𝑀𝑒𝑎𝑛 = 𝜇 1 1 𝑆𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑑𝑒𝑣𝑖𝑎𝑡𝑖𝑜𝑛 = √𝑘 𝜇 T1, T2, …, Tk are k independent random variables with an identical exponential distribution whose mean is 1/(k). T = T1 + T2 + …+ Tk, has an Erlang distribution with parameters and k. Exponential and degenerate (constant) are special cases of Erlang distribution with k=1 and k= respectively. M/Ek/1 Model 2 𝜆 /(𝑘𝜇 2 ) + 𝜌2 1 + 𝑘 𝜆2 𝐿𝑞 = = 2(1 − 𝜌) 2𝑘 𝜇(𝜇 − 𝜆) 𝐿 = 𝜆𝑊 1+𝑘 𝜆 𝑊𝑞 = 2𝑘 𝜇(𝜇 − 𝜆) 1 𝑊 = 𝑊𝑞 + 𝜇 𝑓(𝑡) = 𝑚𝑒𝑚𝑏𝑒𝑟 𝑜𝑓 𝑝𝑟𝑖𝑜𝑟𝑖𝑡𝑦 𝑐𝑙𝑎𝑠𝑠 𝑘, 𝑊𝑘 1 1 𝑊𝑘 = + , 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 𝐴𝐵𝑘−1 𝐵𝑘 𝜇 𝑠−1 𝑠𝜇 − 𝜆 𝑟𝑗 𝑊ℎ𝑒𝑟𝑒 𝐴 = 𝑠! ∑ + 𝑠𝜇 𝑟𝑠 𝑗! 𝑗=0 𝐵0 = 1 ∑𝑘𝑖=1 𝜆𝑖 𝐵𝑘 = 1 − 𝑠𝜇 𝑠 = 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑠𝑒𝑟𝑣𝑒𝑟𝑠 𝜇 = 𝑚𝑒𝑎𝑛 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑟𝑎𝑡𝑒 𝑝𝑒𝑟 𝑏𝑢𝑠𝑦 𝑠𝑒𝑟𝑣𝑒𝑟 𝜆𝑖 = 𝑚𝑒𝑎𝑛 𝑎𝑟𝑟𝑖𝑣𝑎𝑙 𝑟𝑎𝑡𝑒 𝑓𝑜𝑟 𝑝𝑟𝑖𝑜𝑟𝑖𝑡𝑦 𝑐𝑙𝑎𝑠𝑠 𝑖 𝑁 𝜆 = ∑ 𝜆𝑖 𝑖=1 𝑟= 𝑘 𝜆 𝜇 ∑ 𝜆𝑖 < 𝑠𝜇 𝑖=1 𝐿𝑘 = 𝜆𝑘 𝑊𝑘 , 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 𝐴 = 𝑠! 𝑭𝒐𝒓 𝒔𝒊𝒏𝒈𝒍𝒆 𝒔𝒆𝒓𝒗𝒆𝒓, 𝒔 = 𝟏 𝐴 = 𝜇 2 ⁄𝜆 𝑾𝒊𝒕𝒉 𝒅𝒊𝒇𝒇𝒆𝒓𝒆𝒏𝒕 𝒆𝒙𝒑𝒐𝒏𝒆𝒏𝒕𝒊𝒂𝒍 𝒕𝒊𝒎𝒆𝒔 𝜇𝑘 = 𝑚𝑒𝑎𝑛 𝑠𝑒𝑟𝑣𝑖𝑐𝑒 𝑟𝑎𝑡𝑒 𝑓𝑜𝑟 𝑝𝑟𝑖𝑜𝑟𝑖𝑡𝑦 𝑐𝑙𝑎𝑠𝑠 𝑘, 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 𝑎𝑘 1 𝑊𝑘 = + , 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 𝑏𝑘−1 𝑏𝑘 𝜇𝑘 𝑘 𝑊ℎ𝑒𝑟𝑒 𝑎𝑘 = ∑ 𝑏0 = 1 𝑖=1 𝑘 𝑏𝑘 = 1 − ∑ 𝑘 ∑ 𝑖=1 Jackson Networks With m service facilities where facility i (i=1, 2, .,., m) 1. Infinite Queue 2. Customers arriving from outside the system according to a Poisson input process with parameters ai 3. si servers with an exponential service-time distribution with parameter 𝑖=1 𝜆𝑖 𝜇𝑖2 𝜆𝑖 𝜇𝑖 𝜆𝑖 <1 𝜇𝑖 Preemptive Priorities Model 1/𝜇 𝐹𝑜𝑟 𝑠 = 1, 𝑊𝑘 = , 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 𝐵𝑘−1 𝐵𝑘 𝐿𝑘 = 𝜆𝑘 𝑊𝑘 , 𝑓𝑜𝑟 𝑘 = 1, 2, … , 𝑁 20 A Customer leaving facility i is routed next to facility j (j= 1, 2, …, m) with probability pij or departs the system with probability 𝑚 𝑞𝑖 = 1 − ∑ 𝑝𝑖𝑗 𝑗=1 Jackson network behaves as if it were an independent M/M/s queueing system with arrival rate 𝑚 𝜆𝑗 = 𝑎𝑗 + ∑ 𝜆𝑖 𝑝𝑖𝑗 𝑖=1 𝑤ℎ𝑒𝑟𝑒 𝑠𝑗 𝜇𝑗 > 𝜆𝑗 𝜆𝑖 𝜌𝑖 = 𝑠𝑖 𝜇𝑖 21 MARKOV ANALYSIS (i) = vector of state probabilities for period Pij = conditional probability of being in state j in the future given the current state of i i = (1, 2, 3, … , n) P11 P12 P1n P where P22 P2 n 21 P= n = number of states 1, 2, … , n = probability of being in state 1, state 2, …, state n Pm1 Pm 2 Pmn For any period n we can compute the state probabilities for period n + 1 (n + 1) = (n)P Fundamental Matrix F = (I – B)–1 Inverse of Matrix a b P= c d P -1 a b = c d −1 d = r −c r − b r a r r = ad – bc Partition of Matrix for absorbing states I O P= A B I = identity matrix O = a matrix with all 0s Equilibrium condition = P M represent the amount of money that is in each of the nonabsorbing states M = (M1, M2, M3, … , Mn) n = number of nonabsorbing states M1 = amount in the first state or category M2 = amount in the second state or category Mn = amount in the nth state or category Computing lambda and the consistency index −n CI = n −1 Consistency Ratio CI CR = RI Stochastic process {Xt} (t = 0, 1, …) is a Markov chain if it has the Markovian property. Stochastic process {Xt} is said to have the Markovian property if P{Xt+1 = j | X0 = k0, X1 = k1, …, Xt-1 = kt-1, Xt = i} = P{Xt+1=j|Xt=i}, for t=0,1, … and every sequence i, j, k0, k1, …, kt-1. Pij = P{Xt+1 = j | Xt = i} n-step transition probabilities: (𝑛) 𝑃𝑖𝑗 = 𝑃{𝑋𝑡+𝑛 = 𝑗|𝑋𝑡 = 𝑖} (𝑛) 𝑃𝑖𝑗 ≥ 0, 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑖 𝑎𝑛𝑑 𝑗; 𝑛 = 0, 1, 2, … 𝑀 (𝑛) ∑ 𝑃𝑖𝑗 = 1 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑖; 𝑛 = 0, 1, 2, … 𝑗=0 n-step transition matrix: 𝑆𝑡𝑎𝑡𝑒 0 1 … 𝑀 (𝑛) (𝑛) (𝑛) 𝑃00 𝑃01 … 𝑃0𝑀 0 (𝑛) (𝑛) (𝑛) 𝑃(𝑛) = 1 𝑃11 𝑃10 … 𝑃1𝑀 ⋮ … … … … (𝑛) (𝑛) (𝑛) 𝑀 [𝑃𝑀0 𝑃𝑀1 … 𝑃𝑀𝑀 ] 𝐶ℎ𝑎𝑝𝑚𝑎𝑛 − 𝐾𝑜𝑙𝑚𝑜𝑔𝑜𝑟𝑜𝑣 𝐸𝑞𝑢𝑎𝑡𝑖𝑜𝑛𝑠 22 𝑀 (𝑛) 𝑃𝑖𝑗 (𝑚) = ∑ 𝑃𝑖𝑘 (𝑛−𝑚) 𝑃𝑘𝑗 , 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑖 = 0, 1, … , 𝑀; 𝑗 = 0, 1, … , 𝑀; 𝑚 = 1, 2, … , 𝑛 − 1; 𝑛 = 𝑚 + 1, 𝑚 + 2, … 𝑘=0 𝑃 (𝑛) = 𝑃𝑃(𝑛−1) = 𝑃 (𝑛−1) 𝑃 = 𝑃𝑃𝑛−1 = 𝑃𝑛−1 𝑃 = 𝑃𝑛 (𝑛) (𝑛) (𝑛) 𝑈𝑛𝑐𝑜𝑛𝑑𝑖𝑡𝑖𝑜𝑛𝑎𝑙 𝑠𝑡𝑎𝑡𝑒 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑖𝑒𝑠: 𝑃{𝑋𝑛 = 𝑗} = 𝑃{𝑋0 = 0}𝑃0𝑗 + 𝑃{𝑋0 = 1}𝑃1𝑗 + ⋯ + 𝑃{𝑋0 = 𝑀}𝑃𝑀𝑗 (𝑛) lim 𝑝 𝑛→∞ 𝑖𝑗 𝑛 = 𝜋𝑗 > 0 1 (𝑘) lim ( ∑ 𝑝𝑖𝑗 ) = 𝜋𝑗 𝑛→∞ 𝑛 𝑘=1 𝑀 𝜋𝑗 = ∑ 𝜋𝑖 𝑝𝑖𝑗 , 𝑓𝑜𝑟 𝑗 = 0, 1, … , 𝑀 𝑖=0 𝑀 ∑ 𝜋𝑗 = 1 𝑛 𝑗=0 𝑀 1 lim 𝐸 [ ∑ 𝐶(𝑋𝑡 )] = ∑ 𝜋𝑗 𝐶(𝑗) 𝑛→∞ 𝑛 𝑡=1 𝑗=0 (2) (1) 𝐹𝑖𝑟𝑠𝑡 𝑃𝑎𝑠𝑠𝑎𝑔𝑒 𝑇𝑖𝑚𝑒 (1) (1) 𝑓𝑖𝑗 = 𝑝𝑖𝑗 = 𝑝𝑖𝑗 𝑓𝑖𝑗 = ∑ 𝑝𝑖𝑘 𝑓𝑘𝑗 𝑘≠𝑗 (𝑛) 𝑓𝑖𝑗 (𝑛−1) = ∑ 𝑝𝑖𝑘 𝑓𝑘𝑗 ∞ 𝑘≠𝑗 (𝑛) ∑ 𝑓𝑖𝑗 ≤1 𝑛=1 𝐸𝑥𝑝𝑒𝑐𝑡𝑒𝑑 𝑓𝑖𝑟𝑠𝑡 𝑝𝑎𝑠𝑠𝑎𝑔𝑒 𝑡𝑖𝑚𝑒 𝑓𝑟𝑜𝑚 𝑠𝑡𝑎𝑡𝑒 𝑖 𝑡𝑜 𝑠𝑡𝑎𝑡𝑒 𝑗 ∞ (𝑛) ∞ 𝑖𝑓 ∑ 𝑓𝑖𝑗 𝜇𝑖𝑗 = ∞ ∑ {𝑛=1 ∞ (𝑛) 𝑖𝑓 ∑ 𝑓𝑖𝑗 𝑛=1 𝑛=1 (𝑛) 𝑛𝑓𝑖𝑗 <1 ∞ (𝑛) 𝑖𝑓 ∑ 𝑓𝑖𝑗 =1 𝑛=1 = 1, 𝑡ℎ𝑒𝑛 𝜇𝑖𝑗 = 1 + ∑ 𝑝𝑖𝑘 𝜇𝑘𝑗 𝐴𝑏𝑠𝑜𝑟𝑏𝑖𝑛𝑔 𝑆𝑡𝑎𝑡𝑒𝑠 𝑘≠𝑗 𝑀 𝑓𝑖𝑘 = ∑ 𝑝𝑖𝑗 𝑓𝑗𝑘 , 𝑓𝑜𝑟 𝑖 = 0, 1, … , 𝑀, 𝑗=0 𝑆𝑢𝑏𝑗𝑒𝑐𝑡 𝑡𝑜 𝑡ℎ𝑒 𝑐𝑜𝑛𝑑𝑖𝑡𝑖𝑜𝑛𝑠 𝑓𝑘𝑘 = 1 𝑓𝑖𝑘 = 0, 𝑖𝑓 𝑠𝑡𝑎𝑡𝑒 𝑖 𝑖𝑠 𝑟𝑒𝑐𝑢𝑟𝑟𝑒𝑛𝑡 𝑎𝑛𝑑 𝑖 ≠ 𝑘 23 Name When to Use Approximations / Conditions Probability Mass function, Mean and Variance E(X) is Expected Value = Mean; Xi = random variable’s possible values; P(Xi) = Probability of each of the random variable’s possible values n n i =1 i =1 E (X ) = = X i P(X i ) Variance = 2 = [X i − E (X)]2 P(X i ) Standard Deviation = Variance Uniform (Discrete) Equal probability Finite number of possible values Bernoulli Trials: • Each trial is independent • Probability of success in a trial is constant • Only two possible outcomes Unknown: Number of successes Known: Number of trials → Number of trials that result in a success (b + a) 2 2 = Variance = (b − a + 1) 2 − 1 12 If n is large (np > 5, n(1-p) > 5), approximate binomial to normal. P(X x) = P(X x+0.5) P(x X) = P(x-0.5 X) If n is large & p is small, approximate to Poisson as = np Binomial expansion (a + b)n = Negative Binomial (Discrete) Hypergeometri c (Discrete) Poisson (Discrete) Probability of a success in x trials = n! p x q n− x x!(n − x)! x = 0,1,…,n, 0 p 1, n = 1,2,… n n! nC x = C xn = = x x!(n − x)! Expected value (mean) E(X) = = np Variance = V(X) = = np(1 – p) n r n−r p q k =0 r Bernoulli trial; Memoryless f ( x) = (1 − p) x −1 p x = 1,2,…,n, 0 p 1 → Number of trails until first success Expected value (mean) = E(X) = = 1/p Variance = V(X) = = (1 – p)/p2 Unknown: Number of trials x − 1 x−r (1 − p ) p r x = r,r+1,r+2,…, 0 p 1 f (x ) = Known: Number of success r − 1 → Number of trials required to obtain r successes E(X) = = r/p = Variance = r(1-p)/p2 n Geometric (Discrete) i =1 x For a series of n values, f(x) = 1/ n, a b For a range that starts from a and ends with b (a, a+1, a+2, …, b) and a b = Binomial / Bernoulli (Discrete) n Cumulative F(x) = P( X x) = f ( xi ) ; xi Trials are not independent Without replacement → Number of success in the sample Poisson Process: • Probability of more than one event in a subinterval is zero. • Probability of one event in a subinterval is constant & proportional to length of subinterval • Event in each subinterval is independent → Number of events in the interval V(X) = V(X) of binomial * ((N-n)/(N1)) where ((N-n)/(N-1)) is called finite population correction factor n << N or (n/N) < 0.1, hypergeometric is equal to binomial. Approximated to normal if np > 5, n(1-p) > 5 and (n/N) < 0.1 Arrival rate does not change over time; Arrival pattern does not follow regular pattern; Arrival of disjoint time intervals are independent. Approximated to normal if > 5 Z = (X − ) K N − K f (x ) = x n − x N n x =max(0,n-N+k) to min(K,n), K N, n N K objects classed as successes; N – K objects classified as failures; Sample size of n objects E(X) = = np where p = K/N N −n N −1 2 = np(1 − p) f ( x) = P( X = x ) = x e − x! x = 0,1,2,…, 0 < P(X) = probability of exactly X arrivals or occurrences Expected value = Variance = Taylor series: e = k k! k =0 Name When to Use E (X ) = xf ( x)dx − Approximations / Conditions Probability Density function, Mean and Variance P(x1 X x2) = P(x1 X x2) = P(x1 X x2) = P(x1 X x2) 2 = ( x − ) 2 f ( x)dx Uniform Equal probability (Continuous) xa 0 F ( x) = ( x − a) /(b − a) a x b 1 bx Normal Notation: N(,) (Continuous) X is any random variable Cumulative (z) = P(Z < z); Z is standard normal x Cumulative F(x) = P(X x) = − f (u)du ; for - < x < − For a series of n values, f(x) = 1/ (b –a); where a x b For a range that starts from a and ends with b (a, a+1, a+2, …, b) and a b ( a + b) = 2 (b − a)2 =Variance = V ( X ) = 12 2 If n is large (np > 5, n(1-p) > 5), normal is approximated to binomial. f (X) = −( x − )2 1 e 2 2 2 x + 0.5 − np < x < - < < , 0 < P ( X x) = P( X x + 0.5) P Z np ( 1 − p ) E(X) = V(X) = x − 0.5 − np Standard normal means mean = = 0 and variance = P ( x X ) = P( x − 05 X ) P Z = 1 np(1 − p) X − Z= Adding or subtracting 0.5 is called continuity correction. Normal is approximated to Poisson if > 5 Z= X − X − x− P ( X x) = P = P( Z z ) Cumulative distribution of a standard normal variable ( z ) = P(Z z ) - < < + = 68% -2 < < +2 = 95% -3 < < +3 = 99.7% for 0 x f ( x ) = 0 for x < 0 Exponential Memoryless f (x ) = e −x (Continuous) P ( X t1 + t 2 | X t1 ) = P( X t 2 ) P ( X x) =1 − F ( x) = e −x P( X x) = F ( x) = 1 − e −x → distance between successive 1 events of Poisson process with mean P (a X b) = F (b) − F (a) Expected value = = 1 = Average service time Variance = 2 >0 → length until first count in a The probability that an exponentially distributed time (X) required to serve a customer is less than or equal to Poisson process − t time t is given by the formula, P(X t ) = 1 − e 25 Erlang r → shape → scale (Continuous) Time between events are independent → length until r counts in a Poisson process or exponential distribution For mean and variance: Exponential multiplied by r gives Erlang. Expected value = = f (x) = P(X>0.1) = 1 – F(0.1) r x r −1 e −x (r − 1)! r Variance = r 2 for x 0 and r = 1, 2, ... If r = 1, Erlang random variable is an exponential random variable Gamma For r is an integer (r=1,2,…), gamma (Continuous) is Erlang Erlang random variable is time until the rth event in a Poisson process and time between events are independent For = ½, r = ½, 1, 3/2, 2, … gamma is chi-square Weibull Includes memory property (Continuous) - Scale; - Shape → Time until failure of many different physical systems Gamma Function (r ) = x r −1 e − x dx, for r 0 0 f (x) = r −1 − x x e r ( r ) Mean = = r , for x 0, 0 and r 0 and (r ) = (r − 1)!, (1) = 0!= 1, (1/ 2) = 1 / 2 Variance = 2 = r 2 E(X) and V(X) = E(X) and V(X) of exponential distribution multiplied by r −1 =1, Weibull is identical to exponential x x f ( x ) = exp − =2, Weibull is identical to Raleigh x > 0, Cumulative F (x ) = 1 − e = E ( X ) = 1 + 1 where (r ) = (r − 1)! 2 = 1 + − 2 2 Lognormal Includes memory property (Continuous) X = exp(W); W is normally distributed with mean and variance ln(X) = W; X is lognormal Easier to understand than Weibull 1 1 + 2 Weibull can be approximated to lognormal with and ln( x) − ln( x) − F ( x) = P X x = P exp(W ) x = P W ln( x) = P Z = for x 0 F ( X ) = 0, for x 0 (ln x − ) 2 exp − for 0 x 2 2 x 2 2 2 2 E ( X ) = e + 2 V ( X ) = e 2 + e − 1 ( + ) −1 f (x) = x (1 − x) −1 for 0 x 1, 0, 0 ( )( ) f (x) = 1 ( Beta Flexible but bounded over a finite (Continuous) range x − ) 26 E (X ) = Power Law Called as ‘heavy-tailed’ distribution. (Continuous) f(x) decreases rapidly with x but not as rapid as exponential distribution. + V (X ) = ( + ) ( + + 1) 2 A random variable described by its minimum value xmin and a scale parameter > 1 is is said to obey the power law distribution if its density function is given by ( − 1) x f (x) = x min x min − Normalize the function for a given set of parameters to ensure that f ( x ) dx = 1 − Central Limit Theorem If X1, X2, …, Xn is a random sample of size n taken from a population (either finite or infinite) with mean and finite variance 2, and if X is the sample mean, the limiting form of the distribution of Z= X − / n as n → , is the standard normal distribution. 27 Name Two or more Discrete Random Variables f X (x ) = P( X = x) = y Probability Density function, Mean and Variance f Y (y ) = P(Y = y) = f XY ( x, y) f XY ( x, y) E (Y | x ) = y f Y | x ( y) V (Y | x) = ( y − Y | x ) 2 f Y | x ( y) y Independen ce f Y | x ( y ) = f Y | x (y ) = f XY ( x, y) / f X ( x) x y f XY ( x, y) f X ( x) f Y ( y) = = f Y ( y); f X ( x) f X ( x) f XY ( x, y) = f X ( x) f Y ( y) for all x and y Joint Probability Mass Fn: f X1 X 2 X p ( x1 , x 2 , , x p ) = P( X 1 = x1 , X 2 = x 2 , , X p = x p ) for all points (x1,x2,…,xp) in the range of X1,X2,…,Xp Joint Probability Mass For subset: f X1 X 2 X k ( x1 , x 2 , , x k ) = P( X 1 = x1 , X 2 = x 2 , , X k = x k ) = P( X 1 = x1 , X 2 = x 2 , , X k = x k ) for all points in the range of X1,X2,…,Xp for which X1=x1, X2=x2,…, Xk=xk Marginal Probability Mass Function: f X i ( xi ) = P( X i = xi ) = f X1 X 2 X p ( x1 , x 2 , , x p ) Mean: E ( X i ) = xi f X1 X 2 X p ( x1 , x 2 , , x p ) Multinomial Probability Distribution Two or more Continuous Random Variables Variance: V ( X i ) = (x i ) − Xi 2 f X1 X 2 X p ( x1 , x 2 , , x p ) The random experiment that generates the probability distribution consists of a series of independent trials. However, the results from each trial can be categorized into one of k classes. P( X 1 = x1 , X 2 = x 2 , , X k = x k ) = n! p1x1 p 2x2 p kxk x1! x 2 ! x k ! E (X i ) = npi Marginal Probability Density Function: f X (x ) = for x1 + x 2 + + x k = n and p1 + p 2 + + p k = 1 V ( X i ) = npi (1 − pi ) f XY f Y (y ) = f XY ( x, y) dy ( x, y) dy y f Y |x (y ) = f XY ( x, y) for f X ( x) 0 f X ( x) y E (Y | x ) = y f Y | x ( y ) dy V (Y | x) = ( y − Y | x ) 2 f Y | x ( y) dy y y Independence: f Y | x (y ) = f Y ( y); f X | y ( x ) = f X ( x) f XY ( x, y) = f X ( x) f Y ( y) for all x and y f Joint Probability Density Fn: P ( X 1 = x1 , X 2 = x 2 , , X p = x p ) B = X1 X 2 X p ( x1 , x 2 , , x p )dx1 dx 2 dx p B Joint Probability Mass For subset: f X1 X 2 X k ( x1 , x 2 , , x k ) = f X 1 X 2 X p ( x1 , x 2 , , x p )dx1 dx 2 dx p for all points in the range of X1,X2,…,Xp for which X1=x1, X2=x2,…, Xk=xk Marginal Probability Density Function: f X i ( xi ) = f X 1 X 2 X p ( x1 , x 2 , , x p ) dx1 dx 2 dxi −1 dxi +1 dx p where the integral is over all R points of X1,X2,…,Xp for which Xi=xi 28 − − Mean: E ( X i ) = x ( i f X1 X 2 X p ( x1 , x 2 , , x p ) dx1 dx 2 dx p V ( X i ) = xi − X i − ) 2 Variance: f X1 X 2 X p ( x1 , x 2 , , x p ) dx1 dx 2 dx p − Covariance is a measure of linear relationship between the random variables. If the relationship between the random variables is nonlinear, the covariance might not be sensitive to the relationship. Two random variables with nonzero correlation are said to be correlated. Similar to covariance, the correlation is a measure of the linear relationship between random variables. Covariance: XY = E[( X − X )(Y − Y )] = E( XY ) − X Y Correlation: XY = cov( X , Y ) V ( X )V (Y ) = XY XY where − 1 XY +1 If X and Y are independent random variables, XY = XY = 0 Bivariate Normal 2 2 ( x − X )( y − Y ) ( − Y ) 2 − 1 (x − x ) exp − + 2 2 2 2 XY Y 2 X Y 1 − 2(1 − ) X for - < x < and - < y < , with parameters x > 0, y > 0, - < X < , - < Y < and -1 < < 1. Marginal Distribution: If X and Y have a bivariate normal distribution with joint probability density fXY(x, y; X, Y, X, Y, ), the marginal probability distribution of X and Y are normal with means x and y and standard deviation x and y, respectively. Conditional Distribution: If X and Y have a bivariate normal distribution with joint probability density fXY(x, y; X, Y, X, Y, ), the conditional f XY ( x, y; X , Y , X , Y , ) = 1 probability distribution of Y given X = x is normal with mean Y |x = Y − X Linear Functions of random variables Y Y + X and variance 2 Y | X = Y2 (1 − 2 ) X X Correlation: If X and Y have a bivariate normal distribution with joint probability density function fXY(x, y; X, Y, X, Y, ), the correlation between X and Y is If X and Y have a bivariate normal distribution with = 0, X and Y are independent Given random variables X1, X2, …, Xp and constants c1,c2, …, cp, Y = c1X1 + c2X2+ … + cpXp is a linear combination of X1, X2, …, Xp Mean E(Y) = c1E(X1) + c2E(X2) + … + cpE(Xp) Variance: V (Y ) = c12V ( X 1 ) + c 22V ( X 2 ) + + c 2pV ( X p ) + 2 ci c j cov( X i , X j ) i j If X1, X2, …, Xp are independent, variance: V (Y ) = c V ( X 1 ) + c22V ( X 2 ) + + c 2pV ( X p ) 2 1 Mean and variance on average: E ( X ) = ; V (X ) = 2 p E (Y ) = c1 1 + c 2 2 + + c p p ; with E ( X i ) = and V ( X i ) = 2 V (Y ) = c12 12 + c 22 22 + + c 2p p2 29 General Functions of random variables Discrete: y = h(x) and x = u(y): fY (y ) = f X u( y) Continuous: y = h(x) and x = u(y): f Y ( y ) = f X [u ( y)] | J | where J = u ' ( y) is called the Jacobian of the transformation and the absolute value of J is used 30 Name Confidence Interval Sample Size 𝑥̅ 𝑖𝑠 𝑎𝑛 𝑒𝑠𝑡𝑖𝑚𝑎𝑡𝑜𝑟 𝑜𝑓 𝜇; 𝑆 2 𝑖𝑠 𝑠𝑎𝑚𝑝𝑙𝑒 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒 One-Sided Confidence Bounds Type I Error: Rejecting the null Hypothesis H0 when it is true; Type II Error: Failing to reject null hypothesis H0 when it is false. Probability of Type I Error = = P(type I error) = Significance level = -error = -level = size of the test. Probability of Type II Error = = (type II error) Power = Probability of rejecting the null hypothesis H0 when the alternative hypothesis is true = 1 - = Probability of correctly rejecting a false null hypothesis. P-value = Smallest level of significance that would lead to the rejection of the null hypothesis H0. 100(1−∝)% CI on μ is 100(1− 100(1-)% upper confidence bound for ∝)% upper confident that the error x − z∝/2 /√𝑛 ≤ 𝜇 ≤ x + z∝/2 /√𝑛 𝜇 ≤ 𝑢 = x + zσ /√𝑛 |𝑥̅ 100(1-)% lower confidence bound for − 𝜇| will not exceed a specific amount E z∝⁄2 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 100 ∝⁄2 x − zσ ⁄√𝑛 = 𝑙 ≤ 𝜇 when sample size is 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑍𝛼/2 𝜎 2 𝑛𝑜𝑟𝑚𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑛= ( ) 𝐸 Large scale sample size n: Using central limit theorem, X has approximately a normal distribution with mean and variance /n. Z= Normal Distribution with mean unknown and variance known X − S/ n x − z∝/2 Null hypothesis: H0: = 0 Test Statistic: Z 0 = Alternative hypothesis H1: 0 ≤ 𝜇 ≤ x + z∝/2 S √𝑛 Probability of Type II Error for a two-sided test 𝛿 √𝑛 𝛿 √𝑛 𝛽 = 𝜙 (𝑧𝛼/2 − ) − 𝜙 (−𝑧𝛼/2 − ) 𝜎 𝜎 X − 0 / n p-value S √𝑛 2 Rejection Criteria Sample size for a two-sided test 𝑛 = (𝑧𝛼/2 +𝑧𝛽 ) 𝜎 2 𝛿2 2 (𝑧𝛼 +𝑧𝛽 ) 𝜎 2 𝑤ℎ𝑒𝑟𝑒 𝛿 = 𝜇 − 𝜇0 Sample size for a one-sided test 𝑛 = 𝑤ℎ𝑒𝑟𝑒 𝛿 = 𝜇 − 𝜇0 Probability above |z0| & 𝑧0 > zα/2 𝑜𝑟 𝑧0 < 𝛿2 below -|z0|, Large set: for n > 40, replace sample standard deviation s for . −zα/2 |𝜇−𝜇 | |𝛿| P = 2[1 - (|z0|)] Parameter for Operating Characteristic Curves: 𝑑 = 𝜎 0 = 𝜎 Probability above z0 𝑧0 > zα H1: 0 P = 1 - (z0) Probability below z0 𝑧0 < −zα H1: 0 P = (z0) t distribution (Similar to normal in symmetry and unimodal. But t distribution is heavier tails than normal) with n-1 degrees of freedom. k = n-1 T= X − S/ n 𝑓(𝑥) = Γ[(𝑘+1)/2] 1 . √𝜋𝑘Γ(𝑘/2) [(𝑥 2 ⁄𝑘 )+1](𝑘+1)⁄2 Mean = 0 Variance=k/(k-2) for k>2 31 Normal Distribution with mean unknown and variance unknown 100(1−∝)% CI on μ is x − t ∝/2,n−1 s/√𝑛 ≤ 𝜇 ≤ x + t ∝/2,n−1 s/√𝑛 Finding s: Can be obtained only using trial and error as s is unknown until the data is collected. t ∝⁄2,𝑛−1 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 100 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑡 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑛 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 Null hypothesis: H0: = 0 Test Statistic: T0 = Alternative hypothesis H1: 0 H1: 0 H1: 0 Probability of Type II Error for a two-sided test 𝛽 = 𝑃(−t α/2,n−1 ≤ T0 ≤ t α/2,n−1 |𝛿 ≠ 0) = 𝑃(−t α/2,n−1 ≤ 𝑇0′ ≤ t α/2,n−1 ) When the true value of mean = + , the distribution for T0 is called the noncentral t distribution with n-1 degrees of freedom and noncentrality parameter 𝛿√𝑛/𝜎. If = 0, the noncentral t distribution reduces to the usual central t distribution. 𝑇0′ denotes the noncentral t random variable. X − 0 S/ n p-value Rejection Criteria 𝑡0 > t α/2,n−1 𝑜𝑟 𝑡0 < −t α/2,n−1 𝑡0 > t α,n−1 𝑡0 < −t α,n−1 Probability above |t0| & below -|t0| Probability above t0 Probability below t0 Parameter for Operating Characteristic Curves: 𝑑 = |𝜇−𝜇0 | 𝜎 = 100(1-)% upper confidence bound for 𝜇 ≤ 𝑢 = x + t ∝,n−1 s/√𝑛 100(1-)% lower confidence bound for x − t ∝,n−1 𝑠⁄√𝑛 = 𝑙 ≤ 𝜇 |𝛿| 𝜎 Chi − square (χ2 ) distribution with n-1 degrees of freedom. k = n-1 2 = Normal Distribution; CI on variance and Standard deviation (n − 1) S 2 1 𝑓(𝑥) = 2𝑘/2 Γ(𝑘/2) 𝑥 (𝑘/2)−1 𝑒 −𝑥/2 2 Mean = k Variance = 2k 100(1−∝)% CI on σ2 is (𝑛 − 1)𝑠 2 (𝑛 − 1)𝑠 2 2 ≤ 𝜎 ≤ 2 2 𝜒𝛼/2,𝑛−1 𝜒1−𝛼/2,𝑛−1 100(1-)% upper confidence bound for 𝜎 2 (𝑛 − 1)𝑠 2 𝜎2 ≤ 2 𝜒1−𝛼,𝑛−1 100(1-)% lower confidence bound for 𝜎 2 2 2 𝜒𝛼/2,𝑛−1 𝑎𝑛𝑑 𝜒1−𝛼/2,𝑛−1 𝑎𝑟𝑒 𝑡ℎ𝑒 (𝑛 − 1)𝑠 2 ≤ 𝜎2 2 upper and lower 100/2 percentage points of chi-square distribution with 𝜒𝛼,𝑛−1 n-1 degrees of freedom. Null hypothesis: H0 : 2 = 02 2 Test Statistic: 0 = Alternative hypothesis (n − 1) S 2 02 Rejection Criteria 32 2 𝜒02 > 𝜒𝛼/2,𝑛−1 𝑜𝑟 𝜒02 < 2 −𝜒𝛼/2,𝑛−1 2 𝜒02 > 𝜒𝛼,𝑛−1 H1: 2 02 H1: 2 02 2 𝜒02 < −𝜒𝛼,𝑛−1 H1: 2 02 Parameter for Operating Characteristic Curves: 𝑑 = |𝜇−𝜇0 | 𝜎 = |𝛿| 𝜎 Normal approximation for a binomial proportion: If n is large, the distribution of Z= Large scale CI for a population proportion X − np np(1 − p) = pˆ − p p(1 − p) n is approximately standard normal. p̂ is the proportional population. Mean = p. Variance = p(1-p)/n 100(1−∝)% CI on 𝑝𝑟𝑜𝑝𝑜𝑟𝑡𝑖𝑜𝑛 𝑝 is 100(1-)% upper confidence bound for 𝜎 2 𝑍𝛼/2 2 𝑛= ( ) 𝑝(1 − 𝑝) 𝑝̂(1−𝑝) 𝐸 𝑝̂ (1 − 𝑝) 𝑝̂ − 𝑧𝛼 √ 𝑛 ≤ 𝑝 ≤ 𝑝̂ + p can be computed as 𝑝̂ from a 2 𝑝 ≤ 𝑝̂ + 𝑧𝛼 √ 𝑛 𝑝̂(1−𝑝) preliminary sample or use the maximum 𝑧𝛼/2 √ 𝑛 100(1-)% lower confidence bound for 𝜎 2 value of p, which is 0.5. z∝⁄2 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 100 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑛𝑜𝑟𝑚𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 Null hypothesis: H0 : p = p0 Test Statistic: z0 = Alternative hypothesis H1: p p0 H1: p p0 H1: p p0 p-value X − np0 np0 (1 − p0 ) Rejection Criteria 𝑝̂ (1 − 𝑝) 𝑝̂ − 𝑧𝛼 √ ≤𝑝 𝑛 Probability of Type II Error for a two-sided test 𝛽 = 𝜙( 𝑝0 −𝑝+𝑧𝛼/2 √𝑝0 (1−𝑝0 )/𝑛 √𝑝(1−𝑝)/𝑛 )−𝜙( 𝑝0 −𝑝−𝑧𝛼/2 √𝑝0 (1−𝑝0 )/𝑛 √𝑝(1−𝑝)/𝑛 ) Sample size for a two-sided test 𝑛=[ 𝑧𝛼/2 √𝑝0 (1−𝑝0 )+𝑧𝛽 √𝑝(1−𝑝) 𝑝−𝑝0 2 ] Sample size for a one-sided test Probability above |z0| & below -|z0|, P = 2[1 - (|z0|)] Probability above z0 P = 1 - (z0) Probability below z0 P = (z0) 𝑧0 > zα/2 𝑜𝑟 𝑧0 < −zα/2 𝑛=[ 𝑧𝛼 √𝑝0 (1−𝑝0 )+𝑧𝛽 √𝑝(1−𝑝) 𝑝−𝑝0 2 ] 𝑧0 > zα 𝑧0 < −zα 33 100(1−∝)% CI on μ1 − μ2 is z x1 − x2 − 𝛼/2 ≤ 𝜇1 − 𝜇2 ≤ x1 − 2 2 𝜎 √𝜎1 + 2 𝑛1 𝑛2 𝜎2 𝜎2 1 2 x2 + z∝/2 /√𝑛1 + 𝑛2 z∝⁄2 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 100 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑛𝑜𝑟𝑚𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 Inference of the difference in means of two normal distributions , variance known 100(1-)% upper confidence bound for 100(1−∝)% upper confident that the error in estimating 𝜇1 − 𝜇2 by x1 − x2 will not exceed a specific amount E when sample size is 𝑍𝛼/2 2 2 𝑛= ( ) (𝜎1 + 𝜎22 ) 𝐸 Z= 𝜇1 − 𝜇2 ≤ x1 − x2 + zσ √ 𝜎22 + 𝑛1 𝑛2 𝜎21 100(1-)% lower confidence bound for 𝜎12 𝜎2 x1 − x2 − zσ √ + 2 ≤ 𝜇1 − 𝜇2 𝑛1 𝑛2 X 1 − X 2 − (1 − 2 ) 12 n1 + 22 n2 Probability of Type II Error for a two-sided test 𝛽 = 𝜙 𝑧𝛼/2 − Δ−Δ0 2 𝜎2 √𝜎1 + 2 𝑛1 𝑛2 − 𝜙 −𝑧𝛼/2 − ( )2 2 ( 𝜎 +𝜎 Sample size for a two-sided test, with n1 n2, 𝑛 = 𝜎2 /𝑛1 +𝜎22 /𝑛 1 1 2 2 Δ−Δ0 2 2 𝜎 √ 𝜎1 + 2 𝑛1 𝑛2 ) 2 (𝑧𝛼/2 +𝑧𝛽 ) (𝜎12 +𝜎22 ) Sample size for a two-sided test, with n1=n2, 𝑛 = (Δ−Δ0 )2 2 Sample size for a one-sided test, with n1=n2, 𝑛 = Parameter for Operating Characteristic Curves: 𝑑 = (𝑧𝛼 +𝑧𝛽 ) (𝜎12 +𝜎22 ) (Δ−Δ0 )2 |𝜇1 −𝜇2 −Δ0| |Δ−Δ0 | √𝜎12 +𝜎22 = √𝜎12 +𝜎22 Null hypothesis: H0: − = 0 Test Statistic: Z0 = X 1 − X 2 − 0 12 n1 Alternative hypothesis H1: − 0 H1: − 0 + 22 n2 p-value Rejection Criteria Probability above |z0| & below -|z0|, P = 2[1 - (|z0|)] Probability above z0 P = 1 - (z0) 𝑧0 > zα/2 𝑜𝑟 𝑧0 < −zα/2 𝑧0 > zα 34 H1: − 0 Probability below z0 𝑧0 < −zα P = (z0) 100(1−∝)% CI on μ1 − μ2 with assumed equal variance is 100(1−∝)% CI on μ1 − μ2 with assumed unequal variance = is 1 1 x1 − x2 − 𝑡𝛼/2,𝑛1 +𝑛2 −2 𝑆𝑝 √ + ≤ 𝜇1 − 𝜇2 𝑠12 𝑠22 𝑛1 𝑛2 x1 − x2 − 𝑡𝛼/2,𝜐 √ + ≤ 𝜇1 − 𝜇2 𝑛1 𝑛2 1 1 ≤ x1 − x2 + 𝑡𝛼/2,𝑛1 +𝑛2 −2 𝑆𝑝 √ + 𝑠12 𝑠22 𝑛1 𝑛2 ≤ x1 − x2 + 𝑡𝛼/2,𝜐 √ + 𝑛1 𝑛2 𝑡 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 ∝⁄2 𝛼/2,𝑛1 +𝑛2 −2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑡 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑛1 + 𝑛2 − 2 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 Inference of the difference in means of two normal distributions , variance unknown Z= X 1 − X 2 − (1 − 2 ) 1 1 + n1 n2 Pooled estimator of 2, denoted by 𝑆𝑝2 , is: 𝑆𝑝2 = (𝑛1 −1)𝑆12 +(𝑛2 −1)𝑆22 𝑛1 +𝑛2 −2 Null hypothesis: H0: − = 0 H1: − 0 Test Statistic: T0 = X 1 − X 2 − 0 has t distribution with n1+n2-2 degrees of freedom; Called as pooled t-test 1 1 Sp + n1 n2 Alternative hypothesis H1: − 0 H1: − 0 H1: − 0 p-value Rejection Criteria Probability above |t0| 𝑡0 > t α/2,𝑛1 +𝑛2 −2 𝑜𝑟 𝑡0 < −t α/2,𝑛1 +𝑛2 −2 & below -|t0| Probability above t0 𝑡0 > t α,𝑛1 +𝑛2 −2 Probability below t0 𝑡0 < −t α,𝑛1 +𝑛2 −2 If variances are not assumed equal * If H0: − = 0 is true, the statistic T0 = X 1 − X 2 − 0 S12 S22 + n1 n2 with t degrees of freedom given by 35 2 𝜈= Goodness of Fit Test Statistic 𝑆2 𝑆2 (𝑛1 + 𝑛2 ) 1 2 (𝑆12 /𝑛1 )2 (𝑆22 /𝑛2 )2 𝑛1 − 1 + 𝑛2 − 1 𝐼𝑓 𝜈 𝑖𝑠 𝑛𝑜𝑡 𝑎𝑛 𝑖𝑛𝑡𝑒𝑔𝑒𝑟, 𝑟𝑜𝑢𝑛𝑑 𝑑𝑜𝑤𝑛 𝑡𝑜 𝑡ℎ𝑒 𝑛𝑒𝑎𝑟𝑒𝑠𝑡 𝑖𝑛𝑡𝑒𝑔𝑒𝑟 (Oi − Ei )2 X0 = Where Oi is the observed frequency and Ei is the expected frequency in the ith class. E i =1 i Approximated to Chi-square distribution with k-p-1 degrees of freedom. p represents the number of parameters. If test statistic is large, we 2 reject null hypothesis. P-value is 𝑃(𝜒𝑘−𝑝−1 > 𝜒02 ). r 1 c 1 c 1 r uˆi = Oij vˆ j = Oij Expected Frequency of each cell Eij = nuˆi vˆ j = Oij Oij n j =1 n j =1 i =1 n i =1 k 2 r c For large n, the statistic 0 = 2 (Oij − Eij )2 Eij i =1 j =1 100(1−∝)% prediction interval on a single future observation from a normal distribution is Prediction Interval x − t ∝/2,n−1 s√1 + Tolerance Interval Sign Test 2 P-value is 𝑃(𝜒(𝑟−1)(𝑐−1) > 𝜒02 ). 1 1 ≤ 𝑋𝑛+1 ≤ x + t ∝/2,n−1 s√1 + 𝑛 𝑛 Prediction interval for Xn+1 will always be longer than the CI for because there is more variability associated with the prediction error than with the error of estimation. Tolerance interval for capturing at least γ% of the values in a normal distribution with cofidence level 100(1−∝)% is x − 𝑘𝑠, x + 𝑘𝑠 Where k is a tolerance interval factor for the given confidence. ~ differences = r+ Number of positive X i − If P-value is less than some preselected level , we will reject H0. 0 Normal approximation for sign test statistic: z 0 = ~= ~ Null hypothesis: H0 : 0 ~ ~ H : 1 1 0 P-value: P R + r + when p = Wilcoxon Signed-Rank Test ~= ~ One-sided hypothesis: H0 : 0 ~ ~ H : Null hypothesis: H0 : = 0 1 2 0 P-value: P R + r + when p = 1 2 R + − 0.5n 0.5 n ~= ~ Two-sided hypothesis: H0 : 0 ~ ~ H : 1 0 1 2 1 If r+ > n/2, P-value: 2 P R + r + when p = 2 If r+ < n/2, P-value: 2 P R + r + when p = Sort based on X i − 0 differences; Give the ranks the signs of their corresponding differences. Sum of Positive Ranks: W+; Absolute value of Sum of Negative Ranks: W-. W = min(W+,W-) 36 H1 : 0 Reject Null hypothesis, if observed value of statistic w w* For one-sided tests, H1 : 0 Reject Null hypothesis, if observed value of statistic w− w* For one-sided tests, H1 : 0 Reject Null hypothesis, if observed value of statistic w+ w* W + − n(n + 1) / 4 Normal approximation for Wilcoxon signed-rank test statistic: z0 = n(n + 1)(2n + 1) / 24 Arrange all n1 and n2 observations in ascending order of magnitude and assign ranks to them. If two or more observations are tied(identical), use the mean of the ranks that would have assigned if the observations differed. W1 is sum of ranks in smaller sample. W2 is sum of ranks in other sample. 𝑊2 = (𝑛1 +𝑛2 )(𝑛1 +𝑛2 +1) 2 − 𝑊1 Reject null hypothesis, if w1 or w2 is less than or equal to tabulated critical value w. For one-sided hypothesis: H1 : 1 2 reject H0 if w1 w For H1 : 1 2 reject H0 if w2 w Normal approximation when n1 and n2 > 8, Z0 = 𝑊1 −𝜇𝑤1 100(1−∝)% CI on 𝜇𝐷 = μ1 − μ2 is 𝑑̅ − 𝑡𝛼,𝑛−1 𝑆𝐷 /√𝑛 ≤ 𝜇𝐷 ≤ 𝑑̅ + 𝑡𝛼,𝑛−1 𝑆𝐷 /√𝑛 Paired t-test 2 𝜎𝑤1 2 𝑡𝛼,𝑛−1 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑡 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑛 − 1 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 2 Null hypothesis: H0: D = 0 Test Statistic: T0 = Alternative hypothesis H1: D 0 Inference on the variances of two normal distributions (F Distribution) D − 0 ̅ is the sample average of n differences D1, D2, … , Dn, where 𝐷 SD / n and SD is the sample standard deviation of these differences. p-value Rejection Criteria Probability above |t0| 𝑡0 > t α/2,𝑛−1 𝑜𝑟 𝑡0 < −t α/2,n−1 & below -|t0| Probability above t0 𝑡0 > t α,n−1 H1: D 0 Probability below t0 𝑡0 < −t α,n−1 H1: D 0 Let W and Y be independent chi-square random variables with u and v degrees of freedom respectively. Ratio F = 𝑊/𝑢 𝑌/𝑣 has the probability density function 𝑢 + 𝑣 𝑢 𝑢/2 (𝑢2)−1 Γ( )( ) 𝑥 2 𝑣 f(x) = ,0 < 𝑥 < ∞ (𝑢+𝑣)/2 𝑢 𝑣 𝑢 Γ (2) Γ (2) [(𝑣 ) 𝑥 + 1] ν 2𝜈 2 (𝑢+𝜈−2) Mean μ = ν−2 𝑓𝑜𝑟 𝜈 > 2 Variance σ2 = 𝑢(𝜈−2)2 (𝜈−4) , 𝑓𝑜𝑟 𝜈 > 4 Lower tail percentage point, 𝑓1−𝛼,𝑢,𝜈 = 𝑓 1 𝛼,𝜈,𝑢 37 F Distribution 𝐹 = 𝑆12 ⁄𝜎12 𝑆22 ⁄𝜎22 n1-1 numerator degrees of freedom and n2-1 denominator degrees of freedom Null hypothesis: 𝐻0 : 𝜎12 = 𝜎22 𝑆2 Test Statistic: 𝐹0 = 𝑆12 2 Alternative hypothesis H1: 𝜎12 𝜎22 H1: 𝜎12 Rejection Criteria 𝑓0 > fα,𝑛 −1,𝑛2 −1 2 1 𝜎22 𝑜𝑟 𝑓0 < −f1−α,𝑛 𝑓0 > f𝛼,𝑛1 −1,𝑛2 −1 −1,𝑛2 −1 2 1 𝑓0 < f1−𝛼,𝑛1 −1,𝑛2 −1 H1: 𝜎12 𝜎22 P-value is the area (probability) under the F distribution with n1-1 and n2-1 degrees of freedom that lies beyond the computed value of the test statistic f0. σ12 100(1−∝)% CI on the ratio 2 is 𝜎2 𝑠12 σ12 𝑠12 𝑓 𝛼 ≤ ≤ 𝑓𝛼 𝑠22 1−2 ,𝑛2 −1,𝑛1 −1 𝜎22 𝑠22 2 ,𝑛2 −1,𝑛1 −1 𝑓𝛼,𝑛 −1,𝑛 −1 𝑎𝑛𝑑 𝑓1−𝛼,𝑛 −1,𝑛 −1 𝑎𝑟𝑒 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 𝑎𝑛𝑑 𝑙𝑜𝑤𝑒𝑟 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝐹 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑤𝑖𝑡ℎ 𝑛2 − 1 𝑛𝑢𝑚𝑒𝑟𝑎𝑡𝑜𝑟 𝑎𝑛𝑑 𝑛1 2 2 1 2 2 1 − 1 𝑑𝑒𝑛𝑜𝑚𝑖𝑛𝑎𝑡𝑜𝑟 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 𝑜𝑓 𝑓𝑟𝑒𝑒𝑑𝑜𝑚 Null hypothesis: 𝐻0 : 𝑝1 = 𝑝2 𝐻1 : 𝑝1 ≠ 𝑝2 ̂ 𝑃1 − 𝑃̂2 − (𝑝1 − 𝑝2 ) 𝑍= 𝑝 (1 − 𝑝1 ) 𝑝2 (1 − 𝑝2 ) √ 1 + 𝑛 𝑛 Inference on the population proportions 1 Test Statistic: 𝑍0 = Alternative hypothesis H1: p1 p2 H1: p1 p2 H1: p1 p2 2 𝑃̂1 −𝑃̂2 1 1 √𝑃̂(1−𝑃̂)(𝑛 +𝑛 ) 1 2 p-value Probability above |z0| & below -|z0|, P = 2[1 - (|z0|)] Probability above z0 P = 1 - (z0) Probability below z0 P = (z0) Probability of Type II Error for a two-sided test Rejection Criteria for FixedLevel tests 𝑧0 > zα/2 𝑜𝑟 𝑧0 < −zα/2 𝑧0 > zα 𝑧0 < −zα 38 𝛽 = 𝜙[ ̅̅̅̅(1⁄𝑛1 +1⁄𝑛2 )−(𝑝1 −𝑝2 ) 𝑧𝛼/2 √𝑝𝑞 𝜎𝑃1 −𝑃2 𝑛 𝑝 +𝑛 𝑝 Where 𝑝̅ = 1 1 2 2 𝑛1 +𝑛2 ]−𝜙[ and 𝑞̅ = ̅̅̅̅(1⁄𝑛1 +1⁄𝑛2 )−(𝑝1 −𝑝2 ) −𝑧𝛼/2 √𝑝𝑞 𝜎𝑃1 −𝑃2 𝑛1 (1−𝑝1 )+𝑛2 (1−𝑝2 ) 𝑛1 +𝑛2 ] Sample size for a two-sided test 2 𝑛= [𝑧𝛼/2 √(𝑝1 +𝑝2 )(𝑞1 +𝑞2 )/2+𝑧𝛽 √𝑝1 𝑞1 +𝑝2 𝑞2 ] (𝑝1 −𝑝2 )2 where q1 = 1 – p1 and q2 = 1 – p2 100(1−∝)% CI on the difference in the true proportions p1 − p2 is 𝑝̂1 (1 − 𝑝̂1 ) 𝑝̂ 2 (1 − 𝑝̂2 ) 𝑝̂1 (1 − 𝑝̂1 ) 𝑝̂ 2 (1 − 𝑝̂2 ) 𝑝̂1 − 𝑝̂2 − zα/2 √ + ≤ 𝑝1 − 𝑝2 ≤ 𝑝̂1 − 𝑝̂2 + zα/2 √ + 𝑛1 𝑛2 𝑛1 𝑛2 z∝⁄2 𝑖𝑠 𝑡ℎ𝑒 𝑢𝑝𝑝𝑒𝑟 ∝⁄2 𝑝𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑝𝑜𝑖𝑛𝑡 𝑜𝑓 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑛𝑜𝑟𝑚𝑎𝑙 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 Inferences: 1. Population is normal: Sign test or t-test. a. t-test has the smallest value of for a significance level , thus t-test is superior to other tests. 2. Population is symmetric but not normal (but with finite mean): a. t-test will have the smaller (or a higher power) than sign test. b. Wilcoxon Signed-rank test is comparable to t-test. 3. Distribution with heavier tails: a. Wilcoxon Signed-rank test is better than t-test as t-test depends on the sample mean, which is unstable in heavy-tailed distributions. 4. Distribution is not close to normal: a. Wilcoxon Signed-rank test is preferred. 5. Paired observations: a. Both sign test and Wilcoxon Signed-rank test can be applied. In sign test, median of the differences is equal to zero in null hypothesis. In Wilcoxon Signed-rank test, mean of the differences is equal to zero in null hypothesis. 6. A 39 DECISION ANALYSIS Criterion of Realism Expected Monetary Value Weighted average = (best in row) + (1 – )(worst in EMV(alternative) = X i P( X i ) row) Xi = payoff for the alternative in state of nature i For Minimization: P(Xi) = probability of achieving payoff Xi (i.e., Weighted average = (best in row) + (1 – )(worst in probability of state of nature i) row) ∑ = summation symbol EMV (alternative i) = (payoff of first state of nature) x Expected Value with Perfect Information (probability of first state of nature) + (payoff of second EVwPI = ∑(best payoff in state of nature i) state of nature) x (probability of second state of nature) + (probability of state of nature i) … + (payoff of last state of nature) x (probability of last EVwPI = (best payoff for first state of nature) x state of nature) (probability of first state of nature) + (best payoff for second state of nature) x (probability of second state of nature) + … + (best payoff for last state of nature) x (probability of last state of nature) Expected Value of Sample Information EVSI = (EV with SI + cost) – (EV without SI) Utility of other outcome = (p)(utility of best outcome, which is 1) + (1 – p)(utility of the worst outcome, which is 0) Expected Value of Perfect Information EVPI = EVwPI – Best EMV EVSI Efficiency of sample information = 100% EVPI FORECASTING forecast error (error ) 2 Mean Absolute Deviation (MAD)= Mean Squared Error (MSE) = n n error actual Mean Absolute Percent Error (MAPE) = 100% n Y + Yt −1 + ... + Yt − n +1 sum of demands in previous n periods Mean Average Forecast = = Ft +1 = t n n Weighted Moving Average : Ft +1 = (Weight in period i)(Actual value in period) = w Y + w Y w +w (Weights ) + ... + wnYt −n +1 2 + ... + wn 2 t −1 1 t 1 Exponential Smoothing : Ft +1 = Ft + (Yt − Ft ) New forecast = Last period’ s forecast + (Last period’ s actual demand – Last period’ s forecast) Exponential Smoothing with Trend : Yˆ = b0 + b1 X Ft +1 = FITt + (Yt − FITt ) where Yˆ = predicted value Tt +1 = Tt + (Ft +1 − FITt ) b0 = intercept FITt +1 = Ft +1 + Tt +1 b1 = slope of the line X = time period (i.e., X = 1, 2, 3, , n) Tracking signal = RSFE (forecast error) = MAD MAD Yˆ = a + b1 X1 + b2 X 2 + b3 X 3 + b4 X 4 41 INVENTORY CONTROL MODELS Annual ordering cost = Number of orders placed per year Q Average inventory level = 2 (Ordering cost per order) Annual Demand D = Co = Co Numb er of units in each order Q Annual holding cost = Average Inventory Economic Order Quantity Annual ordering cost = Annual holding cost (Carrying cost per unit per year) D Q Order quantity Co = Ch = (Carrying cost per unit per year) Q 2 2 2DCo Q EOQ = Q* = = Ch Ch 2 Total cost (TC) = Order cost + Holding cost Cost of storing one unit of inventory for one year = Ch = IC, where C is the unit price or cost of an inventory item D Q TC = Co + Ch and I is Annual inventory holding charge as a percentage Q 2 of unit price or cost 2DCo Q* = IC ROP without Safety Stock: EOQ without instantaneous receipt assumption Reorder Point (ROP) = Demand per day x Lead Maximum inventory level = (Total produced during the time for a new order in days production run) – (Total used during the production run) =dL = (Daily production rate)(Number of days production) Inventory position = Inventory on hand + – (Daily demand)(Number of days production) Inventory on order = (pt) – (dt) d Q Q = pt – dt = p – d = Q1 – p p p Total produced Q = pt Q d Q d Average inventory = 1 – Annual holding cost = 1 – Ch 2 p 2 p D D Annual setup cost = Cs Annual ordering cost = Co Q Q D = the annual demand in units Q = number of pieces per order, or production run Production Run Model: EOQ without Quantity Discount Model instantaneous receipt assumption 2DCo EOQ = Annual holding cost = Annual setup cost IC Q d D If EOQ < Minimum for discount, adjust the quantity to Q 1 – Ch = Cs 2 p Q = Minimum for discount Total cost = Material cost + Ordering cost + Holding cost 2DCs Q* = D Q Total cost = DC + Co + Ch d Ch 1 – Q 2 p Holding cost per unit is based on cost, so Ch = IC Where I = holding cost as a percentage of the unit cost (C) Safety Stock Safety Stock with Normal Distribution ROP = (Average demand during lead time) + ZsdLT 42 ROP = Average demand during lead time + Safety Stock Service level = 1 – Probability of a stockout Probability of a stockout = 1 – Service level Demand is variable but lead time is constant ROP = d L + Z d L Z = number of standard deviations for a given service level dLT = standard deviation of demand during the lead time Safety stock = ZdLT Demand is constant but lead time is variable ROP = dL + Z (d L ) d = average daily demand d = standard deviation of daily demand L = average lead time L = standard deviation of lead time L = lead time in days Both demand and lead time are variable d = daily demand Total Annual Holding Cost with Safety Stock Total Annual Holding Cost = Holding cost of regular inventory + Holding cost of safety stock Q THC = Ch + (SS) Ch 2 ( ) ROP = d L + Z L d2 + d 2 L2 The expected marginal profit = P(MP) The expected marginal loss = (1 – P)(ML) The optimal decision rule Stock the additional unit if P(MP) ≥ (1 – P)ML P(MP) ≥ ML – P(ML) P(MP) + P(ML) ≥ ML P(MP + ML) ≥ ML ML P ML + MP 43 PROJECT MANAGEMENT Expected Activity Time t = a + 4m + b 6 Earliest finish time = Earliest start time + Expected activity time EF = ES + t Latest start time = Latest finish time – Expected activity time LS = LF – t Slack = LS – ES, or Slack = LF – EF Project standard deviation = T = Project variance Value of work completed = (Percentage of work complete) x (Total activity budget) Crash cost − Normal Cost Crash cost/Time Period = Normal time − Crash time 2 b –a Variance = 6 Earliest start = Largest of the earliest finish times of immediate predecessors ES = Largest EF of immediate predecessors Latest finish time = Smallest of latest start times for following activities LF = Smallest LS of following activities Project Variance = sum of variances of activities on the critical path Due date − Expected date of completion Z= T Activity difference = Actual cost – Value of work completed 44 Upper control limit (UCL) = x + z x STATISTICAL QUALITY CONTROL UCL x = x + A2 R Lower control limit (LCL) = x − z x LCL x = x − A2 R x = mean of the sample means z = number of normal standard deviations (2 for 95.5% confidence, 3 for 99.7%) x = standard deviation of the sampling distribution R = average of the samples A2 = Mean factor x = mean of the sample means of the sample means = x n UCL R = D4 R LCL R = D3 R UCLR = upper control chart limit for the range LCLR = lower control chart limit for the range D4 and D3 = Upper range and lower range p-charts UCL p = p + z p LCL p = p − z p p = mean proportion or fraction defective in the sample Total number of errors p= Total number of records examined z = number of standard deviations p = standard deviation of the sampling distribution p is estimated by ˆ p c-charts The mean is c and the standard deviation is equal to c Estimated standard deviation of a binomial distribution p (1 − p ) ˆ p = n where n is the size of each sample Range of the sample = Xmax - Xmin To compute the control limits we use c 3 c (3 is used for 99.7% and 2 is used for 95.5%) UCL c = c + 3 c LCL c = c − 3 c Control Chart Model k is the distance of control limits from the center line, expressed in Standard Deviation units. Common choice is k = 3. 𝑈𝐶𝐿 = 𝜇𝑊 + 𝑘𝜎𝑊 𝐶𝐿 = 𝜇𝑊 𝐿𝐶𝐿 = 𝜇𝑊 − 𝑘𝜎𝑊 𝜇𝑊 𝑖𝑠 𝑡ℎ𝑒 𝑚𝑒𝑎𝑛 𝑜𝑓 𝑡ℎ𝑒 𝑠𝑎𝑚𝑝𝑙𝑒 𝑊 𝜎𝑊 𝑖𝑠 𝑡ℎ𝑒 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑑𝑒𝑣𝑖𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑠𝑎𝑚𝑝𝑙𝑒 𝑊 𝜎 𝜎𝑊 = √𝑛 𝑅 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 𝑋̅ 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 𝑈𝐶𝐿 = 𝜇 + 3𝜎/√𝑛 𝐿𝐶𝐿 = 𝜇 − 3𝜎/√𝑛 𝐶𝐿 = 𝜇 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝑚𝑒𝑎𝑛 𝑜𝑓 𝑡ℎ𝑒 𝑝𝑜𝑝𝑢𝑙𝑎𝑡𝑖𝑜𝑛 𝜇, 𝑔𝑟𝑎𝑛𝑑 𝑚𝑒𝑎𝑛, 𝑚 1 𝜇̂ = 𝑋̿ = ∑ 𝑋̅𝑖 𝑚 𝑖=1 𝑋̿ 𝑖𝑠 𝑡ℎ𝑒 𝑐𝑒𝑛𝑡𝑒𝑟 𝑙𝑖𝑛𝑒 𝑜𝑛 ̅𝑋 𝑐𝑜𝑛𝑡𝑟𝑜𝑙 𝑐ℎ𝑎𝑟𝑡 ̅𝑋 𝑐𝑜𝑛𝑡𝑟𝑜𝑙 𝑐ℎ𝑎𝑟𝑡 𝑓𝑟𝑜𝑚 𝑅̅ 𝑈𝐶𝐿 = 𝑥̿ + 𝐴2 𝑟̅ 𝐶𝐿 = 𝑥̿ 𝐿𝐶𝐿 = 𝑥̿ − 𝐴2 𝑟̅ 𝑆 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 45 𝑚 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝑚𝑒𝑎𝑛 𝜇𝑅 𝑖𝑠 𝑅̅ = 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝜎, 𝑖𝑠 𝜎̂ = 1 ∑ 𝑅𝑖 𝑚 𝑖=1 𝑅̅ 𝑑2 𝐶𝑜𝑛𝑠𝑡𝑎𝑛𝑡 𝑑2 𝑖𝑠 𝑡𝑎𝑏𝑢𝑙𝑎𝑡𝑒𝑑 𝑓𝑜𝑟 𝑣𝑎𝑟𝑖𝑜𝑢𝑠 𝑠𝑎𝑚𝑝𝑙𝑒 𝑠𝑖𝑧𝑒𝑠 3 𝑈𝐶𝐿 = 𝑋̿ + 𝑅̅ 𝑑2 √𝑛 3 𝐿𝐶𝐿 = 𝑋̿ − 𝑅̅ 𝑑2 √𝑛 3 𝐴2 = 𝑑2 √𝑛 𝑈𝐶𝐿 = 𝐷4 𝑟̅ 𝐶𝐿 = 𝑟̅ 𝐿𝐶𝐿 = 𝐷3 𝑟̅ 𝑊ℎ𝑒𝑟𝑒 𝑟̅ 𝑖𝑠 𝑡ℎ𝑒 𝑠𝑎𝑚𝑝𝑙𝑒 𝑎𝑣𝑒𝑟𝑎𝑔𝑒 𝑟𝑎𝑛𝑔𝑒. 𝐷3 𝑎𝑛𝑑 𝐷4 𝑎𝑟𝑒 𝑡𝑎𝑏𝑢𝑙𝑎𝑡𝑒𝑑 𝑓𝑜𝑟 𝑣𝑎𝑟𝑖𝑜𝑢𝑠 𝑠𝑎𝑚𝑝𝑙𝑒 𝑠𝑖𝑧𝑒𝑠 𝑀𝑜𝑣𝑖𝑛𝑔 𝑅𝑎𝑛𝑔𝑒 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 𝑚 1 ∑|𝑋𝑖 − 𝑋𝑖−1 | 𝑚−1 𝑖=2 ̅̅̅̅̅ ̅̅̅̅̅ 𝑀𝑅 𝑀𝑅 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝜎, 𝑖𝑠 𝜎̂ = = 𝑑2 1.128 𝐶𝐿, 𝑈𝐶𝐿 𝑎𝑛𝑑 𝐿𝐶𝐿 𝑓𝑜𝑟 𝑐𝑜𝑛𝑡𝑟𝑜𝑙 𝑐ℎ𝑎𝑟𝑡 𝑓𝑜𝑟 𝑖𝑛𝑑𝑖𝑣𝑖𝑑𝑢𝑎𝑙𝑠 ̅̅̅̅ 𝑚𝑟 ̅̅̅̅ 𝑚𝑟 𝑈𝐶𝐿 = 𝑥̅ + 3 = 𝑥̅ + 3 𝑑2 1.128 𝐶𝐿 = 𝑥̅ ̅̅̅̅ 𝑚𝑟 ̅̅̅̅ 𝑚𝑟 𝐿𝐶𝐿 = 𝑥̅ − 3 = 𝑥̅ − 3 𝑑2 1.128 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝑐ℎ𝑎𝑟𝑡 𝑓𝑜𝑟 𝑚𝑜𝑣𝑖𝑛𝑔 𝑟𝑎𝑛𝑔𝑒𝑠 𝑈𝐶𝐿 = 𝐷4 𝑚𝑟 ̅̅̅̅ = 3.267𝑚𝑟 ̅̅̅̅ 𝐶𝐿 = 𝑚𝑟 ̅̅̅̅ 𝐿𝐶𝐿 = 𝐷3 𝑚𝑟 ̅̅̅̅ = 0 𝑎𝑠 𝐷3 𝑖𝑠 0 𝑓𝑜𝑟 𝑛 = 2. ̅̅̅̅̅ = 𝑀𝑅 𝑆̅ 𝑐4 𝐶𝑜𝑛𝑠𝑡𝑎𝑛𝑡 𝑐4 𝑖𝑠 𝑡𝑎𝑏𝑢𝑙𝑎𝑡𝑒𝑑 𝑓𝑜𝑟 𝑣𝑎𝑟𝑖𝑜𝑢𝑠 𝑠𝑎𝑚𝑝𝑙𝑒 𝑠𝑖𝑧𝑒𝑠 𝑠̅ 𝑈𝐶𝐿 = 𝑠̅ + 3 √1 − 𝑐42 𝑐4 𝐶𝐿 = 𝑠̅ 𝑠̅ 𝐿𝐶𝐿 = 𝑠̅ − 3 √1 − 𝑐42 𝑐4 𝐸𝑠𝑡𝑖𝑚𝑎𝑡𝑒 𝑜𝑓 𝜎, 𝑖𝑠 𝜎̂ = ̅𝑋 𝑐𝑜𝑛𝑡𝑟𝑜𝑙 𝑐ℎ𝑎𝑟𝑡 𝑓𝑟𝑜𝑚 𝑆̅ 𝑠̅ 𝑈𝐶𝐿 = 𝑥̿ + 3 𝑐4 √𝑛 𝐶𝐿 = 𝑥̿ 𝑠̅ 𝐿𝐶𝐿 = 𝑥̿ − 3 𝑐4 √𝑛 𝑈𝑆𝐿 − 𝐿𝑆𝐿 𝑃𝑟𝑜𝑐𝑒𝑠𝑠 𝐶𝑎𝑝𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑅𝑎𝑡𝑖𝑜 (𝑃𝐶𝑅) = 6𝜎̂ 𝑟̅ 𝜎̂ = 𝑑2 𝑈𝑆𝐿 − 𝜇 𝜇 − 𝐿𝑆𝐿 𝑂𝑛𝑒 − 𝑠𝑖𝑑𝑒𝑑 𝑃𝐶𝑅, 𝑖𝑠 𝑃𝐶𝑅𝑘 = 𝑚𝑖𝑛 [ , ] 3𝜎 3𝜎 𝐿𝑆𝐿 − 𝜇 𝑃(𝑋 < 𝐿𝑆𝐿) = 𝑃(𝑍 < ) 𝜎 𝑈𝑆𝐿 − 𝜇 𝑃(𝑋 > 𝑈𝑆𝐿) = 𝑃(𝑍 > ) 𝜎 𝑃 𝐶ℎ𝑎𝑟𝑡 (𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 𝑓𝑜𝑟 𝑃𝑟𝑜𝑝𝑜𝑟𝑡𝑖𝑜𝑛𝑠) 𝑈 𝐶ℎ𝑎𝑟𝑡 (𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 𝑓𝑜𝑟 𝐷𝑒𝑓𝑒𝑐𝑡𝑠 𝑝𝑒𝑟 𝑈𝑛𝑖𝑡) 1 1 𝑃̅ = ∑ 𝑃𝑖 = ∑ 𝐷𝑖 𝑚 𝑚𝑛 1 ̅= 𝑈 ∑ 𝑈𝑖 𝑚 𝑝̅ (1 − 𝑝̅ ) 𝑈𝐶𝐿 = 𝑝̅ + 3√ 𝑛 𝐶𝐿 = 𝑝̅ 𝑢̅ 𝑈𝐶𝐿 = 𝑢̅ + 3√ 𝑛 𝑝̅ (1 − 𝑝̅ ) 𝐿𝐶𝐿 = 𝑝̅ − 3√ 𝑛 𝑊ℎ𝑒𝑟𝑒 𝑝̅ 𝑖𝑠 𝑡ℎ𝑒 𝑜𝑏𝑠𝑒𝑟𝑣𝑒𝑑 𝑣𝑎𝑙𝑢𝑒 𝑜𝑓 𝑎𝑣𝑒𝑟𝑎𝑔𝑒. 𝑢̅ 𝐿𝐶𝐿 = 𝑢̅ − 3√ 𝑛 𝑚 𝑚 𝑖=1 𝑖=1 𝐴𝑣𝑒𝑟𝑎𝑔𝑒 𝑅𝑢𝑛 𝐿𝑒𝑛𝑔𝑡ℎ, 𝐴𝑅𝐿 = 𝑚 𝑖=1 𝐶𝐿 = 𝑢̅ 1 𝑝 𝑊ℎ𝑒𝑟𝑒 𝑝 𝑖𝑠 𝑡ℎ𝑒 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑡ℎ𝑎𝑡 𝑎 𝑛𝑜𝑟𝑚𝑎𝑙𝑙𝑦 𝑑𝑖𝑠𝑡𝑟𝑖𝑏𝑢𝑡𝑒𝑑 𝑝𝑜𝑖𝑛𝑡 𝑓𝑎𝑙𝑙𝑠 𝑜𝑢𝑡𝑠𝑖𝑑𝑒 𝑡ℎ𝑒 𝑙𝑖𝑚𝑖𝑡𝑠 𝑤ℎ𝑒𝑛 𝑡ℎ𝑒 𝑝𝑟𝑜𝑐𝑒𝑠𝑠 𝑖𝑠 𝑖𝑛 𝑐𝑜𝑛𝑡𝑟𝑜𝑙 𝑊ℎ𝑒𝑟𝑒 𝑢̅ 𝑖𝑠 𝑡ℎ𝑒 𝑎𝑣𝑒𝑟𝑎𝑔𝑒 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑑𝑒𝑓𝑒𝑐𝑡𝑠 𝑝𝑒𝑟 𝑢𝑛𝑖𝑡 𝐶𝑈𝑆𝑈𝑀 𝐶ℎ𝑎𝑟𝑡 (𝐶𝑢𝑚𝑢𝑙𝑎𝑡𝑖𝑣𝑒 𝑆𝑢𝑚 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡) 𝑈𝑝𝑝𝑒𝑟 𝑜𝑛𝑒 − 𝑠𝑖𝑑𝑒𝑑 𝐶𝑈𝑆𝑈𝑀 𝑓𝑜𝑟 𝑝𝑒𝑟𝑖𝑜𝑑 𝑖 𝑆𝐻 (𝑖) = 𝑚𝑎𝑥[0, 𝑥̅𝑖 − (𝜇0 + 𝐾) + 𝑠𝐻 (𝑖 − 1)] 𝐿𝑜𝑤𝑒𝑟 𝑜𝑛𝑒 − 𝑠𝑖𝑑𝑒𝑑 𝐶𝑈𝑆𝑈𝑀 𝑓𝑜𝑟 𝑝𝑒𝑟𝑖𝑜𝑑 𝑖 46 𝐸𝑥𝑝𝑜𝑛𝑒𝑛𝑡𝑖𝑎𝑙𝑙𝑦 𝑊𝑒𝑖𝑔ℎ𝑡𝑒𝑑 𝑀𝑜𝑣𝑖𝑛𝑔 𝐴𝑣𝑒𝑟𝑎𝑔𝑒 𝐶𝑜𝑛𝑡𝑟𝑜𝑙 𝐶ℎ𝑎𝑟𝑡 (𝐸𝑊𝑀𝐴) (1 𝑧𝑡 = 𝜆𝑥̅𝑡 + − 𝜆)𝑧𝑡−1 𝑓𝑜𝑟 𝑒𝑎𝑐ℎ 𝑡𝑖𝑚𝑒 𝑡 𝑈𝐶𝐿 = 𝜇0 + 3 𝜎 𝜆 [1 − (1 − 𝜆)2𝑡 ] √𝑛 2 − 𝜆 𝐶𝐿 = 𝜇0 𝜆 √ [1 − (1 − 𝜆)2𝑡 ] √𝑛 2 − 𝜆 𝑅̅ 𝑆̅ 𝜇̂ 0 = 𝑋̿ 𝑎𝑛𝑑 𝜎̂ = 𝑜𝑟 𝜎̂ = 𝑑2 𝑐4 ̅̅̅̅̅ 𝑀𝑅 𝐹𝑜𝑟 𝑛 = 1, 𝜇̂ 0 = 𝑋̅ 𝑎𝑛𝑑 𝜎̂ = 1.128 𝐿𝐶𝐿 = 𝜇0 − 3 𝜎 √ 𝑆𝐿 (𝑖) = 𝑚𝑎𝑥[0, (𝜇0 − 𝐾) − 𝑥̅𝑖 + 𝑠𝐿 (𝑖 − 1)] 𝑊ℎ𝑒𝑟𝑒 𝑠𝑡𝑎𝑟𝑡𝑖𝑛𝑔 𝑣𝑎𝑙𝑢𝑒𝑠 𝑠𝐻 (0) = 𝑠𝐿 (0) = 0 𝐻 𝑖𝑠 𝑐𝑎𝑙𝑙𝑒𝑑 𝑑𝑒𝑐𝑖𝑠𝑖𝑜𝑛 𝑖𝑛𝑡𝑒𝑟𝑣𝑎𝑙 Δ 𝑅𝑒𝑓𝑒𝑟𝑒𝑛𝑐𝑒 𝑣𝑎𝑙𝑢𝑒, 𝐾 = 2 𝜇1 = 𝜇0 + Δ 𝐾 = 𝑘𝜎; 𝐻 = ℎ𝜎 𝑠𝐻 (𝑖) 𝜇0 + 𝐾 + , 𝑖𝑓 𝑠𝐻 (𝑖) > 𝐻 𝑛𝐻 𝜇̂ = 𝑠𝐿 (𝑖) 𝜇0 − 𝐾 − , 𝑖𝑓 𝑠𝐿 (𝑖) > 𝐻 { 𝑛𝐿 47 OTHERS Computing lambda and the consistency index −n CI = n −1 Consistency Ratio CI CR = RI Fixed cost Price/unit – Variable cost/unit f = s−v P(loss) = P(demand < break-even) P(profit) = P(demand > break-even) Break - even point (units) = The input to one stage is also the output from another stage sn–1 = Output from stage n The transformation function tn = Transformation function at stage n General formula to move from one stage to another using the transformation function sn–1 = tn (sn, dn) The total return at any stage fn = Total return at stage n Transformation Functions sn−1 = (an sn ) + (bn d n ) + cn Return Equations rn = (an sn ) + (bn d n ) + cn Probability of breaking even break - even point − Z= Price Variable cost EMV = – (Mean demand) unit unit − Fixed costs Using the unit normal loss integral, EOL can be K(break - even point – X)for X = BEP Opportunity Loss = computed using $0for X BEP EOL = KN(D) where EOL = expected opportunity loss K = loss per unit when sales are below the break-even point K = loss per unit when sales are below the breakX = sales in units even point = standard deviation of the distribution N(D) = value for the unit normal loss integral for a given value of D – break even point D= a AB = b (d c ad ae e ) = bd be = C cd ce d (a b c ) e = (ad + be + cf ) f a b e c d g f ae + bg af + bh = h ce + dg cf + dh a b c d Determinant Value = (a)(d) – (c)(b) a b c d e f g h i Determinant Value = aei + bfg + cdh – gec – hfa – idb Numerical value of numerator determinan t X= Numerical value of denominator determinan t 48 a b Original matrix = c d Determinant value of original matrix = ad − cb d − c M atrix of cofactors = − b a d − b Adjoint of the matrix = − c a Equation for a line Y = a + bX where b is the slope of the line Given any two points (X1, Y1) and (X2, Y2) Change in Y Y Y2 – Y1 b= = = Change in X X X 2 – X1 d −1 a b = ad − cb −c c d ad − cb −b ad − cb a ad − cb For the Nonlinear function Y = X2 – 4X + 6 Find the slope using two points and this equation Change in Y Y Y2 – Y1 b= = = Change in X X X 2 – X1 Y1 = aX 2 + bX + c Y =C Y = 0 Y2 = a( X + X ) 2 + b( X + X ) + c Y = Xn Y = nX n −1 Y = Y2 − Y1 = b(X ) + 2aX (X ) + c(X ) 2 Y b(X ) + 2aX (X ) + c(X ) 2 = X X X (b + 2aX + cX ) = = b + 2aX + cX X Y = cX n Y = cnX n −1 1 Xn Y = g ( x ) + h( x ) Y = Total cost = (Total ordering cost) + (Total holding cost) + (Total purchase cost) D Q TC = C o + C h + DC Q 2 Q = order quantity D = annual demand Co = ordering cost per order Ch = holding cost per unit per year C = purchase (material) cost per unit Y= Y = g ( x ) − h( x ) Economic Order Quantity dTC – DCo Ch = + dQ Q2 2 2DCo Q= Ch −n X n +1 Y = g ( x) + h( x) Y = g ( x) − h( x) d 2TC DCo = 3 dQ 2 Q 49