Here

advertisement
𝑃(𝐡) = 𝑃(𝐡 ∩ 𝐴) + 𝑃(𝐡 ∩ 𝐴′ ) = 𝑃(𝐡|𝐴)𝑃(𝐴) + 𝑃(𝐡|𝐴′ )𝑃(𝐴′)
𝑃(𝐴 ∩ 𝐡) = 𝑃(𝐡|𝐴)𝑃(𝐴) = 𝑃(𝐴|𝐡)𝑃(𝐡)
𝑃(𝐴 ∪ 𝐡) = 𝑃(𝐴) + 𝑃(𝐡) − 𝑃(𝐴 ∩ 𝐡) ≤ 1
𝐸1 ∩ 𝐸2 = 𝐸1 ∩ 𝐸1′ = 0 – Mutually exclusive
𝐸(πœƒΜ‚) − πœƒ = 0 – Unbiased
𝑉(𝑋) = 𝐸(𝑋 2 ) − (𝐸(𝑋))2
Two events are independent if:
(1) 𝑃(𝐴|𝐡) = 𝑃(𝐴)
(2) 𝑃(𝐡|𝐴) = 𝑃(𝐡)
(3) 𝑃(𝐴 ∩ 𝐡) = 𝑃(𝐴)𝑃(𝐡)
Probability density function:
(1) 𝑓(π‘₯) ≥ 0
∞
(2) ∫−∞ 𝑓(π‘₯)𝑑π‘₯ = 1
Joint probability mass function:
(1) π‘“π‘‹π‘Œ (π‘₯, 𝑦) ≥ 0
(2) ∑π‘₯ ∑𝑦 π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 1
(3) π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 𝑃(𝑋 = π‘₯, π‘Œ = 𝑦)
𝐸(𝑋𝑖 ) = 𝑛𝑝𝑖 , 𝑉(𝑋𝑖 ) = 𝑛𝑝𝑖 = (1 − 𝑝𝑖 )
Marginal probability mass function:
𝑓𝑋 (π‘₯) = 𝑃(𝑋 = π‘₯) = ∑ π‘“π‘‹π‘Œ (π‘₯, 𝑦)
𝑦
π‘“π‘Œ (𝑦) = 𝑃(π‘Œ = 𝑦) = ∑ π‘“π‘‹π‘Œ (π‘₯, 𝑦)
π‘₯
(3) 𝑃(π‘Ž ≤ 𝑋 ≤ 𝑏) = ∫π‘Ž 𝑓(π‘₯)𝑑π‘₯
Probability mass function:
(1) 𝑓(π‘₯𝑖 ) ≥ 0
(2) ∑𝑛𝑖=1 𝑓(π‘₯𝑖 ) = 1
(3) 𝑓(π‘₯𝑖 ) = 𝑃(𝑋 = π‘₯𝑖 )
Mean and variance of the discrete rv:
Conditional probability mass function of Y given X=x is:
π‘“π‘‹π‘Œ (π‘₯, 𝑦)
π‘“π‘Œ|π‘₯ (𝑦) =
,
π‘“π‘œπ‘Ÿ 𝑓𝑋 (π‘₯) > 0
𝑓𝑋 (π‘₯)
Because a conditional probability mass function π‘“π‘Œ|π‘₯ (𝑦) is a probability mass
function, the following prop. are satisfied:
(1) π‘“π‘Œ|π‘₯ ( 𝑦) ≥ 0
(2) ∑𝑦 π‘“π‘Œ|π‘₯ (𝑦) = 1
(3) 𝑃(π‘Œ = 𝑦|𝑋 = π‘₯) = π‘“π‘Œ|π‘₯ (𝑦)
Conditional mean, variance:
πœ‡ = 𝐸(𝑋) = ∑ π‘₯𝑓(π‘₯) = ∫ π‘₯𝑓(π‘₯)𝑑π‘₯
πœ‡π‘Œ|π‘₯ = 𝐸(π‘Œ|π‘₯) = ∑ π‘¦π‘“π‘Œ|π‘₯ (𝑦)
𝑏
∞
π‘₯
−∞
∞
𝜎 2 = 𝑉(𝑋) = ∑ π‘₯ 2 𝑓(π‘₯) − πœ‡ 2 = ∫ π‘₯ 2 𝑓(π‘₯)𝑑π‘₯ − πœ‡ 2
π‘₯
−∞
Discrete uniform distribution, all n has equal probability:
1
𝑏+π‘Ž
(𝑏 − π‘Ž + 1)2 − 1
𝑓(π‘₯𝑖 ) = , 𝐸(𝑋) =
, 𝑉(𝑋) =
𝑛
2
12
Continuous uniform distribution:
1
(π‘Ž + 𝑏)
(𝑏 − π‘Ž)2
𝑓(π‘₯) =
, 𝐸(𝑋) =
, 𝑉(𝑋) =
(𝑏 − π‘Ž)
2
12
A Bernoulli trial (Binomial distribution, success or failure)
𝑛 π‘₯
𝑓(π‘₯) = ( ) 𝑝 (1 − 𝑝)𝑛−π‘₯ (π‘”π‘œπ‘œπ‘‘ π‘“π‘œπ‘Ÿ 𝑛𝑝 > 5 π‘Žπ‘›π‘‘ 𝑛(1 − 𝑝) > 5)
π‘₯
𝐸(𝑋) = 𝑛𝑝,
𝑉(π‘₯) = 𝑛𝑝(1 − 𝑝)
Geometric distribution:
𝑓(π‘₯) = (1 − 𝑝)π‘₯−1 𝑝,
π‘₯ = 1,2, …
1
2
πœ‡ = 𝐸(𝑋) = ,
𝜎 = 𝑉(π‘₯) = (1 − 𝑝)/𝑝2
𝑝
Negative binomial distribution nr of trials until r successes:
π‘₯−1
) (1 − 𝑝)π‘₯−π‘Ÿ π‘π‘Ÿ , π‘Ÿ = 1, 2, … , π‘₯ = π‘Ÿ, π‘Ÿ + 1, π‘Ÿ + 2, …
𝑓(π‘₯) = (
π‘Ÿ−1
πœ‡ = 𝐸(𝑋) = π‘Ÿ/𝑝, 𝜎 2 = 𝑉(𝑋) = π‘Ÿ(1 − 𝑝)/𝑝2
Hypergeometric distribution:
N objects contains K objects = successes, N-K objects = failures. A sample of size
n, X = # of successes in the sample. p=K/N, (good for n/N < 0.1)
(𝐾)(𝑁−𝐾)
𝑓(π‘₯) = π‘₯ 𝑁𝑛−π‘₯ , π‘₯ = max{0, 𝑛 + 𝐾 − 𝑁} π‘‘π‘œ min{𝐾, 𝑛}
(𝑛 )
𝑁−𝑛
)
πœ‡ = 𝐸(𝑋) = 𝑛𝑝, 𝜎 2 = 𝑉(𝑋) = 𝑛𝑝(1 − 𝑝) (
𝑁−1
Poisson distribution:
𝑒 −πœ† πœ†π‘₯
𝑓(π‘₯) =
, π‘₯ = 0, 1, 2, … (π‘”π‘œπ‘œπ‘‘ π‘“π‘œπ‘Ÿ πœ† > 5)
π‘₯!
πœ‡ = 𝐸(𝑋) = 𝑉(𝑋) = πœ†
Cumulative distribution function:
π‘₯
𝐹(π‘₯) = 𝑃(𝑋 ≤ π‘₯) = ∫ 𝑓(𝑒)𝑑𝑒,
π‘“π‘œπ‘Ÿ − ∞ < π‘₯ < ∞
−∞
Expected value of a function of a continuous random value:
∞
𝐸[β„Ž(π‘₯)] = ∫ β„Ž(π‘₯)𝑓(π‘₯)𝑑π‘₯
−∞
Normal distribution:
−(π‘₯−πœ‡)2
1
𝑓(π‘₯) =
𝑒 2𝜎2 , 𝑁(πœ‡, 𝜎 2 )
√2πœ‹πœŽ
Standardizing to calculate a probability:
𝑋−πœ‡ π‘₯−πœ‡
) = 𝑃(𝑍 ≤ 𝑧)
𝑃(𝑋 ≤ π‘₯) = 𝑃 (
≤
𝜎
𝜎
Normal approximation to the binomial distribution:
𝑋 − 𝑛𝑝
𝑍=
√𝑛𝑝(1 − 𝑝)
is approximately a standard normal rv. To approximate a binomial probability
with normal distribution
π‘₯ + 0.5 − 𝑛𝑝
)
𝑃(𝑋 ≤ π‘₯) = 𝑃(𝑋 ≤ π‘₯ + 0.5) ≅ 𝑃 (𝑍 ≤
√𝑛𝑝(1 − 𝑝)
and
π‘₯ − 0.5 − 𝑛𝑝
𝑃(𝑋 ≤ π‘₯) = 𝑃(π‘₯ − 0.5 ≤ 𝑋) ≅ 𝑃 (
≤ 𝑍)
√𝑛𝑝(1 − 𝑝)
The approximation is good for 𝑛𝑝 > 5 and 𝑛(1 − 𝑝) > 5
Normal approximation to the Poisson distribution:
𝑋−πœ†
𝑍=
,
πœ†>5
√πœ†
Exponential distribution:
𝑓(π‘₯) = λe−λx , for 0 ≤ x ≤ ∞
πœ‡ = 𝐸(𝑋) = 1/πœ†, 𝜎 2 = 𝑉(𝑋) = 1/πœ†2
𝑦
2
2
πœŽπ‘Œ|π‘₯
= 𝑉(π‘Œ|π‘₯) = ∑ 𝑦 2 π‘“π‘Œ|π‘₯ (𝑦) − πœ‡π‘Œ|π‘₯
𝑦
For discrete random variable X & Y, if one of the following is true, the other are
also true and X, Y are independent:
(1) π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 𝑓𝑋 (π‘₯)π‘“π‘Œ (𝑦), π‘“π‘œπ‘Ÿ π‘Žπ‘™π‘™ π‘₯ & 𝑦
(2) π‘“π‘Œ|π‘₯ (𝑦) = π‘“π‘Œ (𝑦), π‘“π‘œπ‘Ÿ π‘Žπ‘™π‘™ π‘₯ & 𝑦 π‘€π‘–π‘‘β„Ž 𝑓𝑋 (π‘₯) > 0
(3) 𝑓𝑋|𝑦 (π‘₯) = 𝑓𝑋 (π‘₯), π‘“π‘œπ‘Ÿ π‘Žπ‘™π‘™ π‘₯ & 𝑦 π‘€π‘–π‘‘β„Ž π‘“π‘Œ (𝑦) > 0
(4) 𝑃(𝑋 ∈ 𝐴, π‘Œ ∈ 𝐡) = 𝑃(𝑋 ∈ 𝐴)𝑃(π‘Œ ∈ 𝐡)
Covarience:
∑ π‘₯𝑖
π‘π‘œπ‘£(𝑋, π‘Œ) = πœŽπ‘‹π‘Œ = 𝐸[(𝑋 − πœ‡π‘‹ )(π‘Œ − πœ‡π‘Œ )] = 𝐸(π‘‹π‘Œ) − πœ‡π‘‹ πœ‡π‘Œ , πœ‡π‘‹ =
𝑛
Correlation:
π‘π‘œπ‘£(𝑋, π‘Œ)
πœŽπ‘‹π‘Œ
πœŒπ‘‹π‘Œ =
=
, −1 ≤ πœŒπ‘‹π‘Œ ≤ 1
√𝑉(𝑋)𝑉(π‘Œ) πœŽπ‘‹ πœŽπ‘Œ
If X & Y are independent random variables, πœŽπ‘‹π‘Œ = πœŒπ‘‹π‘Œ = 0 (⇍)
𝑛!
Combinations: (π‘›π‘Ÿ) =
π‘Ÿ!(𝑛−π‘Ÿ)!
𝑛!
Permutations: π‘ƒπ‘Ÿπ‘› = (𝑛−π‘Ÿ)!
Μ‚ is its standard deviation, given by:
The standard error of an estimator Θ
Μ‚)
𝜎ΘΜ‚ = √𝑉(Θ
Μ‚ of the parameter θ is:
The mean squared error of the estimator Θ
Μ‚ ) = 𝐸(Θ
Μ‚ − θ)2
𝑀𝑆𝐸(Θ
Likelihood function: 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 )
Suppose a random experiment consists of a series of n trials. Assume that
1)
The result of each trial is classified into one of k classes.
2)
The probability of a trial generating a result in class 1 to class k is
constant over trials and equal to p1 to pk.
3)
The trials are independent
The random variables X1, X2,…,Xn that denote the number of trials that result in
class 1 to class k have a multinomial distribution and the joint probability mass
function is
𝑛!
π‘₯
π‘₯
π‘₯
𝑃(𝑋1 = π‘₯1 , 𝑋2 = π‘₯2 , … , π‘‹π‘˜ = π‘₯π‘˜ ) =
𝑝 1 𝑝 2 … π‘π‘˜ π‘˜
π‘₯1 ! π‘₯2 ! … π‘₯π‘˜ ! 1 2
π‘“π‘œπ‘Ÿ π‘₯1 + π‘₯2 + β‹― + π‘₯π‘˜ = 𝑛 π‘Žπ‘›π‘‘ 𝑝1 + 𝑝2 + β‹― + π‘π‘˜ = 1
𝐸(𝑋𝑖 ) = 𝑛𝑝𝑖 ,
𝑉(𝑋𝑖 ) = 𝑛𝑝𝑖 (1 − 𝑝𝑖 )
If x1, x2, … , xn is a sample of n observations, the sample variance is:
(∑𝑛𝑖=1 π‘₯𝑖 )2
𝑛
2
∑𝑛𝑖=1(π‘₯𝑖 − π‘₯Μ… )2 ∑𝑖=1 π‘₯𝑖 −
𝑛
𝑠2 =
=
𝑛−1
𝑛−1
Confidence interval on the mean, variance known:
π‘₯Μ… − 𝑧𝛼/2 𝜎/√𝑛 ≤ πœ‡ ≤ π‘₯Μ… + 𝑧𝛼/2 𝜎/√𝑛
𝑧𝛼/2 𝜎 2
𝑋̅ − πœ‡
) , 𝐸 = |π‘₯Μ… − πœ‡|
𝑍=
,
πΆβ„Žπ‘œπ‘–π‘π‘’ π‘œπ‘“ 𝑛 = (
𝐸
𝜎/√𝑛
Confidence interval on the mean, variance unknown:
π‘₯Μ… − 𝑑𝛼/2,𝑛−1 𝑠/√𝑛 ≤ πœ‡ ≤ π‘₯Μ… + 𝑑𝛼/2,𝑛−1 𝑠/√𝑛
𝑋̅ − πœ‡
𝑇=
𝑆/√𝑛
Random sample normal disr. mean=µ, var=σ2, S2=sample var.
(𝑛 − 1)𝑆 2
𝑋2 =
β„Žπ‘Žπ‘  πœ’ 2 𝑑𝑖𝑠𝑑. π‘€π‘–π‘‘β„Ž 𝑛 − 1 π‘‘π‘’π‘”π‘Ÿπ‘’π‘’π‘  π‘œπ‘“ π‘“π‘Ÿπ‘’π‘’π‘‘π‘œπ‘š
𝜎2
Ci on variance, s2=sample variance, σ2 unknown
(𝑛 − 1)𝑠 2
(𝑛 − 1)𝑠 2
≤ 𝜎2 ≤ 2
πœ’π›Ό2,𝑛−1
πœ’1−𝛼,𝑛−1
2
2
Lower and upper confidence bounds on σ2:
(𝑛 − 1)𝑠 2
(𝑛 − 1)𝑠 2
≤ 𝜎 2 π‘Žπ‘›π‘‘ 𝜎 2 ≤ 2
2
πœ’π›Ό,𝑛−1
πœ’1−𝛼,𝑛−1
Binomial proportion:
If n is large, the distribution of
𝑍=
𝑋 − 𝑛𝑝
√𝑛𝑝(1 − 𝑝)
=
𝑃̂ − 𝑝
√𝑝(1 − 𝑝)
𝑛
is approximately standard normal.
Ci on binomial proportion (obs, lower, upper change zα/2 to zα):
𝑝̂ (1 − 𝑝̂ )
𝑝̂ (1 − 𝑝̂ )
≤ 𝑝 ≤ 𝑝̂ + 𝑧𝛼/2 √
𝑛
𝑛
Sample size for a specified error on binomial proportion:
𝑧𝛼 2
𝑛 = ( 2 ) 𝑝(1 − 𝑝), 𝑛 𝑖𝑠 max π‘“π‘œπ‘Ÿ 𝑝 = 0.5
𝐸
Prediction int. on a single future observation from norm. dist:
π‘₯Μ… − 𝑑𝛼/2,𝑛−1 𝑠√1 + 1/𝑛 ≤ 𝑋𝑛+1 ≤ π‘₯Μ… + 𝑑𝛼/2,𝑛−1 𝑠√1 + 1/𝑛
Ci, difference in mean, variances known:
𝑝̂ − 𝑧𝛼/2 √
𝜎12 𝜎22
𝜎12 𝜎22
π‘₯Μ…1 − π‘₯Μ… 2 − 𝑧𝛼/2 √ +
≤ πœ‡1 − πœ‡2 ≤ π‘₯Μ…1 − π‘₯Μ… 2 + 𝑧𝛼/2 √ +
𝑛1 𝑛2
𝑛1 𝑛2
for one-sided, change zα/2 to zα.
Sample size for a c.i. on difference in mean, variances known:
𝑧𝛼/2 2 2
) (𝜎1 + 𝜎22 )
𝑛=(
𝐸
Ci Case 1, difference in mean, variance unknown & equal:
1
1
π‘₯Μ…1 − π‘₯Μ… 2 − 𝑑𝛼/2,𝑛1+𝑛2−2 𝑠𝑝 √ + ≤ πœ‡1 − πœ‡2
𝑛1 𝑛2
1
1
≤ π‘₯Μ…1 − π‘₯Μ… 2 + 𝑑𝛼/2,𝑛1+𝑛2−2 𝑠𝑝 √ +
𝑛1 𝑛2
(𝑛1 − 1)𝑆12 + (𝑛2 − 1)𝑆22
𝑛1 + 𝑛2 − 2
Ci Case 2, difference in mean, variance unknown, not equal:
𝑆𝑝2 =
𝑠12 𝑠22
𝑠12 𝑠22
π‘₯Μ…1 − π‘₯Μ… 2 − 𝑑𝛼/2,𝜈 √ + ≤ πœ‡1 − πœ‡2 ≤ π‘₯Μ…1 − π‘₯Μ… 2 + 𝑑𝛼/2,𝜈 √ +
𝑛1 𝑛2
𝑛1 𝑛2
2
𝑠2 𝑠2
( 1 + 2)
𝑛1 𝑛2
𝜈= 2
(𝑠1 /𝑛1)2 (𝑠22 /𝑛2 )2
+
𝑛1 − 1
𝑛2 − 1
ν is degrees of freedom for tα/2,if not integer, round down.
Ci for µD from paired samples:
𝑑̅ − 𝑑𝛼/2,𝑛−1 𝑠𝐷 /√𝑛 ≤ πœ‡π· ≤ 𝑑̅ + 𝑑𝛼/2,𝑛−1 𝑠𝐷 /√𝑛
Approximate ci on difference in population proportions:
𝑝̂1 (1 − 𝑝̂1 ) 𝑝̂ 2 (1 − 𝑝̂ 2 )
𝑝̂1 − 𝑝̂ 2 − 𝑧𝛼/2 √
+
≤ 𝑝1 − 𝑝2
𝑛1
𝑛2
𝑝̂1 (1 − 𝑝̂1 ) 𝑝̂ 2 (1 − 𝑝̂ 2 )
≤ 𝑝̂1 − 𝑝̂ 2 + 𝑧𝛼/2 √
+
𝑛1
𝑛2
Hypothesis test:
1.
Choose parameter of interest
2.
H0:
3.
H1:
4.
α=
5.
The test statistic is
6.
Reject H0 at α= … if
7.
Computations
8.
Conclusions
Test on mean, variance known
𝑋̅ − πœ‡0
𝐻0 : πœ‡ = πœ‡0 ,
𝑍0 =
𝜎/√𝑛
Alternative hypothesis
Rejection criteria
H1: µ ≠ µ0
z0 > zα/2 or z0 < -zα/2
H1: µ > µ0
z0 > zα
H1: µ < µ0
z0 < -zα
Test on mean, variance unknown
𝑋̅ − πœ‡0
𝐻0 : πœ‡ = πœ‡0 ,
𝑇0 =
𝑆/√𝑛
Alternative hypothesis
Rejection criteria
H1: µ ≠ µ0
t0 > tα/2,n-1 or t0 < -tα/2, n-1
H1: µ > µ0
t0 > tα,n-1
H1: µ < µ0
t0 < -tα,n-1
Test in the variance of a normal distribution:
(𝑛 − 1)𝑆 2
𝐻0 : 𝜎 2 = 𝜎02 ,
πœ’02 =
𝜎02
Alternative hypothesis
Rejection criteria
2
2
H1: σ2 ≠𝜎02
πœ’02 > πœ’π›Ό/2,𝑛−1
π‘œπ‘Ÿ πœ’02 < −πœ’π›Ό/2,𝑛−1
2
H1: σ2 > 𝜎02
πœ’02 > πœ’π›Ό,𝑛−1
2
H1: σ2 < 𝜎02
πœ’02 < −πœ’π›Ό,𝑛−1
Approximate test on a binomial proportion:
𝑋 − 𝑛𝑝0
𝐻0 : 𝑝 = 𝑝0 ,
𝑍0 =
√𝑛𝑝0 (1 − 𝑝0 )
Alternative hypothesis
Rejection criteria (**)
H1: p ≠ p0
z0 > zα/2 or z0 < -zα/2
H1: p > p0
z0 > zα
H1: p < p0
z0 < -zα
App. Sample size for a 2-sided test on a binomial proportion:
2
𝑧𝛼/2 √𝑝0 (1 − 𝑝0 ) + 𝑧𝛽 √𝑝(1 − 𝑝)
] , π‘“π‘œπ‘Ÿ 1 − 𝑠𝑖𝑑𝑒𝑑 𝑒𝑠𝑒 𝑧𝛼
𝑛=[
𝑝 − 𝑝0
Test on the differens in mean, variance known
𝑋̅1 − 𝑋̅2 − βˆ†0
𝐻0 : πœ‡1 − πœ‡2 = βˆ†0 ,
𝑍0 =
𝜎2 𝜎2
√ 1+ 2
𝑛1 𝑛2
Alternative hypothesis
Rejection criteria
H1: πœ‡1 − πœ‡2 ≠ βˆ†0
z0 > zα/2 or z0 < -zα/2
H1: πœ‡1 − πœ‡2 > βˆ†0
z0 > zα
H1: πœ‡1 − πœ‡2 < βˆ†0
z0 < -zα
Sample size, 1-sided test on difference in mean, with power of at least 1-β,
n1=n2=n, variance known:
2
(𝑧𝛼 + 𝑧𝛽 ) (𝜎12 + 𝜎22)
𝑛=
(βˆ† − βˆ†0 )2
Tests on diff. in mean, variances unknown and equal:
𝑋̅1 − 𝑋̅2 − βˆ†0
𝐻0 : πœ‡1 − πœ‡2 = βˆ†0 ,
𝑇0 =
1
1
𝑆𝑝 √ +
𝑛1 𝑛2
Alternative hypothesis
Rejection criteria
H1: πœ‡1 − πœ‡2 ≠ βˆ†0
𝑑0 > 𝑑𝛼/2,𝑛1+𝑛2−2 π‘œπ‘Ÿ
𝑑0 < −𝑑𝛼/2,𝑛1+𝑛2−2
H1: πœ‡1 − πœ‡2 > βˆ†0
𝑑0 > 𝑑𝛼,𝑛1 +𝑛2−2
H1: πœ‡1 − πœ‡2 < βˆ†0
𝑑0 < −𝑑𝛼,𝑛1+𝑛2−2
Tests on diff. in mean, variances unknown and not equal:
𝑋̅1 − 𝑋̅2 − βˆ†0
If 𝐻0 : πœ‡1 − πœ‡2 = βˆ†0 is true, the statistic 𝑇0 =
𝑆2 𝑆2
√ 1+ 2
𝑛1 𝑛2
is distributed app. as t with ν degrees of freedom, ~𝑑(𝜈)
Paired t-test:
Μ… − π›₯0
𝐷
πœ‡π·
πœ‡1 − πœ‡2
𝐻0 : πœ‡π· = πœ‡1 − πœ‡2 = Δ0 ,
𝑇0 =
,
𝑑=
=
𝜎𝐷 √𝜎12 + 𝜎22
𝑆𝐷 /√𝑛
Alternative hypothesis
Rejection criteria
H1: πœ‡π· ≠ Δ0
t0 > tα/2,n-1 or t0 < -tα/2, n-1
H1: πœ‡π· > Δ0
t0 > tα,n-1
H1: πœ‡π· < Δ0
t0 < -tα,n-1
Approximate tests on the difference of two population proportions:
𝑃̂1 − 𝑃̂2
𝑋1 + 𝑋2
𝐻0 : 𝑝1 = 𝑝2 , 𝑍0 =
, 𝑃̂ =
, 𝑠𝑒𝑒 (∗∗) ↑
𝑛1 + 𝑛2
1
1
√𝑃̂ (1 − 𝑃̂ ) ( + )
𝑛1 𝑛2
Goodness of fit:
π‘˜ (𝑂 − 𝐸 )2
𝑖
𝑖
2
𝑋02 = ∑
> πœ’π›Ό,π‘˜−𝑝−1
𝐸𝑖
𝑖=1
Expected frequency: 𝐸𝑖 = 𝑛𝑝𝑖 , 𝑝𝑖 = 𝑃(𝑋 = π‘₯) = 𝑓(π‘₯)
The power of a test: = 1 − 𝛽
Δ − Δ0
𝛽 = Φ zα/2 −
σ2
√ 1
n1
+
Δ − Δ0
− Φ −zα/2 −
σ22
σ2 σ2
√ 1+ 2
n1 n2
n2
(
)
(
)
The P-value is the smallest level of significance that would lead to rejection of
the null hypothesis H0 with the given data.
2[1 − Φ(|𝑧0 |)] π‘“π‘œπ‘Ÿ π‘Ž π‘‘π‘€π‘œ − π‘‘π‘Žπ‘–π‘™π‘’π‘‘ 𝑑𝑒𝑠𝑑: π»π‘œ : πœ‡ = πœ‡0
𝐻1 : πœ‡ ≠ πœ‡0
𝐻1: πœ‡ > πœ‡0
𝑃 = {1 − Φ(z0 ) π‘“π‘œπ‘Ÿ π‘Ž π‘’π‘π‘π‘’π‘Ÿ − π‘‘π‘Žπ‘–π‘™π‘’π‘‘ 𝑑𝑒𝑠𝑑: π»π‘œ : πœ‡ = πœ‡0
Φ(z0 )
π‘“π‘œπ‘Ÿ π‘Ž π‘™π‘œπ‘€π‘’π‘Ÿ − π‘‘π‘Žπ‘–π‘™π‘’π‘‘ 𝑑𝑒𝑠𝑑: π»π‘œ : πœ‡ = πœ‡0
𝐻1 : πœ‡ < πœ‡0
Fitted or estimated regression line: 𝑦̂ = 𝛽̂0 + 𝛽̂1 π‘₯
𝑛
𝑛
1
1
𝛽̂0 = 𝑦̅ − 𝛽̂1 π‘₯Μ… ,
𝑦̅ = ( ) ∑ 𝑦𝑖 ,
π‘₯Μ… = ( ) ∑ π‘₯𝑖 , 𝑆𝑆𝑇 = 𝑆𝑆𝑅 + 𝑆𝑆𝐸
𝑛
𝑛
𝑖=1
𝑖=1
(∑𝑛𝑖=1 𝑦𝑖 )(∑𝑛𝑖=1 π‘₯𝑖 )
𝑛
𝑛
𝑆π‘₯𝑦 ∑𝑖=1 𝑦𝑖 π‘₯𝑖 −
𝑛
𝛽̂1 =
=
, 𝑒𝑖 = 𝑦𝑖 − 𝑦̂𝑖 , 𝑆𝑆𝐸 = ∑ 𝑒𝑖2
𝑛
2
(∑
)
π‘₯
𝑆π‘₯π‘₯
𝑖
𝑖=1
∑𝑛𝑖=1 π‘₯𝑖2 − 𝑖=1
𝑛
𝑛
𝑛
𝑆𝑆𝐸
πœŽΜ‚ 2 =
, 𝑆𝑆𝐸 = 𝑆𝑆𝑇 − 𝛽̂1 𝑆π‘₯𝑦 , 𝑆𝑆𝑇 = ∑ (𝑦𝑖 −𝑦̅)2 = ∑ 𝑦𝑖2 − 𝑛𝑦̅ 2
𝑛−2
𝑖=1
𝑖=1
Ci: 𝛽1 ∈ (𝛽̂1 ± 𝑑𝛼/2,𝑛−2 √
Μ‚2
𝜎
𝑆π‘₯π‘₯
𝑅2 =
𝑆𝑆𝑅
𝑆𝑆𝑇
=1−
𝑆𝑆𝐸
𝑆𝑆𝑇
1
π‘₯Μ… 2
𝑛
𝑆π‘₯π‘₯
) , 𝛽0 ∈ (𝛽̂0 ± 𝑑𝛼/2,𝑛−2 √πœŽΜ‚ 2 [ +
, 𝑆𝑆𝑅 = 𝛽̂1 𝑆π‘₯𝑦 , 𝐹0 =
𝑆𝑆𝑅
1
𝑆𝑆𝐸
𝑛−2
=
𝑀𝑆𝐸
𝑀𝑆𝐸
])
Download