Chapter 3 Descriptive Statistics Population Mean Sample Mean

advertisement
Sample Standard Deviation
Chapter 3 Descriptive Statistics
Population Mean
πœ‡=
∑π‘₯ π‘₯1 + π‘₯2 + π‘₯3 + β‹― + π‘₯𝑁
=
𝑁
𝑁
Sample Mean
π‘₯=
∑π‘₯ π‘₯1 + π‘₯2 + π‘₯3 + β‹― + π‘₯𝑛
=
𝑛
𝑛
∑(π‘₯ − π‘₯)2
𝑠=√
𝑛−1
Computational Formulas for Population Variance
and Standard Deviation
𝜎2 =
(∑ π‘₯ )2
𝑁
𝑁
∑ π‘₯2 −
Interquartile Range
Q3 – Q1
𝜎 = √𝜎 2
Sum of Deviations from the Arithmetic Mean is
Always Zero
∑(π‘₯ − πœ‡) = 0
Computational Formulas for Sample Variance and
Standard Deviation
𝑠2 =
(∑ π‘₯ )2
𝑛
𝑛−1
∑ π‘₯2 −
Mean Absolute Deviation
𝑠 = √𝑠 2
∑|π‘₯ − πœ‡|
𝑀𝐴𝐷 =
𝑁
z Score
𝑧=
Population Variance
𝜎2 =
∑(π‘₯ − πœ‡)2
𝑁
Coefficient of Variation
𝑐𝑣 =
Population Standard Deviation
∑(π‘₯ − πœ‡)2
𝜎=√
𝑁
Empirical Rule*
Distance from the Mean
πœ‡ ± 1𝜎
πœ‡ ± 2𝜎
πœ‡ ± 3𝜎
1
π‘˜2
proportion of the values.
Assumption: k > 1
Sample Variance
𝑠2 =
∑(π‘₯ − π‘₯)2
𝑛−1
∑𝑓 𝑀 ∑𝑓 𝑀
=
∑𝑓
𝑁
𝑓1 𝑀1 + 𝑓2 𝑀2 + β‹― + 𝑓𝑖 𝑀𝑖
=
𝑓1 + 𝑓2+ β‹― + 𝑓𝑖
πœ‡π‘”π‘Ÿπ‘œπ‘’π‘π‘’π‘‘ =
Values within the Distance
68%
95%
99.7%
Chebyshev’s Theorem
Within k standard deviations of the mean, πœ‡ ± π‘˜πœŽ, lie at
1−
𝜎
(100)
πœ‡
Mean of Grouped Data
*Based on the assumption that the data are
approximately normally distributed.
least
π‘₯− πœ‡
𝜎
where
i = the number of classes
f = class frequency
M = class midpoint
N = total frequencies (total number of data values)
Medium of Grouped Data
mediangrouped
where
𝑁
− 𝐹
= 𝑙 + (2
)𝑀
𝑓
l = lower endpoint of the class containing the median
w = width of the class containing the median
f = frequency of the class containing the median
F = cumulative frequency of classes preceding the
class containing the median
N = total frequencies (total number of data values)
Formulas for Population Variance and Standard
Deviation of Grouped Data
Original Formula
Computational Version
𝜎2 =
∑ 𝑓 (𝑀 − πœ‡)
𝑁
2
𝜎2 =
(∑ 𝑓𝑀 )2
𝑁
𝑁
∑ 𝑓𝑀2 −
𝜎 = √𝜎 2
where
f = frequency
M = class midpoint
N = ∑𝑓, or total of the frequencies of the population
πœ‡ = grouped mean for the population
Formulas for Sample Variance and Standard
Deviation of Grouped Data
Original Formula
Computational Version
2
𝑠2 =
∑ 𝑓 (𝑀 − π‘₯Μ… )
𝑛−1
𝑠2 =
(∑ 𝑓𝑀 )2
𝑛
𝑛−1
∑ 𝑓𝑀2 −
𝑠 = √𝑠 2
where
f = frequency
M = class midpoint
N = ∑𝑓, or total of the frequencies of the population
π‘₯Μ… = grouped mean for the sample
Coefficient of Skewness
π‘ π‘˜ =
3(πœ‡ − 𝑀𝑑)
𝜎
where
π‘ π‘˜ = coefficient of skewness
𝑀𝑑 = median
Chapter 4 Probability
Classical Method of Assigning Probabilities
𝑛𝑒
𝑃(𝐸) =
𝑁
where
N = total possible number of outcomes of an
experiment
𝑛𝑒 = the number of outcomes in which the event
occurs out of N outcomes
Range of Possible Probabilities
0 ≤ 𝑃(𝐸) ≤ 1
Probability by Relative Frequency of Occurrence
Number of Times an Event Occurred
Total Number of Opportunities for the Event to Occur
Mutually Exclusive Events X and Y
𝑃(𝑋 ∩ π‘Œ) = 0
Independent Events X and Y
𝑃(𝑋|π‘Œ) = 𝑃(𝑋) and 𝑃 (π‘Œ|𝑋) = 𝑃(π‘Œ)
Probability of the Complement of A
𝑃(𝐴′ ) = 1 − 𝑃(𝐴)
The mn Counting Rule
For an operation that can be done m ways and a second
operation that can be done n ways, the two operations
can then occur, in order, in mn ways. This rule can be
extended to cases with three or more operations.
General Law of Addition
𝑃(𝑋 ∪ π‘Œ) = 𝑃(𝑋) + 𝑃(π‘Œ) − 𝑃(𝑋 ∩ π‘Œ)
where X, Y, are events and (𝑋 ∩ π‘Œ) is the intersection of
X and Y.
Special Law of Addition
If X, Y are mutually exclusive,
𝑃(𝑋 ∪ π‘Œ) = 𝑃(𝑋) + 𝑃(π‘Œ)
General Law of Multiplication
𝑃(𝑋 ∩ π‘Œ) = 𝑃(𝑋) βˆ™ 𝑃(π‘Œ|𝑋) = 𝑃(π‘Œ) βˆ™ 𝑃(𝑋|π‘Œ)
Special Law of Multiplication
If X, Y are independent,
𝑃(𝑋 ∩ π‘Œ) = 𝑃(𝑋) βˆ™ 𝑃(π‘Œ)
Law of Conditional Probability
𝑃(𝑋 | π‘Œ) =
𝑃 (𝑋 ∩ π‘Œ)
𝑃(𝑋) βˆ™ 𝑃(π‘Œ | 𝑋)
=
𝑃(π‘Œ)
𝑃(π‘Œ)
Independent Events X, Y
If X and Y are independent events, the following must be
true:
𝑃(𝑋|π‘Œ) = 𝑃(𝑋) and 𝑃(π‘Œ|𝑋) = 𝑃(π‘Œ)
Bayes’ Rule
𝑃(𝑋𝑖 | π‘Œ) =
𝑃(𝑋𝑖 ) βˆ™ 𝑃(π‘Œ|𝑋𝑖 ) β‹―
𝑃(𝑋1 ) βˆ™ 𝑃(π‘Œ | 𝑋1 ) + 𝑃(𝑋2 ) βˆ™ 𝑃(π‘Œ | 𝑋2 ) + β‹― + 𝑃(𝑋𝑛 ) βˆ™ 𝑃(π‘Œ | 𝑋𝑛 )
Chapter 5 Discrete Distributions
Hypergeometric Formula
Mean or Expected Value of a Discrete Distribution
πœ‡ = 𝐸(π‘₯) = ∑[π‘₯ βˆ™ 𝑃(π‘₯)]
where
E(x) = long-run average
x = an outcome
P(x) = probability of that outcome
Variance of a Discrete Distribution
𝜎 2 = ∑[(π‘₯ − πœ‡)2 βˆ™ 𝑃(π‘₯)]
where
x = an outcome
P(x) = probability of a given outcome
πœ‡= mean
Standard Deviation of a Discrete Distribution
𝜎 = √∑[(π‘₯ − πœ‡)2 βˆ™ 𝑃(π‘₯)]
Assumptions of the Binomial Distribution
- The experiment involves n identical trials.
- Each trial has only two possible outcomes denoted as
success or failure.
- Each trial is independent of the previous trials.
- The terms p and q remain constant throughout the
experiment, where the term p is the probability of
getting a success on any one trial and the term q = 1 –
p is the probability of getting a failure on any one
trial.
Binomial Formula
𝑛𝐢π‘₯ . 𝑝
π‘₯
. π‘ž 𝑛−π‘₯ =
𝑛!
. 𝑝 π‘₯ . π‘ž 𝑛−π‘₯
π‘₯! (𝑛 − π‘₯)!
where
n = the number of trials (or the number being
sampled)
x = the number of successes desired
p = the probability of getting a success in one trial
π‘ž = 1 – p = the probability of getting a failure in one
trial
Mean and Standard Deviation of a Binomial
Distribution
πœ‡= 𝑛 βˆ™ 𝑝
𝜎 = √𝑛 βˆ™ 𝑝 βˆ™ π‘ž
Poisson Formula
𝑃(π‘₯) =
where
πœ†π‘₯ 𝑒 −πœ†
π‘₯!
x = 0, 1, 2, 3, …
λ = long-run average
e = 2.718281…
𝑃(π‘₯) =
𝑛𝐢π‘₯ . 𝑁−𝐴𝐢𝑛−π‘₯
𝑁𝐢𝑛
where
N = size of the population
n = sample size
A = number of successes in the population
x = number of successes in the sample; sampling is
done without replacement
Chapter 6 Continuous Distributions
Probability Density Function of a Uniform
Distribution
1
𝑓(π‘₯) = {𝑏 − 1 for a ≤ x ≤ b
0
for all other values
Mean and Standard Deviation of a Uniform
Distribution
π‘Ž+𝑏
2
πœ‡=
𝑏−π‘Ž
𝜎=
√12
Probabilities in a Uniform Distribution
𝑃(π‘₯) =
where
π‘₯2 − π‘₯1
𝑏−π‘Ž
π‘Ž ≤ π‘₯1 ≤ π‘₯2 ≤ 𝑏
z formula
𝑧=
π‘₯− πœ‡
,𝜎 ≠ 0
𝜎
Exponential Probability Density Function
𝑓(π‘₯) = λ𝑒−λπ‘₯
where
x≥0
λ>0
and e = 2.271828…
Probabilities of the Right Tail of the Exponential
Distribution
𝑃(π‘₯ ≥ π‘₯0 ) = 𝑒−λπ‘₯0
where
π‘₯0 ≥ 0
Chapter 7 Sampling and Sampling
Distributions
Determining the Value of k
π‘˜=
where
𝑁
𝑛
n = sample size
N = population size
k = size of interval for selection
Central Limit Theorem
If samples of size n are drawn randomly from a
population that has a mean of πœ‡ and a standard
deviation of 𝜎, the sample means, π‘₯, are approximately
normally distributed for sufficiently large samples (n ≥
30*) regardless of the shape of the population
distribution. If the population is normally distributed,
the sample means are normally distributed for any
sample size.
From mathematical expectation, it can be shown that
the mean of the sample means is the population mean:
πœ‡π‘₯Μ… = πœ‡
and the standard deviation of the sample means (called
the standard error of the mean) is the standard
deviation of the population divided by the square root
of the sample size:
𝜎
𝜎π‘₯Μ… =
√𝑛
z Formula for Sample Means
𝑧=
π‘₯Μ… − πœ‡
𝜎
√𝑛
z Formula for Sample Means of a Finite Population
𝑧=
π‘₯Μ… − πœ‡
𝜎 √𝑁 − 𝑛
√𝑛 𝑁 − 1
Sample Proportion
𝑝̂ =
where
π‘₯
𝑛
x = number of items in a sample that have the
characteristic
n = number of items in the sample
z Formula for Sample Proportions for n βˆ™ 𝒑 > πŸ“ and n
βˆ™π’’>πŸ“
𝑧=
𝑝̂ − 𝑝
𝑝 βˆ™π‘ž
𝑛
√
where
𝑝̂ = sample proportion
n = sample size
p = population proportion
q=1–p
Chapter 8 Statistical Inference: Estimation for
Single Populations
100(1 – a)% Confidence Interval to Estimate 𝛍: 𝝈
Known (8.1)
π‘₯Μ… − 𝑧∝⁄2
𝜎
√𝑛
≤ πœ‡ ≤ π‘₯Μ… + 𝑧∝⁄2
√𝑛
𝑛=
the area under the normal curve outside the
confidence interval area
 = the area in one end (tail) of the distribution
outside the confidence intervalο€ 
Confidence Interval to Estimate 𝛍 Using the Finite
Correction Factor (8.2)
𝜎 𝑁−𝑛
𝜎 𝑁−𝑛
√
√
≤ πœ‡ ≤ π‘₯Μ… + 𝑧∝⁄2
√𝑛 𝑁 − 1
√𝑛 𝑁 − 1
Confidence Interval to Estimate 𝛍: Population
Standard Deviation Unknown and the Population
Normally Distributed (8.3)
π‘₯Μ… − 𝑑∝⁄2,𝑛−1
𝑠
√𝑛
≤ πœ‡ ≤ π‘₯Μ… + 𝑑∝⁄2,𝑛−1
𝑠
√𝑛
df = 𝑛 − 1
Confidence Interval to Estimate p (8.4)
𝑝̂ . π‘žΜ‚
𝑝̂ . π‘žΜ‚
𝑝̂ − 𝑧∝⁄2 √
≤ 𝑝 ≤ 𝑝̂ + 𝑧∝⁄2 √
𝑛
𝑛
where
𝑝̂ = sample proportion
π‘žΜ‚ = sample size
p = population proportion
n = sample size
𝝌𝟐 Formula for Single Variance (8.5)
(𝑛 − 1)𝑠 2
πœ’2 =
𝜎2
𝑑𝑓 = 𝑛 − 1
Confidence Interval to Estimate the Population
Variance (8.6)
(𝑛 − 1)𝑠 2
(𝑛 − 1)𝑠 2
2
≤
𝜎
≤
2
πœ’π›Ό2⁄2
πœ’1−𝛼
⁄2
𝑛=
𝑧∝2⁄2 𝜎 2
𝑧∝⁄2 𝜎 2
=(
)
2
𝐸
𝐸
Sample Size When Estimating p (8.8)
𝜎
where
=
π‘₯Μ… − 𝑧∝⁄2
Sample Size When Estimating 𝝁 (8.7)
where
𝑧∝2⁄2 𝑝. π‘ž
𝐸2
p = population proportion
q= 1 – p
E = error of estimation
n = sample size
Chapter 9 Statistical Inference: Hypothesis
Testing for Single Populations
z Test for a Single Mean (9.1)
π‘₯Μ… − πœ‡
𝜎
√𝑛
𝑧=
Formula to Test Hypotheses about 𝝁 with a Finite
Population (9.2)
𝑧=
π‘₯Μ… − πœ‡
𝜎 √𝑁 − 𝑛
√𝑛 𝑁 − 1
t Test for 𝝁 (9.3)
𝑑=
π‘₯Μ… − πœ‡
𝑠
√𝑛
𝑑𝑓 = 𝑛 − 1
z Test of a Population Proportion (9.4)
𝑝̂ − 𝑝
𝑧=
𝑝 βˆ™π‘ž
𝑛
√
where
𝑝̂ = sample proportion
p = population proportion
q=1–p
Formula for Testing Hypotheses about a Population
Variance (9.5)
πœ’2 =
(𝑛 − 1)𝑠 2
𝜎2
𝑑𝑓 = 𝑛 − 1
Chapter 10 Statistical Inference: About Two
Populations
z Formula for the Difference in Two Sample Means
(Independent Samples and Population Variances
Known) (10.1)
𝑧=
(π‘₯Μ…1 − π‘₯Μ…2 ) − (πœ‡1 − πœ‡2 )
𝜎2
𝜎2
√ 1 + 2
𝑛1
𝑛2
where
πœ‡1 = mean of population 1
πœ‡2 = mean of population 2
𝑛1 = size of sample 1
𝑛2 = size of sample 2
df = n1 +n1 − 2
t Formula to Test the Difference in two Dependent
Populations (10.6)
t=
Μ… −D
d
sd
√n
df = n − 1
Confidence Interval to Estimate 𝝁𝟏 − 𝝁𝟐 (10.2)
σ2 σ2
σ2 σ2
(xΜ…1 − xΜ…2 ) − zα⁄2 √ 1 + 2 ≤ μ1 − μ2 ≤ (xΜ…1 − xΜ…2 ) + zα⁄2 √ 1 + 2
n1 n2
n1 n2
t Formula to Test the Difference in Means Assuming
𝝈𝟐𝟏 and 𝝈𝟐𝟐 are Equal (10.3)
t=
s 2 (n1 − 1) + s22 (n2 − 1) 1
1
(xΜ…1 − xΜ…2 ) + t√ 1
√ +
n1 + n2 − 2
n1 n2
(xΜ…1 − xΜ…2 ) − (μ1 − μ2 )
1
1
sp √n + n
1
2
df = n1 +n1 − 2
where
n = number of pairs
d = sample difference in pairs
D = mean population difference
sd = standard deviation of sample difference
d = mean sample difference
Formulas for d and sd (10.7 and 10.8)
Μ…=
d
∑d
n
(∑ d)2
Μ… )2 √∑ d2 −
∑(d
−
d
n
sd = √
=
n−1
n−1
Confidence Interval Formula to Estimate the
Difference in Related Populations, D (10.9)
where
s12 (n1 − 1) + s22 (n2 − 1)
sp = √
n1 + n2 − 2
Μ… −t
d
t Formula to Test the Difference in Means (10.4)
df = n − 1
t=
(xΜ…1 − xΜ…2 ) − (μ1 − μ2 )
s2
s2
√ 1 + 2
n1
n2
2
df =
s2
s2
(n1 + n2 )
1
2
2
s12
(n )
1
n1 − 1 +
2
s22
(n )
2
n2 − 1
Confidence Interval to Estimate μ1 − μ2 Assuming
the Population Variances are Unknown and Equal
(10.5)
s 2 (n1 − 1) + s22 (n2 − 1) 1
1
(xΜ…1 − xΜ…2 ) − t√ 1
√ + ≤ μ1 − μ2 ≤
n1 + n2 − 2
n1 n2
sd
√n
Μ… +t
≤D≤ d
sd
√n
z Formula for the Difference in Two Population
Proportions (10.10)
z=
(pΜ‚1 − pΜ‚2 ) − (p1 − p2 )
p .q
p .q
√ 1n 1 + 2n 2
1
2
where
pΜ‚1 = proportion from sample 1
𝑝̂2 = proportion from sample 2
𝑛1 = size of sample 1
𝑛2 = size of sample 2
𝑝1 = proportion from population 1
𝑝2 = proportion from population 2
π‘ž1 = 1 − 𝑝1
π‘ž2 =1 − 𝑝2
z Formula to Test the Difference in Population
Proportions (10.11)
z=
(pΜ‚1 − pΜ‚2 ) − (p1 − p2 )
1
1
√(p
Μ… . qΜ…) (n + n )
1
2
where
p
Μ…=
x1 + x2 (n1 pΜ‚1 + n2 pΜ‚2 )
=
n1 + n2
n1 + n2
and
qΜ… = 1 − p
Μ…
Confidence Interval to Estimate p1 - p2 (10.12)
pΜ‚1 . qΜ‚1 pΜ‚2 . qΜ‚2
(pΜ‚1 − pΜ‚2 ) − z√
+
≤ p1 − p2 ≤
n1
n2
pΜ‚1 . qΜ‚1 pΜ‚2 . qΜ‚2
(pΜ‚1 − pΜ‚2 ) + z√
+
n1
n2
F Test for Two Population Variances (10.13)
F=
s12
s22
dfnumerator = v1 = n1 − 1
dfdenominator = v2 = n2 − 1
Formula for Determining the Critical Value for the
Lower-Tail F (10.14)
F1−∝,v1 ,v2 =
1
F∝,v1 ,v2
Chapter 11 Analysis of Variance and Design of
Experiments
𝐢
𝑛
SSE = ∑ ∑(π‘₯𝑖𝑗 − π‘₯̅𝑗 − π‘₯̅𝑖
𝑖=1 𝑗=1
− π‘₯Μ… )
Formulas for computing a one-way ANOVA
𝑐
2
SST = ∑ ∑(π‘₯𝑖𝑗 − π‘₯Μ… )
SSC = ∑ 𝑛𝑗 (π‘₯̅𝑗 − π‘₯Μ… )
SSE
𝑁−𝑛−𝐢+1
Fπ‘π‘™π‘œπ‘π‘˜π‘  =
2
𝑖=1 𝑗=1
𝑛𝑗 𝑐
𝑅
𝑖=1 𝑗=1
df𝐢 = 𝐢 − 1
𝑖=1
𝐢
SSC = 𝑛𝑅 ∑(π‘₯̅𝑗 − π‘₯Μ… )
df𝐸 = 𝑁 − 𝐢
df𝐼 = (𝑅 − 1)(𝐢 − 1)
2
df𝐸 = 𝑅𝐢(𝑛 − 1)
𝑗=1
df 𝑇 = 𝑁 − 1
𝐢
𝑅
SSI = 𝑛 ∑ ∑(π‘₯̅𝑖𝑗 − π‘₯̅𝑖 − π‘₯̅𝑗 − π‘₯Μ… )
SSC
df𝐢
𝑖=1 𝑗=1
𝑅
𝐢
SSE
MSE =
df𝐸
𝑖=1 𝑗=1 π‘˜=1
𝑅
𝐢
𝑛
SST = ∑ ∑ ∑(π‘₯π‘–π‘—π‘˜ − π‘₯Μ… )
MSC
F=
MSE
2
df 𝑇 = 𝑁 − 1
MSR =
SSR
𝑅−1
MSC =
SSC
𝐢−1
𝑛
SSE = ∑ ∑ ∑(π‘₯π‘–π‘—π‘˜ − π‘₯Μ… 𝑖𝑗 )
2
2
MSI =
SSI
(𝑅 − 1)(𝐢 − 1)
𝑖=1 𝑗=1 π‘˜=1
MSE =
Tukey’s HSD test
MSE
HSD = π‘ž∝,𝐢,𝑁−𝐢 √
𝑛
Tukey-Kramer formula
MSE 1
1
q∝,𝐢,𝑁−𝐢 √
( + )
2 π‘›π‘Ÿ 𝑛 𝑠
Formulas for computing a randomized block design
2
𝑗=1
𝑛
SSR = 𝐢 ∑(π‘₯̅𝑖 − π‘₯Μ… )2
MSE
df𝑅 = 𝑅 − 1
SSR = 𝑛𝐢 ∑(π‘₯̅𝑖 − π‘₯Μ… )2
df𝐢 = 𝐢 − 1
𝐢
MSR
Formulas for computing a two-way ANOVA
2
SST = ∑ ∑(π‘₯𝑖𝑗 − π‘₯Μ… )
SSC = 𝑛 ∑(π‘₯̅𝑗 − π‘₯Μ… )
MSC
MSE
Fπ‘‘π‘Ÿπ‘’π‘Žπ‘‘π‘šπ‘’π‘›π‘‘π‘  =
SSE = ∑ ∑(π‘₯𝑖𝑗 − π‘₯̅𝑗 )
MSC =
MSE =
2
𝑖=1 𝑗=1
𝑗=1
𝑛𝑗 𝑐
SSR
𝑛−1
2
𝐢
𝑛
MSR =
df𝐢 = 𝐢 − 1
df𝑅 = 𝑅 − 1
df𝐸 = (𝐢 − 1)(𝑛 − 1)
= 𝑁−𝑛−𝐢+1
𝑖=1
MSC =
SSC
𝐢−1
SSE
(𝑛
𝑅𝐢 − 1)
F𝑅 =
MSR
MSE
F𝐢 =
MSC
MSE
F𝐼 =
MSI
MSE
Chapter 12 Correlation and Simple Regression
Analysis
(12.1) Pearson product-moment correlation
coefficient
r=
∑(x − xΜ…)(y − yΜ…)
√∑(x − xΜ…)2 ∑(y − yΜ…)2
∑x∑y
∑ xy −
n
=
2
2
(∑
x)
2 − (∑ y) ]
√[∑ x 2 −
y
]
[∑
n
n
Equation of the simple regression line
yΜ‚ = β0 + β1 x
Standard error of the estimate
se = √
SSE
n−2
(12.5) Coefficient of determination
r2 = 1 −
SSE
=1−
SSyy
SSE
(∑ y)2
∑ y2 −
n
Computational formula for r2
r2 =
b12 SSxx
SSyy
Sum of squares
SSxx = ∑ x 2 −
SSyy = ∑ y 2 −
SSxy = ∑ xy −
(∑ x)2
n
(∑ y)2
n
(∑ x)(∑ y)
n
(12.2) Slope of the regression line
b1 =
=
∑(x − xΜ…)(y − yΜ…) ∑ xy − nxΜ… yΜ…
=
∑(x − xΜ…)2
∑ x 2 − nxΜ… 2
(∑ x)(∑ y)
n
(∑ x)2
2
∑x −
n
t=
b1 − β1
sb
sb =
se
√SSxx
(12.6) Confidence interval to estimate E(yx) for a
given value of x
1 (xo − xΜ…)2
yΜ‚ ± t ∝⁄2,n−2 se √ +
n
SSxx
∑ xy −
(12.3) Alternative formula for slope
b1 =
t test of slope
SSxy
SSxx
(12.4) y intercept of the regression line
b0 = yΜ… − b1 xΜ… =
∑y
(∑ x)
− b1
n
n
Sum of squares of error
SSE = ∑(y − yΜ‚)2
= ∑ y 2 − b0 ∑ y − b1 ∑ xy
(12.7) Prediction interval to estimate y for a given
value of x
yΜ‚ ± t ∝⁄2,n−2 se √1 +
1 (xo − xΜ…)2
+
n
SSxx
Chapter 13 Multiple Regression Analysis
The F value
F=
MSreg SSreg⁄dfreg
SSR⁄k
=
=
MSerr SSerr⁄dferr SSE⁄(N − k − 1)
Sum of squares of error
SSE = ∑(y − yΜ‚)2
Standard error of the estimate
se = √
SSE
n−k−1
Coefficient of multiple determination
R2 −
SSR
SSE
=1−
SSyy
SSyy
Adjusted R2
Adjusted R2 = 1 −
SSE⁄(n − k − 1)
SSyy ⁄(n − 1)
Chapter 14 Building Multiple Regression
Models
(14.1) Variance inflation factor
VIF =
1
1 − R2i
Chapter 15 Time-Series forecasting and Index
Numbers
Individual forecast error
et = X t − Ft
Mean absolute deviation
MAD =
∑ |𝑒𝑖 |
Number of Forecasts
Mean square error
∑ e2i
MSE =
Number of Forecasts
Exponential smoothing
Ft+1 =∝. X t − (1−∝). Ft
Durbin-Watson test
D=
∑nt=2(et − et−1 )2
∑nt=1 e2t
Chapter 16 Analysis of Categorical Data
χ2 goodness-of-fit test (16.1)
(π‘“π‘œ − 𝑓𝑒 )2
𝑓𝑒
df = k − 1 − c
πœ’2 = ∑
χ2 test of independence (16.2)
(π‘“π‘œ − 𝑓𝑒 )2
𝑓𝑒
df = (r − 1)(c − 1)
πœ’2 = ∑ ∑
Chapter 17 Nonparametric Statistics
Large-sample runs test
πœ‡π‘… =
𝑐
𝑇𝑗2
12
𝐾=
(∑ ) − 3(𝑛 + 1)
𝑛(𝑛 + 1)
𝑛𝑗
2𝑛1 𝑛2
+1
𝑛1 +𝑛2
𝑗=1
2𝑛1 𝑛2 (2𝑛1 𝑛2 − 𝑛1 − 𝑛2 )
πœŽπ‘… = √
(𝑛1 +𝑛2 )2 (𝑛1 +𝑛2 − 1)
𝑅 − πœ‡π‘…
𝑧=
=
πœŽπ‘…
Kruskal-Wallis test (17.3)
2𝑛 𝑛
𝑅 − ( 1 2 + 1)
𝑛1 +𝑛2
2𝑛 𝑛 (2𝑛1 𝑛2 − 𝑛1 − 𝑛2 )
√ 1 2
(𝑛1 +𝑛2 )2 (𝑛1 +𝑛2 − 1)
Friedman test (17.4)
𝑐
πœ’π‘Ÿ2
12
=
∑ 𝑅𝑗2 − 3𝑏(𝑐 + 1)
𝑏𝑐(𝑐 + 1)
𝑗=1
Spearman’s rank correlation (17.5)
π‘Ÿπ‘  = 1 −
Mann-Whitney U test (small sample)
π‘ˆ1 = 𝑛1 𝑛2 +
𝑛1 (𝑛1 + 1)
− π‘Š1
2
π‘ˆ2 = 𝑛1 𝑛2 +
𝑛2 (𝑛2 + 1)
− π‘Š2
2
π‘ˆ ′ = 𝑛1 𝑛2 − π‘ˆ
Mann-Whitney U test (large sample) (17.1)
πœ‡π‘ˆ =
𝑛1 𝑛2
2
πœŽπ‘ˆ = √
𝑧=
𝑛1 𝑛2 (𝑛1 + 𝑛2 + 1)
12
π‘ˆ − πœ‡π‘ˆ
πœŽπ‘ˆ
Wilcoxon matched-pairs signed rank test (17.2)
πœ‡π‘‡ =
(𝑛)(𝑛 + 1)
4
(𝑛)(𝑛 + 1)(2𝑛 + 1)
πœŽπ‘‡ = √
24
𝑧=
𝑇 − πœ‡π‘‡
πœŽπ‘‡
6 ∑ 𝑑2
𝑛(𝑛2 − 1)
Chapter 18 Statistical Quality Control
Μ… π‚π‘πšπ«π­π¬
𝒙
Centreline:
∑ xΜ…
k
xΜ… =
UCL:
Μ…
xΜ… + A2 R
LCL:
Μ…
xΜ… − A2 R
OR
UCL:
xΜ… + A3 sΜ…
LCL:
xΜ… − A3 sΜ…
R charts
Centreline:
∑R
k
Μ…=
R
UCL:
Μ…
D4 R
LCL:
Μ…
D3 R
p charts
Centreline:
UCL:
LCL:
p=
∑ pΜ‚
k
p + 3√
p. q
n
p − 3√
p. q
n
c charts
Centreline:
cΜ… =
c1 + c2 + c3 … + ci
i
UCL:
cΜ… + 3√cΜ…
LCL:
cΜ… − 3√cΜ…
Chapter 19 Decision Analysis
Bayes’ Rule
𝑃(𝑋𝑖 | π‘Œ) =
𝑃(𝑋𝑖 ) βˆ™ 𝑃(π‘Œ|𝑋𝑖 )
𝑃(𝑋1 ) βˆ™ 𝑃(π‘Œ | 𝑋1 ) + 𝑃(𝑋2 ) βˆ™ 𝑃(π‘Œ | 𝑋2 ) + β‹― + 𝑃(𝑋𝑛 ) βˆ™ 𝑃(π‘Œ | 𝑋𝑛 )
Download