definitions and formulae with statistical tables for elementary

advertisement
DEFINITIONS AND FORMULAE WITH
STATISTICAL TABLES
FOR ELEMENTARY STATISTICS AND
QUANTITATIVE METHODS COURSES
Department of Statistics
University of Oxford
October 2012
Contents
1 Laws of Probability
1
2 Theoretical mean and variance for discrete distributions
1
3 Mean and variance for sums of Normal random variables
1
4 Estimates from samples
1
5 Two common discrete distributions
1
6 Standard errors
2
7 95% confidence limits for population parameters
2
8 z-tests
3
9 t-tests
3
10 The χ2 -test
3
11 Correlation and regression
4
12 Analysis of variance
4
13 Median test for two independent samples
5
14 Rank sum test or Mann-Whitney test
5
15 Sign test for matched pairs
5
16 Wilcoxon test for matched pairs
5
17 Kolmogorov-Smirnov test
5
18 Kruskal-Wallis test for several independent samples
6
19 Spearman’s Rank Correlation Coefficient
6
20 TABLE 1 : The Normal Integral
7
21 TABLE 2 : Table of t
TABLE 3 : Table of χ2
8
22 TABLE 4 : Table of F for P = 0.05
9
23 TABLE 5 : Critical values of R for the Mann-Whitney rank-sum test
9
24 TABLE 6 : Critical values for T in the Wilcoxon Matched-Pairs Signed-Rank test . 10
1
Laws of Probability
Multiplication law
Addition law
Bayes’ Rule
B1
A1
B2
B1
A2
B2
2
P (A and B) = P (A)P (B|A) = P (B)P (A|B)
P (A or B) = P (A) + P (B) − P (A and B)
P (B1 ) = P (B1 |A1 )P (A1 ) + P (B1 |A2 )P (A2 )
Theoretical mean and variance for discrete distributions
σ 2 = Σ(x − µ)2 p(x)
µ = Σxp(x)
3
P (B1 |A1 )P (A1 )
P (B1 )
P (A1 |B1 ) =
Mean and variance for sums of Normal random variables
P
If Xi ∼ N(µi , σi2 ), i = 1, . . . , n and let Y = ni=1 ai Xi then
µY =
n
X
σY2 =
ai µ i
i=1
4
n
X
a2i σi2
i=1
Estimates from samples
Ungrouped data:
Σx
sample mean x̄ = n i
sample variance s2 =
5
Σx2i −nx̄2
n−1
2
for x in a sample of size n ,
x
sample proportion p̂ = n
Two common discrete distributions
(i) Binomial
p(x) =
n
(ii) Poisson
Cx px q n−x , x = 0, 1, . . . , n
µ = np, σ 2 = npq, σ =
n
=
Σf (x−x̄)2
s = Σf −1
Σf x
Grouped data: x̄ = Σf
Counted events:
Σ(xi −x̄)2
n−1
√
p(x) =
npq
e−λ λx
x! ,
µ = λ,
n!
Cx = x!(n−x)!
1
x = 0, 1, . . . , ∞
σ 2 = λ,
σ=
√
λ
6
Standard errors
Single sample of size n
SE(x̄) = √σn or, if σ unknown, √sn
q
p pq
SE(p̂) =
with q = 1 − p ,
n
or, if p unknown,
p̂(1−p̂)
n
Sampling without replacement
When n individuals are sampled from a population of N without replacement, the standard
error is reduced. The standard error for no replacement SENR is related to the standard error
with replacement SEWR by the formula
s
s
σ
n−1
n−1
=√
,
1−
1−
SENR = SEWR
N −1
N −1
n
where σ is the known standard deviation of the whole population.
Two independent samples of sizes, n1 and n2
q
SE(x̄1 − x̄2 ) =
σ12
n1
+
σ22
n2
q
or, if σ1 and σ2 unknown and different,
For common but unknown σ ,
q
SE(p̂1 − p̂2 ) =
p1 q1
n1
+
p2 q2
n2 ,
SE(x̄1 − x̄2 ) = s
q
p̂1 q̂1
n1
7
+
1
n2
with s2 =
s22
n2
.
(n1 −1)s21 +(n2 −1)s22
n1 +n2 −2
p̂2 q̂2
n2
+
For common but unknown p , SE(p̂1 − p̂2 ) =
n1 p̂1 +n2 p̂2
n1 +n2
1
n1
+
or,
if p1 and p2 unknown and unequal,
p defined as p̂ =
q
s21
n1
q
p̂q̂( n11 +
1
n2 )
where p̂ is a pooled estimate of
and q̂ = 1 − p̂.
95% confidence limits for population parameters
Mean:
when σ known use
when σ unknown use
x̄ ± 1.96 √σn
x̄± t √sn
where t is the tabulated two-sided 5% level value with degrees of freedom, d.f. = n − 1
p
Proportion:
p̂ ± 1.96 p̂q̂/n
2
8
z-tests
x̄−µ
z = σ/√n
Single sample test for population mean µ (known σ):
p̂−p
z = √ pq
Single sample test for population proportion p:
n
z=
Two sample test for difference between two means (known σ1 and σ2 ):
Two sample test for difference between two proportions :
z =q
where p̂ is a pooled estimate of p defined as p̂ =
9
2
2
σ1
σ2
n1 + n2
pˆ1 −pˆ2
p̂q̂( n1 + n1 )
1
n1 p̂1 +n2 p̂2
n1 +n2
rx̄1 −x̄2
2
and q̂ = 1 − p̂
t-tests
Population variance σ 2 unknown and estimated by s2
x̄−µ
Single sample test for population mean µ
t = √ with d.f.= n − 1
s/ n
Paired samples : test for zero mean difference, using n pairs (x, y), d = x − y
d¯√
t=
with d.f.= n − 1, where d¯ and sd are the mean and standard deviation of d.
sd / n
Independent samples test for difference between population means µx and µy using nx x’s
and ny y’s. Provided that s2x and s2y are similar values, use the pooled variance estimate,
(nx −1)s2x +(ny −1)s2y
x̄−ȳ
2
s =
with d.f = nx + ny − 2
,
and t = q 1
nx +ny −2
s
+1
nx
10
ny
The χ2 -test
(Note that the two tests in this section are nonparametric tests. There are χ2 tests of variances, not included here, that are parametric.)
χ2 Goodness-of-fit tests using k groups have
d.f.=(k − 1) − p where p is the number of independent parameters estimated
and used to obtain the (fitted) expected values.
χ2 Contingency table tests on two-way tables with r rows and c columns have
d.f.= (r − 1)(c − 1)
For both tests, χ2 = Σ
expected frequency
(O−E)2
E
where O is an observed frequency and E is the corresponding
3
11
Correlation and regression
For n pairs
(xi , yi ), with sample variances s2x and s2y as in section 3, define sample covariance,
P
P
−x̄)(yi −ȳ)
1
n
sxy = (xi(n−1)
. May be computed as sxy = n−1
xi yi − n−1
x̄ȳ.
2
2
Computed from the sample variances sx+y of x + y and sx−y of x − y as sxy = 14 (s2x+y − s2x−y ).
Sample product-moment correlation coefficient,
sxy
r= s s
x y
Test the significance of
√ the correlation coefficient, ρ = 0, or equivalently, of the regression
r n−2
slope β = 0:
t =√
with d.f.= n − 2
1−r2
Linear Regression of y on x. Equation y = α + βx with α and β estimated by a and b.
sxy
a = ȳ − bx̄.
Estimates:
b = s2
x
√
Root-mean-square error of regression prediction given by sy 1 − r2 .
√
b √1−r2
b
Test significance of regression as above or t= SE(b) with d.f.= n − 2 where SE(b) =
r n−2
b ± t.SE(b), where t is the tabulated two-sided
5% level value with d.f.= n − 2
95% confidence limits for the slope β are:
12
Analysis of variance
Single factor or One-way analysis for a completely randomized design.
The test statistic F is calculated as a ratio of two mean squares. If the numbers in the k groups
are n1 , n2 , . . . , nk then the total sample size is Σni = n. Calculate the “total sum of squares”,
T SS = (n − 1)s2T , where s2T is variance of all n observations. Calculate the sample means and
sample variances of the k groups by x̄1 , x̄2 , . . . , x̄k and s21 , s22 , . . . , s2k and then the “within groups
sum of squares” (also known as the“error sum of squares”), ESS = Σ(ni P
−1)s2i . The “between
groups sum of squares” may be computed in two different ways: BSS = ni (x̄i − x̄T )2 , where
x̄T is the mean of all n observations; or BSS = T SS − ESS.
These together with their degrees of freedom are entered into the ANOVA table:
Source of variation
Degrees
of freedom
Sum of
squares
Between samples
k−1
BSS
BM S =
BSS
k−1
Within samples
n−k
ESS
EM S =
ESS
n−k
Total
n−1
T SS
4
Mean
square
F
BM S
EM S
13
Median test for two independent samples
For two independent samples, sizes n1 and n2 , the median of the whole sample of n = n1 + n2
observations is found. The number in each sample above this median is counted and expressed
as a proportion of that sample size. The two proportions are compared using the Z-test as in §8.
14
Rank sum test or Mann-Whitney test
For two independent samples, sizes n1 and n2 , ranked without regard to sample, call the sum
of the ranks in the smaller sample R. If n1 ≤ n2 ≤ 10 referq
to Table 5, otherwise use a Z
n1 n2 (n1 +n2 +1)
test with z = (R − µ)/σ where µ = 21 n1 (n1 + n2 + 1) and σ =
, assuming
12
n1 ≤ n2 . In case of ties, ranks are averaged.
15
Sign test for matched pairs
The number of positive differences from the n pairs is counted. This number is binomially
distributed with p = 21 , assuming a population zero median difference. So apply the Z test
for a binomial proportion with p = 21 .
16
Wilcoxon test for matched pairs
Ignoring zero differences, the differences between the values in each pair are ranked without
regard to sign and the sums of the positive ranks, R+ and of the negative ranks, R− , are
calculated. (Check R+ + R− = 12 n(n + 1), where n is the number of nonzero differences). The
smaller of R+ and R− is called T and may be compared with the critical values in Table 6 for
a two-tailed test. (For one-tailed tests, use R− and R+ with the same table, remembering to
halve P .) In case of ties, ranks are averaged.
17
Kolmogorov-Smirnov test
Two samples of sizes n1 and n2 are each ordered along a scale. At each point on the scale
the empirical cumulative distribution function is calculated for each sample and the difference
between the pairs are recorded as Di . The largest absolute value of the Di is called Dmax and
this value is compared with the 5% one-tailed value
r
n1 + n2
Dcrit = 1.36
.
n1 n2
Single sample version, compares sample with theoretical distribution,
r
1
Dcrit = 1.36
.
n
Should only be used with no ties, but it commonly is used otherwise. With ties, the value of
Dmax tends to be too small, so that the p-value is an overestimate.
5
18
Kruskal-Wallis test for several independent samples
(Analysis of variance for a single factor). For k samples of sizes n1 , n2 , ..nk , comprising a total
of n observations, all values are ranked without regard to sample, from 1 to n. The rank sums
for the samples are calculated as R1 , R2 , .., Rk . (Check ΣRi = 21 n(n + 1). The test statistic is
R2
12
H = n(n+1) Σ ni
i
h
i
−3(n + 1),
which is compared to χ2 table with d.f. = k − 1
19
Spearman’s Rank Correlation Coefficient
If x and y are ranked variables the Spearman Rank Correlation Coefficient is just the sample
product moment correlation coefficient between the pairs of ranks, rs , which may also be
computed by
rs = 1−
6Σd2
n(n2 −1)
where d is the difference
x − y, and n is the number of pairs (x, y).
√
r√
n−2
with d.f.= n − 2
Test rs using t = s
1−rs2
6
20
TABLE 1 : The Normal Integral
Tabled value is P
2P ! 1
P
z
0
!z
0
1!P
1!P
!z 0
z
0.0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.0
1.1
1.2
1.3
1.4
1.5
1.6
1.7
1.8
1.9
z
0
z
0.00
0.5000
0.5398
0.5793
0.6179
0.6554
0.6915
0.7257
0.7580
0.7881
0.8159
0.8413
0.8643
0.8849
0.9032
0.9192
0.9332
0.9452
0.9554
0.9641
0.9713
0.01
5040
5438
5832
6217
6591
6950
7291
7611
7910
8186
8438
8665
8869
9049
9207
9345
9463
9564
9649
9719
0.02
5080
5478
5871
6255
6628
6985
7324
7642
7939
8212
8461
8686
8888
9066
9222
9357
9474
9573
9656
9726
0.03
5120
5517
5910
6293
6664
7019
7357
7673
7967
8238
8485
8708
8907
9082
9236
9370
9484
9582
9664
9732
0.04
5160
5557
5948
6331
6700
7054
7389
7704
7995
8264
8508
8729
8925
9099
9251
9382
9495
9591
9671
9738
0.05
5199
5596
5987
6368
6736
7088
7422
7734
8023
8289
8531
8749
8944
9115
9265
9394
9505
9599
9678
9744
0.06
5239
5636
6026
6406
6772
7123
7454
7764
8051
8315
8554
8770
8962
9131
9279
9406
9515
9608
9686
9750
0.07
5279
5675
6064
6443
6808
7157
7486
7794
8078
8340
8577
8790
8980
9147
9292
9418
9525
9616
9693
9756
0.08
5319
5714
6103
6480
6844
7190
7517
7823
8106
8365
8599
8810
8997
9162
9306
9429
9535
9625
9699
9761
0.09
5359
5753
6141
6517
6879
7224
7549
7852
8133
8389
8621
8830
9015
9177
9319
9441
9545
9633
9706
9767
2.0 0.9772
3.0 0.9987
0.1
9821
9990
0.2
9861
9993
0.3
9893
9995
0.4
9918
9997
0.5
9938
9998
0.6
9953
9998
0.7
9965
9999
0.8
9974
9999
0.9
9981
9999
7
21
1!
2
TABLE 3 : Table of χ2
TABLE 2 : Table of t
P
P
P
1!
2
!t
0
t
X2
Probability P of a value of χ2
greater than:
Probability P of lying outside ±t
d.f.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
40
60
∞
P=0.10
6.31
2.92
2.35
2.13
2.02
1.94
1.90
1.86
1.83
1.81
1.80
1.78
1.77
1.76
1.75
1.75
1.74
1.73
1.73
1.73
1.72
1.72
1.71
1.71
1.71
1.71
1.70
1.70
1.70
1.70
1.68
1.67
1.65
P=0.05
12.71
4.30
3.18
2.78
2.57
2.45
2.37
2.31
2.26
2.23
2.20
2.18
2.16
2.15
2.13
2.12
2.11
2.10
2.09
2.09
2.08
2.07
2.07
2.06
2.06
2.06
2.05
2.05
2.05
2.04
2.02
2.00
1.96
P=0.02
31.82
6.96
4.54
3.75
3.36
3.14
3.00
2.90
2.82
2.76
2.72
2.68
2.65
2.62
2.60
2.58
2.57
2.55
2.54
2.53
2.52
2.51
2.50
2.49
2.49
2.48
2.47
2.47
2.46
2.46
2.42
2.39
2.33
P=0.01
63.7
9.93
5.84
4.60
4.03
3.71
3.50
3.36
3.25
3.17
3.11
3.06
3.01
2.98
2.95
2.92
2.90
2.88
2.86
2.85
2.83
2.82
2.81
2.80
2.79
2.78
2.77
2.76
2.76
2.75
2.70
2.66
2.58
8
d.f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
40
60
P=0.05
3.84
5.99
7.81
9.49
11.07
12.59
14.07
15.51
16.92
18.31
19.68
21.03
22.36
23.68
25.00
26.30
27.59
28.87
30.14
31.41
32.67
33.92
35.17
36.42
37.65
38.89
40.11
41.34
42.56
43.77
55.76
79.08
P=0.01
6.63
9.21
11.34
13.28
15.09
16.81
18.48
20.09
21.67
23.21
24.73
26.22
27.69
29.14
30.58
32.0
33.41
34.81
36.19
37.57
38.93
40.29
41.64
42.98
44.31
45.64
46.96
48.28
49.59
50.90
63.69
88.38
22
TABLE 4 : Table of F for P = 0.05
P = 0.05
F
Variance ratio F = s21 /s22 with ν1 and ν2 degrees of freedom respectively.
ν1
ν2
6
8
10
12
14
16
18
20
30
40
60
∞
23
1
5.99
5.32
4.96
4.75
4.60
4.49
4.41
4.35
4.17
4.08
4.00
3.84
2
3
5.14
4.46
4.10
3.89
3.74
3.63
3.55
3.49
3.32
3.23
3.15
3.00
4.76
4.07
3.71
3.49
3.34
3.24
3.16
3.10
2.92
2.84
2.76
2.60
4
4.53
3.84
3.48
3.26
3.11
3.01
2.93
2.87
2.69
2.61
2.53
2.37
5
6
4.39
3.69
3.33
3.11
2.96
2.85
2.77
2.71
2.53
2.45
2.37
2.21
4.28
3.58
3.22
3.00
2.85
2.74
2.66
2.60
2.42
2.34
2.25
2.10
8
4.15
3.44
3.07
2.85
2.70
2.59
2.51
2.45
2.27
2.18
2.10
1.94
12
4.00
3.28
2.91
2.69
2.53
2.42
2.34
2.28
2.09
2.00
1.92
1.75
∞
24
3.84
3.12
2.74
2.51
2.35
2.24
2.15
2.08
1.89
1.79
1.70
1.52
3.67
2.93
2.54
2.30
2.13
2.01
1.92
1.84
1.62
1.51
1.39
1.00
ν2
6
8
10
12
14
16
18
20
30
40
60
∞
TABLE 5 : Critical values of R for the Mann-Whitney rank-sum test
The pairs of values below are approximate critical values of R for two-tailed tests at levels
P = 0.10 (upper pair) and P = 0.05 (lower pair). (Use relevant P = 0.10 entry for one-tailed
test at level 0.05).
smaller sample
size n1
4
5
6
7
larger sample size, n2
4
5
6
7
12,24 13,27 14,30 15,33
11,25 12,28 12,32 13,35
19,36 20,40 22,43
18,37 19,41 20,45
28,50 30,54
26,52 28,56
39,66
37,68
8
9
10
9
8
16,36
14,38
23,47
21,49
32,58
29,61
41,71
39,73
52,84
49,87
9
17,39
15,41
25,50
22,53
33,63
31,65
43,76
41,78
54,90
51,93
66,105
63,108
10
18,42
16,44
26,54
24,56
35,67
33,69
46,80
43,83
57,95
54,98
69,111
66,114
83,127
79,131
24
TABLE 6 : Critical values for T in the Wilcoxon Matched-Pairs Signed-Rank test .
The values below are the approximate critical values of T for two-tailed tests at level P. For
a significant result, the calculated T must be less than or equal to the tabulated value.
(Values of P are halved for one-tailed tests using R− and R+ .)
n
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
P = 0.10
2
2
3
5
8
10
14
17
21
26
30
36
41
47
53
60
67
75
83
91
100
10
P = 0.05
0
2
3
5
8
10
13
17
21
25
29
34
40
46
52
58
65
73
81
89
Download