Lecture 4

advertisement
STATISTICAL METHODS FOR ASSESSING ERROR
Errors
Small random,
Small systematic
Large random,
Small systematic
Small random,
Large systematic
Large random,
Large systematic
The Mobster wants to buy a
new gun.
Errors. Assume no systematic.
Small random,
Large random,
x
x
You see one bullet hole (marked with an X), but you
can’t see the target.
If there is no systematic error, how well can you tell
where the target is?)
x
Think about the target in 1-d – bullets miss to the right or left.
Make a bunch of measurements of x, each with some random error
target
x
x2 x4
x3
bullet
holes
x5
x1
x
1 N
x   xi
N i1
x  xavg  xmean  x
If
there is only random error, no systematic
error, then in the limit of infinite number of measurements, the average IS
the “target”. Even if you only take five “shots”, the average is usually
closer to the target than most of the typical individual shots.
Make 6 measuremen ts (N  6). xi  2 ,-2, - 1, 2, 5, 0
What is x? (What is average? or, What is mean?)
1 N
1
1
x   xi  (2  2  1  2  5  0)  6  1
N i 1
6
6
x 1
How much does point #3 differ from the average?
x3  x   3
x
3
x2 x4
x3
x5
x1
x
How much, on average, does each point differ from the average?
N
Always 0! Not so enlightening. Could use
average
i
absolute value, but instead, use...
i 1


1
N
 (x  x)  0
1 N
   ( xi  x )2
N i1
1/ 2
1 N
2 1/ 2
2
x  
   ( xi  x ) 
 N i 1

2

2
1 N
  ( xi  x )2
N i1
1/ 2
x  
2 1/ 2
1 N

   ( xi  x ) 2 
 N i 1

x is sometimes defined with a N-1 instead of an N. This is a little
better, I think.
1/ 2
x   2
1/ 2
 1 N

 
( xi  x ) 2 

 N  1 i 1

x “sigma” is the “standard deviation”. Roughly speaking
it tells you by about how much a typical measurement will
deviate from the mean value. Really it is the square root of
the average of the square difference from the mean. 2 is
called the “variance”.
x is a good value to use for x.
it is sometimes called “the standard error.”
In fact,
By how much does a typical point differ from the average?
(root-mean-square difference)
x
+/-x
x2 x4
x3
x5
x1
x
If we measure just one point, (call it xi), the true value of x
is “probably” within about 1  of xi. Oftentimes by “x” we
mean, “one standard deviation”, or “the standard error”, x.
1/ 2
x  
2 1/ 2
 1 N
2

 
(
x

x
)

i
N

1
i 1


Make 6 measuremen ts (N  6). xi  2,-2,-1,2,5,0
x 1
What is standard deviation,  x ?
x 
1 N
( xi  x ) 2 

N  1 i 1


1
(2  x ) 2  (2  x ) 2  (1  x ) 2  (2  x ) 2  (5  x ) 2  (0  x ) 2
N 1

1
(1) 2  (3) 2  (2) 2  (1) 2  (4) 2  (1) 2
N 1

1
1  9  4  1  16  1 
N 1

 6.4  2.5
1
32
N 1


Make 6 measuremen ts (N  6). xi  2,-2,-1,2,5,0
x  1,  x  2.5
If we pick any one point, for instance, x3, which was “-1”, and ask, how far
away is it likely to be from true answer, answer is “probably less than x.”
So if we make just the one measurement, we say x=x, or in this case, 2.5.
But we think, if we take the average of lots of points, that should have a
smaller error. How much smaller? Let’s use the master rule. We can think
of the average of six points as being a function of six points
x  f ( x1, x2 , x3 , x4 , x5 , x6 )
1
f ( x1, x2 , x3 , x4 , x5 , x6 )  ( x1  x2  x3  x4  x5  x6 )
N
x  f ( x1, x2 , x3 , x4 , x5 , x6 )
f ( x1 , x2 , x3 , x4 , x5 , x6 ) 
What is f?
 f

2
 f

2
1
( x1  x2  x3  x4  x5  x6 )
N
 f

2
f  
x1   
x2   
x3   ....
 x1

 x2

 x3

2
2
2
f 
1

1

1

 x1    x2    x3   .... x i are uncorrel., all equal to  x
N

N

N

f 
1

1

1

  x     x     x   .... 
N

N

N

2
f  x 
x
N
2
2
we could write this  x 
x
N
2
1

N   x  .... 
N

or x 
x
N
 x2
N
The “standard error on mean” is smaller than standard error on a single measurement, by
a factor of (Number of points averaged together)1/2 .
The standard error on a single point is the standard deviation on a single measurement
x.
The standard error on the mean (if you are reporting an average value for your final result,
then the error on that average value) is x/(N)1/2
Say we make a measurement, and the true value of the thing we are measuring is “5”.
But there is some purely random noise, so that on average, half the time we measure “0”
and half the time we measure “10”. (weirdly, we never measure exactly “5”. But 5 is
what we get on average.) So <x> is 5, and the error on a single measurement is x = 5.
What if we average ten points? What will be the standard error on the average?
Histogram.
x = 5 +/- 5. x = 5.
Suppose we make ten measurements. We think mean will be “5”, but if we make just ten
measurements, can’t be sure mean will come out just right. What is standard error on the
mean? Let’s do a mass experiment
Heads = “I measure 10”. Tails = “I measure 0”.
Flip the coin ten times. Write down the answers (eg
“THHHTTHTTT”.) Calculate average. For my example, the answer is “4”
(hint: to do this fast, just add up the numbers of heads, multiply by ten, then divide by
ten. Or, you could just add up the number of heads, and stop right there!)
Clicker questions 5.2a and 5.2b. (Don’t answer both questions!)
What is the average of your ten measurements
5.2a
5.2b
a 0, or 10
b1
c2
d3
e4
a. 5
b. 6
c. 7
d. 8
e. 9
# of times
0
1
2
3
4
5
6
7
8
9 10
Value
1
# of times
0
1
2
3
4
5
6
7
Value
“Normal”, or “Gaussian” distribution.
8
9 10
If data points are “normally
distributed”, one expects:
+/-3
expected
# of times
(probability)
99.7%, about 299 out of
300, should be within +/-3 
+/-2
95%, about 19 out of 20,
should be within +/-2 
+/-x
+/- 1
68%, about 2 out of 3,
should be within +/-1 
<x>
0
1
2
3
4
5
6
7
8
9 10
Value
The “Gaussian”, or “normal” distribution
When do we kill the courier?
Shipped quantity = S +/- S
Received quantity = R +/- R
Discrepancy = R-S.
Estimated error on the difference (R-S) = ((R)2 + (S)2)1/2
If we are working with “one sigma” errors, (x = 1* x), and assuming our
courier is “honest”, and our measurements, and our estimates of uncertainty
are all accurate (ie, no “systematic error” in measurement, or in the courier.)
One chance in 3 that the discrepancy will be more than (R-S) (a
“one-sigma” discrepancy)
But only one chance in 20 that the discrepancy will be more than 2*
(R-S) (a “two-sigma” discrepancy)
Only one chance in 300 that the discrepancy will be more than 3*
(R-S) (a “three-sigma” discrepancy)
Download