0 References ISBN 978-86-900284-2-9 10 11 12 13 14 15 16 17 18 19 20 21 [...] 9 9 Measurement uncertainty 8 8 Tests and functions fitting 7 7 Variances In front of you is the Companion with concisely described subjects that are indispensable in all measurements. The Companion consists of about minimal and sufficient set of concepts and methods required to compute a measurement result and its measurement uncertainty. The text is in the form that enables direct applications in computations, manual or spreadsheet, and in writing software. 6 6 Outlines of the statistics 5 5 Measurement standards and calibrations The essence of measurement is the same in all areas. The purpose of this publication is to be a companion to people dealing with measurements, that is experiments or observations, from production to research, from physics and chemistry to biology and medicine. The publication is also intended for students. 4 4 Measurement results, accuracy and errors Goran Kostić 3 3 Devices for measurement 2 2 Measurements From a shop floor to the fundamental sciences, statistical analysis of results of repeated measurements is required. The purpose of the analysis is to determine the value of a measured quantity, as well as the accuracy of the determined value. It is requested that accuracy is described by the standard measurement uncertainty to enable traceability and ensure computing of the confidence level and the confidence interval. 1 1 Quantities and units of measurement From the Preface Goran Kostić Chapters THE NEW SI E-BOOK ON CONCEPTS AND COMPUTATIONS IN MEASUREMENTS Goran Kostić PAGES FROM: Metrology Companion To buy this book, click www.symmetry.rs. Leskovac, 2019 CONTENTS CONTENTS 5 Abbreviations 12 Letter symbols 13 Greek alphabet 14 PREFACE 15 INTRODUCTION 19 PART I FUNDAMENTALS OF MEASUREMENTS 21 1 QUANTITIES AND UNITS OF MEASUREMENT 23 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 1.10 1.11 1.12 1.13 1.14 1.15 1.16 1.17 1.18 1.19 1.20 1.21 Measurable quantity 23 Quantities of the same kind 23 System of quantities 24 Base quantity 25 Derived quantity 26 Ordinal quantity 26 Dimension of a quantity 27 Unit of measurement 28 Base unit of measurement 29 Derived unit of measurement 29 Coherent unit of measurement 29 Value of a quantity 30 Numerical value of a quantity 31 Conventional value of a quantity 31 Reference value 31 Constant 32 Quantity equation 34 Numerical value equation 34 Unit equation 35 Conversion factor between units 35 System of units 36 1.21.1 SI base units 37 1.22 Off-system unit of measurement 41 Kostić • 2019 • Metrology Companion 5 1.23 Decimal units of measurement 41 1.24 Binary unit of measurement 42 1.25 Expressing the value of a quantity in the SI 43 2 MEASUREMENTS 51 2.1 2.2 2.3 2.4 2.5 2.6 2.7 2.8 2.9 2.10 2.11 2.12 2.13 2.14 Measurement 51 Metrology 51 Measurand 52 Principle of measurement 52 Method of measurement 52 Measurement procedure 53 Robustness of procedure 53 Standard operating procedure 53 Standard measurement procedure 54 Qualification and validation 55 Validation of measurement procedure 55 Accreditation and certificate 57 About design of experiment 58 Mathematical model 59 2.14.1 Examples 59 2.15 Simulation 60 3 DEVICES FOR MEASUREMENT 61 3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8 3.9 3.10 3.11 3.12 3.13 3.14 3.15 3.16 3.17 6 Device for measurement 61 Indicating measuring instrument 62 Material measure 62 Measuring chain 63 Measuring system 63 Nominal value 64 Indication of a device for measurement 64 Indication interval 65 Measuring interval 65 Response function 65 Measuring transducer 66 Sensor 66 Detector 66 Measurement signal 66 Resolution 67 Sensitivity 67 Response time 67 Kostić • 2019 • Metrology Companion 3.18 3.19 3.20 3.21 3.22 3.23 3.24 3.25 3.26 3.27 3.28 3.29 3.30 3.31 3.32 3.33 3.34 3.35 3.36 3.37 Limit of detection 68 Limit of quantification 68 Dynamic interval 68 Dead interval 69 Hysteresis 69 Influence quantity 69 Blank indication 69 Stability 70 Selectivity 70 Transparency 70 Operating conditions 71 Reference operating conditions 71 Limiting conditions 71 Gauging of a device for measurement 72 Error of indication of a device for measurement 72 Datum error of a device for measurement 72 Zero error of a device for measurement 72 Bias of a device for measurement 73 Intrinsic error of a device for measurement 73 Accuracy class 73 3.37.1 Examples 73 3.38 Type approval of a device for measurement 74 3.39 Verification of a device for measurement 74 4 MEASUREMENT RESULTS, ACCURACY AND ERRORS 75 4.1 4.2 4.3 4.4 4.5 4.6 4.7 4.8 4.9 4.10 4.11 4.12 4.13 4.14 4.15 4.16 Measurement result 75 Accuracy 75 Absolute error 76 Relative error 77 Random error 78 Systematic error 78 Combined error 79 4.7.1 Example 79 Limits of error 80 Outlier 80 Correction 81 Precision of measurement results 82 Repeatability of measurement results 82 Reproducibility of measurement results 82 Rounding of number 83 Truncation of digits 84 Significant digits 84 Kostić • 2019 • Metrology Companion 7 5 MEASUREMENT STANDARDS AND CALIBRATIONS 85 5.1 5.2 5.3 5.4 5.5 5.6 5.7 5.8 5.9 5.10 5.11 5.12 Measurement standard 85 Reference material 86 Primary measurement standard 88 Secondary measurement standard 88 International measurement standard 88 National measurement standard 89 Transfer measurement standard 89 Reference measurement standard 89 Working measurement standard 89 Relativistic effects on the values of standards 90 Calibration 91 Calibration report 92 PART II STATISTICAL ANALYSIS OF MEASUREMENT RESULTS 93 6 OUTLINES OF THE STATISTICS 95 6.1 6.2 6.3 6.4 6.5 6.6 6.7 6.8 6.9 6.10 6.11 6.12 6.13 6.14 6.15 6.16 6.17 6.18 6.19 6.20 6.21 6.22 6.23 8 About statistical analysis 95 Statistical analysis of results of repeated measurements 97 Probability 98 Random variable 99 Dependent variable and independent variable 100 Correlated variables 100 Sample 101 Maximum likelihood value 102 Best estimate 102 Biased evaluation 103 Distribution of a variable 104 Normal distribution 106 Means 108 Mode 110 Median 111 Arithmetic mean 112 Experimental arithmetic mean 114 Indicators of dispersion of values 115 Natural Deviation 116 Experimental standard deviation 118 Deviation of the arithmetic mean 121 Deviation of the standard deviation 122 Standard deviation from a median, range and sample size 124 Kostić • 2019 • Metrology Companion 6.24 Degrees of freedom 125 6.24.1 Examples 126 6.25 Effective degrees of freedom 127 6.25.1 Example 127 6.26 Degrees of freedom from the deviation of the standard deviation of the arithmetic mean 128 6.26.1 Example 128 6.27 Confidence level and confidence interval 129 6.27.1 Example 131 6.28 Standard deviation and its degrees of freedom from the level and interval of confidence and sample size 131 6.29 T- distribution 133 6.30 Uniform distribution 136 6.31 Triangular distribution 138 6.32 U-distribution 140 6.33 Lognormal distribution 142 6.34 Chi-square distribution 144 6.35 F-distribution 145 6.36 Indicator of asymmetry of a distribution 146 6.37 Indicator of peakedness of a distribution 147 6.38 Histogram 148 6.38.1 Example 150 6.38.2 Example 151 6.39 Convolution 152 6.40 Filtering 154 6.41 Operations on distributions and central limit theorem 155 6.42 Monte Carlo simulation 158 6.42.1 Example of estimate of a value and its statistical parameters 162 6.42.2 Example of estimate of a length of gauge block and its statistical parameters 166 The contents continues on the next page. Kostić • 2019 • Metrology Companion 9 7 VARIANCES 171 7.1 7.2 7.3 7.4 7.5 7.6 Natural variance 171 Experimental standard variance 173 Evaluations of grouped results 175 Experimental standard covariance 176 7.4.1 Example of compution of the standard covariance 179 7.4.2 Example of compution of the standard covariance of the means 180 Standard correlation coefficient 181 Combined variance 183 7.6.1 Example of the combined variance of an evaluation of resistance of electrical circuit element and the maximum possible variance of this variance 191 7.6.2 Example of the combined variance of an evaluation of length of a gauge block and the maximum possible variance of this variance 193 7.6.3 Example of the combined variance of pH value 198 7.6.4 Example of the combined variance of resistance of resistors in a series 199 8 TESTS AND FUNCTIONS FITTING 203 8.1 Detection of outliers 203 8.1.1 Example 205 8.2 Fitting 206 8.3 Method of the least sum of squares 206 8.3.1 Example 209 8.4 P-value 212 8.5 Goodness-of-fit test 213 8.6 Estimating the type and parameters of a distribution 215 8.7 Pearson chi-square test for the goodness of fit of distributions 217 8.7.1 Example 221 8.8 Shapiro-Wilk test for normality of a population 223 8.8.1 Example 225 8.9 White noise test 230 8.10 Analysis of variance (ANOVA) 231 8.10.1 Example 233 8.11 Transformation of a variable 236 8.11.1 Example 237 10 Kostić • 2019 • Metrology Companion 9 MEASUREMENT UNCERTAINTY 239 9.1 9.2 9.3 9.4 9.5 9.6 9.7 9.8 9.9 9.10 9.11 9.12 9.13 9.14 Measurement result and its uncertainty 239 Type A standard measurement uncertainty 241 Type B standard measurement uncertainty 242 9.3.1 Example 243 9.3.2 Example 245 Combined standard measurement uncertainty 247 Expanded standard measurement uncertainty 248 Relative standard measurement uncertainties 249 Evaluations from non-normally distributed results 249 Expressing the standard measurement uncertainty 251 Expressing the standard measurement uncertainty when corrections are not applied 252 General guidance on reporting a measurement result 253 Procedure for estimating the measurement result and associated data 254 Example of calibration of a thermometer 256 9.12.1 The measurement problem 256 9.12.2 Least squares fitting 256 9.12.3 Standard uncertainty of the predicted correction addend 259 9.12.4 Elimination of the correlation between the slope and the intercept point 260 Resistance, reactance and impedance, computed from the same component measurement results 261 9.13.1 The measurement problem 262 9.13.2 Mathematical model of the measurement 262 9.13.3 Final results and their standard uncertainties, standard covariances and standard correlation coefficients - approach a 263 9.13.4 Final results and their standard uncertainties, standard covariances and standard correlation coefficients - approach b 263 Metrological traceability 268 9.14.1 Example 269 REFERENCES 271 Kostić • 2019 • Metrology Companion 11 Preface From a shop floor to the fundamental sciences, statistical analysis of results of repeated measurements is required. The purpose of the analysis is to determine the value of a measured quantity, as well as the accuracy of the determined value. It is requested that accuracy is described by the standard measurement uncertainty to enable traceability and ensure computing of the confidence level and the confidence interval. The essence of measurement is the same in all areas. The purpose of this publication is to be a companion to people dealing with measurements, that is experiments or observations, from production to research, from physics and chemistry to biology and medicine. The publication is also intended for students. In front of you is the Companion with concisely described subjects that are indispensable in all measurements. The Companion consists of about minimal and sufficient set of concepts and methods required to compute a measurement result and its measurement uncertainty. The text is in the form that enables direct applications in computations, manual or spreadsheet, and in writing software. The manner of presentation is encyclopedic. However, the order of presentation allows reading the text as a book or textbook. It seems that the form of encyclopedias is suitable for expert writings because it facilitates understanding of relations among concepts that make the essence of every exact knowledge. Most of this text is a terminologically consistent and compliant synthesis of international recommendations and regulations, articles and encyclopedic texts. References are given in brackets. Special attention was paid to the definitions of terms and their consistent use. Most of the terms are according to the second and third editions of the International vocabulary of metrology (see [VIM 2] and [VIM 3]) which was issued by the Joint Committee for Guides in Metrology (JCGM). These glossaries were not applied when it was thought that they do not correspond to what they should express, or that there are far better terms. Any such exceptions from the glossaries are noted with the references. Kostić • 2019 • Metrology Companion 15 The second part of the Companion, regarding statistical analysis of measurement results, is the basis for computation of the standard measurement uncertainty. It is intended for anyone involved in the analysis of the measurement results. The chapter on the standard measurement uncertainty is according to the instructions published by the JCGM in the Guide to the expression of uncertainty in measurement (see [GUM]). The level and interval of confidence of the measurement result can be considered as a purpose of determination of all measurement errors. An accurate evaluation of the level and interval of confidence is enabled by the standard measurement uncertainty of a traceable value. In this Companion, the traceability is described in the short, but comprehensive text given at the end of the section concerning measurement uncertainty. At the end of the Preface, attention is drawn to the following viewpoint expressed in [GUM]. “The evaluation of uncertainty is neither a routine task nor a purely mathematical one; it depends on detailed knowledge of the nature of the measurand and of the measurement. The quality and utility of the uncertainty quoted for the measurement result therefore ultimately depend on the understanding, critical analysis, and integrity of those who contribute to the assignment of its value.” All equations in the Companion are for the coherent units. The following marking is used in the Companion. ● Bold fonts are used for terms when defined. ● The abbreviated terms that may be used are provided next to the terms, in parentheses, when this will not lead to confusion. ● Underlined text is linked to detailed explanations of a concept which underlined text indicates. A “see...” is also linked to a reference. ● Equations, tables and figures are marked with the number of the chapter, then by a point, and subsequently by their order number in the chapter. ● A blue text needs special attention. For the help in the creation of the Companion, I wish to thank to engineer Siniša Hristov, professor Dr Gligorije Perović, Dr Walter Bich, professor Dr Dragić Banković, professor Dr Vesna Jevremović and Dr Emina Krčmar. To my longtime colleague, engineer Zlatko Sudar, I am grateful for comments which have led to an improved text. 16 Kostić • 2019 • Metrology Companion I am grateful to professor Ljiljana Sudar and professor Dr Živomir Petronijević for their efforts and valuable remarks of the trial readers. Thanks to Ljiljana D. Kostić for her trust and material support. This Companion was created as a part of the development project of measurement and control equipment, which is in progress in the laboratory for development of electronic products, Symmetry, based in Leskovac, Serbia. This is an updated e-edition that comes after the fundamental revision of the International System of Units (SI) as well as one English and two Serbian editions. Please send comments to the symmetry@ptt.rs. Leskovac, 15th August 2019 Goran Kostić Kostić • 2019 • Metrology Companion 17 Introduction The measurements are essential for human activity - from the trade to the fundamental sciences. International trade has been one of the important reasons for the establishment of the International Bureau of Weights and Measures (BIPM). The Bureau contributes to the reduction of technical barriers to trade, ensuring that measurements and tests carried out in different countries are considered equivalent. Scientists create their theories on a basis of measurement results which are for them facts about the material world, true within limits of measurement errors. The International System of Units was conceived after ancient discussions of European scientists about the need for a new system of measurement. The first proposal of a unified system of units, based on the metre (metric) and decimal, was made as early as 1670 by Gabriel Mouton, a priest from Lyon, France. That system replaced national and regional variants that made scientific and commercial communication difficult. The first decimal metric system was established in France in 1799 as an important fruit of the French Revolution. The world became metric in early 1993 when the International System of Units (SI) became primary in the USA government institutions. The current SI has emerged as perhaps the most significant revision of the SI. This revision was voted at the 26th General Conference on Weights and Measures (CGPM) in November 2018. The current definitions of SI base units are fundamentally different from those previously used. Now, the seven SI base units are defined using only seven “defining constants” whose values are taken as exact. The definitions are published in the 9th edition of the SI Brochure and are applied from 20th May 2019. The International Bureau of Weights and Measures was created by the Metre Convention signed in Paris on 20th May 1875. The Convention was amended in 1921 and has remained the basis of the international agreement on measuring units until now. The original signatories of the Convention were representatives of 17 States, and on 15th August 2019, there were 60 Member States of the Convention and 42 Associates of the Kostić • 2019 • Metrology Companion 19 CGPM. Since its establishment, the headquarters of Bureau is in Sevres near Paris. The purpose of Bureau is to ensure worldwide unification of measurements in order to enable traceability of measurement results to the International System of Units. Therefore the task of Bureau is to: ● “represent the worldwide measurement community, aiming to maximize its uptake and impact ● be a centre for scientific and technical collaboration between the Member States, providing capabilities for international measurement comparisons ● be the coordinator of the worldwide measurement system, ensuring it gives comparable and internationally accepted measurement results.” The Bureau operates under the exclusive supervision of the International Committee for Weights and Measures (CIPM) which itself comes under the authority of the General Conference on Weights and Measures. The International Committee and the General Conference was established by the Metre Convention, at the same time as the Bureau. By harmonizing and providing complete coverage of all areas, the International System of Units is an outstanding tool of the modern sciences and practices that created it. [SI] Appendix 4. Part 2, 1.1, PP 117 [Britannica] [NIST] 20 Kostić • 2019 • Metrology Companion 1 Quantities and units of measurement 1.1 Measurable quantity Measurable quantity (quantity) is a property of a body, a substance, or a phenomenon, that may be distinguished qualitatively and determined quantitatively. Quantity is not a body, a substance or a phenomenon. Quantity is independent of measurement and system of quantities. The term quantity may refer to a quantity in a general sense (length, time, mass, temperature...) or to a value of a particular quantity (the length of a given rod, the electric resistance of a given wire, a number of entities...). [VIM 2] 1.1 [VIM 3] 1.1 [Sonin] 2.3 1.2 Quantities of the same kind Quantities of the same kind are quantities whose values may be added or subtracted, but that it has a physical representation. Examples. The addition of values of all forms of energies, such as the amount of heat, kinetic energy and potential energy, has a physical representation. Thus, all forms of energies, are quantities of the same kind. A value of energy and a value of moment of force have the same dimension but there is no physical representation of addition or subtraction of their values, so those quantities are not of the same kind. In a given system of quantities: - quantities of the same kind have the same dimension - quantities of the same dimension are not necessarily of the same kind - quantities of the different dimension are always of a different kind. Quantities of the same kind are grouped into categories of quantities. Examples of the categories of quantities are: - thickness, circumference, wavelength - heat, work, energy. [Quinn] [VIM 2] 1.1 Notes [VIM 3] 1.2, 1.7 Kostić • 2019 • Metrology Companion 23 1.4 Base quantity Base quantity, together with other base quantities of a system of quantities, enables obtaining derived quantities. The selected base quantity always has a physical representation so that it may be realized using its definition. A base quantity is a property for which the following mathematical operations are defined in physical terms: comparison, addition, subtraction, multiplication by a quantity of dimension one and division by a quantity of dimension one. Each of these operations is performed on physical properties of the same kind and yields a physical property of that kind. Each of these physical operation obeys the same rules as the corresponding mathematical operation for pure numbers, that is, quantities of dimension one. This makes possible to describe the physical properties by the language of mathematics. The choice of a set of base quantities is somewhat arbitrary. However, for the selected base quantities it must be possible to define the operations previously specified. It is preferred that the base quantities of a set are mutually independent. The set may be minimal, or somewhat larger, than one out of which it is possible to obtain all necessary derived quantities. Example. The base quantities may be: the length, mass, area, speed and force. A shape and color cannot be base quantities because they lack acceptable addition. Things from nature can be described only by comparison with the several well-known things among which are the base quantities. [Sonin] 2.2, 2.1, 2.8 [SI] 2 Kostić • 2019 • Metrology Companion 25 1.21 System of units System of units is formed by the base units, the derived units and equations that define relations between these units. International System of Units, SI, is based on the International System of Quantities. The General Conference on Weights and Measures recommends the SI. SI coherent units are formed only by seven SI base units (base units of the SI) and twenty-two SI coherent derived units. The complete set of SI units (SI units, or units of the SI) is formed by SI coherent units and SI decimal units. The complete set of SI units is not coherent. The complete set of SI units is divided into the class of SI base units and the class of SI derived units. The definitions of seven SI base units are provided in Table 1.1 and described in the following section 1.21.1. Out of seven SI base units, in almost all areas are used: metre, kilogram and second. The other four units are used to define derived units that are specific to particular areas: ● ampere, for electrical and magnetic units ● kelvin, for heat units ● mole, for units of chemistry and molecular physics ● candela, for light units. The SI coherent derived units with special names and symbols are listed in Table 1.2. These special names and symbols may be used in combination with the names and symbols for base units and for other derived units to express the units of other derived quantities. The “SI Brochure” [SI] recommends using the complete set of SI units and only in particular circumstances (to satisfy the needs of commercial, legal, or specialized scientific interests) non-SI units listed in Table 1.3. When the non-SI unit is used, it is recommended to define the non-SI unit in terms of the SI units. By harmonizing and providing complete coverage of all areas, the SI is an outstanding tool of the modern sciences and practices that created it. [Sonin] 2.7 [SI] 2, 3, 4, in which “decimal multiples and sub-multiples of SI units” is used instead of “SI decimal units” 36 Kostić • 2019 • Metrology Companion 1.21.1 SI base units Seven SI base units are defined using only the seven defining constants. The values of defining constants were fixed after being identified as the best estimates at the time just before the decision on the revision of the SI. In this way the SI is scaled to the seven physical constants of high reproducibility, and these constants become exact. See Figure 1.1 and Table 1.1. The reference definitions of base units are published in the SI Brochure, see [SI]. The current definitions of SI base units have brought important benefits. The most important are: 1) a large number of physical constants become either exactly known or known with higher accuracy; 2) values of quantities that are much smaller or much larger than the base units can be measured with unreduced accuracy; 3) liberation of mass metrology from the unit based on the mass standard which is undoubtedly changing with time; 4) the Josephson effect and Hall effect can be used to directly realize the SI definitions of most electrical units; 5) the definition of the kelvin does not make reference to a water having the defined isotopic composition. Some of the consequences of the current definitions are that the following constants are not exactly known, and must be measured: 1) the magnetic constant, the electric constant and the characteristic impedance of a vacuum; 2) the temperature of triple point of water; 3) the molar mass of carbon 12. The current definitions of SI base units are fundamentally different from those previously used. However, as it was the case in the past revisions of the SI, the transition to the current definitions was realized without noticable impact on daily life, while the measurements based on previous definitions of the units remain valid within their measurement uncertainties. Figure 1.1. The SI base units are defined using the set of only seven defining constants whose values are taken as exact. Each base unit is defined, either, through the corresponding constant from the set, or through the constant from the set and other base units that are also defined through the constants from that set. Kostić • 2019 • Metrology Companion 37 1.25 Expressing the value of a quantity in the SI A numerical value of a quantity and unit symbol are printed in upright typeface. The numerical value always precedes the unit symbol, and a space is always used to separate the number from the unit symbol. The only exceptions, in which no space is left between the number and the unit symbol, are the unit symbols for plane angle: degree, °, minute, ’, and second, ”. A prefix symbol precedes the unit symbol without a space, and it is considered as a new inseparable unit symbol, as well as mathematical entity. A numerical value of a quantity with a unit symbol is regarded as the product, and they are treated using the ordinary rules of algebra. One must neither use the plural of a unit symbol, nor the period after a symbol except at the end of a sentence. A prefix symbol, formed by placing together two or more prefix symbols, are not permitted. This also applies to prefix names. A prefix symbol and a prefix name will not be used with the following symbol and name of the units of time: min, minute; h, hour; d, day. No space or hyphen is used between the prefix name and the unit name. The combination of prefix name plus unit name is a single word. In an expression, only one unit is used. An exception is in expressing the values of time, and of plane angles, using non-SI units. However, for plane angles it is generally preferable to divide the degree decimally, except in fields such as navigation, cartography, astronomy, and in expressing very small angles. Example, one would write 22.20° rather than 22° 12’. Unit symbol is never qualified by further information about the property of the quantity. Extra information on the property of the quantity should be attached to the quantity symbol, and not to the unit symbol. Example, U max = 50 V, but not U = 50 Vmax. In forming products and quotients of unit symbols the normal rules of algebraic multiplication or division apply. Multiplication must be indicated by a space or a half-high dot (⋅). Division is indicated by a horizontal line (――), by a solidus (/), or by negative exponents. When several unit symbols are combined, care should be taken to avoid ambiguities, for example by using brackets or negative exponents. To remove ambiguities, a solidus without brackets, must not be used more than once in a given expression. Among the base units of the SI, the kilogram is the only whose name and symbol, for historical reasons, includes a prefix (“k”). Symbol and name for the decimal unit of mass are formed by attaching prefix symbol to the unit symbol “g”, and prefix name to the unit name “gram”. Example, −6 10 kg is written as 1 mg, not as 1 µkg. [SI] 3, 5 Kostić • 2019 • Metrology Companion 43 Table 1.2 SI coherent derived units with special names and symbols [SI] 2.3.4 Expressed in terms of other SI units SI derived quantity Special name of SI unit Special symbol for SI unit plane angle radian rad m · m −1 = 1 solid angle steradian sr m2 · m − 2 = 1 frequency hertz Hz s−1 force newton N kg · m · s − 2 pressure, stress pascal Pa N / m2 kg · m−1 · s−2 energy, work, amount of heat joule J N·m= W·s kg · m2 · s − 2 power, radiant flux watt W J/s kg · m2 · s − 3 coulomb C F·V A·s volt V W/A kg · m2 · s−3 · A−1 ohm Ω V/A kg · m2 · s−3 · A−2 electric conductance siemens S A/ V kg · m · s · A electric capacitance farad F C/V kg · m · s · A electric charge, amount of electricity (electric) potential difference, electromotive force, voltage, (electric) tension electric resistance electric inductance henry Expressed in terms of SI base units −1 −2 −1 3 −2 2 2 4 −2 2 −2 H Wb / A kg · m · s · A kg · m2 · s−2 · A−1 magnetic flux weber Wb V·s= T · m2 magnetic flux density, magnetic induction tesla T Wb / m2 = N / (A · m) kg · s − 2 · A−1 Celsius temperature degree Celsius °C 4) K 4) K 4) luminous flux lumen lm cd · sr m · m · cd = cd illuminance lux lx lm / m2 m − 2 · cd activity referred to a radionuclide (ionizing radiation) absorbed dose, specific (imparted) energy, kerma (radiation) dose equivalent becquerel Bq gray Gy J / kg m2 · s − 2 sievert Sv J / kg m2 · s − 2 catalytic activity katal kat 2 −2 s −1 mol · s −1 4) The units °C and K are equal only when expressing temperature difference or width of temperature interval. The numerical value of a Celsius temperature expressed in degrees Celsius is related to the numerical value of the thermodynamic temperature expressed in kelvins by the relation: t [°C] = T [ K] − 273.15. [SI] 2.3.1 The kelvin 44 Kostić • 2019 • Metrology Companion 2 Measurements 2.1 Measurement Measurement is a determination of a value of a quantity by its comparison with the known value of the quantity. The value of a quantity may be measured with very high accuracy except perfect. The following are necessary for measurement: a specification of measurand, a description of measurement procedure and a device for measurement. Measurement does not apply to nominal properties. Examples: sex of a human being; ISO two-letter country code; the sequence of amino acids in a polypeptide... [Sonin] 2.3 [GUM] 3.1.1 [VIM 2] 2.1 [VIM 3] 2.1, 1.30 Examples 2.2 Metrology Metrology is the science of measurements. Metrology includes all theoretical and practical aspects of measurements, whatever the measurement accuracy and field of application. Legal metrology is the part of metrology relating to measurements specified in the law. The scope of legal metrology commonly includes: ● protection of the interests of individuals and enterprises ● protection of national interests ● protection of public health and safety, especially those related to the environment and medical services ● meeting the requirements for commerce and trade. [VIM 2] 2.2 [VIM 3] 2.2 [VIML] 1.03 Kostić • 2019 • Metrology Companion 51 2.3 Measurand Measurand is a particular quantity subject to measurement. The requested accuracy determine how many details should be given in a specification of a measurand. For example, if the length of a nominally one-metre long steel bar is to be determined to micrometre accuracy with relative error of about 5 · 10 − 6, its specification should include the temperature and pressure of the environment in which is the bar. [VIM 2] 2.6 [VIM 3] 2.3 [GUM] 3.1.3 2.4 Principle of measurement Principle of measurement (or measurement principle) is a phenomenon serving as the basis of a measurement. Example: the thermoelectric effect applied to a measurement of temperature. [VIM 2] 2.3 [VIM 3] 2.4 2.5 Method of measurement Method of measurement (or measurement method) is a concept of applying the principle of measurement. Primary method of measurement (or primary measurement method, primary method) is a method having the highest metrological qualities, whose procedure is completely explained and described, and for which the combined standard uncertainty traceable to the SI units is determined. Primary ratio method is a method used to determine a quotient of value of measurand and value of measurement standard (where both values are of the same kind). Primary direct method (primary method) is a method in which a value of measurand and values of used measurement standards, are of different kind. [BIPM] [NIST] 52 Kostić • 2019 • Metrology Companion 3 Devices for measurement 3.1 Device for measurement Device for measurement is a device, or combination of devices, designed for measurement. The device for measurement may be a measuring instrument or a measuring transducer. A measuring instrument is an indicating measuring instrument or a material measure. A material measure is: a reference material, weight, volume measure, gauge block, standard electric resistor, standard signal generator... Devices for measurement arranged in approximate order of increasing complexity are: a part, material measure, reference material, measuring transducer, measuring instrument, measuring chain, measuring system, measuring installation. Device for measurement is periodically checked, or qualified, by a user or a recognized laboratory. The checks determine whether an error of a device for measurement, in operating conditions, is within a specified interval and whether it will be within this interval in future. [VIM 2] 4 introduction, 4.1, 4.3, 4.6, 4.2 [VIM 3] 3, 3.1, 3.7, 3.3, 3.6 Figure 3.1. A measuring transducer and a measuring installation. Left: an ion-selective probe for concentration of sodium measurement (Thermo Scientific). Right: a highly automated measuring installation, 18 m long, for biochemical analysis in a hospital laboratory (Roche Diagnostics). Kostić • 2019 • Metrology Companion 61 3.2 Indicating measuring instrument Indicating measuring instrument is a measuring instrument providing an output signal that contains information about the value of a measurand. Examples: a thermometer, micrometer, electronic balance, barograph. Displaying measuring instrument (or meter) providing an output signal in visual form. Recording measuring instrument provides a record of its indication as an output signal. [VIM 2] 4.6 [VIM 3] 3.3 3.3 Material measure Material measure (measure) is a device for measurement intended to supply, or reproduce, one or more known values of a quantity. Examples: a weight, volume measure, gauge block, standard electric resistor, standard signal generator, reference material. An indication of a material measure is its assigned value. An error of a material measure is the error of the indication of the material measure. End gauge is a material measure which supplies one known value. Example. A length end gauge has the form of: parallelepipeds (gauge block), pins, blades, plugs, rings, snaps. Go - no go gauge consists of two end gauges which determine whether a value of a measurand is within an interval given by the end gauges. [VIM 2] 4.2, 3.2 Note 3, 5.20 Note 3 [VIM 3] 3.6, 4.1 62 Kostić • 2019 • Metrology Companion 4 Measurement results, accuracy and errors 4.1 Measurement result Measurement result (or result of measurement, or estimated value) is a value assigned to a measurand on the basis of one or more results of measurements. A measurement result is obtained in the manner specified in the description of the measurement procedure: ● taking a result of a single measurement (or particular result), ● computing the arithmetic mean of the results of repeated measurements, ● taking the value that appears most often in the results of repeated measurements (that is, taking the mode), ● computing on the basis of a functional relationship between the value of a measurand and component values (or component results, or component variables, or component quantities, or input quantities, or components) on which the value of the measurand depends... Results of repeated measurements of the same quantity are represented by the random variable which is characterized by statistical parameters. A complete measurement result includes information about its accuracy. [VIM 2] 3.1 [GUM] [VIM 3] 2.9 4.2 Accuracy Accuracy of a value is a qualitative concept which indicates a closeness of this value to a reference value. Accuracy of a measurement result (accuracy) is a qualitative concept which indicates a closeness of this measurement result to a value of the measurand. Accuracy of a measurement result is determined in the manner specified in the description of the measurement procedure. Kostić • 2019 • Metrology Companion 75 Accuracy of a measurement result depends on errors which are classified into the two groups. The first group is random errors, whose influence on the final measurement result can be reduced by increasing the number of repeated measurements and then computing the arithmetic mean of the measurement results. The second group is systematic errors, whose influence on the final measurement result can be reduced by applying certain correction. A complete measurement result includes information about its accuracy. However, if inaccuracy is considered to be negligible for the intended purpose, the information about accuracy may be omitted. In many fields, the ordinary way of expressing a measurement result does not include the information about accuracy. In order to enable the use of one measurement result and its accuracy as component values of another measurement result and its accuracy, the information specified in 9.10 General guidance on reporting a measurement result should be provided. Results of repeated measurements of the same quantity are represented by the random variable which is characterized by statistical parameters. The description of the accuracy of the result of repeated measurements is its standard measurement uncertainty. The term precision should not be confused with accuracy. [VIM 2] 3.5 [VIM 3] 2.13, 2.15, 2.9 Note 2 [GUM] 3.2 4.3 Absolute error Absolute error (error, or difference, or divergence) of a value is given by formula: ABSOLUTE_ERROR = VALUE – REFERENCE_VALUE. (4.1) VALUE = REFERENCE_VALUE + ABSOLUTE_ERROR (4.2) REFERENCE_VALUE = VALUE – ABSOLUTE_ERROR (4.3) ABSOLUTE_ERROR = = SYSTEMATIC_ERROR + RANDOM_ERROR (4.4) In practice, the conventional value is used as the reference value. The value and its absolute error have the same dimension. 76 Kostić • 2019 • Metrology Companion 5 Measurement standards and calibrations 5.1 Measurement standard Measurement standard (standard, or etalon) is a device for measurement intended to realize, reproduce or define the value of a quantity to serve as a reference in measurements. Examples: a Kibble balance (or a watt balance); 1 kg mass standard; 100 Ω standard resistor; standard ammeter; caesium frequency standard; hydrogen reference electrode; reference solutions of cortisol, in human serum, with specified concentration. Figure 5.1. Left: a Kibble balance in the International Bureau of Weights and Measures. Among other things, the balance is used for measuring a 1 kg mass standard with the combined standard measurement uncertainty of 2 · 10 − 8. The measurements are performed through the Planck constant, gravitational acceleration, velocity, plus electromagnetic quantities, and represent the realization of the definition of the kilogram. Right: a Josephson voltage standard used for driving a stable current through the coil of the above-mentioned Kibble balance using a standard resistor. (Courtesy of the BIPM.) [BIPM] Kostić • 2019 • Metrology Companion 85 Intrinsic measurement standard (intrinsic standard) is based on an inherent and reproducible phenomenon in nature. Examples: the speed of light in vacuum; the elementary charge; a triple-point-of-water cell as a standard of temperature; a Josephson array as a voltage standard; a quantum Hall effect as a standard electric resistor. Collective standard (or ensemble of standards) is a set of standards that, through their combined use, constitute a standard. Group standard is a set of standards that, individually, or in combination, provide a sequence of values of quantities of the same kind. Table 5.1 presents the most accurate standards, or measurements, of the base quantities and their approximate combined standard measurement uncertainties. [VIM 2] 6.1 [VIM 3] 5.1 in which the associated measurement uncertainty of a measurement standard is required, 5.10 5.2 Reference material Reference material, RM, is material with homogeneous, stable and accurately defined properties used as a reference in measurements. Reference material may be a pure or mixed: gas, liquid or solid. Examples: a solution used in calibration of measuring instruments for chemical analyses; a color chart; a fish tissue containing a specified mass concentration of dioxin. Certified reference material, CRM, is the reference material accompanied by a certificate specifying the values of the material properties and traceable standard measurement uncertainty of these values. Example. Human serum accompanied by a certificate specifying the value of the concentration of cholesterol and the associated traceable standard uncertainty. [NIST] [VIM 2] 6.13, 6.14 [VIM 3] 5.13, 5.14 86 Kostić • 2019 • Metrology Companion Table 5.1 The most accurate standards, or measurements, of the base quantities, state in 2019 [BIPM] [Newell 2014] [NIST] [SI realization] Quantity Conditions Absolute standard uncertainty Relative standard uncertainty time The best of primary standards which realizes the SI second in the BIPM. 1 · 10−16 s 1 · 10 − 16 The length of the cubic unit cell edge of pure 28Si (5.431 · 10 −10 m) measured using interferometry. 3 · 10 −18 m 5 · 10 − 9 A steel gauge block of the length of 100 mm measured using optical interferometry. 10 nm 1 · 10 − 7 The Earth-Moon separation (376 203 km) measured by a laser and a retroreflector on the Moon. 1 mm 3 · 10 −12 A mass of single atom (an order of magnitude of 10 − 25 kg) through frequency measurements of photons during absorption or emission of that atom. ca 10−33 kg 2 · 10 − 9 A 1 kg weight measured on a Kibble balance. 20 µg 2 · 10 − 8 electric current Generating a current up to 10 mA using a Josephson voltage standard and a Hall standard resistor. < 6 · 10 thermodynamic temperature −155 °C to 30 °C measured using the relative primary acoustic gas thermometry in the NPL. < 1 mK amount of substance The number of atoms in a single crystal of 1 kg of the pure 28Si (2.153 · 10 25 atoms) using volumetric and interferometric measurements. 4 · 10 atoms 2 · 10 − 8 luminous intensity The light source, from 10 − 3 cd to 10 4 cd, directly measured with a standard photometer. 2 · 10 − 6 cd to 20 cd 2 · 10 − 3 length mass Kostić • 2019 • Metrology Companion −12 A 17 6 · 10 −10 for kelvin scale < 7 · 10 − 3 87 6 Outlines of the statistics 6.1 About statistical analysis Statistical methods are used to determine the representative value of an investigated property for a group of elements. The elements within the group can have significantly different values of the property. Evaluation (or estimation, or “statistic”) is a value estimated by the statistical method. Population is a group of all elements for which we determine the representative value of a property by the statistical method. The population has a finite or infinite number of elements. Figure 6.1. The example of four samples of values which yield: the equal arithmetic means of variables x and y, the equal deviations of x, the same linear regression function obtained by the method of least squares... [Anscombe] Kostić • 2019 • Metrology Companion 95 The investigated property of a population is determined by the statistical method usually on a basis of a sample of the population that truly represents the population. The sample truly represents the population when a number of its elements is sufficiently large and when the elements are randomly selected from the population. In order to prove that the conclusions drawn from the sample are with a certain probability valid for the population can be proved by goodness-of-fit tests. In order to reduce the possibility of errors in statistical analysis, data graph should be necessarily considered. The graph provides an overview of qualitative features of data. The diagrams in Figure 6.1 demonstrate the importance of graphical presentation and visual analysis of data. A property of a population determined by a statistical method is the property of a large number of elements of that population. On a basis of statistical evaluation, we can determine only the probability that a particular element of a population has a certain property. Statistics is subjective; statisticians try to explain or predict the material world through arbitrary, but reasonable way, using probability theory, mathematics and common sense. Unlike the statistics, probability theory for a fully defined problem gives a unique and a reproducible solution. Probability theory is the algebra of inductive reasoning. (Boolean algebra is the algebra of deductive reasoning.) Statistical methods are among the most widespread quantitative methods, with important applications in almost all human activities. [Njegić] 1.1., 1.2. [Ivković] II.1 [Britannica] Probability theory [Drake] 7-1, 7-6 [Lazić] I, 1.2 [Anscombe] 96 Kostić • 2019 • Metrology Companion 6.2 Statistical analysis of results of repeated measurements The purpose of statistical analysis of results of repeated measurements of a particular quantity is to determine the best estimated value of the quantity. And also, to determine an interval about the best estimated value which, with requested probability, encompass a true value of the quantity. Normally, for the best estimated value of the quantity is taken the maximum likelihood value of this quantity. The results of repeated measurements most often have normal distribution because the central limit theorem is most often valid. In the case of normal distribution, and after correction for all significant systematic errors, the maximum likelihood value of the quantity is an arithmetic mean of the results of measurements. For the normal distribution, the interval: ARITHMETIC_MEAN ± DEVIATION_OF_THE_ARITHMETIC_MEAN is the interval in which a true value of the quantity is with the probability of 0.68. [GUM] 3, 4 Kostić • 2019 • Metrology Companion 97 6.24 Degrees of freedom Degrees of freedom, ν, of an evaluation estimated from a sample of size n is given by formula (6.64). The R is a number of evaluations estimated from this sample and used to estimate this evaluation. See also 6.24.1 Examples. ν=n–R (6.64) The least possible sample size is 1, and from this sample it is possible to estimate an evaluation of the degree of freedom equal to 1. Evaluations with the degrees of freedom equal to 1 have the greatest possible dispersion. The dispersion is reduced by increasing the sample size, that is, the degrees of freedom, so the degrees of freedom is the indicator of reliability of an evaluation. The degrees of freedom of an evaluation is equal to the number of elements of the sample above the least number of elements necessary to compute the evaluation, plus, 1. The degrees of freedom defined by (6.64) gives a value which is a parameter of different distributions. For example, a sample from population with the normal distribution, for which (6.64) gives a degrees of freedom ν, has the T-distribution with parameter “degrees of freedom” equal to ν. Also, the standard uncertainty, for which (6.64) gives a degrees of freedom ν, has the T-distribution with parameter “degrees of freedom” equal to ν. If for a sample from a population with the normal distribution, is known its standard deviation s, and a degrees of freedom of this deviation, ν, formula (6.65) gives half-width of confidence interval, Δ, for a confidence level of an interval, P. The coefficient k is the coverage factor which can be determined from the relation among P, ν and k listed in Table 6.3 for the T-distribution. Formula (6.65) is derived from (6.77). Δ = k·s (6.65) The degrees of freedom enables accurate computation of a level and interval of confidence, even in the case of small sample size. The quantities Δ and k from (6.65) can be used as an expanded standard measurement uncertainty and its coverage factor. Without giving a general definition, the concept of the degrees of freedom first started to use Sealy Gosset (“Student”) in 1908 and Ronald Fisher in 1915. Since then, the concept is based on an incomplete theory. A particular problem is the computation of the degrees of freedom. [Njegić] 5.6. [GUM] 4.2.6, G.3.2, G.3.3, G.6.4 [Eisenhauer] Kostić • 2019 • Metrology Companion 125 6.27 Confidence level and confidence interval Confidence level (or level of confidence, or coverage probability) is a probability that a value of variable is from a specified interval. Confidence interval (or interval of confidence, or two-sided confidence interval, or coverage interval) is the interval which has the specified probability that the value of variable is from it. Confidence limits are the limits of confidence interval. The number of measurement results is from a particular interval. The probability that the result is from mentioned interval is equal to the quotient of the number of results from mentioned interval, and the total number of results. The mentioned interval is the confidence interval. The mentioned probability is the confidence level. Generally, there is more than one coverage interval for a specified probability. Shortest confidence interval of a variable is a confidence interval with the narrowest width among all coverage intervals having the same coverage probability. In the case of unimodal distribution, the shortest confidence interval is to the left of the mode and to the right of the mode. Unless otherwise indicated, the confidence interval means the shortest confidence interval. A confidence level, P, of a variable x with a probability density function f(x), for confidence interval [a, b], is given by (6.76). With (6.76) there is a graphical example for the normal distribution. P = P(a ≤ x ≤ b ) = b f( x ) ⋅ d x (6.76) a For a variable with the normal distribution, it can be computed an arithmetic mean x , and standard deviation of the mean, s. For a requested confidence level, P, a confidence interval can be given in the form of: [ x − k ⋅ s , x + k ⋅ s ] or simplified x ± k ⋅ s . (6.77) In the case of the normal distribution, the coefficient k can be determined from function (6.78) which gives the relation between P and k. The values in Table 6.2 are obtained from (6.78). P(k ) = 2 ⋅ π k 0 1 k2 2 ⋅ dk (6.78) e The coefficient k is called the coverage factor. Product k · s is a half-width of confidence interval (or one-sided confidence interval). Kostić • 2019 • Metrology Companion 129 A standard deviation of a variable with the normal distribution is a half-width of a confidence interval of the variable at the confidence level (of the interval) equal to 0.6827. One or more measurement results, from a population with the normal distribution, have the T-distribution. For details and about levels and intervals of confidence, see 6.29 T-distribution. The level and interval of confidence of the measurement result can be considered as a purpose of all determinations of measurement errors. [Ivković] PP 79, PP 81, PP 17 [GUM] 7.2.4, G.3.2, G.5.3, G.1.4 Table 6.2 For the normal distribution, N(x): relations between the confidence level, P, and coverage factor k P 0.079 7 0.100 0 0.158 5 0.200 0 0.300 0 0.382 9 0.400 0 0.500 0 0.600 0 0.682 7 0.700 0 0.800 0 0.850 0 0.900 0 0.950 0 0.954 5 0.970 0 0.980 0 0.990 0 0.995 0 0.997 3 0.998 0 0.999 0 0.999 9 0.999 99 0.999 999 130 k 0.100 0.126 0.200 0.253 0.385 0.500 0.524 0.675 0.842 1.000 1.036 1.282 1.440 1.645 1.960 2.000 2.170 2.326 2.576 2.807 3.000 3.090 3.291 3.891 4.417 4.892 Kostić • 2019 • Metrology Companion 6.38 Histogram The sample of values can be splitted into disjoint subintervals of equal widths (or cells, or bins). The histogram is the graphic representation of a number of the values that fall into each of the subintervals. The histogram is made in the following way. First, split a part of the horizontal axis which contains the values from the sample into disjoint subintervals of equal width, see Figure 6.14 b). Then draw rectangles whose bases overlap subintervals on the horizontal axis and whose heights are proportional to the number of values that fall into the subinterval of their base. The histogram appearance is significantly affected by the subintervals width and their position along the axis. The limits of subintervals can be determined according to the following two recommendations. a) If we make the histogram for a sample with a small number of possible discrete values, the subintervals should be chosen so that each discrete value corresponds to one subinterval. This is the case in Example 6.38.1, in which 8033 discrete measurement results have one of 16 possible discrete values, out of which only 5 values are distinguished by the number of appearances. Histograms, such as in this case, are bases for reliable estimating. b) If we make the histogram for a sample with a large number of possible values, subintervals width and position in the horizontal axis can be determined as follows. If the n values in the sample have the normal distribution and n is larger than about 25, the optimal subinterval width w, is given by formula (6.116) with the parameter C = 3.5. The standard deviation of values in the sample is denoted by s. This formula is derived from the criterion of the minimum of integrated mean squared error (IMSE). w = C⋅ s 3 n (6.116) If the values in the sample do not have the normal distribution, and n is larger than about 25, based on the experience we can also recommend using (6.116) with C = 2. In the case of discrete values, it is necessary to round the subintervals width so that each subinterval contains the same number of possible values (that is, round subintervals width to an integer multiple of a resolution). 148 Kostić • 2019 • Metrology Companion The limits of subintervals should be chosen so that more of the values are around the middle of subintervals. In the histograms described in b), it is classified the case in Example 6.38.2, in which 78 discrete measurement results have one of 30 possible values. This example shows how the subintervals width affects the histogram appearance. In most statistical software a default subinterval width and position are not optimal. The height of a rectangle of a histogram is proportional to the probability that values fall into the subinterval of the rectangle base, so the histogram is the variation of the probability density function. By visual inspection of the well-made histogram, we can estimate the following properties of the distribution: a type; a mode; an existence of outliers; a dispersion of values; an existence of several maximums; a symmetry; and a peakedness. The lack of an evaluation based on the histogram is subjectivity. The approximate arithmetic mean of n values represented by a is given by formula (6.117) derived from (6.19). The bi is a histogram, x, number of values in a subinterval with the middle xi . x ≈ 1 1 n ⋅ (b1 ⋅ x1 + b2 ⋅ x2 + ... + bn ⋅ xn ) = ⋅ bi ⋅ xi n n i =1 (6.117) [Britannica] Statistics [Perović, communication] [Scott] [Hristov, communication] Kostić • 2019 • Metrology Companion 149 6.40 Filtering Filtering is a procedure to reduce random differences of values in a sequence. Filtering is most often performed by averagings. Moving average filtering uses the convolution, it is simple and optimal to reduce random differences of values in a sequence. This filtering retains sharp changes but gives the wrong idea about values and widths of peaks in the sequence. These disadvantages are worsened by decreasing a bandwidth of the filtering. In Figure 6.17 there is an example of filtering by using the convolution. The weight function (that is, window) is the Chebyshev type sequence with seven coefficients (that is, points) scaled so that their sum equals to one. (The moving average filtering using the weight function with equal coefficients.) Filtering is applied to the fluctuating discrete measurement results. These results with the corresponding histogram are also shown in Figure 6.15. In addition to averaging, filtering is also implemented by fitting. Savitzki - Golay filtering uses local polynomial fitting across a moving window within the values in the sequence. The polynomial is often of the fourth degree, and it is fitted by the method of least squares. Such filtering tends to preserve the values and widths of peaks in the sequence. See [Savitzky]. [Savitzky] [Hristov, communication] [Guiñón] 154 Figure 6.17. Filtering of the sequence of discrete measurement results by using the convolution: a) the Chebyshev type weight function, b) the sequence of discrete measurement results and c) the convolution of these two functions. [Hristov, communication] Kostić • 2019 • Metrology Companion 6.41 Operations on distributions and central limit theorem When a die is thrown, each number from 1 to 6 comes up with equal probability. The probability function of these numbers, P(x1), is uniform and discrete, and it is represented pictorially in the following figure. Think about the probabilities to happen values which are the sum of two numbers obtained from two tosses of a die. The possible values of the sum are 2 to 12. There is only one way of making 2, two ways of making 3, etc. See the following picture. The sum 7 is the most probable because it can be achieved in the greatest number of ways. The random variable x1 , which assume values obtained from tosses of the die, has the uniform and discrete probability function. The random variable x2 , whose values are the sum of the two numbers obtained from two tosses of the die, has the triangular and discrete probability function. As an example, using the properties of probability, we can compute the probability that the sum of x2 is 5: P(x2 = 5) = P[(4 1) (3 2) (2 3) (1 4)] = = P(4 1) + P(3 2) + P(2 3) + P(1 4) = = P(4) · P(1) + P(3) · P(2) + P(2) · P(3) + P(1) · P(4) = 2 = 4 · (1/6) = 4/36 = 1/9. We see that P(x2 = 5) can be written as P( x2 = 5) = 4 P(5 − i ) ⋅ P(i ), i =1 which is the convolution of discrete function P with itself. Based on previous example, we can determine the probability function of the x2 , P2 (x2 ): 1/ 6, x1 = 1, 2, ..., 6 P1 ( x1 ) = 0 ≥ x1 ≥ 7 0, P 2 ( x2 ) = ∞ P1 ( x2 − i ) ⋅ P1 (i ) = P1 ( x2 ) ∗ P1 ( x2 ). i = −∞ Kostić • 2019 • Metrology Companion 155 Analogous to the previous discrete probability functions, we can determine the probability function of the sum of variables with continuous probability functions. See [Osgood] 3.6.6. For discrete and for continuous probability density functions the following applies. The probability density function of a sum of two independent random variables is the convolution of the probability density functions of those two variables. That is, if f is the probability density function, and if x1 and x2 are independent variables, then: f (x1 + x2 ) = f (x1 ) * f (x2 ). (6.127) From the example in Figure 6.18, we can see the following. The sum of two independent variables, with the uniform distributions of the same width, has a triangular distribution. Approximately normal distribution has the sum of the variable with the triangular distribution and variable with the uniform distribution, as well as the sum of variables with normal distribution and variable with the uniform distribution. The example in Figure 6.18 shows how from chaotic differences of variables the normal distribution emerges necessarily which suggests the central limit theorem. Central limit theorem states that a sum of independent variables, with any distributions, approaches the normal distribution as a number of the variables increases, but if there is no variable with multiple times greater differences than other differences. See [Osgood] 3.7 and 3.10. [Osgood] 3.6, 3.7, 3.10 [Box 2005] PP 28 156 Kostić • 2019 • Metrology Companion Figure 6.18. The variable x1 in time, its distribution and the distributions of the sums of independent variables xi with the distributions as x1. Kostić • 2019 • Metrology Companion 157 6.42 Monte Carlo simulation Monte Carlo simulation, MCS (or Monte Carlo method, MCM) is used to estimate a value by performing mathematical model computations for a large number of combinations of input values. The combinations of input values are chosen so that intervals of interest are simulated as uniformly as possible and at enough points. The input values can be a sequence of numbers represented in the diagrams below, each with one hundred pairs of the: a) equidistant numbers x and y, b) numbers with the uniform distributions, c) numbers with quasi-random distributions, d) numbers with the normal distributions and e) linearly correlated numbers with the normal distributions. Input numbers with quasi-random distributions are suitable when a model has many variables and depends strongly on some of the variables. If with such model we use numbers with other distributions, and with the same total number of points, we obtain significantly poorer resolution per variable. See [Sobol]. The Monte Carlo simulation is for computer implementation. In the metrology, in order to simulate values of component quantities, input values produce number generators for given best estimated component quantities and distributions of these evaluations. The simulated input values must have all parameters of a distribution equal to parameters of a distribution of simulated evaluation, including possible correlation. This simulation truly represents a measurement procedure. The simulation can be used to: ● determine a distribution of evaluations of measurand ● determine a confidence interval of evaluation ● determine the best estimated value of measurand ● determine a standard measurement uncertainty of this evaluation ● determine the degrees of freedom of this uncertainty ● determine sensitivity coefficients ● to validate procedures according to the GUM. 158 Kostić • 2019 • Metrology Companion The Monte Carlo simulation applies if the conditions to use the primary procedures according to the GUM are not fulfilled, or if it is unclear whether they are fulfilled. The simulation can also be applied if the condition to use the convolution is not fulfilled (the condition is uncorrelated input values). The simulation is commonly used in the following cases: ● linearization of a model of measurement is not satisfactory ● it is inconvenient to provide partial derivatives for a model of measurement ● a model of measurement is too complex ● distribution functions of input values are highly skewed ● standard measurement uncertainties of input values are not roughly equal ● a distribution of output values departs appreciably from the normal distribution ● an output value and its standard measurement uncertainty are roughly equal. Conditions for application of the Monte Carlo simulation: ● the possibility of computing a sufficiently large number of Monte Carlo trials ● in the neighborhood of the best estimated input quantities the model of measurement is continuous (not necessarily the derivatives of the model) ● if for output values an arithmetic mean and deviation exist ● if the confidence interval is to be determined: - the cumulative distribution function of output values is continuous and strictly increasing - the probability density function of output values is continuous over an interval for which it is larger than zero, it is unimodal and it is strictly increasing, or zero, to the left of the mode and strictly decreasing, or zero, to the right of the mode. The primary advantages of the Monte Carlo simulation: ● it is not necessary to determine partial derivatives of a model of measurement ● the simulation gives the distribution of output values ● the simulation has broader applicability than the procedures according to the GUM ● the same kind of simulation results and results of procedures according to the GUM. Kostić • 2019 • Metrology Companion 159 Estimate of a value of a quantity and its statistical parameters can be performed by the Monte Carlo simulation in the following steps. 1) Specify the input data: a model of measurement and component quantities with their parameters. 2) Determine the required number of Monte Carlo trials, that is, sets of input values, M. This number depends on the requested accuracy of the 4 simulation results and on distributions of input values. Often satisfies 10 6 to 10 trials. 3) Produce by number generators M sets of randomly ordered input values with appropriate distributions and possible correlations. See Example 6.42.1 and [GUM-S1]. 4) For each set of input values from 3), by the use of a model of measurement, compute the model output value, ym (m = 1, 2, ..., M ). 5) From the M model output values, the histogram is obtained in the usual way according to 6.38 Histogram. 6) The best estimated value of a quantity is obtained by: ● using the model of measurement and other input data - if the input values have symmetric distributions with approximately equal measurement uncertainties which are relatively small in relation to the means of input values, or ● if the input values do not have distributions as in the preceding clause - based on the distribution of the model output values, and according to 6.9 Best estimate. 7) When n results of repeated measurements of component quantities are simulated, the M output values are divided into J = M / n groups with the n values. 8) The maximum likelihood value of the standard deviation of the value obtained from the n results of repeated measurements is computed in the usual way from (6.56) so that the standard deviation of all M values is divided by n . This deviation is the combined standard measurement uncertainty of the mean of n results of repeated measurements. (The standard deviation of all M values is approximately equal to the average (that is, arithmetic mean) of the deviations of the n values within the groups.) 9) For each group with the n values, the standard deviation of the mean of the group is computed in the usual way from (6.56). See Table 6.4. 10) For all J groups, the deviation of the standard deviations of the mean of the group, from 9), is computed in the usual way from (6.49). See Table 6.4. 11) For all J groups, the standard deviation of the means of the group is computed in the usual way from (6.49). See Table 6.4. 160 Kostić • 2019 • Metrology Companion 12) The degrees of freedom of the standard deviation from 8) is approximately computed from (6.75), using the deviation of the standard deviations of the mean of the group from 10) and the standard deviation of the means of the group from 11). See Table 6.4. 13) In order to determine the needed number of the Monte Carlo trials, repeat, for example, 10 simulations with a different number of trials. Compute the standard deviation for the results of the repeated simulations. Based on these deviations estimate the required number of trials. The simulation on a personal computer often takes only a few seconds. Example In Figure 6.19 the instructive example of Monte Carlo simulation is represented. The simulation gives probability density function of evaluations of a measurand, f( y), for uncorrelated component quantities with the probability density functions f(x1) and f(x2) and for the model of measurement y = x1 + x2 . Based on the distribution f( y), it is determined the mode ymode = 0.96, which is the best estimated measurand, y, as well as the standard measurement uncertainty of this evaluation of y, u( y) = 2.149. The procedures according to the GUM gives the best estimated measurand y = x1 + x2 mode = 0.500 + 0.374 = = 0.874 and its standard measurement uncertainty u(y= ) u2 (x1 ) + u2 (x2 = ) = = 0.16672 + 2.1592 = 2.165. [Hristov, communication] [Sobol] [GUM-S1] Introduction, 1, 8.1.1, 5.11, 5.10, 5.9.6, 7 Kostić • 2019 • Metrology Companion Figure 6.19. The Monte Carlo simulation gives probability density function of estimates, f( y), for uncorrelated component quantities with the density functions f(x1) and f(x2). 161 7 Variances 7.1 Natural variance Natural variance (variance) V , of a group of m values xi (i = 1, 2, ..., m), is an arithmetic mean of squared differences of these values from an accurate arithmetic mean of these values, x : m V = ( xi − x )2 i =1 (7.1) . m The natural variance V , of a group of m values xi , can be computed also from the squared differences of these values from pairs obtained by combinations (without repetition) of second order from m elements: m −1 m V = ( xi − x j )2 i =1 j = i +1 (7.2) . m2 Population variance (variance of a population, or expected 2 variance) σ , of all m values in a population, xi (i = 1, 2, ..., m), is computed from (7.1) by taking for the accurate arithmetic mean of these values, x , the accurate arithmetic mean of a population, μ: m σ2 = ( xi − μ)2 i =1 (7.3) . m The variance Vm , of values xi which arrive continuously, is computed by the recursive method using (7.4) and (7.5), derived from (6.19) and (7.1). This computation does not require memorizing all m received values, but only the last xm , and the previously computed arithmetic mean x m − 1 and the variance Vm − 1 . x m = (m − 1) ⋅ x m − 1 + xm m ⋅ m −1 m −1 Vm = ⋅ Vm − 1 + ⋅ ( xm − x m − 1 )2 m m2 Kostić • 2019 • Metrology Companion (7.4) (7.5) 171 Computation using (7.5) in practice gives more accurate results than (7.1). Using (7.5) is especially appropriate in the following cases: ● when the values for which the variance is computed arrive continuously; ● when the values have a small relative deviation; ● if it is required to determine the variance during a process (for example during Monte Carlo simulation, in which trials are stopped when the variance of the simulation results becomes small enough). The degrees of freedom of the variance is equal to the number of values: ν = m. The variance is an indicator of dispersion of values in a group. It is expressed in units which are the squared unit of these values. The variance is valid for values with any distribution. See Example 6.42.2. The variance of a group of values is equal to the squared deviation of this group. The properties of the variance V are listed in the following text. Variables are denoted by x and y, constants by C and D, function (7.1) by a ), and the standard covariance of x and y, given by (7.21), by V(x, y). V( ● V ≥ 0. (7.6) = 0. ● V(C) (7.7) + x ) = V( x ). ● V(C (7.8) ⋅ x ) = C2 ⋅ V( x ). ● V(C (7.9) ● If x and y are uncorrelated: ⋅ x + D ⋅ y ) = C2 ⋅ V( x ) + D2 ⋅ V( y ). V(C (7.10) ● If x and y are correlated: x ) + D2 ⋅ V( y ) + 2 ⋅ C ⋅ D ⋅ V( x, y ). V (C ⋅ x + D ⋅ y ) = C2 ⋅ V( (7.11) ● If x and y are uncorrelated: x ⋅ y ) ≈ y 2 ⋅ V( x ) + x 2 ⋅ V( y ). V( (7.12) [GUM] C.2.20, C.3.2, C.2.12 [Box 2005] PP 27 [Hristov, communication] [Welford] [Ivković] PP 45 172 Kostić • 2019 • Metrology Companion 7.6 Combined variance Combined variance is a variance of estimated value arising from variances of component variables. The computed combined variance of the evaluation is valid for any distribution of the component variables and the evaluation. Moreover, evaluations approach the normal distribution as the number of component variables increases, but if the variables are independent and if there is no a variable with multiple times greater variance than other variances (central limit theorem). See Example 6.42.2. Let the relation between the evaluation y and J component variables xj be given by the function y = y (x1 , x2 , ..., xJ ), let the component variables be uncorrelated and let the nonlinearity of the function y be insignificant. Then the combined variance of the evaluation y, VC ( y), is computed from function (7.40). The partial derivative ∂ y / ∂ x j is called the sensitivity coefficient to x j and is usually denoted by c (x j ) or cj . The variance of the component variable xj is denoted by V(x j). 2 2 2 ∂y ∂y ∂y VC ( y ) = ⋅ V( xJ ) = ⋅ V( x1 ) + ⋅ V( x2 ) + ... + ∂x1 ∂x2 ∂xJ ∂y = j =1 ∂x j J = 2 ⋅ V( x j ) = J ( c2 ( x j ) ⋅ V( x j )) (7.40) j =1 Function (7.41) gives the variance Vmax (VCy ), which is the maximum possible variance of variance VC ( y ) from (7.40). If the maximum variance of variance, Vmax (VCy ), is several times smaller than the variance VC ( y ), the nonlinearity of the function y is insignificant considering the variances of its component variables, so (7.40) gives a valid value. Partial derivatives of (7.41) are at the changing point ( x1 + r ⋅ V( x1 ) , x2 + r ⋅ V( x2 ) , ..., xJ + r ⋅ V( xJ ) ) which changes by varying value of r from 0 to 1. See Examples 7.6.1 and 7.6.2. Vmax (VCy ) = max ( 0 < r <1 1 J ∂ ∂y ⋅ ( ) ⋅ V( x j ) ) 2 j =1 ∂x j ∂x j Kostić • 2019 • Metrology Companion (7.41) 183 The function y is a linear function if it can be expressed using the constants Cj : y (x1 , x2 , ..., xJ ) = C1 · x1 + C2 · x2 + ... + CJ · xJ . (7.42) If the function y is linear and if components xj are uncorrelated, function (7.40) gives not an approximate, but the exact value of the combined variance. In this case, checking by function (7.41) is not required. From (7.40) it follows that the combined variance V, of the sum or difference of uncorrelated variables, is equal to the sum of the variances of these variables, Vj : V = V1 + V2 + ... + VJ . (7.43) Example. The voltage U1 is measured by subtracting the measured voltage U3 from the measured voltage U2 : U1 = U2 − U3 . The correlation between U2 and U3 is negligible. Compute the combined variance of the evaluation U1, VC (U1). U1 = U2 − U 3 2 2 ∂U1 ∂U1 VC (U1 ) = ⋅ V(U 3 ) ⋅ V(U2 ) + ∂U2 ∂U 3 ∂ U1 ∂ U1 = 1, = −1 ∂ U2 ∂ U3 VC (U1 ) = V(U2 ) + V(U 3 ) The combined variance of an evaluation from uncorrelated component values can be computed in the following steps. Examples are given in Sections 7.6.1, 7.6.2, 7.6.3 and 7.6.4. 1) Write a function which gives an estimated value from component variables. 2) For each of the component variables of the function from 1) find the first and second partial derivative. 3) Compute the combined variance of the estimated value using (7.40). 4) Estimate the maximum possible variance of variance from 3) using (7.41). 5) The combined variance from 3) is assumed as valid, considering the nonlinearity of the function from 1), if its maximum possible variance from 4) is eight or more times smaller than it. 184 Kostić • 2019 • Metrology Companion The variance of variance computed in 4) arises from an error of linear approximation of a nonlinear function. The eight times larger variance than the variance of this variance corresponds to the variance of variance of an arithmetic mean of five values with the T-distribution. The relation from 5) can be changed according to specific applications. See 6.22 Deviation of the standard deviation. If the variance of variance is too large for a specific application and if the component variables are uncorrelated, estimate the combined variance by Monte Carlo simulation, or using (7.44) with a validation check. In practice, there are rare cases of the too large variance of variance, that is, with significant nonlinearity of the function. If the component variables are correlated and if nonlinearity of the function from 1) is not significant, use Monte Carlo simulation, or (7.49), (7.50) and a procedure like the one above. In the case with uncorrelated component variables and with significantly nonlinear function y, the combined variance of the evaluation, VC ( y ), is computed from the function (7.44). 2 J J 2 2 J ∂y ∂y ∂y ∂3y 1 + ⋅ ⋅ ⋅ VC(y ) = ⋅ V(xj ) + ⋅ V( x ) V( x ) j k ∂xj ∂xj ∂xk ∂xj ∂x ∂x 2 2 j =1 j k j =1 k =1 (7.44) Partial derivatives of higher order are computed as follows. ∂ ∂y (y ) = ∂x j ∂x j ∂2 y ∂x j 2 = ∂ ∂y ( ) ∂x j ∂x j ∂2 y ∂ ∂y ∂ ∂y ( )= ( ) = ∂x j ∂xk ∂x j ∂xk ∂xk ∂x j ∂3 y ∂x j ∂xk 2 = ∂ ∂ ∂y ( ( )) ∂x j ∂xk ∂xk (7.45) (7.46) (7.47) (7.48) The accuracy of the variance given by (7.44) can be validated by Monte Carlo simulation. Kostić • 2019 • Metrology Companion 185 In the case with correlated component variables and if nonlinearity of the function y is insignificant, the combined variance of the evaluation, VC ( y ), is computed from function (7.49). Standard covariances for the variables x j and x k are denoted by V (x j , x k ). See 7.4 Experimental standard covariance, as well as Examples 8.3.1, 9.12.3 and 9.12.4. ∂y VC ( y ) = j =1 ∂x j J J ( 2 J −1 ⋅ V( x j ) + 2 ⋅ j =1 J −1 ) = c 2 ( x j ) ⋅ V( x j ) + 2 ⋅ j =1 J ∂y ∂y ∂x ⋅ ∂x ⋅ V( x j , xk ) = j k k = j +1 J ( c( x j ) ⋅ c( xk ) ⋅ V( x j , xk )) j =1 k = j +1 (7.49) The covariances V (x j , x k ) in (7.49) are for all pairs of J component variables, without considering the order of the variables and excluding pairs with two equal variables. The pairs are obtained by combinations (without repetition) of second order from J elements, see 7.4 Experimental standard covariance. Function (7.50) gives variance Vmax (VCy ), which is the maximum possible variance of variance VC ( y ) from (7.49). If the variance of variance, Vmax (VCy ), is several times smaller than the variance VC ( y ), the nonlinearity of the function y is insignificant and (7.49) gives a valid value. Partial derivatives of (7.50) are at the changing point ( x1 + r ⋅ V( x1 ) , x2 + r ⋅ V( x2 ) , ..., xJ + r ⋅ V( xJ ) ) which changes by varying the values of r from 0 to 1. Vmax (VCy ) = max ( 0< r <1 1 J ∂ ∂y J−1 J ∂ ∂y ⋅ ( ) ⋅ V(xj ) + ( ) ⋅ V(xj , xk ) ) (7.50) 2 j =1 ∂xj ∂xj x x ∂ ∂ j =1 k= j +1 j k Example. Let the evaluation y is obtained from three correlated component variables a, b and c, then the combined variance of y, VC (y ), is computed from (7.49) as follows. y = y (a, b, c) 2 2 2 ∂y ∂y ∂y VC ( y ) = ⋅ V(a ) + ⋅ V(b ) + ⋅ V(c ) + ∂a ∂b ∂c ∂y ∂y ∂y ∂y ∂y ∂y + 2⋅ ⋅ ⋅ V(a, b ) + ⋅ ⋅ V(a, c ) + ⋅ ⋅ V(b, c ) ∂a ∂c ∂b ∂c ∂a ∂b The rule (or law) of propagation of deviation, and the rule (or law) of propagation of measurement uncertainty are expressed by function (7.49). 186 Kostić • 2019 • Metrology Companion 8 Tests and functions fitting 8.1 Detection of outliers Results of statistical analyses are valid only if the analyzed values are without outliers. Therefore, if there is a significant possibility that outliers will occur, their detection and rejection should be a mandatory part of the statistical analysis. A good practice is to consider as outliers and reject the values which differ from an arithmetic mean of a sample, x , more than the triple standard deviation of the sample, s. According to this practice, detection of outliers should be carried out by firstly rejecting the values which are supposed to be outliers. It is followed by computing the mean x and the standard deviation s. Finally, it should be determined whether the supposed outliers are out of the interval x ± 3 · s, and if they are, the supposition is true, so x and s can be assumed as valid. If among the supposed outliers they are certain values which are not out of the interval, the procedure should be repeated without rejection of these values. In the case of values with the normal distribution, such practice is justified by the fact that the probability that a valid value is out of x ± 3 · s is equal to 0.27 %. ● In the case of 30 values with the T-distribution, the probability is 0.55 %; ● for 10 values, it is 1.5 %; ● for 5 values, it is 4.0 %. The text continues on the next page. Kostić • 2019 • Metrology Companion 203 If the previous manner of detection of outliers is not sufficiently reliable, the following procedure is recommended for a sample from a population with the normal distribution. This test determines, with a requested probability, whether a value with the largest difference from the arithmetic mean is an outlier. The test is performed in the following steps. 1) Check if the sample is from a population with the normal distribution. Checking can be done according to the instructions in 8.6 Estimating the type and parameters of a distribution. If the sample is from the normal population, the test can be applied. 2) Let n be the number of values in the sample, including the tested extreme value. Let P be the requested probability of the test result. Determine a coverage factor k (ν, P) for the T-distribution with the parameters: a degrees of freedom, ν, and the probability P. This factor can be obtained from Table 6.3. The parameter ν is given by (8.1), and for ν > 100, it can be assumed that ν = ∞. 3) Determine the truth (that is, validity) of the inequality (8.2). The extreme value is denoted by xextreme. An arithmetic mean of the sample without the extreme value is x ', and the standard deviation of the mean is s x '. If the inequality is true, with the probability P or larger, xextreme is an outlier. ν = n−2 xextreme − x ' > k ( ν, P ) ⋅ s x ' (8.1) (8.2) 4) The extreme value is rejected if it is determined to be an outlier. The test is then repeated for each of the following extreme values, until finding a value which is not an outlier. 204 Kostić • 2019 • Metrology Companion 8.1.1 Example The following values were obtained as the results of measurements of pH, and then ordered in the ascending array: 3.75; 3.86; 4.19; 4.19; 4.24; 4.37; 4.37; 4.50; 4.53; 4.53; 4.55; 4.58; 4.76; 4.77 and 5.23. Determine whether the value of 5.23 is an outlier, which can be supposed on the basis of the histogram in Figure 8.1. Solution 1) The population of the results is with the normal distribution because errors of the results are a consequence of a large number of independent individual errors with a small share. (See Section 6.12.) 2) The number of the results is n = 15. From (8.1) we obtain that ν = n − 2 = 15 − 2 = 13. The requested probability of the test is P = 0.95. From Table 6.3 we determine the coverage factor k (ν, P ) = k (13, 0.95) = 2.16. 3) The tested extreme value is xextreme = 5.23. The arithmetic mean of the array without xextreme is x ' = 4.371. The deviation of the arithmetic mean is s x ' = 0.302. Determine the truth of the inequality (8.2): xextreme − x ' = 5.23 − 4.371 = 0.859, k ( ν, P ) ⋅ s x ' = 2.16 ⋅ 0.302 = 0.652, since 0.859 > 0.652, the inequality (8.2) is true. It can be concluded that with a probability larger than 0.95, the result 5.23 is an outlier. Figure 8.1. The histogram of the measurement results of pH. Kostić • 2019 • Metrology Companion 205 8.2 Fitting Fitting is defining a function, or a particular value, on the basis of a sample, so that it meets specific requirements of the goodness of fit to the sample. Fitting is most often performed by a variant of the method of least squares. See Section 8.3 and Example 9.12.2 (from the GUM). Regression analysis (that is, analysis of differences) is used to find a function which describes a relationship between correlated values. Parameters of this regression function (or smoothing function, or regression model) are determined by fitting. The regression function is used to estimate or predict values. Linear regression function is a linear function which describes a relationship between correlated values. [Ivković] PP 49, PP 50 [Njegić] PP 182 8.3 Method of the least sum of squares Method of least squares, LSM, is a procedure of fitting of a function, or a particular value, so that the sum of squares of differences between the determined values and input values in a sample is minimized. The method of least squares gives the maximum likelihood values when input values are with a symmetrically decreasing distribution. An example of fitting of a particular value by the method of least squares is computation of the arithmetic mean. The example of fitting of a function by the method of least squares is given below. If we have n pairs of correlated values (xi , yi ), i = 1, 2, ..., n, their mutual relationship can be described by a function, more or less accurately. A linear function is commonly used: y'(x) = a + b · x . (8.3) The parameters a and b of the linear function can be determined by the method of least squares, that is, by minimizing the sum S which is given by formula (8.4). S is the sum of squares of differences between the values given by the function y'(xi ) and the input values yi . 206 Kostić • 2019 • Metrology Companion S= n n i =1 i =1 [ y'( xi ) − y i ] 2 = [(a + b ⋅ xi ) − y i ] 2 (8.4) The sum S given by (8.4) is minimized when the system of equations given by (8.5) is fulfilled. Therefore, the parameters of the linear function are determined by solving the system for a and b. The system (8.5) always has a unique solution. n ∂Z = ∂a {2 ⋅ [(a + b ⋅ xi ) − y i ]} =0 i =1 n ∂Z = ∂b {2 ⋅ [(a + b ⋅ xi ) − y i ] ⋅ xi } = 0 (8.5) i =1 The system (8.5) can be solved by using the following quantities. n xi x = i =1 n the arithmetic mean of xi (8.6) the arithmetic mean of yi (8.7) n yi y = i =1 n n x2 = xi 2 i =1 n (8.8) n ( xi ⋅ y i ) x⋅y = i =1 n Using these quantities, the system (8.5) can be rewritten as: (8.9) y = a + b ⋅ x (8.10) 2 x ⋅ y = a ⋅ x + b ⋅ x . The parameters a and b are obtained by solving the system (8.10): x ⋅y − x ⋅y b= x2 − x2 (8.11) a = y − b⋅x . See Example 8.3.1 and Example 9.12.2 (from the GUM). In both examples, the input values are equivalent and the solutions are the same. Kostić • 2019 • Metrology Companion 207 The following model describes prediction of the value y by using the linear function y'(x): y(x) = y'(x) + ε(x) = a + b · x + ε(x). (8.12) The summand ε(x) represents the differences between the input values and the values predicted by the function y'(x), that is, negative of an error of prediction by the function y'(x). If there are n input values and assuming that the differences ε(x) have the normal distribution, the standard variance of the differences, Vε, is given by (8.13). The degrees of freedom of the variance, νVε , is given by (8.14). n (a + b ⋅ xi − y i ) 2 Vε = i =1 n−2 νVε = n − 2 (8.13) (8.14) The parameters a and b have the standard variances Va and Vb given by (8.15) and (8.16). Their standard covariance, V (a, b), is given by (8.17). Vε Va = (8.15) x2 ) n ⋅ (1 − x2 Vb = Vε n ⋅ ( x 2 − x 2) V(a, b ) = − x ⋅ Vε n ⋅ ( x 2 − x 2) (8.16) (8.17) A value predicted by the linear function y'(x) has the standard variance Vy' given by function (8.18) and derived in the normal manner explained in 7.6 Combined variance. The degrees of freedom of the variance, νVy' is given by (8.19). Vy' ( x ) = Va + x 2 ⋅ Vb + 2 ⋅ x ⋅ V(a, b ) (8.18) νVy' = n − 2 (8.19) For a given x, it is possible to determine the interval which has a requested probability P that within it be a value from the population of input values. Such probability and interval are the level and interval of confidence in the function of x. For the requested probability P, and for n input values, assuming that ν = n − 2 , from Table 6.3 for the T-distribution, we obtain the 208 Kostić • 2019 • Metrology Companion corresponding coverage factor k. A half-width of the confidence interval, ∆(x), is given by (8.20), while the confidence interval is given by (8.21). ∆ (x) = k ⋅ Vy' (x ) (8.20) [ y'(x) − ∆(x), y'(x) + ∆(x) ] (8.21) [GUM] H.3.4 [Panchenko] 29.1, 29.2, 30.1 [Ivković] PP 49, PP 50, II.6 [Njegić] PP 64 [Pantić] PP 134 [Anscombe] [Britannica] Statistics [Perović, communication] 8.3.1 Example Using a linear function to approximate the calibration function of a measuring device. Then, for the correction addend predicted by this calibration function, estimate the standard deviation and the intervals of confidence. From Table 8.1 use the reference values of the standard and the indications of the measuring device. The reference values of the standard have negligible errors. Neglect the bias of the standard deviations because the number of input pairs is large. The experimental correction addends have the normal distribution, so the method of least squares can be used. Solution Using the values xi and yi from Table 8.1 as the input pairs, the requested parameters is computed by using the method of least squares. n = 11 from Table 8.1, the number of the input pairs (xi , yi ) n xi x= i =1 = 4.008 n from (8.6), the arithmetic mean of xi n yi y= i =1 = − 0.1625 n from (8.7), the arithmetic mean of yi n x2 = xi 2 i =1 = 18.56 n from (8.8) n ( xi ⋅ y i ) x⋅y = i =1 n = − 0.6458 from (8.9) Kostić • 2019 • Metrology Companion 209 b= x⋅y − x ⋅y x2 − x2 = 0.002183 from (8.11), the parameter of function from (8.11), the parameter of function a = y − b ⋅ x = −0.1712 y'(x ) = a + b ⋅ x = −0.1712 + 0.002183 ⋅ x the calibration function n (a + b ⋅ x i − y i ) 2 i =1 = 1,223 ⋅ 10 −5 n−2 from (8.13), the standard variance for ε(x) Vε = Vε Va = x2 n ⋅ (1 − Vb = x2 = 8,281⋅ 10 −6 ) Vε n ⋅( x 2 2 −x ) V (a, b ) = − from (8.15), the standard variance for a = 4,461⋅ 10 −7 x ⋅ Vε 2 2 from (8.16), the standard variance for b = −1,788 ⋅ 10 −6 n ⋅( x − x ) from (8.17), the standard covariance of a and b Vy' ( x ) = Va + x 2 ⋅Vb + 2 ⋅ x ⋅ V(a, b) = = 8,281 ⋅10 −6 + x 2 ⋅ 4,461 ⋅ 10−7 + 2 ⋅ x ⋅ ( −1,7883 ⋅ 10−6 ) from (8.18), the standard variance for the predicted correction addend y'(x) sy' ( x ) = Vy' ( x ) the standard deviation for the predicted correction addend νsy ' = νVy ' = n − 2 = 9 the degrees of freedom of the variance sy' (x) ∆ ( x ) = k ⋅ s y' ( x ) from (8.20), the half-width of the confidence interval for the requested probability P. [GUM] H.3 210 Kostić • 2019 • Metrology Companion Table 8.1 With Example 8.3.1: values for obtaining the calibration function Reading number i Input values 1 2 3 4 5 6 7 8 9 10 11 Reference value of standard Ri 1.350 1.843 2.346 2.844 3.343 3.834 4.357 4.845 5.344 5.849 6.351 Indication of measuring device xi 1.521 2.012 2.512 3.003 3.507 3.999 4.513 5.002 5.503 6.010 6.511 Experimental correction addend yi = Ri − xi −0.171 −0.169 −0.166 −0.159 −0.164 −0.165 −0.156 −0.157 −0.159 −0.161 −0.160 Prediction of obtained function Error of Predicted predicted correction correction addend addend y'(xi ) −εi = y'(xi ) − yi −0.16788 −0.16681 −0.16572 −0.16465 −0.16355 −0.16248 −0.16135 −0.16029 −0.15919 −0.15809 −0.15699 0.00312 0.00219 0.00028 −0.00565 0.00045 0.00252 −0.00535 −0.00329 −0.00019 0.00291 0.00301 Figure 8.2. With Example 8.3.1: input pairs (xi , yi ) and the calibration function, y'(x), with the intervals of confidence for different confidence levels, P. Kostić • 2019 • Metrology Companion 211 8.4 P-value Goodness-of-fit evaluation is an indicator of mutual closeness of functions or values. Let the unfit evaluation of a sample and an expected function, T, gives a function with the distribution t(x). Then the P-value of the goodness of fit of the sample and the expected function, q, is given by the following expression (8.22) and represented in the following diagram. For an example, see 8.7 Pearson chi-square test for the goodness of fit of distributions. q= ∞ T t (x) ⋅ d x = 1 − T 0 t (x) ⋅ d x (8.22) As the unfit evaluation T decreases, the corresponding P-value increases, so the P-value is an indicator of the probability of the fit of the sample and the expected function (a larger P-value corresponds to a greater fit). The P-value is an indicator of the probability of a fit, but it is neither the probability of a fit, nor the probability that a hypothesis is true, nor the probability of wrongly rejecting a hypothesis. Also, the probability of a fit is not proportional to the P-value (except in special cases). Contrary to the statement in the second paragraph, in some cases the fit evaluation is taken for the evaluation T, then the P-value is an indicator of the probability of an unfit (a larger P-value corresponds to a greater unfit). Computation of the P-value is schematized and relatively simple. Therefore, the P-value is used in a wide variety of goodness-of-fit tests. Ronald Fisher advocated using the P-value, and it is widely accepted in many sciences. However, it is very often misunderstood and wrongly used. [Berger] [Sellke] 212 Kostić • 2019 • Metrology Companion 9 Measurement uncertainty 9.1 Measurement result and its uncertainty Measurement uncertainty (uncertainty, or uncertainty of measurement) is a parameter of measurement result that describes the accuracy of the result by an indicator of dispersion of values that can be reasonably taken as the measurement result. A measurement result can be used to compute the level and interval of confidence, if the result is the best estimated value of a measurand, and if that result is associated with the measurement uncertainty which is the standard deviation of that best estimated value, as well as the degrees of freedom of that deviation. The best estimated value of a measurand, the measurement uncertainty of the evaluation, and the degrees of freedom of this uncertainty, can be used as component values in evaluating the uncertainty and degrees of freedom of another value in which the first value is used. This allows traceability. In order to establish uniform estimating and expressing of measurement results, in 1993 ISO published the first edition of Guide to the expression of uncertainty in measurement, GUM. The Guide is applicable at various levels of accuracy, from a shop floor to the fundamental sciences. The basis of the Guide is the internationally adopted Recommendation of CIPM (from 1981) which accepts the Recommendation of BIPM Working Group on the Statement of Uncertainties (from 1980). Today, publishing and correction of the Guide are within the competence of the JCGM. The JCGM member organizations are: BIPM, IEC, IFCC, ILAC, ISO, IUPAC, IUPAP and OIML. The application of measurement uncertainty and traceability are mandatory for accredited laboratories, as well as organizations that have the quality management system according to ISO 9001:2008. Kostić • 2019 • Metrology Companion 239 The summarized overview of the procedures given in the GUM for estimating of a measurement result and its measurement uncertainty is as follows. ● A measurement result is obtained based on the results of repeated measurements by determining the best estimated value of a measurand. ● For this best estimated value, the standard deviation, called the standard measurement uncertainty, is determined. ● For this standard uncertainty, the degrees of freedom is determined. These evaluations can be obtained by using classical statistical methods (including convolution) which are regarded as the primary methods, or where appropriate, by using Monte Carlo simulation. Standard measurement uncertainty (standard uncertainty) is the standard deviation of a measurement result which is the best estimated value of a measurand. Depending on the method for estimating the uncertainties, they are grouped into two categories: ● the standard uncertainties which are computed by statistical analysis of results of repeated measurements, type A; and ● the standard uncertainties which are estimated by statistical analysis of reliable information, type B. The squared standard uncertainty, that is, the standard variance, is the unbiased evaluation of the squared deviation of a population of measurement results, that is, the variance of a population of these results. See 7.2 Experimental standard variance. The standard deviation is the biased evaluation of the deviation of a population of measurement results. The bias of the uncertainty increases as its degrees of freedom decreases. In the case of the normal distribution, the relative error of the evaluation is in the interval from 0 % to +8.5 % for the degrees of freedom from ∞ to 3. See 6.20 Experimental standard deviation. When the standard uncertainty with a small degrees of freedom is taken as the evaluation of the deviation of the population of measurement results, the bias of this uncertainty should be corrected. The standard uncertainty of the best estimated value of a measurand arises only from random effects and uncertainties of corrections. Component standard uncertainty (or standard uncertainty component, or component of standard uncertainty) is a standard uncertainty of a component value. [GUM] 2.2.3, 0, PP v, Foreword, Annex A, 3.2, 4.1.4, 0.7 2) 3) 4), 2.3.1, 4.2.3, 4.2.6, 7.2.1 d), 4.1.5, 4.1.6, 2.3.2, 2.3.3, 8, C.2.21 [VIM 2] 3.9 [VIM 3] 2.26 [IEC 17025] 5.4.6.1, 5.4.6.2, 5.10.4.1 [ILAC] 4.8 [ILAC, OIML] 1., 2. [NVLAP] 5.4.6, Annex B.1 [ISO 10012] 7.3 [ISO 9001] 7.6 [GUM-S1] Introduction, 1 240 Kostić • 2019 • Metrology Companion 9.2 Type A standard measurement uncertainty Type A standard measurement uncertainty (type A standard uncertainty, or type A standard uncertainty of the result of a measurement) is the standard deviation, of the best estimated value of a measurand, computed by statistical analysis of results of repeated measurements. For this uncertainty, we should always also give its degrees of freedom. Unless otherwise indicated, it is understood that the best estimated value of a measurand is the maximum likelihood value of the measurand and that this evaluation has the normal distribution. Type A standard uncertainty is denoted by uA , and its degrees of freedom by νA . The maximum likelihood value of a measurand is most often computed as the experimental arithmetic mean of results of repeated measurements. This mean must be corrected for all significant systematic errors. For this corrected mean, the deviation of the arithmetic mean, sx , is computed according to 6.21 Deviation of the arithmetic mean, and then it is taken as the type A standard uncertainty, uA : u A = sx . (9.1) The degrees of freedom of the type A standard uncertainty is computed according to 6.24 Degrees of freedom. When from n results of repeated measurements we compute the experimental arithmetic mean which is then used to compute the deviation of the arithmetic mean, the degrees of freedom of this deviation, as well as, of this uncertainty, is: νA = n – 1. (9.2) See Section 6.24.1, example d). [GUM] A.2, 0.7, 8, 3.1.2, 3.2.4, 4.2.1, 4.2.3, 4.2.6, 3.2.3, 3.4.4, 3.3.5 Kostić • 2019 • Metrology Companion 241 9.3 Type B standard measurement uncertainty Type B standard measurement uncertainty (type B standard uncertainty) is the standard deviation, of the best estimated value of a measurand, estimated by statistical analysis of reliable information. For this uncertainty, we should also give its degrees of freedom if it is deemed useful. Unless otherwise indicated, it is understood that the best estimated value of a measurand is the maximum likelihood value of the measurand and that this evaluation has the normal distribution. Type B standard uncertainty is denoted by uB , and its degrees of freedom by νB . The type B standard uncertainty is estimated if there is not enough measurement data to compute the type A standard uncertainty. The best estimated value of a measurand for which we give the type B standard uncertainty is often obtained by taking a single measurement result or by estimate on the basis of a small amount of information. The type B standard uncertainty is estimated on the basis of an approximated probability density function obtained by analysis of the probability that an event will occur. The evaluations of distribution, deviation, and degrees of freedom of deviation, must be estimated by scientific judgment based on available information. Among other means, information can be obtained from the following sources: ● previous measurement results ● experience or general knowledge of the properties of materials and devices for measurement ● specifications provided by manufacturers of materials and devices for measurement ● calibration reports or other certificates ● handbooks with measurement uncertainties assigned to the given values. Estimate of the type B standard uncertainty needs to be comprehensive so that the obtained evaluation is approximately as accurate as the type A standard uncertainty. This aim is easily achieved when the type A standard uncertainty is estimated for an evaluation obtained from a small number of averaged results. Section 6.22 Deviation of the standard deviation, Table 6.1, shows that the deviation of the standard deviation of an arithmetic mean is not negligible in practical cases. 242 Kostić • 2019 • Metrology Companion The degrees of freedom of the standard uncertainty of type B can be computed according to: ● 6.24 Degrees of freedom; ● 6.25 Effective degrees of freedom; ● 6.28 Standard deviation and its degrees of freedom from the level and interval of confidence; or ● 6.26 Degrees of freedom from the deviation of the standard deviation of the arithmetic mean. If it is considered that the type B standard uncertainty is absolutely reliable, we assign to it the infinite degrees of freedom. See the following Examples 9.3.1 and 9.3.2. [GUM] 0.7, 8, 3.1.2, 3.2.4, 4.3, 7.2.1 d), 7.2.7 c), G.4.2, 3.3.5, 4.3.2, E.4.3, G.4.2, G.6.4 9.3.1 Example An ambient temperature is measured by a digital thermometer placed in an appropriate position. The thermometer is from a reliable manufacturer, with the resolution of 0.1 ºC. However, information about its accuracy is not available. The ambient temperature is achieved naturally (without using air conditioning). For the two cases given below, estimate the environment temperature, its standard uncertainty and the degrees of freedom of the uncertainty. a) The thermometer constantly gives 25.0 ºC. b) The thermometer alternatively gives 25.0 ºC and 25.1 ºC, in approximately equal durations. Solutions a) The evaluation of the environment temperature ta is given by: ta = thermometer_indication − systematic_error − absolute_error_of_resolution. The furthermore evaluations are as follows. The systematic error has the normal distribution and its maximum likelihood arithmetic mean is equal to zero. The absolute error due to resolution has the uniform distribution and has the maximum likelihood value equal to zero. According to 6.12 Normal distribution, the resulting distribution of the estimated temperature is approximately normal. Therefore, we conclude that the maximum likelihood environment temperature is: ta = 25.0 − 0 − 0 = 25.0 ºC. Low-accuracy thermometers normally have a maximum systematic error equal to their resolution. This maximum error probably matches the double standard deviation, so the estimated standard deviation of the systematic error is: s1 = 0.1 ºC / 2 = 0.050 ºC. Kostić • 2019 • Metrology Companion 243 The standard deviation of the error due to resolution is computed from (6.95): s2 = 0.1 °C / 2 3 = 0.029 °C. The standard uncertainty of the estimated environment temperature is computed from (6.41): uB a = s a = s 12 + s 2 2 = 0.050 2 + 0.029 2 = 0.058 °C. The thermometer constantly indicates 25.0 ºC, and this is equivalent to an infinite number of results, n, from which the evaluations are obtained. The degrees of freedom of the uncertainty uB a is computed from (6.64): νB a = n − 1 = ∞ − 1 = ∞. b) The estimated environment temperature tb , is given by: tb = thermometer_indication − systematic_error − absolute_error_of_evaluation_of_indication. The furthermore evaluations are as follows. The systematic error and the absolute error of evaluation of indication have the normal distributions, and the arithmetic means are equal to zero. According to 6.12 Normal distribution, the resulting distribution of the estimated temperature is approximately normal. Considering the principle of operation of the analog-to-digital converter, the thermometer indication is the arithmetic mean of the two values which the thermometer alternatively gives: 25.0 °C + 25.1 °C = 25.05 °C. 2 So the maximum likelihood environment temperature is: thermometer_indication = tb = 25.05 − 0 − 0 = 25.05 ºC. (Like the case a.) Low-accuracy thermometers normally have a magnitude of the maximum systematic error equal to their resolution. This maximum error probably matches the double standard deviation, so the estimated standard deviation of the systematic error is: s3 = 0.1 ºC / 2 = 0.050 ºC. The thermometer alternatively gives two values that are seemingly of equal durations, probably until the durations start to differ more than 244 Kostić • 2019 • Metrology Companion 20 % (0.2). Therefore, the absolute error of evaluation of indication has the approximate uniform distribution with the half-width equal to (0.2 / 2) · resolution = 0.01 ºC, so the standard deviation of this error is computed from (6.95): s4 = 0.01 °C 3 = 0.0058 °C. The standard uncertainty of the estimated environment temperature is computed from (6.41): uB b = s b = s 32 + s 42 = 0.050 2 + 0.0058 2 = 0.050 °C. The thermometer constantly indicates 25.05 ºC, and this is equivalent to an infinite number of results, n, from which the evaluations are obtained. The degrees of freedom of the uncertainty uB b is computed from (6.64): νB b = n − 1 = ∞ − 1 = ∞. 9.3.2 Example It is estimated that out of 12 measurement results of length of one-type elements, half of the results (which are not individually known) have values in the interval from 10.07 mm to 10.15 mm. The results are from a population with the normal distribution (because errors of the results are a consequence of a large number of independent individual errors with a small share). The evaluation of the interval has a negligible systematic error. Estimate: 1) the most frequent length of the elements, 2) the standard uncertainty of the evaluation of this length and 3) the degrees of freedom of this uncertainty. Solution 1) The most frequent length is the mode. The estimated confidence limits are a = 10.07 mm and b = 10.15 mm. The results of measurement of the length have the T-distribution, so the value of the mode, l, is in the midpoint of the interval from a to b: l= a+b 10.07 mm + 10.15 mm = = 10.11 mm. 2 2 3) The estimate of the standard uncertainty is carried out from total n = 12 results. Based on this, we compute the degrees of freedom of this uncertainty from (6.80): νB = n − 1 = 12 − 1 = 11. Kostić • 2019 • Metrology Companion 245 9.5 Expanded standard measurement uncertainty Expanded standard measurement uncertainty (expanded standard uncertainty, expanded uncertainty) U is the combined standard uncertainty uC multiplied by a coverage factor k : U = k · uC . (9.3) The combined standard uncertainty is multiplied by the coverage factor in order to get the interval given by (9.4) in which the value of a measurand is with the required probability. MEASUREMENT_RESULT ± U (9.4) The interval is the confidence interval, and the probability is the confidence level. See 6.27 Confidence level and confidence interval. Typically, the coverage factor will be in the interval from 2 to 3, which in the case of the normal distribution gives the confidence level from 95.5 % to 99.7 %. See Table 6.2. In the case of the T-distribution, the coverage factor for the required confidence level can be determined from Table 6.3, which gives the relations among the confidence level, degrees of freedom, and coverage factor. If the degrees of freedom is not an integer, the recommendation of [GUM] is to truncate its decimals. In the case of an asymmetric distribution, in order to determine the coverage factor, we can use that real distribution to obtain an accurate confidence interval with the requested confidence level. Then, the confidence interval can be defined using two coverage factors, k1 and k2 : [MEASUREMENT_RESULT – k1 · uC , MEASUREMENT_RESULT + k2 · uC ]. (9.5) However, based on the central limit theorem, the asymmetric distribution can often be approximated by the normal distribution with the arithmetic mean and deviation equal to the maximum likelihood value and deviation of the real distribution. The expanded uncertainty is given in some commercial, industrial and legal documents. In this case, a report of a measurement result gives the expanded standard uncertainty, confidence level and coverage factor, and can also give a type of distribution of the results. [GUM] 6, 0.7 5), G.1.4, G.6 248 Kostić • 2019 • Metrology Companion 9.6 Relative standard measurement uncertainties Relative standard measurement uncertainty (relative standard uncertainty) uR( y) is equal to the quotient of the standard uncertainty of a measurement result, u( y), and the absolute value of the measurement result, │y │: u R (y ) = u( y ) . y (9.6) The symbols of relative standard uncertainties are: ● uR , relative standard uncertainty ● uA R , relative type A standard uncertainty ● uB R , relative type B standard uncertainty ● uC R , relative combined standard uncertainty ● UR , relative expanded standard uncertainty. [GUM] J 9.7 Evaluations from non-normally distributed results Table 9.1 gives an overview of common distributions of measurement results and the corresponding evaluations of the maximum likelihood values of a measurand, as well as the type A standard uncertainties of these maximum likelihood values. The formulas in the table are valid only when measurement results are corrected for all significant systematic errors. A non-normal distribution of results can often be approximated by the normal distribution with the arithmetic mean and deviation equal to the maximum likelihood value and deviation of the real distribution (based on the central limit theorem). See Example 6.42.2 Example of estimate of a length of gauge block and its statistical parameters. Kostić • 2019 • Metrology Companion 249 Table 9.1 The common distributions of n results and the evaluations from the results [GUM] 4.2.3, 4.4.5, 4.4.6, F.2.4.4 [NASA] 3.2.5 [Yuan] [Panchenko] 4.1 [Njegić] 3.1. The distribution of measurement results The maximum likelihood value of a measurand 9) Lognormal distribution from (6.16) and (6.17): The type A standard uncertainty of the maximum likelihood value xmode = = most_often_value = = x from (6.56) and (6.49): max L ( x ) Normal distribution u= sx = n n ( xi − x ) = T-distribution 2 i =1 n ⋅ (n − 1) from (6.30): x = Uniform distribution Triangular distribution U-distribution 9) x1 + x2 + ... + xn n Note from (6.56) and (6.95): In the case of the uniform distribution and U-distribution, x is the best estimated value of a measurand according to the criterion of the smallest absolute value of the maximum possible error. u= from (6.56) and (6.100): u= a 6⋅n from (6.56) and (6.105): u= 250 a 3⋅n a 2⋅n Kostić • 2019 • Metrology Companion 9.11 Procedure for estimating the measurement result and associated data According to [GUM], the steps that should be followed for estimating the measurement result and associated data may be summarized as follows. 1) Write the functional relationship, symbolically denoted as f, which gives the value of measurand, Y, from its component values Xi : Y = f ( X1 , X2 , …, XN ). (9.8) The function f should contain all quantities that significantly increase the uncertainty of the measurement result. See Examples 2.14.1, 4.7.1 and 9.13.2. The functional relationship can be extremely complex or can never be written down explicitly. In such case, it should be given as a computer program, or an algorithm, that can be computed numerically. 2) Determine the best estimated values xi of the component values Xi ; preferably on the basis of measurement results, or by other means. See 6.9 Best estimate. 3) Determine the standard uncertainty u( xi ) of each estimated value xi . Depending on the kind of available data, the standard uncertainty is estimated according to: 9.2 Type A standard measurement uncertainty; 9.3 Type B standard measurement uncertainty; or according to this same procedure. 4) Determine the standard covariances of any component values that are correlated. See 7.4 Experimental standard covariance. 5) Determine the degrees of freedom, νi , of each standard uncertainty u( xi ). See 6.24 Degrees of freedom; 6.25 Effective degrees of freedom; 6.26 Degrees of freedom from the deviation of the standard deviation of the arithmetic mean; or 6.28 Standard deviation and its degrees of freedom from the level and interval of confidence and sample size. 6) Compute the measurement result, y, of the value of measurand, Y, from the functional relationship symbolically denoted as f in (9.8), or from other relationship. For this, use the evaluations xi , instead of the component values Xi . Estimate the distribution of the measurement result according to 8.6 Estimating the type and parameters of a distribution. 254 Kostić • 2019 • Metrology Companion 7) Compute the combined standard uncertainty u C ( y), of the measurement result, y, from the standard uncertainties and standard covariances of the estimated values xi . See 9.4 Combined standard measurement uncertainty. If the measurement determines several measurement results from the same estimated values xi , also compute the covariances of these results. See 7.4 Experimental standard covariance, and also Example 9.13 Resistance, reactance and impedance, computed from the same component measurement results. If it is necessary to give an expanded standard uncertainty U on the basis of the required level of confidence, P, select coverage factor k, then compute the uncertainty U. See 9.5 Expanded standard measurement uncertainty. 8) Estimate the degrees of freedom, νC , of the combined standard uncertainty u C ( y). See 6.25 Effective degrees of freedom; 6.24 Degrees of freedom; 6.26 Degrees of freedom from the deviation of the standard deviation of the arithmetic mean; or 6.28 Standard deviation and its degrees of freedom from the level and interval of confidence and sample size. [GUM] 0.7, 8, 7.2.7 Note, 4.1.2, 4.1.3, 5.2.3, 5.2.5, 4.2.6, 7.2.1 d) Kostić • 2019 • Metrology Companion 255 9.14 Metrological traceability Metrological traceability (traceability) is the distinction of a measurement result which means that its specification of combined standard uncertainty includes uncertainties of all its component values beginning from a source of traceability. The combined standard uncertainty of a traceable value is computed based on a relation between component values and the traceable value, normally, as given in 9.4 Combined standard measurement uncertainty. The value is traceable if its combined standard uncertainty is computed based on standard uncertainties of all its component values which are traceable. For such value, it is said that it is traceable to quantities of a specified source of traceability, to which its component values are traceable. Example. A meter is calibrated by a measurement standard whose value is traceable to a national measurement standard. The result of this meter, with its combined standard uncertainty, is therefore, traceable to the national measurement standard. The combined standard uncertainty of the traceable value is due to the uncertainties of all its component values beginning from the source of traceability. The combined standard uncertainty of the traceable value enables accurate computation of the level and interval of confidence. See 9.1 Measurement result and its uncertainty. Traceability chain is an unbroken sequence of calibrations and traceable results of these calibrations. The traceability chain is defined through a calibration hierarchy. Source of traceability is a quantity that is at the beginning of the traceability chain. The source of traceability has the least measurement uncertainty in the traceability chain. The measurement uncertainties of traceable values increase by starting from the source of traceability, see also 5.11 Calibration. For the traceable value, the source of traceability must be declared. “Traceable to SI” means “metrologically traceable to SI units”. A laboratory can choose the source of traceability which is suitable for its needs. [NIST] [VIM 3] 2.41, 2.42, 2.43. [VIM 2] 6.10. 268 Kostić • 2019 • Metrology Companion References [Allan] David W. Allan; Should the classical variance be used as a basic measure in standards metrology; IEEE Transactions on Instrumentation and Measurement, Vol. 1M-36, No. 2, 1987. [Anscombe] F. J. Anscombe; Graphs in statistical analysis; The American Statistician, Vol. 27, No. 1, 1973. [Aslan] B. Aslan, G. Zech; Comparison of different goodness-of-fit tests; In: M. R. Whalley, L. Lyons (editors), Proceedings of Conference on Advanced Statistical Techniques in Particle Physics, Durham, 2002. [Audoin] Claude Audoin, Bernard Guinot; The Measurement of Time: Time, Frequency and the Atomic Clock; Cambridge University Press, Cambridge, 2001. [Berger] James O. Berger; Could Fisher, Jeffreys and Neyman have agreed on testing?; Duke University, Durham, 2002. [BIPM] Internet site of the BIPM: http://www.bipm.org/; 2000 to 2019. [Box] G. E. P. Box, D. R. Cox; An Analysis of Transformations; Journal of the Royal Statistical Society, Series B (Methodological), Vol. 26, No. 2, 1964. [Box 2005] George E. P. Box, Stuart Hunter, William G. Hunter; Statistics for experimenters: design, innovation, and discovery (2nd ed.); John Wiley & Sons, 2005. [Britannica] Britannica Encyclopaedia (CD ROM); 1997. [Bronshtein] I. N. Bronshtein et al; Handbook of Mathematics; Springer-Verlag, Berlin Heidelberg, 2015. [Chamberlain] Richard Chamberlain; Computer systems validation for the pharmaceutical and medical device industries (2nd ed.); Alaren Press, Libertyville, 1994. [CODATA] Peter J. Mohr, Barry N. Taylor, David B. Newell; CODATA recommended values of the fundamental physical constants: 2014; Rev. Mod. Phys., Vol. 88, No. 3, 2016. [Cohen] Morris R. Cohen, Ernest Nagel; An introduction to logic and scientific method; Mr. Sunil Sachdev (Allied Publishers Limited), New Delhi, 1998. Kostić • 2019 • Metrology Companion 271 Goran Kostić Metrology Companion the new SI e-book on concepts and computations in measurements 1st English e-edition, 2nd English updated edition in total Leskovac, 2019 ● English editors Jelena Jerković, PhD, Senior Lecturer in English at Faculty of Technology, University of Novi Sad, Novi Sad, Serbia Bojana Komaromi, PhD, Lecturer in English at Faculty of Agriculture, University of Novi Sad, Novi Sad, Serbia ● Serbian editor Suzana Petković, Graduated Philologist ● Cover design Pavle Ristić, Graphic Designer ● Publisher Goran Kostić, Jovana Cvijića 5, RS -16 000 Leskovac, Serbia ● Editor Goran Kostić ● Production Produced by the publisher. Circulation, 1000. Release date, August 2019. ● Copyright © 2018, 2019 by the publisher, all rights reserved. ● ISBN 978-86-900284-2-9 286 Kostić • 2019 • Metrology Companion 0 References ISBN 978-86-900284-2-9 10 11 12 13 14 15 16 17 18 19 20 21 [...] 9 9 Measurement uncertainty 8 8 Tests and functions fitting 7 7 Variances In front of you is the Companion with concisely described subjects that are indispensable in all measurements. The Companion consists of about minimal and sufficient set of concepts and methods required to compute a measurement result and its measurement uncertainty. The text is in the form that enables direct applications in computations, manual or spreadsheet, and in writing software. 6 6 Outlines of the statistics 5 5 Measurement standards and calibrations The essence of measurement is the same in all areas. The purpose of this publication is to be a companion to people dealing with measurements, that is experiments or observations, from production to research, from physics and chemistry to biology and medicine. The publication is also intended for students. 4 4 Measurement results, accuracy and errors Goran Kostić 3 3 Devices for measurement 2 2 Measurements From a shop floor to the fundamental sciences, statistical analysis of results of repeated measurements is required. The purpose of the analysis is to determine the value of a measured quantity, as well as the accuracy of the determined value. It is requested that accuracy is described by the standard measurement uncertainty to enable traceability and ensure computing of the confidence level and the confidence interval. 1 1 Quantities and units of measurement From the Preface Goran Kostić Chapters THE NEW SI E-BOOK ON CONCEPTS AND COMPUTATIONS IN MEASUREMENTS