In general, the result of a measurement is only an approximation or estimate of the true value of the quantity subject to measurement, and thus the result is of limited value unless accompanied by a statement of its uncertainty. Much (but not all) of the scientific data appearing in the literature does include some indication of the uncertainty, but this may be stated in many different ways and is often explained poorly. In an effort to encourage consistency in uncertainty statements, the International Committee for Weights and Measures (CIPM) of the Bureau International des Poids et Mesures initiated a project, in collaboration with several other international organizations, to prepare a set of guidelines expressing international consensus on the recommended method of stating uncertainties. This project resulted in the publication of the Guide to the Expression of Uncertainty in Measurement (Refs. 1 and 2), which is often referred to as GUM. The recommendations of GUM have been summarized by the National Institute of Standards and Technology in NIST Technical Note 1297, Guidelines for Evaluating the Uncertainty of NIST Measurement Results (Ref. 3).
In the notation of GUM, we are concerned with the measurand, i.e., the quantity that is being measured. In physics and chemistry this is usually called a physical quantity and represents some inherent characteristic of a material, system, or process that can be expressed in numerical terms — specifically as the product of a number and a reference, commonly called a unit. Thus, the density of water at room temperature is (approximately) 0.998 g/mL (grams per milliliter) or, alternatively 998 kg m–3 (kilograms per meter cubed). This statement gives the most likely value of the measurand, to this level of precision, but gives no information on how much the stated value might differ from the true value. A more detailed discussion of measurement terminology is given in the International Vocabulary of Metrology (VIM) (Ref. 4).
It is important to differentiate between the terms error and uncertainty. The error in a measurement is the difference between the measured value and the true value; the error can be stated if the true value is known (to some level of accuracy). The uncertainty is an estimate of the maximum reasonable extent to which the measured value is believed to deviate from the true value, in a situation where the true value is not known (most often the case). The result of a measurement can unknowably be very close to the true value, and thus have negligible error, even though its uncertainty is large.
The uncertainty of the result of a measurement generally consists of several components, which may be grouped in two types according to the method used to estimate their numerical values:
The terms “random uncertainty” and “systematic uncertainty” are often used, but these terms do not always correspond in a simple way to the A and B categories. This is because the nature of an uncertainty component is conditioned by how the quantity appears in the mathematical model that describes the current measurement process. An uncertainty component arising from a systematic effect may in some cases be evaluated by methods of Type A while in other cases by methods of Type B.
In the GUM formulation, each component of uncertainty, whether in the A or B category, is represented by an estimated standard deviation, termed standard uncertainty, symbol ui , and equal to the positive square root of the estimated variance ui2.
For an uncertainty component of Type A, ui = si , where si is the statistically estimated standard deviation, as determined from a series of observations by appropriate statistical analysis. Any valid statistical method may be used. Examples are calculating the standard deviation of the mean of a series of independent observations; using the method of least squares to fit a curve to data in order to estimate parameters of the curve and their standard deviations; and carrying out an analysis of variance (ANOVA) in order to identify and quantify random effects in certain types of measurements. Details of statistical analysis are given in Refs. 5–9 and many other places.
In a similar manner, each uncertainty component of Type B is represented by a quantity uj , which is obtained from an assumed probability distribution based on all the available information about the measurement process. Because uj is treated like a standard deviation,the standard uncertainty in each Type B component is simply uj . The evaluation of uj is usually based on scientific judgment using all the relevant information available, which may include
The specific approach to evaluating the standard uncertainty uj of a Type B uncertainty will depend on the detailed model of the measurement process. The following are examples of steps that may be used:
When all the standard uncertainties of Type A and Type B have been determined in this way, they should be combined to produce the combined standard uncertainty (suggested symbol uc), which may be regarded as the estimated standard deviation of the measurement result. This process, often called the law of propagation of uncertainty or “root-sum-of-squares,” involves taking the square root of the sum of the squares of all the ui . In many practical measurement situations, the probability distribution characterized by the measurement result y and its combined standard uncertainty uc(y) is approximately normal (Gaussian). When this is the case, uc(y) defines an interval y – uc(y) to y + uc(y) about the measurement result y within which the value of the measurand Y estimated by y is believed to lie with a level of confidence of approximately 68%. That is, it is believed with an approximate level of confidence of 68% that y – uc(y) ≤ Y ≤ y + uc(y), which is commonly written as Y = y ± uc(y).
In fundamental metrological research (involving physical constants, calibration standards, and the like) the combined standard uncertainty uc is normally used as the statement of uncertainty in a measurement. In most cases, however, it is desirable to use a measure of uncertainty that defines an interval about the measurement result y within which the value of the measurand Y is confidently believed to lie. The measure of uncertainty intended to meet this requirement is termed expanded uncertainty, suggested symbol U, and is obtained by multiplying uc (y) by a coverage factor, suggested symbol k. Thus U = kuc(y) and it is believed with high confidence that y – U ≤ Y ≤ y + U, which is commonly written as Y = y ± U. The value of the coverage factor k is chosen on the basis of the desired level of confidence to be associated with the interval defined by U = kuc. Typically, k is in the range 2 to 3. When the normal distribution applies, U = 2uc (i.e., k = 2) defines an interval having a level of confidence of approximately 95%, and U = 3uc defines an interval having a confidence level greater than 99%. In current international practice it is most common to use k = 2, corresponding to about 95% confidence, but the value of k should be stated in each case to avoid confusion. See Refs. 1 and 3 for methods of calculating k when a value other than k = 2 is needed for a specific requirement.
It should be noted that the International Union of Pure and Applied Chemistry (IUPAC) is reviewing recommendations on metrological and quality concepts in analytical chemistry (Ref. 10).