# Measurement Uncertainty in Force Measurement

## Preface

This article presents general ways of evaluating uncertainty in force measurement applications. It also applies this information to the uncertainty figures given with load cell product specifications. The information in this article is from published journals and from related documents provided by standard bodies: European Association of National Metrology Institutes (EURAMET), National Institute of Standards and Technology (NIST), American Society for Testing and Materials (ASTM), and the Joint Committee on Guides in Metrology (JCGM). The concept of uncertainty is somewhat new in measurement science. According to JCGM, as of 1977, there was no international consensus on the expression of measurement uncertainty [1]. This led the International Committee on Weights and Measures (abbreviated from the French, to CIPM) to have its sub-authority, the International Bureau of Weights and Measures (BIPM, again from French) to collaborate with various national standard laboratories to develop such standards. The resulting guidelines are applicable to a large spectrum of measurements, including force measurements [1].

This article includes the following topics:

## Measurement Uncertainty In a Load Cell Data Sheet

For a strain gauge, the datasheet specifies uncertainty as a percentage range of full-scale output (FSO). The FSO is specified as . Mathematically, this translates to (since x ), or a range of to .

Ideally, to derive this uncertainty, nationally and internationally accepted procedures are used over repeated measurements with known masses to determine the differences between this load cell’s outputs and the expected measurement result. As explained later, these differences will generally follow a Gaussian distribution around the expected value. The specified uncertainty value is generally two or three standard deviations above/below this expected value, meaning the load cell should give an output within this range with a confidence of 95% (when the uncertainty represents two standard deviations from the mean) or 99.7% (at three standard deviations). From this we then know that the FSO for the specific load cell in this section will be in the range of to about 95% of the time, around a mean expected value of , when statistical evaluations are used to derive uncertainty. The represents two standard deviations of the data of measurement samples. The number of standard deviations used to determine the confidence interval is also known as the “coverage factor,” as will be discussed later.

In reality, if every load cell were tested this way, the cost would be prohibitive. Instead, standards bodies such as OIML and NIST set guidelines for the acceptable tolerances of various classes of load cell. Then manufacturers assign one of these classes to their load cells to assert which of these tolerance values a particular design adheres to for the various parameters on load cell data sheets. Users can therefore expect output values of that load cell to fall withing those tolerances provided ambient conditions, loading, mounting and maintenance are all per manufacturer’s guidelines. According to NIST, “Tolerance values are so fixed that permissible errors are sufficiently small that there is no serious injury to either buyer or seller of commodities, yet not so small as to make manufacturing or maintenance costs of equipment disproportionately high [8].”

## Misconceptions Between Error and Uncertainty

Often the terms “error” and “uncertainty” are used interchangeably. However their distinction is important. *Measurement error* is the difference between the observed value of a quantity and its actual value. Error is a theoretical concept that can never be truly known since it’s impossible to determine the actual value of a quantity. The article, Calibrating the Force Measuring System, explains various ways error is introduced to a measurement; most are attributable to improper use of the measuring device or improper calibration.

Measurement error differs from *measurement uncertainty* because error is a single value (which cannot be known), while uncertainty is a range of measurement values that will occur with a probability known as the *confidence interval*. That is, uncertainty represents a manufacturer’s confidence in how close the measurement is to the actual value. Intuitively we understand this because even when all the sources of error have been identified and adjustments made, there is still uncertainty about the reliability of the measured value.

It is interesting to note that in the uncertainty calculation by NIST given as an example below, one of the components of combined uncertainty in their model is influenced by errors due to creep, hysteresis, loading angle and other factors controllable by proper use and calibration. Therefore the final combined uncertainty figure in [5] is actually given as a best case and worst case range, when these factors are, and are not mitigated.

## The Importance of Determining Force Measurement Uncertainty

The following points express the importance of quantifying the uncertainty in force measurements

1.

The uncertainty value of a force measurement provides a basis for the comparison of the results obtained with those obtained in other laboratories or by national standards.

2.

It helps to properly interpret the results obtained under different conditions. For example, calibration maybe performed under laboratory conditions, while the measurement using the force transducer maybe under totally different conditions. Consequently, there will be differences in the results. These conditions can be grouped into geometrical, mechanical, temporal, electrical, and environmental effects [4]. Accounting for these differences is in the expression of the uncertainty of each result.

3.

The results of the evaluation of uncertainty of measurement can also serve as a statement of compliance, if a customer or regulation requires such a statement.

4.

It is a means of determining the capability of the force measurement system to provide accurate measurement results.

5.

Specifically, the uncertainty components that form the combined uncertainty value can help pinpoint the measurement variables needing improvement.

6.

Understanding the principles of measurement uncertainty used by the laboratory in combination with practical experiences can improve methods of force measurement.

7.

The principles of evaluating force measurement uncertainty can help maintain and improve product quality and quality assurance.

## Standards Procedures for Estimating Measurement Uncertainty

This section explains the general steps followed in evaluating measurement uncertainty in accordance with OIML’s *Guide to the Expression of Uncertainty in Measurement *(GUM) standard [1]. The steps are numbered in the headings.

### 1. Modeling the Measurement

To understand this step, it is important to understand that force measurement is not a direct measurement. An example of a direct measurement is determining the width of a pipe with calipers. The measuring system bases its reading on the actual width of the pipe. An indirect measurement is one where the measuring system translates an input to a different form of output, such as the input force to a load cell being converted to electrical energy, which then is translated by a display device to a reading. This conversion introduces additional uncertainty in the measurement.

Modeling the measurement is simply a way of mathematically representing this relationship between the output of a measuring system, and the input quantities . This notation of upper case and lower case variables is used to distinguish the actual value (upper case) which, recall, can never be truly known, with its best estimate (lower case). Since the output measurement is a function of all of the input measurements, it is represented by the expression below, where is the output estimate, and the input estimates are .

The important thing to understand about this is simply that, since the output estimate is a function of estimates for the inputs, and that each input estimate has a level of uncertainty associated with its value, these input uncertainties cumulatively determine the *combined standard uncertainty* associated with the output of the measuring system, denoted as .

In [5] we see how NIST calculates combined standard uncertainty as a function of the uncertainties due to the applied force, the display, and the device’s response. In turn each of these uncertainties have their own components of uncertainty, such as uncertainty in dead weight mass, acceleration due to gravity and air density as components of the uncertainty due to applied force. The standard uncertainties for each of these components is calculated per the next section.

### 2. Evaluating the Standard Uncertainty of Each Input Quantity

The standard uncertainty of each input estimate () is itself a standard deviation of the mean, or multiple thereof, given a distribution of possible values for the true or actual inputs ().

There are two methods of evaluating the standard uncertainty [7], assuming multiple types of distributions for the possible true input values. In the first method, the distribution of output values is derived from actual data, whereas in the second, the distribution of output values is theoretical.

**Type A Evaluation:** This method involves performing a series of repeated measurements of the inputs. It involves calculating the mean of all the samples, calculating the experimental standard deviation (the square root of the sum of squared differences between each sample and the mean) and finally calculating the standard deviation of the mean. The *standard deviation of the mean* is what the standard uncertainty becomes. The illustration below should explains in greater detail.

**Mean**

**Standard Deviation**

**Standard Deviation of Mean**

**Type B Evaluation:** This method involves determining the standard uncertainty through the use of existing information regarding possible true inputs, which gives the uncertainty sources and their values.

These values may come from the calibration certificate, authoritatively published quantity values, certified reference materials and handbooks, or personal experiences and general knowledge of the instrument. This method is ideal when performing repeated measurements is impractical (e.g., due to cost or time constraints). While this method relies on the triangle and rectangular distributions, Type A is associated with the normal distribution. Figure 1 below shows these distributions and their functions; in each “” is the upper and lower bound of possible values and is the uncertainty around the expected value (which is the midpoint of a distribution curve).

Clearly the normal distribution gives the best picture of uncertainty. However, this involves costly and time consuming experimentation.

A rectangular distribution is used when no confidence interval is stated for an uncertainty, and the assumption is made that all values of the input will fall into a range of values with equal probability. While this is an oversimplification, this assumption can give a valid highest bound of uncertainty, expressed by the equation below the distribution.

A triangular distribution is used when the model of possible true inputs should reflect the fact that values close to the mean are most possible, and the probability of values declines as they get further from the mean. It is an approximation of a normal distribution in the absence of actual measurements. The equation for the standard uncertainty for this assumed distribution is given below the figure.

### 3. Determining the Combined Standard Uncertainty

The most commonly used method for this is the GUM’s law of propagation of uncertainty (LPU). LPU involves the expansion of the mathematical model in a Taylor series and simplifying based on only the first order terms.

The combined standard uncertainty is simply the appropriate combination of the standard uncertainties obtained for each input quantities in step 2. The right expression for the combined standard uncertainty depends on whether the input quantities are independently or interdependently correlated. If they are independent, the expression is:

If they are interdependent, the expression below is added to the right-hand side of the one above:

In the equation above, the partial derivatives are called the *sensitivity coefficients* which are the standard uncertainty of the *i*-th input quantity. The unsquared value of the product of the partial derivatives with the standard uncertainty is called *the uncertainty component*.

### 4. Determining the Expanded Uncertainty

The expanded uncertainty widens the confidence interval of an expected measurement result, by multiplying the combined standard uncertainty by an integer known as the coverage factor, . It ensures the result encompasses a large fraction of the distributed values that could be reasonably attributed to the measurand. It is denoted by and expressed as:

Therefore, the measurement result can be expressed conveniently as . That is, . Recall the example of the FSO taken to be as .

A coverage factor of two gives a confidence interval for the uncertainty of about 95% and a coverage factor of three gives a confidence interval of uncertainty of about 99.7%, meaning the results will fall within the mean result about 95% or 99.7% of the time respectively.

## Lab Procedures for Determining Force Measurement Uncertainty

The above procedures are outlined by OIML and are internationally accepted procedures for calculating any measurement uncertainty. This section gives an example of how these general procedures are applied in a real life laboratory, in this case the procedures used by NIST to calibrate their own testing equipment and certify manufacturer’s load cells [5].

**The Combined Standard Uncertainty Model:** According to NIST, the sources of combined uncertainty (denoted ) considered to be attributable to the transducer force measurement are the applied force (denoted ) , the calibration of the indicating instrumentation (denoted ), and the fit of the measured data to the model equation (denoted ).

Because the combined standard uncertainty is a function of the uncertainties due to these three factors, NIST expresses this relationship as:

The standard combined uncertainty is simply the square root of either side of this equation. The 95% confidence interval is this number multiplied by a coverage factor of 2, and the 99.7% confidence interval is this number multiplied by a coverage factor of 3.

Each of these components of uncertainty on the right side of the equation above also have their own components of uncertainty, like the proverbial layers of an onion. Step 3 from the OIML recommendations is applied below to address how to obtain these figures.

### 1. Modeling the Measurement

The polynomial equation modeling the force transducer response is given below:

Where R is the response, F is the applied force and the is the coefficient calculated by applying the least-square fit method to the data set. This equation will become most relevant in step 2 below when we model its uncertainty, which is the third factor () of the combined standard uncertainty above.

### 2. Evaluating the Standard Uncertainty of Each Input Quantity

#### The Uncertainty in Applied Force

The components most significantly contributing to the uncertainty of the applied force () are threefold. The uncertainty in the dead weights themselves are one factor, because, recall from above, the true weight of any entity can never be known. The remaining factors contributing to applied force uncertainty are the uncertainty in the acceleration due to gravity at the altitude of the test (since force is a function of the mass and gravitational acceleration, or as we learn in high school physics, ), and the air density at the specific location. These factors may seem minuscule but for very precise systems they must be taken into account.

The standard uncertainty of the applied force, is then modeled by the equation:

where is the standard uncertainty associated with the mass, is the acceleration due to gravity, and is the standard uncertainty due to air density. The value of is substituted in the equation in step 3 below.

Each of these quantities has been determined by NIST through Type A evaluation for their specific lab location, and can be found in [5].

#### The Uncertainty of The Voltage Ratio Instrumentation

Recall that force measurement is an indirect measurement. Therefore the output of the voltage transducer must be read by a display. This inherently introduces error, since the display or other indicating instrument has its own associated uncertainty.

NIST includes the following standard uncertainty sources in this figure:

- The calibration factor of the multimeter, denoted as ; this is the ratio of the voltage indicated by the multimeter to that of a reference voltage
- The uncertainty associated with the multimeter’s linearity and resolution, denoted as , that affects its least square fit to a model curve.
- The uncertainty associated with the results of the primary calibration of the multimeter using a primary transfer standard such as a precision load cell simulator, denoted as .

In total, the standard uncertainty associated with the instrument is a combination of its sources

The value of becomes the second term in the equation in step 3 below. Its value has been derived by NIST using Type A evaluation in their laboratory and can be found in [5].

#### The Uncertainty Due to the Deviation of the Observed Data from the Fitted Curve

In step 1 above, the polynomial equation modeling the force transducer response was given as:

This equation models what we expect the output of the tested transducer to look like; it is a theoretical model given its electrical components. However, the actual readings from the transducer will deviate from this curve for various applied forces. The differences between the theoretical and measured values for a given input create the standard uncertainty of the response, denoted as and calculated as:

Where are the differences between the measured response and those calculated using the model equation, is the number of individual measurements in the calibration data set, and is the order of the polynomial modeling the theoretical output, plus one.

The value of becomes the third term in the equation in step 3 below. Again, its value has been derived by NIST using Type A evaluation in their laboratory and can be found in [5].

### 3. Determining the Combined Standard Uncertainty

As given in the introduction to this section:

The terms derived in step 2 are placed in this equation, and the combined standard uncertainty, , is calculated by taking the square root of each side.

### 4. Determining the Expanded Uncertainty

As previously stated, the expanded uncertainty is the product of the combined uncertainty and the coverage factor, . For NIST’s desired confidence interval of 95%, the value of is 2, and the expanded uncertainty is therefore:

.

## Mitigating Sources of Uncertainty

Some components of the uncertainty modeled in force measurement can be reduced with proper technique or load cell design. For example proper axial loading, the load cell measurement resolution, the ability to reduce unwanted noise, and the repeatability of the display results all contribute to the deviations in the response vs. the theoretical response curve. These can be mitigated with a robust design. Moreover, hysteresis and creep contribute greatly to the uncertainty of the response, , but are controllable with proper maintenance and calibration (See “Calibrating the Force Measuring System“).

## Conclusion

Measurement uncertainty is an important concept to understand when selecting and also when calibrating and maintaining load cells. Whereas load cell specification sheets give values of uncertainty, they do not explain their derivation. This article explains globally accepted standard procedures for calculating measurement uncertainty.

In practice, manufacturers design to given, accepted tolerances for a particular class of load cells. Then they specify a load cell measuring range or range of forces that is narrower than the maximum and minimum forces the cell is truly capable of measuring, since within that range the potential errors are generally well within the bounds of the load cell’s stated tolerances.

## References

##### [1]

Joint Committee on Guides in Metrology (JCGM), “Evaluation of Measurement Data – Guide to the Expression of Uncertainty in Measurement,” JCGM, 2008.

##### [2]

Joint Committee for Guides in Metrology (JCGM), “JCGM 200:2008 International vocabulary of metrology – Basic and general concepts and associated terms (VIM),” Joint Committee for Guides in Metrology, 2008.

##### [3]

A. G. Piyal Aravinna, “Basic Concepts of Measurement Uncertainty,” 2018.

##### [4]

Dirk Röske, Jussi Ala-Hiiro, Andy Knott, Nieves Medina, Petr Kaspar, Mikołaj Woźniak, “Tools for uncertainty calculations in force measurement,” *ACTA IMEKO, *vol. 6, pp. 59-63, 2017.

##### [5]

Thomas W. Bartel, “Uncertainty in NIST Force Measurements,” *Journal of Research of the National Institute of Standards and Technology, *vol. 110, no. 6, pp. 589-603, 2005.

##### [6]

European Association of National Metrology Institutes (EURAMET), “Uncertainty of Force Measurements,” 2011.

##### [7]

Jailton Carreteiro Damasceno and Paulo R.G. Couto, “Methods for Evaluation of Measurement Uncertainty,” IntechOpen, 2018.

##### [8]

NIST Handbook 44, “Specifications, Tolerances and Other Technical Requirements for Weighing and Measuring Devices,” Appendix A, “Fundamental Considerations,” 2018 Edition