#### Abstract

The traditional measurement theory interprets the variance as the dispersion of a measured value, which is actually contrary to a general mathematical concept that the variance of a constant is 0. This paper will fully demonstrate that the variance in measurement theory is actually the evaluation of probability interval of an error instead of the dispersion of a measured value, point out the key point of mistake in the traditional interpretation, and fully interpret a series of changes in conceptual logic and processing method brought about by this new concept.

#### 1. Introduction

Human scientific research begins with the measurement of various physical quantities, and measurement is the basis of modern scientific research. Especially for the research of artificial intelligence, it is an essential process to obtain natural information and evaluate its authenticity, which itself is the research area of measurement. Therefore, a rigorous measurement theory should be an important part of the modern scientific theory system. This paper points out the main problems existing in the traditional measurement theory and gives the correct interpretation method for the measurement theory.

In traditional measurement theory, the measured value (or observed value) is considered as a random variable, and variance is interpreted as the dispersion of the measured value, both the precision and uncertainty are defined as the dispersion concept of measured value (or observed value) [1–5], so that people can hardly make clear the conceptual difference between them.

However, in any measurement, both the measured value and every observed value are numerical values. According to probability theory, the variance of a numerical value (constant) is zero. So, how does a numerical value show dispersion? Next, we illustrate the contradictory expression of “variance” in traditional theory.

For example, the measured value of Mount Everest elevation in 2005 is , and its precision is . But in fact, this mathematical expression gives an equation , which violates the basic mathematical concept, because the equation inevitably leads to the equation , and according to the concept of in probability theory, there must be . Although many other measured values , , …, can be obtained by repeatedly measuring the height of the Mount Everest, and there can be , the equations and still exist. Therefore, the equation can never be consistent with mathematical concepts.

It can be seen that in traditional theories, the definition of precision and uncertainty is the dispersion of all possible measured values, but their mathematical expression is the dispersion of a single measured value (a single numerical value). This approach is actually a stealth change of concept and will inevitably lead to a series of conceptual logic troubles.

A question arises. All measured values that diverge from each other can definitely be used to describe a random variable, but a measured value must be a numerical value and belong to a constant. So, how should the measurement theory be interpreted?

So far, in the measurement industry, there is no literature questioning the conceptual category of measured values, such as the recent literature [6].

In references [7–10], the authors proposed some new concepts to reinterpret measurement theory. Ye et al. [7] proposed a new error epistemology that all errors follow a random distribution and cannot be classified as systematic error and random error. Ye et al. [8] pointed out that the standard deviation (variance) is the evaluation value of the probability interval of error, any error has a variance for evaluating its uncertainty, and so on. Ye et al. [9] pointed out that the dispersion and deviation of repeated observations are determined by the changing rules of repeated measurement conditions and it is possible and correct to handle errors according to the function model or the random model. Ye and Ding [10] pointed out that the measured value is a numerical value whose variance is zero and the dispersion of a measured value is an incorrect concept. According to these new concepts, the mathematical expression of the Mount Everest elevation case should be and , where represents the measured value and represents its error.

Although these new concepts have been proposed to reinterpret measurement theory, the root of these concepts and the interpretation process have not been fully described mathematically. Therefore, in this paper, the authors follow strict mathematical concept to point out the misunderstanding of the traditional concepts, give a clear interpretation for the origin of these new measurement concepts, and systematically explain a series of changes in theoretical logic and mathematical processing.

#### 2. Constant and Random Variable

In probability theory, a constant is a numerical value, such as 100, 150, *x* = 100, *x* = 8844.43, and so on.

Unlike constants, a random variable is an unknown quantity whose actual value cannot be given. Because the random variable is unknown or uncertain, we can only describe the probability range of its value. In order to study its probability range, it is necessary to study the distribution range of all its possible values (sample space), while all possible values refer to the set of test values of random variables under all permitted possible test conditions (the random test does not have the same conditions). Mathematical expectation and variance are the numerical expressions of probability range of the random variable.

For a random variable with all possible values , , is the probability that each is (continuous random variables correspond to the probability density function ), and its mathematical expectation is defined as follows:and its variance is defined as the dispersion of its possible values :

This means that the random variable exists within a probability interval with mathematical expectation and variance or that mathematical expectation and variance are the evaluation values of its probability interval. Note that describing a random variable requires two parameters: mathematical expectation and variance, both of which are indispensable.

Now, suppose that there is a constant , and and , then

By substituting , we obtain .

Therefore, .

That is, when the variance of a random variable is reduced to zero, it becomes a constant . In other words, for a constant , because all its possible values are itself, we obtain

That is to say, a constant is a special random variable, and its mathematical expectation and variance are itself and 0, respectively. Of course, a constant is a known quantity and usually does not need to be expressed in terms of probability.

It can be seen that both constant and random variable have their own mathematical expectation and variance. Therefore, if we can give the variance of a quantity but cannot give its mathematical expectation, there must be a conceptual mistake.

It should be noted that for the random variable , its basic feature is that its value is unknown, but for a sample , because it is a numerical value, it is still a constant rather than a random variable, and and . Obviously, and . That is, constant and random variable are distinguished by whether they have a numerical value, and the sample is a numerical value, which is a constant and cannot be described by the entire set . The conceptual differences between random variable and sample are shown in Table 1.

*Example 1. *A dice has six faces corresponding to the values 1, 2, 3, 4, 5, and 6, respectively. After the dice is thrown, we do not know its display value, which is a random variable . Then, according to definitions (1) and (2), and . Its meaning is that although the display value *L* is unknown, it exists within a probability interval with 3.5 as the center and 2.92 as the width evaluation. Obviously,

*Example 2. *The exam scores of all students in a school are , in which the exam score of a student A is . It is reasonable to express an unknown score with the mathematical expectation and variance of , but it is illogical to impose on because is a known constant and has no need to be expressed by variance and mathematical expectation. Moreover, is obviously not the dispersion of future exam scores of student A. That is, .

*Example 3. *Someone’s salary is , and the salaries of all the employees in his company form a sample sequence . By making the statistics of , we can obtain the mathematical expectation and variance . In exactly the same way, although , we cannot force to belong to a random variable at all because is known.

That is to say, it is reasonable to evaluate the probability of an unknown event with the statistic values of a group of known events, but it is illogical to use it to evaluate the “probability” of a known event.

Now, we suppose there is a random variable with mathematical expectation and , and then, there isFor the constant , there areFor the random variable , there areThis is to say, a random variable with the mathematical expectation can be viewed as the superposition of a constant and a random variable taking 0 as its mathematical expectation. And it should be noted that the variance or always has nothing to do with constant .

#### 3. Origin of Conceptual Troubles in Traditional Theory

Figure 1 shows the schematic diagram of the measurement concept in traditional measurement theory. Because people notice that a measured value is in a state of random change in repeated measurement, the measured value and the random error are considered as random variables, and the variance is the dispersion of the measured value or random error. Besides, the systematic error and the true value are constant in repeated measurement, so the systematic error and the true value are considered as constants which have no variances (or the variance is zero). In this way, according to formula (2), . Therefore, traditional textbooks [11–13] usually use the form of or to express the variance. However, these are obviously inconsistent with the meanings of random variables and constants described in Section 2.

Besides, in actual measurement, we always have to give a numerical value as the final measured value. Therefore, the actual schematic diagram is shown in Figure 2. According to the concepts in Section 2, although the measured value is a sample within a random distribution, because the measured value is a numerical value and and , it is illogical to replace with .

In addition, simply replacing with cannot express a complete mathematical meaning because the traditional theory cannot submit the mathematical expectation .

On the contrary, the systematic error and the true value are unknown and are regarded as constants by traditional theory. However, according to formula (4), the mathematical expectation of a constant is itself, so it is impossible to give the numerical values of mathematical expectations of the systematic error and true value. Therefore, this so-called constant is obviously not the same concept as the constant in the probability theory, and it is also a conceptual trouble in the traditional measurement theory.

In short, in the traditional theory, except for the conceptual trouble of violating the concept that the variance of a constant is zero, the conceptual trouble of missing mathematical expectations is shown in Table 2.

Because random errors have variance but systematic errors cannot be quantitatively evaluated, traditional theories believe that trueness and accuracy are both qualitative concepts, and the relationship between the uncertainty concept and them is of course very difficult to explain.

#### 4. Probability Expression of Basic Measurement Concepts

It can be seen that because of the wrong understanding of the concept of the random variable, the conceptual logic of traditional theory actually has systematic troubles. Therefore, we need to reorganize the basic measurement conceptual logic according to the concepts in Section 2.

##### 4.1. Measured Value

In Figure 2, the measured value is an observed value within all possible observed values and is a numerical value. According to formulas (4) and (5), there are

Please note that any quantity with a numerical value, including the error sample, the measured value of error, the detected value of instrument error, the value of mathematical expectation, and the value of variance, is a constant.

##### 4.2. Error

As shown in Figure 2, the true value of the measurand is , and the error of the final measured value , which is an unknown deviation , can be divided into and .

First of all, the error is a random variable, and . Moreover, the sample space of error is also , so error can be used to represent , that is, . According to the formulas (1) and (2), there are

That is, although deviation is unknown, it exists within a probability interval with 0 as center and as width evaluation. In other words, is the evaluation of probability interval of deviation , which expresses the degree that a surveyor cannot determine the value of deviation .

Taking the normal distribution as an example, the variance expresses that the deviation is within the interval of under the confidence probability of 68%. Variance is actually a concept of error range with probability meaning and expresses an error’s possible deviation degree.

In formula (13), the single deviation is a member within all its possible values, and the dispersion interval of all its possible values is the probability interval of this deviation . An unknown deviation follows a random distribution, which means that all possible values of the deviation follow a random distribution.

This principle obviously can be extended to the error . In fact, when we trace back to the upstream measurement of forming error , we will find that the formation principle of error is similar to that of current error and that the error is also a member within all its possible values. Therefore, there is also variance to evaluate the probability interval of error , and also .

For example, the multiplicative constant error *R* of a geodimeter [14, 15] comes from the frequency error of the quartz crystal and is always viewed as a systematic error without variance by traditional measurement theory. However, it is the output error in the field of instrument manufacturing, and the submission process of its variance will be demonstrated in a case in Section 6.

Obviously, according to the principle of formulas (7)–(10), if the mathematical expectation of an error is *C* rather than 0, then *C* must be corrected to the final measured value, and the mathematical expectation of the remaining unknown error is still 0. That is, for any unknown error , there is always

Thus, the error’s variance is expressed as follows:

in formulas (14) and (15) can express not only the deviation between the measured value and the mathematical expectation, but also the deviation between the mathematical expectation and the true value. It can even express the deviation between the measured value and the true value.

In this way, according to formulas (1) and (2), there are

Because the final measured value is unique and constant, both and are unknown deviations. In addition, both of them have their own variance; hence, it is incorrect that the traditional measurement theory considers as the random error and considers as the systematic error. Moreover, the corresponding concepts of precision and trueness are also incorrect.

It should be emphasized that formula (14) means that the mean value of all possible values of an unknown error is 0, which expresses the probabilities that an unknown error takes a positive value and a negative value are equal in our subjective cognition. From a statistical perspective, all possible values of an error refer to the set of all error values under all possible measurement conditions permitted by measurement specification, so the traditional concept of “repeated measurement under the same conditions” must be abandoned [8, 9]; otherwise, a unique error value obtained under a particular condition is only one sample within all possible values and does not represent all possible values, which is very easy to cause the illusion of .

In short, being different from the measured value, the error is unknown; because the error is unknown, we can only study its probability range; because of studying its probability range, we must study all possible values of error; because of studying all possible values of error, error samples must come from all the possible measurement conditions permitted by measurement specification, and the traditional concept of “repeated measurement under the same measurement conditions” must be abandoned.

##### 4.3. Variance of Regular Error

Any error has its variance, including the regular error, because the regular error also has all its possible values.

For example, the periodic error of a phase photoelectric distance meter [14, 15] conforms to the sine regularity, and its function model is . However, when we only observe the density distribution of all its possible values, this cyclic error’s probability density function can be derived as

As shown in Figure 3, furthermore, its variance can be derived as , and its mathematical expectation can be derived as .

Another example, the rounding error is a sawtooth cycle regularity function of a true value . However, when we only observe the density distribution of all possible values, the error also follows a random distribution, as shown in Figure 4, and its probability density function is

Its variance can be derived as , and its mathematical expectation can be derived as .

That is to say, the regularity and the randomness are the effect of observing all possible values of error from different perspectives, there is no opposition between them, and there is actually no need to dwell on the error’s regularity in the discussion of error evaluation. In other words, when a regular error is unknown, we can still use the mathematical expectation and the variance to describe its probability range. Furthermore, traditional measurement theories use the regularity and the randomness to classify errors into systematic errors and random errors, which is also proved to be incorrect.

In addition, when the measurement conditions vary with the repeated measurement, the corresponding regular errors will cause the dispersion of the repeated observations, which is exactly the same as the dispersion caused by the noise which varies with time conditions [9]. This dispersion is exactly the means by which we obtain its variance. It can be seen that, in order to obtain the variance of an error, we need to collect error samples under all possible measurement conditions.

##### 4.4. Probability Expression of True Value

With previous statements, we already know that for the measured value , there are and ; for the error , there are and . Because the error is the difference between the measured value and the true value, that is, , there are

The probability expressions of the true value , the measured value , and the error are summarized in Table 3 [10].

The above is the case where an observed value is used as the final measured value. If the mean value of *n* observations is taken as the final measured value, it can be inferred that the variance will decrease by *n* times (Section 6.1).

#### 5. Covariance Propagation

It can be seen from Table 3 that after a measured value is given, only variance needs to be studied.

##### 5.1. Covariance

Considering that two correlated errors have common component, formula (15) can be extended to any two errors. That is,

Thus, for the error sequence , the definition of variance is

That is,

Obviously, the true definition of the variance is formula (21). Formula (15) is only a special case of formula (21) when *t* = 1, and formula (13) is only a special case of formula (15) when interpreting as .

So, what is the meaning of covariance?

It is assumed that the errors *k*, *p*, and *q* are uncorrelated with each other, and that their variance are , and , respectively. Now, there are two errors and , and they contain a communal error component . Therefore, we can obtain

According to the definition of covariance,with the assumption that the errors *k*, *p*, and *q* are irrelevant from each other, we can obtain , , and , and equation (24) becomes

The covariance is actually the variance of their communal error component . That is to say, the mathematical meaning of covariance is the probability evaluation of the communal error component contained in two errors. As long as there are communal error components among different errors, there must be a covariance between them. Of course, the symbol and coefficient of communal error component should be considered in the actual measurement.

For example, the two measured value’s errors measured by the same instrument are correlated, and the errors of two instruments calibrated by the same benchmark are also correlated.

Moreover, like the above principle, when two errors are associated with the same measurement condition, there is also a covariance between them. For example, both the error of light speed in atmosphere and the thermal expansion error of metal are functions of temperature, and there is a correlation between all possible values of the two errors.

##### 5.2. Law of Covariance Propagation

Because any error has all its possible value and has its variance, the law of covariance propagation is extended to any error. In addition, the law of covariance propagation can only be interpreted as the propagation law of error’s probability interval and cannot be interpreted as propagation law of measured value’s dispersion.

Here is a measurement equation:

We can derive the total differential of equation (26):where

According to formula (21), the covariance matrix of the error sequence is

Equation (29) is the law of covariance propagation. The relationship between equations (27) and (29) is as follows:(1)In the error equation (27), the direct participants of synthesis are the error itself, each error is a deviation, and error synthesis always follows the algebra rule.(2)In the variance equation (29), the participants of synthesis are all possible values of each error instead of each error itself. It expresses the propagation relation of dispersion of all possible values of errors and also the propagation relation of probability intervals between errors.

#### 6. Statistical Calculation of Variance

Since the number of error samples is always limited in actual measurement, formula (15) can be approximated as

Formula (30) is also the source of least squares principle. That is to say, from the perspective of the new concept, only the concept of error evaluation changes, while the principle of the least square method used to obtain the best measured values does not change.

In actual measurement, in order to achieve the reduction and evaluation of the measurement error, a large number of observations should be carried out. Because errors make a large number of observations contradict from each other, the optimal measured values must be given by adjustment process, and the errors of the measured values should also be evaluated. We only discuss the case of the least square adjustment in this section.

##### 6.1. Direct Measurement for Single Measurand

A measurand is directly measured by *n* times, and *n* observations are obtained. In this way, using to represent the best measured value, the error equations are

According to the least squares method, the final measured value is

The measurement model of this measurement method is , and , , and are the samples of random variables , , and , respectively.

Taking the total differentiation of equation (32), the error propagation equation is

Now, we only discuss the variances of the error components and and make and .

Because is unknown, is a random variable, and has the same sample space as , can be represented by , that is, . Similarly, .

By applying the law of covariance propagation to formula (33), there is

According to the measurement model , there is

Therefore, according to the definition of variance, there is

Substituting and formula (34) into equation (36), we obtain

Therefore,

##### 6.2. Indirect Measurement for Single Measurand

Different from the direct measurement, each observation in the indirect measurement is the measured data of times of measurand. The error equations of the repeated measurement are

According to the least squares method, the final measured value is

The measurement model is , and , , and are the samples of random variables , and , respectively.

Similarly, for the errors and , the covariance propagation relationship is

Similarly, according to the measurement model , there is

According to the definition of variance, there is

Therefore,

##### 6.3. Indirect Measurement for Multiple Measurands

In this measurement mode, there are *t* different measurands, and each observation value is obtained by measuring the linear superposition value of multiple measurands. The error equations of the repeated measurements are

That is,

According to the principle of the least squares, its measured values are

The measurement model is , and , , and are the samples of random variables , and , respectively.

The error propagation equation is

Similarly, for the errors and , the covariance propagation relationship is

Similarly, according to the measurement model, there is

According to the definition of variance, there is

Omitting the tedious algebraic calculation process, the final result is

As you can see, the change in Bessel’s formula is that is written as , and represents the dispersion of all possible values of the deviation . Also, the standard deviation or , which is given by formula (34), (41), or (49), is also the evaluation of probability interval of the deviation or . Obviously, it is incorrect to express Bessel’s formula as in the literatures [1, 4] because is a numerical value.

For example, the measured frequency values of a quartz crystal at different temperatures are shown in Table 4, and the nominal value of frequency is . Now, we need to give a temperature correction model for the frequency error and evaluate the standard deviation of the residual error after correction.

We use the first 4 terms of the Taylor series as the temperature model of the frequency error, that is, .

In this way, the error equation set is

According to the least squares method, there is

Substituting the values in Table 4 into the above equation, there are

Solving the equations, the following equation is obtained:

Therefore, the frequency error’s function model is fitted as

Figure 5 is the comparison curve between the model and the actual error. According the formula (52), the standard deviation of residual error is

Finally, the frequency of quartz crystal is given as follows:

That is, temperature-frequency error can be corrected by the measured value of a temperature sensor, and a more accurate frequency value can be calculated. Residual error (as shown in Figure 6), which is still a regular error, is also processed by statistical rules, and the standard deviation of the residual error is ±2.3 × 10^{−6}. This error processing method has been widely used in the manufacture of photoelectric geodimeter [14, 15].

#### 7. Uncertainty

According to Figure 2, the total error of the final measured value iswhere is the deviation between measured value and expectation and is the deviation between expectation and true value.

Because the two errors are usually irrelevant, according to the law of covariance propagation (29), there is

This total standard deviation is the evaluation of probability interval of total error (the dispersion of all possible values of total error ). It can be seen that formula (61) is consistent with the traditional uncertainty evaluation (but the expression is changed from to ). Therefore, this total standard deviation is actually the uncertainty, which expresses the probability range of the total error of final measured value. And the uncertainty concept, which is interpreted as the dispersion of the measured value (constant) in the traditional measurement theory, is also proved to be incorrect.

It can be seen from Table 3 that the uncertainty is also the possible degree that the true value deviates from the measured value. That is, the uncertainty is not only the uncertainty of the error but also the uncertainty of the true value but is not the uncertainty of the measured value. The measured value, which is a certain value, has no uncertainty.

According to the interpretation of the traditional theory, and are referred as the uncertainty of Type A and the uncertainty of Type B, respectively. However, the current is actually of historical upstream measurement, and the current can also be used as in future downstream measurement. This kind of interpretation with A/B classification of the uncertainty evaluation is obviously too rigid.

Moreover, currently widely used formula (61) is only applicable to the direct repeated measurement model in Section 6.1 but has no use at all for the indirect repeated measurement in Sections 6.2 and 6.3 because in indirect repeated measurement, there are usually some error sources which not only contribute to dispersion of repeated observations but also contribute to their deviation, and it is difficult to distinguish them with A/B classification method. Therefore, the A/B classification method is not universal in practice.

Formula (61) comes from the covariance propagation law (29). Thus, the basic principle of uncertainty synthesis is covariance propagation law (29), and the uncertainty synthesis does not need to apply the interpretation of A/B classification mechanically. Here is a simple example to illustrate this principle, which is also a comparison with the traditional practice.

For example, four points A, B, C, and D are located on a straight line (Figure 7), and the observation data of distances obtained by geodimeter [14, 15] are shown in Table 5. Please solve the final measured values of each line segment and the uncertainty of each error.

Using , , and to express the final measured values of AB, BC, and CD, respectively, and using to express the measured value of the additive constant error of geodimeter, the observation error equation is

According to the least squares method, there are

Substituting the numerical values of all observed values into equation (63), we obtain

For the traditional measurement theory, the next step is to substitute (64) into (62), and six residual are obtained. Then is obtained by using the Bessel formula , and , , , and are obtained by using the covariance propagation law. Finally, , , , and are called as precision or uncertainty of Type A, but the uncertainty of Type B is almost impossible to discuss.

However, from the perspective of the new conceptual theory, there are three conceptual troubles in the above variance submission process: (1) The degree of the freedom is too small, so it is meaningless to apply the Bessel formula. (2) In Table 5, each observed value is a numerical value, and according to equation (64), each measured value is also a numerical value, so their variances should be 0. (3) The contribution of the covariance between the errors of each observation value has not been taken into account at all (uncertainty synthesis issue).

The following is the variance submission process of the new conceptual theory for this case.

Taking the total differential of equation (63), the error propagation equation is obtained as follows:

Applying covariance propagation law (29) to equation (65), the covariance propagation equation is obtained as follows:

The acquisition process of the covariance matrix is as follows.

For the observed value , its error is composed of three parts: additive constant error , multiplication constant error , and uneven indexing error . That is,

Its variance iswhere , , and are obtained by consulting instrument instructions or the tolerance standard in instrument specification. Furthermore, according to formula (21), can be deduced as follows:

Finally, the covariance matrix is obtained by substituting equation (69) into equation (66), where is called as uncertainty.

Assuming that , , and , we can obtain

Therefore, the uncertainties are

It can be seen that . Besides, it can be seen that the indication errors not only lead to the dispersion of repeated observations , but also lead to their overall deviation. If we entangle in the influence characteristics (A/B classification) of indication error on repeated observations, it will not only be unable to express, but also will not help to solve the problem. Moreover, whether errors or error are all deviations and have variances used to evaluate their probability intervals, the systematic error without variance does not exist.

#### 8. Conclusions

In short, different from the traditional theory, the new theory follows rigorous mathematical concepts and regards both the observed value and measured value as constants and the error and true value as random variables so that the conceptual logic of new theory has changed in an overall way (shown in Table 6). First, the variance (standard deviation) or the uncertainty is the evaluation of probability interval of a single error (deviation), but not the dispersion of a measured value, and any regular error’s size degree can be evaluated by them. Second, any error follows a random distribution, has variance which can be used to evaluate its size, and cannot be classified as the systematic error or the random error. Third, the error synthesis follows the algebraic rule, the variance synthesis follows the probability principle, and there is no need to use those old concepts such as systematic error, random error, precision, trueness, and accuracy. As a result, the revision of measurement textbooks and metrological terminology will become issues that need to be rediscussed in the future.

#### Data Availability

The data of the two cases in this paper are derived from historical data and are only used to describe the process of error processing. Other uses are invalid.

#### Conflicts of Interest

The authors declare that they have no conflicts of interest.