1. Introduction
Measurements of a variable are made using numerical scales. Usually, the value of a measurement is associated with a real number, but this association is not exact because of imperfect or partial knowledge due to uncertainty, vagueness or indiscernibility.
Incomplete knowledge comes from limited reliability of technical devices, partial knowledge, an insufficient number of observations, or other causes [
1]. Among the different types of uncertainty, we find imprecision, vagueness, or indiscernibility. Vagueness, in the colloquial sense of the term, refers to ambiguity, which remains in a datum due to lack of precision, although its meaning is understood. An example could be the measurement of a person’s weight using a scale, which provides a value within a range of scale accuracy, e.g., between 75 and 75.2 kg. Uncertainty refers to imperfect or unknown information. For example, it is known that the weight of a car is within limits (1000–1500 kg), but the exact value is unknown due to missing information, such as the number of occupants and the load.
The problems of vagueness and uncertainty have received attention for long time by philosophers and logicians (e.g., [
2,
3]). Computational scientists have also provided new tools for dealing with uncertainty and vagueness, such as interval analysis, either classic intervals [
4,
5] or modal intervals [
6,
7], fuzzy set theory [
8,
9] and rough set theory [
10].
Indiscernibility has also received the attention of philosophers. The identity of indiscernible [
11] states is that no two distinct things are exactly alike. It is often referred to as “Leibniz’s Law” and is usually understood to mean that no two objects have exactly the same properties. The identity of indiscernibility is interesting because it raises questions about the factors that individuate qualitatively identical objects. The marks, which are presented and developed in this paper, are outlined to address indiscernibility.
For example, the temperature in a room can be measured with a thermometer at different location points within the room to obtain a spatial distribution. So the temperature is not, for example, 20 °C but an interval of values, between, for example 19 and 21 that represents the different values of the temperature in the room. It does not represent the temperature of the room, which might be necessary for the modeling of an air-conditioning system. The temperature is at this time one value of the interval [19,21], considering the points of this interval as being indistinguishable. This is a known issue in handling “lumped” or “distributed” quantities. Moreover, the thermometer used as a measurement device has a specific precision and provides a reading on a specific scale, which is likely translated to another digital scale to be used in computations. Therefore, a real number or even an interval is not able to represent the read temperature.
Until the 20th century, the preferred theory for modeling uncertainty was probability theory [
12], but the introduction of fuzzy sets by Zadeh [
8] had a profound impact on the notion of uncertainty. At present, sources of uncertainty remain an active challenge for the scientific community, and different research efforts are directed toward finding solutions to deal with these uncertainties, such as using Bayesian inference for predictions of turbulent flows in aerospace engineering [
13], fuzzy sets for time series forecasting based on particle swarm optimization techniques [
14], modal intervals for prediction modeling in grammatical evolution [
15], interval analysis method based on Taylor expansion for distributed dynamic load identification [
16] or rough sets to evaluate the indoor air quality [
17].
In this article, marks are presented as a framework to deal with quantities represented in digital scales because this methodology can take into account many the sources of uncertainty. In any use of a mathematical model of a physical system, such as simulation, fault detection, or control, the system of marks provides values for the state variables and, simultaneously, their corresponding granularities, which represent a measure of the accumulated errors in the successive computations. This performance leads to the following:
Define intervals of variation for these variables values.
Decide the valid values, i.e., which have a meaning, provide by the semantic theorem.
Warns from which simulation step the obtained values will be meaningless because the granularity is greater than the previously fixed tolerance.
In the following sections of this paper, we present and review marks theory and basic arithmetic operations. The main contributions of this paper are the following:
A new characterization for the difference operator;
A new characterization for the difference operator;
A new implementation of the marks library and software developments that are needed to apply these methodologies.
To demonstrate the applicability and potential of marks, a well-known benchmark in process control in which the problems of uncertainty, imprecision, and indiscernibility are present is introduced. After introducing the benchmark, three different problems built on it are presented and solved, using marks: simulation, fault detection, and control.
2. Marks
An approach to deal with the inaccuracy associated with any measurement or computation process with physical quantities is built by means of an interval tool: marks, which define intervals in which it is not possible to make any distinction between its elements, i.e., indiscernibility intervals. Marks are computational objects; however, initial marks can come from either direct or indirect readings from a measurement device. Therefore, it is necessary to represent them on a computational scale to acquire suitable computation items. When a measurement reading or a computation is obtained as a number on a numerical scale, the resulting value has to be associated to a subset of R of indiscernible numbers from this value. Each point of this subset has the same right to be considered as the value of the measurement or computation. This kind of subsets leads to the concept of a mark considered a “spot” of certain size, called the “granularity” of the mark.
Let be a digital scale in floating point notation. A mark on is a tuple of the following five real numbers:
Center: a number c on the digital scale, which is the reading of the mark on .
Number of digits: the number of digits n on the digital scale , computing in double precision.
Base: the of the number system b, usually .
Tolerance: an uncertainty that expresses the relative separation among the points of the scale that the observer considers as indiscernible from the center. It is a relative value greater or equal than , which is the minimum computational error on the digital scales. It is a measure to decide if the mark is to be accepted as valid.
Granularity: an error g, coming from the inaccurate measure or computation, which will increase in every computational process. It must include the imprecision of the measure, devices calibration errors, inaccuracy of the physical system concerned, etc., and it is always expressed in relative terms, i.e., a number between 0 and 1, less than tolerance.
As the numbers n and b are specific on the digital scale , and the value of the tolerance t is assigned for the user, the mark will be denoted by . Tolerance t, number of digits n and base b define the of the mark. The set of marks of the same type is denoted by .
The center of a mark is a reading in a measurement device, so it carries an error and an uncertainty associated to the problem under study. Both yield the value of its granularity. The center and granularity define the mark on the digital scale. Now, it is possible to start the computations required by the mathematical model. The errors in each step of the computations will increase the value of the granularity.
The granularity and tolerance must satisfy a minimum condition of validity of the mark as follows:
These concepts are developed in [
7,
18], which contain the definition of a mark, its components (center, tolerance, granularity, and base number of the numerical scale), features (valid or invalid mark and associated intervals), relationships (equality and inequality), basic operators (max, min, sum, difference, product, and quotient) and general functions of marks and semantic interpretations through associated intervals to the operands and the results.
2.1. Basic Operators of Marks
Operations between marks are defined for marks of the same type and the result is also of the same type as the data. In this way, the tolerance is constant along with any computation, but the granularity increases, reflecting the step-by-step loss of information, which constitutes the deviation of the computed value from the exact value. An extract definition of elementary operators is presented after this, together with a different characterization of the operator difference.
The extension
of a basic operator
to two given marks
and
is the following:
where
is the digital computation of the function
f at
on the scale
, supposing a minimum relative displacement of
with regard to the exact value
; and
is the granularity of the result, which has to be
and is specified as the following:
where the value
is the computation error of
, which is the minimum error of any computation. Any computation has to carry this error by adding
. The term
is the smallest number, verifying the following:
where
is the
modal syntactic extension of
f [
7].
For the operators min and max, no computations with the centers are necessary and the resulting granularity is
. The results are transformed to the following:
where
and
are think positive for the operators + and −.
The difference in the particular case of two marks with equal centers is 0 without any computation. There is a similar situation with the maximum and minimum operators. In these cases, the granularity of the result must be the greatest of the operands’ granularities.
Two shortcomings in the use of marks as numerical entities are evident. The first one concerns the granularity of the difference between two marks
and
with positive but near centers. In its formula, the difference between the two centers appears as the denominator in a fraction. Consequently, when the centers are close, the granularity can be large enough to invalidate the mark, i.e., when granularity is larger than the tolerance, thus invalidating any further result. This can be avoided by taking into account the two real numbers
x and
y:
hence, it is possible to compute the difference
as the following:
where
m is a natural number large enough so that
and
are not near. The center of the resulting mark is
, but its granularity is both different and lesser.
Firstly, to avoid overflows in the computations of
and
it is convenient to normalize, a priori, the two marks to be divided by the greater mark
. For example, if,
then,
, and the normalized marks are the following:
with
, and the difference is at this time the following:
The power of one mark
with
m a natural number is a particular case of a product with the same factors
then, by induction it is the following:
Then, the numerator of (
9) is as follows:
and can be calculated by means of the former formula (
5). If
, then the following holds:
interchanging
and
when
.
will be a valid mark when its granularity is small, i.e., when the term
is less than a fixed small number
, for example
. So,
The computation of the denominator of the Equation (
9) is not problematic because all their terms are positive:
it results to
Dividing
and
and, eventually de-normalizing by multiplying by
,
For example, for
,
,
, the formula (
5) gives the following:
which is an invalid mark because the granularity is larger than 1. From (
11),
and (
13) give the following:
a valid mark.
The second shortcoming is the necessity to calculate the elementary functions (exp, log, power, trigonometric,…) for marks. This is possible with power series but when the convergence is slow, run time can belong. A better alternative is to use the routines ©FDLIBM developed at SunSoft, Sun Microsystems, to approximate these functions using polynomials. The computational processes for marks were developed and integrated into MATLAB in the MICELab research group (Institute of Informatics and Applications, University of Girona). The code to perform computations with marks can be found in [
19].
2.2. Associated Intervals
The theory of marks is a by-product of modal intervals theory, linked by the “improper” [
7] associated interval to a mark denoted by
:
where * is the product of a real number by an interval. Its domain, or set of its points, is referred to as
and called the indiscernibility margin of
. Another related interval is the external shadow defined by the following:
necessary to obtain the semantic meaning of a computation made using marks. As
, the external shadow is an improper interval that verifies the inclusion of
.
2.3. Semantic Theorem for Marks
The associated intervals allow the semantic properties of Modal Intervals to be applied to the results of functions of marks. Given the marks
, the continuous
to
function
and the modal syntactic extension
of
f, then the following holds:
This inclusion confirms the important Semantic Theorem for a function of marks, which provides meaning to any valid result in the evaluation of a function. If the mark
is the calculus of a function of marks,
, supposing that all the involved marks are valid, then we have the following:
So, every point of the external shadow is a true value of the function f for some values of the variables in the intervals .
The external shadow interval depends on the tolerance and the granularity of the mark, which shrinks the interval width with the unavoidable increase of the granularity. As the value approaches the tolerance (the center), the interval width tends to zero. This effect causes a loss of significance, which in many cases, is possible to avoid by performing a “translation” to avoid small values of the state variables. For example, adding a constant to the values of the state variables and scaling the common tolerance, if it depends on these values.
3. Benchmark
3.1. Benchmark Description
The popular three-tank benchmark problem is used to exemplify the usefulness of marks in the context of uncertainty, vagueness, and indiscernibility [
20,
21]. It consists of three cylindrical tanks of liquid connected by pipes of circular section, as depicted in
Figure 1. The first tank has an incoming flow, which can be controlled using a pump (actuator) and the outflow is located in the last tank.
The model in form of difference equations for this systems is
where the state variables
,
,
are the level of liquid in the tanks,
,
,
their respective areas,
,
,
the incoming flows,
,
,
the valves constants which represent the flux between the tanks and
is the simulation step, in seconds.
The following values were considered: the three tanks are the same heights
m, areas
m
, and intermittent inputs of the maximum incoming flows (in m
/s) are the following:
For the valves’ constants (in m
/s) the values are
and the initial liquid levels (in m) are
As an example of the application of marks for this benchmark model, we present three general problems related to many mathematical models: simulation, fault detection, and control to show the suitability of the marks for dealing with mathematical models with uncertainty and indiscernibility.
3.2. Simulation
Two different types of simulations have been performed: using real numbers and using marks, with a simulation step of
and 1000 steps of simulations (5000 s in all). The results using real numbers, for the three state variables are represented in
Figure 2. The intermittent input flow gives the sawtooth shape for the values of
.
For the simulations using marks, all magnitudes are considered as marks whose centers are the former real numbers and granularities have been fixed to , for all the marks. The levels of liquid into the tanks are influenced by perturbations and the dynamics of the inputs and outputs of liquid. Calling this variation, the tolerance can be calculated by . Taking , the common tolerance for all the simulations is . Unlike the granularities, the tolerances have to be equal for all the marks.
Results are shown in
Figure 3 that contains the intervals associated to the marks, drawn in form of little vertical segments. Together, they are represented by the dark band of the figure. The run time for the
s of the simulation is 200 s.
In accordance with the semantic of marks (
16) these results mean that, for one instant, for example
(time
s), the model outputs the result in these marks and related associated intervals:
An experimental state value like
is contained in them and, thus, consistent with the model (
17).
However, for
s the results are the following:
with the associated intervals too narrow to obtain reasonable results with the related semantic (
16).
The heights values contained in the associated intervals (
14) are consistent with the model, but these intervals depend on the value of the center of the mark because their widths tend to zero, as shown in the final parts of the graphics in
Figure 3. To avoid this effect, in this benchmark, it is possible to change to a physical system (
Figure 4), where the common height of the tanks is
. The behaviour of the liquid levels along the simulations is the same for the two physical systems.
To do this in the simulation algorithm, it is sufficient to add to the initial values of the state variables , and and to scale the tolerance to . This scaled tolerance depends on , which when increased for small granularities, the width of all the associated intervals moves near to .
For the case
, the tolerance is
. The simulation results after subtracting
to the final values of
,
and
can be found in
Figure 5.
Now the outputs for the step number 500 (
s) are the following:
and for the step number 1000 (
) are the following:
where the effect of the small center values over the associated intervals widths has disappeared. Associated intervals are truncated to avoid negative values for the heights values.
The model results and the initial granularities are strongly dependent on one another because of the inevitable increase of the granularities throughout the simulation. So, if the initial granularity is changed to
, then in the simulation step
(time
s) the resulting marks are invalid (
1) (the granularity is larger than tolerance) and all the subsequent results are invalid. For
in the simulation step,
and
are only valid for the eight first simulation steps.
As a rule of thumb, starting from a granularity , if to arrive until lasts p simulation steps, then to get a granularity 10 times larger, , lasts near to steps more. This quasi-exponential dependence causes invalidity of non-small granularities to be reached quickly, independently from the fixed value for the tolerance.
3.3. Fault Detection
The goal is to detect the presence of a fault in a system and indicate the location of the detected fault (fault isolation). It is assumed that only the measurements of the liquid levels, which are influenced by leakage in one tank or clogging in a valve are available.
In accordance with the semantic of marks (
16), a fault is detected when a measurement (within an interval of uncertainty) is outside the estimated band of associated intervals obtained by simulation using marks, indicating that it is not consistent with the model. Therefore, if the model is correct, the measurement is not. These measurements are generated simulating the behavior of the system in the following situations:
The system is non-faulty until
s, and from this time on, there is a leakage in tank 1 of approximately
% of the water inflow. The results are shown in
Figure 6, where the bands are the results of the simulation using marks and the line is the values of the heights of the liquid in the tanks for the faulty system. The comparison shows the effect of the leakage. The line is below the band from the instant
s for tank 1 and over the band for tanks 2 and 3.
The system is non-faulty until
s and from this time on, there is a clogging between tanks 2 and 3 of 50% of the nominal flow.
Figure 7 shows the effect of the clogging. The line is below the band from the instant
s for tank 3 and over the band for tanks 2 and 1, some instants later.
The simulations were performed until s to underline the comparisons.
3.4. Control
An elementary open-loop control was developed to show the usefulness of the model using marks in control systems. The process variables to be achieved are target heights of the liquid in the tanks, (or, in this test, arbitrarily chosen from the real simulation with the model, see
Figure 2). For example, the ones represented by black dots in
Figure 8. These outputs are controlled by the input flows to each tank in form of percentages
of the maximal flows
.
When at instant t, the later and nearest target height is above the interval associated to the output mark or or , then or or . When it is under, then or or . Finally, when it is inside the interval, then or or , where is the relative distance between the target height and the center of the interval (the same for and ).
The result is a set of percentages for every instant,
t. The output of the model is the band of associated intervals in
Figure 9. The target heights are very close to being contained in the corresponding marks. Therefore, they can be considered consistent with the model for the inputs flows defined by
,
and
.
4. Conclusions
Real numbers are the “ideal” framework for dealing with quantities associated with physical phenomena. However, real numbers are not attainable and “disappear” when the observer obtains the value of a quantity. The alternative is digital numbers but a measurement value becomes a point on a digital scale depending on the phenomenon itself, the accuracy, correctness, and errors of the devices used in the measurement, and the numbers of digits used to represent it on the digital scale.
A mark represents, in a consistent procedure, the point information provided by a digital scale. The system of marks has an internal structure that reflects not only the losses of information inherent in the readings on a digital scale and the evolution of the computations from them, but also the indiscernibility of the observed phenomena.
The computations performed using marks also reflect the gradual loss of information, due to numerical errors and truncations, and give relevant warnings for decision making, either on the acceptability of the results or on the usefulness of seeking more precision to achieve the necessary validity.
In conclusion, marks are an appropriate framework for any iterative process, within current research conditions and certain assumptions, where uncertainty is significant and can be generalized when needed. For example, if the process is a long simulation with many steps or an iterative approach with slow convergence, it is necessary to control the accumulation of experimental, scaling, and computational errors so as not to let it exceed the tolerance set by the observer.
The benchmark presented is a good example of the use of marks for an iterative process. Marks prove to be a correct and satisfactory tool for modeling physical systems with uncertainties in their variables and parameters. Marks provide a double contribution to any computational process: (1) the granularity as a good timely test for the validity of any result, and (2) they provide meaning to any valid result by means of the semantic theorem. The final semantics of a simulation using marks is just the one needed for problems like fault detection, control, or parameter identification (via optimization) of a mathematical model against a set of experimental data with uncertainties. This opens a wide field of applications for marks.
Author Contributions
Conceptualization, M.A.S., L.J., R.C., I.C and J.V.; methodology, M.A.S., L.J. and R.C.; software, M.A.S. and I.C.; validation, M.A.S. and I.C.; formal analysis, M.A.S., L.J. and R.C.; investigation, M.A.S., R.C., L.J. and J.V.; resources, M.A.S., R.C., L.J. and J.V.; data curation, M.A.S. and I.C.; writing—original draft preparation, M.A.S., R.C. and L.J.; writing—review and editing, M.A.S., L.J., R.C., I.C. and J.V.; visualization, M.A.S., L.J., R.C., I.C. and J.V.; supervision, M.A.S., L.J., R.C., I.C. and J.V.; project administration, R.C., I.C. and J.V.; funding acquisition, J.V. All authors have read and agreed to the published version of the manuscript.
Funding
This work was partially supported by the Spanish Ministry of Science and Innovation through grant PID2019-107722RB-C22/AEI/10.13039/501100011033 and the Government of Catalonia under 2017SGR1551.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data available at reference [
19].
Conflicts of Interest
The authors declare no conflict of interest.
References
- Kruse, R.; Schwecke, J.H. Uncertainty and Vagueness in Knowledge Based Systems Numerical Methods; Springer: Berlin/Heidelberg, Germany, 1991. [Google Scholar]
- Russel, B. Vagueness. Australas. J. Philos. 1923, 1, 84–92. [Google Scholar] [CrossRef]
- Black, M. Vagueness. Philos. Sci. 1937, 2, 427–455. [Google Scholar] [CrossRef]
- Moore, R.E. Interval Analysis; Prentice-Hall: Englewood Cliffs, NJ, USA, 1966. [Google Scholar]
- Moore, R.E. Methods and Applications of Interval Analysis; SIAM: Philadelphia, PA, USA, 1979. [Google Scholar]
- Gardenyes, E.; Mielgo, H.; Trepat, A. Modal intervals: Reasons and ground semantics. Lect. Notes Comput. Sci. 1986, 212, 27–35. [Google Scholar]
- Sainz, M.A.; Armengol, J.; Calm, R.; Herrero, P.; Jorba, L.; Vehi, J. Modal Interval Analysis: New Tools for Numerical Information; Springer: Berlin/Heidelberg, Germany, 2014; Volume 2091. [Google Scholar]
- Zadeh, L. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef] [Green Version]
- Moore, R.; Lodwick, W. Interval analysis and fuzzy set theory. Fuzzy Sets Syst. 2003, 135, 5–9. [Google Scholar] [CrossRef]
- Pawlak, Z. Rough Set. Int. J. Comput. Inf. Sci. 1982, 11, 341–356. [Google Scholar] [CrossRef]
- Leibniz, G.W. Discourse on Metaphysics and the Monadology (Trans. George R. Montgomery); Prometeus Books (First Published by Open Court, 1902): New York, NY, USA, 1992. [Google Scholar]
- Valdes-Lopez, A.; Lopez-Bastida, E.; Leon-Gonzalez, J. Methodological approaches to deal with uncertainty in decision making processes. Univ. Soc. 2020, 12, 7–17. [Google Scholar]
- Xiao, H.; Cinnella, P. Quantification of model uncertainty in RANS simulations: A review. Prog. Aerosp. Sci. 2019, 108, 1–31. [Google Scholar] [CrossRef] [Green Version]
- Chen, S.M.; Zou, X.Y.; Gunawan, G.C. Fuzzy time series forecasting based on proportions of intervals and particle swarm optimization techniques. Inf. Sci. 2019, 500, 127–139. [Google Scholar] [CrossRef]
- Contreras, I.; Calm, R.; Sainz, M.A.; Herrero, P.; Vehi, J. Combining Grammatical Evolution with Modal Interval Analysis: An Application to Solve Problems with Uncertainty. Mathematics 2021, 9, 631. [Google Scholar] [CrossRef]
- Wang, L.; Liu, Y.; Liu, Y. An inverse method for distributed dynamic load identification of structures with interval uncertainties. Adv. Eng. Softw. 2019, 131, 77–89. [Google Scholar] [CrossRef]
- Lei, L.; Chen, W.; Xue, Y.; Liu, W. A comprehensive evaluation method for indoor air quality of buildings based on rough sets and a wavelet neural network. Build. Environ. 2019, 162, 106296. [Google Scholar] [CrossRef]
- Jorba, L. Intervals de Marques. Ph.D Thesis, Universitat de Barcelona, Barcelona, Spain, 2003. Available online: http://hdl.handle.net/2445/42085 (accessed on 31 August 2021).
- Sainz, M.A.; Calm, R.; Jorba, L.; Contreras, I.; Vehi, J. Marks: A New Interval Tool. GitHub Repository. 2021. Available online: https://github.com/MiceLab/MarksLibrary (accessed on 31 August 2021).
- Sainz, M.; Armengol, J.; Vehi, J. Fault detection and isolation of the three-tank system using the modal interval analysis. J. Process Control 2001, 12, 325–338. [Google Scholar] [CrossRef]
- Amira. Documentation of the Three Tank System; Amira GmbH: Duisburg, Germany, 1994. [Google Scholar]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).