Next Article in Journal
Localization of Rolling Element Faults Using Improved Binary Particle Swarm Optimization Algorithm for Feature Selection Task
Previous Article in Journal
Characterization of Traveling Waves Solutions to an Heterogeneous Diffusion Coupled System with Weak Advection
Previous Article in Special Issue
Stieltjes and Hamburger Reduced Moment Problem When MaxEnt Solution Does Not Exist
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quasi-Deterministic Processes with Monotonic Trajectories and Unsupervised Machine Learning

by
Andrey V. Orekhov
Faculty of Applied Mathematics and Control Processes, Saint Petersburg State University, 7–9 Universitetskaya Embankment, 199034 Saint Petersburg, Russia
Mathematics 2021, 9(18), 2301; https://doi.org/10.3390/math9182301
Submission received: 29 July 2021 / Revised: 11 September 2021 / Accepted: 11 September 2021 / Published: 17 September 2021

Abstract

:
This paper aims to consider approximation-estimation tests for decision-making by machine-learning methods, and integral-estimation tests are defined, which is a generalization for the continuous case. Approximation-estimation tests are measurable sampling functions (statistics) that estimate the approximation error of monotonically increasing number sequences in different classes of functions. These tests make it possible to determine the Markov moments of a qualitative change in the increase in such sequences, from linear to nonlinear type. If these sequences are trajectories of discrete quasi-deterministic random processes, then moments of change in the nature of their growth and qualitative change in the process match up. For example, in cluster analysis, approximation-estimation tests are a formal generalization of the “elbow method” heuristic. In solid mechanics, they can be used to determine the proportionality limit for the stress strain curve (boundaries of application of Hooke’s law). In molecular biology methods, approximation-estimation tests make it possible to determine the beginning of the exponential phase and the transition to the plateau phase for the curves of fluorescence accumulation of the real-time polymerase chain reaction, etc.

1. Introduction

Many questions originating from the classical works of P. L. Chebyshev and A. A. Markov on approximation methods are of significant theoretical and great practical importance. The generalized Markov moment problem is closely related to problems of geometry, algebra, and function theory. From this point of view, it is possible to study not only tasks of the Petersburg mathematical school “on the limiting values of integrals”, but also problems of the theory of approximation, interpolation and extrapolation in various classes of functions [1].
It is also interesting that not only the main theorems and statements and the accompanying auxiliary and preparatory results help solve applied problems. It turned out to be possible to use some of them to solve modern artificial intelligence problems, such as machine-learning methods. Today, several types of machine learning are known: supervised, semi-supervised, unsupervised, and reinforcement learning. Unsupervised learning is a type of machine learning that makes decisions or searches for patterns in a set of unlabeled data. He is designed to solve three types of problems: clustering, dimensionality reduction, and anomaly detection.
Cluster analysis is a multivariate statistical procedure for training a set of labels on a sample of unlabeled data. The range of applications of cluster analysis is extensive: it is used in archeology, medicine, psychology, chemistry, biology, public administration, philology, anthropology, marketing, sociology, geology, and other disciplines. However, the universality of application has led to the emergence of many conflicting terms, methods, and approaches. This universality complicates the unambiguous use and consistent interpretation of cluster analysis. There are many clustering algorithms, and it is difficult to say which one will be the most adequate for a given dataset.
Hierarchical methods form a tree-like structure of cluster formation, which is commonly called a dendrogram. New clusters are formed from the previously formed clusters [2,3]. Another approach to solving the clustering problem in Euclidean space is based on the estimation of the distribution density of the elements of the sample population, for example, the methods DBSCAN [4] and OPTICS [5,6]. Splitting methods break objects into a priori a given number of clusters. They are used to optimize the similarity function of objective criteria, for example, when distance is the main parameter. The most famous of these algorithms is the K-means method [2,3]. Grid-based methods form the data space as a finite number of cells that form a grid-like structure. All clustering operations performed on these grids are independent of the number of data [7].
In statistics, machine learning, and information theory, dimensionality reduction is the transformation of data, which reduces the number of variables by deriving the principal variables. Dimensionality reduction of the data removes redundant or highly correlated features and reduces the amount of noise in the data. This transformation of data allows simpler mathematical models that are easier and more transparent to interpret. In practice, the following methods are commonly used for dimensionality reduction: Principal Component Analysis, Uniform Manifold Approximation and Projection, Discriminant Analysis, and Autoencoders [8,9].
The task of anomaly detection is to identify data significantly different from the typical elements of some set. Anomalies are also referred to as outliers, novelties, noise, deviations, and exceptions [10,11]. In particular, in the context of network intrusion detection, objects of interest are often not rare objects but unexpected bursts of activity. This pattern does not follow the general statistical definition of an outlier as a rare object, and many outlier detection methods (in particular, uncontrolled methods) do not work with such data [12].
Sometimes we can consider anomaly detection by unsupervised machine-learning methods as a sequential statistical analysis of monotone trajectories y t of a discrete quasi-deterministic process ξ . For example, solving cluster analysis problems, solid mechanics problems, molecular biology, switched systems, etc.
The main idea of the approach proposed in this article has a pioneering character and is as follows. If a quasi-deterministic process ξ with monotone trajectories y t is studied, then the moment t 0 of the change in the character of their increase from linear to nonlinear type may coincide with a qualitative change ξ . Analytically, this moment can be determined by comparing the squared error of the linear approximation of the trajectory y t with the squared errors of the nonlinear approximation of the same trajectory. The difference in such errors is a quadratic form. This quadratic form changes the sign at the point t 0 if the point is of qualitative change of the trajectory y t . The coefficients of the approximating functions are sought using the least-squares method by of the y t 0 k , , y t 0 2 , y t 0 1 , from the left semi-neighborhood of the point t 0 .
It is noteworthy that in our study, we use a whole set of mathematical concepts bearing the name of the outstanding Russian scientist A. A. Markov. In addition to the constructions accompanying the generalized problem of Markov moments, this is a Markov decision process, a Markov moment in time in the theory of random processes, and a Markov chain with memory.

2. Quasi-Deterministic Processes, Markov Moments and Markov Decision Process

Let T = 1 , m 1 ¯ be a finite subset of the sequence of natural numbers. The family ξ = { ξ t , t T } of random variables ξ t = ξ t ( ω ) , given on the probability space ( Ω , F , P ) is called a discrete random process [13,14]. Each random variable ξ t generates an σ -algebra, which we will denote by as F ξ t [15]. The σ -algebra generated by the random process ξ = { ξ t , t T } is the minimal σ -algebra containing all, F ξ t i.e.,
σ ( ξ ) = σ t = 1 m 1 F ξ t .
If we fix a time, t then obtain a random variable ξ t . If we fix a random event ω 0 , then obtain a trajectory of the random process ξ , which is a random sequence y t = ξ t ( ω 0 ) .
Random processes can be divided into two classes: quasi-deterministic processes and non-deterministic processes. We will consider only quasi-deterministic processes with monotonic trajectories. A random process ξ = ξ ( t , ω ) is called quasi-deterministic if all sequences y t are functions of time of a given form, but each of them depends on a random parameter ω . In general, ω can be a randomly selected number, a random vector, a random sequence, or a random function. The main property of any quasi-deterministic process is that each random event ω corresponds to only one trajectory y t of the random process ξ = ξ ( t , ω ) [16].

2.1. Markov Stopping Time

We will consider the binary problem of testing the statistical hypotheses H 0 and H 1 . The null hypothesis H 0 —the sequence y t increases linearly, and the alternative hypothesis H 1 —the sequence y t increases nonlinearly.
To test a statistical hypothesis, it will be necessary to construct a criterion allowing it to be accepted or rejected. Statistical criteria are based on a random set X. Two variants are possible; in the first, the sample X is extracted from the n -dimensional Euclidean space E n at once, i.e., it has a fixed volume. The second, when the sample X is generated over a period of time, and its size is a random variable [17]. A combined case is possible when X is extracted from E n at once, and then its variable size subsample is studied, which is formed by mapping X into itself. In the last two cases, one speaks of sequential statistical analysis and sequential statistical test [18,19].
Decision-making at a certain moment of time can be based only on the known values of the discrete process ξ = ξ ( t , ω ) . If we use a formal approach, then the events under study should be measurable in a non-decreasing sequence of σ -algebras F k generated by the process ξ = ξ ( t , ω ) [20]. On the probability space, ( Ω , F , P ) such a sequence is considered to be a family of σ -algebras F = { F t , t T } and is called a filtration if for i , j T | i < j : F i F j F . The map τ : Ω T is called the Markov moment with respect to the filtration F , if for t T is the preimage of the set { τ t } F t . If moreover Pr ( τ < + ) = 1 , then τ is called the Markov stopping time [21].
For example, a random event ω belongs to Ω from the previously introduced probability space ( Ω , F , P ) . If this random event ω is the extraction of a finite set X from the n-dimensional Euclidean space, E n then any point x ¯ E n can belong to the set X. By definition, the σ -algebra from the ( Ω , F , P ) contains all E n . In addition, this σ -algebra contains any finite set X from the space E n , all possible countable unions of such sets and complements to hims. We denote this system of sets as S E n . The same reasoning is valid for any σ -algebra F ξ t , therefore σ ( ξ ) = S E n . These σ -algebra S E n will be filtration for ξ = ξ ( t , ω ) [22,23].
If a random event ω Ω is the extraction of a countable set X from the n-dimensional Euclidean space E n , then the filtration for ξ = ξ ( t , ω ) will again be the σ -algebra S E n . If the event ω Ω is some random function f ( x ) , then the filtration for ξ = ξ ( t , ω ) is Borel’s σ -algebra B E n .
In all these cases, the Markov stopping time is the minimum value τ at which the null hypothesis is rejected— H 0 (the trajectory of the quasi-determinate process ξ = ξ ( t , ω ) increases linearly) and an alternative hypothesis is accepted— H 1 (the trajectory of the quasi-determinate process ξ = ξ ( t , ω ) increases nonlinearly).

2.2. Approximation-Estimation Tests

To test the statistical hypotheses H 0 and H 1 , we use the approximation-estimation tests [24,25]. We will construct quadratic forms of approximation-estimation tests as the difference between the quadratic error of the linear approximation of the numerical sequence y t and the quadratic error of the nonlinear approximation of y t in various classes of nonlinear functions.
Let us use the concept of an approximating function. Ordered pairs ( i , y i ) are the knots of approximation for the numerical sequence y t where i is a natural argument, y i is the corresponding value of the sequence y t . We will be called ( i , y i ) the natural knot of approximation [24,25].
The segment of the real axis [ y 0 , y k 1 ] , on which the knots y 0 , y 1 , , y k 1 are located, will be called the “current interval of approximation”.
Let a real function f ( t ) belong to class Y. The function f ( t ) approximates a numerical sequence y t by the least-squares method if
δ 2 = mín f Y i = 0 k 1 ( f ( i ) y i ) 2 .
There is always such the minimum since δ 2 is a positive definite quadratic form.
The quadratic error of approximation of the numerical sequence y t by an arbitrary nonlinear function f ( t ) is equal to the sum of the squares of the differences y t and f ( t ) at the knots y 0 , y 1 , , y k 1 for the corresponding values of the natural argument.
δ f 2 ( k 0 ) = i = 0 k 1 ( f ( i ) y i ) 2 .
The quadratic error of y t linear approximation with respect to the same knots is equal to:
δ l 2 ( k 0 ) = i = 0 k 1 ( a · i + b y i ) 2 .
If in our reasoning, the specific number of approximation knots does not play a role, then the quadratic errors will be denoted by δ f 2 and δ l 2 . These errors calculate from Formulas (3) and (4) respectively.
Let us introduce the notation: m = mín ( δ l 2 , δ f 2 ) .
We assume by definition that the increase in the numerical sequence y t along the knots y 0 , y 1 , , y k 1 is linear if m = δ l 2 . Otherwise, the increase in y t has a nonlinear character. If δ l 2 = δ f 2 , then the point y k 1 is called “critical”.
When constructing quadratic forms of approximation-estimation tests, can use a technique that facilitates calculations. The values of y t can be taken at the points, y 0 , y 1 , , y k 1 assuming that y 0 = 0 . This condition can be easily achieved at any approximation step using the transformation:
y 0 = y j y j , y 1 = y j + 1 y j , , y k 1 = y j + k 1 y j .
It is possible to construct several approximation-estimation tests, for example, logarithmic, parabolic, exponential, etc. In the general case, the approximation-estimation test can be formulated as follows.
Let δ 2 ( k 0 ) = δ l 2 ( k 0 ) δ f 2 ( k 0 ) . We will say that at the left semi-neighborhood of the point k (between points k 1 and k) is the character of the increase of the sequence y t has changed from linear to nonlinear, if the following conditions hold: for the knots y 0 , y 1 , , y k 1 is true δ 2 ( k 0 ) 0 , and for the knots y 1 , y 2 , , y k is true δ 2 ( k 0 ) > 0 . If we use terms of sequential statistical analysis, then the Markov stopping time for a quasi-deterministic process ξ = ξ ( t , ω ) , with a random parameter ω Ω and a monotonically increasing trajectory y t is
τ = mín { k δ 2 ( k 0 ) > 0 } .
where in the null hypothesis H 0 is rejected, and the alternative hypothesis H 1 is accepted.

2.3. Markov Decision Process

Let us move to decisions by unsupervised machine-learning methods. We will use the formal apparatus Markov decision process (MDP) for this. MDP is a discrete-time stochastic control process. It is used to make decisions in situations where the results are partly random and partly under the control of the decision-maker [26,27]. The MDP is in some state s at every moment, and the decision-maker can choose any action a available in the state s. In response, the MDP randomly transitions to the new state s at the next time step and gives the decision-maker a reward R a ( s , s ) .
The probability of the process entering the new s state depends on the action chosen. This probability is given by the function of transition to a new state P a ( s , s ) . Thus, the next state of s depends on the current state of s and the action of the decision-maker, a. Therefore, we can say that the next state does not depend on all previous states and actions, i.e., MDP satisfies the Markov property.
Markov’s decision-making processes are generalizations of Markov chains; the difference lies in adding actions and rewards.
Formally, the MDP is an ordered set of four elements ( S , A s , P a , R a ) , where:
  • S—set of states (state space),
  • A s —set of actions (action space) available from the state s,
  • P a ( s , s ) = Pr ( s t + 1 = s s t = s , a t = a ) —the probability that the action a in state s and at time t will lead to state s at time t + 1 ,
  • R a ( s , s ) —reward received after executing action a and transition from state s to state s .
The goal of the Markov decision process is to build a “good strategy“ for the decision-maker. This strategy is made to maximize some cumulative random reward function.
The MDP for quasi-deterministic random processes with monotone trajectories degenerates into a special case when all states S are linearly ordered into a sequence of a given form y t , i.e., for each state S there is only one action, the transition from y t to, y t + 1 with the probability either 0 or 1. Rewards also take two values, which, by definition, will be considered equal to either 0 or 1. When using approximation-estimation tests, the decision on the amount of remuneration is based on the Markov stopping time, for determining which are used y t 0 k , , y t 0 2 , y t 0 1 , from the left semi-neighborhood of the point t 0 .
By definition, a Markov chain with memory of order k is a process that satisfies the condition:
Pr ( X n = x n X n 1 = x n 1 , X n 2 = x n 2 , , X 1 = x 1 ) =
= Pr ( X n = x n X n 1 = x n 1 , X n 2 = x n 2 , , X n k = x n k ) .
where n > k , i.e., the future state depends on k the past states and the set Y n is such that Y n = X n , X n 1 , , X n k + 1 has the Markov property [28,29].
In the case of using approximation-estimation tests, the formation of a new set of points y t 0 k , , y t 0 2 , y t 0 1 , from the left semi-neighborhood points t 0 , can be considered to be some random event Ω t 0 , and it will correspond to a certain value of the quadratic form of the approximation-estimation test, which we denote as δ t 0 2 .
Consider a sequence of random events: Ω k , , Ω t , with a two-element set of outcomes { C , B } , where outcome C is the event δ t 0 2 0 and B—event δ t 0 2 > 0 . Since the probability of occurrence of either C or B depends only on the set y t 0 k , , y t 0 2 , y t 0 1 , then the sequence of random events Ω k , , Ω t , is a Markov chain with memory of order k. Consequently, the MDP, based on approximation-estimation tests for quasi-deterministic random processes with monotone trajectories, can be considered a Markov chain with memory of order k. In this case, decisions are made at the expense of the Markov stopping time, and, by definition, the decision-maker receives a maximum remuneration equal to 1.

3. Parabolic Approximation-Estimation Tests

We will distinguish between the complete parabolic approximation in the class of functions Q ( t ) = a t 2 + b t + c , and incomplete parabolic approximation in the class of functions q ( t ) = c t 2 + d . Let us denote the quadratic error of the complete parabolic approximation over k knots y 0 , y 1 , , y k 1 as
δ Q 2 ( k 0 ) = i = 0 k 1 ( a · i 2 + b · i + c y i ) 2 .
In addition, let us denote the quadratic error of incomplete parabolic approximation for the same knots as
δ q 2 ( k ) = i = 0 k 1 ( c · i 2 + d y i ) 2 .
It is known that a complete parabolic approximation is always no worse than a linear one, i.e., the inequality is true δ Q 2 δ l 2 . If we consider the incomplete parabolic approximation, then it is always no better than the complete parabolic approximation, i.e., the inequality δ q 2 δ Q 2 holds.
When comparing δ l 2 and δ q 2 three cases are possible: δ q 2 < δ l 2 , δ q 2 > δ l 2 , δ q 2 = δ l 2 .

3.1. Quadratic Errors of Linear Approximation of Natural Knots

The least-squares method will be using for calculating the coefficients a , b of the function of two variables
f l ( a , b ) = i = 0 k 1 ( a · i + b y i ) 2 .
We calculate the partial derivatives of this function
f l a = 2 a i = 0 k 1 i 2 + 2 b i = 0 k 1 i 2 i = 0 k 1 i · y i ,
f l b = 2 a i = 0 k 1 i + 2 b i = 0 k 1 1 2 i = 0 k 1 y i
and we are solving the corresponding system of linear equations.
k ( k 1 ) ( 2 k 1 ) 6 · a + k ( k 1 ) 2 · b = i = 0 k 1 i · y i , k ( k 1 ) 2 · a + k · b = i = 0 k 1 y i .
According to Cramer’s rule, a = Δ a a Δ a , b = Δ b b Δ b , where
Δ = k 2 ( k 2 1 ) 12 ; Δ a = . k . 2 i = 0 k 1 ( 2 i + 1 k ) y i ; Δ b = k ( k + 1 ) 6 i = 0 k 1 ( 2 k 1 3 i ) y i ,
then
a = 6 k ( k 2 1 ) i = 0 k 1 ( 2 i + 1 k ) y i , b = 2 k ( k + 1 ) i = 0 k 1 ( 2 k 1 3 i ) y i .
We use Formulas (4) and (15), to explicitly write down linear approximating functions over natural knots. Let us calculate the square-law errors for three, four, five, six, and seven natural knots.
For natural knots, y 0 , y 1 , y 2 the linear approximating function has the form:
a t + b = 1 6 3 y 2 · t + ( 2 y 1 y 2 ) .
Then
δ l 2 ( 3 0 ) = i = 0 2 3 y 2 · i + ( 2 y 1 y 2 ) 6 y i 2 = 1 6 y 2 2 y 1 2 .
Similarly for knots y 0 , y 1 , y 2 , y 3
δ l 2 ( 4 0 ) = 1 10 7 y 1 2 + 7 y 2 2 + 3 y 3 2 2 y 1 ( 2 y 2 + y 3 ) 8 y 2 y 3 .
For knots y 0 , y 1 , y 2 , y 3 , y 4
δ l 2 ( 5 0 ) = 1 10 7 y 1 2 + 8 y 2 2 + 7 y 3 2 + 4 y 4 2 2 y 1 ( 2 y 2 + y 3 ) 4 y 2 ( y 3 + y 4 ) 8 y 3 y 4 .
For knots y 0 , y 1 , y 2 , y 3 , y 4 , y 5
δ l 2 ( 6 0 ) = 2 105 ( 37 y 1 2 + 43 y 2 2 + 43 y 3 2 + 37 y 4 2 + 25 y 5 2 y 1 ( 22 y 2 + 13 y 3 + 4 y 4 5 y 5 ) y 2 ( 16 y 3 + 13 y 4 + 10 y 5 ) y 3 ( 22 y 4 25 y 5 ) 40 y 4 y 5 ) .
And for knots y 0 , y 1 , y 2 , y 3 , y 4 , y 5 , y 6
δ l 2 ( 7 0 ) = 1 28 ( 20 y 1 2 + 23 y 2 2 + 24 y 3 2 + 23 y 4 2 + 20 y 5 2 + 15 y 6 2 4 y 1 ( 3 y 2 + 2 y 3 + y 4 y 6 ) 2 y 2 ( 4 y 3 + 3 y 4 + 2 y 5 + y 6 ) 8 y 3 ( y 4 + y 5 + y 6 ) 2 y 4 ( 6 y 5 + 7 y 6 ) 20 y 5 y 6 ) .

3.2. Quadratic Forms of Parabolic Approximation-Estimation Tests

The least-squares method will be using for calculating the coefficients c , d of the function of two variables
f q ( c , d ) = i = 0 k 1 ( c · i 2 + d y i ) 2 .
Let us calculate the partial derivatives:
f q c = 2 c i = 0 k 1 i 4 + 2 d i = 0 k 1 i 2 2 i = 0 k 1 i 2 · y i ,
f q d = 2 c i = 0 k 1 i 2 + 2 d i = 0 k 1 1 2 i = 0 k 1 y i ,
and solve the system of two linear equations for the unknowns c , d
k ( k 1 ) ( 2 k 1 ) ( 3 k 2 3 k 1 ) 30 · c + k ( k 1 ) ( 2 k 1 ) 6 · d = i = 1 k 1 i 2 · y i , k ( k 1 ) ( 2 k 1 ) 6 · c + k · d = i = 0 k 1 y i .
c = 30 k ( k 1 ) ( 2 k 1 ) ( 8 k 2 3 k 11 ) i = 0 k 1 ( 6 i 2 ( k 1 ) ( 2 k 1 ) ) y i ,
d = 6 k ( 8 k 2 3 k 11 ) i = 0 k 1 ( 3 k ( k 1 ) 1 5 i 2 ) y i .
We use Formulas (9), (26) and (27), and write incomplete parabolic (without linear term) approximating functions for natural knots. Let us calculate the quadratic errors for them, and then, taking into account the corresponding errors of the linear approximation (Formulas (17)–(21)), write quadratic forms for the parabolic approximation-estimation tests δ l q 2 for three, four, five, six and seven knots.
For knots y 0 , y 1 , y 2 obtain
c t 2 + d = 2 52 ( 7 y 2 2 y 1 ) · t 2 + ( 12 y 1 3 y 2 ) .
Then
δ q 2 ( 3 0 ) = i = 0 2 2 52 ( 7 y 2 2 y 1 · i 2 + ( 12 y 1 3 y 2 ) y i 2 = 1 26 y 2 4 y 1 2 .
Hence,
δ l q 2 ( 3 0 ) = δ l 2 ( 3 0 ) δ q 2 ( 3 0 ) = 1 39 2 y 1 2 14 y 2 y 1 + 5 y 2 2 .
Likewise for knots y 0 , y 1 , y 2 , y 3
δ q 2 ( 4 0 ) = 1 98 61 y 1 2 + 73 y 2 2 + 13 y 3 2 44 y 1 y 2 + 6 y 1 y 3 60 y 2 y 3 .
δ l q 2 ( 4 0 ) = δ l 2 ( 4 0 ) δ q 2 ( 4 0 ) = 1 245 19 y 1 2 11 y 2 2 + 41 y 3 2 + 12 y 1 y 2 64 y 1 y 3 46 y 2 y 3 .
For knots y 0 , y 1 , y 2 , y 3 , y 4
δ q 2 ( 5 0 ) = 1 870 ( 571 y 1 2 + 676 y 2 2 + 651 y 3 2 + 196 y 4 2 2 y 1 ( 224 y 2 + 99 y 3 76 y 4 ) 288 y 2 y 3 148 y 2 y 4 648 y 3 y 4 ) .
δ l q 2 ( 5 0 ) = δ l 2 ( 5 0 ) δ q 2 ( 5 0 ) = 1 435 ( 19 y 1 2 + 10 y 2 2 21 y 3 2 + 76 y 4 2 + + 2 y 1 ( 25 y 2 + 6 y 3 38 y 4 ) 10 y 2 ( 3 y 3 + 10 y 4 ) 24 y 3 y 4 ) .
For knots y 0 , y 1 , y 2 , y 3 , y 4 , y 5
δ q 2 ( 6 0 ) = 2 2849 ( 987 y 1 2 + 1107 y 2 2 + 1187 y 3 2 + 1047 y 4 2 + 435 y 5 2 y 3 ( 468 y 4 + 459 y 5 ) 7 y 1 ( 104 y 2 + 69 y 3 + 20 y 4 43 y 5 ) y 2 ( 480 y 3 + 263 y 4 16 y 5 ) 1124 y 4 y 5 ) .
δ l q 2 ( 6 0 ) = δ l 2 ( 6 0 ) δ q 2 ( 6 0 ) = 4 42735 ( 127 y 1 2 + 448 y 2 2 152 y 3 2 323 y 4 2 + 1825 y 5 2 + + y 1 ( 983 y 2 + 977 y 3 + 236 y 4 1240 y 5 ) + y 2 ( 344 y 3 673 y 4 2155 y 5 ) y 3 ( 967 y 4 + 1645 y 5 ) + 290 y 4 y 5 ) .
And for knots y 0 , y 1 , y 2 , y 3 , y 4 , y 5 , y 6
δ q 2 ( 7 0 ) = 1 1092 ( 792 y 1 2 + 855 y 2 2 + 920 y 3 2 + 927 y 4 2 + 792 y 5 2 + 407 y 6 2 24 y 1 ( 22 y 2 + 17 y 3 + 10 y 4 + y 5 10 y 6 ) y 2 ( 384 y 3 + 258 y 4 + 96 y 5 102 y 6 ) y 3 ( 288 y 4 + 216 y 5 + 128 y 6 ) y 4 ( 384 y 5 + 450 y 6 ) 864 y 5 y 6 ) .
δ l q 2 ( 7 0 ) = δ l 2 ( 7 0 ) δ q 2 ( 7 0 ) = 1 546 ( 6 y 1 2 + 21 y 2 2 + 8 y 3 2 15 y 4 2 6 y 5 2 + 89 y 6 2 + + y 2 ( 36 y 3 + 12 y 4 30 y 5 90 y 6 ) y 3 ( 12 y 4 + 48 y 5 + 92 y 6 ) y 4 ( 42 y 5 + 48 y 6 ) + 42 y 5 y 6 ) .
In the general case, the parabolic approximation-estimation test has the form:
δ l q ( k 0 ) = δ l 2 ( k 0 ) δ q 2 ( k 0 ) .
If the inequality δ l q ( k 0 ) 0 holds for natural knots y 0 , y 1 , , y k 1 , but for knots y 1 , y 2 , , y k , holds δ l q ( k 0 ) > 0 , then we can say that near the point y k the character of the increase y t changed from linear to parabolic [24,25], i.e., under the condition δ l q ( k 0 ) > 0 the null hypothesis H 0 is rejected, the alternative hypothesis H 1 is accepted, and the Markov stopping time for the parabolic approximation-estimation test will be
τ = mín { k δ l q ( k 0 ) > 0 } .
It is important to note that in all cases, the point y k is an “upper estimate” for the corresponding critical value of the monotonically increasing numerical sequence y t .

3.3. Solution of Inverse Problem and Remark on Finite Differences

Of particular interest is the solution of the “inverse problem”. Specifically, let the values of the sequence y t at the knots y 0 , y 1 , , y k 2 be known, and it is required to determine at what value in the knot y k 1 we can say that the character of the increase of the y t has changed from linear to parabolic. In other words, it is necessary to calculate at what numerical value the point y k 1 will become critical.
Let us solve this problem for the knots y 0 , y 1 , y 2 , y 3 .
Equated to zero the quadratic form δ 2 ( 4 0 ) , and replace y 3 by x obtain the quadratic equation
41 x 2 64 y 1 + 46 y 2 x + 19 y 1 2 + 12 y 2 y 1 11 y 2 2 = 0 ,
for which
x 1 , 2 = 32 y 1 + 23 y 2 ± 7 5 y 1 + 2 y 2 41 .
Considering that 7 5 15.65 and 0 y 1 y 2 y 3 we will obtain
y 3 = 32 y 1 + 23 y 2 + 7 5 y 1 + 2 y 2 41 .
Now we can answer the frequently asked questions. First, “Why can’t we limit ourselves to a simple comparison of finite differences?” Here we mean that when passing to a parabolic increase, the finite difference Δ Y k 1 = y k y k 1 is greater than the previous finite difference Δ Y k 2 = y k 1 y k 2 , i.e., Δ y k 1 Δ y k 2 = K > 1 . The following second question arises: “What is equal K at the critical value of y k 1 ”?
Let us look at an example that answers these questions.
We will use Formula (43) to determine the critical value wheat the knot y 3 if known knots y 0 , y 1 , y 2 .
First, let y 0 = 0 , y 1 = 0.1, y 2 = 0.2. Then, by Formula (43) y 3 = 0.381128. Let us calculate the ratio of finite differences for the obtained critical value y 3 and obtain that K = Δ Y 2 Δ Y 1 = 1.8112. The value K = 1.8113 corresponds to “already” parabolic increase, and the ratio of finite differences K = 1.8111 is to a linear change in the sequence y t .
For knots y 0 = 0 , y 1 = 1.1, y 2 = 2.3 by Formula (43) y 3 = 4.32486, the quotient K = Δ Y 2 Δ Y 1 = 1.687385. In addition, finally for y 0 = 0 , y 1 = 0.1, y 2 = 0.3 by Formula (43) y 3 = 0.513579, the quotient K = Δ Y 2 Δ Y 1 = 1.06785.
Thus, for three different sets of knots y 0 , y 1 , y 2 and the critical value at the point y 3 , were received different the ratio of the first order finite differences. Consequently, the comparison of finite differences of elements of the sequence y t cannot be used to obtain a criterion determining the point of change of character of increase for the y t [25].

4. Approximation-Estimation Tests with Irrational Coefficients

Let us construct approximation-estimation tests for four classes of nonlinear functions: of exponential— p exp t + q , of logarithms— g ln ( t + 1 ) + h , of arctangents— w arctan t + v , and of square-roots m t + l .
In the general case, for all these functions, the coefficients of the corresponding quadratic forms are irrational numbers. Therefore, in contrast to parabolic approximation-estimation tests, these coefficients can be calculated only approximately.
It is easy to see that all four approximating functions have the same structure concerning unknown coefficients α and β : α φ ( t ) + β . Using the least-squares method, we will calculate these coefficients.
Let us find the local minimum of a function of two variables
f ( α , β ) = i = 0 k 1 ( α φ ( i ) + β y i ) 2 .
First, we will calculate the partial derivatives of the function f ( α , β ) and equate them to zero.
f α = 2 i = 0 k 1 φ ( i ) ( α φ ( i ) + β y i ) ;
f e β = 2 i = 0 k 1 ( α φ ( i ) + β y i ) .
α · i = 0 k 1 φ ( i ) 2 + β · i = 0 k 1 φ ( i ) = i = 1 k 1 y i φ ( i ) ; α · i = 0 k 1 φ ( i ) + k β = i = 1 k 1 y i .
Then we solve the system of linear equations for the unknowns α and β
α = k · i = 1 k 1 y i φ ( i ) i = 0 k 1 φ ( i ) · i = 1 k 1 y i k · i = 0 k 1 φ ( i ) 2 i = 0 k 1 φ ( i ) 2 ;
β = i = 1 k 1 y i · i = 0 k 1 φ ( i ) 2 i = 0 k 1 φ ( i ) · i = 1 k 1 y i φ ( i ) k · i = 0 k 1 φ ( i ) 2 i = 0 k 1 φ ( i ) 2 .
We will construct the approximation-estimation tests by three, four, and five natural knots for exponentials, logarithms, argtangentials functions, and square-roots.

4.1. Exponential Approximation-Estimation Tests

The quadratic approximation error by natural knots in the class of exponential functions p exp t + q is
δ e 2 ( k 0 ) = i = 0 k 1 ( p exp i + q y i ) 2 .
When we use Formulas (48) and (49) and we obtain:
p = k · i = 1 k 1 y i exp i i = 0 k 1 exp i · i = 1 k 1 y i k · i = 0 k 1 exp 2 i i = 0 k 1 exp i 2 ;
q = i = 1 k 1 y i · i = 0 k 1 exp 2 i i = 0 k 1 exp i · i = 1 k 1 y i exp i k · i = 0 k 1 exp 2 i i = 0 k 1 exp i 2 .
In generalized, the exponential approximation-estimation test has the form:
δ l e ( k 0 ) = δ l 2 ( k 0 ) δ e 2 ( k 0 ) .
If for knots y 0 , y 1 , , y k 1 is valid δ l e ( k 0 ) 0 , and for knots y 1 , y 2 , , y k the inequality δ l e ( k 0 ) > 0 , then the character of growth of y t changed from linear to exponential, i.e., in this case, the null hypothesis H 0 is rejected, the alternative hypothesis H 1 is accepted. Then the Markov stopping time for the exponential approximation-estimation test will be
τ = mín { k δ l e ( k 0 ) > 0 } .
Similar to the construction of parabolic approximation-estimation tests, we will calculate the coefficients of quadratic forms for exponential, logarithmic, arctangential, and square-root tests.
For an exponential test by three knots: y 0 , y 1 , y 2 we obtain:
δ e 2 ( 3 0 ) 0.6224 y 1 2 0.33476 y 1 y 2 + 0.045015 y 2 2 .
δ l e 2 ( 3 0 ) = δ l 2 ( 3 0 ) δ e 2 ( 3 0 ) 0.044302 y 1 2 0.33191 y 1 y 2 + 0.12165 y 2 2 .
For knots y 0 , y 1 , y 2 , y 3
δ e 2 ( 4 0 ) 0.6344 y 1 2 + 0.749 y 2 2 + y 1 0.5186 y 2 + 0.05939 y 3 0.4549 y 2 y 3 + 0.0735 y 3 2 .
δ l e 2 ( 4 0 ) = δ l 2 ( 4 0 ) δ e 2 ( 4 0 ) 0.06563 y 1 2 0.04925 y 2 2 +
+ y 1 0.1186 y 2 0.2594 y 3 0.3451 y 2 y 3 + 0.2265 y 3 2 .
And for knots y 0 , y 1 , y 2 , y 3 , y 4
δ e 2 ( 5 0 ) 0.694 y 1 2 + 0.752 y 2 2 + 0.796 y 3 2 + y 2 0.371 y 3 0.02968 y 4 + + y 1 0.543 y 2 0.357 y 3 + 0.1474 y 4 0.511 y 3 y 4 + 0.0904 y 4 2 .
δ l e 2 ( 5 0 ) = δ l 2 ( 5 0 ) δ e 2 ( 5 0 ) 0.00556 y 1 2 + 0.0483 y 2 2 0.0957 y 3 2 + + y 2 0.02895 y 3 0.370 y 4 + y 1 0.1428 y 2 + 0.1572 y 3 0.1474 y 4 0.2890 y 3 y 4 + 0.3096 y 4 2 .

4.2. Logarithmic Approximation-Estimation Tests

The quadratic approximation error by natural knots in the class of logarithm functions g ln ( t + 1 ) + h is
δ n 2 ( k 0 ) = i = 0 k 1 ( g ln ( i + 1 ) + h y i ) 2 .
When we use Formulas (48) and (49) and we obtain:
g = k · i = 1 k 1 y i ln ( i + 1 ) i = 0 k 1 ln ( i + 1 ) · i = 1 k 1 y i k · i = 0 k 1 ln 2 ( i + 1 ) i = 0 k 1 ln ( i + 1 ) 2 ;
h = i = 1 k 1 y i · i = 0 k 1 ln 2 ( i + 1 ) i = 0 k 1 ln ( i + 1 ) · i = 1 k 1 y i ln ( i + 1 ) k · i = 0 k 1 ln 2 ( i + 1 ) i = 0 k 1 ln ( i + 1 ) 2 .
In generalized, the logarithmic approximation-estimation test has the form:
δ l n ( k 0 ) = δ l 2 ( k 0 ) δ n 2 ( k 0 ) .
If for knots y 0 , y 1 , , y k 1 the inequality δ l n ( k 0 ) 0 is valid, and for knots y 1 , y 2 , , y k the inequality δ l n ( k 0 ) > 0 , then the character of growth of y t changed from linear to logarithmic. When this condition is met, the null hypothesis H 0 is rejected, the alternative hypothesis H 1 is accepted, and the Markov stopping time for the logarithmic approximation-estimation test will be
τ = m i n { k δ l n ( k 0 ) > 0 } .
For knots: y 0 , y 1 , y 2 we obtain
δ n 2 ( 3 0 ) 0.65177 y 1 2 0.82244 y 1 y 2 + 0.25945 y 2 2 .
δ l n 2 ( 3 0 ) = δ l 2 ( 3 0 ) δ n 2 ( 3 0 ) 0.0148974 y 1 2 + 0.155775 y 1 y 2 0.092785 y 2 2 .
For knots: y 0 , y 1 , y 2 , y 3
δ n 2 ( 4 0 ) 0.74052 y 1 2 + 0.66471 y 2 2 + + y 1 0.44314 y 2 0.38934 y 3 0.83197 y 2 y 3 + 0.42699 y 3 2 .
δ l n 2 ( 4 0 ) = δ l 2 ( 4 0 ) δ n 2 ( 4 0 ) 0.040523 y 1 2 + 0.035294 y 2 2 + + y 1 0.043138 y 2 + 0.18934 y 3 + 0.031966 y 2 y 3 0.12699 y 3 2 .
And for knots: y 0 , y 1 , y 2 , y 3 , y 4
δ n 2 ( 5 0 ) 0.75674 y 1 2 + 0.78767 y 2 2 + 0.68619 y 3 2 + 0.53691 y 4 2 0.74609 y 3 y 4 y 2 0.47491 y 3 + 0.51389 y 4 y 1 0.35382 y 2 + 0.25967 y 3 + 0.18664 y 4 .
δ l n 2 ( 5 0 ) = δ l 2 ( 5 0 ) δ n 2 ( 5 0 ) 0.056743 y 1 2 + 0.0123264 y 2 2 + 0.0138145 y 3 2 0.136906 y 4 2 + y 2 0.074911 y 3 + 0.113895 y 4 + + y 1 0.046182 y 2 + 0.059668 y 3 + 0.18664 y 4 0.053914 y 3 y 4 .

4.3. Arctangential Approximation-Estimation Tests

The quadratic approximation error by natural knots in the class of arctangential functions w arctan t + v is
δ a 2 ( k 0 ) = i = 0 k 1 ( w arctan i + v y i ) 2 .
When we use Formulas (48) and (49) and we obtain:
w = k · i = 1 k 1 y i arctan i i = 0 k 1 arctan i · i = 1 k 1 y i k · i = 0 k 1 arctan 2 i i = 0 k 1 arctan i 2 ;
v = i = 1 k 1 y i · i = 0 k 1 arctan 2 i i = 0 k 1 arctan i · i = 1 k 1 y i arctan i k · i = 0 k 1 arctan 2 i i = 0 k 1 arctan i 2 .
In generalized, the arctangential approximation-estimation test has the form:
δ l a ( k 0 ) = δ l 2 ( k 0 ) δ a 2 ( k 0 ) .
If for knots y 0 , y 1 , , y k 1 the inequality δ l a ( k 0 ) 0 is valid, and for knots y 1 , y 2 , , y k the inequality δ l a ( k 0 ) > 0 , then the character of growth of y t has changed from linear to arctangential. When this condition is met, the null hypothesis H 0 is rejected, the alternative hypothesis H 1 is accepted, and the Markov stopping time for the arctangential approximation-estimation test will be
τ = mín { k δ l a ( k 0 ) > 0 } .
For knots: y 0 , y 1 , y 2 we obtain
δ a 2 ( 3 0 ) 0.62985 y 1 2 0.89361 y 1 y 2 + 0.31696 y 2 2 .
δ l a 2 ( 3 0 ) = δ l 2 ( 3 0 ) δ a 2 ( 3 0 ) 0.036820 y 1 2 + 0.226946 y 1 y 2 0.150292 y 2 2 .
For knots: y 0 , y 1 , y 2 , y 3
δ a 2 ( 4 0 ) 0.75 y 1 2 + 0.63932 y 2 2 0.5 y 1 y 2 + y 3 0.81898 y 2 y 3 + 0.52017 y 3 2 .
δ l a 2 ( 4 0 ) = δ l 2 ( 4 0 ) δ a 2 ( 4 0 ) 0.05 y 1 2 + 0.06068 y 2 2 + + y 1 0.1 y 2 + 0.3 y 3 + 0.01898 y 2 y 3 0.22017 y 3 2 .
And for knots: y 0 , y 1 , y 2 , y 3 , y 4
δ a 2 ( 5 0 ) 0.79001 y 1 2 + 0.76095 y 2 2 + 0.69185 y 3 2 y 2 0.52998 y 3 + 0.55804 y 4 y 1 0.36049 y 2 + 0.33425 y 3 + 0.32005 y 4 0.66300 y 3 y 4 + 0.64011 y 4 2 .
δ l a 2 ( 5 0 ) = δ l 2 ( 5 0 ) δ a 2 ( 5 0 ) 0.090007 y 1 2 + 0.039054 y 2 2 + 0.0081498 y 3 2 0.24011 y 4 2 + y 2 0.129980 y 3 + 0.15804 y 4 + + y 1 0.039511 y 2 + 0.134249 y 3 + 0.32005 y 4 0.136998 y 3 y 4 .

4.4. Square-Root Approximation-Estimation Tests

The quadratic approximation error by natural knots in the class of square-roots functions m t + l is
δ s 2 ( k 0 ) = i = 0 k 1 ( m i + l y i ) 2 .
When we use Formulas (48) and (49) and we obtain:
m = k · i = 1 k 1 y i i i = 0 k 1 i · i = 1 k 1 y i k 2 ( k 1 ) 2 i = 0 k 1 i 2 ;
l = k 2 ( k 1 ) 2 i = 1 k 1 y i i = 0 k 1 i · i = 1 k 1 y i i k 2 ( k 1 ) 2 i = 0 k 1 i 2 .
In generalized, the square-root approximation-estimation test has the form:
δ l s ( k 0 ) = δ l 2 ( k 0 ) δ s 2 ( k 0 ) .
If for knots y 0 , y 1 , , y k 1 the inequality δ l s ( k 0 ) 0 is valid, and for knots y 1 , y 2 , , y k the inequality δ l s ( k 0 ) > 0 , then character of growth of y t has changed from linear to parabolic of degrees the half. When this condition is met, the null hypothesis H 0 is rejected, the alternative hypothesis H 1 is accepted, and the Markov stopping time for the square-root approximation-estimation test will be
τ = mín { k δ l s ( k 0 ) > 0 } .
For knots: y 0 , y 1 , y 2 we obtain
δ s 2 ( 3 0 ) 0.6306 y 1 2 0.8918 y 1 y 2 + 0.31530 y 2 2 .
δ l s 2 ( 3 0 ) = δ l 2 ( 3 0 ) δ s 2 ( 3 0 ) 0.036065 y 1 2 + 0.22514 y 1 y 2 0.14863 y 2 2 .
For knots: y 0 , y 1 , y 2 , y 3
δ s 2 ( 4 0 ) 0.7492 y 1 2 + 0.6662 y 2 2 y 1 0.4838 y 2 + 0.4701 y 3 0.8086 y 2 y 3 + 0.4658 y 3 2 .
δ l s 2 ( 4 0 ) = δ l 2 ( 4 0 ) δ s 2 ( 4 0 ) 0.04921 y 1 2 + 0.033788 y 2 2 + + y 1 0.08377 y 2 + 0.27012 y 3 + 0.008612 y 2 y 3 0.16583 y 3 2 .
And for knots: y 0 , y 1 , y 2 , y 3 , y 4
δ s 2 ( 5 0 ) 0.77850 y 1 2 + 0.78601 y 2 2 + 0.69659 y 3 2 y 2 0.47608 y 3 + 0.51663 y 4 y 1 0.36531 y 2 + 0.30570 y 3 + 0.25544 y 4 0.71704 y 3 y 4 + 0.55700 y 4 2 .
δ l s 2 ( 5 0 ) = δ l 2 ( 5 0 ) δ s 2 ( 5 0 ) 0.078502 y 1 2 + 0.0139938 y 2 2 + 0.0034103 y 3 2 0.157003 y 4 2 + y 2 0.076082 y 3 + 0.116627 y 4 + + y 1 0.034690 y 2 + 0.105699 y 3 + 0.25544 y 4 0.082961 y 3 y 4 .

5. Integral-Estimation Tests

Let us consider a generalization of the approximation-estimation tests for the continuous case. We will consider continuous and non-decreasing functions y = f ( t ) on the segment [ 0 , b ] . The class of such functions will be denoted by M, i.e., f ( t ) M [ 0 , b ] .
Let us study the problem of classifying f ( t ) M [ 0 , b ] by the character of their increase on the segment [ 0 , b ] . We will choose the following etalons for such a classification by analogy with the approximation-estimation tests: linear functions — φ 1 ( t ) = a t , parabolic functions— φ 2 ( t ) = c t 2 , exponential functions— φ 3 ( t ) = p ( exp t 1 ) , logarithmic functions— φ 4 ( t ) = g ln ( 1 + t ) , arctangent functions— φ 5 ( t ) = w arctan t and square-roots— φ 6 ( t ) = m t .
All these functions are equal to zero at the point 0. Any function f ˜ ( t ) M [ 0 , b ] can be changed using the transformation f ( t ) = f ˜ ( t ) f ˜ ( 0 ) so that it also satisfies the condition f ( 0 ) = 0 . Additionally, we accept the agreement that undefined coefficients for etalon functions must satisfy the condition f ( b ) = φ i ( b ) and then:
a = f ( b ) b , c = f ( b ) b 2 , g = f ( b ) ln ( 1 + b ) , p = f ( b ) exp b 1 , w = f ( b ) ( b ) , m = f ( b ) b .
The purpose of classification will be to determine the nature of the increase in f ( t ) M [ 0 , b ] . We note that the integral characteristic of the speed is the distance traveled. Therefore, if being compared to two speeds, then the criterion could be the difference in the distance traveled. Based on this, we will define the criterion comparison rate of change of two functions f ( t ) , φ ( t ) M [ 0 , b ] as an integral
S ( b ) = 0 b f ( t ) φ ( t ) 2 d t .
The geometric meaning of this criterion is that S ( b ) estimates the area of a flat closed region bounded by the continuous curves f ( t ) and φ ( t ) .
Let us introduce the notation
S i = 0 b f ( t ) φ i ( t ) 2 d t ,
and by definition, we say that the continuous function f ( t ) “almost linearly” increases on the segment [ 0 , b ] if the inequality is true: S 1 < S i , where index i 2 , 6 ¯ , otherwise, the change in f ( t ) will be considered nonlinear.
To determine the point at which the linear increase of f ( t ) becomes nonlinear, it is necessary to compare this function with the etalons φ 1 ( t ) , φ i ( t ) , where i 2 , 6 ¯ . We will calculate two integrals with variable upper limit:
I 1 ( x ) = 0 x f ( t ) φ 1 ( t ) 2 d t , I 2 ( x ) = 0 x f ( t ) φ i ( t ) 2 d t .
Then the solution to the equation
I 1 ( x ) I 2 ( x ) = 0
with respect to the unknown x is the “critical point” b, at which the linear increase of f ( t ) changes to nonlinear.

5.1. Tangential Test for a Discrete Case

We will consider a simple computational experiment will be called the tangential test. We use tan t as a model function. As is known, in a neighborhood of zero tan t t . When the argument t tends to π / 2 on the left, the function tan t is infinitely large and approaches the vertical asymptote t = π / 2 . Therefore, on the interval 0 , 1.571 there must be a point to the left of which the ascending tan t is closer to linear, and to the right of it, the function tan t is more precisely approximated by a parabola. According to a previous convention, such a point is called the critical value of the argument.
Table 1 (adapted from [25]) shows the results of the tangential test for the approximation-estimation tests by three, four, five, six, and seven approximation knots; the values of the δ 2 were calculated by Formulas (30)–(38). It is important to note that the character of an increase in any variable does not depend on the scale. Therefore, before using the indicated formulas, the corresponding similarity transformation was performed so that the discreteness step becomes one.
In the left column of the table, the initial (before similarity transformation) value of the discreteness step is indicated, in the following columns, the value of the minimum upper estimate for the critical value of the argument for three, four, five, six, and seven knots of approximation, respectively.
It is easy to see that as the discreteness step decreases, the critical point shifts to the right, closer to the vertical asymptote. At the same time, with an increase in the number of approximation knots, the critical point at the same discreteness step shifts to the left.
Let us pay attention to the following fact: sometimes, when the discreteness step decreases, monotonically of the increase in the minimum upper bound for the critical value of the argument is violated. The explanation for this phenomenon is simple: for example, for seven points with a discreteness step of 0.1 , the critical value of the argument for tan t lies in the interval ( 0.9 , 1.0 ) . When the discreteness step decreases to 0.09 , such a point, although it shifts to the right, is within the interval ( 0.9 , 0.99 ) [25].

5.2. Tangential Test for a Continuous Case

We will be based on the results of the discrete tangential test. We can assume that as the number of approximation knots increases, the upper bound for the critical value of the function argument tan t will tend to the minimum value. Integral-estimation tests generalize approximation-estimation tests in the continuous case. It can be assumed that they are corresponding to an infinite number of approximation knots. For the parabolic approximation-estimation tests, it was shown that the minimum upper estimate for the critical point tan t is equal to 0.99 . Let us solve the same problem using the integral-estimation test and compare the results.
Let us solve the same problem using the integral-estimation test, comparing tan t with etalons functions φ 1 ( t ) = a t and φ 2 ( t ) = c t 2 .
We will be to consider two integrals with the variable upper limit:
I 1 ( x ) = 0 x tan t a t 2 d t , I 2 ( x ) = 0 x tan t c t 2 2 d t .
Let us calculate I 1 as a function of x.
I 1 ( x ) = 0 x tan 2 t 2 a t tan t + a 2 t 2 d t = 0 x tan 2 t d t 2 a 0 x t tan t d t + a 2 0 x t 2 d t .
For x [ 0 , π / 2 ) the function tan t can be expanded in a power series [30]:
tan t = t + . t 3 3 + 2 15 t 5 + + 2 2 n ( 2 2 n 1 ) B n ( 2 n ) ! t 2 n 1 + ,
where B n are Bernoulli numbers, then
t tan t d t = . t 3 3 + . t 5 15 + + 2 2 n ( 2 2 n 1 ) B n ( 2 n + 1 ) ! t 2 n + 1 + ,
I 1 ( x ) = tan t t | 0 x 2 a t 3 3 + t 5 15 + o ( t 6 ) | 0 x + a 2 t 3 3 | 0 x = = x + v x 3 3 + 2 15 x 5 x + a 2 2 a 3 x 3 2 a 15 x 5 + o ( x 6 ) .
Consequently
I 1 ( x ) = ( 1 a ) 2 3 x 3 + 2 ( 1 a ) 15 x 5 + o ( x 6 ) .
Likewise
I 2 ( x ) = 0 x tan 2 t 2 c t 2 tan t + c 2 t 4 d t = 0 x tan 2 t d t 2 c 0 x t 2 tan t d t + c 2 0 x t 4 d t .
t 2 tan t d t = . t 4 4 + . t 6 18 + + 2 2 n 1 ( 2 2 n 1 ) B n ( n + 1 ) ( 2 n ) ! t 2 n + 2 + ,
Then
I 2 ( x ) = tan t t | 0 x 2 c t 4 4 + t 6 18 + o ( t 6 ) | 0 x + . c 2 t 5 5 | 0 x = = x + v x 3 3 + 2 15 x 5 x c x 4 2 + x 6 9 + . c 2 x 5 5 + o ( x 6 ) .
Consequently
I 2 ( x ) = v x 3 3 . c . 2 x 4 + 2 + 3 c 2 15 x 5 . c . 9 x 6 + o ( x 6 ) .
The solution to Equation (98) will be the critical point at which the “almost linear” increase in tan t changes to “almost parabolic”. Using (94) for expressions of undefined coefficients a, b and Formulas (104) and (108) we obtain the transcendental equation:
I 1 ( x ) I 2 ( x ) = x tan x 2 15 tan x . x . 6 . x 3 45 = 0
with a unique solution x 0.885 (Figure 1).
From which it follows that on the segment [ 0 , 0.885 ] the function tan t increases “almost linearly”, and to the right, its growth becomes nonlinear. Thus, the determination of the critical growth point for tan t using the integral-estimation test qualitatively coincides with the solution of the same problem using the approximation-estimation test, but the integral-estimation test gives a more minimum estimate.

6. Discussion

When applied problems to solve, discrete quasi-deterministic random processes with monotonically increasing trajectories are encountered quite often. For example, these are sequences of minimum distances at clustering by agglomerative methods, deformation diagrams of various materials, experimental creep curves, fluorescence accumulation curves for a real-time polymerase chain reaction, the dependence of network activity on generalized time (milliseconds, iterations, time to live, etc.) during DoS and DDoS attacks.

6.1. Analytical Generalization of the “Elbow Method” Heuristic

One of the main problems at the cluster analysis is the determination of the preferred number of clusters. Finding the moment of completion of the process itself is associated with the solution of this issue [31,32]. Usually, the decision on the number of clusters is made during the process of clustering, but sometimes before it starts (for example, when using the k-means method) [2,3].
At present, the problem of determining the number of clusters is open. For instance, Baxter and Everitt argue that the approach to establish the true number of clusters is to use subjective criteria based on expert judgment [3,33]. Nevertheless, cluster analysis remains one of the main methods of preliminary typology [34], and this necessitates the derivation of formal criteria for completing the clustering process and the rules for calculating the number of clusters.
In the overwhelming majority of modern works devoted to studying and solving these problems, the authors consider not general but various exceptional cases of clustering. We can highlight the article [35], which describes an algorithm based on finding and evaluating jumps of the so-called index functions. Developing the ideas presented in [35], it was proposed to use randomized algorithms for approximation of jumps of index functions [36] to find the number of clusters.
Quite often, the determination of the number of clusters during the execution of the hierarchical agglomerative clustering process is based on a visual analysis of dendrograms [37,38]. However, this approach is heuristic, but heuristic methods are based on some plausible assumptions and not on rigorous mathematical inferences.
Another heuristic approach to determining the preferred number of clusters under hierarchical agglomerative clustering is called the “elbow method” [39]. The main idea of this heuristic is that if the graph of some variable describing the clustering process resembles a hand, then the “elbow” (the point of the sharp bend in the graph) is a good indicator that we received of the preferred number of clusters. From a formal point of view, if a value increased linearly, its increase became nonlinear at the “elbow” point.
For the “elbow method” application, can be used the sequence of minimum distances between the elements of the set X. This sequence linearly ordered with respect to numeric values: 0 F 1 F 2 F m 1 . In Euclidean spaces, when clusters merge, there should be a sharp jump in the numerical value of the minimum distance, which is the moment of completion of the clustering process. In Figure 2 shows a graph of the sequence of minimum distances for the model set X. The figure shows that this jump (point F 30 ) is better approximated not by a linear function, but of a parabolic or exponential. Moreover, this point is the “elbow” of the graph.
Clustering study as a quasi-deterministic process, as its trajectories, one can consider monotonically changing numerical characteristics, including the sequence of minimum distances, where time is the iteration number of the agglomerative process. To construct statistical criteria for completing the clustering process, one can use quadratic forms of approximation-evaluative criteria, which are an analytical generalization of the heuristic of the “elbow method”. It is essential to emphasize that determining the number of clusters with their help is based not on heuristic conclusions but sequential statistical analysis.
Several clusters automatic determination is an urgent problem in many cases of preliminary typologization of empirical data. For example, in cytometric blood analysis [40], in automatic analysis of texts [41] both on topics [23,42], and by emotional coloring [43], as well in all other cases when the number of clusters is a priori unknown.

6.2. Limits of Application of Hooke’s Law

The main requirement for any structure is that it has strength, rigidity, and stability. The calculations of these parameters are based on experimental data and assumptions made within disciplines such as structural mechanics, the strength of materials, and elasticity theory.
The main goal of studying the properties of materials under the influence of external forces is to describe the relationship between stresses and strain. Sometimes strains are determined by stresses, and sometimes, conversely, stresses are determined by strain. The elastic deformation is described by Hooke’s law, which states that the stress is proportional to the strain, i.e., the relationship between them is linear [44,45].
Elastic deformation study is critical primarily since it is with it that any deformation process begins: plastic deformation, highly elastic deformation, or brittle fracture. The behavior in the elastic region is of great practical importance both for the brittle states of solids and for the plastic states of materials, for which elastic deformation has a significant effect on the development of inelastic processes.
The elastic properties of a solid are related to the nature of the adhesion forces, intermolecular bonds, etc. The nature of the interatomic forces indicates that Hooke’s law is approximate, and the directly proportional relationship between stresses and deformations is an idealized scientific abstraction. However, it has been shown experimentally that Hooke’s law is observed with sufficient accuracy for most materials, but only within certain limits. If the relationship between stresses and strains becomes nonlinear, then Hooke’s law becomes inapplicable [46].
The currently used graphical methods for determining the boundaries of the elastic zone using the stress strain curve are rather primitive. They are intended for a situation where stress is a function of deformation in the one-dimensional case [47]. Then the transition from elastic to plastic state is characterized by a change in the type of stress increase from linear to logarithmic or arctangential (Figure 3) [48].
However, cases where, on the contrary, stress is a function of deformation are also of great importance. Then the transition from elastic to plastic state is characterized by a change in the type of deformation growth from linear to parabolic or exponential (Figure 4) [49].
Various types of approximation-estimation tests can be used to determine the scope of Hooke’s law [22].

6.3. Singular Points of the Creep (Cold Flow) Curve

As a rule, creep (cold flow) is slow, occurring in time, deformation of a solid under the influence of a constant load. Cold flow is described by the so-called “creep curve”, which depends on deformation on time at constant temperature and force. Creep mechanisms depend both on the type of material and on the conditions under which it occurs. Its physical mechanism is predominantly of a diffusion nature, which makes it different from plastic deformation, which is associated with fast sliding along the atomic planes of grains of the polycrystal [50].
The creep curve is conventionally divided into three sections (stages). The first stage is a section of unsteady creep when the creep rate slows down. The second stage is a section of steady-state creep when deformation occurs at a constant rate. The third stage is an area of accelerated creep. Creep curves have the same form (Figure 5) for a wide range of materials: metals, alloys, semiconductors, polymers, ice, etc.
In the first stage of creep, the initial velocity, given by the instantaneous initial deformation, gradually decreases to a particular minimum value. At the second stage, creep occurs at a constant speed [50]. This piece of the curve can be compared to a graph of the locus of points, in which the relationship between the abscissa and the ordinate is first linear, then logarithmic, and then linear again.
We will call the moment of change in the nature of the increase in deformation from linear to logarithmic—“the first singular point of the creep curve”.
At the third stage, the growth of deformation occurs at an increasing rate and ends with the destruction of the material. The transition from the second stage to the third is characterized by a change in the rate of increase in deformation from linear to parabolic, or exponential [50].
We will call the moment of change in the nature of the increase in deformation from linear to parabolic or exponential—“the second singular point of the creep curve”.
Methods of analytical determination of singular points of the creep curve can be of practical and theoretical interest in both physical and computational experiments. A possible approach to solving this problem can be the use of approximation-estimation tests.

6.4. Characteristic Points for the Curves of the Real-Time PCR

Real-time polymerase chain reaction (real-time PCR, qPCR), based on the PCR method [51], allows you to determine the presence of the target nucleotide sequence in the sample and measure the number of its copies. The amount of amplified DNA is measured after each cycle of amplification using fluorescent labels. Evaluation can be quantitative (measuring the number of copies of the template) and relative (measuring relative to the introduced DNA or additional calibration genes). For real-time PCR, the amount of product formed is monitored during the reaction by monitoring the fluorescence of the dyes introduced into the reaction. The number of fluorophores is proportional to the amount of the resulting DNA product. Assuming a specific amplification efficiency, which is usually close to doubling the number of molecules per amplification cycle, it is possible to calculate the number of DNA molecules initially present in the sample. Thanks to highly efficient detection chemistry, sensitive instrumentation, and optimized analysis methods, DNA molecules can be quantified with unprecedented precision [52].
The fluorescence accumulation curve (fluorescence graph) for real-time PCR has a characteristic form (Figure 6); it consists of a baseline, an exponential phase, and a plateau phase [53].
For initial cycles, when the fluorescent signal is below the value that the instrument can register, the amplification graph slowly increases as a linear function. Then, as the product accumulates, the signal increases exponentially and then reaches a plateau, similar to an arctangent. The plateau is due to the lack of one or another component of the reaction. In a standard real-time PCR reaction, all samples will plateau and reach approximately the same signal level. On the other hand, in the exponential phase, differences in the growth rate of the amount of product can be traced. Differences in the initial number of molecules affect the number of cycles required to raise the fluorescence level above the noise level.
There are many mathematical models describing the fluorescence graph for real-time PCR [54,55]. In some cases, theoretical and practical interest is not a heuristic inference, but a formal determination of the moments of transition of the fluorescence accumulation curve from linear growth to exponential growth, and then, reaching the plateau [56,57].

6.5. Switched Systems

Continuous dynamics and discrete events interconnect many systems encountered in practice. Systems in which these two types of dynamics coexist and interact are usually called hybrid systems. Their study entails interesting theoretical problems that are important for solving many applied problems. Due to its interdisciplinary nature, this field attracts the attention of a wide range of scientists. Researchers with experience and interest in continuous-time systems and control theory are primarily interested in the properties of continuous dynamics, such as Lyapunov stability. At the same time, it is necessary to study the discrete behavior of such systems. To describe the specifics of discrete dynamics, it is helpful to analyze a more general situation that contains a specific model as a particular case. We can is done this by considering systems with continuous-time and discrete switching events from a particular class. Such systems are called switched systems and are considered higher-level abstractions than hybrid systems. It should be borne in mind that such systems often demonstrate non-trivial switching behavior and, thus, go beyond the traditional control theory [58].
For example, we can consider the security issues on the Internet of Things (IoT), which connects devices and users without limiting time and place. Latency, availability, and reliability are critical metrics for the efficient use of IoT data and services. IoT devices equipped with embedded controllers are often targeted by denial-of-service attacks. Denial-of-service (DoS) attacks and Distributed Denial-of-Service (DDoS) attacks are the most common attacks on the Internet of Things [59]. They can take many forms and are defined as attacks that can undermine the network’s ability to perform its function. During DoS and DDoS attacks, several “malicious” computer systems “flood” the selected server with a huge volume of concurrent requests and cause a denial of service for users and devices on the network [60].
In these cases, the system must change “its behavior” depending on the situation and use specialized hardware and software methods to deal with them. A possible approach, in this case, would be to use machine-learning methods in the IoT [61]. The fact is that random functions, which are mappings between the generalized time and the amount of information (the number of requests), at the beginning of the attacks change the character of their increase from linear to nonlinear form, as a rule—exponential (Figure 7). Approximation-estimation tests can be used to definition Markov stopping time corresponding to a change in strategy in response to an attack or overload.
The flexibility of the set of approximation-estimation tests allows them to make decisions on changing the control strategy based on any available metric. Network congestion control can be thought of as a switching system [62,63]. In this case, the primary mechanism for congestion avoidance is to determine the available system bandwidth. If the network is overloaded, it is necessary to automatically switch to emergency mode with a decrease in unacknowledged packets.
One of the most critical aspects of the considered tools is that the decision to change the strategy is made based on data on changes in the nature of network activity and not based on overcoming specific fixed load values. This allows these algorithms to be applied without additional configuration on systems of any size.

7. Conclusions

At the end of the article, it should be noted that modern computers do not learn anything. Typical machine learning boils down to finding and using “mathematical formulas” that produce the desired results when applied to a set of inputs. Just as artificial intelligence is not intelligence, machine learning is not learning. The term was coined for marketing reasons; in the 1960s, IBM used it to attract customers and talents [64]. Nevertheless, the term machine learning has gained worldwide acceptance and is understood as the theory and practice of creating hardware and software that can perform practical actions without direct human intervention.
The practical application of approximation-estimation tests for solid mechanics, molecular biology, and switching systems can be attributed to a particular case of detecting anomalies. The same applies to their use in cluster analysis; the definition of “elbow” on the graph identifies an anomaly. Specific implementations of the proposed formulas and algorithms and an assessment of their computational complexity are beyond the scope of this article, but there are no prerequisites to believe that this aspect will cause significant difficulties in implementation. Moreover, the implementation of algorithms using approximation-estimation tests is possible in both hardware and software.
Approximation-estimation tests and integral-estimation tests open up new purely mathematical problems. For example, we can be studying their extreme properties. Approximation-estimation tests let obtain an upper bound for the critical point of the transition of the trajectory of a quasi-deterministic process from linear to nonlinear growth. Moreover, if the trajectory is convex, then an increase in the approximation knots shifts this upper bound to the left. Integral and differential calculus tools provide more accurate results than discrete methods. Therefore, it can be assumed that integral-estimation tests allow one to obtain more accurate estimates for critical points of monotonic trajectories than approximation-estimation tests. The next problem, by the way, is related to the previous one. Behavior study of approximation-estimation tests and integral-estimation tests if an increasing variable has not one inflection point, such as a sigmoid or a logit, but a countable set of such points, etc.
And in the end, for the sake of fairness, it should still be noted that integral-estimation tests are primarily of academic interest since practical implementation in digital technologies is possible mainly for discrete solutions.

Funding

This research received no external funding.

Acknowledgments

The author acknowledges senior researcher, laboratory of “Photoactive Nanocomposite Materials”, of St. Petersburg State University, Artemev Yu.M. for help in preparing the paper.

Conflicts of Interest

The author declare no conflict of interest.

References

  1. Crane, M.G.; Nudelman, A.A. Markov Moment Problem and Extremal Problems; Fizmatlit: Moscow, Russia, 1973; p. 552. [Google Scholar]
  2. Hartigan, J.A. Clustering Algorithms; John Wiley & Sons: New York, NY, USA, 1975; p. 351. [Google Scholar]
  3. Everitt, B.S. Cluster Analysis; John Wiley & Sons Ltd.: Chichester, UK, 2011; p. 352. [Google Scholar]
  4. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  5. Ankerst, M.; Breunig, M.M.; Kriegel, H.-P.; Sander, J. OPTICS: Ordering Points To Identify the Clustering Structure. ACM Sigmod Rec. 1999, 28, 49–60. [Google Scholar] [CrossRef]
  6. Schubert, E.; Gertz, M. Improving the Cluster Structure Extracted from OPTICS Plots; LWDA: Mannheim, Germany, 2018; pp. 318–329.
  7. Charu, C.A.; Chandan, K.R. (Eds.) Data Clustering. Algorithms and Applications; Imprint Chapman and Hall/CRC: Boca Raton, FL, USA, 2014; p. 652. [Google Scholar]
  8. Pawlus, M.; Devine, R. Hands-On Deep Learning with R; Packt: Birmingham, UK, 2020; p. 330. [Google Scholar]
  9. van der Maaten, L.; Postma, E.; van den Herik, J. Dimensionality Reduction: A Comparative Review. J. Mach. Learn. Res. 2009, 10, 66–71. [Google Scholar]
  10. Zimek, A.; Schubert, E. Outlier Detection. In Encyclopedia of Database Systems; Springer: New York, NY, USA, 2017; pp. 1–5. [Google Scholar]
  11. Hodge, V.J.; Austin, J. A Survey of Outlier Detection Methodologies. Artif. Intell. Rev. 2004, 22, 85–126. [Google Scholar] [CrossRef] [Green Version]
  12. Dokas, P.; Ertoz, L.; Kumar, V.; Lazarevic, A.; Srivastava, J.; Tan, P.-N. Data mining for network intrusion detection. In Proceedings of the NSF Workshop on Next Generation Data Mining, Marriott, Inner Harbor, MD, USA, 1–3 November 2002; Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.331.6701&rep=rep1&type=pdf (accessed on 27 July 2021).
  13. Krishnan, V. Probability and Random Processes; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2006; p. 723. [Google Scholar]
  14. Ibe, O. Fundamentals of Applied Probability and Random Processes; Academic Press: San Diego, CA, USA, 2014; p. 456. [Google Scholar]
  15. Bulinsky, A.V.; Shiryaev, A.N. Theory of Random Processes; Fizmatlit: Moscow, Russia, 2003; p. 400. [Google Scholar]
  16. Levin, B.R. Theoretical Foundations of Statistical Radio Engineering; Radio and Communication: Moscow, Russia, 1989; p. 656. [Google Scholar]
  17. Lehmann, E.L.; Romano, J.P. Testing Statistical Hypotheses; Springer: New York, NY, USA, 2005; p. 786. [Google Scholar]
  18. Wald, A. Sequential Analysis; John Wiley & Sons: New York, NY, USA, 1947; p. 212. [Google Scholar]
  19. Chow, Y.S. Great Expectations: The Theory of Optimal Stopping, 1st ed.; Houghton Mifflin: Boston, MA, USA, 1971; p. 139. [Google Scholar]
  20. Mazalov, V.V. Mathematical Game Theory and Applications; Publishing House “Lan”: St. Petersburg, Russia, 2017; p. 448. [Google Scholar]
  21. Shiryaev, A.N. Optimal Stopping Rules; Springer: Berlin/Heidelberg, Germany, 2008; p. 220. [Google Scholar]
  22. Orekhov, A.V. Statistical criteria for the limits of application of Hooke’s law. Vestn. St.-Peterbg. Univ. Prikl. Mat. Inform. Protsessy Upr. 2020, 16, 391–401. [Google Scholar]
  23. Bodrunova, S.S.; Orekhov, A.V.; Blekanov, I.S.; Lyudkevich, N.S.; Tarasov, N.A. Topic Detection Based on Sentence Embeddings and Agglomerative Clustering with Markov Moment. Future Internet 2020, 12, 144. [Google Scholar] [CrossRef]
  24. Orekhov, A.V. Criterion for estimation of stress-deformed state of SD-materials. In AIP Conference Proceedings; AIP Publishing LLC: College Park, MD, USA, 2018; Volume 1959, p. 70028. [Google Scholar]
  25. Orekhov, A.V. Approximation-evaluation tests for a stress-strain state of deformable solids. Vestn. St.-Peterbg. Univ. Prikl. Mat. Inform. Protsessy Upr. 2018, 14, 230–242. [Google Scholar] [CrossRef]
  26. Bellman, R. A Markovian decision process. J. Math. Mech. 1957, 6, 679–684. [Google Scholar] [CrossRef]
  27. Howard, R.A. Dynamic Programming and Markov Processes, 1st ed.; The MIT Press: Cambridge, MA, USA, 1960; p. 136. [Google Scholar]
  28. Chung, K.L. Lectures from Markov Processes to Brownian Motion; Springer: New York, NY, USA, 1982; p. 242. [Google Scholar]
  29. Sheng-Jhih, W.; Moody, T.C. Markov chains with memory, tensor formulation, and the dynamics of power iteration. Appl. Math. Comput. 2017, 303, 226–239. [Google Scholar]
  30. Dwight, H.B. Tables of Integrals and other Mathematical Data, 4th ed.; Macmillan Company: New York, NY, USA, 1966; p. 336. [Google Scholar]
  31. Orekhov, A.V. Markov stopping time of an agglomerative clustering process in Euclidean space. Vestn. St.-Peterbg. Univ. Prikl. Mat. Inform. Protsessy Upr. 2019, 15, 76–92. [Google Scholar]
  32. Orekhov, A.V. Agglomerative Method for Texts Clustering. In Proceedings of the 5th International Conference on Internet Science (INSCI 2018), St. Petersburg, Russia, 24–26 November 2018; Bodrunova, S., Ed.; Lecture Notes in Computer Science (LNCS). Springer: Cham, Switzerland, 2019; Volume 11551, pp. 19–32. [Google Scholar]
  33. Baxter, M.J. Exploratory Multivariate Analysis in Archaeology; Edinburgh University Press: Edinburgh, UK, 1994; 307p. [Google Scholar]
  34. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; John Wiley & Sons Ltd.: New York, NY, USA, 2000; p. 688. [Google Scholar]
  35. Sugar, C.A.; James, G.M. Finding the number of clusters in a dataset. J. Am. Stat. Assoc. 2003, 98, 750–763. [Google Scholar] [CrossRef]
  36. Granichin, O.N.; Shalymov, D.S.; Avros, R.; Volkovich, Z. A randomized algorithm for estimating the number of clusters. Autom. Rem. Contr. 2011, 72, 754–765. [Google Scholar] [CrossRef]
  37. Calirnski, T.; Harabasz, J. A dendrite method for cluster analysis. Commun. Stat. 1974, 3, 1–27. [Google Scholar]
  38. Aldenderfer, M.S.; Blashfield, R.K. Cluster Analysis: Quantitative Applications in the Social Sciences; SAGE Publications, Inc.: Beverly Hills, CA, USA, 1984; p. 88. [Google Scholar]
  39. Thorndike, R.L. Who Belongs in the Family? Psychometrika 1953, 18, 267–276. [Google Scholar] [CrossRef]
  40. Zurochka, A.V.; Khaydukov, S.V.; Kudryavtsev, I.V.; Chereshnev, V.A. Flow Cytometry in Medicine and Biology, 2nd ed.; Ural Branch of the Russian Academy of Sciences Publ.: Yekaterinburg, Russia, 2014; p. 574. [Google Scholar]
  41. Lappin, S.; Fox, C. The Handbook of Contemporary Semantic Theory, 2nd ed.; Wiley-Blackwell: Malden, MA, USA, 2015; p. 776. [Google Scholar]
  42. Orekhov, A.V.; Kharlamov, A.A.; Bodrunova, S.S. Network presentation of texts and clustering of messages. In Proceedings of the 6th International Conference on Internet Science, Perpignan, France, 2–5 December 2019; El Yacoubi, S., Bagnoli, F., Pacini, G., Eds.; Lecture Notes in Computer Science (LNCS). Springer: Cham, Switzerland, 2019; Volume 11938, pp. 235–249. [Google Scholar]
  43. Kharlamov, A.A.; Orekhov, A.V.; Bodrunova, S.S.; Lyudkevich, N.S. Social Network Sentiment Analysis and Message Clustering. In Proceedings of the 6th International Conference on Internet Science, Perpignan, France, 2–5 December 2019; El Yacoubi, S., Bagnoli, F., Pacini, G., Eds.; Lecture Notes in Computer Science (LNCS). Springer: Cham, Switzerland, 2019; Volume 11938, pp. 18–31. [Google Scholar]
  44. Timoshenko, S. Strength of Materials, 3rd ed.; Krieger Pub Co.: Malabar, FL, USA, 1983; p. 1010. [Google Scholar]
  45. Beer, F.; Russell Johnston, E., Jr.; De Wolf, J.; Mazurek, D. Mechanics of Materials, 7th ed.; McGraw-Hill Education: New York, NY, USA, 2014; p. 896. [Google Scholar]
  46. Friedman, Y.B. Mechanical Properties of Metals. Part 1. The Deformation and Fracture; Mechanical Engineering: Moscow, Russia, 1974; p. 472. [Google Scholar]
  47. Atlas of Stress-strain Curves; ASM International: Geauga County, OH, USA, 2002; p. 816.
  48. Schneider, S.; Schneider, S.G.; da Silva, H.M.; de Moura Neto, C. Study of the non-linear stress-strain behavior in Ti-Nb-Zr alloys. Mat. Res. 2005, 8, 435–438. [Google Scholar] [CrossRef] [Green Version]
  49. Pavilaynen, G.V.; Yushin, R.U. An approximate solution of elastic-plastic problem of circular strength different (SD) plates. In Proceedings of the 2017 Constructive Nonsmooth Analysis and Related Topics (Dedicated to the Memory of V.F. Demyanov), CNSA 2017, St. Petersburg, Russia, 22–27 May 2017. [Google Scholar]
  50. Rabotnov, Y.N. Mechanics of a Deformable Solid; Nauka: Moscow, Russia, 1979; p. 743. [Google Scholar]
  51. Higuchi, R.; Dollinger, G.; Walsh, P.S.; Griffith, R. Simultaneous Amplification and Detection of Specific DNA Sequences. Nat. Biotechnol. 1992, 10, 413–417. [Google Scholar] [CrossRef]
  52. Provenzano, M.; Mocellin, S. Complementary techniques: Validation of gene expression data by quantitative real time PCR. Adv. Exp. Med. Biol. 2007, 593, 66–73. [Google Scholar]
  53. Kubista, M.; Andrade, J.M.; Bengtsson, M.; Forootan, A.; Jonák, J.; Lind, K.; Sindelka, R.; Sjöback, R.; Sjxoxgreen, B.; Strxoxmbom, L.; et al. The real-time polymerase chain reaction. Mol. Asp. Med. 2006, 27, 95–125. [Google Scholar] [CrossRef]
  54. Gevertz, J.L.; Dunn, S.M.; Roth, C.M. Mathematical model of real-time PCR kinetics. Biotechnol. Bioeng. 2005, 92, 346–355. [Google Scholar] [CrossRef] [Green Version]
  55. Rebrikov, D.V.; Trofimov, D.Y. Real-time PCR: Approaches to data analysis (Review). Appl. Biochem. Microbiol. 2006, 42, 520–528. [Google Scholar] [CrossRef]
  56. Rutledge, R.G.; Cote, C. Mathematics of quantitative kinetic PCR and the application of standard curves. Nucleic Acids Res. 2003, 31, e93. [Google Scholar] [CrossRef] [PubMed]
  57. Liu, W.; Saint, D.A. A new quantitative method of real time reverse transcription polymerase chain reaction assay based on simulation of polymerase chain reaction kinetics. Anal. Biochem. 2002, 302, 52–59. [Google Scholar] [CrossRef] [Green Version]
  58. Liberzon, D. Switching in Systems and Control; Birkhäuser: Boston, MA, USA, 2003; p. 246. [Google Scholar]
  59. Daferovic, E.; Sokol, A.; Almisreb, A.A.; Mohd Norzeli, S. DoS and DDoS vulnerability of IoT: A review. Sustain. Eng. Innov. 2019, 1, 43–48. [Google Scholar] [CrossRef] [Green Version]
  60. Chen, Q.; Chen, H.; Cai, Y.; Zhang, Y.; Huang, X. Denial of Service Attack on IoT System. In Proceedings of the 2018 9th International Conference on Information Technology in Medicine and Education (ITME), Hangzhou, China, 19–21 October 2018. [Google Scholar]
  61. Piccialli, F.; Casolla, G.; Cuomo, S.; Giampaolo, F.; di Cola, V.C. Decision Making in IoT Environment through Unsupervised Learning. IEEE Intell. Syst. 2020, 35, 27–35. [Google Scholar] [CrossRef]
  62. Shorten, R.; Wirth, F.; Mason, O.; Wulff, K.; King, C. Stability criteria for switched and hybrid systems. SIAM Rev. 2007, 49, 545–592. [Google Scholar] [CrossRef] [Green Version]
  63. Hespanha, J.P. Stochastic hybrid systems: Application to communication networks. In Hybrid Systems: Computation and Control. HSCC 2004; Alur, R., Pappas, G.J., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2004; Volume 2993. [Google Scholar]
  64. Burkov, A. The Hundred-Page Machine Learning Book; Publisher Andriy Burkov: Quebec, QC, Canada, 2019; p. 160. [Google Scholar]
Figure 1. Numerical-graphic solution of the equation I 1 ( x ) I 2 ( x ) = 0 .
Figure 1. Numerical-graphic solution of the equation I 1 ( x ) I 2 ( x ) = 0 .
Mathematics 09 02301 g001
Figure 2. Graph of F i values, the abscissa represents the iteration numbers, the ordinate represents the sequence of minimum distances.
Figure 2. Graph of F i values, the abscissa represents the iteration numbers, the ordinate represents the sequence of minimum distances.
Mathematics 09 02301 g002
Figure 3. Sketch of a stress–strain curve, on the abscissa—strain, on the ordinate—stress values.
Figure 3. Sketch of a stress–strain curve, on the abscissa—strain, on the ordinate—stress values.
Mathematics 09 02301 g003
Figure 4. Sketch of the inverse strain diagram (strain–stress curve), on the abscissa—stress values, on the ordinate—strain.
Figure 4. Sketch of the inverse strain diagram (strain–stress curve), on the abscissa—stress values, on the ordinate—strain.
Mathematics 09 02301 g004
Figure 5. Classic view of the creep curve (according to Rabotnov), on the abscissa—time, on the ordinate—strain values.
Figure 5. Classic view of the creep curve (according to Rabotnov), on the abscissa—time, on the ordinate—strain values.
Mathematics 09 02301 g005
Figure 6. Fluorescence accumulation curve (fluorescence graph) for real-time PCR, abscissa—amplification cycles, ordinate—fluorescence brightness.
Figure 6. Fluorescence accumulation curve (fluorescence graph) for real-time PCR, abscissa—amplification cycles, ordinate—fluorescence brightness.
Mathematics 09 02301 g006
Figure 7. Sketch of the network activity curve during DoS or DDoS attacks, on the abscissa—generalized time, on the ordinate—the number of requests.
Figure 7. Sketch of the network activity curve during DoS or DDoS attacks, on the abscissa—generalized time, on the ordinate—the number of requests.
Mathematics 09 02301 g007
Table 1. Tangential test for the parabolic approximation-estimation test.
Table 1. Tangential test for the parabolic approximation-estimation test.
Step y 0 y 2 y 0 y 3 y 0 y 4 y 0 y 5 y 0 y 6
0.1 1.3 1.2 1.1 1.1 1.0
0.09 1.35 1.26 1.17 1.08 0.99
0.08 1.36 1.28 1.2 1.12 1.04
0.07 1.4 1.33 1.19 1.12 1.12
0.06 1.44 1.32 1.26 1.2 1.14
0.05 1.45 1.35 1.3 1.25 1.2
0.04 1.48 1.4 1.36 1.28 1.24
0.03 1.5 1.44 1.41 1.35 1.32
0.02 1.52 1.48 1.46 1.42 1.4
0.01 1.55 1.53 1.51 1.5 1.48
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Orekhov, A.V. Quasi-Deterministic Processes with Monotonic Trajectories and Unsupervised Machine Learning. Mathematics 2021, 9, 2301. https://doi.org/10.3390/math9182301

AMA Style

Orekhov AV. Quasi-Deterministic Processes with Monotonic Trajectories and Unsupervised Machine Learning. Mathematics. 2021; 9(18):2301. https://doi.org/10.3390/math9182301

Chicago/Turabian Style

Orekhov, Andrey V. 2021. "Quasi-Deterministic Processes with Monotonic Trajectories and Unsupervised Machine Learning" Mathematics 9, no. 18: 2301. https://doi.org/10.3390/math9182301

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop