1. Introduction
The keys to evaluating an experimental result—e.g., compare it with the result anticipated by theories—require first the right selection of potential statistical tools and techniques for correctly processing and analyzing this result. This “processing and analyzing” involves two general types of approximation problems: One problem concerns a function fitting to given set of data. The other problem arises when a function is given analytically by an explicit mathematical type but we would like to find an alternative function with simpler form.
Let
, with
and
, denote a multi-parametrical approximating function [
1,
2,
3,
4,
5,
6], symbolized as
, for short.
The widely used, traditional fitting method of least squares involves minimizing the sum of the squares of the residuals, i.e., the squares of the differences between the function and the approximating function that represents the statistical model, . However, the least-square method is not unique. For instance, the absolute deviations minimization can also be applied. Generally, as soon as the desired norm of the metric space is given, the respective method of deviations minimization is defined. The least-square method is based on the Euclidean norm, while the alternative absolute deviations method is based on the uniform or Taxicab norm. In general, an infinite number of fitting methods can be defined, based on the metric space induced by the Lq-norm; this case is studied here in detail.
Given the metric induced by the
Lq-norm, the functional of the total
Lq-normed residuals [
7,
8,
9,
10,
11,
12], noted also as total deviations (TD), between the fixed
and the approximating
functions in the domain
D, is given by:
The functional of total deviations,
, is expanded (Taylor series) near its local minimum:
where
is the total deviation function at its global minimum, while
is the Hessian matrix at this minimum, where all the components are positive, i.e.,
,
.
By expanding the approximating function
near the TD’s minimum, [
7] showed the following equations:
and
where
The normal equations are given by
where we set
for short.
The purpose of this paper is to present the geometric interpretation of the errors of the optimal parameter values, derived from a multi-parametrical fitting, based on a metric induced by the non-Euclidean
Lq-norm. In
Section 2, we derive the smallest possible value of the variation of the total deviations from its minimum,
, also called, the error of the total deviations value. In
Section 3, we describe the geometric interpretation of the errors of the optimal parameter values, while in
Section 4, we use this geometry to derive the exact equations that provide these errors. In
Section 5, we apply the developed formulation for the 1-dim and 2-dim cases. Finally,
Section 6 summarizes the conclusions.
2. The Error of the Total Deviation Values
The total deviations functional, , has a minimum value . The difference between these functionals cannot be arbitrarily small. Here we derive the smallest possible value of the variation of the total deviations from its minimum, , also called, the error of the total deviations value.
First, we mention that the transition of the continuous to the discrete way for describing the values of
x, can be realized as follows:
while the expression of the total deviations is given by
for large values of
N, where
L is the total length of the domain
D, and the resolution of
x-values is
. In the discrete case, it is sufficient to express the total deviations simply by
where we set
,
and
.
Then, we calculate the error of the total deviations values, δTD, near the local minimum of , that is, for , . Thus,
In the case of a large number of sampling elements, we adopt the continuous description, i.e.,
and
is the distribution of
u-values in their domain
, that is,
since
x-values are equidistributed in their domain
D. Therefore,
where the number of the sampling elements,
N, can be varied by 1, thus
δN = 1. Hence,
where
Moreover, we show that the far right part of Equation (12) is zero. Indeed:
where
and thus, we obtain:
leading to the set of the following
n normal equations:
Similarly, for the continuous way of
x-values, we have:
The result of Equation (20d) will be used in
Section 4 on the expression of the optimal errors.
3. The Uncertainty Manifold
We define the deviation of the total deviations functional from its minimum,
, which is expressed with the quadratic form:
where we set
,
.
Given a particular value of
, each of these parameter deviations, e.g., the
k-th component
, has a maximum value
. This maximum value
of each parameter deviation
, depends on the value of
. The smallest possible value of
is deduced when
also reaches its smallest value. The smallest possible value of
interprets the error
of the optimal parameter values
,
; this is achieved when the particular value
is given by the smallest possible value of a deviation from the
TD’s minimum,
. In
Section 2, we showed that
equals:
There are cases, where the total deviations value is subject to an experimental, reading, or any other type of a non-statistical error; this is, in general, called the resolution value . Then, the smallest possible value is meaningful only when it stays above the threshold of ; in other words, or, if , then . Hence,
The quadratic form in Equation (21) is positive definite, and thus it defines an
n-dimensional paraboloid (hypersurface with a local minimum) immersed into an (
n+1)-dimensional space. The corresponding
n + 1 axes are given by the
n parameter deviations
and the deviation
, describing thus, the (
n+1)-dimensional space as
where
is the domain of the deviation values,
.
Given a fixed value of , and that can be the value of the smallest deviation, i.e., , the set of the parameter deviations defines a locus of an n-dimensional ellipsoid, rotated with respect to the axes . This n-dimensional ellipsoid is bounded by the (n − 1)-dimensional locus of intersection between the n-dimensional paraboloid and the n-dimensional hyperplane .
The n-dimensional ellipsoid is called uncertainty manifold, denoted by , for short. This is a manifold with an edge, meaning thus, its boundary, denoted by . In general, the edge of an n-dimensional manifold is an (n − 1)-dimensional manifold. Here, the edge involves the (n − 1)-dimensional locus of intersection between the n-dimensional paraboloid and the n-dimensional hyperplane . The n-dimensional cuboid, which encloses the uncertainty manifold’s edge , is also a manifold with an edge and is denoted by . Its edge is an (n − 1)-dimensional manifold denoted by .
For example, consider the case of two-parametrical approximating functions,
. Then, the quadratic form of Equation (21) defines the two-dimensional paraboloid
, immersed into the three-dimensional space with Cartesian axes given by
. The two-dimensional ellipsoid is defined by the space bounded by the locus
, which is the intersection of the two-dimensional paraboloid
and the two-dimensional hyperplane
. For visualizing this example, see
Figure 1.
Next, we will use the concept of the hyper-dimensional uncertainty manifold to derive the expressions of the errors of the optimal parameter values.
4. Derivation of the Errors of the Optimal Parameter Values
The expressions of the errors of the optimal parameter values—or simply, optimal errors—are well-known in the case of the least-square and other Euclidean based fitting methods. In [
7], we have used the error expression, which is caused by the curvature, in order to have an estimate of the optimal errors (for applications, see [
11,
12,
13,
14,
15,
16,
17,
18]). Here, we will see the formal geometric derivation of the optimal error expressions.
First, we note that the edge of the uncertainty manifold, , has a number of n extrema, denoted by , and they are related to the errors of the parameters optimal values, , as follows:
The position vector of the corresponding point , , consists of n components each, i.e., . Thus,
These components are given by the condition:
, with
.
The above (
n − 1) equations, given in Equation (26), together with
are sufficient for the calculation of the
n unknown components of
.
The (n − 1) equations, given in Equation (26), arise from the fact that each of the errors of parameters optimal values, i.e., the k-th, is derived from the maximum value of the corresponding component , that is, . Hence, the existence of maximum values , i.e., , : , , leads to the errors estimation, . These maximum values, are located on the edge of the uncertainty manifold , that is the hypersurface , i.e., .
The maximization of the μ-th parameter deviation within uncertainty manifold is derived as follows: Since , we can express in terms of , with , i.e., . Then, the procedure of finding the maximum value of , that is, , involves finding all the derivatives , with of , or equivalently, of (implicit derivatives). The (n − 1) equations lead to the (n − 1) relationships , with , which together with , leads to the specific values of . Yet, only the μ-th component gives the error , i.e.,
In
Appendix A, we solve Equations (26) and (27), where, we concluded that
or
and thus, from Equation (28), we finally derive the errors:
.
Finally, taking into account the resolution of each parameter value,
, we have:
where
.
5. Formulation of the Cases of n = 1 and n = 2 Dimensional Uncertainty Manifold
5.1. The Case of n = 1
Let us begin with the case of a one-dimensional paraboloid, given simply by the parabola
corresponding to uni-parametrical approximating functions. The locus of intersection between this parabola and the line
(that is, the one-dimensional hyperplane) are the two points
. The uncertainty manifold
is the one-dimensional ellipsoid, defined by the line segment
, which is enclosed by the points
. In this case, the edge of the uncertainty manifold
is restricted to the zero-dimensional space composed only by the two points
. The manifolds
and
coincide (similarly with their edges,
and
, respectively). Hence,
5.2. The Case of n = 2
The case of bi-parametrical approximating functions is characterized by the two-dimensional paraboloid,
which is illustrated in
Figure 1. The locus of intersection between this paraboloid and the plane
is given by the rotated ellipse:
written suitably as
after the rotation transformation
where
Then,
where the diagonal matrix
has the following elements
which are the eigenvalues of the matrix
. The ellipsis’ major/minor axes in Equation (35) are:
while the rotation angle
θ in Equation (36) is given by
The uncertainty manifold
is the rotated 2-dim ellipsoid in the
axes, defined by:
or, in the rotated axes
, is simply given by:
which is enclosed by the ellipse corresponding to the equal sign of Equation (45), that is the edge of the uncertainty manifold,
.
Finally, the errors are
and
6. Conclusions
The paper presented the geometric interpretation of the errors of the optimal parameter values, derived from a multi-parametrical fitting, based on a metric induced by the non-Euclidean Lq-norm. Typical fitting methods are mostly developed based on Euclidean norms, leading to the traditional least–square method. On the other hand, the theory of general fitting methods based on non-Euclidean norms, is still under development; the normal equations can provide the optimal values of the fitting parameters, while this paper completed the puzzle by improving understanding the derivations and geometric meaning of the errors.
In particular, we showed that the statistical errors of the optimal parameter values are given by the axes of the ellipsoid called uncertainty manifold, that is, the intersection of the paraboloid of the residuals’ expansion along the deviations , with the hyperplane The constant represents the smallest possible value of a deviation from the TD’s minimum, also mentioned as an error of the value of the total deviations.
In summary, the
Lq-normed fitting involves minimizing:
where
The normal equations are given by:
where we set
.
Finally, we summarize the concluding relationships of the paper:
with special cases: