Next Article in Journal
Confidence Sets for Statistical Classification (II): Exact Confidence Sets
Previous Article in Journal
Linear Regression with Optimal Rotation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geometric Interpretation of Errors in Multi-Parametrical Fitting Methods Based on Non-Euclidean Norms

by
George Livadiotis
Southwest Research Institute, Space Science & Engineering, San Antonio, TX 78238, USA
Stats 2019, 2(4), 426-438; https://doi.org/10.3390/stats2040029
Submission received: 29 September 2019 / Revised: 20 October 2019 / Accepted: 25 October 2019 / Published: 29 October 2019

Abstract

:
The paper completes the multi-parametrical fitting methods, which are based on metrics induced by the non-Euclidean Lq-norms, by deriving the errors of the optimal parameter values. This was achieved using the geometric representation of the residuals sum expanded near its minimum, and the geometric interpretation of the errors. Typical fitting methods are mostly developed based on Euclidean norms, leading to the traditional least–square method. On the other hand, the theory of general fitting methods based on non-Euclidean norms is still under development; the normal equations provide implicitly the optimal values of the fitting parameters, while this paper completes the puzzle by improving understanding the derivations and geometric meaning of the optimal errors.

1. Introduction

The keys to evaluating an experimental result—e.g., compare it with the result anticipated by theories—require first the right selection of potential statistical tools and techniques for correctly processing and analyzing this result. This “processing and analyzing” involves two general types of approximation problems: One problem concerns a function fitting to given set of data. The other problem arises when a function is given analytically by an explicit mathematical type but we would like to find an alternative function with simpler form.
Let V ( x ; p 1 , p 2 , , p n ) , with x D and ( p 1 , p 2 , , p n ) { D p 1 D p 2 D p n } n , denote a multi-parametrical approximating function [1,2,3,4,5,6], symbolized as V ( x ; { p k } ) , for short.
The widely used, traditional fitting method of least squares involves minimizing the sum of the squares of the residuals, i.e., the squares of the differences between the function f ( x ) and the approximating function that represents the statistical model, V ( x ) . However, the least-square method is not unique. For instance, the absolute deviations minimization can also be applied. Generally, as soon as the desired norm of the metric space is given, the respective method of deviations minimization is defined. The least-square method is based on the Euclidean norm, while the alternative absolute deviations method is based on the uniform or Taxicab norm. In general, an infinite number of fitting methods can be defined, based on the metric space induced by the Lq-norm; this case is studied here in detail.
Given the metric induced by the Lq-norm, the functional of the total Lq-normed residuals [7,8,9,10,11,12], noted also as total deviations (TD), between the fixed f ( x ) and the approximating V ( x ; { p k } ) functions in the domain D, is given by:
T D q ( { p k } ) q = x D | V ( x ; { p k } ) f ( x ) | q d x   .  
The functional of total deviations, T D q ( { p k } ) q , is expanded (Taylor series) near its local minimum:
T D q ( { p k } ) q = A 0 ( q ) + k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) ( p k 1 p k 1 * ) ( p k 2 p k 2 * ) + O ( | p k p k * | 3 )   ,
where
A 0 ( q ) = T D q ( { p k * } ) q   ,  
is the total deviation function at its global minimum, while
A 2 , k 1 k 2 ( q ) = 1 2 2 p k 1 p k 2 T D q ( { p k } ) q | { p k } = { p k * }   ,  
is the Hessian matrix at this minimum, where all the components are positive, i.e., A 0 , A 2 , k 1 k 2 0 , k 1 , k 2 = 1 , 2 , , n .
By expanding the approximating function V ( x ; { p k } ) near the TD’s minimum, [7] showed the following equations:
A 0 ( q ) = x D | u | q d x   ,  
and
A 2 , k 1 k 2 ( q ) = δ 1 q γ k 1 k 2 + q 2 x D [ | u | q 1 sgn ( u ) 2 V ( x ; { p k * } ) p k 1 p k 2 + ( q 1 ) | u | q 2 V ( x ; { p k * } ) p k 1 V ( x ; { p k * } ) p k 2 ] d x ,
where
γ k 1 k 2 i : u ( x i ) = 0 1 | u ( x i ) | V ( x i ; { p k * } ) p k 1 V ( x i ; { p k * } ) p k 2   .  
The normal equations are given by
x D | u | q 1 sgn ( u ) V ( x ; { p k * } ) p l d x = 0 ,   l = 1 , 2 , , n   .  
where we set u = u ( x ) V ( x ; { p k * } ) f ( x ) for short.
The purpose of this paper is to present the geometric interpretation of the errors of the optimal parameter values, derived from a multi-parametrical fitting, based on a metric induced by the non-Euclidean Lq-norm. In Section 2, we derive the smallest possible value of the variation of the total deviations from its minimum, δ T D , also called, the error of the total deviations value. In Section 3, we describe the geometric interpretation of the errors of the optimal parameter values, while in Section 4, we use this geometry to derive the exact equations that provide these errors. In Section 5, we apply the developed formulation for the 1-dim and 2-dim cases. Finally, Section 6 summarizes the conclusions.

2. The Error of the Total Deviation Values

The total deviations functional, T D a ( { p k } ) q , has a minimum value A 0 ( q ) . The difference between these functionals cannot be arbitrarily small. Here we derive the smallest possible value of the variation of the total deviations from its minimum, δ T D , also called, the error of the total deviations value.
First, we mention that the transition of the continuous to the discrete way for describing the values of x, can be realized as follows:
a b ( ) d x = lim N b a N i = 1 N ( )   ,  
while the expression of the total deviations is given by
T D q ( { p k } ) q = x D | V ( x ; { p k } ) f ( x ) | q d x L N i = 1 N | V ( x i ; { p k } ) f ( x i ) | q = x r e s i = 1 N | V ( x i ; { p k } ) f ( x i ) | q ,
for large values of N, where L is the total length of the domain D, and the resolution of x-values is x r e s = L / N . In the discrete case, it is sufficient to express the total deviations simply by
T D q ( { p k } ) q = i = 1 N | u i | q   ,  
where we set u i u ( x i ; { p k } ) V ( x i ; { p k } ) f ( x i ) , V i V ( x i ; { p k } ) and y i f ( x i ) .
Then, we calculate the error of the total deviations values, δTD, near the local minimum of T D q ( { p k } ) q , that is, for { p k } = { p k * } , k = 1 , 2 , , n . Thus,
δ T D = [ δ T D q ( { p k } ) q ] | { p k } = { p k * } = [ δ i = 1 N | u i | q ] | { p k } = { p k * }   .  
In the case of a large number of sampling elements, we adopt the continuous description, i.e.,
i = 1 N | u i | q N | u | q   ,   with   | u | q = u D u | u | q S ( u ) d u   ,
and S ( u ) is the distribution of u-values in their domain D u , that is,
| u | q = 1 L x D | V ( x ; { p k } ) f ( x ) | q d x   ,  
since x-values are equidistributed in their domain D. Therefore,
δ i = 1 N | u i | q δ N | u | q + N δ | u | q = | u | q + N δ | u | q 1 N i = 1 N | u i | q + N δ | u | q ,  
where the number of the sampling elements, N, can be varied by 1, thus δN = 1. Hence,
[ δ i = 1 N | u i | q ] | { p k } = { p k * } 1 N [ i = 1 N | u i | q ] | { p k } = { p k * } + N [ δ | u | q ] | { p k } = { p k * } ,  
where
[ i = 1 N | u i | q ] | { p k } = { p k * } = A 0 ( q ) .
Moreover, we show that the far right part of Equation (12) is zero. Indeed:
δ | u | q = 1 L δ x D | V ( x ; { p k } ) f ( x ) | q d x = 1 L k = 1 n A 1 , k ( q ) δ p k   ,  
where
A 1 , k ( q ) = q x D | V ( x ; { p k } ) f ( x ) | q 1 s i g n [ V ( x ; { p k } ) f ( x ) ] V ( x ; { p k } ) p k d x   ,  
and thus, we obtain:
[ δ | u | q ] | { p k } = { p k * } = 1 L k = 1 n A 1 , k ( q ) | { p k } = { p k * } δ p k * = 0   ,  
leading to the set of the following n normal equations:
0 = A 1 , k ( q ) | { p k } = { p k * } = q x D | V ( x ; { p k * } ) f ( x ) | q 1 s i g n [ V ( x ; { p k * } ) f ( x ) ] V ( x ; { p k } ) p k | { p k } = { p k * } d x .  
Finally,
[ δ i = 1 N | u i | q ] | { p k } = { p k * } = 1 N A 0 ( q )   ,  
hence:
δ T D ( q ) = 1 N A 0 ( q )   .  
Similarly, for the continuous way of x-values, we have:
T D q ( { p k } ) q = x D | u | q d x x r e s i = 1 N | u i | q   ,  
δ T D x r e s [ δ i = 1 N | u i | q ] | { p k } = { p k * } 1 N x r e s [ i = 1 N | u i | q ] | { p k } = { p k * } = 1 N A 0 ( q )   ,  
A 0 ( q ) = [ x D | u | q d x ] | { p k } = { p k * } x r e s [ i = 1 N | u i | q ] | { p k } = { p k * }   ,  
δ T D ( q ) = 1 N A 0 ( q ) x r e s L A 0 ( q )   .  
The result of Equation (20d) will be used in Section 4 on the expression of the optimal errors.

3. The Uncertainty Manifold

We define the deviation of the total deviations functional from its minimum, Δ T D T D q ( { p k } ) q T D q ( { p k * } ) q > 0 , which is expressed with the quadratic form:
Δ T D ( { δ p k } ) = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ p k 2 ,  
where we set δ p k p k p k * , k = 1 , , n .
Given a particular value of Δ T D , each of these parameter deviations, e.g., the k-th component δ p k , has a maximum value δ p k , max . This maximum value δ p k , max of each parameter deviation δ p k , depends on the value of Δ T D . The smallest possible value of δ p k , max is deduced when Δ T D also reaches its smallest value. The smallest possible value of δ p k , max interprets the error δ p k * of the optimal parameter values p k * , k = 1 , 2 , , n ; this is achieved when the particular value Δ T D is given by the smallest possible value of a deviation from the TD’s minimum, δ T D . In Section 2, we showed that δ T D equals:
δ T D ( q ) = 1 N A 0 ( q )   .
There are cases, where the total deviations value is subject to an experimental, reading, or any other type of a non-statistical error; this is, in general, called the resolution value T r e s . Then, the smallest possible value δ T D is meaningful only when it stays above the threshold of T r e s ; in other words, δ T D T r e s or, if A 0 / N T r e s , then δ T D = T r e s . Hence,
δ T D ( q ) = { 1 N A 0 ( q ) ,        if     A 0 N > T r e s    , T r e s ,       if     A 0 N T r e s    .
The quadratic form in Equation (21) is positive definite, and thus it defines an n-dimensional paraboloid (hypersurface with a local minimum) immersed into an (n+1)-dimensional space. The corresponding n + 1 axes are given by the n parameter deviations { δ p k } k = 1 n and the deviation Δ T D , describing thus, the (n+1)-dimensional space as
( δ p 1 , δ p 2 , , δ p n ; Δ T D ) { D p 1 D p 2 D p n } D Δ T D n + 1 ,  
where D Δ T D = { Δ T D δ T D > 0 \ Δ T D } is the domain of the deviation values, Δ T D .
Given a fixed value of Δ T D , and that can be the value of the smallest deviation, i.e., δ T D = Δ T D ( { δ p k } ) , the set of the parameter deviations { δ p k } k = 1 n defines a locus of an n-dimensional ellipsoid, rotated with respect to the axes { δ p k } k = 1 n . This n-dimensional ellipsoid is bounded by the (n − 1)-dimensional locus of intersection between the n-dimensional paraboloid Δ T D = Δ T D ( { δ p k } ) and the n-dimensional hyperplane Δ T D = δ T D .
The n-dimensional ellipsoid is called uncertainty manifold, denoted by U n , for short. This is a manifold with an edge, meaning thus, its boundary, denoted by U n . In general, the edge of an n-dimensional manifold is an (n − 1)-dimensional manifold. Here, the edge U n involves the (n − 1)-dimensional locus of intersection between the n-dimensional paraboloid Δ T D = Δ T D ( { δ p k } ) and the n-dimensional hyperplane Δ T D = δ T D . The n-dimensional cuboid, which encloses the uncertainty manifold’s edge U n , is also a manifold with an edge and is denoted by Uc n . Its edge is an (n − 1)-dimensional manifold denoted by Uc n .
For example, consider the case of two-parametrical approximating functions, V ( x ; p 1 , p 2 ) . Then, the quadratic form of Equation (21) defines the two-dimensional paraboloid Δ T D = Δ T D ( δ p 1 , δ p 2 ) , immersed into the three-dimensional space with Cartesian axes given by ( x δ p 1 , y δ p 2 , z Δ T D ) . The two-dimensional ellipsoid is defined by the space bounded by the locus δ T D = Δ T D ( δ p 1 , δ p 2 ) , which is the intersection of the two-dimensional paraboloid Δ T D = Δ T D ( δ p 1 , δ p 2 ) and the two-dimensional hyperplane Δ T D = δ T D . For visualizing this example, see Figure 1.
Next, we will use the concept of the hyper-dimensional uncertainty manifold to derive the expressions of the errors of the optimal parameter values.

4. Derivation of the Errors of the Optimal Parameter Values

The expressions of the errors of the optimal parameter values—or simply, optimal errors—are well-known in the case of the least-square and other Euclidean based fitting methods. In [7], we have used the error expression, which is caused by the curvature, in order to have an estimate of the optimal errors (for applications, see [11,12,13,14,15,16,17,18]). Here, we will see the formal geometric derivation of the optimal error expressions.
First, we note that the edge of the uncertainty manifold, U n , has a number of n extrema, denoted by { C ( k ) } k = 1 n , and they are related to the errors of the parameters optimal values, { δ p k * } k = 1 n , as follows:
The position vector Δ ( μ ) of the corresponding point C ( μ ) , μ = 1 , 2 , , n , consists of n components each, i.e., Δ ( μ ) = ( Δ ( μ ) 1 , Δ ( μ ) 2 , , Δ ( μ ) n ) . Thus,
C ( 1 ) :    ( Δ ( 1 ) 1 , Δ ( 1 ) 2 , , Δ ( 1 ) n ) = Δ ( 1 ) C ( 2 ) :    ( Δ ( 2 ) 1 , Δ ( 2 ) 2 , , Δ ( 2 ) n ) = Δ ( 2 ) C ( n ) :    ( Δ ( n ) 1 , Δ ( n ) 2 , , Δ ( n ) n ) = Δ ( n ) .
These components are given by the condition:
{ Δ T D ( { δ p k } ) δ p ν } | { δ p k } = { Δ ( μ ) k } = { δ p ν k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ p k 2 } | { δ p k } = { Δ ( μ ) k } = 0 ,  
ν = 1 , 2 , , n , with ν μ .
The above (n − 1) equations, given in Equation (26), together with
δ T D = Δ T D ( { δ p k } = { Δ ( μ ) k } ) = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) Δ ( μ ) k 1 Δ ( μ ) k 2 ,
are sufficient for the calculation of the n unknown components of Δ ( μ ) = ( Δ ( μ ) 1 , Δ ( μ ) 2 , , Δ ( μ ) n ) .
The (n − 1) equations, given in Equation (26), arise from the fact that each of the errors of parameters optimal values, i.e., the k-th, δ p k * is derived from the maximum value of the corresponding component δ p k , that is, δ p k , max . Hence, the existence of maximum values { δ p k * } k = 1 n , i.e., k = 1 , 2 , , n , δ p k * U n : δ p k U n , δ p k * δ p k , leads to the errors estimation, { δ p k * } k = 1 n = { δ p k , max } k = 1 n . These maximum values, are located on the edge of the uncertainty manifold U n , that is the hypersurface δ T D = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ p k 2 , i.e., δ p k * U n .
The maximization of the μ-th parameter deviation δ p μ within uncertainty manifold is derived as follows: Since δ T D = Δ T D ( { δ p k } ) , we can express δ p μ in terms of δ p ν , ν = 1 , 2 , , n with ν μ , i.e., δ p μ = δ p μ ( { δ p k } k = 1 , k μ n ) . Then, the procedure of finding the maximum value of δ p μ , that is, δ p μ * = δ p μ | max , involves finding all the derivatives / δ p ν , ν = 1 , 2 , , n with ν μ of δ p μ = δ p μ ( { δ p k } k = 1 , k μ n ) , or equivalently, of δ T D = Δ T D ( { δ p k } ) (implicit derivatives). The (n − 1) equations δ p μ ( { δ p k } k = 1 , k μ n ) / δ p ν = 0 lead to the (n − 1) relationships δ p ν = δ p ν ( δ p μ ) , ν = 1 , 2 , , n with ν μ , which together with δ p μ = δ p μ ( { δ p k } k = 1 , k μ n ) , leads to the specific values of { Δ ( μ ) k = δ p k } k = 1 n . Yet, only the μ-th component Δ ( μ ) μ = δ p μ gives the error δ p μ * = δ p μ | max , i.e.,
δ p μ * = Δ ( μ ) μ   .  
In Appendix A, we solve Equations (26) and (27), where, we concluded that
Δ ( m 1 ) m 2 = σ 2 m 1 m 2 σ 2 m 1 m 1 = δ T D ( A 2 ( q ) 1 ) m 1 m 2 ( A 2 ( q ) 1 ) m 1 m 1   ,
or
Δ ( m 1 ) m 2 = { 1 N A 0 ( q ) ( A 2 ( q ) 1 ) m 1 m 2 ( A 2 ( q ) 1 ) m 1 m 1 ,    if   A 0 N > T r e s , T r e s ( A 2 ( q ) 1 ) m 1 m 2 ( A 2 ( q ) 1 ) m 1 m 1 ,       if   A 0 N T r e s ,
and thus, from Equation (28), we finally derive the errors:
δ p k * = Δ ( k ) k = δ T D ( q ) ( A 2 ( q ) 1 ) k k = { 1 N A 0 ( q ) ( A 2 ( q ) 1 ) k k ,    A 0 N > T r e s , T r e s ( A 2 ( q ) 1 ) k k ,        A 0 N T r e s ,
k = 1 , 2 , , n .
Finally, taking into account the resolution of each parameter value, { p r e s , k } k = 1 n , we have:
δ p k * = M a x ( p r e s , k , δ T D ( q ) ( A 2 ( q ) 1 ) k k ) ,
where
δ T D ( q ) ( A 2 ( q ) 1 ) k k = { 1 N A 0 ( q ) ( A 2 ( q ) 1 ) k k ,    if     A 0 N > T r e s    , T r e s ( A 2 ( q ) 1 ) k k ,        if     A 0 N T r e s    ,
k = 1 , 2 , , n .

5. Formulation of the Cases of n = 1 and n = 2 Dimensional Uncertainty Manifold

5.1. The Case of n = 1

Let us begin with the case of a one-dimensional paraboloid, given simply by the parabola
Δ T D ( δ p ) = A 2 ( q ) δ p 2   ,  
corresponding to uni-parametrical approximating functions. The locus of intersection between this parabola and the line Δ T D = δ T D (that is, the one-dimensional hyperplane) are the two points δ p ± = ± δ T D / A 2 ( q ) . The uncertainty manifold U 1 is the one-dimensional ellipsoid, defined by the line segment δ p δ p δ p + , which is enclosed by the points δ p ± . In this case, the edge of the uncertainty manifold U 1 is restricted to the zero-dimensional space composed only by the two points δ p ± . The manifolds U 1 and Uc 1 coincide (similarly with their edges, U 1 and Uc 1 , respectively). Hence,
δ p * = δ T D ( q ) A 2 ( q ) = { 1 N A 0 ( q ) / A 2 ( q ) ,    A 0 N > T r e s , T r e s / A 2 ( q ) ,        A 0 N T r e s .

5.2. The Case of n = 2

The case of bi-parametrical approximating functions is characterized by the two-dimensional paraboloid,
Δ T D ( δ p 1 , δ p 2 ) = A 11 ( q ) δ p 1 2 + 2 A 12 ( q ) δ p 1 δ p 2 + A 22 ( q ) δ p 2 2 ,  
which is illustrated in Figure 1. The locus of intersection between this paraboloid and the plane Δ T D ( δ p 1 , δ p 2 ) = δ T D is given by the rotated ellipse:
δ T D = A 11 ( q ) δ p 1 2 + 2 A 12 ( q ) δ p 1 δ p 2 + A 22 ( q ) δ p 2 2 ,  
written suitably as
( δ p 1 b 1 ) 2 + ( δ p 2 b 2 ) 2 = 1 ,  
after the rotation transformation
( δ p 1 δ p 2 ) = ( cos θ sin θ sin θ cos θ ) ( δ p 1 δ p 2 )   ,   or
δ p = R δ p ,
where
R = R ( θ ) = ( cos θ sin θ sin θ cos θ ) ,   δ p ( δ p 1 δ p 2 ) ,   δ p ( δ p 1 δ p 2 ) .
Then,
δ T D = δ p t A 2 δ p = δ p t ( R t A 2 R ) δ p = δ p t B δ p ,
where the diagonal matrix
B = B ( q ) = ( B 11 ( q ) 0 0 B 22 ( q ) ) ,
has the following elements
B ± = 1 2 [ A 11 + A 22 ± ( A 11 A 22 ) 2 + 4 A 12 2 ] ,   B 11 = B + ,   B 22 = B ,
which are the eigenvalues of the matrix A 2 ( q ) . The ellipsis’ major/minor axes in Equation (35) are:
b 1 = δ T D B 11 ,   b 2 = δ T D B 22   .
while the rotation angle θ in Equation (36) is given by
tan 2 θ = 2 A 12 A 22 A 11   .
The uncertainty manifold U 2 is the rotated 2-dim ellipsoid in the ( δ p 1 , δ p 2 ) axes, defined by:
A 11 ( q ) δ p 1 2 + 2 A 12 ( q ) δ p 1 δ p 2 + A 22 ( q ) δ p 2 2 δ T D   ,
or, in the rotated axes ( δ p 1 , δ p 2 ) , is simply given by:
( δ p 1 b 1 ) 2 + ( δ p 2 b 2 ) 2 1   ,
which is enclosed by the ellipse corresponding to the equal sign of Equation (45), that is the edge of the uncertainty manifold, U 2 .
Finally, the errors are
δ p 1 * = δ T D ( q ) ( A 2 ( q ) 1 ) 1 1 = δ T D ( q ) ( A 2 ) 2 2 ( q ) D A 2 ( q ) = { 1 N A 0 ( q ) ( A 2 ) 2 2 ( q ) D A 2 ( q ) ,    A 0 N > T r e s , T r e s ( A 2 ) 2 2 ( q ) D A 2 ( q ) ,        A 0 N T r e s ,
and
δ p 2 * = δ T D ( q ) ( A 2 ( q ) 1 ) 2 2 = δ T D ( q ) ( A 2 ) 1 1 ( q ) D A 2 ( q ) = { 1 N A 0 ( q ) ( A 2 ) 1 1 ( q ) D A 2 ( q ) ,    A 0 N > T r e s , T r e s ( A 2 ) 1 1 ( q ) D A 2 ( q ) ,        A 0 N T r e s .

6. Conclusions

The paper presented the geometric interpretation of the errors of the optimal parameter values, derived from a multi-parametrical fitting, based on a metric induced by the non-Euclidean Lq-norm. Typical fitting methods are mostly developed based on Euclidean norms, leading to the traditional least–square method. On the other hand, the theory of general fitting methods based on non-Euclidean norms, is still under development; the normal equations can provide the optimal values of the fitting parameters, while this paper completed the puzzle by improving understanding the derivations and geometric meaning of the errors.
In particular, we showed that the statistical errors of the optimal parameter values are given by the axes of the ellipsoid called uncertainty manifold, that is, the intersection of the paraboloid of the residuals’ expansion Δ T D ( { δ p k } ) T D q ( { p k = p k * + δ p k } ) q T D q ( { p k * } ) q along the deviations { δ p k } k = 1 n , with the hyperplane Δ T D ( { δ p k } ) = δ T D = c o n s t . The constant δ T D represents the smallest possible value of a deviation from the TD’s minimum, also mentioned as an error of the value of the total deviations.
In summary, the Lq-normed fitting involves minimizing:
T D q ( { p k } ) q = A 0 ( q ) + k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) ( p k 1 p k 1 * ) ( p k 2 p k 2 * ) + O ( | p k p k * | 3 ) ,
where
A 0 ( q ) = x D | u | q d x   ,  
A 2 , k 1 k 2 ( q ) = δ 1 q γ k 1 k 2 + q 2 x D [ | u | q 1 sgn ( u ) 2 V ( x ; { p k * } ) p k 1 p k 2 + ( q 1 ) | u | q 2 V ( x ; { p k * } ) p k 1 V ( x ; { p k * } ) p k 2 ] d x ,
γ k 1 k 2 i : u ( x i ) = 0 1 | u ( x i ) | V ( x i ; { p k * } ) p k 1 V ( x i ; { p k * } ) p k 2   .  
The normal equations are given by:
x D | u | q 1 sgn ( u ) V ( x ; { p k * } ) p l d x = 0 ,   l = 1 , 2 , , n   ,  
where we set u = u ( x ) V ( x ; { p k * } ) f ( x ) .
Finally, we summarize the concluding relationships of the paper:
Δ T D ( { δ p k } ) T D q ( { p k = p k * + δ p k } ) q T D q ( { p k * } ) q = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ p k 2   ,
Δ T D min M a x ( δ T D ( q ) , T r e s )   ,   δ T D ( q ) = 1 N A 0 ( q )   ,   A 0 ( q ) T D q ( { p k * } ) q   ,  
δ p k * = M a x ( p r e s , k , δ T D ( q ) ( A 2 ( q ) 1 ) k k )   ,  
with special cases:
- For n = 1:
δ p * = δ T D ( q ) A 2 ( q ) = { 1 N A 0 ( q ) / A 2 ( q ) ,    A 0 N > T r e s , T r e s / A 2 ( q ) ,        A 0 N T r e s .
- For n = 2:
δ p 1 * = δ T D ( q ) ( A 2 ( q ) 1 ) 1 1 = δ T D ( q ) ( A 2 ) 2 2 ( q ) D A 2 ( q ) = { 1 N A 0 ( q ) ( A 2 ) 2 2 ( q ) D A 2 ( q ) ,    A 0 N > T r e s , T r e s ( A 2 ) 2 2 ( q ) D A 2 ( q ) ,        A 0 N T r e s ,
and
δ p 2 * = δ T D ( q ) ( A 2 ( q ) 1 ) 2 2 = δ T D ( q ) ( A 2 ) 1 1 ( q ) D A 2 ( q ) = { 1 N A 0 ( q ) ( A 2 ) 1 1 ( q ) D A 2 ( q ) ,    A 0 N > T r e s , T r e s ( A 2 ) 1 1 ( q ) D A 2 ( q ) ,        A 0 N T r e s .

Funding

This research was funded by NASA’s HGI Program, grant number NNX17AB74G.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A. Extrema of the Uncertainty Manifold

Here we calculate the position vectors { Δ ( k ) } k = 1 n and the maximum points { C ( k ) } k = 1 n , used for the derivations of errors of the optimal parameter values in Section 4.
The following (n − 1) equations:
[ δ p ν Δ T D ( { δ p k } ) ] | { δ p k } = { Δ ( μ ) k } = 0 ,
ν = 1 , 2 , , n , with ν μ , together with the one of
δ T D = Δ T D ( { δ p k } = { Δ ( μ ) k } ) = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) Δ ( μ ) k 1 Δ ( μ ) k 2 ,
are sufficient for the calculation of the n unknown components of Δ ( μ ) = ( Δ ( μ ) 1 , Δ ( μ ) 2 , , Δ ( μ ) n ) . Then,
{ δ p ν [ Δ T D ( { δ p k } ) ] } | { δ p k } = { Δ ( μ ) k } = { δ p ν [ k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ p k 2 ] } | { δ p k } = { Δ ( μ ) k } = { k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 1 δ k 2 ν + k 1 , k 2 = 1 n A 2 , k 1 k 2 ( q ) δ p k 2 δ k 1 ν } | { δ p k } = { Δ ( μ ) k } = 2 { k = 1 n A 2 , ν k ( q ) δ p k } | { δ p k } = { Δ ( μ ) k } = 2 k = 1 n A 2 , ν k ( q ) Δ ( μ ) k ,
(where we used the Kronecker’s delta, δmn = 1 for m = n and 0 for mn). Hence,
k = 1 n A 2 , ν k ( q ) Δ ( μ ) k = 0   ,
ν = 1 , 2 , , n , with ν μ . Setting:
k = 1 n A 2 , μ k ( q ) Δ ( μ ) k ζ μ ,
we have
ζ μ e ^ μ = ( A 2 )    Δ ( μ )   ,
where
e ^ μ ( 0 0 1 0 0 ) = ( { e ^ μ , k = 0 } k = 1 , k μ n ; e ^ μ , μ = 1 )   .
Inversing Equation (A6), we obtain
Δ ( μ ) = ζ μ ( A 2 1 ) e ^ μ   .
or
Δ ( μ ) m = ( A 2 1 ) m μ ζ μ   ,
m = 1 , 2 , , n . Then,
δ T D = k 1 , k 2 = 1 n A 2 , k 1 k 2 Δ ( μ ) k 1 Δ ( μ ) k 2 = k 1 , k 2 = 1 n A 2 , k 1 k 2 ( A 2 1 ) k 1 μ ζ μ ( A 2 1 ) k 2 μ ζ μ = ζ μ 2 k 1 , k 2 = 1 n A 2 , k 1 k 2 ( A 2 1 ) k 1 μ ( A 2 1 ) k 2 μ = ζ μ 2 k 2 = 1 n δ k 2 μ ( A 2 1 ) k 2 μ = ζ μ 2 ( A 2 1 ) μ μ ,
thus,
ζ μ = δ T D ( A 2 1 ) μ μ   ,
or
Δ ( μ ) m = ( A 2 1 ) m μ δ T D ( A 2 1 ) μ μ   ,
m = 1 , 2 , , n , and by refreshing the indices, we end up with
Δ ( m 1 ) m 2 = δ T D ( A 2 ( q ) 1 ) m 1 m 2 ( A 2 ( q ) 1 ) m 1 m 1   .

References

  1. Kenney, J.F.; Keeping, E.S. Linear Regression and Correlation. In Mathematics of Statistics, 3rd ed.; Part 1; Van Nostrand: Princeton, NJ, USA, 1962; pp. 252–285. [Google Scholar]
  2. McCullagh, P. What is statistical model? Ann. Stat. 2002, 30, 1225–1310. [Google Scholar] [CrossRef]
  3. Adèr, H.J. Modelling. In Advising on Research Methods: A Consultant’s Companion; Adèr, H.J., Mellenbergh, G.J., Eds.; Johannes van Kessel Publishing: Huizen, The Netherlands, 2008; pp. 271–304. [Google Scholar]
  4. Melissinos, A.C. Experiments in Modern Physics; Academic Press Inc.: London, UK, 1966; pp. 438–464. [Google Scholar]
  5. Burden, R.L.; Faires, J.D. Numerical Analysis; PWS Publishing Company: New Orleans, LA, USA, 1993; pp. 437–438. [Google Scholar]
  6. Livadiotis, G.; McComas, D.J. Fitting method based on correlation maximization: Applications in Astrophysics. J. Geophys. Res. 2013, 118, 2863–2875. [Google Scholar] [CrossRef]
  7. Livadiotis, G. Approach to general methods for fitting and their sensitivity. Physica A 2007, 375, 518–536. [Google Scholar] [CrossRef]
  8. Livadiotis, G. Expectation values and variance based on Lp norms. Entropy 2012, 14, 2375–2396. [Google Scholar] [CrossRef]
  9. Livadiotis, G. Chi-p distribution: Characterization of the goodness of the fitting using L p norms. J. Stat. Distr. Appl. 2014, 1, 4. [Google Scholar] [CrossRef]
  10. Livadiotis, G. Non-Euclidean-normed Statistical Mechanics. Physica A 2016, 445, 240–255. [Google Scholar] [CrossRef] [Green Version]
  11. Livadiotis, G. On the convergence and law of large numbers for the non-Euclidean Lp-means. Entropy 2017, 19, 217. [Google Scholar] [CrossRef]
  12. Livadiotis, G.; Moussas, X. The sunspot as an autonomous dynamical system: A model for the growth and decay phases of sunspots. Physica A 2007, 379, 436–458. [Google Scholar] [CrossRef]
  13. Frisch, P.C.; Bzowski, M.; Livadiotis, G.; McComas, D.J.; Mӧbius, E.; Mueller, H.-R.; Pryor, W.R.; Schwadron, N.A.; Sokól, J.M.; Vallerga, J.V.; et al. Decades-long changes of the interstellar wind through our solar system. Science 2013, 341, 1080. [Google Scholar] [CrossRef] [PubMed]
  14. Schwadron, N.A.; Mobius, E.; Kucharek, H.; Lee, M.A.; French, J.; Saul, L.; Wurz, P.; Bzowski, M.; Fuselier, S.; Livadiotis, G.; et al. Solar radiation pressure and local interstellar medium flow parameters from IBEX low energy hydrogen measurements. Astrophys. J. 2013, 775, 86. [Google Scholar] [CrossRef]
  15. Livadiotis, G.; McComas, D.J. Evidence of large scale phase space quantization in plasmas. Entropy 2013, 15, 1118–1132. [Google Scholar] [CrossRef]
  16. Fuselier, S.A.; Allegrini, F.; Bzowski, M.; Dayeh, M.A.; Desai, M.; Funsten, H.O.; Galli, A.; Heirtzler, D.; Janzen, P.; Kubiak, M.A.; et al. Low energy neutral atoms from the heliosheath. Astrophys. J. 2014, 784, 89. [Google Scholar] [CrossRef]
  17. Zirnstein, E.J.; McComas, D.J. Using kappa functions to characterize outer heliosphere proton distributions in the presence of charge-exchange. Astrophys. J. 2015, 815, 31. [Google Scholar] [CrossRef]
  18. Fuselier, S.A.; Dayeh, M.A.; Livadiodis, G.; McComas, D.J.; Ogasawara, K.; Valek, P.; Funsten, H.O. Imaging the development of the cold dense plasma sheet. Geophys. Res. Lett. 2015, 42, 7867–7873. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) Scheme of the n-dimensional paraboloid, Δ T D ( { δ p k } ) , which is constructed as the deviation of the total residuals or deviations functional, total deviations (TD), from its minimum value, A0, as shown in Equation (21); the illustrated example is for n = 2. (b) The intersection between the paraboloid and a constant hyper-plane Δ T D = δ T D is a rotated n-dimensional ellipsoid, or a rotated ellipsis for the case of n = 2, that is the uncertainty manifold U 2 (enclosing U 2 ). The rectangular adjusted on the extrema of U 2 denotes the manifold Uc 2 (enclosing Uc 2 ).
Figure 1. (a) Scheme of the n-dimensional paraboloid, Δ T D ( { δ p k } ) , which is constructed as the deviation of the total residuals or deviations functional, total deviations (TD), from its minimum value, A0, as shown in Equation (21); the illustrated example is for n = 2. (b) The intersection between the paraboloid and a constant hyper-plane Δ T D = δ T D is a rotated n-dimensional ellipsoid, or a rotated ellipsis for the case of n = 2, that is the uncertainty manifold U 2 (enclosing U 2 ). The rectangular adjusted on the extrema of U 2 denotes the manifold Uc 2 (enclosing Uc 2 ).
Stats 02 00029 g001

Share and Cite

MDPI and ACS Style

Livadiotis, G. Geometric Interpretation of Errors in Multi-Parametrical Fitting Methods Based on Non-Euclidean Norms. Stats 2019, 2, 426-438. https://doi.org/10.3390/stats2040029

AMA Style

Livadiotis G. Geometric Interpretation of Errors in Multi-Parametrical Fitting Methods Based on Non-Euclidean Norms. Stats. 2019; 2(4):426-438. https://doi.org/10.3390/stats2040029

Chicago/Turabian Style

Livadiotis, George. 2019. "Geometric Interpretation of Errors in Multi-Parametrical Fitting Methods Based on Non-Euclidean Norms" Stats 2, no. 4: 426-438. https://doi.org/10.3390/stats2040029

APA Style

Livadiotis, G. (2019). Geometric Interpretation of Errors in Multi-Parametrical Fitting Methods Based on Non-Euclidean Norms. Stats, 2(4), 426-438. https://doi.org/10.3390/stats2040029

Article Metrics

Back to TopTop