1. Introduction
Nowadays, the use of 3D laser scanning technology is widespread in various applications for digitizing real world objects. The acquired point cloud consists of measured coordinates in
of numerous points. The number of points can be up to several millions, resulting in a huge data set. Moreover, the points are considered unorganized and
scattered, which means that there is no specific structure or topological relation between them. Hence, the points and their coordinates are not sufficient to provide additional information about the object. As Bureick et al. [
1] mentioned, only by modelling the geometric information can further analysis be performed, and several questions which could not be addressed before can be answered. For instance, in Computer-Aided Design (CAD) or in deformation analysis, a surface approximation with a continuous mathematical function is required, as, e.g., pointed out by Bureick et al. [
1]. Therefore, modelling the point cloud is a necessity and surface approximation is taken into account.
There are numerous approaches to approximate a surface involving functions such as polynomials, splines, B–splines, etc. For instance, Bureick et al. [
1] presented surface approximations based on polynomials, Bézier, B-splines and Non-Uniform Rational B-splines (NURBS). Each of these approaches has its advantages and some disadvantages. In this article, the Radial Basis Function (RBF) approach is discussed. This choice is motivated by typical tasks in Geodesy and Geoinformation Science. There, the challenge is often to model point clouds that include a very large number of points in the 3D space dimension with unorganized points (scattered points), which point clouds have been captured with laser scanners or photogrammetric methods. As Buhmann [
2] (p. 1ff.) explained, when the functions to be approximated (a) depend on many variables or parameters, (b) are defined by possibly very many data, and (c) the data are scattered in their domain, then the Radial Basis Function approach is especially well-suited.
Another advantage of the Radial Basis Function approach is that it is a
meshfree approach. In contrast to some other approaches that require a mesh such as those using wavelets, multivariate splines, finite elements, etc., as Wendland [
3] (p. ix) explained. According to Fasshauer [
4] (p. 1), mesh generation can be the most time-consuming part of any mesh-based numerical approach. Moreover, he continued, a meshfree approach is often better suited to cope with changes in the geometry of the domain of interest (e.g., free surfaces and large deformations) than approaches such as finite elements.
Mainly, in this article, the approximation of point clouds is considered, since with high point density and the fact that measured coordinates are subject to random measurement errors, interpolation is no longer appropriate. In the case of interpolation, measurement errors and other abrupt changes in the data points would be modelled, resulting in a strongly oscillating modelled surface. Nevertheless, some numerical examples of interpolation, with a few data points, will be presented to reveal the effect of the applied technique.
For the scattered data approximation problem with the RBF approach, an approximation function is built that consists of a linear combination of radial basis functions. Each radial “basis” function
is centered at each fixed center point, and by composition with the Euclidean norm
,
is radially (or spherically) symmetric about this center point. A formal definition of a radial function is given by Fasshauer [
4] (p. 17). Furthermore, each radial “basis” function is a
multivariate function which is based on a
univariate continuous “basic” function
. The basic function is used to derive all the radial basis functions in the approximation function. The terminology “basis” and “basic” function is adopted from the textbook by Fasshauer [
4] (p. 18). Based on the contribution by Flyer et al. [
5],
Table 1 shows some of the well-known basic functions, where
is a
shape parameter defined by the user.
In this article, the basic function used, namely
thin-plate spline, is chosen for the important advantage that no shape parameter is required. Based on the analysis of Fasshauer [
4] (p. 23ff.), the value of the shape parameter has a great influence on the numerical stability of the interpolation problem, and if not chosen appropriately, can result in severe ill-conditioning of the interpolation matrix. A sophisticated method is, e.g., presented in the paper by Zheng et al. [
6], who used a test differential equation for the selection of a good shape parameter. In the following, a shape parameter-free basic function is preferred. In addition to the previous advantage, the thin-plate spline function, according to Fasshauer [
4] (p. 170), has the tendency to produce “visually pleasing” smooth and tight surfaces.
The first appearance of the thin-plate spline function in an interpolation problem was presented by Harder and Desmarais [
7] under the name “surface spline” for applications related to aircraft design. Duchon [
8] studied their method and extended the mathematical representation based on Hilbert kernel theory, which had a major impact on subsequent research on the interpolation of spline functions with radial basis functions. In his article, Duchon [
8] was interested in an interpolating surface with minimal bending energy, which geometrically means that he was aiming for an interpolation that was as flat as possible, that is, as close to a plane as possible. He composed, for this purpose, the interpolating function as a linear combination of thin-plate spline functions with an addition of linear polynomials. Furthermore, condition equations were taken into account to ensure a unique solution.
As Harder and Desmarais [
7] explained, the thin-plate spline function has a physical interpretation which is a plate of infinite extend that deforms in bending only. Since the bending energy is represented by the integral of the partial derivatives, the minimal bending energy of the surface that Duchon [
8] was interested in is approximately equivalent to the minimal integral curvature, and hence with the minimum of the functional of second derivatives. For this minimisation, Duchon [
8] defined a semi-norm in the infinitely dimensional vector subspace of functions of finite energy. Furthermore, since linear polynomials defining a plane are equal to zero in their second derivatives, they are also included in the interpolation function. Finally, Duchon [
8] proved that the functions, thin-plate splines and polynomials can be used as a basis for surface interpolation, and the space of this basis is a subspace of the infinite dimensional Hilbert vector space of all real functions of two real variables. Later on, this form of interpolation presented by Duchon [
8] was widely used in the literature, for instance, by Buhmann [
9], by Wendland [
3] (p. 116ff.), by Fasshauer [
4] (p. 70ff.), and by Flyer et al. [
5], to name a few, when interpolating with thin-plate spline functions.
Here, it should be clarified that this form of interpolation that is adding a series of polynomials to the interpolation function and also considering condition equations is an interpolation technique that was first introduced by Hardy [
10], but with different basis functions, under the title “osculating mode”. Specifically, in his article, the interpolation function is constructed as a linear combination of
multiquadric functions plus a series of polynomials. Additionally, condition equations were used. Hardy [
10] applied this technique for the problem of representing a topographic surface by a continuous function while a set of few discrete data on the surface was given. In his application, he was seeking to minimize horizontal and vertical displacements of the maximum (hilltops) and minimum (depressions) points in the resulting surface of topography. Therefore, he chose an osculating, as he named it, interpolation technique, where the surface coordinates collocate with the data point coordinates, but also surface tangents coincide at specified points. The latter was achieved by the usage of low degree polynomials in the interpolation function. Later on, this interpolation technique was used in various applications, as were described collectively by Hardy [
11]. In this article, Hardy [
11] named this interpolation technique the
multiquadric method.
An additional clarification follows here. In the literature on RBF, the term “condition” equations is usually used, while in the geodetic literature these equations are referred to as “constraint” equations since only unknown parameters (and possibly error-free values) are considered and related to each other. Mikhail [
12] (p. 213ff.) explained that constraints for the parameters occur when some or all of the parameters must conform to some relationships arising from either geometric or physical characteristics of the problem. Therefore, in the remainder of this article, these equations will only be referred to as constraint equations.
In this article, it is shown that the approximation problem with the RBF approach, consisting of a linear combination of thin-plate splines, has a unique solution without modifying the approximation function, i.e., without additional polynomials and without constraint equations. The unknown parameters are estimated with the use of the least squares method. Additionally, it is shown that when constraint equations must be introduced, a least squares solution with constraint equations for the unknown parameters can be computed. Moreover, some interpolation problems of a few points, with thin-plate splines, without additional polynomial terms and constraint equations, revealed that they are well-posed. This article is organized as follows.
In
Section 2, the theoretical foundations for the solution of the approximation problem with thin-plate splines are shown. The result is a linear least squares adjustment problem, which is well-known, e.g., in Geodesy.
In
Section 3, a numerical example from the textbook by Fasshauer [
4] is presented. The solution technique used by Fasshauer [
4] (p. 170) is described and discussed.
In
Section 4, the numerical example of
Section 3 is used again. However, the solution is now performed using the adjustment technique presented in
Section 2, which allows for the rigorous consideration of the constraint equations.
In
Section 5, the numerical example of
Section 3 is solved again, but without the addition of the polynomial terms and the constraint equations for the unknown parameters. The solution is obtained based on the theory presented in
Section 2.
Section 7 contains a comparison of eight interpolation problems with thin-plate splines of four data sets. Finally, in
Section 8, conclusions are drawn and some considerations for further investigations are made.
3. A Constrained Solution of RBF Surface Approximation from the Literature
In the following, a constrained solution of RBF surface approximation from the textbook by Fasshauer [
4] is presented and discussed using a numerical example presented by Fasshauer [
4] (p. 170ff.) in Section 19.4 under the title “Least squares smoothing of noisy data”. The approximation problem described is dealing with noisy data, for instance, measurements. The noisy data are created by sampling Franke’s function, given by Fasshauer [
4] (p. 20) as
at a set
, with
, of the given data, to which
of normally distributed random noise is added, and then the “measurements” are stored in a file. In addition, a smaller set
, with
, of uniformly distributed data is given where every
is a center of a radial basis function in two dimensions. Both sets of data were provided on a CD supplied with the textbook.
Figure 1 shows the positions of the data points together with the coordinates of the center points.
The problem is overdetermined with
and the approximation is performed with the use of the least squares method. The approximation function (
27) is built as a linear combination of radial basis functions with an addition of three linear polynomial terms. The chosen basic function for this problem is the thin-plate spline function, as it is given in (
2). According to Fasshauer [
4] (p. 71), the added polynomials are necessary in order for an interpolation problem using thin-plate splines to be
well-posed. Nevertheless, an approximation problem is considered here and Fasshauer [
4] (p. 170ff.) did not give a clear justification for the addition of the polynomials in this approximation problem.
Since the functions of this problem are not given explicitly in the aforementioned Section 19.4 of the textbook, the notation has been derived with the help of Section 6.2 (“Example: Reproduction of linear functions using Gaussian RBFs”) in Fasshauer [
4] (p. 55ff.), the information given in Fasshauer [
4] (p. 70ff.) and the source code given in Fasshauer [
4] (p. 171). Therefore, the approximation function is formed as
with
and
. Now, to the number
m of the unknown parameters
are added three additional unknown parameters, namely
. This results in the total number of unknown parameters of the problem being
. However, the problem remains overdetermined with
.
Additionally to the function in (
27), three constraint equations are considered in this approximation problem, given by
which, according to Fasshauer [
4] (p. 58), are needed to ensure a unique solution. Here it must be noted that in Fasshauer [
4] (p. 58), in this particular page of the textbook, an interpolation problem with added polynomial terms is considered, where
. Obviously, the three constraint equations are introduced because three additional unknown parameters are added to the function used for the interpolation. Since
, adding three unknown parameters leads to
, and the interpolation matrix would be singular because there would be a rank deficiency of three. In order to eliminate this rank deficiency, three constraint equations are added.
Back to the approximation problem described in Fasshauer [
4] (p. 170ff.). The augmented system
derived from the code given in Fasshauer [
4] (p. 171), needs to be solved. Here, the matrix
contains the radial basis functions. Since the coordinates of the center points
are different from each other,
applies. Thus, the columns of matrix
(
30) are linearly independent and
has full (column) rank. The observations (or the noisy data) are denoted with
while the matrix
is a
zero matrix.
In order for Fasshauer [
4] (p. 170ff.) to solve this linear system of equations with the rectangular matrix
, he extended it, as shown in (
29), by using the data points
in the right upper block of the augmented matrix and the center points
in the left lower block of the same matrix. Afterwards, he applied the built-in function of Matlab, known as “backslash operator” (∖ or
mldivide), where according to the official documentation of Matlab, if the matrix
is rectangular, then
returns a least squares solution to the system of equations
.
For the solution of the example by Fasshauer [
4] (p. 170ff.), the residuals
cf. (
14), and the objective function of the least squares method (
10) were additionally calculated. The numerical result of the objective function is
.
The condition number, see (
23), of
which was used to obtain the solution, results in
, while the condition number of the normal matrix
cf. (
18) is
. Furthermore, the normal matrix has full (maximum) rank with
, since this number is equal to the number of the unknown parameters of the problem, which is
. Additionally, the matrix
has full (column) rank with
. Moreover, the eigenvalues of matrix
are all real and positive, which shows that the normal matrix is positive definite.
It should be mentioned that Fasshauer [
4] (p. 170ff.) computes a least squares solution without explicitly building the normal matrix. Setting up this matrix and calculating the condition number is only presented here to compare it with the numerical results presented in
Section 4 and
Section 5.
For the example considered, Fasshauer [
4] (p. 170ff.) did not explain how the set of the center points was selected. However, in Fasshauer [
4] (p. 168), it is described that the approximation problem will have a unique solution if the rectangular matrix
has full rank. In particular, if the center points are chosen to form a subset of the data points, then the matrix
will have a square matrix
, which will be non-singular.
The equations given in (
28) were defined as constraint equations, which means that they are supposed to be rigorously fulfilled by the estimated unknown parameters
. In fact, the results are the following
and it can be seen that all three equations are not equal to zero, as was requested in (
28). This means that these equations are not defined as constraint equations but as additional equations to the function given in (
27). This conclusion can also be derived from the extension of the matrix
, where different points were used for each block of this matrix, as was explained previously. Moreover, the last three zero entries in the extended vector of the observations
, given in (
29), are introduced as “pseudo-observations” and hence residuals must be taken into account for these observations as well.
Therefore, the functions that describe this numerical example must be rewritten correctly as
where, according to (
35), the observation equations can be formed as
with
. Here, the zero values are additional observations, and residuals are considered for these observations as well. This, however, leads finally to the fact that the constraint equations are not strictly satisfied by the estimated parameters, see (
34). This gives the motivation to develop a solution in which the constraint equations are rigorously satisfied, which will be shown in the following
Section 4.
4. A Rigorous Constrained Solution of RBF Surface Approximation
The introduction of constraint equations in adjustment calculations can be traced far back in the geodetic literature, see, e.g., the textbook by Helmert [
18] (p. 195ff.). Constraint equations are always introduced when a function of the estimated parameters must fulfil certain geometric or physical properties. In the case of a singular adjustment problem,
d constraint equations can be introduced to eliminate a rank defect
d of the design matrix. An example of this is the so-called free network adjustment, where the geodetic datum is fixed by introducing constraint equations for the coordinates to be determined.
For the approximation function (
27), in which three additional unknown parameters were introduced by considering the polynomial terms, the design matrix
results, here, in
Despite the introduction of additional unknowns, the rank of this matrix is maximum with . Thus, the problem can be solved without additional information, i.e., without the introduction of constraint equations. Furthermore, the problem remains overdetermined, although three additional unknown parameters were introduced.
Nevertheless, in the following it will be shown, by means of the numerical example from
Section 3, how the three constraint equations (
28) can be taken into account in a least squares adjustment problem in such a way that they are rigorously fulfilled by the estimated parameters.
Therefore, as shown in
Section 2, the matrix
has to be set up, which contains the coefficients with which the unknown parameters in the constraint equations (
28) are multiplied. This matrix results, in this case, in
where zero values are included since in the constraint equations no polynomial terms are present.
With the design matrix
in (
37), the normal matrix
according to (
18) can be formed, and thus the extended normal equation system
cf. (
25) can be set up. The right hand side vector of the system
is calculated as in (
19). Since the three constraint equations given in (28) must each yield the value zero, a zero vector of the dimension
is appended. The extended vector
contains the coefficients of the unknown parameters, the additional unknown parameters of the polynomials and the Lagrange multipliers
. The residuals
can be computed according to (
14). The numerical result of the objective function (
10) is
, which is slightly larger (the difference starts in the sixth decimal place) than the numerical result of the objective function given in the previous
Section 3.
Inserting the estimated parameters into the constraint equations, the following results
are obtained, and it can be stated that within the range of the machine precision (in this case double precision), these yield the value zero as a result. Thus, the constraint equations are rigorously fulfilled by the presented adjustment technique.
Additionally, the design matrix
, the normal matrix
and the extended normal matrix
also have, in this case, full rank. This, according to the theory in
Section 2, means that the solution is unique under the consideration of the constraint equations. Finally, the condition number of the design matrix is
and the condition number of the extended normal matrix is
.
Obviously, the condition number of the extended normal matrix is not much better than the condition number of the normal matrix given in
Section 3. Furthermore, the calculation of the eigenvalues of the extended normal matrix
shows that all eigenvalues are real, of which three values are negative and the others are positive. This is expected since three constraint equations were introduced into this problem. This phenomenon, that the number of negative eigenvalues corresponds to the number of constraint equations introduced, can also be found in typical geodetic problems such as free network adjustment.
5. RBF Surface Approximation without Polynomial Terms
In
Section 3 and
Section 4, the RBF surface approximation problem was presented considering three linear polynomial terms. As already mentioned in the introduction of this article, this addition of polynomial terms along with thin-plate spline functions has been widely used in the RBF literature, mainly for interpolation, and for almost any application. Fasshauer [
4] (p. 55) explained that in some cases it is desirable that the interpolant exactly reproduces certain types of functions, e.g., constant or linear. Particularly in applications, he continued, related to partial differential equations (finite element methods), or in specific engineering applications (exact calculation of constant stress and strain), the reproduction of these simple linear polynomials is a necessity.
This leads to the important conclusion that polynomial terms are added when the application has certain requirements and it is motivated solely by the user’s intention to obtain a solution with certain properties. This conclusion can also be drawn from the article by Duchon [
8] where he was interested in an interpolating surface as close to a plane as possible. Similarly, Hardy [
10] added polynomial terms with multiquadric functions to the interpolation function to obtain an interpolating surface where surface tangents coincide at specified points. Finally, there seem to be (infinite dimensional) problems in physics where what is interpolated or approximated has almost a planar form with some small deviations. For all these applications, the added polynomial terms can be very useful.
However, the thin-plate spline function can be used as a basis for an interpolation or an approximation problem with the RBF approach without the polynomials. In particular, for the geodetic application of point cloud modelling, the data (geodetic) do not have the characteristics of the aforementioned problems. Hence, there is no need for these polynomials. Moreover, due to the high point density of the point clouds, the added polynomials would not significantly smooth the approximated surface.
In this section, the same numerical example will be considered, as in
Section 3, but in this case, no polynomials will be used. Therefore, the function without additional polynomial terms reads
In the overdetermined case, this function can only be fulfilled by the (theoretical) true values for the measurements
and the unknown parameters
, and the functional model is written
Since the true values are, of course, not known in practical applications, a solution according to the method of least squares is to be determined, for which the observation equations
are set up, where
are the measurements and
are the residuals.
The design matrix
, which contains the coefficients associated with the unknown parameters
, is calculated as already shown in (
30). The normal equations given in (17) are built with the normal matrix
(
18). The solution of the normal equations is derived (
20).
The numerical result of the objective function (
10) is
, which is smaller (the difference starts in the third decimal place) than the numerical result of the objective function from the techniques in
Section 3 and
Section 4. The normal matrix
and the design matrix
both also have in this case full rank, which is equal to the number of the unknown parameters
, and thus the solution is unique.
The condition number of the design matrix is
and the condition number of the normal matrix is
. These condition numbers appear smaller compared to the condition numbers given in the previous two sections but they still have the same order of magnitude as the others. Furthermore, the eigenvalues of the normal matrix are all real and positive, thus the normal matrix is positive definite. The following
Figure 2a shows the original (exact) surface derived from the Franke’s function given in
Section 3.
Figure 2b shows the resulting approximated surface calculated in this section.
6. Comparison and Discussion of the Investigated Techniques for Surface Approximation with Thin-Plate Splines
In
Section 3,
Section 4 and
Section 5, the same numerical example was computed in each surface approximation technique. For this numerical example, the measurements were “created” using Franke’s function. Thus, the original surface is known, see
Figure 2a, and the numerical differences between the values of the original surface and the values of the approximated surfaces can be calculated. Therefore, the
absolute error
is computed, with
points of the approximated surface and
points of the original surface. The points
are equally spaced points, here of
, covering the
square in two dimensions, and they are called
evaluation points.
Figure 3 illustrates the absolute errors
for the different solutions from
Section 3,
Section 4 and
Section 5.
Additionally,
Table 2 contains the maximum value of the absolute error from the three solutions.
Besides the absolute error, the
Root-Mean-Square error (
RMS–error) can be computed as well. According to Fasshauer [
4] (p. 10), the
RMS–error is given by
Table 3 contains the
RMS–error for the solutions given in the previous three sections.
From
Table 2 and
Table 3 it can be seen that the solution presented in
Section 5 gives the smallest values for both the maximum absolute error and the
RMS–error. The value of the maximum absolute error of the solution presented in
Section 4 is slightly smaller (differs in the fifth digit after the comma) than the value of this error from the solution in
Section 3, see
Table 2. On the contrary, the value of the
RMS–error of the solution in
Section 3 is slightly smaller (differs in the seventh digit after the comma) than the value of this error from the solution in
Section 4, see
Table 3.
Equivalent to the
RMS–error, the numerical value of the objective function of the least squares method, see
Table 4, can be used as a measure of the goodness of fit.
As for the
RMS–error, the smallest value of the objective function results from the solution from
Section 5 where no polynomials and constraint equations were introduced into the problem.
In addition, condition numbers for the design matrices and the normal matrices were given for the solution of each section, which are compiled in
Table 5 and
Table 6.
The condition numbers of the design matrices in
Table 5 are similar in numerical quantity, as they are of the same order of magnitude,
. According to Trefethen and Bau [
17] (p. 95), if
is ill-conditioned, one always expects to “lose
digits” in computing the solution. For the numerical examples in
Section 3,
Section 4 and
Section 5, the matrices of all the cases are mildly ill-conditioned and
for every case. Alternatively, working with double precision in Matlab with 64 bits, the
machine epsilon is
, which that means double precision numbers can be accurate to about 16 decimal places. Multiplying the machine epsilon with the condition numbers in
Table 5, the result for every case is
. This means that at least 12 digits in the solutions of the previous sections are correct. Hence, 4 digits must be ignored. Similar conclusions can also be drawn for the condition numbers in
Table 6 since the condition number of the normal matrix (or the extended normal matrix) computed as
is equal to the square of the condition number
.
Considering the fact that erroneous data (measurements) are involved in the approximation problems in
Section 3,
Section 4 and
Section 5, the linear system of equations, in each case, with respect to the solution, presents low sensitivity. That is, the solution can be “trusted” up to 12 decimal places for every case. Moreover, the introduction of polynomial terms and constraint equations into the problem showed no improvement with respect to the sensitivity of the computation process. This means that the condition number of the design matrices in
Section 3 and
Section 4 was not of a smaller order of magnitude than the condition number of the design matrix in
Section 5.
Furthermore, the errors in
Table 2 and
Table 3, and the objective functions in
Table 4, showed that the solution from
Section 5 provides a better approximation in terms of smallest deviations than the solutions from
Section 3 and
Section 4. It should be noted that the solution in
Section 5 was determined with a different approximation function than the solutions in
Section 3 and
Section 4, as no polynomial terms and constraint equations were introduced. By adding polynomials, see (
27), or by considering additional equations, see (
35), modification of the mathematical function(s) of the problem is made. This leads to the solution of a different problem, since the mathematical function(s) chosen must describe the current problem as well as possible.
The investigations presented up to this point were carried out for a noise level of
for the observations. In order to determine how a larger noise level affects the surface approximation with and without polynomials and constraint equations, the previously described investigations were carried out with a noise level of
. A visualisation of the approximated surfaces based on the techniques in
Section 4 and
Section 5 is shown in
Figure 4.
Obviously, there are no significant differences between the results from both solution approaches. This is also reflected in the a posteriori parameters in
Table 7.
It can be observed that all values show no significant deviations. Therefore, it can be concluded that even for data sets with a higher noise level, the introduction of polynomial terms and the corresponding constraints does not lead to a significant smoothing of the approximated surface for dense point clouds.
Since these findings are so far based on the evaluation of only one data set representing a dense point cloud, another numerical example will be considered in the following. Using the function
the “measurements” (noisy data) are created taking into account a set
, with
, to which
of normally distributed random noise is added. The number of data points is 25, thus resulting in a sparse point cloud of a more curved surface. Moreover, a set
, with
, of uniformly distributed center points, is given. The number of center points is 9. The files with the data points and the center points are taken from the CD in the appendix of the textbook by Fasshauer [
4].
Figure 5 shows the positions of the data points together with the coordinates of the center points.
The original surface and the approximated ones with and without polynomials and constraint equations are shown in
Figure 6.
In contrast to the previous example, it can already be recognised visually that the surface approximated using polynomials and constraint equations is smoother than the one without taking them into account. However, a stronger smoothing of the approximated surface immediately leads to the fact that it no longer fits the original data set so well. This can also be observed in the comparison of the a posteriori parameters in
Table 8. Since the smoother surface leads to larger residuals, the value
, for example, is significantly larger than the value for an approximation without polynomials and the corresponding constraint equations.
Regarding applications with very high noise levels in the input data, in addition to the investigations with noise levels
and
, the previously described calculations are carried out with higher noise levels of up to
.
Figure 7 shows the value of the objective function
for different noise levels.
Firstly, it can be stated that the value of the objective function (
10) increases with increasing noise level, because the residuals become larger. Comparing the solutions with and without consideration of polynomials and corresponding constraint equations, it can be observed that up to a noise level of
, the value of the objective function
is larger with consideration of polynomials and constraint equations than without their consideration. This indicates a stronger smoothing of the approximated surface, which then results in larger residuals. At a noise level of
, the solution with polynomials and constraint equations yields a slightly lower value of the objective function, above which the situation described previously is again apparent.
To evaluate how the solutions at different noise levels for the observations affect the shape of the approximated surface, the
RMS–error (
46) can be used, see
Figure 8, as it describes the deviation of the approximated surface from the original one given by (
47).
It can be observed that for a noise level up to , the resulting surfaces are almost identical, both for the solutions with and without polynomials and constraint equations. For noise levels from to , there is still a good agreement with the original surface, whereas at , a larger deviation from it can be observed. From a noise level of onwards, it can be seen in the investigated example that the solution with polynomials and constraint equations leads to a lower RMS–error. This indicates that the polynomials and the constraint equations can have a positive effect on the resulting approximated surface for sparse point clouds with a very high noise level.
The numerical investigations in this section have shown that the consideration of polynomials and the associated constraint equations in the case of dense point clouds have no significant influence on the approximated surface compared to a solution without polynomials. Neighbouring points are so close to each other that the polynomials do not have a significant smoothing effect on the approximated surface. For sparse point clouds, on the other hand, a smoothing effect could be observed. It should be noted, however, that in this case, larger residuals result, since the approximated surface no longer fits the given data points so well. Based on the RMS–error it could be observed that the approximated surface agrees very well with the original surface known in the numerical example up to a noise level of , both with and without consideration of polynomials and constraint equations. The influence of the polynomial terms on the resulting surface is further investigated in the following section by means of the interpolation of sparse point clouds.
7. Comparison of Interpolation Problems
Based on the interpolation technique that Hardy [
10] introduced, eight interpolation problems are considered here with very few discrete data points. Four different sets of points
, with
and
, are given. For the interpolation process, the center points
with
, of each problem, are equal and identical to the data sites
.
For every set of given points, the interpolation problem with thin-plate splines is solved twice. First is case (a), by using the function given in (
42) where no added polynomials or constraint equations are considered, and then is case (b), by using the function given in (
27) where three linear polynomials and three constraint equations given in (28) are involved. This results in eight solutions for the four given data sets. For each problem, the interpolation condition is defined as
For case (a) the system of linear equations has the form
where
is a square matrix of dimension
and it is set up as in (
30). The vector
contains the unknown parameters and
.
For case (b), the system of linear equations with additional constraint equations
is built with
to be set up as shown in (
30) while
is a
zero matrix. Also, in this case,
is a square matrix of dimension
. The vector
contains the unknown parameters associated with the radial basis functions and the unknown parameters of the polynomials, see (
27).
Table 9,
Table 10,
Table 11 and
Table 12 show the given data sets for each numerical example.
Figure 9,
Figure 10,
Figure 11 and
Figure 12 present the interpolated surfaces obtained from both cases where in case (a) no polynomials were added to the interpolation function and no constraint equations were introduced, and in case (b) three polynomials were added to the interpolation function and three constraint equations were taken into account.
In
Figure 9,
Figure 10,
Figure 11 and
Figure 12, the same phenomenon can be observed for each interpolation problem. This is, when the interpolation was performed without additional polynomials in the interpolation function and without constraint equations, see case (a), the interpolated surface has larger curvatures. On the other hand, when the interpolation was performed with additional polynomials and constraint equations, see case (b), the interpolated surface is smoother, and in two solutions, even has the shape of a plane, see
Figure 9b and
Figure 11b.
The goal of determining an interpolated surface that is as smooth as possible was probably the motivation for Duchon [
8]’s investigations. Hardy [
10] did not explicitly look for a surface that was as smooth as possible, but was rather interested in the surface tangents to coincide at specified points. Using numerical examples, it could be shown that the addition of polynomials and constraint equations can influence the smoothness of the interpolated surface, in some cases considerably, depending on the data set used.
Furthermore, it can be determined from the numerical investigations that the interpolation matrices
in case (a) (without consideration of polynomials and constraint equations) as well as
in case (b) (with consideration of polynomials and constraint equations) have full rank in each case. The condition numbers (
23) of the matrices
and
are listed in
Table 13 and
Table 14.
The values in
Table 13 and
Table 14 show the interesting fact that the condition numbers in case (b) (with consideration of polynomials and constraint equations) each have a larger value than for case (a) (without consideration of polynomials and constraint equations). In particular, for the examples with data sets no. 1 and no. 2, the condition number yields
in case (a), which means that the matrix
is well-conditioned in both examples. In contrast, the condition numbers of matrices
for the examples of the same data sets in case (b) are five times larger and the matrices are thus worse conditioned than in case (a). Similar conclusions can be drawn for the condition numbers of the examples with data sets no. 3 and no. 4.
In case (a), the examples with data sets no. 1 and no. 2 have interpolation matrices that are well-conditioned and this means that no digits are lost in the calculation of the solution with . For the examples with data sets no. 3 and no. 4, only 1 digit is lost when computing the solution with . In case (b), 1 digit is lost in the computation of the solution for the examples with data sets no. 1 and no. 2 with , while 2 digits are lost for the examples with data sets no. 3 and no. 4 with .
These numerical investigations have shown that by adding polynomials to the interpolation function and considering constraint equations, the interpolated surface is smoother, but the condition number of the interpolation matrix increases, leading to a loss of digits in the solution.
Finally, some considerations about the well-posedness of interpolation problems should be made. According to Isaacson and Keller [
19] (p. 21), a problem is defined as well-posed if, and only if, three requirements are met. These are:
A solution exists for the given data.
The solution is unique (meaning that when the computation is performed several times with the same data, identical results are obtained).
The solution depends continuously on the data with a constant that is not too large (meaning that “small” changes in the data, should result in only “small” changes in the solution).
ad 1. and 2.: For all interpolation problems, either of case (a) or of case (b), a solution exists. Additionally, this solution is unique since the interpolation matrix has, for each problem, full rank.
ad 3.: The numerical examples with data sets no. 1 and no. 2, of case (a), yielded a well-conditioned interpolation matrix. This means that for these examples, the solution depends continuously on the data. In contrast, in all other examples where the interpolation matrix is conditioned worse, an arbitrary “small” change in the data can lead to a “not so small” change in the solution for the parameters to be determined.
Consequently, based on the three requirements listed above, the problems with data sets no. 1 and no. 2, of case (a), are well-posed.
Regarding the well-posedness of interpolation with thin-plate splines, the following interesting statement by Fasshauer [
4] (p. 71) is cited:
“There is no result that states that interpolation with thin-plate splines (or any other strictly conditionally positive definite function of order ) without the addition of an appropriate degree polynomial is well-posed.”
However, the investigations carried out in this section have shown that surface interpolation with thin-plate splines can lead to a well-posed interpolation problem even without the addition of polynomial terms and corresponding constraint equations. Therefore, the above statement by Fasshauer [
4] (p. 71) cannot be generalized for every interpolation problem with thin-plate splines.
8. Conclusions and Outlook
In this article, the scattered data approximation problem with thin-plate splines using the RBF approach was investigated. Using a numerical example chosen from the literature, it was shown that a solution is possible without the addition of polynomial terms and corresponding constraint equations. In addition, it was shown that in the solution selected from the literature, the constraint equations were merely introduced as additional observation equations in a determination of the unknown parameters according to the method of least squares. As a result, the constraint equations were not strictly fulfilled by the estimated parameters. Consequently, a solution of the approximation problem according to the method of least squares was presented under rigorous consideration of the constraint equations. Furthermore, it was shown that the addition of polynomial terms to the approximation function did not lead to an improvement of the condition number for the systems of equations to be solved.
The impact of the addition of polynomials and the corresponding constraint equations on the shape of the resulting surface in approximation and interpolation with thin-plate splines was investigated using numerical examples. The investigations carried out led to the conclusion that the addition of polynomial terms and corresponding constraint equations is
not a necessary requirement for both surface interpolation and approximation to compute a solution at all, since the coefficient matrix (
30) has full (column) rank. Modification of the functional model of the problem, i.e., by adding polynomials, should only be motivated by the user’s intention to obtain a solution with certain properties.
In the case of
sparse point clouds, a significant smoothing of the resulting surface can be achieved by introducing polynomial terms and corresponding constraint equations. In the case of approximation, it was found that the resulting surface using polynomials and constraint equations was smoother than the surface without consideration of polynomials, see
Figure 6. In the case of interpolation, the consideration of polynomials and constraint equations could yield a smoothing to such an extent that a plane resulted as interpolating surface, see
Figure 9 and
Figure 11.
In the case of
dense point clouds, such as those obtained from terrestrial laser scanning, only a small influence of the polynomial terms on the approximating surface is to be expected. This can be seen, for example, when comparing the approximation with polynomials in
Figure 3b with the one without polynomials in
Figure 3c on the basis of the absolute numerical differences between the values of the original surface and the values of the approximated ones. Even when the noise level was increased from
to
, there were no significant differences in the resulting approximated surfaces with and without polynomials and constraint equations, see
Figure 4.
In the investigations carried out, the data sites were regarded as fixed (error-free) values, which leads to a linear adjustment problem that is quite easy to solve. In future investigations, the data sites should also be considered as measurements that are subject to random errors, which then leads to a non-linear adjustment problem.