Abstract
The theory of reproducing kernel Hilbert spaces (RKHSs) has been developed into a powerful tool in mathematics and has lots of applications in many fields, especially in kernel machine learning. Fractal theory provides new technologies for making complicated curves and fitting experimental data. Recently, combinations of fractal interpolation functions (FIFs) and methods of curve estimations have attracted the attention of researchers. We are interested in the study of connections between FIFs and RKHSs. The aim is to develop the concept of smooth fractal-type reproducing kernels and RKHSs of smooth FIFs. In this paper, a linear space of smooth FIFs is considered. A condition for a given finite set of smooth FIFs to be linearly independent is established. For such a given set, we build a fractal-type positive semi-definite kernel and show that the span of these linearly independent smooth FIFs is the corresponding RKHS. The nth derivatives of these FIFs are investigated, and properties of related positive semi-definite kernels and the corresponding RKHS are studied. We also introduce subspaces of these RKHS which are important in curve-fitting applications.
1. Introduction
The basic theory of reproducing kernel Hilbert spaces (RKHSs) was studied by Aronszajn [1], and it has been developed to be a powerful tool in operator theory, differential equations, integral equations, probability, statistics, and learning theory. See the excellent monographs [2,3,4,5,6,7] and references therein. Nowadays, learning algorithms in RKHSs play an important role in the development of machine learning. A linear model that can only be formulated as an inner product can be transformed into a nonlinear model by replacing the inner product with a symmetric positive semi-definite kernel. We use a kernel to map input data implicitly into a high-dimensional feature space to improve the performance of a learning algorithm. This is a widely used principle called the “kernel trick”. In many applications, data are arising from unknown functions, and it is required to generate a function to interpolate or approximate these data. The concept of RKHSs has been widely applied to such regression problems.
Various types of functions, such as polynomials, splines, rational functions, trigonometric functions, and wavelets have been widely-used in real-world data-fitting. However, sampled signals may have irregular forms in many practical problems, and fractal theory provides new technologies for making complicated curves and fitting experimental data. The theory of fractal interpolation functions (FIFs) is developed for the interpolation problem with a class of fractal functions. It generalizes traditional interpolation techniques through the property of self-similarity. The concept of FIFs defined through an iterated function system was introduced by Barnsley [8,9]. See also these books [10,11,12] for developments of the theory of FIFs and their applications.
Smooth FIFs have also been discussed by many authors. A construction of -FIFs was given in [13]. Based on this work, -Hermite FIFs were obtained in [14,15], -cubic spline FIFs were discussed in [16], and smooth rational cubic FIFs were investigated in [17,18,19,20]. Error bounds, shape-preserving properties, and parameter identification of smooth FIFs have been extensively discussed in the literature.
In [21,22,23], linear FIFs and recurrent linear FIFs were applied to model discrete sequences. In [24], estimations in RKHSs with dependent data were investigated. Recently, the combination of FIFs and other types of curve estimations has attracted the attention of researchers. In [25], the authors studied a fractal perturbation of a type of nonparametric curve estimation. In [26], a training set of samples was used to train an SVM model, and then a linear FIF was constructed based on the predicted data of SVM. In [27], fractal-type reproducing kernels and RKHSs of fractal functions were established. In [28], the author showed that a set of FIFs is an RKHS under two different types of inner products, and then apply such RKHSs to curve-fitting problems. Through the work given in [27,28], connections between FIFs and RKHSs are clearer and we see a new direction of research in the theory of FIFs and RKHSs. Since smooth FIFs have been studied by many researchers, it is natural to develop the concept of RKHSs of smooth FIFs. The purpose of this paper is to construct smooth fractal-type reproducing kernels and RKHSs of smooth FIFs.
Throughout this paper, let be a given set of real numbers such that , where is a positive integer and , and let , For each , let . We will denote by the set of all real-valued continuous functions defined on . Define for . For a nonnegative integer , let denote the space of all real-valued functions whose th derivatives exist and are continuous on . We also denote by in the case .
This paper is structured as follows. In Section 2, we give a brief introduction to the construction of smooth FIFs by the approach given in [13]. For given numbers and a nonnegative integer , a class of FIFs in is established. In Section 3, suppose that and are fixed, and all vertical scaling factors in the construction of FIFs are also fixed numbers. We consider a linear space of smooth FIFs since linear combinations of smooth FIFs are also smooth FIFs. A condition for a set of smooth FIFs to be linearly independent is given. Then we establish a fractal-type positive semi-definite kernel by functions in , and show that is a RKHS with the reproducing kernel . A subspace of , which is important in curve fitting problems, is considered. To investigate the space of th derivatives of functions in , we consider the space and show that is a RKHS with a reproducing kernel defined by , where is the th derivative of . Two subspaces of , and , are considered. We investigate connections between and , and prove that if the ranks of and are both equal to , then . Hence for any function in , we have two equivalent representations for the th derivative of .
2. Construction of Smooth Fractal Interpolation Functions
The construction of smooth FIFs given here has been treated in [13]. We show the details here to make our paper more self-contained.
Let be a nonnegative integer. For , suppose , and define and by
Here , and these numbers are called vertical scaling factors. Assume that
We also define
for , and . Here is the th derivative of and .
In the case , , and it is proved in ([8] Theorem 1) that if
and
then determines a FIF . We call a FIF corresponding to the set of scaling factors and functions . Moreover, the obtained FIF satisfies , and
In the case , it is proved in ([13] Theorem 2) that if
then determines a FIF , and is the FIF determined by for . Moreover, for ,
We see that is a FIF corresponding to the set
By and Equation , conditions and imply that
which are consistent with that given in Equation with . The conditions for can be reduced to
By Equation, condition Equation can be reduced to
for and . Note that in the case , Equation is reduced to Equation .
3. Reproducing Kernel Hilbert Spaces of Smooth Fractal Interpolants
Throughout this section, suppose that is a fixed nonnegative integer, is a fixed positive integer, and , are all fixed numbers. Let be given by Equation .
3.1. Linear Spaces of FIFs and a Condition for Linearly Independent FIFs
Suppose that is a set of functions in such that each in is a -FIF corresponding to a set , where each is a function in , and Equation holds for and . By the results given in Section 2 and Equation , we see that is a FIF corresponding to the set .
Proposition 1.
is a linear space.
Proof.
It is easy to see that the zero function is in with for each . Suppose . Then and are -FIFs corresponding to the sets and , respectively, where for , and satisfy Equation . Therefore, for and ,
and the following conditions hold for and :
Let . Then satisfies
for and . Here and satisfies
for and . Therefore and is corresponding to the set . Moreover, is a FIF corresponding to the set . □
When considering a subspace of , we are usually interested in a set of linearly independent functions in . Each FIF in is defined by Equation and is not given in an explicit form. To determining whether a set of FIFs is linearly independent may not be a trivial task. Here we investigate this problem. Let be a positive integer. Suppose that, for , and is corresponding to the set , where Equation holds with being replaced by , . In the following, we give a condition for to be linearly independent.
Proposition 2.
If are linearly independent on for some , then are linearly independent on .
Proof.
For , is in and is corresponding to the set . By Equation with , we see that also satisfies
for .
If are linearly dependent on , then there exist , not all zero, such that on . This implies on and hence are linearly dependent on for each . Therefore, if are linearly independent on for some , then we have a set of linearly independent FIFs on . □
Similarly, we have
for , . If are linearly independent on for some , then are linearly independent on .
3.2. Fractal-Type Positive Semi-Definite Kernels and RKHSs of FIFs
Suppose that an inner product on is defined. Let be a positive integer. Let , where each is a FIF corresponding to a set , where Equation holds with being replaced by . Suppose that are linearly independent on for some . Then by Proposition 2, are linearly independent on . Let
Then is a finite-dimensional Hilbert space with a basis . Let . By ([6] Proposition 2.23), we see that is a positive definite matrix and hence is invertible.
Define
where the matrix is the inverse of . The following theorem shows that is a reproducing kernel Hilbert space and is the reproducing kernel. Similar results can be found in [27,28].
Proposition 3.
The function defined by Equation is positive semi-definite. The space is a finite-dimensional reproducing kernel Hilbert space with the reproducing kernel .
Proof.
Since is symmetric, let be the matrix such that . Let be any positive integer and let be any choice of distinct points in . Let . Then for any column vector in ,
This shows that is positive semi-definite.
Let and we write in the form
Then for and ,
We also have for .
By Equation with , we have
Define:
It is easy to see that . In fact, for , we can write in the form and then by Equation ,
The following well-known result shows the role of the subspace . □
Proposition 4.
For any , there exists a function such that for , and
Proof.
For , let be the orthogonal projection of on . Then is orthogonal to the subspace . By Equation with , , we have
This proposition is proved by choosing . □
For , . If is a function in , then can be represented by . We write in the form , then
3.3. RKHSs Defined by the Derivatives of Functions in
In Section 3.2, we establish a positive semi-definite kernel by a set of linearly independent functions in , and show that the span of these functions is a reproducing kernel Hilbert space and is the reproducing kernel. Since all are functions in and derivatives of are still FIFs, it is quite nature to investigate those RKHSs which are spanned by the derivatives of .
Suppose and . If are linearly independent on for some , then are linearly independent on . Let and
Each is a FIF corresponding to the set . Define
where the matrix is the inverse of the matrix . We see that . By a similar approach given in the proof of Theorem 3, we see that is positive semi-definite and we can write in the form
For and for ,
Therefore, we have the following theorem.
Proposition 5.
The space is a finite-dimensional reproducing kernel Hilbert space with the reproducing kernel defined by Equation .
By Equation with , we have
Define
It is easy to see that . In fact, for , we can write and then
The following result shows the role of the subspace .
Proposition 6.
For any , there exists a function such that for , and .
Proof.
For , let be the orthogonal projection of on . Then is orthogonal to the subspace . By Equation with , , we have
This proposition is proved by choosing . □
If is a function in , then can be represented by . For , we write as the form , and then
A function in can be written as the form . We have , where is the th derivative of defined by Equation and it can be written as
Since is a linear combination of , it is natural to consider the space
when we discuss the th derivative of . Recall that . Is the subspace identical to the subspace ? We discuss this question below. Here both sets of functions and are supposed to be linearly independent.
We first consider functions in . Suppose . We can write
Then by Equations and ,
Since are linearly independent on , we have
The system Equation can be written in the matrix form
where , , and , are matrices , , respectively. Here is given in Equation and is given in Equation . Hence if and only if there exist vectors and that satisfy Equation , and in this case, is given by Equation .
Since is invertible, we have
If the rank of is equal to , then this equation is consistent for any vector . Conversely, since is also invertible, Equation implies that
If the rank of is equal to , then this equation is consistent for any vector . We have the following theorem.
Proposition 7.
If the ranks of and are both equal to , then .
Proof.
For any , we write in the form . If the rank of is equal to , then by setting , the Equation has a solution , and can be written as . This shows that . Conversely, for any , we write in the form . If the rank of is equal to , then by setting , the Equation has a solution , and can be written as . This implies that . □
For , we write and . If is a solution of Equation , then we can write and
We can investigate the relationship between and by Equation . If is the vector such that and for , where , then the left-hand side of Equation is reduced to the -th column of the matrix , and if is a solution of Equation , then
If is the vector such that and for , where , then the right-hand side of Equation is reduced to the -th column of the matrix , and if is a solution of Equation , then
Consider the particular case . If the ranks of and are both equal to , then and are both invertible. By Equation ,
4. Conclusions
The theory of RKHSs and the concept of FIFs play important roles in mathematics and have variety of applications in many fields. The work given in [27,28] bridged the gap between RKHSs and FIFs. In this paper we study RKHSs of smooth FIFs and their th derivatives .
First, a linear space of -FIF is established. Then we consider a linearly independent set of functions in . A condition for the independent property of functions in is established. We show that the space is a finite-dimensional RKHS with the reproducing kernel defined by Equation . Define , which is a subspace of and plays an important role in curve fitting applications. If is a function in , then can be represented by .
To investigate the RKHS of th derivatives of functions in , we define and . is a finite-dimensional RKHS with the reproducing kernel defined by Equation . For , . We also define a subspace of , where . We prove that if the ranks of and are both equal to , then . Here and are both matrices. If is a function in , then can be represented by . Note that can be represented by and also by , the two representations are connected by the Formula . Moreover, it is clear to write each as a linear combination of , and to write each as a linear combination of .
In this paper we establish a finite-dimensional RKHS spanned by a set of linearly independent FIFs, and the RKHS spanned by the th derivatives of these basis functions. An important subspace is introduced, and the subspace of the th derivative of in is investigated. We believe that results established in this paper enrich the theory of RKHSs and FIFs, and may have applications in many fields.
Author Contributions
Conceptualization, D.-C.L. and L.-Y.H.; methodology, D.-C.L.; validation, L.-Y.H.; formal analysis, L.-Y.H.; investigation, D.-C.L.; resources, L.-Y.H.; writing—original draft preparation, D.-C.L.; writing—review and editing, L.-Y.H.; visualization, L.-Y.H.; supervision, D.-C.L.; project administration, D.-C.L. and L.-Y.H.; funding acquisition, D.-C.L. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the Ministry of Science and Technology, R.O.C., under Grant MOST 110-2115-M-214-002.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
References
- Aronszajn, N. Theory of reproducing kernels. Trans. Am. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
- Vapnik, V.N. Statistical Learning Theory; Wileys: New York, NY, USA, 1998. [Google Scholar]
- Berlinet, A.; Thomas-Agnan, C. Reproducing Kernel Hilbert Spaces in Probability and Statistics; Springer: New York, NY, USA, 2004. [Google Scholar]
- Rasmussen, C.E.; Williams, C.K.I. Gaussian Processes of Machine Learning; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
- Cucker, F.; Zhou, D.X. Learning Theory: An Approximation Theory Viewpoint; Cambridge University Press: New York, NY, USA, 2007. [Google Scholar]
- Paulsen, V.I.; Raghupathi, M. An Introduction to the Theory of Reproducing Kernel Hilbert Spaces; Cambridge University Press: Cambridge, UK, 2016. [Google Scholar]
- Pereverzyev, S. An Introduction to Artificial Intelligence Based on Reproducing Kernel Hilbert Spaces; Birkhäuser: Cham, Switzerland, 2022. [Google Scholar]
- Barnsley, M.F. Fractal functions and interpolation. Constr. Approx. 1986, 2, 303–329. [Google Scholar] [CrossRef]
- Barnsley, M.F. Fractals Everywhere; Academic Press: Orlando, FL, USA, 1988. [Google Scholar]
- Massopust, P.R. Fractal Functions, Fractal Surfaces, and Wavelets; Academic Press: San Diego, CA, USA, 2004. [Google Scholar]
- Massopust, P.R. Interpolation and Approximation with Splines and Fractals; Oxford University Press: New York, NY, USA, 2010. [Google Scholar]
- Banerjee, S.; Gowrisankar, A. Frontiers of Fractal Analysis Recent Advances and Challenges; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar]
- Barnsley, M.F.; Harrington, A.N. The calculus of fractal interpolation functions. J. Approx. Theory 1989, 57, 14–34. [Google Scholar] [CrossRef]
- Chand, A.K.B.; Navascués, M.A. Generalized Hermite fractal interpolation. Rev. R. Acad. Cienc. Zaragoza 2009, 64, 107–120. [Google Scholar]
- Navascués, M.A.; Sebastián, M.V. Generalization of Hermite functions by fractal interpolation. J. Approx. Theory 2004, 131, 19–29. [Google Scholar] [CrossRef]
- Chand, A.K.B.; Kapoor, G.P. Generalized cubic spline fractal interpolation functions. SIAM J. Numer. Anal. 2006, 44, 655–676. [Google Scholar] [CrossRef]
- Balasubramani, N. Shape preserving rational cubic fractal interpolation function. J. Comput. Appl. Math. 2017, 319, 277–295. [Google Scholar] [CrossRef]
- Balasubramani, N.; Prasad, M.G.P.; Natesan, S. Shape preserving α-fractal rational cubic splines. Calcolo 2020, 57, 21. [Google Scholar] [CrossRef]
- Viswanathan, P.; Chand, A.K.B. A fractal procedure for monotonicity preserving interpolation. Appl. Math. Comput. 2014, 247, 190–204. [Google Scholar] [CrossRef]
- Viswanathan, P.; Chand, A.K.B.; Agarwal, R.P. Preserving convexity through rational cubic spline fractal interpolation function. J. Comput. Appl. Math. 2014, 263, 262–276. [Google Scholar] [CrossRef]
- Mazel, D.S.; Hayes, M.H. Using iterated function systems to model discrete sequences. IEEE Trans. Signal Process. 1992, 40, 1724–1734. [Google Scholar] [CrossRef]
- Mazel, D.S. Representation of discrete sequences with three-dimensional iterated function systems. IEEE Trans. Signal Process. 1994, 42, 3269–3271. [Google Scholar] [CrossRef]
- Marvasti, M.A.; Strahle, W.C. Fractal geometry analysis of turbulent data. Signal Process. 1995, 41, 191–201. [Google Scholar] [CrossRef]
- Sancetta, A. Estimation in reproducing kernel Hilbert spaces with dependent data. IEEE Trans. Inform. Theory 2021, 67, 1782–1795. [Google Scholar] [CrossRef]
- Luor, D.-C.; Liu, C.-W. Fractal perturbation of the Nadaraya-Watson estimator. Fractal Fract. 2022, 6, 680. [Google Scholar] [CrossRef]
- Wang, H.-Y.; Li, H.; Shen, J.-Y. A novel hybrid fractal interpolation-SVM model for forecasting stock price indexes. Fractals 2019, 27, 1950055. [Google Scholar] [CrossRef]
- Bouboulis, P.; Mavroforakis, M. Reproducing kernel Hilbert spaces and fractal interpolation. J. Comput. Appl. Math. 2011, 235, 3425–3434. [Google Scholar] [CrossRef]
- Luor, D.-C. Reproducing kernel Hilbert spaces of fractal interpolation functions for curve fitting problems. Fractals 2022, 30, 2250037. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).