1. Introduction
The classical regression model plays an important role in many practical applications. The definition of this model is shown by
The aim of this conventional regression model is to estimate the unknown regression function
by observed data
. For this classical regression model, many important and interesting results have been obtained by Hart [
1], Kerkyacharian and Picard [
2], Chesneau [
3], Reiß [
4], Yuan and Zhou [
5], and Wang and Politis [
6].
Recently, Chesneau et al. [
7] studied the following regression model
where
are independent and identically distributed random variables,
f is an unknown function defined on
,
are
n identically distributed random vectors,
and
are identically distributed random variables. Moreover,
and
are independent,
and
are independent for any
. The aim of this model is to estimate the unknown function
by the observed data
.
For the above model (
1), it reduces to the classical regression model when
. In other words, (
1) can be viewed as an extension of the classical regression problem. In addition, model (
1) becomes the classical heteroscedastic regression model when
is a function of
(
). Then, the function
is called a variance function in a heteroscedastic regression model, which plays a crucial role in financial and economic fields (Cai and Wang [
8], Alharbi and Patili [
9]). Furthermore, the regression model (
1) is also widely used in Global Positioning Systems (Huang et al. [
10]), Image processing (Kravchenko et al. [
11], Cui [
12]), and so on.
For this regression model, Chesneau et al. [
7] propose two wavelet estimators and discuss convergence rates under the mean integrated square error over Besov space. However, this study only focuses on the global error of wavelet estimators. There is a lack of pointwise risk estimation for this model. In this paper, two new wavelet estimators are constructed, and the convergence rates over the pointwise error of wavelet estimators in local Hölder space are considered. More importantly, those wavelet estimators can all obtain the optimal convergence rate under pointwise error.
2. Assumptions, Local Hölder Space and Wavelet
In this paper, we will consider model (
1) with
. Additional technical assumptions are formulated below.
A1: is bounded for any .
A2: .
A3: .
A4: has a moment of order 2.
A5: and are independent for any .
A6: , where is known and bounded.
For the above assumptions, it is easy to see that A5 and A6 are reversed. Hence, we will define the following two sets, H1 and H2, of the above assumptions
Note that the difference between H1 and H2 is the relationship between and . Since the above assumptions are separated into two sets, H1 and H2; the estimators of the function should be constructed under different condition sets, respectively.
This paper will consider nonparametric pointwise estimation in local Hölder space. Now, we introduce the concept of local Hölder space. Recall the classic Hölder condition
,
Let
be a neighborhood of
and a function space
be defined as
where
is a fixed constant. Clearly,
must be contained in
. However, the converse does not hold.
For
with
and
(the nonnegative integer set), we define the local Hölder space as
Furthermore, it follows from the definition of that
In order to construct wavelet estimators in later sections, we introduce some basic theories of wavelets.
Definition 1. A multiresolution analysis (MRA) is a sequence of closed subspaces of the square-integrable function space satisfying the following properties:
- (i)
;
- (ii)
(the space is dense in ;
- (iii)
if and only if for each ;
- (iv)
There exists (scaling function) such that forms an orthonormal basis of .
Let ϕ be a scaling function, and ψ be a wavelet function such thatconstitutes an orthonormal basis of , where is a positive integer, and . In this paper, we choose the Daubechies wavelets. Then for any , it has the following expansionwhere , . Further details can be found in Meyer [13] and Daubechies [14]. Let be the orthogonal projection operator from onto the space with the orthonormal basis . Then for and , In this position, we give an important lemma, which will be used in later discussions. Here and after, we adopt the following symbol: denotes for some constant ; means ; stand for both and .
Lemma 1 (Liu and Wu [
15]).
If , with , then for and ,- (i)
;
- (ii)
;
- (iii)
.
3. Linear Wavelet Estimator
In this section, a linear wavelet estimator is given by using the wavelet method, and the order of pointwise convergence of this estimator is studied in local Hölder space. Now we define our linear wavelet estimator
where
According to the definition of , it is clear that the structure of this linear wavelet estimator depends on the reverse conditions of A5 and A6. Some of the lemmas needed in this section and their proofs are given below.
Lemma 2. For model (1), if H1 or H2 hold, Proof. According to the definition of
Since
is independent from
and
, respectively,
In addition, condition A3 implies that
. Then one gets
It follows from A5, A2 and A4 that
On the other hand, we obtain
with condition A6.
Finally, according to the assumption of A3 and A2,
□
In order to estimate , we need the following Rosenthal’s inequality.
Rosenthal’s inequality Let be independent random variables such that and ,
- (i)
- (ii)
Lemma 3. Let be defined by (3). If H1 or H2 hold and , then for Proof. By (
5) and the definition of
,
with
. It is clear that
. Using the definition of
and A1, there exists a constant
such that
When
, according to Rosenthal’s inequality,
Note that
Furthermore, it follows from A1 and the property of
that
Then it can be easily seen that
By (
8) and (
9), we obtain
It follows from (
7), (
10) and (
11) that
This with
implies that
□
Now the convergence rate of the linear wavelet estimator is proved in the following.
Theorem 1. Let with . Then for each , the linear wavelet estimator defined in (2) with satisfies Remark 1. Note that is the optimal convergence rate over pointwise error for nonparametric functional estimation (Brown and Low [16]). The above result yields that the linear wavelet estimator can obtain the optimal convergence rate. Proof. The triangular inequality gives
The bias term
. According to Lemma 1,
The stochastic term
. Note that
with
. According to the Hölder inequality, Lemma 3 and
, the above inequality reduces to
Combining (
13), (
14) and (
15), one has
Furthermore, by the given choice
,
□
4. Nonlinear Wavelet Estimator
According to the definition of the linear wavelet estimator, we can easily find that the scale parameter
of the linear wavelet estimator depends on the smooth parameter
s of the function
to be estimated, so the linear estimator is not adaptive. In this section, we will solve this problem by constructing a nonlinear wavelet estimator with the hard thresholding method. Now we define our nonlinear wavelet estimator
where
is defined by (
3),
and
,
denotes the indicator function over an event G. The positive integer
, and
will be given in Theorem 2.
Remark 2. Compared with the structure of in Chesneau et al. [7], the definition of in this paper does not need a thresholding algorithm. In other words, this paper reduces the complexity of the nonlinear wavelet estimator. Lemma 4. For model (1), if H1 or H2 hold, then Lemma 5. Let be defined by (17). If H1 or H2 hold and , then for The proof methods of Lemmas 4 and 5 are similar to that of Lemmas 2 and 3, so the proofs are omitted here. For nonlinear wavelet estimation, Bernstein’s inequality plays a crucial role.
Bernstein’s inequality Let
be independent random variables such that
,
and
, then for each
Lemma 6. Let be defined by (17), and . If H1 or H2 hold, then for each , there exists a constant such that Proof. According to the definition of
,
with
. Clearly,
. Furthermore, by A1 and the property of
,
and
. Note that
Using Bernstein’s inequality,
and
,
Then one chooses a large enough
such that
□
Theorem 2. Let with . Then for each , the nonlinear wavelet estimator defined in (16) with and satisfies Remark 3. Compared with the linear wavelet estimator, the nonlinear wavelet estimator does not depend on the smooth parameter of . Hence, the nonlinear estimator is adaptive. More importantly, the nonlinear estimator can also achieve the optimal convergence rate up to an factor.
Proof. By the definition of
and
, one has
For
. It follows from (
15) and
that
For
. Using Lemma 1 and
, one gets
Then equality (
19) will be proven if we can show
According to Hölder inequality,
Hence, one can obtain that
where
For
. By Hölder inequality
and
Furthermore, using the Cauchy–Schwarz inequality, Lemmas 5 and 6, one has
This with (
22) yields that
Hence,
where
is chosen to be large enough such that
in Lemma 6. This with the choice
shows that
For
. Let us first define
Clearly,
. Note that
Similar to the argument of (
15), one gets
On the other hand, by Hölder inequality
and Lemma 1
Combing (
26), (
27) and
, one gets
Then according to (
29), (
30) and
, one can obtain
Furthermore, together with (
25) and (
28), this yields
Finally, it follows from (
20), (
21) and (
32) that
which completes the proof of Theorem 2. □
5. Conclusions
This paper studies the pointwise estimations of an unknown function in a regression model with multiplicative and additive noise. Under some different assumptions, linear and nonlinear wavelet estimators are constructed. It is clear that those wavelet estimators have diverse forms with different conditions. The convergence rates over the pointwise risk of two wavelet estimators are proposed by Theorems 1 and 2. It should be pointed out that the linear and nonlinear wavelet estimators can all obtain the optimal convergence rate of pointwise nonparametric estimation. More importantly, the nonlinear wavelet estimator is adaptive. In other words, the conclusions of asymptotic and theoretical performance are clear in this paper. However, it is a difficult problem to give numerical experiments, which need more investigations and new skills. We will study it in the future.
Author Contributions
Writing—original draft, J.K. and Q.H.; Writing—review and editing, H.G. All authors have read and agreed to the published version of the manuscript.
Funding
Junke Kou is supported by the National Natural Science Foundation of China (12001133) and Guangxi Natural Science Foundation (2019GXNSFFA245012). Huijun Guo is supported by the National Natural Science Foundation of China (12001132), and Guangxi Colleges and Universities Key Laboratory of Data Analysis and Computation.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
The authors would like to thank the anonymous reviewers for their helpful comments.
Conflicts of Interest
The authors state that there is no conflict of interest.
References
- Hart, J.D. Kernel regression estimation with time series errors. J. R. Stat. Soc. Ser. B 1991, 53, 173–187. [Google Scholar]
- Kerkyacharian, G.; Picard, D. Regression in random design and warped wavelets. Bernoulli 2004, 10, 1053–1105. [Google Scholar] [CrossRef]
- Chesneau, C. Regression with random design: A minimax study. Stat. Probab. Lett. 2007, 77, 40–53. [Google Scholar] [CrossRef]
- Reiß, M. Asymptotic equivalence for nonparametric regression with multivariate and random design. Ann. Stat. 2008, 36, 1957–1982. [Google Scholar] [CrossRef]
- Yuan, M.; Zhou, D.X. Minimax optimal rates of estimation in high dimensional additive models. Ann. Stat. 2016, 44, 2564–2593. [Google Scholar] [CrossRef]
- Wang, L.; Politis, D.N. Asymptotic validity of bootstrap confidence intervals in nonparametric regression without an additive model. Electron. J. Stat. 2021, 15, 392–426. [Google Scholar] [CrossRef]
- Chesneau, C.; Kolei, S.E.; Kou, J.K.; Navarro, F. Nonparametric estimation in a regression model with additive and multiplicative noise. J. Comput. Appl. Math. 2020, 380, 112971. [Google Scholar] [CrossRef]
- Cai, T.T.; Wang, L. Adaptive variance function estimation in heteroscedastic nonparametric regression. Ann. Stat. 2008, 36, 2025–2054. [Google Scholar] [CrossRef]
- Alharbi, Y.F.; Patili, P.N. Error variance function estimation in nonparametric regression models. Commun. Stat. Simul. Comput. 2018, 47, 1479–1491. [Google Scholar] [CrossRef]
- Huang, P.; Pi, Y.; Progri, I. GPS signal detection under multiplicative and additive noise. J. Navig. 2013, 66, 479–500. [Google Scholar] [CrossRef]
- Kravchenko, V.F.; Ponomaryov, V.I.; Pustovoit, V.I.; Palacios-Enriquez, A. 3D Filtering of images corrupted by additive-multiplicative noise. Dokl. Math. 2020, 102, 414–417. [Google Scholar] [CrossRef]
- Cui, G. Application of addition and multiplication noise model parameter estimation in INSAR image Processing. Math. Probl. Eng. 2022, 2022, 3164513. [Google Scholar] [CrossRef]
- Meyer, Y. Wavelets and Operators; Cambridge University Press: Cambridge, UK, 1992. [Google Scholar]
- Daubechies, I. Ten Lecture on Wavelets; SIAM: Philadelphia, PA, USA, 1992. [Google Scholar]
- Liu, Y.M.; Wu, C. Point-wise estimation for anisotropic densities. J. Multivar. Anal. 2019, 171, 112–125. [Google Scholar] [CrossRef]
- Brown, L.D.; Low,, M.G. A constrained risk inequality with applications to nonparametric functional estimation. Ann. Stat. 1996, 24, 2524–2535. [Google Scholar] [CrossRef]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).