Next Article in Journal
A Time-Fractional Differential Inequality of Sobolev Type on an Annulus
Next Article in Special Issue
Excess Lifetime Extropy of Order Statistics
Previous Article in Journal
Existence of Multiple Weak Solutions to a Discrete Fractional Boundary Value Problem
Previous Article in Special Issue
Modified Maximum Likelihood Estimation of the Inverse Weibull Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Regression Estimation with Errors in the Variables via the Laplace Transform

School of Mathematics and Computational Science, Guilin University of Electronic Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(10), 992; https://doi.org/10.3390/axioms12100992
Submission received: 14 September 2023 / Revised: 13 October 2023 / Accepted: 18 October 2023 / Published: 19 October 2023
(This article belongs to the Special Issue Mathematical and Statistical Methods and Their Applications)

Abstract

:
This paper considers nonparametric regression estimation with errors in the variables. It is a standard assumption that the characteristic function of the covariate error does not vanish on the real line. This assumption is rather strong. In this paper, we assume the covariate error distribution is a convolution of uniform distributions, the characteristic function of which contains zeros on the real line. Our regression estimator is constructed via the Laplace transform. We prove its strong consistency and show its convergence rate. It turns out that zeros in the characteristic function have no effect on the convergence rate of our estimator.

1. Introduction

This paper considers a regression model with errors in the variables. Suppose observations ( W 1 , Y 1 ) , , ( W n , Y n ) are i.i.d. (independent and identically distributed) random variables generated by the model
W j = X j + δ j , Y j = m ( X j ) + ϵ j , j = 1 , , n .
The i.i.d. random variables δ j are independent of X j and Y j . ϵ j are independent of X j , E ϵ j = 0 and E ϵ j 2 < + . The functions f δ (known) and f X (unknown) stand for the densities of δ j and X j , respectively. The goal is to estimate the regression function m ( x ) from the observations ( W 1 , Y 1 ) , , ( W n , Y n ) . Errors-in-variables regression problems have been extensively studied in the literature, see, for example, ([1,2,3,4,5,6,7]). Regression models with errors in the variables play an important role in many areas of science and social science ([8,9,10]).
Nadaraya and Watson ([11,12]) propose a kernel regression estimator for the classical regression model ( δ j = 0 ) . Since the Fourier transform can transform a complex convolution to an ordinary product, it is a common method to deal with the deconvolution problem. Fan and Truong [4] generalize the Nadaraya–Watson regression estimator from the classical regression model to the regression model (1) via the Fourier transform. They study the convergence rate by assuming the integer order derivatives of f X and m to be bounded. Compared to integer-order derivatives, it is more precise to describe the smoothness by the Hölder condition. Meister [6] shows the convergence rate under the local Hölder condition.
The above references on model (1) both assume that the characteristic function of the covariate errors δ j does not have zeros on the real line. The assumption is rather strong. For example, if f δ is of uniform density on [−1, 1], it vanishes at v = k π , k = ± 1 , ± 2 , in the Fourier domain. Delaigle and Meister [1] consider the regression model (1) with a Fourier-oscillating noise, which means the Fourier transform of f δ vanishes periodically. They show that if f X and m are compact, then they can be estimated with the standard rate, as in the case where f δ does not vanish in the Fourier domain. Guo and Liu ([13,14,15]) extend Delaigle and Meister [1]’s work to multivariate cases.
The compactness is the cost of eliminating the zero points effect in the Fourier domain. Belomestny and Goldenshluger [16] apply the Laplace transform to construct a deconvolution density estimator without assuming the density to be compact. They provide sufficient conditions under which the zeros of the corresponding characteristic function have no effect on the estimation accuracy. Goldenshluger and Kim [17] also construct a deconvolution density estimator via the Laplace transform; they study how zero multiplicity affects the estimation accuracy. Motivated by the above work, we apply the Laplace transform to study the regression model (1) with errors following a convolution of uniform distributions.
The organization of the paper is as follows. In Section 2, we present some knowledge about the covariate error distribution and functional classes. Section 3 introduces the kernel regression estimator via the Laplace transform. The consistency and convergence rate of our estimator are discussed in Section 4 and Section 5, respectively.

2. Preparation

This section will introduce the covariate error distribution and functional classes.
For a integrable function f, the bilateral Laplace transform [18] is defined by
f ^ ( z ) : = + f ( t ) e z t d t .
The Laplace transform f ^ ( z ) is an analytic function in the convergence region Σ f , which is a vertical strip:
Σ f : = { z C : σ f < Re ( z ) < σ f + } , for some σ f < σ f + + .
The inverse Laplace transform is given by the formula
f ( t ) = 1 2 π i s i s + i f ^ ( z ) e z t d z = 1 2 π + f ^ ( s + i v ) e ( s + i v ) t d v , s ( σ f , σ f + ) .
Let the covariate error distribution be a γ -fold convolution of the uniform distribution on [ θ , θ ] , θ > 0 . This means
δ = Z 1 + + Z γ ,
where Z i ( i = 1 , 2 , , γ ) are i.i.d and Z i U ( θ , θ ) with density f Z . Hence,
f ^ δ ( z ) = f ^ Z ( z ) γ = sinh ( θ z ) θ z γ = ( 1 e 2 θ z ) γ ( 2 θ z ) γ e γ θ z , z C .
Here, f ^ δ ( z ) is the product of two functions; the function ( 1 e 2 θ z ) γ has zeros only on the imaginary axis, the function 1 ( 2 θ z ) γ e γ θ z does not have zeros for the analyticity of ( 2 θ z ) γ e γ θ z . The zeros of f ^ δ ( z ) are z k = i k π θ , where k Z { 0 } .
Now, we introduce some functional classes.
Definition 1. 
For A > 0 , σ > 0 , and β > 0 , a function f : R R is said to satisfy the local Hölder condition with smoothness parameter β if f is k times continuously differentiable and
| f ( k ) ( y ) f ( k ) ( y ˜ ) | A | y y ˜ | β 0 , y , y ˜ [ x σ , x + σ ] ,
where β = k + β 0 and 0 < β 0 1 . All these functions are denoted by H σ , β ; x ( A ) .
If (3) holds for any y , y ˜ R , f satisfies the Hölder condition with smoothness parameter β . All these functions are denoted by H β ( A ) .
Clearly, k in Definition 1 equals max { l N : l < β } . In later discussions, β : = max { l N : l < β } .
Example 1. 
Function
f 1 ( x ) : = 1 | x | , | x | 1 , 0 , | x | > 1 .
Then, f 1 H 1 ( A ) and f 1 H σ , 1 ; x ( A ) .
It is easy to see that f H β ( A ) must be contained in H σ , β ; x ( A ) for each x R . However, the reverse is not necessarily true.
Example 2 
([19]). Consider the function
f 2 ( x ) : = l = 0 ( 1 2 l | x 2 l | ) χ l ( x ) ,
where χ l ( x ) is the indicator function on the interval [ 2 l 2 l , 2 l + 2 l ] for a non-negative integer l. Then, f 2 H σ , 1 ; x ( A ) for each x R . However, f 2 H 1 ( A ) .
Note that (3) is a local Hölder condition around x R . When we consider the pointwise estimation, it is natural to assume the unknown function to satisfy a local smoothness condition.
Definition 2. 
Let r > 0 and B > 0 be real numbers. We say that a function f belongs to the functional class M r ( B ) if
max { f , max 0 < r 1 r + | x | r 1 | f ( x ) | d x } B .
We denote F σ , β , r ; x ( A , B ) = H σ , β ; x ( A ) M r ( B ) .

3. Kernel Estimator

This section will construct the kernel regression estimator. Two kernels K and L s , h will be used.
Assume that the kernel K : R R satisfies the following conditions:
(i) 1 1 K ( x ) d x = 1 , K C ( R ) and supp ( K ) [ 1 , 1 ] ;
(ii) There exists a fixed positive integer k 0 such that
1 1 x j K ( x ) d x = 0 , j = 1 , , k 0 .
Example 3 
([20]). Function
K ( x ) = a φ ( x ) ,
where
φ ( x ) : = e 1 1 x 2 , | x | < 1 0 , | x | 1
and a = + φ ( x ) d x 1 . Then, the kernel K ( x ) satisfies conditions (i) and (ii) with k 0 = 1 .
Motivated by Belomestny and Goldenshluger [16], we will construct the regression estimator via the Laplace transform. Note that f ^ δ ( z ) does not have zeros out of the imaginary axis. Then, the kernel L s , h is defined by the inverse Laplace transform
L s , h ( t ) : = 1 2 π i s i s + i K ^ ( z h ) f ^ δ ( z ) e z t d z = 1 2 π + K ^ ( ( s + i v ) h ) f ^ δ ( s i v ) e ( s + i v ) t d v ,
where s 0 , h > 0 and K ^ ( · ) is the Laplace transform of kernel K with the convergence region Σ K = C . There is a complex-valued improper integral in (4). One can use the property of the Laplace transform to compute it, see [18].
The following lemma provides a infinite series of kernel L s , h ( t ) . It is a specific form of Lemma 2 in [16]. In order to explain the construction of the estimator, we give the details of the proof.
Lemma 1. 
Let (2) hold and + | K ^ ( i v h ) | | v | γ d v < .
(a) If s > 0 , then
L + , h ( t ) : = L s , h ( t ) = ( 2 θ ) γ 2 π l 1 = 0 l γ = 0 + K ^ ( i v h ) ( i v ) γ e i v [ t γ θ 2 θ ( l 1 + + l γ ) ] d v .
(b) If s < 0 , then
L , h ( t ) : = L s , h ( t ) = ( 2 θ ) γ 2 π l 1 = 0 l γ = 0 + K ^ ( i v h ) ( i v ) γ e i v [ t + γ θ + 2 θ ( l 1 + + l γ ) ] d v .
Proof. 
(a) If s > 0 , we have
1 1 e 2 θ ( s + i v ) = l = 0 e 2 θ l ( s + i v ) .
Therefore,
1 [ 1 e 2 θ ( s + i v ) ] γ = l = 0 e 2 θ l ( s + i v ) γ = l 1 = 0 l γ = 0 e 2 θ ( s + i v ) ( l 1 + + l γ ) .
By (2) and (4),
L + , h ( t ) : = L s , h ( t ) = 1 2 π + K ^ ( ( s + i v ) h ) [ 2 θ ( s + i v ) ] γ [ 1 e 2 θ ( s + i v ) ] γ e ( s + i v ) ( t γ θ ) d v = ( 2 θ ) γ 2 π l 1 = 0 l γ = 0 + K ^ ( ( s + i v ) h ) ( s + i v ) γ e ( s + i v ) [ t γ θ 2 θ ( l 1 + + l γ ) ] d v = ( 2 θ ) γ 2 π l 1 = 0 l γ = 0 + K ^ ( i v h ) ( i v ) γ e i v [ t γ θ 2 θ ( l 1 + + l γ ) ] d v .
(b) If s < 0 , then
1 e 2 θ ( s + i v ) γ = 1 e 2 θ ( s + i v ) e 2 θ ( s + i v ) γ = ( 1 ) γ e 2 γ θ ( s + i v ) [ 1 e 2 θ ( s + i v ) ] γ = ( 1 ) γ e 2 γ θ ( s + i v ) l 1 = 0 l γ = 0 e 2 θ ( s + i v ) ( l 1 + + l γ ) .
Similarly,
L , h ( t ) : = L s , h ( t ) = ( 2 θ ) γ 2 π l 1 = 0 l γ = 0 + K ^ ( i v h ) ( i v ) γ e i v [ t + γ θ + 2 θ ( l 1 + + l γ ) ] d v .
This ends the proof. □
The truncation is used to deal with infinite series. Select parameter N so that N γ N + . The cut-off kernels are defined by
L + , h ( N ) ( t ) : = ( 2 θ ) γ 2 π l 1 = 0 N γ l γ = 0 N γ + K ^ ( i v h ) ( i v ) γ e i v [ t γ θ 2 θ ( l 1 + + l γ ) ] d v ,
L , h ( N ) ( t ) : = ( 2 θ ) γ 2 π l 1 = 0 N γ l γ = 0 N γ + K ^ ( i v h ) ( i v ) γ e i v [ t + γ θ + 2 θ ( l 1 + + l γ ) ] d v .
Denote
L s , h ( N ) ( t ) : = L + , h ( N ) ( t ) , s > 0 , L , h ( N ) ( t ) , s < 0 .
Motivated by the Nadaraya–Watson regression estimator, we define the regression estimator of m ( x ) as
m ˜ s , h ( N ) ( x ) : = p ˜ s , h ( N ) ( x ) f ˜ X , s , h ( N ) ( x ) ,
where
f ˜ X , s , h ( N ) ( x ) : = 1 n j = 1 n L s , h ( N ) ( W j x ) and p ˜ s , h ( N ) ( x ) : = 1 n j = 1 n Y j L s , h ( N ) ( W j x ) .
In what follows, we will write m ˜ + , h ( N ) ( x ) and m ˜ , h ( N ) ( x ) for the estimator (7) associated with s > 0 and s < 0 , respectively. Finally, our regression estimator is denoted by
m ˜ h ( N ) ( x ) : = m ˜ + , h ( N ) ( x ) , x 0 , m ˜ , h ( N ) ( x ) , x < 0 .

4. Strong Consistency

In this section, we investigate the consistency of the regression estimator (9). Roughly speaking, consistency means that the estimator m ˜ h ( N ) ( x ) converges to m ( x ) as the sample size tends to infinity.
Theorem 1 
(Strong consistency). Consider the model (1) with (2). Suppose f X , p : = m f X M r ( B ) ( r > 1 2 ) , E | Y 1 | 8 ( γ + 1 ) < + and kernel function K satisfies condition (i). If x is the Lebesgue point of both f X and p ( f X ( x ) 0 ) , then m ˜ h ( N ) ( x ) satisfies
lim n m ˜ h ( N ) ( x ) = a . s . m ( x )
with h = n 1 6 ( γ + 1 ) and n 1 3 ( γ + 1 ) N 2 n 1 3 ( γ + 1 ) .
Proof. 
( 1 ) We consider the estimator m ˜ + , h ( N ) ( x ) for x 0 .
Note that m ˜ + , h ( N ) ( x ) = p ˜ + , h ( N ) ( x ) f ˜ X , + , h ( N ) ( x ) , m ( x ) = p ( x ) f X ( x ) and f X ( x ) 0 . Then, it is sufficient to prove lim n p ˜ + , h ( N ) ( x ) = a . s . p ( x ) and lim n f ˜ X , + , h ( N ) ( x ) = a . s . f X ( x ) .
Now, we prove lim n p ˜ + , h ( N ) ( x ) = a . s . p ( x ) . For any ϵ > 0 ,
P | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ P | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | > ϵ 2 + χ ( ϵ 2 , ) ( | E p ˜ + , h ( N ) ( x ) p ( x ) | ) .
By Markov’s inequality, we obtain
P | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ c 1 ϵ s E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s + χ ( ϵ 2 , ) ( | E p ˜ + , h ( N ) ( x ) p ( x ) | )
for s : = 4 ( γ + 1 ) . This motivates us to derive an upper bound on E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s . Combining (5) with (8), we have
p ˜ + , h ( N ) ( x ) = 1 n j = 1 n Y j L + , h ( N ) ( W j x ) = ( 2 θ ) γ 2 π n j = 1 n + K ^ ( i v h ) ( i v ) γ ( Y j e i v W j ) e i v ( x + γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v ,
and
E p ˜ + , h ( N ) ( x ) = ( 2 θ ) γ 2 π n j = 1 n + K ^ ( i v h ) ( i v ) γ E ( Y j e i v W j ) e i v ( x + γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v .
We obtain
E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s = ( 2 θ ) γ 2 π n 2 s E | j = 1 n + K ^ ( i v h ) ( i v ) γ Ψ j ( v ) Φ + , N ( x , v ) d v | 2 s ,
where Ψ j ( v ) : = Y j e i v W j E ( Y j e i v W j ) , Φ + , N ( x , v ) : = e i v ( x + γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) .
Thus,
E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s = ( 2 θ ) γ 2 π n 2 s j 1 = 1 n j 2 s = 1 n E [ ( + + k = 1 s K ^ ( i v 2 k 1 h ) K ^ ( i v 2 k h ) ( i v 2 k 1 ) γ ( i v 2 k ) γ × Φ + , N ( x , v 2 k 1 ) Φ + , N ( x , v 2 k ) Ψ j 2 k 1 ( v 2 k 1 ) Ψ j 2 k ( v 2 k ) d v 1 d v 2 s ) ] = ( 2 θ ) γ 2 π n 2 s j 1 = 1 n j 2 s = 1 n + + [ k = 1 s K ^ ( i v 2 k 1 h ) K ^ ( i v 2 k h ) ( i v 2 k 1 ) γ ( i v 2 k ) γ × Φ + , N ( x , v 2 k 1 ) Φ + , N ( x , v 2 k ) ] E k = 1 s Ψ j 2 k 1 ( v 2 k 1 ) Ψ j 2 k ( v 2 k ) d v 1 d v 2 s ( 2 θ ) γ 2 π n 2 s j 1 = 1 n j 2 s = 1 n + + k = 1 s | K ^ ( i v 2 k 1 h ) | | K ^ ( i v 2 k h ) | | v 2 k 1 | γ | v 2 k | γ × ( N γ + 1 ) 2 γ s | E k = 1 s Ψ j 2 k 1 ( v 2 k 1 ) Ψ j 2 k ( v 2 k ) | d v 1 d v 2 s .
Let # A denote the number of elements contained in the set A. If # { j 1 , , j 2 s } > s , at least one of Ψ j l is independent of all other Ψ j l , l l . Hence,
E k = 1 s Ψ j 2 k 1 ( v 2 k 1 ) Ψ j 2 k ( v 2 k ) = 0 .
On the other hand, if # { j 1 , , j 2 s } = s 1 for s 1 s , by Jensen’s inequality, we obtain
| E k = 1 s Ψ j 2 k 1 ( v 2 k 1 ) Ψ j 2 k ( v 2 k ) | = | E Ψ j 1 ( t 1 ) λ 1 Ψ j s 1 ( t s 1 ) λ s 1 | E | Ψ 1 ( t 1 ) | 2 s E ( | Y 1 | + E | Y 1 | ) 2 s 4 s E | Y 1 | 2 s ,
where λ 1 + + λ s 1 = 2 s . Let J n = { ( j 1 , , j 2 s ) : # { j 1 , , j 2 s } s , j i { 1 , , n } , i = 1 , , 2 s } . Then,
E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s ( 2 θ ) γ 2 π n 2 s ( c 2 N 2 γ s ) ( 4 s E | Y 1 | 2 s ) j J n + | K ^ ( i v h ) | | v | γ d v 2 s .
Since | v | k | K ^ ( i v ) | c ( k ) for all k, we obtain that + | K ^ ( i v h ) | | v | γ d v c 3 h ( γ + 2 ) for k = γ + 2 . This, with # J n c 4 n s , leads to
E | p ˜ + , h ( N ) ( x ) E p ˜ + , h ( N ) ( x ) | 2 s c 5 n s N 2 γ s h 2 s ( γ + 2 ) c 6 n s 3 ( γ + 1 ) .
Inserting this into (10), we obtain
P | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ c 7 n s 3 ( γ + 1 ) + χ ( ϵ 2 , ) ( | E p ˜ + , h ( N ) ( x ) p ( x ) | ) .
Note that ( W j , Y j ) are identically distributed. Then, it follows from (11) and E ( Y j e i v W j ) = E ( Y j e i v X j ) E ( e i v δ j ) that
E p ˜ + , h ( N ) ( x ) = ( 2 θ ) γ 2 π + K ^ ( i v h ) ( i v ) γ E ( Y 1 e i v X 1 ) E ( e i v δ 1 ) e i v ( x + γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v ,
where
l 1 = 0 N γ l γ = 0 N γ e i 2 θ v l 1 + + l γ = l = 0 N γ e i v ( 2 θ l ) γ = 1 ( e i 2 θ v ) N γ + 1 1 e i 2 θ v γ .
By (2), we have
E p ˜ + , h ( N ) ( x ) = 1 2 π + E ( Y 1 e i v X 1 ) K ^ ( i v h ) e i v x 1 ( e i 2 θ v ) N γ + 1 γ d v = 1 2 π + E e i v X 1 · E ( Y 1 | X 1 ) K ^ ( i v h ) e i v x l = 0 γ γ l ( 1 ) l ( e i 2 θ v ) l ( N γ + 1 ) d v = l = 0 γ γ l ( 1 ) l 1 2 π + + p ( t ) e i v t d t K ^ ( i v h ) e i v [ x + 2 θ l ( N γ + 1 ) ] d v = l = 0 γ γ l ( 1 ) l 1 h + p ( t ) K t x 2 θ l ( N γ + 1 ) h d t = + 1 h K ( t x h ) p ( t ) d t + T + , N γ ( p ; x ) ,
where
T + , N γ ( p ; x ) : = l = 1 γ γ l ( 1 ) l 1 h + p ( t ) K t x 2 θ l ( N γ + 1 ) h d t = l = 1 γ γ l ( 1 ) l 1 1 K ( y ) p y h + x + 2 θ l ( N γ + 1 ) d y .
Hence,
| E p ˜ + , h ( N ) ( x ) p ( x ) | | + 1 h K ( t x h ) p ( t ) d t p ( x ) | + | T + , N γ ( p ; x ) | = | 1 1 K ( y ) [ p ( x + y h ) p ( x ) ] d y | + | T + , N γ ( p ; x ) | .
Since p M r ( B ) and considering the boundedness of K,
| T + , N γ ( p ; x ) | c 8 l = 1 γ γ l 1 1 p y h + x + 2 θ l ( N γ + 1 ) d y c 9 B h ( x + 2 θ N · 1 γ ) r c 10 B h N r
holds for an h that is small enough. It follows from r > 1 2 , h = n 1 6 ( γ + 1 ) and N n 1 3 ( γ + 1 ) that
| T + , N γ ( p ; x ) | n 0 .
Note that the kernel function K satisfies condition (i) and p L ( R ) , then
| 1 1 K ( y ) [ p ( x + y h ) p ( x ) ] d y | h 0 0
holds for each Lebesgue point x of p. Hence, for an n that is sufficiently large, the term χ ( ϵ 2 , ) ( | E p ˜ + , h ( N ) ( x ) p ( x ) | ) vanishes. This, with (14), shows
P | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ c 7 n s 3 ( γ + 1 )
for an n that is large enough. Since s = 4 ( γ + 1 ) , we have
n = 1 P | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ < .
For any ϵ > 0 , it follows from the Borel–Cantelli lemma that
P lim n ¯ | p ˜ + , h ( N ) ( x ) p ( x ) | > ϵ = 0 .
Thus,
lim n p ˜ + , h ( N ) ( x ) = a . s . p ( x ) .
When putting Y j 1 almost surely, we have
lim n f ˜ X , + , h ( N ) ( x ) = a . s . f X ( x ) .
Hence,
lim n m ˜ + , h ( N ) ( x ) = a . s . m ( x ) .
(2) We consider the estimator m ˜ , h ( N ) ( x ) for x < 0 . Inserting (6) into (8), we obtain
p ˜ , h ( N ) ( x ) = 1 n j = 1 n Y j L , h ( N ) ( W j x ) = ( 2 θ ) γ 2 π n j = 1 n + K ^ ( i v h ) ( i v ) γ ( Y j e i v W j ) e i v ( x γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v ,
and
E p ˜ , h ( N ) ( x ) = ( 2 θ ) γ 2 π n j = 1 n + K ^ ( i v h ) ( i v ) γ E ( Y j e i v W j ) e i v ( x γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v .
We obtain
E | p ˜ , h ( N ) ( x ) E p ˜ , h ( N ) ( x ) | 2 s = ( 2 θ ) γ 2 π n 2 s E | j = 1 n + K ^ ( i v h ) ( i v ) γ Ψ j ( v ) Φ , N ( x , v ) d v | 2 s ,
where Φ , N ( x , v ) : = e i v ( x γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) . Similar to (12) and (13), we obtain
E | p ˜ , h ( N ) ( x ) E p ˜ , h ( N ) ( x ) | 2 s c 11 n s 3 ( γ + 1 ) .
By (21), we have
E p ˜ , h ( N ) ( x ) = ( 2 θ ) γ 2 π + K ^ ( i v h ) ( i v ) γ E ( Y 1 e i v X 1 ) E ( e i v δ 1 ) e i v ( x γ θ ) × l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v ,
where
l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) = 1 ( e i 2 θ v ) N γ + 1 1 e i 2 θ v γ .
By 1 ( 1 e i 2 θ v ) γ = 1 ( e i 2 θ v ) γ ( 1 e i 2 θ v ) γ and (2), we have that ( i 2 θ v ) γ e i v ( γ θ ) ( 1 e i 2 θ v ) γ = 1 f ^ δ ( i v ) . So,
E p ˜ , h ( N ) ( x ) = 1 2 π + K ^ ( i v h ) E ( Y 1 e i v X 1 ) e i v x 1 ( e i 2 θ v ) N γ + 1 γ d v .
Similar to (17), we obtain
E p ˜ , h ( N ) ( x ) = + 1 h K ( t x h ) p ( t ) d t + T , N γ ( p ; x ) ,
where
T , N γ ( p ; x ) : = l = 1 γ γ l ( 1 ) l 1 h + p ( t ) K t x + 2 θ l ( N γ + 1 ) h d t = l = 1 γ γ l ( 1 ) l 1 1 K ( y ) p y h + x 2 θ l ( N γ + 1 ) d y .
Thus, we have
| E p ˜ , h ( N ) ( x ) p ( x ) | | 1 1 K ( y ) [ p ( x + y h ) p ( x ) ] d y | + | T , N γ ( p ; x ) | .
Since p M r ( B ) and considering the boundedness of K,
| T , N γ ( p ; x ) | c 12 l = 1 γ γ l 1 1 p y h + x 2 θ l ( N γ + 1 ) d y c 13 B h ( x + 2 θ N · 1 γ ) r c 14 B h N r
holds for an h that is small enough.
Similar to x 0 , we get
lim n m ˜ , h ( N ) ( x ) = a . s . m ( x ) .
This completes the proof. □
Remark 1. 
Theorem 1 shows the strong consistency of kernel estimator m ˜ h ( N ) ( x ) . It is different from the work of Meister [6] in that the density function of our covariate error δ contains zeros in the Fourier domain. Our covariate error belongs to the Fourier oscillating noise considered by Delaigle and Meister [1]. Compared to their work, we construct a regression estimator via the Laplace transform without assuming f X and m to be compact.

5. Convergence Rate

In this section, we focus on the convergence rate in the weak sense. Meister [6] introduces the weak convergence rate by modifying the concept of weak consistency. A regression estimator m ^ n ( x ) is said to attain the weak convergence rate ε n if
lim C lim sup n sup ( m , f X ) P P | m ^ n ( x ) m ( x ) | 2 C · ε n = 0 .
The set P is the collection of all pairs ( m , f X ) that satisfy some conditions. The order of limits is first n , and then C . Here, C is independent of n.
Define the set
P β , r ; x : = { ( m , f X ) : f X , m f X F σ , β , r ; x ( A , B ) , | m ( x ) | C 1 , f X ( x ) C 2 , m ( · ) C 3 } ,
where C 1 , C 2 , C 3 > 0 .
The following Lemma is used to prove the theorem in this section.
Lemma 2 
([6]). If p : = m f X , m ^ n ( x ) = p n ( x ) f X , n ( x ) , | m ( x ) | < + and f X ( x ) 0 . Then, for a small enough ϵ > 0 ,
P | m ^ n ( x ) m ( x ) | 2 > ϵ P | p n ( x ) p ( x ) | 2 > c 1 ϵ + P | f X , n ( x ) f X ( x ) | 2 > c 2 ϵ
with two positive constants, c 1 and c 2 .
Theorem 2. 
Consider the model (1) with (2). Assume that ( m , f X ) P β , r ; x with r = 2 γ 2 if γ > 1 , and r > 0 if γ = 1 . Suppose kernel K satisfies conditions (i), (ii) with k 0 β . Let h = n 1 2 β + 2 γ + 1 , N n β + 1 r ( 2 β + 2 γ + 1 ) . Then,
lim C lim sup n sup ( m , f X ) P β , r ; x P | m ˜ h ( N ) ( x ) m ( x ) | 2 C · ε n = 0 ,
where ε n = n 2 β 2 β + 2 γ + 1 .
Proof. 
(1) We assume that x 0 and consider the estimator m ˜ + , h ( N ) ( x ) . Applying Lemma 2 and Markov’s inequality, we obtain
P | m ˜ + , h ( N ) ( x ) m ( x ) | 2 C · ε n c 3 C ε n E | p ˜ + , h ( N ) ( x ) p ( x ) | 2 + E | f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 ,
where c 3 is the larger of 1 c 1 and 1 c 2 , and c 1 , c 2 appear in Lemma 2. Then,
E | p ˜ + , h ( N ) ( x ) p ( x ) | 2 = var [ p ˜ + , h ( N ) ( x ) ] + | E p ˜ + , h ( N ) ( x ) p ( x ) | 2 ,
and
E | f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 = var [ f ˜ X , + , h ( N ) ( x ) ] + | E f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 .
First, we estimate | E p ˜ + , h ( N ) ( x ) p ( x ) | 2 and | E f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 . By (18), we have
| E p ˜ + , h ( N ) ( x ) p ( x ) | 2 2 | 1 1 K ( y ) [ p ( x + y h ) p ( x ) ] d y | 2 + | T + , N γ ( p ; x ) | 2 .
By Taylor expansion of p with the degree β 1 , there exists 0 < η < 1 such that
1 1 K ( y ) p ( x + y h ) p ( x ) d y = 1 1 K ( y ) j = 1 β ( y h ) j j ! p ( j ) ( x ) + ( y h ) β β ! p ( β ) ( x + η y h ) p ( β ) ( x ) d y | 1 1 K ( y ) j = 1 β ( y h ) j j ! p ( j ) ( x ) d y | + | 1 1 K ( y ) ( y h ) β β ! p ( β ) ( x + η y h ) p ( β ) ( x ) d y | .
Since kernel K satisfies condition (ii) and β k 0 , we have
1 1 K ( y ) j = 1 β ( y h ) j j ! p ( j ) ( x ) d y = 0 .
By p H σ , β ; x ( A ) , we find that
1 1 K ( y ) ( y h ) β β ! p ( β ) ( x + η y h ) p ( β ) ( x ) d y 1 1 | K ( y ) | | y h | β β ! p ( β ) ( x + η y h ) p ( β ) ( x ) d y 1 1 | K ( y ) | A | y h | β η | β β β ! d y c 4 A h β
holds for an h that is small enough. Equations (19) and (28) imply the following upper bound:
| E p ˜ + , h ( N ) ( x ) p ( x ) | 2 c 5 A 2 h 2 β + B 2 h 2 N 2 r .
Now, we estimate the term | E f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 . By (8) and (5),
E f ˜ X , + , h ( N ) ( x ) = E L + , h ( N ) ( W 1 x ) = ( 2 θ ) γ 2 π + K ^ ( i v h ) ( i v ) γ E ( e i v X 1 ) E ( e i v δ 1 ) e i v ( x + γ θ ) l 1 = 0 N γ l γ = 0 N γ e i 2 θ v ( l 1 + + l γ ) d v .
Note that E ( e i v X 1 ) = + f X ( t ) e i v t d t . Then, similar arguments to (15)–(17) show
E f ˜ X , + , h ( N ) ( x ) = + 1 h K ( t x h ) f X ( t ) d t + T + , N γ ( f X ; x ) .
Similar to (27)–(29), we have
| E f ˜ X , + , h ( N ) ( x ) f X ( x ) | 2 c 6 A 2 h 2 β + B 2 h 2 N 2 r .
Now, we estimate var [ p ˜ + , h ( N ) ( x ) ] and var [ f ˜ X , + , h ( N ) ( x ) ] . By (8), we have
var [ p ˜ + , h ( N ) ( x ) ] 1 n E | Y 1 | 2 | L + , h ( N ) ( W 1 x ) | 2 = 1 n E E | Y 1 | 2 | L + , h ( N ) ( W 1 x ) | 2 | X 1 = 1 n + E ( | Y 1 | 2 | X 1 = t ) E | L + , h ( N ) ( t + δ 1 x ) | 2 f X ( t ) d t .
Note that var ( Y 1 | X 1 = t ) = E ( | Y 1 | 2 | X 1 = t ) m 2 ( t ) . It follows from var ( Y 1 | X 1 = · ) = E ϵ j 2 and m ( · ) C 3 that E ( | Y 1 | 2 | X 1 = · ) var ( Y 1 | X 1 = · ) + m 2 ( · ) c 7 . Then,
var [ p ˜ + , h ( N ) ( x ) ] c 7 n + E | L + , h ( N ) ( t + δ 1 x ) | 2 f X ( t ) d t = c 7 n E E ( | L + , h ( N ) ( W 1 x ) | 2 | X 1 ) = c 7 n + | L + , h ( N ) ( ω x ) | 2 f W ( ω ) d ω .
It follows from (5) that
L + , h ( N ) ( t ) = ( 2 θ ) γ h γ + 1 l 1 = 0 N γ l γ = 0 N γ K ( γ ) t γ θ 2 θ ( l 1 + + l γ ) h .
Therefore,
var [ p ˜ + , h ( N ) ( x ) ] c 7 n + ( 2 θ ) γ h γ + 1 l 1 = 0 N γ l γ = 0 N γ K ( γ ) ω x γ θ 2 θ ( l 1 + + l γ ) h 2 × f W ( ω ) d ω c 7 ( 2 θ ) 2 γ n h 2 γ + 2 + l 1 = 0 N γ l γ = 0 N γ K ( γ ) ω x γ θ 2 θ ( l 1 + + l γ ) h 2 × f W ( ω ) d ω .
Let
C l , γ : = l + γ 1 γ 1 ,
where C l , γ is the number of weak compositions of l in γ parts [21]. Note that
l 1 = 0 N γ l γ = 0 N γ K ( γ ) ω x γ θ 2 θ ( l 1 + + l γ ) h l = 0 N C l , γ K ( γ ) ω x γ θ 2 θ l ) h .
Then,
var [ p ˜ + , h ( N ) ( x ) ] c 7 ( 2 θ ) 2 γ n h 2 γ + 2 l = 0 N j = 0 N C l , γ C j , γ + | K ( γ ) ω x θ ( γ + 2 l ) h | × | K ( γ ) ω x θ ( γ + 2 j ) h | f W ( ω ) d ω .
By supp ( K ) [ 1 , 1 ] , we have supp K ( γ ) ω x θ ( γ + 2 l ) h [ x + θ ( γ + 2 l ) h , x + θ ( γ + 2 l ) + h ] . Denote I + , l ( x ) : = [ x + θ ( γ + 2 l ) h , x + θ ( γ + 2 l ) + h ] . If h < θ , the intervals I + , l ( x ) and I + , j ( x ) are disjointed for l j . For an h that is small enough, we obtain
var [ p ˜ + , h ( N ) ( x ) ] c 7 ( 2 θ ) 2 γ n h 2 γ + 2 l = 0 N C l , γ 2 + K ( γ ) ω x θ ( γ + 2 l ) h 2 f W ( ω ) d ω c 8 ( 2 θ ) 2 γ n h 2 γ + 1 l = 0 N C l , γ 2 h I + , l ( x ) f W ( ω ) d ω .
Denote ξ + , l : = x + θ ( γ + 2 l ) . By supp ( f δ ) [ γ θ , γ θ ] and f δ c 9 θ ,
1 h I + , l ( x ) f W ( ω ) d ω c 9 θ h I + , l ( x ) γ θ γ θ f X ( ω t ) d t d ω = c 9 θ h + f X ( u ) + χ ( u γ θ , u + γ θ ) ( ω ) χ ( ξ + , l h , ξ + , l + h ) ( ω ) d ω d u .
Since h < γ θ , we have
1 h I + , l ( x ) f W ( ω ) d ω c 9 θ [ h h ( 1 + t h ) f X ( t + ξ + , l γ θ ) d t + 2 h γ θ h + γ θ f X ( t + ξ + , l ) d t + h h ( 1 t h ) f X ( t + ξ + , l + γ θ ) d t ] c 10 θ [ h h f X ( t + x + 2 θ l ) d t + γ θ γ θ f X ( t + x + θ ( γ + 2 l ) ) d t + h h f X ( t + x + 2 θ ( γ + l ) ) d t ] .
This, with (37), leads to
var [ p ˜ + , h ( N ) ( x ) ] c 8 c 10 ( 2 θ ) 2 γ n h 2 γ + 1 l = 0 N C l , γ 2 θ [ h h f X ( t + x + 2 θ l ) d t + γ θ γ θ f X ( t + x + θ ( γ + 2 l ) ) d t + h h f X ( t + x + 2 θ ( γ + l ) ) d t ] .
When γ > 1 , we obtain
l = 0 N C l , γ 2 θ h h f X ( t + x + 2 θ l ) d t = 1 θ x h x + h f X ( t ) d t + l = 1 N C l , γ 2 θ x + 2 θ l h x + 2 θ l + h f X ( t ) d t 1 θ + c 11 B θ 2 γ + 1
by f X M 2 γ 2 ( B ) and similar arguments to [16]. Similarly,
l = 0 N C l , γ 2 θ γ θ γ θ f X ( t + x + θ ( γ + 2 l ) ) d t 1 θ + c 12 B θ 2 γ + 1 ,
and
l = 0 N C l , γ 2 θ h h f X ( t + x + 2 θ ( γ + l ) ) d t 1 θ + c 13 B θ 2 γ + 1 .
When γ = 1 , we have that
l = 0 N 1 θ h h f X ( t + x + 2 θ l ) d t = l = 0 N 1 θ x + 2 θ l h x + 2 θ l + h f X ( t ) d t 1 θ x h x + 2 θ N + h f X ( t ) d t 1 θ
holds for h < θ . Similar to (41), for h < θ , we have
l = 0 N 1 θ θ θ f X ( t + x + θ ( 2 l + 1 ) ) d t 1 θ and l = 0 N 1 θ h h f X ( t + x + 2 θ ( l + 1 ) ) d t 1 θ .
Hence,
var [ p ˜ + , h ( N ) ( x ) ] c 14 ( 2 θ ) 2 γ n h 2 γ + 1 B θ 2 γ + 1 + 3 θ c 15 ( B θ + θ 2 γ 1 ) ( n h 2 γ + 1 ) 1 .
Similar to estimate var [ p ˜ + , h ( N ) ( x ) ] , we have
var [ f ˜ X , + , h ( N ) ( x ) ] 1 n E | L + , h ( N ) ( W 1 x ) | 2 c 16 ( B θ + θ 2 γ 1 ) ( n h 2 γ + 1 ) 1 .
By (29), (32), (42) and (43) with (24)–(26), we obtain
P | m ˜ + , h ( N ) ( x ) m ( x ) | 2 C · ε n c 17 C ε n A 2 h 2 β + B 2 h 2 N 2 r + ( B θ + θ 2 γ 1 ) ( n h 2 γ + 1 ) 1 .
Since h = n 1 2 β + 2 γ + 1 and N n β + 1 r ( 2 β + 2 γ + 1 ) ,
P | m ˜ + , h ( N ) ( x ) m ( x ) | 2 C · ε n c 17 ( A 2 + B 2 + B θ + θ 2 γ 1 ) C ε n · n 2 β 2 β + 2 γ + 1 .
Note that ε n = n 2 β 2 β + 2 γ + 1 . Then,
sup ( m , f X ) P β , r ; x P | m ˜ + , h ( N ) ( x ) m ( x ) | 2 C · ε n c 17 ( A 2 + B 2 + B θ + θ 2 γ 1 ) C 1 .
This leads to the result of Theorem 2 for x 0 .
( 2 ) We consider the estimator m ˜ , h ( N ) ( x ) for x < 0 . By (22), (23) and (28), we have
| E p ˜ , h ( N ) ( x ) p ( x ) | 2 2 | 1 1 K ( y ) [ p ( x + y h ) p ( x ) ] d y | 2 + | T , N γ ( p ; x ) | 2 c 18 A 2 h 2 β + B 2 h 2 N 2 r .
Similar arguments to (30)–(32) show
| E f ˜ X , , h ( N ) ( x ) f X ( x ) | 2 c 19 A 2 h 2 β + B 2 h 2 N 2 r .
Similar to (33),
var [ p ˜ , h ( N ) ( x ) ] c 20 n + | L , h ( N ) ( ω x ) | 2 f W ( ω ) d ω ,
and from (6),
L , h ( N ) ( t ) = ( 2 θ ) γ h γ + 1 l 1 = 0 N γ l γ = 0 N γ K ( γ ) t + γ θ + 2 θ ( l 1 + + l γ ) h .
Similar arguments to (34)–(37) show
var [ p ˜ , h ( N ) ( x ) ] c 20 ( 2 θ ) 2 γ n h 2 γ + 2 + l 1 = 0 N γ l γ = 0 N γ K ( γ ) ω x + γ θ + 2 θ ( l 1 + + l γ ) h 2 × f W ( ω ) d ω c 21 ( 2 θ ) 2 γ n h 2 γ + 1 l = 0 N C l , γ 2 h I , l ( x ) f W ( ω ) d ω
holds for an h that is small enough, where I , l ( x ) : = [ x θ ( γ + 2 l ) h , x θ ( γ + 2 l ) + h ] . Denote ξ , l : = x θ ( γ + 2 l ) . Similar to (38),
1 h I , l ( x ) f W ( ω ) d ω c 22 θ [ h h ( 1 + t h ) f X ( t + ξ , l γ θ ) d t + 2 h γ θ h + γ θ f X ( t + ξ , l ) d t + h h ( 1 t h ) f X ( t + ξ , l + γ θ ) d t ] c 23 θ [ h h f X ( t + x 2 θ ( γ + l ) ) d t + γ θ γ θ f X ( t + x θ ( γ + 2 l ) ) d t + h h f X ( t + x 2 θ l ) ) d t ] .
By similar arguments to (39)–(42), we have
var [ p ˜ , h ( N ) ( x ) ] c 24 ( B θ + θ 2 γ 1 ) ( n h 2 γ + 1 ) 1 ,
and
var [ f ˜ X , , h ( N ) ( x ) ] 1 n E | L , h ( N ) ( W 1 x ) | 2 c 25 ( B θ + θ 2 γ 1 ) ( n h 2 γ + 1 ) 1 .
Similar to (45),
sup ( m , f X ) P β , r ; x P | m ˜ , h ( N ) ( x ) m ( x ) | 2 C · ε n c 26 ( A 2 + B 2 + B θ + θ 2 γ 1 ) C 1 .
This leads to the result of Theorem 2 for x < 0 .
This completes the proof. □
Remark 2. 
Our convergence rate is the same as that in the ordinary smoothness case of Meister [6], where the density function of the covariate error does not vanish in the Fourier domain. Compared to Delaigle and Meister [1], we do not assume f X and m to be compact.
Remark 3. 
Belomestny and Goldenshluger [16] consider the density deconvolution problem with non-standard error distributions. They assume the density function to be estimated satisfies the Hölder condition. It is natural to assume a local smooth condition in point estimation. Hence, f X and m f X are assumed to satisfy the local Hölder condition in our discussion.
Remark 4. 
Theorem 1 shows the strong consistency of the regression estimator without the smoothness assumption. The main tool used is the Borel–Cantelli lemma which requires a convergent series. It is easy to see from (13) and (20) that the choice of h is not unique. Theorem 2 gives a weak convergence rate, which is defined by modifying the weak consistency. It is natural to assume the smoothness condition when discussing the convergence rate. In Theorem 2, the choice of h is related to the smoothness index β. It follows from our proof (44) that the choice of h is unique in the sense of a constant difference.
Remark 5. 
In our discussion, f ^ δ ( i v ) = sinh ( i θ v ) i θ v γ = sin ( θ v ) θ v γ . Substituting this into the proof of Theorem 3.5 in [6], one can obtain the optimality of convergence rate in our Theorem 2. This means that there does not exist an estimator m ˜ ( x ) of the regression function m ( x ) based on i.i.d data ( W 1 , Y 1 ) , , ( W n , Y n ) generated by model (1) with (2), which satisfies
lim C lim sup n sup ( m , f X ) P β , r ; x P | m ˜ ( x ) m ( x ) | 2 C · n 2 β 2 β + 2 γ + 1 = 0 .
It would be interesting to study the numerical illustration of our estimation. We shall investigate this in the future.

Author Contributions

Writing—original draft preparation, H.G. and Q.B.; Writing—review and editing, H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the National Natural Science Foundation of China (No. 12001132), the Guangxi Colleges and Universities Key Laboratory of Data Analysis and Computation, and the Center for Applied Mathematics of Guangxi (GUET).

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editor and reviewers for their important comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Delaigle, A.; Meister, A. Nonparametric function estimation under Fourier-oscillating noise. Stat. Sin. 2011, 21, 1065–1092. [Google Scholar] [CrossRef]
  2. Dong, H.; Otsu, T.; Taylor, L. Bandwidth selection for nonparametric regression with errors-in-variables. Econom. Rev. 2023, 42, 393–419. [Google Scholar] [CrossRef]
  3. Di Marzio, M.; Fensore, S.; Taylor, C.C. Kernel regression for errors-in-variables problems in the circular domain. Stat. Methods Appl. 2023. [Google Scholar] [CrossRef]
  4. Fan, J.Q.; Truong, Y.K. Nonparametric regression with errors in variables. Ann. Stat. 1993, 21, 1900–1925. [Google Scholar] [CrossRef]
  5. Hu, Z.R.; Ke, Z.T.; Liu, J.S. Measurement error models: From nonparametric methods to deep neural networks. Stat. Sci. 2022, 37, 473–493. [Google Scholar] [CrossRef]
  6. Meister, A. Deconvolution Problems in Nonparametric Statistics; Springer: Berlin, Germany, 2009. [Google Scholar]
  7. Song, W.X.; Ayub, K.; Shi, J.H. Extrapolation estimation for nonparametric regression with measurement error. Scand. J. Stat. 2023. [Google Scholar] [CrossRef]
  8. Carroll, R.J.; Delaigle, A.; Hall, P. Non-parametric regression estimation from data contaminated by a mixture of Berkson and classical errors. J. R. Stat. Soc. Ser. B Stat. Methodol. 2007, 69, 859–878. [Google Scholar] [CrossRef] [PubMed]
  9. Zhou, S.; Pati, D.; Wang, T.Y.; Yang, Y.; Carroll, R.J. Gaussian processes with errors in variables: Theory and computation. J. Mach. Learn. Res. 2023, 24, 1–53. [Google Scholar]
  10. Delaigle, A.; Hall, P.; Jamshidi, F. Confidence bands in non-parametric errors-in-variables regression. J. R. Stat. Soc. Ser. B Stat. Methodol. 2015, 77, 149–169. [Google Scholar] [CrossRef]
  11. Nadaraya, E.A. On estimating regression. Theory Probab. Its Appl. 1964, 9, 141–142. [Google Scholar] [CrossRef]
  12. Watson, G.S. Smooth regression analysis. Sankhyā Indian J. Stat. 1964, 26, 359–372. [Google Scholar]
  13. Guo, H.J.; Liu, Y.M. Strong consistency of wavelet estimators for errors-in-variables regression model. Ann. Inst. Stat. Math. 2017, 69, 121–144. [Google Scholar] [CrossRef]
  14. Guo, H.J.; Liu, Y.M. Convergence rates of multivariate regression estimators with errors-in-variables. Numer. Funct. Anal. Optim. 2017, 38, 1564–1588. [Google Scholar] [CrossRef]
  15. Guo, H.J.; Liu, Y.M. Regression estimation under strong mixing data. Ann. Inst. Stat. Math. 2019, 71, 553–576. [Google Scholar] [CrossRef]
  16. Belomestny, D.; Goldenshluger, A. Density deconvolution under general assumptions on the distribution of measurement errors. Ann. Stat. 2021, 49, 615–649. [Google Scholar] [CrossRef]
  17. Goldenshluger, A.; Kim, T. Density deconvolution with non-standard error distributions: Rates of convergence and adaptive estimation. Electron. J. Stat. 2021, 15, 3394–3427. [Google Scholar] [CrossRef]
  18. Oppenheim, A.V.; Willsky, A.S.; Nawab, H.S. Signals & Systems, 2nd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 1996. [Google Scholar]
  19. Liu, Y.M.; Wu, C. Point-wise estimation for anisotropic densities. J. Multivar. Anal. 2019, 171, 112–125. [Google Scholar] [CrossRef]
  20. Stein, E.M.; Shakarchi, R. Real Analysis: Measure Theory, Integration, and Hilbert Spaces; Princeton University Press: Princeton, NJ, USA, 2005. [Google Scholar]
  21. Stanley, R.P. Enumerative Combinatorics; Cambridge University Press: Cambridge, UK, 1997; Volume 1. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, H.; Bai, Q. Regression Estimation with Errors in the Variables via the Laplace Transform. Axioms 2023, 12, 992. https://doi.org/10.3390/axioms12100992

AMA Style

Guo H, Bai Q. Regression Estimation with Errors in the Variables via the Laplace Transform. Axioms. 2023; 12(10):992. https://doi.org/10.3390/axioms12100992

Chicago/Turabian Style

Guo, Huijun, and Qingqun Bai. 2023. "Regression Estimation with Errors in the Variables via the Laplace Transform" Axioms 12, no. 10: 992. https://doi.org/10.3390/axioms12100992

APA Style

Guo, H., & Bai, Q. (2023). Regression Estimation with Errors in the Variables via the Laplace Transform. Axioms, 12(10), 992. https://doi.org/10.3390/axioms12100992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop