Next Article in Journal
Bases of G-V Intuitionistic Fuzzy Matroids
Previous Article in Journal
A Proposal of an Adaptive Neuro-Fuzzy Inference System for Modeling Experimental Data in Manufacturing Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Wavelet Estimations in the Convolution Structure Density Model

1
School of Mathematics and Information Science, Weifang University, Weifang 261061, China
2
School of Mathematics, Faculty of Science, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(9), 1391; https://doi.org/10.3390/math8091391
Submission received: 23 June 2020 / Revised: 31 July 2020 / Accepted: 17 August 2020 / Published: 19 August 2020

Abstract

:
Using kernel methods, Lepski and Willer study a convolution structure density model and establish adaptive and optimal L p risk estimations over an anisotropic Nikol’skii space (Lepski, O.; Willer, T. Oracle inequalities and adaptive estimation in the convolution structure density model. Ann. Stat. 2019, 47, 233–287). Motivated by their work, we consider the same problem over Besov balls by wavelets in this paper and first provide a linear wavelet estimate. Subsequently, a non-linear wavelet estimator is introduced for adaptivity, which attains nearly-optimal convergence rates in some cases.

1. Introduction

The estimation of a probability density from independent and identically distributed (i.i.d.) random observations X 1 , X 2 , , X n of X is a classical problem in statistics. The representative work is Donoho et al. [1], they established an adaptive and nearly-optimal estimate (up to a logarithmic factor) over Besov spaces using wavelets.
However, the observed data are always polluted by noises in many real-life applications. One of the important problems is the density estimation with an additive noise. Let Z 1 , Z 2 , , Z n be i.i.d. random variables and have the same distribution as
Z = X + Y ,
where X denotes a real-valued random variable with unknown probability density function f and Y stands for an independent random noise (error) with a known probability density g. The problem is to estimate f by Z 1 , Z 2 , , Z n in some sense. Moreover, it is also called a denconvolution problem (model), because the density h of Z equals the convolution of f and g. Fan and Koo [2] studied the MISE performance ( L 2 -risk) of linear wavelet deconvolution estimator over a Besov ball. The L risk optimal wavelet estimations were investigated by Lounici and Nickl [3]. Furthermore, Li and Liu [4] provided L p ( 1 p ) risk optimal deconvolution estimations using wavelet bases.
In this paper, we consider a generalized deconvolution model introduced by Lepski & Willer [5,6]. More precisely, let ( Ω , F , P ) be a probability space and Z 1 , Z 2 , , Z n be i.i.d. random variables having the same distribution as
Z = X + ε Y ,
where the symbols X and Y are same as model (1), f and g are the corresponding densities respectively. Moreover, the biggest difference with model (1) is that a Bernoulli random variable ε { 0 , 1 } with P { ε = 1 } = α is added in (2), and α [ 0 , 1 ] is known. The problem is also to estimate f by the observed data Z 1 , Z 2 , , Z n in some sense.
When α = 1 , model (2) reduces to the deconvolution one (see [2,3,4,7,8] et al.), while α = 0 corresponds to the classical density model with no errors (see [1,9,10,11] et al.). Clearly, the density function h of Z in (2) satisfies
h = ( 1 α ) f + α f g .
Here, f g stands for the convolution of f and g. Furthermore, when the function G α ( t ) : = 1 α + α g f t ( t ) 0 for t R , we have
f f t ( t ) = [ ( 1 α ) + α g f t ( t ) ] 1 h f t ( t ) = ( G α ( t ) ) 1 h f t ( t ) ,
where g, α are known and f f t is the Fourier transform of f L 1 ( R ) given by
f f t ( t ) : = R f ( x ) e i t x d x .
Based on the model (2) with some mild assumptions on G α , Lepski and Willer [5] provided a lower bound estimation over L p risk on an anistropic Nikol’skii space. Moreover, they investigated an adaptive and optimal L p estimate by using kernel method in Ref. [6]. Recently, Wu et al. [12] established a pointwise lower bound estimation for model (2) under the local Hölder condition.
When compared with the classical kernel estimation of density functions, the wavelet estimations provide more local information and fast algorithm [13]. We will consider the L p ( 1 p < ) risk estimations under the model (2) over Besov balls by using wavelets and expect to obtain the corresponding convergence rates.
The same as Assumption 4 in [6], we also need the following condition on Y,
| G α ( t ) | ( 1 + | t | 2 ) β ( α ) 2
with β ( α ) = β 0 for α = 1 and β ( α ) = 0 for others. It is reasonable, because it holds automatically for α = 0 , while the same condition for α = 1 is necessary for the deconvolution estimations [4,7]. In addition, when 0 < α < 1 2 , β ( α ) = 0 and | G α ( t ) | 1 α α | g f t ( t ) | 1 thanks to g f t 1 . In fact, the condition (3) is necessary to prove Lemmas 2 and 3 in Section 2. Here and after, A B denotes A c B for a fixed constant c > 0 ; A B means B A ; A B stands for both A B and A B .
It is well-known that the wavelet estimation depends on an orthonormal wavelet expansion in L 2 ( R ) , even in L p ( R ) . Let { V j : j N } be a classical Multiresolution Analysis of L 2 ( R ) with scaling function φ and ψ being the corresponding wavelet. Subsequently, for f L 2 ( R ) ,
f = k Z α j 0 k φ j 0 k + j = j 0 k Z β j k ψ j k ,
where α j k : = f , φ j k , β j k : = f , ψ j k and ϑ j k ( · ) : = 2 j 2 ϑ ( 2 j · k ) ( ϑ = φ o r ψ ) . A scaling function φ is called m-regular ( m N ) , if φ C m ( R ) and | φ ( k ) ( x ) | ( 1 + | x | 2 ) l for each l N ( k = 0 , 1 , , m ). Clearly, the m-regularity of φ implies that of the corresponding ψ , and | φ f t ( t ) | ( 1 + | t | 2 ) m 2 due to the integration by parts. An important example is Daubechies’ function D 2 N with N large enough.
As usual, let P j be the orthogonal projection from L 2 ( R ) onto V j ,
P j f : = k Z α j k φ j k .
If φ is m-regular, then P j f is well-defined for f L p ( R ) . Moreover, the identity (4) holds in L p ( R ) for 1 p .
The following lemma is needed for later discussions.
Lemma 1
([13]). Let ϑ be an orthogonal scaling function or a wavelet satisfying m-regularity. Subsequently, there exist C 2 C 1 > 0 , such that, for λ = { λ k } l p ( Z ) and 1 p ,
C 1 2 j ( 1 2 1 p ) λ l p k Z λ k ϑ j k p C 2 2 j ( 1 2 1 p ) λ l p .
One of the advantages of wavelet bases is that they can characterize Besov spaces, which contain the L 2 -Sobolev spaces and Hölder spaces as special examples.
Proposition 1
([13]). Let scaling function φ be m-regular with m > s > 0 and ψ be the corresponding wavelet. Afterwards, for r , q [ 1 , ] and f L r ( R ) , the following conditions are equivalent:
(i). 
f B r , q s ( R ) ;
(ii). 
{ 2 j s P j f f r } j 0 l q ; and,
(iii). 
α 0 · l r + ( 2 j ( s + 1 2 1 r ) β j · l r ) l q < .
The Besov norm can be defined by
f B r , q s : = α j 0 · l r + ( 2 j ( s + 1 2 1 r ) β j · l r ) l q .
When s > 0 and 1 r , p , q , it is well-known that
(1)
B r , q s B r , s B , s 1 r for s > 1 r ;
(2)
B r , q s B p , q s for r p and s 1 r = s 1 p ,
where A B stands for a Banach space A continuously embedded in another Banach space B. More precisely, u B c u A ( u A ) holds for some c > 0 .
In this paper, we use the notation B r , q s ( L , M ) with some constants L , M > 0 to stand for a Besov ball, i.e.,
B r , q s ( L , M ) : = { f L r ( R ) , f is a density , f B r , q s L and supp f [ M , M ] } .
Next, we will estimate f with L p risk by constructing wavelet estimators from the observed data Z 1 , Z 2 , , Z n . To introduce wavelet estimators, we take φ having compact support and m-regularity with m β ( α ) + 2 in this paper. Moreover, denote
α ^ j k : = 1 n l = 1 n ( K φ j k ) ( Z l ) ,
where
( K φ j k ) ( x ) : = 1 2 π R e i t x ( φ j k ) f t ( t ) ¯ G α 1 ( t ) d t
and K ψ j k is defined by the way. Clearly, E α ^ j k = α j k due to the Plancherel formula. Subsequently, the linear wavelet estimator is given by
f ^ n l i n : = k Λ j 0 α ^ j 0 k φ j 0 k ,
where Λ j : = { k Z , supp f supp φ j k } { k Z , supp f supp ψ j k } . In particular, the cardinality of Λ j satisfies that | Λ j | 2 j , when f and φ have compact supports.
Now, we are in a position to state the first result of this paper.
Theorem 1.
For r , p [ 1 , + ) , q [ 1 , + ] and 1 r < s < m , the estimator f ^ n l i n in (7) with 2 j 0 n 1 2 s + 2 β ( α ) + 1 satisfies
sup f B r , q s ( L , M ) E f ^ n l i n f p p n s p 2 s + 2 β ( α ) + 1 ,
where s = s ( 1 r 1 p ) + and a + = max { a , 0 } .
Remark 1.
When α = 1 , β ( α ) = β , the conclusion of Theorem 1 reduces to Theorem 3 of Li & Liu [4].
Note that the estimator f ^ n l i n is non-adaptive, because the choice of j 0 depends on the unknown parameter s. To obtain an adaptive estimate, define
β ^ j k : = 1 n l = 1 n ( K ψ j k ) ( Z l ) and β ˜ j k : = β ^ j k I { | β ^ j k | > τ j , n } .
Here, τ j , n = c γ 2 j β ( α ) j n and the constants c , γ will be determined later on. Subsequently, the non-linear wavelet estimator is defined by
f ^ n n o n : = k Λ j 0 α ^ j 0 k φ j 0 k + j = j 0 j 1 k Λ j β ˜ j k ψ j k ,
where j 0 , j 1 are positive integers satisfying 2 j 0 n 1 2 m + 2 β ( α ) + 1 and 2 j 1 n ln n respectively. Clearly, j 0 and j 1 do not depend on the unknown parameters s , r , q , which means that the estimator f ^ n n o n in (9) is adaptive.
Theorem 2.
Let r , p [ 1 , + ) , q [ 1 , + ] and 1 r < s < m . Then the estimator f ^ n n o n in (9) satisfies
sup f B r , q s ( L , M ) E f ^ n n o n f p p ( ln n ) p ln n n θ p ,
where θ : = min { s 2 s + 2 β ( α ) + 1 , s 1 r + 1 p 2 ( s 1 r ) + 2 β ( α ) + 1 } .
Remark 2.
When α = 0 , β ( α ) = 0 and θ = min { s 2 s + 1 , s 1 r + 1 p 2 ( s 1 r ) + 1 } , the convergence rate of Theorem 2 coincides with that of Theorem 3 in Donoho et al. [1]. On the other hand, β ( α ) = β and θ = min { s 2 s + 2 β + 1 , s 1 r + 1 p 2 ( s 1 r ) + 2 β + 1 } for the case α = 1 , while the conclusion of Theorem 4 in Li & Liu [4] can follow directly from this theorem.
Remark 3.
When comparing the result of Theorem 2 with Theorem 1, we find easily that for the case r p , the convergence rate of non-linear estimator is better than that of the linear one with n s p 2 s + 2 β ( α ) + 1 and s = s 1 r + 1 p .
Remark 4.
The convergence rates of Theorem 2 with the cases α = 0 and α = 1 are nearly-optimal (up to a logarithmic factor) by Donoho et al. [1] and Li & Liu [4] respectively. However, it is not clear whether the estimation in Theorem 2 is optimal (nearly-optimal) or not for α ( 0 , 1 ) . Therefore, one of our future work is to determine a low bound estimate for model (2) with α ( 0 , 1 ) . This problem may be much more complicated than the cases of α = 0 and α = 1 .

2. Preliminaries

This section is devoted to introduce some useful lemmas. The following inequality is necessary in the proof of Lemma 2.
Rosenthal’s inequality ([13]). Let p > 0 and X 1 , X 2 , , X n be the independent random variables such that E X l = 0 and E | X l | p < ( l = 1 , 2 , , n ) . Subsequently, there exists C ( p ) > 0 , such that
E l = 1 n X l p C ( p ) l = 1 n E | X l | p I { p > 2 } + l = 1 n E X l 2 p 2 .
Lemma 2.
Let 1 2 j n and f 1 . Then for p [ 1 , + ) ,
E | α ^ j k α j k | p n p 2 2 j β ( α ) p a n d E | β ^ j k β j k | p n p 2 2 j β ( α ) p .
Proof. 
Obviously, one only needs to prove the first inequality and the second one is similar. Define ξ l : = ( K φ j k ) ( Z l ) E ( K φ j k ) ( Z l ) ( l = 1 , , n ). Subsequently, { ξ l } l = 1 n are i.i.d. samples and E ξ l = 0 ( l = 1 , , n ). By the definitions of α j k and α ^ j k , E α ^ j k = α j k and
E | α ^ j k α j k | p = 1 n p E l = 1 n [ ( K φ j k ) ( Z l ) E ( K φ j k ) ( Z l ) ] p = 1 n p E l = 1 n ξ l p .
According to (6), one obtains that
| ( K φ j k ) ( Z l ) | R | ( φ j k ) f t ( t ) G α 1 ( t ) | d t = 2 j 2 R | φ f t ( 2 j t ) G α 1 ( t ) | d t .
This with (3), m ( m β ( α ) + 2 ) regularity of φ and 2 j 1 shows
| ( K φ j k ) ( Z l ) | 2 j 2 R ( 1 + | t | 2 ) m 2 ( 1 + | 2 j t | 2 ) β ( α ) 2 d t 2 j ( 1 2 + β ( α ) ) .
Hence, for l = 1 , 2 , , n ,
ξ l 2 j ( 1 2 + β ( α ) ) .
On the other hand, h = ( 1 α ) f + α ( f g ) 1 follows from f 1 . Afterwards,
E | ξ l | 2 E ( K φ j k ) 2 ( Z l ) = R | K φ j k ( z ) | 2 h ( z ) d z R | K φ j k ( z ) | 2 d z .
Furthermore, E | ξ l | 2 R | ( φ j k ) f t ( z ) G α 1 ( z ) | 2 d z due to the Plancherel formula. The same arguments as (11) imply that
E | ξ l | 2 R ( 1 + | t | 2 ) m ( 1 + | 2 j t | 2 ) β ( α ) d t 2 2 j β ( α ) .
By Rosenthal’s inequality, E | l = 1 n ξ l | p l = 1 n E | ξ l | p I { p > 2 } + ( l = 1 n E ξ l 2 ) p 2 . This with (12) and (13) and 2 j n shows
E l = 1 n ξ l p n p 2 2 j β ( α ) p [ ( n 1 2 j ) p 2 1 I { p > 2 } + 1 ] n p 2 2 j β ( α ) p .
Finally, the desired conclusion is concluded from (10) and (14). □
We state another classical inequality, before giving the proof of Lemma 3.
Bernstein’s inequality ([13]).Let X 1 , X 2 , , X n be independent random variables with E X l = 0 , E X l 2 M and | X l | X ( l = 1 , 2 , , n ) . Then for each ϵ > 0 ,
P 1 n l = 1 n X l > ϵ 2 exp n ϵ 2 2 M + X ϵ / 3 .
Lemma 3.
If j 2 j n , 2 j 1 and f 1 , then there exists some constat c > 0 such that for any γ > 0 ,
P | β ^ j k β j k | > τ j , n 2 2 γ j ,
where τ j , n = c γ 2 j β ( α ) j n .
Proof. 
Denote η l : = ( K ψ j k ) ( Z l ) E ( K ψ j k ) ( Z l ) , l = 1 , , n . Then { η l } l = 1 n are i.i.d., E η l = 0 and | β ^ j k β j k | = | 1 n l = 1 n η l | by (8). Moreover, the same arguments as (12) and (13) show
η l H 1 2 j ( 1 2 + β ( α ) ) and E η l 2 H 2 2 2 j β ( α )
with some positive constants H 1 , H 2 .
According to Bernstein’s inequality,
P | β ^ j k β j k | > τ j , n 2 2 exp n τ j , n 2 8 [ H 2 2 2 j β ( α ) + H 1 2 j ( 1 2 + β ( α ) ) τ j , n / 6 ] .
Note that for j 2 j n ,
8 [ H 2 2 2 j β ( α ) + H 1 2 j ( 1 2 + β ( α ) ) τ j , n / 6 ] H 2 2 j β ( α ) ( 1 + c γ j 2 j n ) ( 1 + c γ ) H 2 2 j β ( α )
with some positive constant H. Hence,
n τ j , n 2 8 [ H 2 2 2 j β ( α ) + H 1 2 j ( 1 2 + β ( α ) ) τ j , n / 6 ] n c 2 γ 2 j n 1 2 2 j β ( α ) ( 1 + c γ ) H 2 2 j β ( α ) = c 2 γ 2 j ( 1 + c γ ) H .
Furthermore, (15) reduces to
P | β ^ j k β j k | > τ j , n 2 2 e c 2 γ 2 j ( 1 + c γ ) H 2 γ j
by choosing c max { 2 H ln 2 , 2 H γ 1 ln 2 } . This completes the proof. □

3. Proofs of Main Results

We shall show the proofs of Theorems 1 and 2 in this section.
Proof of Theorem 1.
It is sufficient to prove the case for r p . In fact, when r > p and f has a compact support, f ^ n l i n does because of φ having the same property. Subsequently, it follows from Hölder inequality and Jensen’s inequality that
sup f B r , q s ( L , M ) E f ^ n l i n f p p sup f B r , q s ( L , M ) ( E f ^ n l i n f r r ) p r n s p 2 s + 2 β ( α ) + 1 .
According to (5) and (7), one easily finds that
E f ^ n l i n f p p P j 0 f f p p + E f ^ n l i n E f ^ n l i n p p .
Clearly, by Proposition 1, P j 0 f f p p 2 j 0 s p thanks to the well-known embedding theorem B r , q s B p , q s for r p .
On the other hand, s > 1 r implies f 1 and Lemma 2 tells that E | α ^ j k α j k | p 2 j β ( α ) p n p 2 . This with Lemma 1 and | Λ j 0 | 2 j 0 shows that
E f ^ n l i n E f ^ n l i n p p 2 j 0 ( p 2 1 ) k Λ j 0 E | α ^ j 0 k α j 0 k | p 2 j 0 ( β ( α ) + 1 2 ) p n p 2 .
Finally, (16) reduces to
E f ^ n l i n f p p 2 j 0 s p + 2 j 0 ( β ( α ) + 1 2 ) p n p 2 n s p 2 s + 2 β ( α ) + 1
because of 2 j 0 n 1 2 s + 2 β ( α ) + 1 . The proof is done. □
Now, we give a proof of Theorem 2, which is the most important result.
Proof of Theorem 2.
The same arguments as the proof of Theorem 1, one only needs to prove the case for r p .
According to (4), (5) and (9), E f ^ n n o n f p p A n + B n + C n , where
A n = E k Λ j 0 ( α ^ j 0 k α j 0 k ) φ j 0 k p p , B n = E j = j 0 j 1 k Λ j ( β ˜ j k β j k ) ψ j k p p
and C n = f P j 1 f p p .
It is known that B r , q s B p , q s for r p with s = s 1 r + 1 p . Hence, C n = f P j 1 f p p 2 j 1 s p due to Proposition 1. Moreover, it follows from the choice of 2 j 1 n ln n , s > 1 r and θ : = min { s 2 s + 2 β ( α ) + 1 , s 1 r + 1 p 2 ( s 1 r ) + 2 β ( α ) + 1 } that
C n ln n n s p ln n n θ p .
Similar to (17), one obtains
A n = E k Λ j 0 ( α ^ j 0 k α j 0 k ) φ j 0 k p p 2 j 0 ( β ( α ) + 1 2 ) p n p 2 ln n n θ p
by 2 j 0 n 1 2 m + 2 β ( α ) + 1 , s < m and the definition of θ .
Next, the main work of this proof is to estimate B n . By Lemma 1,
B n = E j = j 0 j 1 k Λ j ( β ˜ j k β j k ) ψ j k p p j 1 p 1 j = j 0 j 1 2 j ( p 2 1 ) k Λ j E | β ˜ j k β j k | p .
Define B ^ j : = { k : | β ^ j k | > τ j , n } , B j : = { k : | β j k | > τ j , n 2 } , C j : = { k : | β j k | > 2 τ j , n } and ξ j k : = β ^ j k β j k . Subsequently, B n ( ln n ) p 1 l = 1 4 E e l , where
e 1 : = j = j 0 j 1 2 j ( p 2 1 ) k Λ j | ξ j k | p I { k B ^ j B j c } , e 2 : = j = j 0 j 1 2 j ( p 2 1 ) k Λ j | ξ j k | p I { k B ^ j B j } , e 3 : = j = j 0 j 1 2 j ( p 2 1 ) k Λ j | β j k | p I { k B ^ j c C j } , e 4 : = j = j 0 j 1 2 j ( p 2 1 ) k Λ j | β j k | p I { k B ^ j c C j c }
with A c denoting the complement of A in Z . Hence, it suffices to prove E e l ln n ( ln n n ) θ p ( l = 1 , 2 , 3 , 4 ) .
To estimate E e 1 , note that | β ^ j k β j k | τ j , n 2 follows from k B ^ j B j c . Moreover, E | ξ j k | p I { k B ^ j B j c } ( E | ξ j k | 2 p ) 1 2 P 1 2 ( | ξ j k | > τ j , n 2 ) by the Hölder inequality. Afterwards, for large γ ,
E e 1 j = j 0 j 1 2 j ( p 2 1 ) k Λ j n p 2 2 j β ( α ) p 2 γ j 2 2 j 0 ( p 2 + β ( α ) p γ 2 ) n p 2 ln n n θ p
thanks to | Λ j | 2 j , Lemmas 2 and 3.
For e 3 , | β ^ j k β j k | τ j , n 2 due to k B ^ j c C j . Combining this with Lemma 3, one knows E I { k B ^ j c C j } P ( | ξ j k | τ j , n 2 ) 2 γ j . Therefore, by | β j k | 2 j 2 , | Λ j | 2 j and large γ ,
E e 3 j = j 0 j 1 2 j ( p 2 1 ) k Λ j 2 p j 2 2 γ j 2 ( γ p ) j 0 ln n n θ p .
In order to estimate e 2 and e 4 , one defines j 0 , j 1 Z satisfying
2 j 0 n ln n 1 2 θ 2 β ( α ) + 1 and 2 j 1 n ln n θ s 1 r + 1 p .
Recall that θ : = min { s 2 s + 2 β ( α ) + 1 , s 1 r + 1 p 2 ( s 1 r ) + 2 β ( α ) + 1 } , 2 j 0 n 1 2 m + 2 β ( α ) + 1 and 2 j 1 n ln n . Then j 0 j 0 j 1 j 1 due to 1 r < s < m and r p .
To estimate E e 2 , one divides e 2 into
e 2 = ( j = j 0 j 0 + j = j 0 + 1 j 1 ) 2 j ( p 2 1 ) k Λ j | ξ j k | p I { k B ^ j B j } : = e 21 + e 22 .
Similar to (17), according to 2 j 0 ( n ln n ) 1 2 θ 2 β ( α ) + 1 ,
E e 21 2 j 0 ( β ( α ) + 1 2 ) p n p 2 ln n n θ p .
Define
ω : = s + β ( α ) + 1 2 r β ( α ) + 1 2 p .
When ω 0 , θ : = min { s 2 s + 2 β ( α ) + 1 , s 1 / r + 1 / p 2 ( s 1 / r ) + 2 β ( α ) + 1 } = s 2 s + 2 β ( α ) + 1 . It is easy to see that | β j k | τ j , n / 2 > 1 for k B j . Hence,
E e 22 j = j 0 + 1 j 1 2 j ( p 2 1 ) k Λ j E | ξ j k | p | β j k | τ j , n / 2 r .
In addition, E | ξ j k | p 2 j β ( α ) p n p 2 by Lemma 2 and k | β j k | r 2 j ( s r + r 2 1 ) due to Proposition 1 and f B r , q s . These with (18), τ j , n 2 j β ( α ) j n and 2 j 0 ( n ln n ) 1 2 θ 2 β ( α ) + 1 lead to
E e 22 j = j 0 + 1 j 1 ln n n p r 2 2 j ω ln n n p r 2 2 j 0 ω ln n n θ p .
For the case ω < 0 , θ : = min { s 2 s + 2 β ( α ) + 1 , s 1 / r + 1 / p 2 ( s 1 / r ) + 2 β ( α ) + 1 } = s 1 / r + 1 / p 2 ( s 1 / r ) + 2 β ( α ) + 1 . Denote r 1 : = ( 1 2 θ ) p . Then r r 1 p . Hence, the same arguments as (18) show that
E e 22 j = j 0 + 1 j 1 2 j ( p 2 1 ) k Λ j E | ξ j k | p | β j k | τ j , n / 2 r 1 .
On the other hand, k | β j k | r 1 ( k | β j k | r ) r 1 r 2 j r 1 ( s + 1 2 1 r ) follows from f B r , q s and Proposition 1. Furthermore, E | ξ j k | p 2 j β ( α ) p n p 2 thanks to Lemma 2. These with (19) and τ j , n 2 j β ( α ) j n imply that
E e 22 j = j 0 + 1 j 1 2 j [ p 2 1 + β ( α ) p β ( α ) r 1 r 1 ( s + 1 2 1 r ) ] ln n n p r 1 2 ln n ln n n θ p ,
because p r 1 2 = θ p and p 2 1 + β ( α ) p β ( α ) r 1 r 1 ( s + 1 2 1 r ) = 0 .
Finally, it remains to estimate E e 4 . Obviously, θ = s 2 s + 2 β ( α ) + 1 for ω 0 . Thus,
e 4 = ( j = j 0 j 0 + j = j 0 + 1 j 1 ) 2 j ( p 2 1 ) k Λ j | β j k | p I { k B ^ j c C j c } : = e 41 + e 42 .
Because k C j c and τ j , n 2 j β ( α ) j n , | β j k | < 2 τ j , n 2 j β ( α ) j n . This with | Λ j | 2 j implies that
E e 41 j = j 0 j 0 2 j ( p 2 1 ) k Λ j τ j , n p 2 j 0 p ( β ( α ) + 1 2 ) ln n n p 2 .
Note that 2 τ j , n | β j k | > 1 by k C j c , and k | β j k | r 2 j ( s r + r 2 1 ) due to f B r , q s . Subsequently, with r p and τ j , n 2 j β ( α ) j n ,
E e 42 j = j 0 + 1 j 1 2 j ( p 2 1 ) k | β j k | p 2 τ j , n | β j k | p r 2 j 0 ω ln n n p r 2 .
Combining (20)–(22) with 2 j 0 ( n ln n ) 1 2 θ 2 β ( α ) + 1 and θ = s 2 s + 2 β ( α ) + 1 ( ω 0 ) , one obtains that
E e 4 = E e 41 + E e 42 ln n n θ p
holds for ω 0 .
At last, one shows E e 4 ( ln n n ) θ p for the case ω < 0 . Subsequently,
e 4 = ( j = j 0 j 1 + j = j 1 + 1 j 1 ) 2 j ( p 2 1 ) k Λ j | β j k | p I { k B ^ j c C j c } : = e 41 + e 42 .
The same arguments as (22), one finds
E e 41 2 j 1 ω ln n n p r 2 .
On the other hand, k | β j k | p ( k | β j k | r ) p r 2 j p ( s + 1 2 1 r ) by f B r , q s ( r p ) and Proposition 1. Furthermore,
E e 42 j = j 1 + 1 j 1 2 j ( p 2 1 ) k Λ j | β j k | p 2 j 1 p ( s 1 r + 1 p )
because of s > 1 r . This with (23) and (24) and 2 j 1 ( n ln n ) θ s 1 r + 1 p leads to
E e 4 = E e 41 + E e 42 ln n n θ p
thanks to θ = s 1 / r + 1 / p 2 ( s 1 / r ) + 2 β ( α ) + 1 for ω < 0 . The proof is completed. □

4. Conclusions

This current paper shows L p ( 1 p < ) risk estimations of both linear and non-linear wavelet estimators under a convolution structure density model over Besov balls. The corresponding conclusions are introduced by Theorems 1 and 2 in this paper, which can be seen as an extension of the works of Donoho et al. [1] and Li & Liu [4].
It should be pointed out that the non-linear wavelet estimator is adaptive, and the convergence rate of non-linear estimator are better than that of linear one for the case r p . However, it is not clear whether the estimations are optimal (nearly-optimal) or not for α ( 0 , 1 ) . Therefore, one of our future work is to determine a low bound estimate for model (2) with α ( 0 , 1 ) .

Author Contributions

Conceptualization, K.C. and X.Z.; methodology, K.C. and X.Z.; validation, X.Z.; formal analysis, K.C.; writing–original draft preparation, K.C.; writing–review and editing, K.C. and X.Z.; supervision, X.Z.; funding acquisition, X.Z.; project administration, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the National Natural Science Foundation of China (No. 11901019) and the Science and Technology Program of Beijing Municipal Commission of Education (No. KM202010005025).

Acknowledgments

The authors would like to thank the referees for their very helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Donoho, D.L.; Johnstone, I.M.; Kerkyacharian, G.; Picard, D. Density estimation by wavelet thresholding. Ann. Stat. 1996, 24, 508–539. [Google Scholar] [CrossRef]
  2. Fan, J.; Koo, J.-Y. Wavelet deconvolution. IEEE Trans. Inform. Theory 2002, 48, 734–747. [Google Scholar]
  3. Lounici, K.; Nickl, R. Global uniform risk bounds for wavelet deconvolution estimators. Ann. Stat. 2011, 39, 201–231. [Google Scholar] [CrossRef] [Green Version]
  4. Li, R.; Liu, Y.M. Wavelet optimal estimations for a density with some additive noises. Appl. Comput. Harmon. Anal. 2014, 36, 416–433. [Google Scholar] [CrossRef]
  5. Lepski, O.; Willer, T. Lower bounds in the convolution structure density model. Bernoulli 2017, 23, 884–926. [Google Scholar] [CrossRef] [Green Version]
  6. Lepski, O.; Willer, T. Oracle inequalities and adaptive estimation in the convolution structure density model. Ann. Stat. 2019, 47, 233–287. [Google Scholar] [CrossRef] [Green Version]
  7. Liu, Y.M.; Zeng, X.C. Asymptotic normality for wavelet deconvolution density estimators. Appl. Comput. Harmon. Anal. 2020, 48, 321–342. [Google Scholar] [CrossRef]
  8. Pensky, M.; Vidakovic, B. Adaptive wavelet estimator for nonparametric density deconvolution. Ann. Stat. 1999, 27, 2033–2053. [Google Scholar]
  9. Cao, K.K.; Liu, Y.M. On the Reynaud-Bouret–Rivoiard–Tuleau-Malot problem. Int. J. Wavelets Multiresolut. Inf. Process. 2018, 16, 1850038. [Google Scholar] [CrossRef]
  10. Kerkyacharian, G.; Picard, D. Density estimation in Besov spaces. Stat. Probab. Lett. 1992, 13, 15–24. [Google Scholar] [CrossRef]
  11. Liu, Y.M.; Wang, H.Y. Convergence order of wavelet thresholding estimator for differential operators on Besov spaces. Appl. Comput. Harmon. Anal. 2012, 32, 342–356. [Google Scholar] [CrossRef] [Green Version]
  12. Wu, C.; Wang, J.R.; Zeng, X.C. A pointwise lower bound for generalized deconvolution density estimation. Appl. Anal. 2019. [Google Scholar] [CrossRef]
  13. Härdle, W.; Kerkyacharian, G.; Picard, D.; Tsybakov, A. Wavelets, Approximation and Statistical Applications; Springer: New York, NY, USA, 1998. [Google Scholar]

Share and Cite

MDPI and ACS Style

Cao, K.; Zeng, X. Adaptive Wavelet Estimations in the Convolution Structure Density Model. Mathematics 2020, 8, 1391. https://doi.org/10.3390/math8091391

AMA Style

Cao K, Zeng X. Adaptive Wavelet Estimations in the Convolution Structure Density Model. Mathematics. 2020; 8(9):1391. https://doi.org/10.3390/math8091391

Chicago/Turabian Style

Cao, Kaikai, and Xiaochen Zeng. 2020. "Adaptive Wavelet Estimations in the Convolution Structure Density Model" Mathematics 8, no. 9: 1391. https://doi.org/10.3390/math8091391

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop