Next Article in Journal
Common Environmental Effects on Quantum Thermal Transistor
Previous Article in Journal
The Listsize Capacity of the Gaussian Channel with Decoder Assistance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sparse Density Estimation with Measurement Errors

1
School of Mathematics and Statistics, Chaohu University, Hefei 238000, China
2
Department of Mathematics, Faculty of Science and Technology, University of Macau, Macau 999078, China
3
UMacau Zhuhai Research Institute, Zhuhai 519031, China
4
Department of Statistics, North Carolina State University, Raleigh, NC 27695, USA
5
Graduate School of Arts and Science, Yale University, New Haven, CT 06510-8034, USA
*
Author to whom correspondence should be addressed.
Xiaowei Yang, Huiming Zhang and Haoyu Wei are co-first authors which contributes equally to this work.
Entropy 2022, 24(1), 30; https://doi.org/10.3390/e24010030
Submission received: 16 November 2021 / Revised: 13 December 2021 / Accepted: 15 December 2021 / Published: 24 December 2021
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
This paper aims to estimate an unknown density of the data with measurement errors as a linear combination of functions from a dictionary. The main novelty is the proposal and investigation of the corrected sparse density estimator (CSDE). Inspired by the penalization approach, we propose the weighted Elastic-net penalized minimal 2 -distance method for sparse coefficients estimation, where the adaptive weights come from sharp concentration inequalities. The first-order conditions holding a high probability obtain the optimal weighted tuning parameters. Under local coherence or minimal eigenvalue assumptions, non-asymptotic oracle inequalities are derived. These theoretical results are transposed to obtain the support recovery with a high probability. Some numerical experiments for discrete and continuous distributions confirm the significant improvement obtained by our procedure when compared with other conventional approaches. Finally, the application is performed in a meteorology dataset. It shows that our method has potency and superiority in detecting multi-mode density shapes compared with other conventional approaches.

1. Introduction

Over the years, the mixture models have been extensively applied to model unknown distributional shapes in astronomy, biology, economics, and genomics (see [1] and references therein). The distributions of real data involving potential complex variables often show multi-mode and heterogeneity. Due to the flexibility, it also appears in various distribution-based statistical techniques, such as cluster analysis, discriminant analysis, survival analysis, and empirical Bayesian inference. Flexible mixture models can naturally represent how the data are generated as mathematical artifacts. Theoretical results show that the mixture can approximate any density in the Euclidean space well, and the amount of the mixture can also be finite (for example, a mixture of several Gaussian distributions). Although the mixture model is inherently attractive to the statistical modeling, it is well-known that it is difficult to infer (see [2,3]). From the computational aspect, the optimization problems of mixture models are non-convex. Although existing computational methods, such as EM and various MCMC algorithms, can make the mixture model fit the data relatively easily. It should be emphasized that the mixture problems are essentially challenging, even unrecognizable, and the number of components (says, the order selection) is hard to determine (see [4]). There is a large amount of literature on its approximation theory, and various methods have been proposed to estimate the components (see [5] and references therein).
The nonparametric method and combinatorial method in density estimation were well studied in [6,7], as well as [8]. These can consistently estimate the number of the mixture’s components when the components have known functional forms. When the number of candidate components is large, the non-parametric method becomes computationally infeasible. Fortunately, the high-dimensional inference would compensate for this gap and guarantee the corrected identification of the mixture components with a probability tending to one. With the advancement of technology, high-dimensional problems have been applied at the forefront of statistical research. The high-dimensional inference method has been applied to the infinite mixture models with a sparse mixture of p components, which is an interesting and challenging problem (see [9,10]). We propose an improvement of the sparse estimation strategy proposed in [9], in which Bunea et al. propose a 1 -type penalty [11] to obtain a sparse density estimate (SPADES). At the same time, we add a 2 -type penalty and extend the oracle-inequality results to our new estimator.
In the real data, we often encounter the situation that the i.i.d. samples X i = Z i + ε i are contained by some zero-mean measurement errors { ε i } i = 1 n ; see [12,13,14,15,16]. For density estimation of { Z i } i = 1 n , if there exists orthogonal basis functions, the estimation method is quite easy. In the measurement errors setting, however, finding an orthogonal-based density function is not easy (see [17]). Ref. [17] suggests the assumption that the conditional distribution function of X i given Z i is known. This condition is somewhat strong since most conditional distributions cannot obtain the explicit formula (except the Gaussian distribution). To address this predicament, particularly with nonorthogonal base functions, the SPADES model is attractive and makes the situation easier to deal with. Based on the SPADES method, our approach is an Elastic-net calibration approach, which is simpler and more interpreted than the conditional inference procedure proposed by [17]. In this paper, we proposed the corrected loss function to debiasing the measurement errors, and this is motivated by [18]. The main problem of considering measurement errors in various statistical models is that they are responsible for the bias of the classical statistical estimates; this is true, e.g., in linear regression, which has traditionally been the main focus of studies related to measurement errors. Debiasing represents an important task in various statistical models. In linear regression, it can be performed in the basic measurement errors (ME) model, which is also denoted as the Errors-in-variables (EIV or EV) model, if it is possible to estimate the variability of measurement errors (see [19,20]). We derive the real variable selection consistency based on weighted 1 + 2 penalty [21]. At the same time, some theoretical results of SPADES only contain the situation of the equal weights setting, which is not plausible in the sense of adaptive (data-dependent) penalized estimation. Moreover, we perform the Poisson mixture model to approximate the complex discrete distribution in the simulation part, while existing papers only emphasize the performance of continuous distribution models. Note that the multivariate kernel density estimator can only deal with a continuous distribution, and it requires a multivariate bandwidths section, while our method is dimensional-free (the number of the required tuning parameters is only two). There has been quite a lot of work in this area, starting with [22].
There are several differences between our article and [9]. The first point is that the upper bound of non-asymptotic oracle inequality in in our Theorems 1 and 2 is tighter than Theorems 1 and 2 in [9], and the optimal weighted tuning parameters are derived. The second point is that the 1 -penalized techniques are applied in [9] to estimate the sparse density. Still, this paper considers the estimation of density functions in the presence of a classical measurement error. We opt to use an Elastic-net criterion function to estimate the density, which is taken to be approximated by a series of basis functions. The third point is that the tuning parameters are chosen by the coordinate descent algorithm in [9], and the mixture weights are calculated by the generalized bisection method (GBM). However, this paper directly calculates the optimal weights, so our algorithm is more accessible to implement than [9].
This paper is presented as follows. Section 2 introduces the density estimator, which can deal with measurement errors. This section introduces data-dependent weights for the Lasso penalty, and the weights are derived by the event of KKT conditions holding a high probability. In Section 3, we give a condition that can accurately estimate the weights of the mixture, with a probability tending to 1. We show that, in an increasing dimensional mixture model under the local coherence assumption, if the tuning parameter is higher than the noise level, the recovery of the mixture components can hold with a high probability. In Section 4, we study the performance of our approach on artificial data generated from mixture Gaussian or Poisson distributions compared with other conventional methods, which indeed shows the improvement by employing our procedures. Moreover, the simulation also demonstrates that our method is better than the traditional EM algorithm, even under a low-dimensional model. Considering the multi-modal density aspect of the meteorology dataset, our proposed estimator has a stronger ability to detect multiple modes for the underlying distribution compared with other methods, such as SPADES or un-weighted Elastic-net estimator. Section 5 is the summary, and the proof of theoretical results is delivered in the Appendix A.

2. Density Estimation

2.1. Mixture Models

Suppose that { Z i } i = 1 n R d are independent random variables with a common unknown density h. However, the observations are contaminated with measurement errors { ε i } i = 1 n as latent variables, the observed data are actually X i = Z i + ε i . Let { h j } j = 1 W be a series of density functions (such as Gaussian density or Poisson mass function), and { h j } j = 1 W are also called basis functions. Assume that the h belongs to the linear combination of { h j } j = 1 W . The W : = W n is a function of n, which is particularly intriguing for us since there may be W n (the high-dimensional setting). Let β * : = ( β 1 * , , β W * ) R W be the unknown true parameter. Assume that
  • (H.1): the h : = h β * is defined as
    Z h ( z ) : = h β * ( z ) = j = 1 W β j * h j ( z ) , with j = 1 W β j * = 1 .
If the base is orthogonal and there are no measurement errors, a perfectly natural method is to estimate h by an orthogonal series of estimators in the form of h β ˜ , where β ˜ has the coordinates β ˜ j = 1 n i = 1 n h j ( X i ) (see [17]). However, this estimator depends on the choice of W, and a data-driven selection of W or the threshold needs to be adaptive. This estimator can only be applied to W n . Nevertheless, we want to solve more general problems for W > n , and the base functions { h j } j = 1 W may not orthogonal.
We aim to achieve the best convergence for the estimator when the W is not necessarily less than n. Theorem 33.2 in [5] states that any smooth density can be well-approximated by a finite mixture of some continuous functions. However, Theorem 33.2 in [5] does not confirm how many components W are required for the mixture. Thus, the hypothesis of the increasing-dimensional W is reasonable. For discrete distributions, there is also a similar mixture density approximation—see Remark of Theorem 33.2 in [5].

2.2. The Density Estimation with Measurement Errors

This subsection aims to construct a sparse estimator for the density h ( z ) : = h β * ( z ) as a linear combination of known densities.
Recall the definition of the L 2 ( R d ) norm f = R d f 2 ( x ) d x 1 2 . For f , g L 2 ( R d ) , let < f , g > = R d f ( x ) g ( x ) d x be the inner product. If two functions f and g satisfy < f , g > = 0 , then we call these two functions are orthogonal. Note that if the density h ( z ) belongs to L 2 ( R d ) and assume that { X i } i = 1 n has the same distribution X, for any f L 2 , we have < f , h > = R d f ( x ) h ( x ) d x = E f ( X ) . If h ( x ) is the density function for a discrete distribution, the integral is replaced by summation, and we can define the inner product as < f , h > : = k Z d f ( k ) h ( k ) .
For true observations { Z i } i = 1 n , we minimize the h β h 2 on β R W to obtain the estimate of h ( z ) : = h β * ( z ) , i.e., minimizing
h β h 2 = h 2 + h β 2 2 < h β , h > = h 2 + h β 2 2 E h β ( Z ) 2 E h β ( Z ) + h β 2 ,
which implies that minimizing the h β h 2 is equivalent to minimizing
2 E h β ( Z ) + h β 2 2 n i = 1 n h β ( Z i ) + h β 2 .
It is plausible to assign more constrains for the candidate set of β in the optimization, for example, the 1 constrains β 1 a , where a is the tuning parameter. More adaptively, we prefer to use the weighted 1 restriction j = 1 W ω j | β j | a , where the weights ω j ’s are data-dependent and will be specified later. From [23], we add Elastic-net penality 2 j = 1 W ω j | β j | + c j = 1 W β j 2 with tuning parameter c, which is in regards to the measurement errors (see [24,25]) for a similar purpose. We would have c = 0 in the situation without measurement errors. The c indeed becomes larger if the measurement errors become more serious, i.e., we can say that the c is proportional to the increasing variability of the measurement errors. It is different from SPADES since adjusting for the measurement errors is important for accurately describing the relationship between the observed varables and the outcomes of interest.
From the discussion above, now we propose the following Corrected Sparse Density Estimator (CSDE):
β ^ : = β ^ ( ω 1 , , ω W ) = arg min β R W 2 n i = 1 n h β ( X i ) + h β 2 + 2 j = 1 W ω j | β j | + c j = 1 W β j 2
where the c is the tuning parameter for 2 -penalty, and the c also presents the correction for adjusting the measurement errors in our observations.
For CSDE, if { h j } j = 1 W is an orthogonal system, it can be clearly seen that the CSDE estimator is consistent with the soft thresholding estimator, and the explicit solution is β ^ j = ( 1 ω j / | β ˜ j | ) + β ˜ j 1 + c , where β ˜ j = 1 n i = 1 n h j ( X i ) and x + = max ( 0 , x ) . In this case, we can see that ω j is the threshold of the j-th component of the simplest mean estimator β ˜ = ( β ˜ 1 , , β ˜ W ) .
From the sub-differential of the convex optimization, the corresponding Karush–Kuhn–Tucker conditions (necessary and sufficient first-order condition) for the minimizer in Equation (3) is
Lemma 1
(KKT conditions, Lemma 4.2 of [26]). Let k { 1 , 2 , , W } and c > 0 . Then, a necessary and sufficient condition for CSDE to be a solution of Equation (3)
  • β ^ k : 0 if 1 n i = 1 n h k ( X i ) j = 1 W β ^ j < h j , h k > c β ^ k = w k sign ( β ^ k ) .
  • β ^ k = 0 if 1 n i = 1 n h k ( X i ) j = 1 W β ^ j < h j , h k > c β ^ k w k .
Since all values of β j * are non-negative, when conducting minimization in Equation (3), we have to put a non-negative restriction for optimizing Equation (3).
Due to the computational feasibility and optimal first-order conditions, we prefer an adaptively weighted Lasso penalty as a convex adaptive 1 penalization. We require that the larger weights are assigned to the coefficients of unimportant covariates, while significant covariates accompany the smaller weights. So, the weights represent the importance of the covariates. The larger (smaller) weights shrink to zero more easily (difficultly) than the unweighted Lasso, with appropriate or even optimal weights, leading to less bias and more efficient variable selection. The derivation of the weights will be given in Section 3.1.
In the end of this part, we will illustrate that in the mixture models, even without measurement errors, Equation (1) cannot be partially transformed into the linear model Y = X T β + ε , where Y is the n-dimensional response variables, X is the W × n -dimensional fixed design matrix, β is a W-dimensional vector of model parameters, the ε is a n × 1 -dimensional vector for random error terms with zero mean and finite variance. Consider the least square objective function U ( β ) for estimating β , U ( β ) = ( Y X T β ) T ( Y X T β ) = 2 Y T X T β + β T X X T β + Y T Y . Minimizing U ( β ) is equivalent to minimizing U * ( β ) in Formula (4)
U * ( β ) = 2 Y T X T β + β T X X T β .
Comparing the objective function in Formula (4) with Equation (2), it is easy to obtain Y = ( 1 n , 1 n , , 1 n ) T , β = ( β 1 , β 2 , , β W ) T , X = h 1 ( X 1 ) h 1 ( X n ) h W ( X 1 ) h W ( X n ) . Substituting Y, X and β into a linear regression model, we obtain
1 n 1 n n × 1 = h 1 ( X 1 ) h W ( X 1 ) h 1 ( X n ) h W ( X n ) n × W β 1 β W W × 1 + ε 1 ε n n × 1 .
Then,
ε i = 1 n j = 1 W β j h j ( X i ) , i = 1 , 2 , , n .
It can be seen from Equation (5) that the value of ε i is no longer random if X was the fixed design matrix. Furthermore, even for a random design X , take the expectation on both sides of Equation (5), and one can find that the left side is not equal to the right side, that is, E ( ε i ) = 0 = 1 n j = 1 W β j E h j ( X i ) . It leads to an additional requirement j = 1 W β j E h j ( X i ) = 1 n 0 , which is meaningless as n , since all β j and h j are positive. This is a contradiction to j = 1 W β j E h j ( X i ) > 0 for all n.
Both of the two situations above contradict the definition of the assumed linear regression model. Hence, we cannot convert the estimation of Equation (1) into the estimation problem of linear models. Thus, the existing oracle inequalities are not applicable anymore, and we will propose new ones later. However, we can transform the mixture models to a corrected score Dantzig selector, such as in [27]. Although [10] studies the oracle inequalities for adaptive Dantzig density estimation, their study does not contain the error-in-variables framework and the support recovery content.

3. Sparse Mixture Density Estimation

In this section, we will present the oracle inequalities for estimators β ^ and h β ^ . The core of this section consists of five main results corresponding to the oracle inequalities for estimated density (Theorems 1 and 2), upper bounds on 1 -estimation error (Corollaries 1 and 2) and support consistency (Theorem 3) as the byproduct of Corollary 2.

3.1. Data-Dependent Weights

The weights ω j ’s are chosen adequately such that the KKT conditions for stochastic optimization problems have a high probability of being satisfied.
As mentioned before, the weights in Equation (3) rely on the observed data since we calculate the weights, ensuring the KKT conditions hold with a high probability. The weighted Lasso estimates could have less 1 estimation error than Lasso estimates (see the simulation part and [28]). Next, we need to consider what kind of data-dependent weight configuration can enable the KKT conditions to be satisfied with a high probability. A way to obtain data-dependent weights is to apply a concentration inequality for a weighted sum of independent random variables. Moreover, the weights should be a known data function without any unknown parameters. A criterion can help obtain the weights grounded on McDiarmid’s inequality (see [29] for more details).
Lemma 2.
Suppose X 1 , , X n are independent random variables, and all values belong to a set A. Let f : A n R be a function and satisfy the bounded difference conditions
sup x 1 , , x n , x s A | f ( x 1 , , x n ) f ( x 1 , , x s 1 , x s , x s + 1 , , x n ) | C s ,
then for all t > 0 , P | f ( X 1 , , X n ) E f ( X 1 , , X n ) | t 2 exp 2 t 2 s = 1 n C s 2 .
We define the KKT conditions of optimization evaluated at β * (it is from the sub-gradient of the optimization function evaluated at β * ) by the events below:
F k ( ω k ) : = 1 n i = 1 n h k ( X i ) j = 1 W β j * < h j , h k > c β k * ω k , k = 1 , 2 , , W .
Assume that
  • (H.2): L k > 0 s.t. h k = max 1 i n | h k ( X i ) | 2 L k ;
  • (H.3): 0 < max 1 j W | β j * | B .
(H.2) is an assumption in sparse 1 estimation, and the assumption (H.3) is a classical compact parameter space assumption in sparse high-dimensional regressions (see [9,25]).
Next, we check that the event F k ( ω k ) is hold with high probability. Note that E h k ( X i ) = j = 1 W β j * < h j , h k > (which is free of X i ), we find
1 n i = 1 n h k ( X i ) ( i s n h k ( X i ) + h k ( X s ) ) = 1 n h k ( X s ) h k ( X s ) 1 n ( | h k ( X s ) | + | h k ( X s ) | ) 4 L k n ,
where the last inequality is due to | h k ( X i ) | max 1 i n | h k ( X i ) | 2 L k .
Next, we apply the McDiarmid’s inequality on the event F k c ( ω k ) by (H.3). Then
P ( F k c ( ω k ) ) = P 1 n i = 1 n h k ( X i ) j = 1 W β j * < h j , h k > c β k * ω k ( by ( H . 3 ) ) P 1 n i = 1 n h k ( X i ) E h k ( X i ) ω k c B ( define ω ˜ k : = ω k c B > 0 ) 2 exp 2 ω ˜ k 2 16 L k 2 / n = 2 exp n ω ˜ k 2 8 L k 2 = : δ W , 0 < δ < 1 .
Considering the previous line,
ω k : = 2 2 L k 1 n log 2 W δ + c B = : 2 2 L k v ( δ / 2 ) + c B , where v = v ( δ ) : = 1 n log W δ .
The weight ω k in our paper is different from [9], which gives the un-shift version ( ω k ˇ = 4 L k 1 n log W δ / 2 ), due to the Elastic-net penalty. Define the modified KKT conditions:
K k ( ω k ) : = { | 1 n i = 1 n h k ( X i ) j = 1 W β j * < h j , h k > | ω ˜ k } , k = 1 , 2 , , W
which hold with probability of at least 1 2 exp n ω ˜ k 2 8 L k 2 .

3.2. Non-Asymptotic Oracle Inequalities

Introduced by [30], oracle inequality is a powerful non-asymptotic and analytical tool that seeks to provide the distance between the obtained estimator and a true estimator. The sharp oracle inequality connects the optimal convergence of an obtained estimator compared with the true parameter (see [31,32]).
For β R W , let I ( β ) = { j { 1 , , W } : β j 0 } be the indices corresponding to the non-zero components of the vector β , i.e., the support in mathematical jargon. If there is no ambiguity, we would like to write I ( β * ) as I * for simplicity. Define W ( β ) = j = 1 W I ( β j 0 ) as the number of its non-zero components, where I ( · ) represents the indicative function. Let σ j 2 = V a r ( h j ( X 1 ) ) , 1 j W .
Below, we will state the non-asymptotic oracle inequalities for h β ^ (with the probability at least 1 δ ( W , n ) for any integer W and n), which measures the L 2 distance between h β ^ and h. For β R W , define the correlation for the two base densities: h i and h j , ρ W ( i , j ) = < h i , h j > h i h j , i , j = 1 , , W . Our results will be established under the local coherence condition, and we define the maximal local coherence as:
ρ ( β ) = max i I ( β ) max j i | ρ W ( i , j ) | .
It is easy to see that ρ ( β ) measures the separation of the variables in the set I ( β ) from one another and the rest. The degree of separation is measured in terms of the size of the correlation coefficients. However, the regular condition introduced by this coherence may be too strong. It may exclude cases that the “correlation” can be relatively significant for a small number of pairs ( i , j ) and almost zero otherwise. Thus, we consider the definition of the cumulative local coherence given by [9]: ρ * ( β ) = i I ( β ) j > i | ρ W ( i , j ) | . Define H ( β ) = max j I ( β ) ω j v ( δ / 2 ) h j , F = max 1 j W v ( δ / 2 ) h j ω ˜ j = max 1 j W h j 2 2 L j , where ω ˜ j : = 2 2 L j v ( δ / 2 ) .
By using the definition of ρ * ( β ) and the notations above, we present the main results of this paper, which lays the foundation for the oracle inequality of the estimated mixture coefficients.
Theorem 1.
Under (H.1)–(H.3), let c = min 1 j W { ω ˜ j } B and a given constant 0 < γ 1 . If the true base functions { h j } j = 1 W conform to the cumulative local coherence assumption for all β R W ,
12 F H ( β ) ρ * ( β ) W ( β ) γ ,
then the β ^ of the optimization problem in Equation (3) has the following oracle inequality with a probability at least 1 δ ,
h β ^ h 2 + α o p t 1 ( 1 γ ) ( α o p t 1 1 ) j = 1 W ω ˜ j | β ^ j β j | + α o p t 1 α o p t 1 1 j = 1 W c ( β ^ j β j ) 2 α o p t 1 + 1 α o p t 1 1 h β h 2 + 18 α o p t 1 2 α o p t 1 1 H 2 ( β ) v 2 ( δ / 2 ) W ( β ) ,
where α o p t 1 = 1 + 1 + h β h 2 9 H 2 ( β ) v 2 ( δ / 2 ) W ( β ) .
It is worthy to note that here we use W ( β ) instead of W ( β ) , and the latter is used in [9]. The upper bound of the oracle inequality by Theorem 1 is sharper than the upper bound of Theorem 1 in [9]. Further, we give the value of the optimal α o p t 1 , but [9] did not give it. The reason for this phenomenon is quite clean actually: from the proof, it is due to ineuqality (A5). Now, let us address the sparse Gram matrix Ψ W = ( < h i , h j > ) 1 i , j W with a small number of non-zero elements in off-diagonal positions, define ψ W ( i , j ) as the element ( i , j ) -th of position ψ W . Condition (8) in Theorem 1 can be transformed to the condition
12 S H ( β ) W ( β ) γ ,
where the number S is called the sparse index of matrix Ψ W , which is defined as S = | { ( i , j ) : i , j { 1 , , W } , i > j and ψ W ( i , j ) 0 } | , where | A | is the number of elements of set A.
Sometimes the assumption in Condition (8) does not imply the positive definiteness of Ψ W . Next, we give a similar oracle inequality that is valid under the hypothesis that the Gram matrix Ψ W is positive definite.
Theorem 2.
Under the assumption of (H.1)–(H.3) and that the Gram matrix Ψ W is positive definite with a minimum eigenvalue greater than or equal to λ W > 0 . For all β R W , the β ^ of the optimization problem in Equation (3) has the following oracle inequality with probability at least 1 δ ,
|| h β ^ h 2 + α o p t 2 α o p t 2 1 j = 1 W ω ˜ j | β ^ j β j | + α o p t 2 α o p t 2 1 j = 1 W c ( β ^ j β j ) 2 α o p t 2 + 1 α o p t 2 1 h β h 2 + 576 α o p t 2 2 α o p t 2 1 G λ W v 2 ( δ / 2 ) ,
where G = G ( β ) : = j I ( β ) L j 2 and α o p t 2 = 1 + 1 + h β h 2 288 G λ W v 2 ( δ / 2 ) .
Remark 1.
The argument and result of Theorem 1 in this paper is more refined than the conclusion of Theorem 1 in [9] for Lasso by letting γ = 1 / 2 and c = 0 . In addition, Theorems 1 and 2 of this paper, respectively, give the optimal α value of the density estimation oracle inequalities, namely α o p t 1 , α o p t 2 . It provides a potentially sharper bound for the 1 -estimation error bound.
Next, we will present the 1 -estimation error for the estimator β ^ by Equation (3), and the weights are defined by Equation (6). For the technical point, we consider that h j = 1 for all j in Equation (3), i.e., the base functions are normalized. This normalization mimics the covariates’ standardization procedure when doing penalized estimations in generalized linear models. For simplicity, we put L : = max 1 j W L j .
For any other choice of v ( δ / 2 ) greater than or equal to 1 n log 2 W δ , the conclusions of Section 3 are valid with a high probability. It imposes a restriction on the predictive performance of CSDE. As pointed out in [33], for the 1 -penalty in the regression, the adjusted sequence ω j required for the corrected selection is usually larger than the adjusted sequence ω j that produces a good prediction. The selection of the mixture density shown below is also true. Specifically, we will take the value β = β * and v = v ( δ / 2 W ) = log ( 2 W 2 / δ ) n , then α o p t 1 , α o p t 2 = 2 . Below, we give the Corollaries of Theorems 1 and 2.
Corollary 1.
Given the same conditions as Theorem 1 with h j = 1 for all j, let α o p t 1 = 2 , then we have the following 1 -estimation error oracle inequality:
j = 1 W | β ^ j β j * | 72 2 v ( δ / 2 W ) W ( β * ) 1 γ ( L + L min ) 2 L min
with probability at least 1 δ / W , where L min = min 1 j W L j .
Corollary 2.
Given the same conditions as Theorem 2 with h j = 1 for all j, let α o p t 2 = 2 , then we have the following oracle inequality, with probability at least 1 δ / W ,
j = 1 W | β ^ j β j * | 288 2 v ( δ / 2 W ) G * L min λ W ,   w h e r e   G * = j I * L j 2 .
If the number W ( β * ) of the mixture indicator elements is much smaller than n , then inequality (9) guarantees that the estimated β ^ is close to the true β * , and the 1 -estimation error will be presented in the numerical simulation in Section 4. Our results of Corollaries 1 and 2 are non-asymptotic for any W and n. The oracle inequalities are guiders for us to find an optimal tuning parameter with order O ( log W n ) for a sharper estimation error and better prediction performance. This is also an intermediate and crucial result, which leads to the main results of correctly identifying the mixture components in Section 3.3. In the following section, we turn to cope with the identification of I * . Corrected components are selected by the proposed oracle inequality for the weighted 1 + 2 penalty.

3.3. Corrected Support Identification of Mixture Models

In this section, we will study the results of the support recovery of our CSDE estimator. There are few results on support recovery, while most of the results are the consistency of the 1 -error and prediction errors. Here, we borrow the framework of [25,33]. They give many proof techniques to deal with the corrected support identification in linear models by 1 + 2 regularization. Let I ^ be the set of indicators consisting of non-zero elements of β ^ in the given Equation (3). In other words, I ^ is an estimate of the true support set I ( β * ) : = I * . We will study P ( I ^ = I ( β * ) ) 1 ε for a given 0 < ε < 1 under some mild conditions.
To identify the I * consistently, we need more assumptions about some special correlation conditions than 1 -error consistency.
Condition (A): ρ * ( β * ) L L min λ W 288 G * .
Moreover, we need an additional condition that the minimal signal should be higher than a threshold level and quantified by order of tuning parameter. Therefore, we state it as follows:
Condition (B): min j I * | β j * | 4 2 v ( δ 2 W ) L , where v ( δ 2 W ) : = 1 n log 2 W 2 δ .
When performing simulation, Condition (B) is the theoretical guarantee that the minor magnitude of β j must be greater than a threshold value as a minimal signal condition. It is also called the Beta-min condition (see [26]).
Theorem 3.
Let 0 < δ < 1 2 and define ϵ k : = | E [ h k ( X 1 ) ] E [ h k ( Z 1 ) ] | . Assume that both conditions (A) and (B) are true and give the same conditions as Corollary 2, then
P ( I ^ = I * ) 1 4 W δ 2 W 2 ( 1 ϵ k * ) 2 + 2 δ ,   w h e r e   ϵ k * = ϵ k / 2 v ( δ / 2 W ) L .
Under the Beta-min condition, the support estimation is very close to the true support of β j * . The probability of the event { I ^ = I * } is high when W is growing. The β ^ recovers the corrected support with probability at least 1 ( 4 W ( δ 2 W 2 ) ( 1 ϵ k * ) 2 + 2 δ ) . The result is non-asymptotic and it is true for any fixed W and n. There is a similar conclusion about support consistency (see Theorem 6 of [25]).

4. Simulation and Real Data Analysis

Ref. [9] proposes the SPADES estimation to deal with the samples for sparse mixture density, and they also derive an algorithm from complementing their theoretical result. Their findings successfully handle the high-dimensional adaptive density estimation to some degree. However, their algorithm is costly and unstable. In this section, we deal with the tuning parameter directly and compare our CSDE method with the SPADES method in [9] and other similar methods. In all cases here, we fix n = 100 for W = 81 , 131 , 211 , 321 , which is known as the dimension of the unknown parameter β * . The performance of each estimator is evaluated by the 1 -estimation error and the total variation (TV) distance between the estimator and the true value of β * . The total variation (TV) error is defined as: TV ( h β * , h β ^ ) = | h β * ( x ) h β ^ ( x ) | d x .

4.1. Tuning Parameter Selection

In [9], the λ 1 is chosen by the coordinate descent method, while the mixture weights are detected by GBM. However, in our article, the optimal weights can be computed directly. Thus, it is much easier to carry out than [9]. The 1 -penalty term j = 1 W ω j | β j | with optimal weights are defined by ω k : = 2 2 L k v ( δ / 2 ) + c B , where L j = h j , which usually can be computed easily for a continuous h j .
For a discrete base density { h j } j = 1 W , it can be estimated as the following approximation by using concentration inequalities from Exercise 4.3.3 of [34]: | med ( X ) E ( X ) | 2 V a r ( X ) , x ¯ x m e d 1 + O ( n 1 ) h 1 ( L j ) 1 + O ( n 1 ) , where x ¯ and x m e d represent the sample mean and sample median, respectively, in each simulation, then we only need to select the λ 1 and c = λ 2 , and they can be detected by the nesting coordinate descent method. Moreover, the precision level is assigned as ξ = 0.001 in our simulation.

4.2. Multi-Modal Distributions

First, we examine our method in a multi-modal Gaussian model that is similar to the first model in [9]. However, our mixture Gaussian has a different variance, which leads the meaningful weights to our estimation. The density function for the i.i.d. sample Z 1 , , Z n is assigned as follows:
h β * ( x ) = j = 1 W β j * ϕ x | a j , σ j ,
where ϕ x | a j , σ j is the density of N ( a j , σ j 2 ) . However, to estimate β * , we only observe i.i.d. data X 1 , , X n with density g β j * ( x ) = j = 1 W β j * ϕ ( x | a j , 1.1 σ j ) . Put a = 0.5 , n = 100 and
β * = 0 8 T , 0.2 , 0 10 T , 0.1 , 0 5 T , 0.1 , 0 10 T , 0.1 , 0 10 T , 0.1 , 0 5 T , 0.15 , 0 10 T , 0.15 , 0 10 T , 0.1 , 0 W 76 T T ,
with σ = 1 20 T , 0.8 6 T , 0.6 11 T , 0.4 11 T , 0.6 6 T , 0.8 11 T , 1.2 W 76 T T .
We replicate the simulation N = 100 times. Simulation results are presented in Table 1, from which we can see our method has more and more excellent performances as the W increases, which matches the non-asymptotic results in the previous section. The best performance is far better than the other three methods when W = 321 . It is worthy to note that the better approximation follows the increase in W, matching Equation (7) and Theorem 3 in our previous section.
We plot the solution path to compare the performance of the four estimators in β j I ( β ) for every W in Figure 1 (the result of Elastic-net in W = 321 is not be shown due to its poor performance.). These figures also provide strong support for the above analysis. Meanwhile, we plot the probability densities of the several estimators and the true density to complement the visual sensory of the advantage in our method in Figure 2. The robust competency of detecting the multi-mode is shown (whereas other methods only find the most strongest signal, ignoring other meaningful but relatively weak signals).

4.3. Mixture of Poisson Distributions

We study the mixture of discrete distribution: the mixture Poisson distribution
h β * ( x ) = j = 1 W β j * p x | λ j = a · j ,
where p x | λ j = a · j is the probability mass function (p.m.f.) of the Poisson distribution with mean λ j . We set a = 0.1 and
β * = 0 8 T , 0.2 , 0 10 T , 0.1 , 0 5 T , 0.1 , 0 10 T , 0.1 , 0 10 T , 0.1 , 0 5 T , 0.15 , 0 10 T , 0.15 , 0 10 T , 0.1 , 0 W 75 T T .
The adjusted weights are calculated by Equation (3), and in discrete distributions, we define f , g = k = 1 f ( k ) g ( k ) . Meanwhile, the Poisson random variable with measurement errors can be treated as a negative binomial random variable. Let n ( x | λ j , r ) be the p.m.f. of the Poisson distribution with the mean λ j and dispersion parameter r. Suppose the observed data with sample size n = 100 has the p.m.f.
g β * ( x ) = j = 1 W β j * n x | λ j = a · j , r ,
where r = 6 , which leads to an increment of variance from Poisson to the negative binomial distribution. Similarly, we replicate each simulation to estimate the parameter N = 100 times with the sample from the mixture negative binomial distribution above. The result is shown in Table 2. The result is actually akin to that in the previous mixture Gaussian distribution, while the strong performance of our method is shown clearly when W is considerable.

4.4. Low-Dimensional Mixture Model

Surprisingly, our method has more competitive efficacy than some popular methods (such as EM algorithm), even the dimension W is relatively small. To see this, we introduce the following numerical experiments to estimate the weights of the low-dimensional Gaussian mixture model: the samples X 1 , , X n come from the model: h β * ( x ) = j = 1 W β j * ϕ x | μ j , σ j . The updated equation for the EM algorithm in t-th step is:
ω i j ( t ) = p j ( t ) ϕ x i ; μ t , σ t s = 1 W p s ( t ) ϕ x i ; μ s , σ s , β j ( t + 1 ) = i = 1 W ω i j ( t ) i = 1 n j = 1 W ω i j ( t ) .
Here, we consider two scenarios:
( 1 ) W = 6 , β = ( 0.3 , 0 , 0 , 0.3 , 0 , 0.4 ) T , μ = ( 0 , 10 , 20 , 30 , 40 , 50 ) T , σ = ( 1 , 2 , 3 , 4 , 5 , 6 ) T ; ( 2 ) W = 7 , β = ( 0.1 , 0 , 0 , 0.8 , 0 , 0 , 0.1 ) T , μ = ( 0 , 1 , 2 , 3 , 4 , 5 , 6 ) T , σ = ( 0.3 , 0.2 , 0.2 , 0.1 , 0.2 , 0.2 , 0.3 ) T .
For each scenario n = 50 , and the fitter levels (cessation level) in the EM approach and our method are both ξ = 10 4 . A well-advised initial value in the EM approach is the equal weight.
We replicate the simulation N = 100 times, and the optimal tuning parameters stem from the cross-validation (CV). Thus, under each simulation, they are not the same, albeit they are very close to each other. The result can be seen in Table 3.

4.5. Real Data Examples

Practically, we consider using our method to estimate some densities in the environmental science field. Wind, which is mercurial, has been an advisable object to study for a long time in meteorology. Please note that the wind’s speed at one specific location may not be diverse so we will use the wind’s azimuth angle with a more sparse density at two sites in China. Many types of research about the estimated density for wind exist, so there is a possibility of using our approach to cope with some difficulties in meteorology science.
There have been some very credible meteorological dataset. We used the ERA5 hourly data in [35] to continue our analysis. We want chose a continental area and a coastal area in China, so we chose Beijing Nongzhanguan and Qingdao Coast. The locations of these two areas are: (116.3125 ° E, 116.4375 ° E) × (39.8125 ° N, 39.8125 ° N). Take notice that the wind in one day may be highly correlated. Therefore, using the data at a specific time point of each day in a consecutive period as i.i.d. samples is reasonable. The sample histograms at 6 am in Beijing Nongzhanguan and at midnight on the Qingdao Coast are shown in Figure 3. Here, we used the data from 1 January 2013 to 12 December 2015.
As we can see, their density does multi-peak (we used 1095 samples). Now, we can use our approach to estimate the multi-mode densities based on a relatively small size of samples, which is only a tiny part of the whole data from 1 January 2013 to 12 December 2015. Because one year has about 360 days, we may assume that every day is a latent factor that forms the base density. Thus, the model is designed as h β * ( x ) = j = 1 360 β j * ϕ x | μ j , σ j with the mean and variance parameters μ = ( 1 , 2 , , 360 ) T , σ = t · 1 360 T , where t is the bandwidth (or tuning parameter). With the different sub-samples, the computed values are different.
Another critical issue is how to choose the tuning parameters λ 1 and λ 2 . Then, we apply the cross-validation criterion, namely choosing λ i , to minimize the difference between the two estimators derived from the separated samples in a random dichotomy.
Now, start to construct the samples for the estimating procedure. Assume that an observatory wants to figure out some information about the two areas’ wind. However, it does not have intact data due to the limited budget at its inception. The only samples it has are several days’ information each month for the two areas, and these days scatter randomly. Furthermore, sample size n = 168 exactly. These imperfect data increase the challenge of estimating a trustworthy density. We compared our method with other previous methods, in which appraising the difference between the complete data sample histogram and the estimated density under each method is for the evaluation. Notice that the samples are only a tiny part of the data, so the n = 168 < M ( = 360 ) is relatively small. The small sample and large dimension setting coincide with the non-asymptotic theory provided in the previous section. The estimated density has been shown in Figure 4.
In this practical application, our method vindicates its more efficient estimating performance and stability from its propinquity of the complete sample histogram, namely the productive capacity of detecting the shape of the multi-mode density and the stronger inclination to bear a resemblance to each other sub-sample (although some subtle nuances do exist because of the different sub-sample). An alternative approach can be to consider principles and tools of circular statistics, which has been reviewed in [36].

5. Summary and Discussions

The paper deals with the deconvolution problem using Lasso-type methods: the observations X 1 , , X n are independent and generated from X i = Z i + ε i , and the goal is to estimate the unknown density h of the Z i . We assume that the function h can be written as h ( · ) = h β * ( · ) = j = 1 W β j * h j ( · ) based on some functions { h j } j = 1 W from a specific dictionary and propose estimating the coefficients of this decomposition with the Elastic-net method. For this estimator, we show that under some classical assumptions of the model, such as coherence of the Gram matrix, finite sample bounds for the estimation and the prediction errors valid with a relatively high probability can be obtained. Moreover, we prove a variable selection consistency result under a beta-min condition and conduct an extensive numerical study. The following estimation problem is also similar to the CSDE.
For future study, it is also interesting and meaningful to do hypothesis testing about the coefficients β * R W in sparse mixture models. For a general function h : R W R m and a nonempty closed set Ω R m , we can consider
H 0 : h β * Ω vs . H 1 : h β * Ω .
It is possible to use [37] as a general approach to hypothesis testing within models with measurement errors.

Author Contributions

Conceptualization, H.Z.; Data curation, X.Y.; Formal analysis, X.Y. and S.Z.; Funding acquisition, X.Y.; Investigation, X.Y. and S.Z.; Methodology, X.Y., H.Z., H.W. and S.Z.; Project administration, H.Z.; Resources, H.W.; Software, H.W.; Supervision, X.Y. and H.Z.; Visualization, H.W.; Writing—original draft, X.Y., H.Z., H.W. and S.Z.; Writing—review and editing, H.Z. and H.W. All authors have read and agreed to the published version of the manuscript.

Funding

Xiaowei Yang is supported in part by the General Research Project of Chaohu University (XLY-201906), Chaohu University Applied Curriculum Development Project (ch19yykc21), Key Project of Natural Science Foundation of Anhui Province Colleges and Universities (KJ2019A0683), Key Scientific Research Project of Chaohu University (XLZ-202105). Huiming Zhang is supported in part by the University of Macau under UM Macao Talent Program (UMMTP-2020-01). This work also is supported in part by the National Natural Science Foundation of China (Grant No. 11701109, 11901124) and the Guangxi Science Foundation (Grant No. 2018GXNSFAA138164).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors would like to thank Song Xi Chen’s Group https://songxichen.com/(accessed on 20 December 2021) for sharing the meteorological dataset.

Acknowledgments

The Appendix includes the proofs of the lemmas, corollaries and theorems in the main body.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

For convenience, we first give a preliminary lemma and proof. Define the random variables M j = 1 n i = 1 n h j ( X i ) E h j ( X i ) . Consider the event E by E = j = 1 W { 2 | M j | ω ˜ j } , where ω ˜ k : = 2 2 L k 1 n log W δ / 2 = : 2 2 L k v ( δ / 2 ) . Then, we have the following lemma, which is cornerstone for the proofs in below.
Lemma A1.
Suppose max 1 j W L j < and c = min 1 j W { ω ˜ j } B , for any β R W on the event E , we have h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + 6 j I ( β ) ω j | β ^ j β j | .

Appendix A.1. Proof of Lemma A1

According to the definition of β ^ , for any β R W , we find 2 n i = 1 n h β ^ ( X i ) + h β ^ 2 + 2 j = 1 W ω j | β ^ j | + c j = 1 W β ^ j 2 2 n i = 1 n h β ( X i ) + h β 2 + 2 j = 1 W ω j | β j | + c j = 1 W β j 2 . Then
h β ^ 2 h β 2 2 n i = 1 n h β ^ ( X i ) 2 n i = 1 n h β ( X i ) + 2 j = 1 W ω j | β j | 2 j = 1 W ω j | β ^ j | + c j = 1 W β j 2 c j = 1 W β ^ j 2 .
Note that
h β ^ h 2 = h β ^ h β + h β h 2 = h β ^ h β 2 + h β h 2 + 2 < h β h , h β ^ h β > = h β h 2 2 < h , h β ^ h β > + 2 < h β , h β ^ h β > + h β ^ h β 2 = h β h 2 2 < h , h β ^ h β > + h β ^ 2 h β 2 .
Combining the two result above, we obtain
h β ^ h 2 h β h 2 + 2 j = 1 W ω j | β j | 2 j = 1 W ω j | β ^ j | + c j = 1 W β j 2 c j = 1 W β ^ j 2 2 < h , h β ^ h β > + 2 n i = 1 n h β ^ ( X i ) 2 n i = 1 n h β ( X i ) .
According to the definition of h β ( x ) , it gives h β ( x ) = j = 1 W β j h j ( x ) with β = ( β 1 , , β W ) . For the three terms in Equation (A1), we have
2 < h , h β ^ h β > + 2 n i = 1 n h β ^ ( X i ) 2 n i = 1 n h β ( X i ) = 2 · 1 n i = 1 n j = 1 W β ^ j h j ( X i ) j = 1 W β j h j ( X i ) 2 E ( h β h β ) ( X i ) | β = β ^ = 2 j = 1 W 1 n i = 1 n h j ( X i ) ( β ^ j β j ) 2 j = 1 W E [ h j ( X i ) ] ( β ^ j β j ) = 2 j = 1 W 1 n i = 1 n h j ( X i ) E [ h j ( X i ) ] ( β ^ j β j ) .
Then
h β ^ h 2 h β h 2 + 2 j = 1 W 1 n i = 1 n h j ( X i ) E [ h j ( X i ) ] ( β ^ j β j ) + 2 j = 1 W ω j | β j | 2 j = 1 W ω j | β ^ j | + c j = 1 W β j 2 c j = 1 W β ^ j 2 .
Conditioning on E , we have h β ^ h 2 h β h 2 + j = 1 W ω ˜ j | β ^ j β j | + 2 j = 1 W ω j ( | β j | | β ^ j | ) + c j = 1 W ( β j 2 β ^ j 2 ) . We add j = 1 W ω ˜ j | β ^ j β j | + c j = 1 W ( β j β ^ j ) 2 to both sides of the inequality, it gives
h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + c j = 1 W ( β j β ^ j ) 2 h β h 2 + 2 j = 1 W ω ˜ j | β ^ j β j | + 2 j = 1 W ω j ( | β j | | β ^ j | ) + c j = 1 W ( β j 2 β ^ j 2 ) + c j = 1 W ( β j β ^ j ) 2 .
Note that
c [ j = 1 W ( β j 2 β ^ j 2 ) + j = 1 W ( β j β ^ j ) 2 ] = c [ j = 1 W ( β j 2 β ^ j 2 + β j 2 2 β j β ^ j + β ^ j 2 ) ] = 2 c j = 1 W β j ( β j β ^ j ) = 2 c j I ( β ) β j ( β j β ^ j ) 2 c B j I ( β ) | β j β ^ j | 2 j I ( β ) ω ˜ j | β j β ^ j | ,
where the last inequality is due to the assumption c = min 1 j W { ω ˜ j } B ω ˜ j B . Thus, we obtain
h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + c j = 1 W ( β ^ j β j ) 2 h β h 2 + 2 j = 1 W ω ˜ j | β ^ j β j | + 2 j = 1 W ω j ( | β j | | β ^ j | ) + 2 j I ( β ) ω ˜ j | β ^ j β j | h β h 2 + 2 j = 1 W ω j | β ^ j β j | + 2 j = 1 W ω j ( | β j | | β ^ j | ) + 2 j I ( β ) ω j | β ^ j β j | ,
where the last inequality follows from ω ˜ j ω j for all j.
We know β j 0 if j I ( β ) , and β j = 0 if j I ( β ) . Considering | β j | | β ^ j | | β ^ j β j | for all j, we have 2 j = 1 W ω j | β ^ j β j | + 2 j = 1 W ω j ( | β j | | β ^ j | ) 4 j I ( β ) ω j | β ^ j β j | . Then
|| h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + c j = 1 W ( β ^ j β j ) 2 h β h 2 + 4 j I ( β ) ω j | β ^ j β j | + 2 j I ( β ) ω j | β ^ j β j | = h β h 2 + 6 j I ( β ) ω j | β ^ j β j | .

Appendix A.2. Proof of Theorems

According to ω ˜ j = 2 2 L j 1 n log 2 W δ in Equation (6), the sum of the independent random variables ζ i j = h j ( X i ) E h j ( X i ) is determined by Hoeffding’s inequality, and | h j ( X i ) | 2 L j . We obtain
P ( E c ) = P j = 1 W { 2 | M j | > ω ˜ j } j = 1 W P ( 2 | M j | > ω ˜ j ) 2 j = 1 W exp 2 n 2 · ω ˜ j 2 / 4 4 n L j 2 = 2 j = 1 W exp log 2 W δ = 2 W · δ 2 W = δ .

Appendix A.3. Proof of Theorem 1

By Lemma A1, we need an upper bound on j I ( β ) ω j | β ^ j β j | . For easy notation, let q j = β ^ j β j , Q ( β ) = j I ( β ) | q j | h j , Q = j = 1 W | q j | h j . According to the definition of H ( β ) , that is, H ( β ) = max j I ( β ) ω j v ( δ / 2 ) h j , we have
j I ( β ) ω j | β ^ j β j | v ( δ / 2 ) H ( β ) Q ( β ) .
Let Q * ( β ) : = j I ( β ) q j 2 h j 2 . Using the definition of h β ( x ) , we obtain Q * 2 ( β ) = j I ( β ) q j 2 h j 2 = h β ^ h β 2 i , j I ( β ) q i q j < h i , h j > ( 2 i I ( β ) j I ( β ) q i q j < h i , h j > + i , j I ( β ) , i j q i q j < h i , h j > ) . As i , j I ( β ) , β i = β j = 0 , it is easy to see i , j I ( β ) < h i , h j > q i q j 0 . Observe that
2 i I ( β ) j I ( β ) q i q j < h i , h j > + i , j I ( β ) , i j q i q j < h i , h j > = 2 i I ( β ) j I ( β ) q i q j < h i , h j > + 2 i , j I ( β ) , j > i q i q j < h i , h j > = 2 i I ( β ) , j > i q i q j < h i , h j > .
By the definitions of ρ W ( i , j ) and ρ * ( β ) , then
Q * 2 ( β ) h β ^ h β 2 + 2 i I ( β ) , j > i | q i | | q j | h i h j < h i , h j > h i h j h β ^ h β 2 + 2 ρ * ( β ) max i I ( β ) , j > i | q i | h i | q j | h j .
By max i I ( β ) | q i | h i j I ( β ) q j 2 h j 2 = Q * ( β ) , max i I ( β ) , j > i | q j | h j j = 1 W | q j | h j ,
Q * 2 ( β ) h β ^ h β 2 + 2 ρ * ( β ) Q * ( β ) j = 1 W | q j | h j = h β ^ h β 2 + 2 ρ * ( β ) Q * ( β ) Q .
By Equation (A3), we can obtain Q * 2 ( β ) 2 ρ * ( β ) Q * ( β ) Q h β ^ h β 2 0 .
To find the upper bound of Q * ( β ) , applying the properties of the quadratic inequality to the above formula, we obtain that
Q * ( β ) ρ * ( β ) Q + ρ * 2 ( β ) Q 2 + h β ^ h β 2 ρ * ( β ) Q + [ ρ * ( β ) Q + h β ^ h β ] 2 ρ * ( β ) Q + h β ^ h β .
Note that W ( β ) = | I ( β ) | = j = 1 W I ( β j 0 ) , employing the Cauchy–Schwarz inequalities, we have
W ( β ) j I ( β ) | q j | 2 h j 2 = j I ( β ) I 2 ( j I ( β ) ) j I ( β ) | q j | 2 h j 2 ( j I ( β ) I ( { j I ( β ) ) | q j | h j ) 2 = Q 2 ( β ) .
Then, Q * 2 ( β ) = j I ( β ) | q j | 2 h j 2 Q 2 ( β ) / W ( β ) . In combination with Equation (A4), we can obtain Q ( β ) / W ( β ) Q * ( β ) 2 ρ * ( β ) Q + h β ^ h β . Therefore,
Q ( β ) 2 ρ * ( β ) W ( β ) Q + W ( β ) h β ^ h β .
By Lemma A1 and Equation (A2), we have the following inequality with probability exceeding 1 δ ,
h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + 6 j I ( β ) ω j | β ^ j β j | h β h 2 + 6 v ( δ / 2 ) H ( β ) Q ( β ) h β h 2 + 6 v ( δ / 2 ) H ( β ) [ 2 ρ * ( β ) W ( β ) j = 1 W | q j | h j + W ( β ) h β ^ h β ] ( by ( A 5 ) ) = h β h 2 + 12 v ( δ / 2 ) H ( β ) ρ * ( β ) W ( β ) j = 1 W ω ˜ j | β ^ j β j | h j ω ˜ j + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h β h β h 2 + 12 F H ( β ) ρ * ( β ) W ( β ) j = 1 W ω ˜ j | β ^ j β j | + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h β h β h 2 + γ j = 1 W ω ˜ j | β ^ j β j | + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h β ,
where the second last inequality follows from the definition of F : = max 1 j W v ( δ / 2 ) h j ω ˜ j , and the last inequality is derived by the assumption 12 F H ( β ) ρ * ( β ) W ( β ) γ , ( 0 < γ 1 ) .
Further, we can find that, with probability at least 1 δ ,
h β ^ h 2 + ( 1 γ ) j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h β = h β h 2 + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h + h h β h β h 2 + 6 v ( δ / 2 ) H ( β ) W ( β ) h β ^ h + 6 v ( δ / 2 ) H ( β ) W ( β ) h h β .
Using the elementary inequality 2 s t s 2 / α + α t 2 ( s , t R , α > 1 ) to the last two terms of the above inequality, it yields
2 { 3 v ( δ / 2 ) H ( β ) W ( β ) } h β ^ h α · 9 v 2 ( δ / 2 ) H 2 ( β ) W ( β ) + h β ^ h 2 / α , 2 { 3 v ( δ / 2 ) H ( β ) W ( β ) } h β h α · 9 v 2 ( δ / 2 ) H 2 ( β ) W ( β ) + h β h 2 / α .
Thus,
h β ^ h 2 + ( 1 γ ) j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + 18 α v 2 ( δ / 2 ) H 2 ( β ) W ( β ) + h β ^ h 2 / α + h β h 2 / α .
Simplifying, we have
h β ^ h 2 + α ( 1 γ ) ( α 1 ) j = 1 W ω ˜ j | β ^ j β j | + α α 1 j = 1 W c ( β ^ j β j ) 2 α + 1 α 1 h β h 2 + 18 α 2 α 1 H 2 ( β ) v 2 ( δ / 2 ) W ( β ) , α > 1 , 0 < γ 1 .
Optimizing α to obtain the sharp upper bounds for the above oracle inequality
α o p t 1 : = arg min α > 1 α + 1 α 1 h β h 2 + 18 α 2 α 1 H 2 ( β ) v 2 ( δ / 2 ) W ( β ) = 1 + 1 + h β h 2 9 H 2 ( β ) v 2 ( δ / 2 ) W ( β )
by the first order condition. To date, Theorem 1 is proved by substituting α o p t 1 into Equation (A6).

Appendix A.4. Proof of Theorem 2

By the minimal eigenvalue assumption for ψ W , we have
h β 2 = || j = 1 W β j h j ( x ) || 2 = β T ψ W β λ W β 2 λ W j I ( β ) β j 2 .
Using the definition of ω j and assumption L min : = min 1 j W L j > 0 ,
ω j = 2 L j 2 log ( 2 W / δ ) n + c B 2 L j 2 L j 2 log ( 2 W / δ ) n + c B 2 L min .
Since c B = ω ˜ min = 2 2 L min v ( δ / 2 ) and v ( δ / 2 ) = log ( 2 W / δ ) n , we have ω j 4 2 L j v ( δ / 2 ) .
Let G ( β ) = j I ( β ) L j 2 , by the Cauchy–Schwartz inequality, we obtain
6 j I ( β ) ω j | β ^ j β j | 24 2 v ( δ / 2 ) j I ( β ) L j | β ^ j β j | 24 2 v ( δ / 2 ) j I ( β ) L j 2 j I ( β ) ( β ^ j β j ) 2 24 2 v ( δ / 2 ) G ( β ) λ W h β ^ h β ,
where the last inequality above is from Equation (A7) due to
h β ^ h β 2 = 1 i , j W ( β ^ i β i ) ( β ^ j β j ) < h i , h j > λ W j I ( β ) ( β ^ j β j ) 2 .
Let b ( β ) : = 12 2 v ( δ / 2 ) G ( β ) λ W , Lemma 2 implies
h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + 2 b ( β ) h β ^ h β = h β h 2 + 2 b ( β ) ( h β ^ h + h h β ) h β h 2 + 2 b ( β ) h β ^ h + 2 b ( β ) h β h .
Using the inequality 2 s t s 2 / α + α t 2 ( s , t R , α > 1 ) for the last two terms on the right side of the above inequality, we find
2 b ( β ) h β ^ h + 2 b ( β ) h β h h β ^ h 2 / α + b 2 ( β ) α + h β h 2 / α + b 2 ( β ) α = h β ^ h 2 / α + h β h 2 / α + 2 b 2 ( β ) α .
Thus,
h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 h β h 2 + h β ^ h 2 / α + h β h 2 / α + 2 b 2 ( β ) α
gives α 1 α h β ^ h 2 + j = 1 W ω ˜ j | β ^ j β j | + j = 1 W c ( β ^ j β j ) 2 α + 1 α h β h 2 + 2 α b 2 ( β ) . Therefore,
h β ^ h 2 + α α 1 j = 1 W ω ˜ j | β ^ j β j | + α α 1 j = 1 W c ( β ^ j β j ) 2 α + 1 α 1 h β h 2 + 2 α 2 α 1 b 2 ( β ) = α + 1 α 1 h β h 2 + 576 α 2 α 1 G ( β ) λ W v 2 ( δ / 2 ) .
To obtain the sharp upper bounds for the above oracle inequality, we optimize α
α o p t 2 : = arg min α > 1 α + 1 α 1 h β h 2 + 576 α 2 α 1 G ( β ) λ W v 2 ( δ / 2 ) = 1 + 1 + h β h 2 288 G ( β ) λ W v 2 ( δ / 2 ) ,
by the first-order condition. This completes the proof of Theorem 2.

Appendix A.5. Proof of Corollory 1

Let ω ˜ min : = min 1 j W ω ˜ j . We replace v ( δ / 2 ) in Theorem 1 by the larger value v ( δ / 2 W ) . Substituting β = β * in Theorem 1, we have
α o p t 1 ( 1 γ ) α o p t 1 1 j = 1 W ω ˜ j | β ^ j β j * | 18 α o p t 1 2 α o p t 1 1 H 2 ( β * ) v 2 ( δ / 2 W ) W ( β * )
by h = h β * . Since ω ˜ j ω ˜ min for all j, we obtain
j = 1 W | β ^ j β j * | 18 α o p t 1 1 γ · 1 ω ˜ min · max j I ( β ) ω j 2 h j 2 · W ( β * ) .
In this case, α o p t 1 = 2 , and h j = 1 ; thus,
β ^ β * 36 1 γ · max j I ( β ) ω j 2 ω ˜ min · W ( β * ) = 72 2 v ( δ / 2 W ) W ( β * ) 1 γ max j I ( β ) ( L j + L min ) 2 L min 72 2 v ( δ / 2 W ) W ( β * ) 1 γ ( L + L min ) 2 L min
from ω ˜ min = 2 2 v ( δ / 2 W ) L min and
ω j 2 = [ 2 2 v ( δ / 2 W ) ] 2 L j + ω ˜ min 2 2 v ( δ / 2 W ) 2 = [ 2 2 v ( δ / 2 W ) ] 2 [ L j + L min ] 2 . This completes the proof of Corollary 1.

Appendix A.6. Proof of Corollary 2

Let β = β * in Theorem 2, with α o p t 2 = 2 , we replace v ( δ / 2 ) in Theorem 2 by the larger value v ( δ / 2 W ) , then j = 1 W ω ˜ min | β ^ j β j * | j = 1 W ω ˜ j | β ^ j β j * | 576 α o p t 2 G * λ W v 2 ( δ / 2 W ) . By the definition of ω ˜ min , we can obtain
j = 1 W | β ^ j β j * | 576 α o p t 2 G * v 2 ( δ / 2 ) ω ˜ min λ W = 576 · 2 G * v 2 ( δ / 2 W ) 2 2 v ( δ / 2 W ) L min λ W = 288 2 G * v ( δ / 2 W ) L min λ W .
This concludes the proof of Corollary 2.

Appendix A.7. Proof of Theorem 3

The following lemma is by virtue of the KKT conditions. It derives a bound of P ( I * I ^ ) , which is easily analyzed.
Lemma A2
(Proposition 3.3 in [33]). P ( I * I ^ ) W ( β * ) max k I * P ( β ^ k = 0 and β k * 0 ) .
To present the proof of Theorem 3, we first notice that P ( I ^ I * ) P ( I * I ^ ) + P ( I ^ I * ) . Next, we control the probability on the right side of the above inequality.
For the control of P ( I * I ^ ) , by Lemma A2, it remains to bound P ( β ^ k = 0 and β k * 0 ) .
Below, we will use the conclusion of Lemma 2 (KKT conditions). Recall that E [ h k ( Z 1 ) ] = j I * β j * < h k , h j > = j = 1 W β j * < h k , h j > . Since we assume that the density of Z 1 is the mixture density h β * = j I * β j * h j . Therefore, for k I * , we have,
P ( β ^ k = 0 and β k * 0 ) = P 1 n i = 1 n h k ( X i ) j = 1 W β ^ j < h j , h k > 2 2 v ( δ / 2 W ) L k ; β k * 0 = P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] + E [ h k ( Z 1 ) ] j = 1 W β ^ j < h j , h k > 2 2 v ( δ / 2 W ) L k ; β k * 0 = P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] j = 1 W ( β ^ j β j * ) < h j , h k > 2 2 v ( δ / 2 W ) L k ; β k * 0 = P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] j k W ( β ^ j β j * ) < h j , h k > + β k * h k 2 2 2 v ( δ / 2 W ) L k P | β k * h k 2 2 2 v ( δ / 2 W ) L k 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] + j k W ( β ^ j β j * ) < h j , h k > (A9) P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] | β k * | h k 2 2 2 v ( δ / 2 W ) L k (A10) + P j k W ( β ^ j β j * ) < h j , h k > | β k * | h k 2 2 2 v ( δ / 2 W ) L k .
Similar to Lemma 2, for Equation (A9), we use Hoeffding’s inequality. Since h k = 1 for all k. Put ϵ k : = | E [ h k ( X 1 ) ] E [ h k ( Z 1 ) ] | . Consider Condition (B), min k I * | β k * | 4 2 v ( δ / 2 W ) L and L max 1 k W L k , then we have
P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] | β k * | h k 2 2 2 v ( δ / 2 W ) L k = P 1 n i = 1 n h k ( X i ) E [ h k ( X 1 ) ] + E [ h k ( X 1 ) ] E [ h k ( Z 1 ) ] | β k * | h k 2 2 2 v ( δ / 2 W ) L k P 1 n i = 1 n h k ( X i ) E [ h k ( X 1 ) ] | β k * | 2 2 v ( δ / 2 W ) L ϵ k P 1 n i = 1 n h k ( X i ) E [ h k ( X 1 ) ] 2 2 v ( δ / 2 W ) L 2 v ( δ / 2 W ) L ϵ k = P 1 n i = 1 n h k ( X i ) E [ h k ( X 1 ) ] 2 v ( δ / 2 W ) L ϵ k = P 1 n i = 1 n h k ( X i ) E [ h k ( X 1 ) ] 2 v ( δ / 2 W ) L ( 1 ϵ k * ) ( let ϵ k * = ϵ k / 2 v ( δ / 2 W ) L ) 2 exp 4 n 2 v 2 ( δ / 2 W ) L 2 ( 1 ϵ k * ) 2 4 n L 2 = 2 exp n ( 1 ϵ k * ) 2 log ( 2 W 2 / δ ) n = 2 δ 2 W 2 ( 1 ϵ k * ) 2 .
For the upper bound of Equation (A10), using Condition (A) and Condition (B), by the definitions of ρ * ( β * ) and W ( β * ) , we obtain
P j k W ( β ^ j β j * ) < h j , h k > | β k * | h k 2 2 2 v ( δ / 2 W ) L k = P j k W ( β ^ j β j * ) < h j , h k > | β k * | 2 2 v ( δ / 2 W ) L k P j k W ( β ^ j β j * ) < h j , h k > h j h k · h j h k 2 2 v ( δ / 2 W ) L 2 v ( δ / 2 W ) L P ρ * ( β * ) j k W β ^ j β j * 2 v ( δ / 2 W ) L P j = 1 W β ^ j β j * 2 v ( δ / 2 W ) L ρ * ( β * ) P j = 1 W β ^ j β j * 288 2 G * v ( δ / 2 W ) L min λ W δ W .
where the second last inequality is by Condition (A), and the last inequality above is by using the 1 -estimation oracle inequality in Corollary 2.
Therefore, by the definition of W ( β * ) , W ( β * ) = | I * | W , we find
P ( I * I ^ ) W ( β * ) max k I * P ( β ^ k = 0 ) W ( β * ) 2 δ 2 W 2 ( 1 ϵ k * ) 2 + W ( β * ) δ W 2 W δ 2 W 2 ( 1 ϵ k * ) 2 + W δ W = 2 W δ 2 W 2 ( 1 ϵ k * ) 2 + δ .
For the control of P ( I ^ I * ) , let
η ˜ = arg min η R W ( β * ) z ( η ) ,
where z ( η ) = 2 n i = 1 n j I * η j h j ( X i ) + j I * η j h j 2 + j I * ( 4 2 v ( δ / 2 ) L j + 2 c B ) | η j | + c j I * η j 2 . Consider the following random event
k I * 1 n i = 1 n h k ( X i ) + j I * η ˜ j < h j , h k > 2 2 v ( δ / 2 ) L k k I * 1 n i = 1 n h k ( X i ) + j I * η ˜ j < h j , h k > 2 2 v ( δ / 2 W ) L : = Ψ .
Let η ¯ R W be a vector corresponding to the component of the index set I * having η ˜ given by Equation (A12), and the component at other corresponding positions is 0. By Lemma 1, we know that η ¯ R W is a solution of Equation (3) on the event Ψ . It is recalled that β ^ R W , which is also a solution of Equation (3). Through the definition of the indicator set I ^ , we have β ^ k 0 for k I ^ . By construction, we obtain η ˜ k 0 for some subset T I * . The KKT conditions indicate that any two solutions have non-zero components at the same positions. Therefore, I ^ = T I * on the event Ψ . Further, we can write
P ( I ^ I * ) P ( Ψ c ) = P k I * 1 n i = 1 n h k ( X i ) + j I * η ˜ j < h j , h k > 2 2 v ( δ / 2 W ) L k I * P 1 n i = 1 n h k ( X i ) + j I * η ˜ j < h j , h k > 2 2 v ( δ / 2 W ) L = k I * P 1 n i = 1 n h k ( X i ) + E [ h k ( Z 1 ) ] E [ h k ( Z 1 ) ] + j I * η ˜ j < h j , h k > 2 2 v ( δ / 2 W ) L = k I * P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] j I * ( η ˜ j β j * ) < h j , h k > 2 2 v ( δ / 2 W ) L (A14) k I * P 1 n i = 1 n h k ( X i ) E [ h k ( Z 1 ) ] 2 v ( δ / 2 W ) L (A15) + k I * P j I * | η ˜ j β j * | < h j , h k > 2 v ( δ / 2 W ) L .
According to the previously proven Formula (A11), we find
k I * P 1 n i = 1 n h k ( Z i ) E h k ( Z 1 ) 2 v ( δ / 2 W ) L k I * P 1 n i = 1 n h k ( X i ) E h k ( X 1 ) 2 v ( δ / 2 W ) L ϵ k = k = 1 W P 1 n i = 1 n h k ( X i ) E h k ( X 1 ) 2 v ( δ / 2 W ) L ( 1 ϵ k * ) 2 W ( δ 2 W 2 ) ( 1 ϵ k * ) 2 .
For the upper bound of Equation (A15), observe Theorem 2, we can use a larger v ( δ / 2 W ) instead of v ( δ / 2 ) . Consider the construction of η ˜ in Equation (A12), we obtain
P j I * | η ˜ j β j * | 288 2 G * v ( δ / 2 W ) L min λ W δ W .
Similarly, we have
k I * P j I * | η ˜ j β j * | < h j , h k > 2 v ( δ / 2 W ) L k = 1 W P j I * | η ˜ j β j * | < h j , h k > h j h k h j h k 2 v ( δ / 2 W ) L k = 1 W P j I * | η ˜ j β j * | ρ * ( β * ) 2 v ( δ / 2 W ) L = k = 1 W P j I * | η ˜ j β j * | 2 v ( δ / 2 W ) L ρ * ( β * ) ( using Condition ( A ) ) k = 1 W P j I * | η ˜ j β j * | 288 2 G * v ( δ / 2 W ) L min λ W k = 1 W δ W = δ .
Combining all the bounds above, we can obtain
P ( I ^ I * ) P ( I * I ^ ) + P ( I ^ I * ) 2 W δ 2 W 2 ( 1 ϵ k * ) 2 + δ + 2 W δ 2 W 2 ( 1 ϵ k * ) 2 + δ = 4 W δ 2 W 2 ( 1 ϵ k * ) 2 + 2 δ .
This completes the proof of Theorem 3.

References

  1. McLachlan, G.J.; Lee, S.X.; Rathnayake, S.I. Finite mixture models. Ann. Rev. Stat. Appl. 2019, 6, 355–378. [Google Scholar] [CrossRef]
  2. Balakrishnan, S.; Wainwright, M.J.; Yu, B. Statistical guarantees for the EM algorithm: From population to sample-based analysis. Ann. Stat. 2017, 45, 77–120. [Google Scholar] [CrossRef]
  3. Wu, Y.; Zhou, H.H. Randomly initialized EM algorithm for two-component Gaussian mixture achieves near optimality in O( n ) iterations. arXiv 2019, arXiv:1908.10935. [Google Scholar]
  4. Chen, J.; Khalili, A. Order selection in finite mixture models with a nonsmooth penalty. J. Am. Stat. Assoc. 2008, 103, 1674–1683. [Google Scholar] [CrossRef] [Green Version]
  5. DasGupta, A. Asymptotic Theory of Statistics and Probability; Springer: New York, NY, USA, 2008. [Google Scholar]
  6. Devroye, L.; Lugosi, G. Combinatorial Methods in Density Estimation; Springer: New York, NY, USA, 2001. [Google Scholar]
  7. Biau, G.; Devroye, L. Density estimation by the penalized combinatorial method. J. Multivar. Anal. 2005, 94, 196–208. [Google Scholar] [CrossRef] [Green Version]
  8. Martin, R. Fast Nonparametric Estimation of a Mixing Distribution with Application to High Dimensional Inference. Ph.D. Thesis, Purdue University, West Lafayette, IN, USA, 2009. [Google Scholar]
  9. Bunea, F.; Tsybakov, A.B.; Wegkamp, M.H.; Barbu, A. Spades and mixture models. Ann. Stat. 2010, 38, 2525–2558. [Google Scholar] [CrossRef]
  10. Bertin, K.; Le Pennec, E.; Rivoirard, V. Adaptive Dantzig density estimation. Annales de l’IHP Probabilités et Statistiques 2011, 47, 43–74. [Google Scholar] [CrossRef]
  11. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Methodological 1996, 58, 267–288. [Google Scholar] [CrossRef]
  12. Hall, P.; Lahiri, S.N. Estimation of distributions, moments and quantiles in deconvolution problems. Ann. Stat. 2008, 36, 2110–2134. [Google Scholar] [CrossRef]
  13. Meister, A. Density estimation with normal measurement error with unknown variance. Stat. Sinica 2006, 16, 195–211. [Google Scholar]
  14. Cheng, C.L.; van Ness, J.W. Statistical Regression with Measurement Error; Wiley: New York, NY, USA, 1999. [Google Scholar]
  15. Zhu, H.; Zhang, R.; Zhu, G. Estimation and Inference in Semi-Functional Partially Linear Measurement Error Models. J. Syst. Sci. Complex. 2020, 33, 1179–1199. [Google Scholar] [CrossRef]
  16. Zhu, H.; Zhang, R.; Yu, Z.; Lian, H.; Liu, Y. Estimation and testing for partially functional linear errors-in-variables models. J. Multivar. Anal. 2019, 170, 296–314. [Google Scholar] [CrossRef]
  17. Bonhomme, S. Penalized Least Squares Methods for Latent Variables Models. In Advances in Economics and Econometrics: Volume 3, Econometrics: Tenth World Congress; Cambridge University Press: Cambridge, UK, 2013; Volume 51, p. 338. [Google Scholar]
  18. Nakamura, T. Corrected score function for errors-in-variables models: Methodology and application to generalized linear models. Biometrika 1990, 77, 127–137. [Google Scholar] [CrossRef]
  19. Buonaccorsi, J.P. Measurement error. In Models, Methods, and Applications; Chapman & Hall/CRC: Boca Raton, FL, USA, 2010. [Google Scholar]
  20. Carroll, R.J.; Ruppert, D.; Stefanski, L.A.; Crainiceanu, C.M. Measurement error in nonlinear models. In A Modern Perspective, 2nd ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2006. [Google Scholar]
  21. Zou, H.; Zhang, H. On the adaptive elastic-net with a diverging number of parameters. Ann. Stat. 2009, 37, 1733–1751. [Google Scholar] [CrossRef] [Green Version]
  22. Aitchison, J.; Aitken, C.G. Multivariate binary discrimination by the kernel method. Biometrika 1976, 63, 413–420. [Google Scholar] [CrossRef]
  23. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  24. Rosenbaum, M.; Tsybakov, A.B. Sparse recovery under matrix uncertainty. Ann. Stat. 2010, 38, 2620–2651. [Google Scholar] [CrossRef]
  25. Zhang, H.; Jia, J. Elastic-net regularized high-dimensional negative binomial regression: Consistency and weak signals detection. Stat. Sinica 2022, 32. [Google Scholar] [CrossRef]
  26. Buhlmann, P.; van de Geer, S. Statistics for High-Dimensional Data: Methods, Theory and Applications; Springer: New York, NY, USA, 2011. [Google Scholar]
  27. Belloni, A.; Rosenbaum, M.; Tsybakov, A.B. Linear and conic programming estimators in high dimensional errors-in-variables models. J. R. Stat. Soc. Series B Stat. Methodol. 2017, 79, 939–956. [Google Scholar] [CrossRef] [Green Version]
  28. Huang, H.; Gao, Y.; Zhang, H.; Li, B. Weighted Lasso estimates for sparse logistic regression: Non-asymptotic properties with measurement errors. Acta Math. Sci. 2021, 41, 207–230. [Google Scholar] [CrossRef]
  29. Zhang, H.; Chen, S.X. Concentration Inequalities for Statistical Inference. Commun. Math. Res. 2021, 37, 1–85. [Google Scholar]
  30. Donoho, D.L.; Johnstone, J.M. Ideal spatial adaptation by wavelet shrinkage. Biometrika 1994, 81, 425–455. [Google Scholar] [CrossRef]
  31. Deng, H.; Chen, J.; Song, B.; Pan, Z. Error bound of mode-based additive models. Entropy 2021, 23, 651. [Google Scholar] [CrossRef]
  32. Bickel, P.J.; Ritov, Y.A.; Tsybakov, A.B. Simultaneous analysis of Lasso and Dantzig selector. Ann. Stat. 2009, 37, 1705–1732. [Google Scholar] [CrossRef]
  33. Bunea, F. Honest variable selection in linear and logistic regression models via 1 and 1 + 2 penalization. Electron. J. Stat. 2008, 2, 1153–1194. [Google Scholar] [CrossRef]
  34. Chow, Y.S.; Teicher, H. Probability Theory: Independence, Interchangeability, Martingales, 3rd ed.; Springer: New York, NY, USA, 2003. [Google Scholar]
  35. Hersbach, H.; de Rosnay, P.; Bell, B.; Schepers, D.; Simmons, A.; Soci, C.; Abdalla, S.; Alonso-Balmaseda, M.; Balsamo, G.; Bechtold, P.; et al. Operational Global Reanalysis: Progress, Future Directions and Synergies with NWP; European Centre for Medium Range Weather Forecasts: Reading, UK, 2018. [Google Scholar]
  36. Fisher, N.I. Statistical Analysis of Circular Data; Cambridge University Press: Cambridge, UK, 1995. [Google Scholar]
  37. Broniatowski, M.; Jureckova, J.; Kalina, J. Likelihood ratio testing under measurement errors. Entropy 2018, 20, 966. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The simulation result in Section 4.2. The estimated support of β * by the four types of estimators, and the W is varying. The circles represent the means of the estimators under the four specific approaches, while the half vertical lines mean the standard deviations.
Figure 1. The simulation result in Section 4.2. The estimated support of β * by the four types of estimators, and the W is varying. The circles represent the means of the estimators under the four specific approaches, while the half vertical lines mean the standard deviations.
Entropy 24 00030 g001
Figure 2. The simulation results in Section 4.2. The density map of the four estimators. The result of Elastic-net in W = 321 is not be shown due to its poor performance.
Figure 2. The simulation results in Section 4.2. The density map of the four estimators. The result of Elastic-net in W = 321 is not be shown due to its poor performance.
Entropy 24 00030 g002
Figure 3. The sample histogram of the azimuth in Beijing Nongzhanguan at 6am and Qingdao Coast at 12 am.
Figure 3. The sample histogram of the azimuth in Beijing Nongzhanguan at 6am and Qingdao Coast at 12 am.
Entropy 24 00030 g003
Figure 4. The density map of the four estimators’ approaches for the three random sub-samples from the real-world data in Section 4.5.
Figure 4. The density map of the four estimators’ approaches for the three random sub-samples from the real-world data in Section 4.5.
Entropy 24 00030 g004
Table 1. The simulation results in Section 4.2. The mean and standard deviation of the errors in the four estimators of β * under N = 100 simulations, with n = 100 . The quasi-optimal λ 2 is c = 0.002 for Elastic-net, while c = 0.027 is for the CSDE.
Table 1. The simulation results in Section 4.2. The mean and standard deviation of the errors in the four estimators of β * under N = 100 simulations, with n = 100 . The quasi-optimal λ 2 is c = 0.002 for Elastic-net, while c = 0.027 is for the CSDE.
W λ 1 L 1 Error TV Error
Lasso810.0652.133 (2.467)1.137 (1.115)
Elastic-net2.061 (1.439)1.114 (0.805)
SPADES0.0531.922 (2.211)1.258 (1.296)
CSDE2.191 (4.812)1.405 (2.329)
Lasso1310.0682.032 (0.985)1.352 (0.712)
Elastic-net2.236 (2.498)1.409 (1.056)
SPADES0.0561.880 (2.644)0.972 (1.204)
CSDE1.635 (0.342)0.863 (0.402)
Lasso2110.0712.572 (4.187)1.605 (2.702)
Elastic-net2.061 (1.883)1.353 (1.516)
SPADES0.0581.764 (1.041)0.832 (0.610)
CSDE1.648 (0.168)0.791 (0.415)
Lasso3210.0742.120 (2.842)1.146 (1.115)
Elastic-net10.173 (82.753)7.839 (67.887)
SPADES0.0612.106 (4.816)0.818 (1.565)
CSDE1.623 (0.085)0.634 (0.199)
Table 2. The simulation result in Section 4.3. The mean and standard deviation of the errors in the four estimators of β * under N = 100 simulations. The λ 2 is chosen as c = 0.005 for Elastic-net, while c = 0.203 for the CSDE.
Table 2. The simulation result in Section 4.3. The mean and standard deviation of the errors in the four estimators of β * under N = 100 simulations. The λ 2 is chosen as c = 0.005 for Elastic-net, while c = 0.203 for the CSDE.
W λ 1 L 1 Error TV Error
Lasso810.0481.796 (0.006)0.002 (0.001)
Elastic-net1.796 (0.006)0.002 (0.001)
SPADES0.1381.811 (0.013)0.002 (0.005)
CSDE1.806 (0.008)0.003 (0.005)
Lasso1310.0511.828 (0.006)0.003 (0.001)
Elastic-net1.830 (0.009)0.004 (0.002)
SPADES0.1451.880 (0.006)0.002 (0.005)
CSDE1.854 (0.006)0.002 (0.004)
Lasso2110.0531.935 (0.010)0.005 (0.003)
Elastic-net2.061 (0.014)0.007 (0.008)
SPADES0.1521.935 (0.008)0.005 (0.003)
CSDE1.861 (0.005)0.003 (0.002)
Lasso3210.0551.927 (0.031)0.005 (0.002)
Elastic-net2.123 (0.026)0.009 (0.009)
SPADES0.1581.938 (0.008)0.005 (0.003)
CSDE1.852 (0.002)0.002 (0.001)
Table 3. The low-dimensional simulation result in Section 4.4.
Table 3. The low-dimensional simulation result in Section 4.4.
L 1 Error TV Error
Scenario 1EM0.255 (0.122)0.205 (0.098)
CSDE0.206 (0.145)0.185 (0.104)
Scenario 2EM0.111 (0.055)0.111 (0.055)
CSDE0.109 (0.037)0.108 (0.037)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, X.; Zhang, H.; Wei, H.; Zhang, S. Sparse Density Estimation with Measurement Errors. Entropy 2022, 24, 30. https://doi.org/10.3390/e24010030

AMA Style

Yang X, Zhang H, Wei H, Zhang S. Sparse Density Estimation with Measurement Errors. Entropy. 2022; 24(1):30. https://doi.org/10.3390/e24010030

Chicago/Turabian Style

Yang, Xiaowei, Huiming Zhang, Haoyu Wei, and Shouzheng Zhang. 2022. "Sparse Density Estimation with Measurement Errors" Entropy 24, no. 1: 30. https://doi.org/10.3390/e24010030

APA Style

Yang, X., Zhang, H., Wei, H., & Zhang, S. (2022). Sparse Density Estimation with Measurement Errors. Entropy, 24(1), 30. https://doi.org/10.3390/e24010030

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop