Next Article in Journal
Effort and Cost Estimation Using Decision Tree Techniques and Story Points in Agile Software Development
Previous Article in Journal
Distributed Optimization Control for Heterogeneous Multiagent Systems under Directed Topologies
Previous Article in Special Issue
Monitoring of Linear Profiles Using Linear Mixed Model in the Presence of Measurement Errors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Restricted Distance-Type Gaussian Estimators Based on Density Power Divergence and Their Applications in Hypothesis Testing

Department of Statistics and Operational Research, Complutense University of Madrid, 28040 Madrid, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(6), 1480; https://doi.org/10.3390/math11061480
Submission received: 15 February 2023 / Revised: 10 March 2023 / Accepted: 15 March 2023 / Published: 17 March 2023
(This article belongs to the Special Issue Advances in Statistical Analysis and Applications in Engineering)

Abstract

:
In this paper, we introduce the restricted minimum density power divergence Gaussian estimator (MDPDGE) and study its main asymptotic properties. In addition, we examine it robustness through its influence function analysis. Restricted estimators are required in many practical situations, such as testing composite null hypotheses, and we provide in this case constrained estimators to inherent restrictions of the underlying distribution. Furthermore, we derive robust Rao-type test statistics based on the MDPDGE for testing a simple null hypothesis, and we deduce explicit expressions for some main important distributions. Finally, we empirically evaluate the efficiency and robustness of the method through a simulation study.

1. Introduction

Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with probability density function f θ ( y ) , θ Θ R d , where θ is the vector of unknown parameters, Θ is the corresponding parameter space and d is the dimension of θ , d 1 . We denote,
E θ Y = μ θ and C o v θ Y = Σ θ .
The log-likelihood function of the assumed model is given by
l θ = i = 1 n log f θ ( y i )
for y 1 , , y n observations of the m-dimensional random vectors Y 1 , , Y n . Then, the maximum likelihood estimator (MLE) is computed as
θ ^ MLE = max θ Θ l θ .
In many real-life situations, the underlying density function, f θ ( · ) , is unknown or its computation is quite difficult but contrariwise, the mean vector and variance–covariance matrices of the underlying distribution of the data as a function of θ , namely μ ( θ ) and Σ θ , are known.
In this case, Zhang [1] proposed a general procedure based on the Gaussian distribution for estimating the model parameter vector θ . In [1], it is assumed that the m-dimensional random vector Y comes from a multidimensional normal distribution with mean vector μ θ and variance-covariance matrix Σ θ . From a statistical point of view, this procedure can be justified on the basis of the maximum-entropy principle (see [2]), as the multidimensional normal distribution has maximum uncertainty in terms of Shannon entropy and is as well consistent with the given information, i.e., vector mean and variance–covariance matrix.
Then, an estimator of the model parameter θ based on the Gaussian distribution can be obtained by maximizing the log-likelihood function as defined in (2) but using as f θ ( · ) the probability density function of a normal distribution with known mean vector μ ( θ ) and variance–covariance matrix Σ θ , corresponding to the true mean vector and variance–covariance matrix of the underlying distribution. That is, the Gaussian-based likelihood function of θ is given by
l G θ = n m 2 log 2 π n 2 log Σ θ 1 2 i = 1 n y i μ θ T Σ θ 1 y i μ θ
for any y 1 , , y n independent observations of the population Y , and the Gaussian MLE of θ is defined by
θ ^ G = arg max θ Θ l G θ .
The Gaussian estimator is an MLE and thus inherits all the good properties of the likelihood estimators. Consequently, it works well in terms of the asymptotic efficiency, but it has important robustness problems. That is, in the absence of contamination in data, the MLE consistently estimates the true value of the model parameter, but it may be quite heavily affected by outlying observations in the data. For this reason, Castilla and Zografos extended in [3] the concept of Gaussian estimator and defined a robust version of the estimator based on the density power divergence (DPD) introduced in Basu et al. in [4]. The DPD robustly quantifies the statistical difference between two distributions, and it has been widely used for developing robust inferential methods in many different statistical models. Given a set of observations, the robust minimum DPD estimator (MDPDE) is computed as the minimizer of the DPD between the assumed model distribution and the empirical distribution of the data. The MDPDE enjoys good asymptotic properties and produces robust estimators under general statistical models, as discussed later. The minimum density power divergence Gaussian estimator (MDPDGE) of the parameter θ is defined for τ 0 as
θ ^ G τ = arg max θ Θ R d H n τ θ ,
where
H n τ θ = τ + 1 τ 2 π m τ / 2 Σ θ τ / 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ τ 1 + τ ( m / 2 ) + 1 1 τ = a Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b 1 τ ,
and
a = τ + 1 τ 2 π m τ / 2 and b = τ 1 + τ ( m / 2 ) + 1 .
The MDPDGE family is indexed by a tuning parameter τ controlling the trade-off between robustness and efficiency; the greater the value of τ , the more robust the resulting estimator is, but the efficiency decreases. It has been shown in the literature that values of the tuning parameter above 1 do not provide sufficiently efficient estimators, and so, the tuning parameter would be chosen in the [ 0 , 1 ] interval. Furthermore, at τ = 0 , the MDPDGE reduces to the Gaussian estimator of [1],
θ ^ G = arg max θ Θ R d H n 0 θ
with
H n 0 θ = lim τ 0 H n τ θ = n 2 log Σ θ 1 2 i = 1 n y i μ θ T Σ θ 1 y i μ θ .
Note that the above objective function does not perfectly match with the likelihood function of the model stated in (2), as it lacks the first term of the likelihood. However, this term does not depend on the parameter θ , and thus, both loss functions will lead to the same maximizer. Indeed, the loss in Equation (7) corresponds to the Kullback–Leiber divergence between the assumed normal distribution and the empirical distribution of the data, which justifies the MLE from the point of view of information theory (see [5,6,7]).
Furthermore, the MDPDGE is consistent and asymptotically normal, that is, given Y 1 , , Y n independent and identically distributed vectors from the m-dimensional random vector Y , the MDPDGE, θ ^ G τ , defined in (4) satisfies
n θ ^ G τ θ L n N ( 0 d , J τ ( θ ) 1 K τ ( θ ) J τ ( θ ) 1 ) ,
being
J τ ( θ ) = J τ i j θ i , j = 1 , , d and K τ ( θ ) = K τ i j θ i , j = 1 , , d .
and the elements J τ i j θ and K τ i j θ of the matrices J τ θ and K τ θ are given by
J τ i j θ = 1 2 π m / 2 Σ θ 1 / 2 τ 1 1 + τ ( m / 2 ) + 2 × τ + 1 t r a c e Σ θ 1 μ θ θ i μ θ θ j T + Δ τ i Δ τ j + 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i
and
K τ i j θ = 1 2 π m / 2 Σ θ 1 / 2 2 τ 1 1 + 2 τ ( m / 2 ) + 2 Δ 2 τ i Δ 2 τ j + 1 + 2 τ t r a c e Σ θ 1 μ θ θ i μ θ θ j T + 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i 1 2 π m / 2 Σ θ 1 / 2 2 τ 1 1 + τ m + 2 Δ τ i Δ τ j ,
with Δ τ i = τ 2 t r a c e Σ θ 1 Σ θ θ i . By t r a c e ( A ) , we mean the trace of matrix A.
The above asymptotic distribution follows from Theorem 2 in [8], where matrices J τ θ and K τ θ are defined for general statistical models.
The asymptotic distribution of θ G ^ has been considered in many papers, see e.g., [9,10,11,12,13].
On the other hand, in some situations, we may have additional knowledge about the true parameter space. Then, these constraints should be included in the definition of the parameter space Θ . Here, we will consider restricted parameter spaces of the form
Θ 0 = θ Θ / g ( θ ) = 0 r ,
where g : R d R r is a vector-valued function mapping such that the d × r matrix
G θ = g T ( θ ) θ
exists and is continuous in θ and rank G θ = r , and 0 r denotes the null vector of dimension r. The notation Θ 0 clues the use of the present restricted estimator for defining test statistics under composite null hypothesis.
The most popular estimator of θ satisfying the constraints in (11) is the restricted MLE (RMLE), which is naturally defined as the maximizer of the log-likelihood function of the model but is subject to the parameter space restrictions g ( θ ) = 0 r (see [14,15,16,17]). Unfortunately, the RMLE has the same robustness problems as the MLE, and so robust alternatives should be adopted in the presence of contamination in data. Several robust restricted estimators have been considered in the statistical literature to overcome the robustness drawback of the RMLE. For example, Pardo et al. introduced in [18] the restricted minimum Phi-divergence estimator and studied its properties. In [8], the restricted minimum density power divergence estimators (RMDPDE) are presented, and some applications on the testing hypothesis are studied. In [19], the theoretical robustness properties of the RMDPDE were studied. In [20,21], the restricted Rényi pseudodistance estimator is considered, and robust Rao-type tests are derived from it and developed. More recently, in [22], the RMDPD under normal distributions is studied, and independence tests under the normal assumption are developed, and in [23] the RMDPDE is applied in the context of independent but not identically distributed variables under heterocedastic linear regression models. Other interesting papers related to multivariate analysis are [24,25,26,27].
In this paper, we introduce and study the restricted minimum density power divergence Gaussian estimator (RMDPDGE). The aim of this study is to introduce an estimator dealing with situations where Gaussian estimators are useful; there are additional constraints on the parameter space, and the estimator should be robust in terms of contamination. To show the robustness, we compute the corresponding influence function (IF), showing that in general, it is bounded. As an application of these restricted estimators, we develop Rao-type test statistics. The idea in this case is to consider the additional constraints as the constraints defining the null hypothesis of the test. Finally, we show the robustness via a simulation study, in which the good behavior of these estimators is shown comparing it with the behavior of classical restricted estimators based on MLE. We compare the loss in efficiency when comparing these estimators with more specific estimators dealing with the real distribution, showing that the loss is affordable.
The rest of the paper is organized as follows: In Section 2, we introduce the RMDPDGE and we obtain its asymptotic distribution. Section 3 presents the influence function of the RMDPDGE and theoretically proves the robustness of the proposed estimators. Some statistical applications for testing are presented in Section 4, and an explicit expression of the Rao-type test statistics based on the RMDPDGE under exponential, Poisson and Lindley models are given. Section 5 empirically demonstrates the robustness of the method through a simulation study, and the advantages and disadvantages of the Gaussian assumption are discussed there. Section 6 presents some conclusions. The proofs of the main results stated in the paper are included in Appendix A.

2. Restricted Minimum Density Power Divergence Gaussian Estimators

In this section, we present the RMDPDGE under general equality non-linear constraints and we study its asymptotic distribution, showing the consistency of the estimator.
Definition 1.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . The RMDPDGE, θ ˜ G τ , is defined by
θ ˜ G τ = arg max Θ 0 H n τ θ ,
where H n τ θ is as given in Equation (5) and Θ 0 = { θ Θ / g ( θ r ) = 0 r } is the restricted parameter space defined in (11).
Before presenting the asymptotic distribution of the MDPDGE, we present some previous results whose proofs are included in Appendix A.
Proposition 1.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Then,
n 1 τ + 1 H n τ θ θ L n N ( 0 d , K τ ( θ ) ) ,
where K τ ( θ ) was defined in (10).
Proof. 
See Appendix A.2. □
Proposition 2.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Then,
2 H n τ θ θ θ T P n τ + 1 J τ ( θ ) ,
where J τ ( θ ) was defined in (9).
Proof. 
See Appendix A.3. □
Next, we present the asymptotic distribution of θ ˜ G τ .
Theorem 1.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Suppose the true distribution of Y belongs to the model, and we consider θ Θ 0 . Then, the RMDPDGE θ ˜ G τ of θ obtained under the constraints g ( θ ) = 0 r satisfies
n 1 / 2 ( θ ˜ G τ θ ) L n N ( 0 d , M τ ( θ ) ) ,
where
M τ ( θ ) = P τ * ( θ ) K τ θ P τ * ( θ ) T ,
P τ * ( θ ) = J τ ( θ ) 1 Q τ ( θ ) G ( θ ) T J τ ( θ ) 1 ,
Q τ ( θ ) = J τ 1 ( θ ) G ( θ ) G ( θ ) T J τ ( θ ) 1 G ( θ ) 1 ,
and J τ ( θ ) and K τ θ were defined in (9) and (10), respectively.
Proof. 
The estimating equations for the RMDPDGE are given by
θ H n τ ( θ ) + G ( θ ) λ n = 0 d , g ( θ ˜ G τ ) = 0 r ,
where λ n is a vector of Lagrangian multipliers. Now, we consider θ n = θ + m n 1 / 2 , where | | m | | < k , for 0 < k < . We have,
θ H n τ ( θ ) θ = θ n = θ H n τ ( θ ) + 2 θ T θ H n τ ( θ ) θ = θ * ( θ n * θ )
and
n 1 / 2 θ H n τ ( θ ) θ = θ n = n 1 / 2 θ H n τ ( θ ) + 2 θ T θ H n τ ( θ ) θ = θ * n 1 / 2 ( θ n θ ) ,
where θ * belongs to the segment joining θ and θ n . Since
lim n 2 θ T θ H n τ ( θ ) = τ + 1 J τ ( θ ) ,
we obtain
n 1 / 2 θ H n τ ( θ ) θ = θ n = n 1 / 2 θ H n τ ( θ ) τ + 1 n 1 / 2 J τ ( θ ) ( θ n θ ) + o p ( 1 ) .
Taking into account that G ( θ ) is continuous in θ
n 1 / 2 g ( θ n ) = G ( θ ) T n 1 / 2 ( θ n θ ) + o p ( 1 ) .
The RMDPDGE θ ˜ G τ must satisfy the conditions in (16), and in view of (18), we have
n 1 / 2 θ H n τ ( θ ) τ + 1 J τ ( θ ) n 1 / 2 ( θ ˜ G τ θ ) + G ( θ ) n 1 / 2 λ n + o p ( 1 ) = 0 p .
From (19), it follows that
G T ( θ ) n 1 / 2 ( θ ˜ G τ θ ) + o p ( 1 ) = 0 r .
Now, we can express Equations (20) and (21) in matrix form as
τ + 1 J τ ( θ ) G ( θ ) G T ( θ ) 0 n 1 / 2 ( θ ˜ G τ θ ) n 1 / 2 λ n = n 1 / 2 θ H n τ ( θ ) 0 + o p ( 1 ) .
Therefore,
n 1 / 2 ( θ ˜ G τ θ ) n 1 / 2 λ n = τ + 1 J τ ( θ ) G ( θ ) G T ( θ ) 0 1 n 1 / 2 θ H n τ ( θ ) 0 r + o p ( 1 ) .
However,
τ + 1 J τ ( θ ) G ( θ ) G T ( θ ) 0 1 = L τ * ( θ ) Q τ ( θ ) Q τ ( θ 0 ) T R τ ( θ ) ,
where
L τ * ( θ ) = 1 τ + 1 J τ ( θ ) 1 Q τ ( θ ) G ( θ ) T J τ ( θ ) 1 = 1 τ + 1 P τ * ( θ ) , Q τ ( θ ) = J τ 1 ( θ ) G ( θ ) G ( θ ) T J τ ( θ ) 1 G ( θ ) 1 , R τ ( θ ) = G ( θ ) T J τ ( θ ) 1 G ( θ ) ,
and P τ * ( θ 0 ) and Q τ ( θ 0 ) are as given in (14) and (15), respectively. Then,
n 1 / 2 ( θ ˜ G τ θ ) = τ + 1 1 P τ * ( θ ) n 1 / 2 θ H n τ ( θ ) + o p ( 1 ) ,
and we know by Proposition 1 that
n 1 / 2 τ + 1 1 θ H n τ ( θ ) L n N 0 , K τ θ .
Now, by (22) and (23), we have the desired result. □
Remark 1.
Notice that the result in (8) is a special case of the previous theorem when there are no restrictions on the parameter space, in the sense that G defined in (12) is the null matrix. In this case, matrix P τ * ( θ ) given in (14) becomes P τ * ( θ ) = J τ ( θ ) 1 . Therefore, the asymptotic variance–covariance matrix of the unrestricted estimator, i.e., the MDPDGE, may be reconstructed from the previous theorem.
In order to compute the MDPDGE, we note that it is an optimum of a differentiable function H n τ , so it must annul its first derivatives. We will use that
Σ θ τ / 2 θ = τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ ,
and
θ y i μ θ T Σ θ 1 y i μ θ = 2 μ θ θ T Σ θ 1 y i μ θ y i μ θ T × Σ θ 1 Σ θ θ Σ θ 1 y i μ θ .
Now, taking derivatives in (5), for a certain fixed τ , we have that
θ H n τ ( θ ) = 1 n i = 1 n a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ × exp τ 2 y i μ θ T Σ θ 1 y i μ θ + b a τ 2 Σ θ τ / 2 × t r a c e Σ θ 1 Σ θ θ + a τ 2 Σ θ τ / 2 × exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ Σ θ 1 y i μ θ = 1 n i = 1 n Ψ τ ( y i ; θ ) ,
with
Ψ τ ( y i ; θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ + 2 μ θ θ T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ Σ θ 1 y i μ θ × exp τ 2 y i μ θ T Σ θ 1 y i μ θ + b t r a c e Σ θ 1 Σ θ θ .
Therefore, the estimating equations of the MDPDGE for a fixed parameter τ are given by
i = 1 n Ψ τ ( y i ; θ ) = 0 d .
The previous estimating equations characterize the MDPDGE as an M-estimator and so its asymptotic distribution could have been also derived from the general theory of M-estimators. In particular, the MDPDGE, θ ^ G τ , satisfies for any τ 0
n θ ^ G τ θ L n N 0 d , S 1 MS 1 ,
with
S = E 2 H n τ θ θ θ T and M = C o v n θ H n τ ( θ ) .
Based on Propositions 1 and 2, we can express the previous matrices as
S = τ + 1 J τ θ and M = τ + 1 2 K τ θ ,
and we obtain the expressions established in (8). The asymptotic convergence in (26) offers an alternative proof of the asymptotic distribution of MDPDGE developed in [3] in terms of the transformed matrices S and M in Equation (27).

3. Influence Function for the RMDPDGE

To analyze the robustness of an estimator, Hampel et al. introduced in [28] the concept of an Influence Function (IF). Since then, the IF has been widely used in the statistical literature to measure robustness in different statistical contexts. Intuitively, the IF describes the effect of an infinitesimal contamination of the model on the estimation. Robust estimators should be less affected by contamination, and thus, IFs associated to locally robust (B-robust) estimators should be bounded.
The IF of an estimator, θ ˜ G τ , is defined in terms of its statistical functional T ˜ τ satisfying T ˜ τ ( g ) = θ ˜ G τ , where g is the true density function underlying the data. Given the density function g, we define its contaminated version at the point perturbation y 0 as,
g = ( 1 ε ) g + ε Δ y 0 ,
where ε is fraction of contamination and Δ y 0 denotes the indicator function at y 0 . Then, the IF of θ ˜ G τ is defined as the derivative of the functional at ε = 0
I F ( y 0 , T ˜ τ ) = T ˜ τ ( g ε ) ε | ε = 0 .
Hence, the above derivative quantifies the rate of change of the sample estimator when contamination occurs.
Let us now obtain the IF of RMDPDE. We consider the contaminated model
g ε ( y ) = ( 1 ε ) f θ ( y ) + ε Δ y 0 ,
where f θ is the assumed probability density function of a normal population. The MDPDGE for the contaminated model is then given by θ ˜ G , ε τ = T ˜ τ ( g ε ) .
By definition, θ ˜ G , ε τ is the maximizer of H n τ ( θ ) in (5), subject to the constraints g ( θ ˜ G , ε τ ) = 0 . Using the characterization of the MDPDGE as an M-estimator, we have that the influence function of the MDPDGE is given by
I F ( y , T ˜ τ , θ ) = J τ ( θ ) 1 Ψ τ ( y ; θ ) ,
where J τ ( θ ) was defined in (9) and Ψ τ ( y ; θ ) in (24). The influence function of the RMDPDGE will be obtained with the additional condition g ( θ ˜ G , ε τ ) = 0 . Differentiating this last equation gives, at ε = 0 ,
G θ T I F ( y , T ˜ τ , θ ) = 0 .
Based on (29) and (30), we have
J τ ( θ ) G θ T I F ( y , T ˜ τ , θ ) = Ψ τ ( y ; θ ) 0 .
Therefore,
J τ ( θ ) T , G θ J τ ( θ ) G θ T I F ( y , T ˜ τ , θ ) = J τ ( θ ) T Ψ τ ( y ; θ ) ,
and the influence function of the RMDPDGE, θ ˜ G τ , is given by
J τ ( θ ) T J τ ( θ ) ) + G θ G θ T 1 J τ ( θ ) T Ψ τ ( y ; θ ) .
We can observe that the influence function of θ ˜ G τ , obtained in (31), will be bounded if the influence function of the MDPDGE, θ ^ G τ , given in (29) is bounded. In general, it is not easy to see if it is bounded or not, but in particular situations, this can be solved. On the other hand, if there are no restrictions, G θ = 0 , and therefore, (31) coincides with (29).
In Section 4.1, we shall present the expression of J τ ( θ ) and Ψ τ ( y , θ ) for some models. Based on that explicit expressions, Figure 1 presents the influence function of the MDPDGE, θ ^ G τ , for the exponential model with θ = 4 and τ = 0 , 0.2 and 0.8 . At τ = 0 , the influence function of θ ^ G τ is not bounded, whereas it is bounded at the positive values of the tuning parameter, τ = 0.2 and 0.8 . This fact illustrates the robustness of the MDPDGE for τ > 0 for the exponential model.

4. Rao-Type Tests Based on RMDPDGE

Recently, many robust test statistics based on minimum distance estimators have been introduced in the statistical literature for testing under different statistical models. Among them, density power divergence and Rényi’s pseudodistance-based test statistics have shown very competitive performance with respect to classical tests in many different problems. Distance-based test statistics are essentially of two types: Wald-type tests and Rao-type tests. Some applications of these tests can be seen at [8,20,21,22,23,29,30,31,32,33,34,35,36,37] and references therein. In this section, we introduce the Rao-type tests based on RMDPDGE, and we study their asymptotic properties, proving the consistency of the tests.
We analyze here a simple null hypothesis of the form
H 0 : θ = θ 0 vs. H 1 : θ θ 0 .
Definition 2.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d , and consider the testing problem defined in (32). The Rao-type test statistic based on RMDPDGE is defined by
R τ θ 0 = 1 n U n τ ( θ 0 ) T K τ θ 0 1 U n τ ( θ 0 ) ,
where
U n τ ( θ ) = 1 τ + 1 i = 1 n Ψ τ 1 ( y i ; θ ) , , 1 τ + 1 i = 1 n Ψ τ d ( y i ; θ ) T
is the score function defining the estimating equations of the MDPDGE and
Ψ τ ( y i ; θ ) = Ψ τ 1 ( y i ; θ ) , . , Ψ τ d ( y i ; θ ) .
The next result establishes the asymptotic behavior of the proposed Rao-type test statistic.
Theorem 2.
Let Y 1 , , Y n be independent and identically distributed observations from an m- dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Under the null hypothesis given in (32), it holds
R τ θ 0 L n χ d 2 .
Proof. 
First, note that we can rewrite
1 n i = 1 n 1 τ + 1 Ψ τ ( y i ; θ ) = n 1 τ + 1 θ H n τ ( θ ) ,
and hence, by Proposition 1, we can establish the asymptotic distribution of the τ -score function U n τ ( θ )
1 n i = 1 n 1 τ + 1 Ψ τ ( y i ; θ ) L n N 0 p , K τ θ .
Hence, as the τ -score function is asymptotically normal,
1 n U n τ ( θ ) = 1 n i = 1 n 1 τ + 1 Ψ τ ( y i ; θ ) = n 1 τ + 1 θ H n τ ( θ ) L n N 0 , K τ θ .
Then, applying a suitable transformation, the result follows. □
Remark 2.
The Rao-type statistic relies on the τ-score function U n τ ( θ ) defining the estimating equations U n τ ( θ ) = 0 . Therefore, if the simple null hypothesis holds, the τ-score function vanishes and conversely, if the true parameter is far from the null hypothesis, large τ-scores will be produced. Based on Theorem 2, for large enough sample sizes, one can use the 100 ( 1 α ) percentile of the chi-square distribution with d degrees of freedom χ d , α 2 satisfying
Pr χ d 2 > χ d , α 2 = α ,
to define the reject region of the test with null hypothesis in (32) as
R C = { R τ θ 0 > χ d , α 2 } .
For illustrative purposes, we present here the application of the proposed method in elliptical distributions.
Example 1. (Elliptical distributions). The m-dimensional random vector Y follows an elliptical distribution if its characteristic function has the form
φ Y t = exp i t 2 μ ψ 1 2 t 2 Σ t ,
where μ is an m-dimensional vector, Σ is a positive definite matrix and ψ ( t ) denotes the so-called characteristic generator function. The function ψ may depend on the dimension of random vector Y . In general, it does not hold that Y has a joint density function, f Y ( y ) , but if this density exists, it is given by
f Y ( y ) = c m Σ 1 2 g m 1 2 y μ T Σ 1 y μ
for some density generator function g m which could depend on the dimension of the random vector. Moreover, if the density exists, the parameter c m is given explicitly by
c m = 2 π m 2 Γ m 2 x m 2 1 g m x d x 1 .
The elliptical distribution family is in the following denoted by E m μ , Σ , g m . For more details about the elliptical family E m μ , Σ , g m , see [38,39,40,41,42] and references therein. In [40], for instance, it can be seen that the mean vector and variance–covariance matrix can be obtained as
E Y = μ and C o v Y = c Y Σ ,
where c Y = 2 ψ ( 0 ) .
For the elliptical model, the parameter to be estimated is θ = μ T , Σ whose dimension is s = m + m m + 1 2 . In the following, we denote μ ( θ ) instead of μ and Σ θ instead of Σ, in order to be consistent with the paper notation.
Let us consider the testing problem
H 0 : μ ( θ ) , Σ θ = μ 0 , Σ 0 vs. H 1 : μ ( θ ) , Σ θ μ 0 , Σ 0 ,
where μ 0 and Σ 0 are known. The Rao-type test statistic based on the MDPDGE for the elliptical model is given as
R τ μ 0 , Σ 0 = 1 n U n τ ( μ 0 , Σ 0 ) T K τ μ 0 , Σ 0 1 U n τ ( μ 0 , Σ 0 ) ,
where
U n τ ( μ 0 , Σ 0 ) = i = 1 n 1 τ + 1 Ψ τ ( y i ; μ 0 , Σ 0 ) ,
with Ψ τ ( y i ; μ 0 , Σ 0 ) and K τ μ 0 , Σ 0 as defined in (24) and (10), respectively, but replacing Σ θ by c Y Σ and μ θ by μ. Then, the null hypothesis in (34) should be rejected if
R τ μ 0 , Σ 0 > χ m + m ( m + 1 ) 2 , α 2 ,
with χ m + m ( m + 1 ) 2 , α 2 the ( 1 α ) upper quantile of a chi-square with m + m ( m + 1 ) 2 degrees of freedom.
We finally prove the consistency of the Rao-type test based on RMDPDGE. To simplify the statement of the next result, we first define the vector
Y τ ( θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ exp τ 2 Y μ θ T Σ θ 1 Y μ θ + b t r a c e Σ θ 1 Σ θ θ + exp τ 2 Y μ θ T Σ θ 1 Y μ θ 2 μ θ θ T Σ θ 1 Y μ θ + Y μ θ T Σ θ 1 Σ θ θ Σ θ 1 Y μ θ ,
where a and b were defined in (6). We can observe that θ H n ( θ ) is the sample mean of a random sample of size n from the m-dimensional population Y τ ( θ ) .
Theorem 3.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Let θ Θ with θ θ 0 , with θ 0 defined in (32), and let us assume that E θ Y τ ( θ 0 ) 0 d . Then,
lim n P θ R τ θ 0 > χ d , α 2 = 1 .
Proof. 
From the previous results, it holds that
1 n U n τ ( θ 0 ) = 1 n i = 1 n 1 τ + 1 Ψ τ ( Y i ; θ 0 ) = 1 τ + 1 θ H n τ ( θ 0 ) P n 1 τ + 1 E θ Y τ ( θ 0 ) ,
where Y τ ( θ 0 ) is as defined in (35). Therefore,
P θ R τ θ 0 > χ d , α 2 = P θ 1 n R τ θ 0 > 1 n χ d , α 2 n I 1 τ + 1 2 E θ Y τ ( θ 0 ) K τ 1 θ E θ T Y τ ( θ 0 ) > 0 = 1 ,
where I ( · ) is the indicator function. □
A natural question that arises here is how the asymptotic power of different test statistics considered for testing the hypothesis in (34) could be compared. Lehmann [43] stated that contiguous alternative hypotheses are of great interest for such purposes, as their associated power functions do not converge to 1. In this regard, we next derive the asymptotic distribution of R τ θ 0 under local Pitman-type alternative hypotheses of the form
H 1 , n : θ = θ n : = θ 0 + n 1 / 2 l ,
where l is a d-dimensional normal vector. The next result determines the asymptotic power of the Rao-type test based on RMDPDGE under a contiguous alternative hypothesis.
Theorem 4.
Let Y 1 , , Y n be independent and identically distributed observations from an m-dimensional random vector Y with E θ Y = μ θ and C o v θ Y = Σ θ , θ Θ R d . Under the contiguous alternative hypothesis of the form
H 1 , n : θ n = θ 0 + n 1 / 2 l ,
the asymptotic distribution of the Rao-type test based on RMDPDGE, R τ θ 0 , is a non-central chi-square distribution with d degrees of freedom and non-centrality parameter given by
δ τ ( θ 0 , l ) = l T J τ θ 0 K τ 1 θ 0 J τ θ 0 l .
Proof. 
Consider the Taylor series expansion
1 n U n τ ( θ n ) = 1 n U n τ ( θ 0 ) + 1 n U n τ ( θ ) θ T θ = θ n * l ,
where θ n * belongs to the line segment joining θ 0 and θ 0 + 1 n l . Now, by Proposition 2
1 n U n τ ( θ ) θ T = 1 τ + 1 2 H n τ θ θ θ T P n J τ ( θ ) .
Therefore,
1 n U n τ ( θ ) θ = θ 0 + n 1 / 2 d L n N J τ θ 0 l , K τ θ 0 ,
and
R τ θ 0 L n χ p 2 δ τ ( θ 0 , l ) ,
with δ τ ( θ 0 , d ) given by
δ τ ( θ 0 , l ) = l T J τ θ 0 K τ 1 θ 0 J τ θ 0 l .
Hence, the result holds. □
Remark 3.
The previous result can be used for defining an approximation to the power function under any alternative hypothesis, θ Θ \ Θ 0 , given as
θ = θ θ 0 + θ 0 = n 1 n θ θ 0 + θ 0 = θ 0 + n 1 / 2 l ,
with l = n θ θ 0 .
Remark 4.
In this section, we have dealt with a Rao-type test for a simple null hypothesis. This family can be extended to a composite null hypothesis. If we are interested in testing H 0 : θ Θ 0 = θ Θ / g ( θ ) = 0 r , we can consider the family of Rao-type tests given by
R τ θ ˜ G τ = 1 n U n τ ( θ ˜ G τ ) T Q τ ( θ ˜ G τ ) Q τ ( θ ˜ G τ ) K τ θ ˜ G τ Q τ ( θ ˜ G τ ) 1 Q τ ( θ ˜ G τ ) T U n τ ( θ ˜ G τ ) .
However, the extension of the presented results for the family of robust test statistics defined in (36) is not trivial, and it will be established in future research.
In particular, the simple null hypothesis in (32) can be written as a composite null hypothesis with g ( θ ) = θ θ 0 . In this case, G ( θ ) reduces to the identity matrix of dimension p and the restricted estimator θ ˜ G τ coincides with θ 0 . In this case, the Rao-type test statistic R τ θ ˜ G τ coincides with the proposed R τ θ 0 given in (33). Rao-type test statistics based on RMDPDE have been developed in [22].

4.1. Rao-Type Tests Based on MDPDGE for Univariate Distributions

Let Y 1 , , Y n be a random sample from the population Y , with
E Y = μ θ and V a r Y = σ 2 θ .
Based on (25), the estimating equation is given by
i = 1 n Ψ τ y i , θ = 0 ,
with
Ψ τ y i , θ = τ + 1 σ 2 θ τ / 2 2 2 π τ / 2 log σ 2 θ θ + log σ 2 θ θ y i μ θ σ 2 θ 2 + 2 μ θ θ y i μ θ 1 σ 2 θ × exp τ 2 σ 2 θ y i μ θ 2 + τ 1 + τ 3 / 2 log σ 2 θ θ .
Moreover, the expressions of J τ θ and K τ θ are, respectively, given by
J τ θ = 1 2 π σ θ 2 τ 2 1 1 + τ 5 / 2 τ + 1 σ 2 θ μ θ θ 2 + τ 2 4 log σ 2 θ θ 2 + 1 2 log σ 2 θ θ 2
and
K τ θ = 1 2 π 1 / 2 σ θ 2 τ 1 1 + 2 τ 5 / 2 τ 2 log σ 2 θ θ 2 + 1 + 2 τ σ 2 θ μ θ θ 2 + 1 2 log σ 2 θ θ 2 τ 2 4 1 + τ 3 log σ 2 θ θ 2 .
Therefore, if we are interesting in testing
H 0 : θ = θ 0 vs. H 1 : θ θ 0 ,
the Rao-type tests based on RMDPDGE are given by
R τ θ 0 = 1 n U n τ ( θ 0 ) 2 K τ θ 0 1 ,
where
U n τ ( θ 0 ) = 1 τ + 1 i = 1 n Ψ τ y i , θ
and Ψ τ y i , θ and K τ θ are given in (37) and (38). The null hypothesis is rejected if
R τ θ 0 > χ 1 , α 2 ,
where χ 1 , α 2 is the upper 1 α quantile of a chi-square distribution with 1 degree of freedom.
We finally derive explicit expressions of the Rao-type test statistics under Poisson, exponential and Lindley models.

4.1.1. Poisson Model

Let us assume that the random variable Y is Poisson with parameter θ . In this case, it is well known that E Y = V a r Y = θ , and so the RMDPDGE, for τ > 0 , is given by
θ ^ G τ = arg max θ τ + 1 τ 2 π θ τ 2 1 n i = 1 n exp τ 2 θ y i θ 2 τ 1 + τ 3 / 2 1 τ .
At τ = 0 , the RMDPDGE reduces to the Gaussian MLE,
θ ^ G = arg max θ 1 2 log 2 π 1 2 log θ 1 n i = 1 n 1 2 θ y i θ 2 .
On the other hand, the score function Ψ τ · is given by
Ψ τ y i , θ = τ + 1 2 2 π θ τ 2 θ 2 θ 2 θ + y i 2 exp τ 2 θ y i θ 2 + τ θ 1 + τ 3 2 ,
and naturally, at τ = 0 , we obtain the score function of the Gaussian MLE presented in [1]
Ψ 0 y i , θ = 1 2 θ 2 2 θ 2 + y i 2 .
On the other hand, the matrix K τ θ under the Poisson model has the explicit expression
K τ θ = 1 2 π τ 1 2 θ 2 + τ 1 1 + 2 τ 5 / 2 2 τ 2 + 2 θ + 4 θ τ + 1 τ 2 2 1 + τ 3 ,
and hence, the Rao-type tests based on RMDPDGE, R τ θ 0 , for testing a simple null hypothesis is given, for τ > 0 , by
R τ θ 0 = 1 n 1 2 2 π θ τ 2 θ 2 2 i = 1 n 2 θ 0 2 + y i 2 exp τ 2 θ 0 y i θ 0 2 + τ θ 1 + τ 3 2 2 × 2 π τ ( 2 θ 2 + τ ) 1 + 2 τ 5 / 2 2 τ 2 + 2 θ + 4 θ τ + 1 τ 2 2 1 + τ 3 1 .
Again, for τ = 0 , we obtain the expression of the classical Rao test based on the Gaussian MLE,
R 0 θ 0 = 1 4 n i = 1 n 2 θ 0 2 + y i 2 θ 0 2 2 2 θ 0 2 2 θ 0 + 1 .
The null hypothesis is rejected if
R τ θ 0 > χ 1 , α 2 .

4.1.2. Exponential Model

Let us assume now that the random variable Y comes from an exponential distribution with probability density function
f θ ( x ) = 1 θ exp x θ , x > 0 .
In this case, the true mean and variance are given by E Y = θ and V a r Y = θ 2 . The RMDPDGE under the exponential model, for τ > 0 , is given by
θ ^ G τ = arg max θ τ + 1 τ 1 θ 2 π τ 1 n i = 1 n exp τ 2 y i θ θ 2 τ 1 + τ 3 / 2 1 τ ,
and for τ = 0 , we have the Gaussian MLE
θ ^ G = arg max θ 1 2 log 2 π log θ 1 n i = 1 n 1 2 y i θ θ 2 .
On the other hand, the score function is
Ψ τ y i , θ = τ + 1 θ τ + 3 2 π τ y i 2 y i θ θ 2 exp τ 2 y i θ θ 2 + τ θ 2 1 + τ 3 2 ,
and for τ = 0 , we recover the score function of the Gaussian MLE,
Ψ 0 y i , θ = 1 θ 3 y i 2 y i θ + θ 2 .
The value K τ θ has the expression
K τ θ = 1 2 π τ θ 2 τ + 1 1 1 + 2 τ 5 / 2 4 τ 2 + 2 τ + 3 τ 2 1 + τ 3 ,
and at τ = 0
K 0 θ = 2 θ 2 .
Correspondingly, the Rao-type tests based on RMDPDGE for testing
H 0 : θ = θ 0 vs. H 1 : θ θ 0 ,
is given, for τ > 0 , by
R τ θ 0 = 1 n 1 θ 0 2 τ + 6 2 π τ i = 1 n y i 2 y i θ 0 θ 0 2 exp τ 2 y i θ 0 θ 0 2 + τ θ 0 2 1 + τ 3 2 2 × 2 π τ θ 2 τ + 1 1 1 + 2 τ 5 / 2 4 τ 2 + 2 τ + 3 τ 2 1 + τ 3 1 ,
and by
R 0 θ 0 = 1 2 n i = 1 n y i 2 y i θ 0 θ 0 2 2 θ 0 4
for τ = 0 .

4.1.3. Lindley Model

Let us finally assume that the random variable Y comes from a Lindley distribution [44] with probability density function
f θ ( x ) = θ 2 θ + 1 ( 1 + x ) exp θ x , x > 0 , θ > 0 .
In this case,
E Y = θ + 2 θ θ + 1 and V Y = θ 2 + 4 θ + 2 θ 2 θ + 1 2 .
The RMDPGE under the Lindlley model, for τ > 0 , is given by
θ ^ G = arg max θ τ + 1 τ 1 2 π τ θ 2 + 4 θ + 2 θ 2 θ + 1 2 τ 2 1 n i = 1 n exp τ 2 y i θ + 2 θ θ + 1 2 θ 2 + 4 θ + 2 θ 2 θ + 1 2 1 τ 1 + τ 3 2 1 τ ,
and for τ = 0 , we have
θ ^ G = arg max θ 1 2 ln 1 2 π 1 2 ln θ 2 + 4 θ + 2 θ 2 θ + 1 2 1 n i = 1 n θ y i θ y i θ 2 + 2 2 2 θ 2 + 4 θ + 2 .
On the other hand, the score funtion, Ψ τ ( y i , θ ) is given by
τ + 1 2 2 π τ 2 θ 2 + 4 θ + 2 θ 2 θ + 1 2 τ 2 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ θ θ y i θ 2 y i + 2 2 θ 2 + 4 θ + 2 2 y i θ + 2 θ θ + 1 exp τ 2 θ θ y i θ 2 y i + 2 2 θ 2 + 4 θ + 2 τ 1 + τ 3 2 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ
and K τ θ has the expression
θ 2 θ + 1 2 2 π θ 2 + 4 θ + 2 τ 1 1 + 2 τ 5 / 2 τ 2 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ 2 + 1 + 2 τ θ 2 + 4 θ + 2 θ 2 θ + 1 2 + 1 2 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ 2 τ 2 4 1 + τ 3 2 θ 3 + 12 θ 2 + 12 θ + 4 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ 2
and for τ = 0 ,
K 0 θ = 2 θ 3 + 12 θ 2 + 12 θ + 4 2 2 θ 4 + 5 θ 3 + 6 θ 2 + 2 θ 2 + θ + 2 θ θ + 1 .
Therefore, the Rao-type test based on RMDPDGE for testing
H 0 : θ = θ 0 vs. H 1 : θ θ 0
is given by
R τ θ 0 = 1 n U n τ ( θ 0 ) 2 K τ θ 0 1 > χ 1 , α 2 ,
where
U n τ ( θ 0 ) = 1 τ + 1 i = 1 n Ψ τ y i , θ ,
with Ψ τ y i , θ is defined in (42) and K τ θ in (43).

5. Simulation Study

We analyze here the performance of the Rao-type tests based on the MDPDGE, R τ θ 0 , in terms of robustness and efficiency. We compare the proposed general method assuming Gaussian distribution with Rao-type test statistics based on the true parametric distribution underlying the data.
We consider the exponential model with density function f θ 0 ( x ) given in (39). For the exponential model, the Rao-type test statistics based on MDPDGE is for τ > 0 given in (40) and for τ = 0 given in (41). To evaluate the robustness of the tests, we generate samples from an exponential mixture,
f θ 0 ε ( x ) = ( 1 ε ) f θ 0 ( x ) + ε f 2 θ 0 ( x ) ,
where θ 0 denotes the parameter of the exponential distribution and ε is the contamination proportion. The uncontaminated model is thus obtained by setting ε = 0 .
For comparison purposes, we have also considered the robust Rao-type tests based on the restricted MDPDE, which was introduced and studied [30]. The efficiency loss caused by the Gaussian assumption should be advertised by the poorer performance of the Rao-type tests based on the restricted MDPDGE with respect to their analogous based on the restricted MDPDE. For the exponential model, the family Rao-type test statistics based on the restricted MDPDE is given, for β > 0 , by
S n β ( θ 0 ) = 4 β 2 + 1 ( 2 β + 1 ) 3 β 2 ( β + 1 ) 4 1 1 n 1 θ 0 i = 1 n y i θ 0 exp β y i θ 0 + n β ( β + 1 ) 2 2 .
For β = 0 , the above test reduces to the classical Rao test given by
S n θ 0 = S β = 0 , n θ 0 = n X ¯ n θ 0 θ 0 2 .
We consider the testing problem
H 0 : θ = θ 0 vs. H 1 : θ θ 0 ,
and we empirically examine the level and power of both Rao-type test statistics, the usual test based on the parametric model and the Gaussian-based test by setting the true value of the parameter θ 0 = 2 and θ 0 = 1 , respectively. Different sample sizes were considered, namely n = 10 , 20 , 30 , 40 , 50 , 60 , 70 , 80 , 90 , 100 and 200 , but simulation results were quite similar and so, for brevity, we only report here results for n = 20 and n = 40 .
The empirical level of the test is computed
α ^ n ε = Number of times R n τ θ 0 ( or S n β θ 0 ) > χ 1 , 0.05 2 = 3.84146 Number of simulated samples .
We set ε = 0 % , 5 % , 10 % and 20 % of contamination proportions and perform the Monte-Carlo study over R = 10,000 replications. The tuning parameters τ and β are fixed from a grid of values, namely { 0 , 0.1 , , 0.7 } .
Simulation results are presented in Table 1 and Table 2 for n = 20 and n = 40 , respectively. The empirical powers are denoted by π ^ n ε . The robustness advantage in terms of level of both Rao-type tests considered, R τ θ 0 and S n β ( θ 0 ) with positive values of the turning parameter with respect to the test statistics with τ = 0 and β = 0 is clearly shown, as their simulated levels are closer to the nominal value in the presence of contamination.
Regarding the power of the tests for uncontaminated scenarios, there are values at least as good as the corresponding to τ = 0 and β = 0 , and for contaminated data, the power corresponding to τ > 0 and β > 0 is higher.
The loss of efficiency caused by the Guassian assumption can be measured by the discrepancy of the estimated levels and powers between the family of Rao-type tests based on the restricted MDPDGE and the restricted MDPDE. As expected, empirical levels of the test statistics based on the restricted MDPDGE are higher than the corresponding levels of the test based on the restricted MDPDE. However, the test statistic based on the parametric model, S n β ( θ 0 ) , is quite conservative and so the corresponding powers are higher than those of the proposed tests, R τ θ 0 . Based on the presented results, it seems that the proposed Rao-type tests, R τ θ 0 , performs reasonably well and offers an appealing alternative for situations where the probability density function of the true model is unknown or it is very complicated to work with it.

6. Conclusions

In this paper, we have considered the situation in which the parametric distribution of the variables is unknown and the only available information is given in terms of the mean vector and the variance–covariance matrix. Hence, we only know that the mean vector and variance–covariance matrix depend on the unknown values of a parameter vector θ . To deal with this problem, Zhang [1] proposed a procedure to estimate θ assuming that the underlying distribution was Gaussian. This assumption is justified in terms of the maximum entropy of the unknown distribution. However, the estimator developed in [1] is not robust. This procedure was extended using DPD leading to robust estimations of θ . In this paper, we have dealt with the case in which additional constraints must be imposed to the estimated parameters, thus leading to a new family of estimators that we have named RMDPGE. For these estimators, we have derived their asymptotic distribution and we have studied their robustness properties in terms of the corresponding IF. As an application, we have developed robust Rao-type test statistics under the null hypothesis, where the null hypothesis is indeed the restricted version of the estimator. Finally, we have tested the performance of these test statistics via a simulation study. From the results of this study, we empirically showed that the Rao-type tests considered in this paper seem to have a good performance in terms of efficiency and are more robust than the corresponding approach in [1].
There are several problems to be treated in future research. The most natural seems to develop Rao-type test statistics for composite null hypothesis and study the results obtained in terms of efficiency and robustness.

Author Contributions

Conceptualization, Á.F., M.J., P.M. and L.P.; methodology, Á.F., M.J., P.M. and L.P.; software, Á.F., M.J., P.M. and L.P.; validation, Á.F., M.J., P.M. and L.P.; formal analysis, Á.F., M.J., P.M. and L.P.; investigation, Á.F., M.J., P.M. and L.P.; resources, Á.F., M.J., P.M. and L.P.; data curation, Á.F., M.J., P.M. and L.P.; writing—original draft preparation, Á.F., M.J., P.M. and L.P.; writing—review and editing, Á.F., M.J., P.M. and L.P.; visualization, Á.F., M.J., P.M. and L.P.; supervision, Á.F., M.J., P.M. and L.P.; project administration, Á.F., M.J., P.M. and L.P.; funding acquisition, Á.F., M.J., P.M. and L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Spanish Grant PID2021-124933NB-I00.

Data Availability Statement

Not applicable.

Acknowledgments

Jaenada, M., Miranda P. and Pardo L. are members of the Interdisciplinary Mathematics Institute.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLEMaximum likelihood estimator
MDPDEMinimum density power divergence estimator
MDPDGEMinimum density power divergence Gaussian estimator
RMDPDERestricted minimum density power divergence estimator
RMDPDGERestricted minimum density power divergence Gaussian estimator

Appendix A. Derivatives Calculation and Proofs of the Main Results

Appendix A.1. Previous Results

In different parts of Appendix, the following results are applied.
Lemma A1.
The following results can be shown:
  • Σ θ θ i = Σ θ t r a c e Σ θ 1 Σ θ θ i .
  • t r a c e Σ θ θ i = t r a c e Σ θ θ i .
  • Σ θ θ i 1 = Σ θ 1 Σ θ θ i Σ θ 1 .
Lemma A2.
Let Y be a normal population with vector mean μ and variance–covariance Σ . Then, we have
E Y μ T A Y μ = T r a c e A Σ ,
E Y μ T A Y μ Y μ T B Y μ = T r a c e A Σ B + B T Σ + T r a c e A Σ T r a c e B Σ ,
E Y μ T A Y μ Y μ = 0 .
For more details about these results, see for instance [45].

Appendix A.2

Proof of Proposition 1.
The expression of H n τ ( θ ) introduced in (5) is given by
H n τ ( θ ) = a Σ θ τ / 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b 1 τ .
Consider the d-dimensional random vector Y τ ( θ ) defined in (35). Applying the Central Limit Theorem, we have
n θ H n τ ( θ ) = 1 n i = 1 n Ψ τ ( y i ; θ ) L n N ( 0 m , S τ ( θ 0 ) ) ,
with
S τ ( θ 0 ) = C o v Y τ ( θ ) = E Y τ ( θ ) T Y τ ( θ ) ,
because
E Y τ ( θ ) = 0 d .
To see that E Y τ ( θ ) = 0 d , consider
E Y τ ( θ ) = a τ 2 Σ θ τ / 2 E t r a c e Σ θ 1 Σ θ θ exp τ 2 Y μ θ T Σ θ 1 Y μ θ + b t r a c e Σ θ 1 Σ θ θ + exp τ 2 Y μ θ T Σ θ 1 Y μ θ 2 μ θ θ T Σ θ 1 Y μ θ + Y μ θ T Σ θ 1 Σ θ θ Σ θ 1 Y μ θ = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ 1 τ + 1 m / 2 + τ τ + 1 m 2 + 1 t r a c e Σ θ 1 Σ θ θ + 1 τ + 1 m 2 + 1 t r a c e Σ θ 1 Σ θ θ = 0 d .
We can observe that Y τ ( θ ) is a d-dimensional vector whose j-th component is
Y τ j ( θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ j exp τ 2 y μ θ T Σ θ 1 y μ θ + b t r a c e Σ θ 1 Σ θ θ j + exp τ 2 y μ θ T Σ θ 1 y μ θ 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ , j = 1 , . . . , d .
Therefore, the element ( i , j ) of the matrix S τ ( θ 0 ) is given by
E Y τ i ( θ ) Y τ j ( θ ) .
We are going to obtain Y τ i ( θ ) Y τ j ( θ ) . First,
Y τ i ( θ ) Y τ j ( θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i exp τ 2 y i μ θ T Σ θ 1 y i μ θ + b t r a c e Σ θ 1 Σ θ θ i + exp τ 2 y i μ θ T Σ θ 1 y i μ θ 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ j exp τ 2 y μ θ T Σ θ 1 y μ θ + b t r a c e Σ θ 1 Σ θ θ j + exp τ 2 y μ θ T Σ θ 1 y μ θ 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ .
Therefore, Y τ i ( θ ) Y τ j ( θ ) is given by
a 2 τ 2 2 Σ θ τ t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j b × exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ t r a c e Σ θ 1 Σ θ θ i exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ b t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j exp τ 2 y μ θ T Σ θ 1 y μ θ + b 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + b t r a c e Σ θ 1 Σ θ θ i exp τ 2 y μ θ T Σ θ 1 y μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ t r a c e Σ θ 1 Σ θ θ i exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ + b t r a c e Σ θ 1 Σ θ θ i exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ + exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ .
Consequently, we can write Y τ i ( θ ) Y τ j ( θ ) by
Y τ i ( θ ) Y τ j ( θ ) = a 2 τ 2 2 Σ θ τ C 1 + C 2 + C 3 + C 4 + C 5 + C 6 + C 7 + C 8 + C 9 ,
and thus,
E Y i ( θ ) Y j ( θ ) = a 2 τ 2 2 Σ θ τ × E C 1 + C 2 + C 3 + C 4 + C 5 + C 6 + C 7 + C 8 + C 9 .
Now, we are going to calculate the different expectations appearing in Equation A2. We have
C 1 = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ .
Therefore,
E C 1 = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 2 π m / 2 Σ θ 1 / 2 exp 1 2 y μ θ T Σ θ 1 y μ θ × exp 1 2 y μ θ T Σ θ 2 τ 1 y μ θ d y = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 2 τ + 1 m 2 1 2 π m / 2 Σ θ 2 τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ 2 τ + 1 1 y μ θ d y = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 2 τ + 1 m 2 .
The expression of C 2 is given by
C 2 = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j b exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ ,
and thus,
E C 2 = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 2 π m / 2 Σ θ 1 / 2 exp 1 2 y μ θ T Σ θ 1 y μ θ × exp 1 2 y μ θ T Σ θ τ 1 y μ θ d y = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 1 + τ m 2 1 2 π m / 2 Σ θ τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ τ + 1 1 y μ θ d y = τ 1 + τ m + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
The expression of C 3 is given by
C 3 = t r a c e Σ θ 1 Σ θ θ i exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ .
Then,
E C 3 = t r a c e Σ θ 1 Σ θ θ i 1 1 + 2 τ m / 2 × 1 2 π m / 2 Σ θ 2 τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ 1 y μ θ × y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ d y = t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 1 + 2 τ m 2 + 1 .
Related to C 4 , we have
C 4 = b t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j exp τ 2 y μ θ T Σ θ 1 y μ θ ,
and
E C 4 = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 2 π m / 2 Σ θ 1 / 2 exp 1 2 y μ θ T Σ θ 1 y μ θ × exp τ 2 y μ θ T Σ θ 1 y μ θ d y = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 1 + τ m 2 1 2 π m / 2 Σ θ τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ τ + 1 1 y μ θ d y = τ 1 + τ m + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
Related to C 5 , we have
C 5 = b 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j ,
and
E C 5 = τ 2 1 + τ m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
The expression of C 6 is given by
C 6 = b t r a c e Σ θ 1 Σ θ θ i exp τ 2 y μ θ T Σ θ 1 y μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ ,
and
E C 6 = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 1 + τ m 2 + 1 = τ 1 + τ m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
The expression of C 7 is
C 7 = t r a c e Σ θ 1 Σ θ θ i exp 2 τ 2 y μ θ T Σ θ 1 y μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ ,
and
E C 7 = τ 1 + 2 τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
Related to C 8 ,
C 8 = b t r a c e Σ θ 1 Σ θ θ i exp τ 2 y μ θ T Σ θ 1 y μ θ × 2 μ θ θ j T Σ θ 1 y μ θ + y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ ,
and
E C 8 = τ 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 1 + τ m 2 + 1 = τ 1 + τ m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
Finally,
C 9 = exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ × 2 μ θ θ j T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ .
Therefore,
C 9 = exp 2 τ 2 y i μ θ T Σ θ 1 y i μ θ 4 y i μ θ T Σ θ 1 μ θ θ i T × μ θ θ j Σ θ 1 y i μ θ + 2 μ θ θ i T Σ θ 1 y i μ θ y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ + 2 y i μ θ T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ μ θ θ j T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ = A 1 + A 2 + A 3 + A 4 ,
and
E C 9 = E A 1 + E A 4
because E A 2 = E A 3 = 0 . We have
E A 1 = E exp 2 τ 2 Y μ θ T Σ θ 1 Y μ θ 4 Y μ θ T Σ θ 1 × μ θ θ i T μ θ θ j Σ θ 1 Y μ θ = 4 1 2 π m / 2 Σ θ 1 / 2 exp 1 2 y μ θ T Σ θ 2 τ + 1 1 y μ θ y μ θ T × Σ θ 1 μ θ θ i T μ θ θ j Σ θ 1 y μ θ d y = 4 1 2 τ + 1 m / 2 1 2 π m / 2 Σ θ 2 τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ 2 τ + 1 1 y μ θ × y μ θ T Σ θ 1 μ θ θ i T μ θ θ j Σ θ 1 y μ θ d y = 4 1 2 τ + 1 m / 2 t r a c e Σ θ 1 μ θ θ i T μ θ θ j Σ θ 1 Σ θ 2 τ + 1 = 4 2 τ + 1 m / 2 + 1 t r a c e Σ θ 1 μ θ θ i T μ θ θ j .
Finally, we are going to obtain E A 4 .
E A 4 = E exp 2 τ 2 Y μ θ T Σ θ 1 Y μ θ Y μ θ T Σ θ 1 Σ θ θ i Σ θ 1 × Y μ θ Y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 Y μ θ = 1 2 π m / 2 Σ θ 1 / 2 exp 1 2 y μ θ T Σ θ 2 τ + 1 1 y μ θ y μ θ T × Σ θ 1 Σ θ θ i Σ θ 1 y μ θ y μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y μ θ d y = 1 2 τ + 1 m / 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ 2 τ + 1 × Σ θ 1 Σ θ θ i Σ θ 1 + Σ θ 1 Σ θ θ j Σ θ 1 Σ θ 2 τ + 1 + 1 2 τ + 1 m / 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ 2 τ + 1 × t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ 2 τ + 1 = 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j + 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j T + 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j = 2 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j + 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
Based on the previous results, we have
E Y τ i ( θ ) Y τ j ( θ ) = a 2 τ 2 2 Σ θ τ 1 2 τ + 1 m 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j τ τ + 1 m + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 2 τ + 1 m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j τ τ + 1 m + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + τ 2 τ + 1 m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + τ τ + 1 m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j 1 2 τ + 1 m 2 + 1 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + τ τ + 1 m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + 4 2 τ + 1 m 2 + 1 t r a c e Σ θ 1 μ θ θ i T μ θ θ j + 2 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j + 1 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
The previous expression can be written as
E Y τ i ( θ ) Y τ j ( θ ) = a 2 τ 2 2 Σ θ τ t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 2 τ + 1 m 2 1 2 τ + 1 m 2 + 1 1 2 τ + 1 m 2 + 1 + 1 2 τ + 1 m 2 + 2 + t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × τ τ + 1 m + 1 τ τ + 1 m + 1 + τ 2 τ + 1 m + 2 + τ τ + 1 m + 2 + τ τ + 1 m + 2 + 4 2 τ + 1 m 2 + 1 t r a c e Σ θ 1 μ θ θ i T μ θ θ j + 2 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j .
Therefore,
E Y τ i ( θ ) Y τ j ( θ ) = τ + 1 τ 2 π m τ / 2 2 τ 2 2 Σ θ τ × 4 τ 2 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j τ 2 1 + τ m + 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + 4 4 2 τ + 1 m 2 + 1 t r a c e Σ θ 1 μ θ θ i T μ θ θ j + 2 2 τ + 1 m 2 + 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j .
Finally,
E Y τ i ( θ ) Y τ j ( θ ) = τ + 1 2 1 2 π m τ / 2 Σ θ 1 / 2 2 τ 1 2 τ + 1 m 2 + 2 × Δ 2 τ i Δ 2 τ j + 2 τ + 1 t r a c e Σ θ 1 μ θ θ i T μ θ θ j + 1 2 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ j 1 2 π m τ / 2 Σ θ 1 / 2 2 τ 1 1 + τ m + 2 Δ τ i Δ τ j = τ + 1 2 K τ i j θ ,
where K τ i j θ was defined in (10). Then,
n θ H n ( θ ) L n N 0 , τ + 1 2 K τ θ ,
and
n 1 τ + 1 θ H n ( θ ) L n N 0 , K τ θ .

Appendix A.3

Proof of Proposition 2.
Note that
θ i H n τ ( θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b + a Σ θ τ / 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ τ 2 × 2 μ θ θ i T Σ θ 1 y i μ θ y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b + a τ 2 Σ θ τ / 2 τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ × y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ .
Therefore,
2 θ i θ j H n τ ( θ ) = θ j L 1 τ ( θ ) + θ j L 2 τ ( θ ) ,
being
L 1 τ ( θ ) = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b
and
L 2 τ ( θ ) = a τ 2 Σ θ τ / 2 τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ × y i μ θ T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ .
We are going to obtain θ j L 1 τ ( θ ) .
θ j L 1 τ ( θ ) = a τ 2 τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ j t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + Σ θ 1 2 Σ θ θ i θ j × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ τ 2 × 2 μ θ θ i T Σ θ 1 y i μ θ y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ = D 1 + D 2 + D 3 ,
being
D 1 = a τ 4 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b ,
D 2 = a τ 2 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + Σ θ 1 2 Σ θ θ i θ j × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ b ,
and
D 3 = a τ 2 4 Σ θ τ / 2 t r a c e Σ θ 1 Σ θ θ i × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ .
Now, we are going to see some result that will be important in order to obtain convergence in probability of D 1 , D 2 and D 3 .
Lemma A3.
We have
l = 1 n i = 1 n exp τ 2 Y i μ θ T Σ θ 1 Y i μ θ P n 1 1 + τ m / 2 .
Proof. 
It is clear that
l P n E N μ θ , Σ θ exp τ 2 Y μ θ T Σ θ τ Y μ θ = exp τ 2 Y μ θ T Σ θ τ Y μ θ f N μ θ , Σ θ y d y = 1 1 + τ m / 2 1 2 π m / 2 1 Σ θ τ + 1 1 / 2 exp τ + 1 2 y μ θ T Σ θ τ y μ θ d y = 1 1 + τ m / 2 .
Lemma A4.
We have
m = 1 n i = 1 n exp τ 2 Y i μ θ T Σ θ 1 Y i μ θ b P n 1 1 + τ m 2 + 1 .
Proof. 
Applying the previous Lemma
m P n 1 1 + τ m / 2 τ 1 + τ m 2 + 1 = 1 1 + τ m 2 + 1 .
Lemma A5.
If we denote
n = 1 n i = 1 n exp τ 2 Y i μ θ T Σ θ 1 Y i μ θ Y i μ θ T A Y i μ θ ,
we have
n P n t r a c e A Σ θ 1 + τ m 2 + 1 .
Proof. 
It is clear that
n P n E N μ θ , Σ θ exp 1 2 Y μ θ T Σ θ τ 1 Y μ θ × Y μ θ T A Y μ θ = 1 1 + τ m / 2 1 2 π m / 2 1 Σ θ τ + 1 1 / 2 exp 1 2 y μ θ T Σ θ τ + 1 1 y μ θ × y μ θ T A y μ θ d y = 1 1 + τ m / 2 E N μ θ , Σ θ 1 + τ Y μ θ T A Y μ θ = 1 1 + τ m 2 + 1 t r a c e A Σ θ .
Based on the previous results we have in relation to D 1 ,
D 1 P n τ 4 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
With respect to D 2 ,
D 2 P n 1 2 π m / 2 Σ θ τ 2 1 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i Σ θ τ 2 2 π m / 2 1 2 t r a c e Σ θ 1 2 Σ θ θ j θ i .
In a similar way, we obtain for D 3 that
D 3 P n τ 4 Σ θ τ 2 2 π m τ / 2 t r a c e Σ θ 1 Σ θ θ i 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j .
Therefore, we have
θ j L 1 τ ( θ ) P n 1 2 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i 1 2 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 Σ θ 1 2 Σ θ θ j θ i .
Now, we have
θ j L 2 τ ( θ ) = a τ 2 4 Σ θ τ 2 t r a c e Σ θ 1 Σ θ θ j × 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ 2 μ θ θ i T × Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ + a τ 2 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × τ 2 θ j y i μ θ T Σ θ 1 y i μ θ 2 μ θ θ i T Σ θ 1 y i μ θ × y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ + a τ 2 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × θ j 2 μ θ θ i T Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ = C 1 + C 2 + C 3 .
Now,
C 1 = a τ 2 4 Σ θ τ 2 t r a c e Σ θ 1 Σ θ θ j 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 μ θ θ i T Σ θ 1 y i μ θ × y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ .
It is clear that
C 1 P n τ 4 1 1 + τ m / 2 Σ θ τ 2 2 π m τ / 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j .
Next,
C 2 = a τ 2 4 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ S 1 + S 2 + S 3 + S 4 = L 1 * + L 2 * + L 3 * + L 4 * .
First,
L 1 * = a τ 2 4 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 4 y i μ θ T Σ θ 1 μ θ θ j μ θ θ i T Σ θ 1 y i μ θ ,
and
L 1 * P n Σ θ τ 2 2 π m τ / 2 τ 1 + τ m / 2 t r a c e Σ θ 1 μ θ θ j μ θ θ i T .
It is clear that
L 2 * P n 0 and L 3 * P n 0 .
Finally,
L 4 * P n τ + 1 2 π m τ / 2 τ 4 Σ θ τ 2 1 1 + τ m / 2 × t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i Σ θ 1 Σ θ θ i Σ θ 1 + Σ θ 1 Σ θ θ i Σ θ 1 × Σ θ 1 + τ + t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ 1 + τ t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ 1 + τ = 2 τ 4 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ i Σ θ 1 Σ θ Σ θ θ i + τ 4 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j t r a c e Σ θ 1 Σ θ θ i .
Therefore,
C 2 = L 1 * + L 2 * + L 3 * + L 4 * P n R ,
where
R = τ Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 t r a c e Σ θ 1 μ θ θ j μ θ θ j T + τ 2 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + τ 4 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j t r a c e Σ θ 1 Σ θ θ i .
Finally,
C 3 = a τ 2 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × 2 θ j μ θ θ i T Σ θ 1 y i μ θ + 2 μ θ θ i T Σ θ 1 θ i y i μ θ 2 μ θ θ i T Σ θ 1 μ θ θ j T 2 μ θ θ i T Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ + y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i Σ θ 1 + Σ θ 1 2 Σ θ θ i θ j Σ θ 1 Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i Σ θ 1 y i μ θ = A 1 * + A 2 * + A 3 * + A 4 * + A 5 * .
It is clear that
A 1 * P n 0 A 2 * P n 0 and A 4 * P n 0 .
On the other hand,
A 3 * = a τ 2 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ 2 μ θ θ i T Σ θ 1 μ θ θ i ,
and
A 3 * P n ( τ + 1 ) Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 μ θ θ i T Σ θ 1 μ θ θ j .
Related to A 5 * , we have
A 5 * = a τ 2 Σ θ τ 2 1 n i = 1 n exp τ 2 y i μ θ T Σ θ 1 y i μ θ × y i μ θ T Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i Σ θ 1 + Σ θ 1 2 Σ θ θ i θ j Σ θ 1 Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ j Σ θ 1 y i μ θ ,
and
A 5 * P n 1 2 π m τ / 2 1 2 Σ θ τ 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + 1 2 π m τ / 2 1 2 Σ θ τ 2 1 1 + τ m / 2 t r a c e Σ θ 1 2 Σ θ θ i θ 1 2 π m τ / 2 1 2 Σ θ τ 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i .
Therefore,
C 3 P n τ + 1 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 μ θ θ i T Σ θ 1 μ θ θ j Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 2 Σ θ θ i θ j Σ θ τ 2 2 π m τ / 2 1 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i .
We are going to join all the previous expressions in order to obtain θ j L 2 τ ( θ ) ,
θ j L 2 τ ( θ ) = C 1 + C 2 + C 3 = C 1 + L 1 * + L 2 * + L 3 * + L 4 * + C 3 = C 1 + L 1 * + L 2 * + L 3 * + L 4 * + A 1 * + A 2 * + A 3 * + A 4 * + A 5 * .
Then,
θ j L 2 τ ( θ ) P n τ 4 1 1 + τ m / 2 Σ θ τ 2 2 π m τ / 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + τ Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + τ 2 Σ θ τ 2 2 π m τ / 2 1 + τ m 2 + 1 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + τ 4 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j t r a c e Σ θ 1 Σ θ θ i τ + 1 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 μ θ θ i T Σ θ 1 μ θ θ j Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 2 Σ θ θ i θ j Σ θ τ 2 2 π m τ / 2 1 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i .
Based on the previous results, we have
2 θ i θ j H n τ ( θ ) = θ j L 1 τ ( θ ) + θ j L 2 τ ( θ ) = D 1 + D 2 + D 3 + C 1 + C 2 + C 3 = D 1 + D 2 + D 3 + C 1 + L 1 * + L 2 * + L 3 * + L 4 * + A 1 * + A 2 * + A 3 * + A 4 * + A 5 *
and
2 θ i θ j H n τ ( θ ) P n 1 2 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i 1 2 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 Σ θ 1 2 Σ θ θ j θ i τ 4 1 1 + τ m / 2 Σ θ τ 2 2 π m τ / 2 t r a c e Σ θ 1 Σ θ θ i t r a c e Σ θ 1 Σ θ θ j + τ Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + τ 2 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + τ 4 Σ θ τ 2 2 π m τ / 2 1 + τ m / 2 + 1 t r a c e Σ θ 1 Σ θ θ j t r a c e Σ θ 1 Σ θ θ i τ + 1 Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 μ θ θ i T Σ θ 1 μ θ θ j Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i + Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 2 Σ θ θ i θ j Σ θ τ 2 2 π m τ / 2 1 1 + τ m / 2 1 2 t r a c e Σ θ 1 Σ θ θ j Σ θ 1 Σ θ θ i .
After some algebra, we have
2 θ i θ j H n τ ( θ ) P n τ + 1 J τ i j θ .

References

  1. Zhang, T. General Gaussian estimation. J. Multivar. Anal. 2019, 169, 234–247. [Google Scholar] [CrossRef]
  2. Kapur, J.N. Maximum Entropy Models in Science and Engineering; Wiley: New Delhi, India, 1989. [Google Scholar]
  3. Castilla, E.; Zografos, K. On distance-type Gaussian estimation. J. Multivar. Anal. 2022, 188, 22. [Google Scholar] [CrossRef]
  4. Basu, A.; Harris, I.R.; Hjort, N.L.; Jones, M.C. Robust and efficient estimation by minimising a density power divergence. Biometrika 1998, 85, 549–559. [Google Scholar] [CrossRef] [Green Version]
  5. Pardo, L. Statistical Inference Based on Divergence Measures; CRC Press: Boca Raton, FL, USA, 2006. [Google Scholar]
  6. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley and Sons, Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  7. Guiasu, S. Information Theory with Applications; McGraw Hill: New York, NY, USA, 1977. [Google Scholar]
  8. Basu, A.; Mandal, A.; Martin, N.; Pardo, L. Testing composite hypothesis based on the density power divergence. Sankhya B 2018, 80, 222–262. [Google Scholar] [CrossRef] [Green Version]
  9. Giraitis, L.; Taniguchi, M.; Taqqui, M.S. Asymptotic normality of quadratic forms of martingale difference. Stat. Inference Stoch. Process. 2017, 20, 315–327. [Google Scholar] [CrossRef] [Green Version]
  10. Götze, F.; Tikhomirov, A. Asymptotic distribution of quadratic forms and applications. J. Theor. Probab. 2002, 15, 423–475. [Google Scholar] [CrossRef]
  11. Wu, W.B.; Shao, X. A limit theorem for quadratic forms and its applications. Econom. Theory 2007, 23, 930–950. [Google Scholar] [CrossRef] [Green Version]
  12. Eubank, R.L.; Wang, S. A central limit theorem for the sum of generalized linear and quadratic forms. Statistics 1999, 33, 85–91. [Google Scholar] [CrossRef] [Green Version]
  13. Horváth, L.; Shao, Q.M. Limit theorems for quadratic forms with applications to Whittle’s estimate. Ann. Appl. Probab. 1999, 9, 146–187. [Google Scholar] [CrossRef]
  14. Silvey, S.D. Reprinting, Monographs on Statistical Subjects; Chapman and Hall: London, UK, 1975. [Google Scholar]
  15. Sen, P.K.; Singer, J.M.; Pedroso De Lima, A.C. From Finite Sample to Asymptotic Methods in Statistics. Cambridge Series in Statistical and Probabilistic Mathematics; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  16. Ferguson, T.S. A Course in Large Sample Theory; Chapman & Hall: London, UK, 1996. [Google Scholar]
  17. Aitchison, J.; Silvey, S.D. Maximum Likelihood Estimation of Parameters Subject to Restraints. Ann. Math. Stat. 1958, 29, 813–828. [Google Scholar] [CrossRef]
  18. Pardo, J.A.; Pardo, L.; Zografos, K. Minimum ϕ-divergence estimators with constraints in multinomial populations. J. Stat. Plan. Inference 2002, 104, 221–237. [Google Scholar] [CrossRef]
  19. Ghosh, A. Influence function analysis of the restricted minimum divergence estimators: A general form. Electron. J. Stat. 2015, 9, 1017–1040. [Google Scholar] [CrossRef] [Green Version]
  20. Jaenada, M.; Pardo, L. Robust statistical inference in generalized linear models based on minimum Renyi’s pseudodistance estimators. Entropy 2022, 24, 123. [Google Scholar] [CrossRef] [PubMed]
  21. Jaenada, M.; Miranda, P.; Pardo, L. Robust test statistics based on restricted minimum Rényi’s pseudodistance estimators. Entropy 2022, 24, 616. [Google Scholar] [CrossRef] [PubMed]
  22. Martin, N. Rao’s Score Tests on Correlation Matrices. arXiv 2021, arXiv:2012.14238. [Google Scholar]
  23. Martin, N. Robust and efficient Breusch-Pagan test-statistic: An application of the beta-score Lagrange multipliers test for non-identically distributed individuals. arXiv 2023, arXiv:2301.07245. [Google Scholar]
  24. Huang, C.; Liu, H.; Shi, H.; Chen, X.; Xiao, M.; Wang, Z.; Cao, J. Bifurcations in a fractional-order neural network with multiple leakage delays. Neural Netw. 2020, 131, 115–126. [Google Scholar] [CrossRef]
  25. Huang, C.; Wang, J.; Chen, X.; Cao, J. Bifurcations in a fractional-order BAM neural network with four different delays. Neural Netw. 2021, 141, 344–354. [Google Scholar] [CrossRef]
  26. Xu, C.; Mu, D.; Liu, Z.; Pang, Y.; Liao, M.; Aouiti, C. New insight into bifurcation of fractional-order 4D neural networks incorporating two different time delays. Commun. Nonlinear Sci. Numer. Simul. 2023, 118, 107043. [Google Scholar] [CrossRef]
  27. Xu, C.; Zhang, W.; Aouiti, C.; Liu, Z.; Yao, L. Bifurcation insight for a fractional-order stage-structured predator-prey system incorporating mixed time delays. Math. Methods Appl. Sci. 2023. early view. [Google Scholar] [CrossRef]
  28. Hampel, F.R. Contributions to the Theory of Robust Estimation. Ph.D. Thesis, University of California, Berkeley, CA, USA, 1968. [Google Scholar]
  29. Basu, A.; Chakraborty, S.; Ghosh, A.; Pardo, L. Robust density power divergence based tests in multivariate analysis: A comparative overview of different approaches. J. Multivar. Anal. 2022, 188, 104846. [Google Scholar] [CrossRef]
  30. Basu, A.; Ghosh, A.; Martin, N.; Pardo, L. A robust generalization of the Rao test. J. Bus. Econ. Stat. 2022, 40, 868–879. [Google Scholar] [CrossRef]
  31. Castilla, E.; Jaenada, M.; Pardo, L. Estimation and testing on independent not identically distributed observations based on Rényi’s pseudodistances. IEEE Trans. Inf. Theory 2022, 68, 4588–4609. [Google Scholar] [CrossRef]
  32. Castilla, E.; Jaenada, M.; Martín, N.; Pardo, L. Robust approach for comparing two dependent normal populations through Wald-type tests based on Rényi’s pseudodistance estimators. Stat. Comput. 2022, 32, 34. [Google Scholar] [CrossRef]
  33. Ghosh, A.; Basu, A.; Pardo, L. Robust Wald-type tests under random censoring. Stat. Med. 2021, 40, 1285–1305. [Google Scholar] [CrossRef] [PubMed]
  34. Menéndez, M.; Morales, D.; Pardo, L.; Vajda, I. Divergence-based estimation and testing of statistical models of classification. J. Multivar. Anal. 1995, 1995 54, 329–354. [Google Scholar] [CrossRef] [Green Version]
  35. Toma, A.; Broniatowski, M. Dual Divergence Estimators and Tests: Robustness Results. J. Multivar. Anal. 2011, 102, 20–36. [Google Scholar] [CrossRef] [Green Version]
  36. Toma, A.; Leoni-Aubin, S. Robust Tests Based on Divergence Estimators and Sadlepoint Approximations. J. Multivar. Anal. 2010, 101, 1143–1155. [Google Scholar] [CrossRef] [Green Version]
  37. Qin, Y.; Priebe, C.E. Robust Hypothesis Testing via Lq-Likelihood. Stat. Sin. 2017, 27, 1793–1813. [Google Scholar] [CrossRef] [Green Version]
  38. Cambanis, S.; Huang, S.; Simons, G. On the theory of elliptically contoured distributions. J. Multivar. Anal. 1981, 11, 368–385. [Google Scholar] [CrossRef] [Green Version]
  39. Fang, K.T.; Zhang, K.T. Generalized Multivariate Analysis; Science Press Beiging: Beiging, China, 1990. [Google Scholar]
  40. Fang, K.T.; Kotz, S.; Ng, K.W. Symmetric Multivariate and Related Distributions; Chapman & Hall: London, UK, 1987. [Google Scholar]
  41. Gupta, A.K.; Varga, T. Elliptically Contoured Models in Statistics; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1993. [Google Scholar]
  42. Veroniki, A.A. Multivariate Elliptically Contoured Distributions: Estimation-Testing Hypothesis. A Review and an Application. Master’s Thesis, Department of Mathematics, University of Ioannina, Ioannina, Greece, 2010. [Google Scholar]
  43. Lehman, E.L. Testing Statistical Hypothesis; John Wiley and Sons: Hoboken, NJ, USA, 1959. [Google Scholar]
  44. Lindley, D.V. Fiducial distributions and Bayes’ theorem. J. R. Stat. Soc. 1958, 20, 102–107. [Google Scholar] [CrossRef]
  45. Harville, D.A. Matrix Algebra From a Statistician’s Perspective; Springer: Berlin/Heidelberg, Germany, 1997. [Google Scholar]
Figure 1. Influence function of the MDPDGE for the exponential model with τ = 0 (red), τ = 0.2 (black) and τ = 0.8 (green).
Figure 1. Influence function of the MDPDGE for the exponential model with τ = 0 (red), τ = 0.2 (black) and τ = 0.8 (green).
Mathematics 11 01480 g001
Table 1. Simulated levels for different contamination proportions and different tuning parameters τ , β = 0 , 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 for the Rao-type tests R τ θ 0 and S n β ( θ 0 ) for n = 20 .
Table 1. Simulated levels for different contamination proportions and different tuning parameters τ , β = 0 , 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 for the Rao-type tests R τ θ 0 and S n β ( θ 0 ) for n = 20 .
τ α ^ n 0 α ^ n 0.05 α ^ n 0.10 α ^ n 0.20 π ^ n 0 π ^ n 0.1 π ^ n 0.15 π ^ n 0.20
0.00.26010.30930.34530.46610.92780.67910.68870.5088
0.10.18950.17480.15610.19890.95440.72130.73010.0595
0.20.21200.17760.14170.11740.97470.83980.84300.6395
0.30.25320.21130.16600.12750.98260.89630.89610.7301
0.40.29630.24470.19860.14710.98630.92280.92570.7893
0.50.32430.27730.23070.16950.98750.93630.93860.8254
0.60.35120.30550.25990.18990.98850.94410.94370.8434
0.70.37510.32580.27620.20600.98840.94660.94690.8541
β
0.00.04530.06820.10480.19090.72000.43650.43840.2323
0.10.04760.06020.07800.14170.77990.52230.52670.3029
0.20.04980.05520.06670.11030.79220.57510.57800.3558
0.30.04940.05170.05840.08970.78820.59970.60240.3878
0.40.04890.05050.05350.07730.77790.60670.60580.4106
0.50.04940.04980.05040.06920.76340.60480.60370.4221
0.60.04910.05040.04970.06470.74920.60080.59860.4265
0.70.05020.04950.04940.06130.73480.59320.59190.4259
Table 2. Simulated levels for different contamination proportions and different tuning parameters τ , β = 0 , 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 for the Rao-type tests R τ θ 0 and S n β ( θ 0 ) for n = 40 .
Table 2. Simulated levels for different contamination proportions and different tuning parameters τ , β = 0 , 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 for the Rao-type tests R τ θ 0 and S n β ( θ 0 ) for n = 40 .
τ α ^ n 0 α ^ n 0.05 α ^ n 0.10 α ^ n 0.20 π ^ n 0 π ^ n 0.1 π ^ n 0.15 π ^ n 0.20
0.00.30140.35880.44070.59190.99480.80640.75910.5957
0.10.23930.19340.17570.20320.99910.95400.92290.7712
0.20.77120.25590.19700.13170.99950.99160.98460.9204
0.30.42570.34850.27820.17530.99970.99970.99530.9694
0.40.50210.42940.35720.23880.99990.99890.99780.9851
0.50.56420.49200.42530.29930.99990.99920.99860.9908
0.60.60840.54150.47420.34911.00000.99920.99940.9935
0.70.64160.57550.50810.38311.00000.99940.99940.9948
β
0.00.04670.07580.13090.27280.98380.80930.74830.4905
0.10.04690.06230.09590.19870.98700.87700.83170.6072
0.20.04640.05540.08000.15260.98620.90100.86870.6778
0.30.04810.05290.07040.12200.98460.90840.88040.7169
0.40.04830.05180.06490.10360.98080.90590.88090.7316
0.50.05000.05190.06180.09290.97560.90080.87420.7338
0.60.05000.05010.05770.08580.96890.89140.86620.7317
0.70.05040.05190.05620.08010.96340.88130.85620.7258
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Felipe, Á.; Jaenada, M.; Miranda, P.; Pardo, L. Restricted Distance-Type Gaussian Estimators Based on Density Power Divergence and Their Applications in Hypothesis Testing. Mathematics 2023, 11, 1480. https://doi.org/10.3390/math11061480

AMA Style

Felipe Á, Jaenada M, Miranda P, Pardo L. Restricted Distance-Type Gaussian Estimators Based on Density Power Divergence and Their Applications in Hypothesis Testing. Mathematics. 2023; 11(6):1480. https://doi.org/10.3390/math11061480

Chicago/Turabian Style

Felipe, Ángel, María Jaenada, Pedro Miranda, and Leandro Pardo. 2023. "Restricted Distance-Type Gaussian Estimators Based on Density Power Divergence and Their Applications in Hypothesis Testing" Mathematics 11, no. 6: 1480. https://doi.org/10.3390/math11061480

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop