Next Article in Journal
A Method for Reducing Sub-Divisional Errors in Open-Type Optical Linear Encoders with Angle Shift Pattern Main Scale
Next Article in Special Issue
Birth–Death Processes with Two-Type Catastrophes
Previous Article in Journal
CLG: Contrastive Label Generation with Knowledge for Few-Shot Learning
Previous Article in Special Issue
Modified Block Bootstrap Testing for Persistence Change in Infinite Variance Observations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymptotic Form of the Covariance Matrix of Likelihood-Based Estimator in Multidimensional Linear System Model for the Case of Infinity Number of Nuisance Parameters

by
Alexander Varypaev
Institute of Mathematics and Informatics, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria
Mathematics 2024, 12(3), 473; https://doi.org/10.3390/math12030473
Submission received: 11 December 2023 / Revised: 16 January 2024 / Accepted: 19 January 2024 / Published: 1 February 2024
(This article belongs to the Special Issue Probability, Statistics and Random Processes)

Abstract

:
This article is devoted to the synthesis and analysis of the quality of the statistical estimate of parameters of a multidimensional linear system (MLS) with one input and m outputs. A nontrivial case is investigated when the one-dimensional input signal of MLS is a deterministic process, the values of which are unknown nuisance parameters. The estimate is based only on observations of MLS output signals distorted by random Gaussian stationary m -dimensional noise with a known spectrum. It is assumed that the likelihood function of observations of the output signals of MLS satisfies the conditions of local asymptotic normality. The n -consistency of the estimate is established. Under the assumption of asymptotic normality of an objective function, the limiting covariance matrix of the estimate is calculated for case where the number of observations tends to infinity.

1. Introduction

The model of observations of the output MLS signals, which is used to estimate MLS parameters, has the form:
z u , t = y u , t + ξ t ,   y u , t = τ = 0 h t τ u s t ,   t 1 , n ¯ ,
where z u , t m , t 1 , n ¯ , and n > q are observations of the output MLS signals y u , t distorted by the noise ξ t ; h τ u = h k , τ u ,   k 1 , m ¯ ,   τ 0 , ¯ is the impulse response of MLS; and s t 1 , t is the input signal of MLS, whose values are unknown nuisance parameters.
The value u of the parameter of observations z u , t is unknown and belongs to the open set U ˜ of the q -dimensional vector space u U ˜ q . The functions h τ u and τ + are supposed to be known. The noise ξ t = ξ k , t ,   k 1 , m ¯ is the m-dimensional random Gaussian stationary time series with zero mean value and known complex matrix power spectral density (MPSD) F ˙ ξ λ m × m , λ π , π . We also suppose that the noise ξ t is the regular random process of the maximal rank satisfying the strong mixing condition [1,2].
In this article, we synthesized the estimate u ^ z ¯ u , n of the value u in a situation where there is no detailed prior information about the MLS input signal s t . Such problems arise in many technologies, such as radio engineering, acoustics, and seismology, when it is necessary to determine the characteristics of a physical linear medium based on the results of its sounding by propagating waves emitted by natural or man-made sources [3,4,5,6]. It is impossible in such problems to observe the medium deformations in the inner regions of the Earth’s crust, caused by the action of an unknown probing signal s t . Therefore, the characteristics of the medium can only be determined by analyzing the wave field excited by the signal s t and recorded by an array of spatially distributed sensors located at the Earth’s surface.
The problems of estimating the characteristics of the Earth’s medium that arise in seismology are, as a rule, the most complex. If the propagation of waves in the Earth’s medium can be described by a system of linear partial differential equations, then the signals recorded by an array consisting of m seismic sensors can be interpreted as output signals of an MLS with one input and m outputs. In this case, the physical characteristics of the Earth’s medium can be considered mathematically as some parameters of this MLS.
Because the input signal s t of the MLS belongs to a one-dimensional subspace 1 of the m-dimensional space of the MLS output signals, the observations z u , t m , even distorted by interferences ξ t , provide enough information about the value u q , q < m of the MLS parameters to construct a consistent statistical estimate of this value.
The problem of statistical estimation of parameters of multivariate observations, closest in formulation to our problem, was considered in [7]. This is the problem of estimating the parameter B m × m of the model of observations known as “multivariate linear functional relationship”:
z t = B s t + ξ t ,   t 1 , n ¯ ;   x t = s t + ζ t ,   t 1 , n ¯ ,
where s t m is a sequence of unknown nuisance parameters and the errors ξ t ,   ζ t are independent Gaussian vectors with a mean equal to zero and a covariance matrix equal to I . That is, the parameter B should be estimated from the observations z t ,   x t ,   t 1 , n ¯ .
As noted in [7], if a priori restrictions on an infinite sequence of nuisance parameters are not imposed, then the nonparametric model z t ,   x t ,   t 1 , n ¯ does not correspond to traditional nonparametric models, where observations belong to some metric space and for which asymptotically efficient (AE) estimates can be constructed. Nevertheless, in [7], the local asymptotically minimax estimate of the parameter B was constructed, which has been proposed in [8].
The model of observations z u , t differs significantly from the model z t ,   x t , since it does not assume observations x t = s t + ζ t , t 1 , n ¯ of the nuisance parameters s t 1 ,   t . Also, for z u , t , we have additive time series ξ t , t 1 , n ¯ , a sample from the stationary Gaussian random time series with a known MPSD F ˙ ξ λ m × m , while for z t ,   x t , ξ t ,   ζ t are independent Gaussian vectors. In addition, the estimated parameter u q is “hidden” in the impulse response h t u m of the MLS. For these reasons, it is impossible to construct a local asymptotically minimax estimate for the parameter u in our problem. Nevertheless, in this problem, it is possible to construct a n -consistent estimate of the parameter value u. This will be performed in detail in Section 3.

2. Basics of Efficiency Criteria for Statistical Estimates

In denoting z ¯ u , n = z u , t ,   t 1 , n ¯ as a criterion of quality for estimates u ^ n z ¯ u , n , we will use the mean square error (MSE) matrix
K n u ^ n = E u u ^ n u u ^ n u T
It is said that one estimate u ^ n , 1 is better than u ^ n , 2 when the corresponding MSE matrices satisfy the inequality,
K n u ^ n , 1 K n u ^ n , 2
where inequality A B means that A B is a nonnegative semi-definite matrix.
If u U ˜ q is a vector of true values of parameters for so-called regular parametric probability models of the observations, a lower boundary for matrices K n u ^ n in a class of unbiased estimators u ^ n z ¯ u , n exists; it is defined by Fisher’s information matrix J n u [9]:
K n u ^ n J n 1 u ,
where J n u = R m n u p z ¯ u , n ;   u u p z ¯ u , n ;   u T p 1 z ¯ u , n ;   u d   z ¯ u , n ;
p z ¯ u , n ;   u is the probability density of random observations z ¯ u , n ; and u p z ¯ u , n ;   u = u k p z ¯ u , n ;   u ,   k 1 , q ¯ T .
In this case, an estimate is said to be statistically effective if the following equality holds:
K n ef u = E u u ^ ef z ¯ u , n u u ^ ef z ¯ u , n u T = J n 1 u
However, effective estimates exist, even theoretically, only in some special parametric probability models of the observations, which rarely correspond to practical needs. Instead, in some cases, it is possible to construct the asymptotically efficient (AE) estimate u ^ ae z ¯ n , which has a limit error covariance matrix K ae u equal to the limit of the inverse Fisher’s information matrix J n 1 u   n [10]:
K ae u = lim n   n E u ^ ae z ¯ u , n u u ^ ae z ¯ u , n u T = lim n   n J n 1 u
For instance, in certain applications, s t ,   t 1 , n ¯ can be considered a realization of the stationary Gaussian random process s t ,   t with a zero mean and known power spectral density (PSD) g s λ . Hence, the Gaussian probability density p z ¯ u , n ;   u of the random sample z ¯ u , n , whose n m × n m -matrix covariance function depends on the parameter u , does not belong to the exponential family of distributions [11]. Therefore, in this case, there is no efficient estimate u ^ ef z ¯ u , n that has an error covariance matrix K n ef u equal to the inverse Fisher’s information matrix J n 1 u at each number of observations n, but an AE estimate exists, and its analytical form is given in [12]. However, the most common situation occurs when s t ,   t are unknown, and the observation model for z ¯ u , n becomes nonparametric and the AE estimate cannot be constructed in the sense of J n 1 u .

3. The Estimate of MLS Parameters in the Case of Unknown Input MLS Signal

In many applications, the values of the MLS input signal s t ,   t are unknown and are not observed. In this case, the estimation of u based on the sample of observations z ¯ u , n of MLS output signals becomes a statistical problem with nuisance parameters, which are the unknown values of the signal s t ,   t . This problem was studied in [13] for cases where the number of nuisance parameters does not increase with the increase in the number n of observations. In our case, the number of nuisance parameters s t ,   t 1 , n ¯ is equal to the number n of observations z ¯ u , n , and it is impossible to construct a consistent estimate of the informative parameter u without introducing some constraints on the asymptotical properties of the nuisance parameters s t ,   t 1 , n ¯   n + . Such constraints were proposed in [11]:
1.a. The signal s t has finite average power: lim n   2 n 1 t = n n s t 2 = C < .
1.b. For any n + , the signal s t ,   t 1 , n ¯ satisfies the inequality max t 1 , n ¯ s t < n β , where β 0 , 1 / 2 .
Note that almost all sample functions of stationary random processes satisfy such restrictions [14].
It is shown in [11] that, in cases where all the values of the input MLS signal s t ,   t are known and constraints 1.a and 1.b are satisfied (and also some restrictions on the MLS impulse response h τ u and the noise MPSD F ˙ ξ λ ), the likelihood function ln p z ¯ u , n ; u of observations z ¯ n admits the LAN [15] expansion in the vicinity of value u U ˜ :
ln p n z ¯ u , n ; u + n 1 / 2 w = ln p n z ¯ u , n ;   u + w T Δ n z ¯ u , n ;   u 1 2 w T Γ n u w + α n z ¯ u , n ;   u , w ,
where w < C ; C is any constant; P - l i m n α n z ¯ u , n ;   u , w = 0 ; z ¯ u , n = z u , t ,   t 1 , n ¯ ;
Δ n z ¯ u , n ; u = Δ k , n x ¯ ˙ u , n ; u , s ˙ 1 , , s ˙ n ,   k 1 , q ¯ ;   Γ n u = Γ k , l , n u , s ˙ 1 , , s ˙ n ,   k , l 1 , q ¯ ;
Δ k , n x ¯ ˙ u , n ; u , s ˙ 1 , , s ˙ n = n 1 / 2 j = 1 n x ˙ u , j h ˙ j u s ˙ j F ˙ ξ , j 1 h ˙ k , j u s ˙ j ;   x ¯ ˙ u , n   is   the   DFFT   of   z ¯ u , n ;
Γ k , l , n u , s ˙ 1 , , s ˙ n = n 1 j = 1 n h k , j u F ˙ ξ , j 1 h l , j u s ˙ j 2 ;   lim n   Γ n u = Γ u ;
h ˙ j u = h ˙ λ j ; u ;   h ˙ k , j u = h ˙ λ j ; v v k v = u ;   s ˙ j ,   j 1 , n ¯   are   the   DFFTs   of   s t ,   t 1 , n ¯ ;
and the probability distribution of the random function Δ n z ¯ u , n ;   u converges as n to the q -dimensional Gaussian distribution with the moments 0 , Γ u .
When the values s t ,   t 1 , n ¯ are unknown for any n + , it is impossible to consistently estimate all unknown parameters of the observations z ¯ u , n (i.e., the MLS parameter value u together with the nuisance parameters s t , t 1 , n ¯ ) using only the main terms (2) of the LAN expansion (1) of the likelihood function of the observations. But in our problem, it is necessary to estimate only the MLS parameter value u , and it is not necessary to estimate the nuisance parameters s t ,     t 1 , n ¯ . The approach to solving such an unconventional estimation problem was first proposed in [16]. With some modifications, this approach was implemented in [12] to solve our problem. The method proposed in these publications was as follows:
It is easy to show that if the values s t ,     t 1 , n ¯ are known, the family of statistics Δ n x ¯ ˙ u , n ; u = Δ k , n x ¯ ˙ u , n ; u ,   k 1 , q ¯ in Equation (2) is the gradient of the function ln p ˜ n x ¯ ˙ n ;   v , s ˙ 1 , , s ˙ n with respect to parameters v k ,     k 0 , k ¯ :
Δ n x ¯ ˙ u , n ;   u = n 1 / 2 g r a d v U ln p ˜ n x ¯ ˙ u , n ; v , s ˙ 1 , , s ˙ n v = u ,
where
ln p ˜ n x ¯ ˙ u , n ; v , s ˙ 1 , , s ˙ n = ln j = 1 n 2 π det F ˙ ξ , j 1 / 2 exp x ˙ u , j h ˙ j v s ˙ j * F ˙ ξ , j 1 x ˙ u , j h ˙ j v s ˙ j .
Note that the function p ˜ n x ¯ ˙ u , n ; v , s ˙ 1 , , s ˙ n in Equation (3) is the asymptotic approximation of the probability density for spectral observations of x ˙ u , j ,     j 1 , n ¯ , which are asymptotically mutually independent [1].
Hence, the AE estimate u ^ n ae x ¯ ˙ u , n of parameters u in cases of known s t ,   t 1 , n ¯ can be obtained as the root of equation Δ n x ¯ ˙ u , n ;   v = 0 and has the form:
u ^ n ae x ¯ ˙ u , n = arg max v U j = 1 n x ˙ u , j h ˙ j v s ˙ j * F ˙ ξ , j 1 x ˙ u , j h ˙ j v s ˙ j .
When the values of s ˙ j ,   j 1 , n ¯ are unknown, one can, according to [14], construct some estimate of the value u of the informative parameter together with the unknown nuisance parameters s ˙ j ,   j 1 , n ¯ by formally applying the maximum likelihood approach to the right side of Equation (4). That is, this estimate can be obtained by maximizing the right side of Equation (4) through all unknown parameters φ = v 1 , , v q , s ˙ 1 , , s ˙ n V , where V is a bounded set of q + n :
φ ^ x ¯ ˙ u , n = arg max φ V   ln p ˜ n x ¯ ˙ u , n ; φ = arg max φ V j = 1 n x ˙ u , j h ˙ j v s ˙ j * F ˙ ξ , j 1 x ˙ u , j h ˙ j v s ˙ j .
Estimates (5) can be calculated by solving the following system of equations:
  2 .       v k ln   p ˜ n x ¯ ˙ u , n ; φ = 0 ; k 1 , q ¯ . 1 .   s r r e ln p ˜ n x ¯ ˙ u , n ; φ = 0 ; s r i m ln p ˜ n x ¯ ˙ u , n ; φ = 0 .
By representing a positively definite Hermitian matrix F ˙ ξ , j 1 in the form F ˙ ξ , j 1 = F ˙ ξ , j 1 / 2 F ˙ ξ , j 1 / 2 , we obtain:
ln p ˜ n x ¯ ˙ u , n ; φ = j = 1 n ρ ˙ u , j d ˙ j v s ˙ j 2 ,
where ρ ˙ u , j = F ˙ ξ , j 1 / 2 x ˙ u , j ;   d ˙ j v = F ˙ ξ , j 1 / 2 h ˙ j v .
Then, subsystem (6.1) can be written in the form:
2     r e d ˙ j v ρ ˙ u , j d ˙ j v 2 s ˙ j r e = 0 . 2   i m d ˙ j v ρ ˙ u , j d ˙ j v 2 s ˙ j i m = 0 .
The system of Equation (8) has the following solution:
s ˜ ˙ j x ˙ u , j ; v = d ˙ j v ρ ˙ u , j d ˙ j v 2 ;       j 1 , n ¯ .
To construct an estimate of the parameter value u , it is necessary to substitute the estimate s ˜ j x ˙ u , j ; v according to Equation (9) into subsystem (6.2):
v k ln   p x ¯ ˙ u , n ; φ = v k j = 1 n ρ ˙ u , j d ˙ j v d ˙ j v d ˙ j v 2 ρ ˙ u , j 2 = = v k j = 1 n I d ˙ j v d ˙ j v d ˙ j v 2 ρ ˙ u , j 2 = 0 ;   k 1 , q ¯ .
Thus, we define some estimate u ˜ n δ x ¯ ˙ n of the value u of the MLS parameter as a solution to the nonlinear system of equations:
v k j = 1 n   Π ˙ j v ρ ˙ u , j 2 = v k j = 1 n ρ ˙ u , j Π ˙ j v Π ˙ j v ρ ˙ u , j = 0 ;       k 1 , q ¯ ,     v U ,
where Π ˙ j v = I d ˙ j v d ˙ j v d ˙ j v 2 .
It is easy to check that the matrices Π ˙ j v are the idempotent Hermitian matrices, i.e., Π ˙ j * v Π ˙ j v = Π ˙ j v for all v U and j 1 , n ¯ . Hence, Equation (11) can be rewritten in the following form:
j = 1 n   ρ ˙ u , j v k Π ˙ j v ρ ˙ u , j = j = 1 n   ρ ˙ u , j Π ˙ k , j v ρ ˙ u , j = 0 ;       k 1 , q ¯ .
Therefore, the estimate of the value u of the MLS parameter at which the observations ρ ˙ u , j = F ˙ ξ , j 1 / 2 x ˙ u , j were obtained can be found by maximizing the objective function
Q x ¯ ˙ u , n ; v = j = 1 n ρ ˙ u , j Π ˙ j v ρ ˙ u , j ,
that is,
u ^ n δ x ¯ ˙ u , n = arg max v U Q x ¯ ˙ u , n ; v .
According to the definitions of quantities ρ ˙ u , j ,       Π ˙ j v , we can write objective Function (13) as:
Q x ¯ ˙ u , n ; v = j = 1 n x ˙ u , j Λ ˙ λ j ; v x ˙ u , j = j = 1 n h ˙ j * v F ˙ ξ , j 1 x ˙ u , j 2 h ˙ j * v F ˙ ξ , j 1 h ˙ j v ,
where Λ ˙ λ ; v = F ˙ ξ 1 λ h ˙ λ ; v h ˙ λ ; v F ˙ ξ 1 λ h ˙ j * λ ; v F ˙ ξ 1 λ h ˙ λ ; v ; x ˙ u , j = h ˙ λ j ; u s ˙ λ j + ξ ˙ λ j , in which u is the value of parameter v U under which the sample x ¯ ˙ u , n = x ˙ u , j ,   j 1 , n ¯ (DFFT of the sample z ¯ u , n = z u , j ,   j 1 , n ¯ ) was obtained; and λ j = 2 π j n 1 .
It is important to note that estimate (14) does not depend on the unknown values of the nuisance parameters, i.e., the input MLS signals s t ,     t 1 , n ¯ . For calculating estimate (14), we must process only the spectral observations x ˙ u , j = y ˙ u , j + ξ ˙ j ,     j 1 , n ¯ at the output of MLS.
In what follows, we will take into account the assumptions under which estimate (14) was obtained in [12]:
A. The MLS frequency response h ˙ λ ; v has second partial derivatives in the components of the vector v , and these derivatives are continuous in v U q , λ π , π :
h k , l λ ; v = 2 v k v l h λ ; v ,   k , l 1 , q ¯ ,   h k , l λ ; v < c   for   v U ,   λ π , π .
B.  det F ˙ ξ λ > 0 .
Under assumptions A and B, the matrix functions
Λ ˙ k λ ; u = Λ ˙ λ ; u u k ,   Λ ˙ k , l u = 2 Λ ˙ λ ; u u k u l
exist, are uniformly bounded in norm, and are continuous in v U , λ π , π for all k , l 1 , q ¯ .
Let us note an equation that will be needed later. From (11) and (15), it follows that
Π ˙ k , j u = F ˙ 1 / 2 λ j Λ ˙ k , j λ j ; u F ˙ 1 / 2 λ j
Assumptions A and B may have a physical explanation. The considered MLS model arises in tasks such as wave field source localization in acoustics, radio, or slowness vector estimation in seismology, where the vector function h ˙ λ ; v describes wave propagation. As a rule, such functions are sufficiently smoothed to have second and even third partial derivatives in the components of the vector v . Assumption B is related to the definition of regularity [1] of additive noise, which always holds for stationary Gaussian processes.
Estimate (14) belongs to the class of M-estimates. M-estimates may have the property of robustness, i.e., their accuracy depends less on changes in the probability distribution of observations, unlike AE-estimates [17]. For this reason, M-estimators are used in many applications of mathematical statistics in the natural sciences and in econometrics for statistical estimation problems where complete probabilistic models of observations are not known [10,18]. In these problems, the estimates are found by maximizing some objective function Q n x ¯ ˙ u , n ;   v , v U instead of the likelihood function. Despite the fact that likelihood-based estimate (14) is obtained from a parametric observation model, taking into account an infinite number of estimated parameters φ , it is not an AE estimate but could still be robust.
The asymptotic properties of M-estimates were studied in [10] for the problem of estimating the distribution parameter of a one-dimensional random variable from a sample of independent observations of this variable, while determining the probability distribution of estimate (14) is a rather difficult task. This task will be simplified if we consider the equivalent problem of determining the probability distribution of the root u ^ n δ x ¯ ˙ u , n of the equation g r a d v U Q n x ¯ ˙ u , n ;   v =   0 .
The analysis of the asymptotic statistical properties of the random estimate u ^ n δ x ¯ ˙ u , n when n is based on the following theorem proved in [19], Theorem 1 and Corollary 1.
Theorem 1. 
Let an objective function  Q n x ¯ ˙ u , n ;   v , v U q  satisfy the following conditions:
A. The objective function  Q n x ¯ ˙ u , n ;   v ,   v U  admits the following asymptotic expansion in a small vicinity of the value  u U ˜ int U :
Q n x ¯ ˙ u , n ;   u + n 1 / 2 w = Q n x ¯ ˙ u , n ;   u + n 1 / 2 δ n x ¯ ˙ u , n ; u w 2 n 1 w T Φ n u w + β n x ¯ ˙ u , n ; w ,   w < C ,
where n 1 / 2 δ n x ¯ ˙ u , n ; u = n 1 / 2 δ k , n x ¯ ˙ u , n ; u = w k Q n x ¯ ˙ u , n ;   u + n 1 / 2 w w = 0 ,   k 1 , q ¯ ;
n 1 Φ n x ¯ ˙ u , n ;   u = n 1 Φ k , l , n x ¯ ˙ u , n ;   u = 2 w k w l Q n x ¯ ˙ u , n ;   u + n 1 / 2 w w = 0 ,   k , l 1 , q ¯ ;
P n , u - lim n   n 1 Φ n x ¯ ˙ u , n ; u = Φ u ;   inf u U   d e t   Φ u > C > 0 ;   Φ 1 u < C ;
P - lim n   β n x ¯ ˙ u , n ; w = 0 .
B. The vector statistic  n 1 / 2 δ n x ¯ ˙ u , n ;   u  has a Gaussian limiting distribution for  n  with the following moments:
lim n   E n 1 / 2 δ n x ¯ ˙ u , n ; u = 0 ;   lim n   E n 1 δ n x ¯ ˙ u , n ; u δ n T x ¯ ˙ u , n ; u = Ψ u ;   Ψ u < C .
C. Let some statistic  u ^ n δ x ¯ ˙ u , n q  have the following properties:
a. The statistic  u ^ n δ x ¯ ˙ u , n  is a solution of the system of equations  δ n x ¯ ˙ u , n ; v = 0  for any  n > m  and almost each  x ¯ ˙ u , n  with respect to the probability distribution  d P u , n x ¯ ˙ u , n :
P u , n δ n x ¯ ˙ u , n ; u ^ n δ x ¯ ˙ u , n = 0 1 .
b. The statistic  u ^ n δ x ¯ ˙ u , n  is the  n -consistent estimate of the value  u U ˜ .
Then, the random variable  ζ n = n u ^ n δ x ¯ ˙ u , n u  has a Gaussian distribution in the asymptotic  n  with the following moments:
lim n   E ζ u , n = 0 ;   l i m n   E ζ u , n ζ u , n T = D u = Φ 1 u Ψ u Φ 1 u ;   D   ( u ) < C ,   u U ˜ int U
.
It has been proven in Section 5 that objective Function (13) satisfies the conditions A and C of Theorem 1. The proof of the asymptotic normality of n 1 / 2 δ n x ¯ ˙ u , n ;   u , which is part of condition B, is quite challenging (it will be discussed in the end), but we will assume that it holds. Then, the random variable ζ u , n = n u ˜ n δ x ¯ ˙ u , n u has Gaussian distribution in the asymptotic n with the following moments:
lim n   E ζ u , n = 0 ; lim n   E ζ u , n ζ u , n T = D u = Φ 1 u Ψ u ¯ Φ 1 u ,
where Φ u = Φ k , l u ;   k , l 1 , q ¯ ; Ψ u = Ψ k , l u ;   k , l 1 , q ¯ ; Φ k , l u = 0 2 π tr Λ k , l λ ; u F ξ λ d λ ; Ψ k , l u = 2 π π tr Λ k λ ; u F ˙ ξ λ Λ ˙ l λ ; u F ˙ ξ λ d λ  +  4 π π h ˙ * λ ; u Λ ˙ k λ ; u F ˙ ξ λ Λ ˙ l λ ; u   h ˙ λ ; u d w s λ ; Λ k λ ; u = v k Λ λ ; v v = u ; Λ k , l λ ; u = 2 v k v l Λ λ ; v v = u ; Λ ˙ λ ; v = F ˙ ξ 1 λ F ˙ ξ 1 λ h ˙ λ ; v h ˙ * λ ; v F ˙ ξ 1 λ h ˙ j * λ ; v F ˙ ξ 1 λ h ˙ λ ; v   v U ; w s λ = lim n   n 1 j = 1 ν λ s ˙ 2 π j n 1 2 ; and ν λ = λ is the largest integer j for which 2 π j n 1 λ .

4. Numerical Comparison of u ^ n δ x ¯ ˙ u , n and the SRP-PHAT Estimator

As mentioned in the introduction, the particular case of MLS parameter estimation problem is wavefield source localization by antenna arrays. The SRP-PHAT estimator of source coordinates is the most popular, well-recommended, and robust method, as described in [3,6]. In order to demonstrate the effectiveness of u ^ n δ x ¯ ˙ u , n given by (14) compared to SRP-PHAT, Monte-Carlo simulations for two different cases of noise ξ t properties were performed in [4]. A set of 150 MLS outputs was considered in this experiment. For the known value of the estimated parameter (x*, y*), a set of 110 mixtures of 150 MLS outputs and time series ξ t was simulated.
A simple and known metric was used to numerically compare the effectiveness of the algorithms:
ρ R M S E v ^ n = 1 p j = 1 p i = 1 q v i j v i * 2 1 / 2
where v i * , i = 1 , q ¯ is the known value of the unknown parameter v = v 1 , , v q and v i j , i = 1 , q ¯ , j = 1 , p ¯ is the set of independent estimated values of the parameter v = v 1 , , v q . In the current modeling, q = 2 and p = 110. Two different sets of v i j , i = 1 , q ¯ , j = 1 , p ¯ were obtained by two different estimates: u ^ n δ x ¯ ˙ u , n and SRP-PHAT.
Below, in Figure 1, the empirical two-dimensional probability density functions for averaged (among MLS outputs) signal-to-noise ratio SNR = 0.05 are obtained for 110 independent estimates of source coordinates v = v 1 , v 2 in the presence of real correlated noise with matrix power spectral density F ˙ ξ λ . In Figure 2, similar probability density functions are provided for F ˙ ξ λ = c o n s t * I , where I is the unit matrix. That means that ξ t is represented as multidimensional white Gaussian noise with equal power spectral densities.
As can be seen from Figure 1 and Figure 2, in both cases, the ρ R M S E value of u ^ n δ x ¯ ˙ u , n is approximately two times greater than the ρ R M S E value of SRP-PHAT. That is, the changing properties of additive noise ξ t lead estimate u ^ n δ x ¯ ˙ u , n to be more efficient than SRP-PHAT. But while varying the additive noise properties, both estimators show nondramatic changes in estimation accuracy, so they are both robust. More detailed information and a straight numerical comparison are given in [4].

5. Asymptotic Properties of the Objective Function Q n x ¯ ˙ u , n ; v

The proof that the objective function Q n x ¯ ˙ u , n ;   v satisfies conditions A and C of Theorem 1 consists of a sequence of lemmas.
Lemma 1.
The objective function, Q n x ¯ ˙ u , n ;   v , v U , satisfies conditions A  of Theorem 1.
Proof. 
Let us write the Tailor expansion of the objective function Q n x ¯ ˙ u , n ;   u + n 1 / 2 w in the vicinity n 1 / 2 w , w < C of the parameter value u with the remainder term in the Lagrange form:
j = 1 n   ρ ˙ u , j * Π ˙ j u + n 1 / 2 w ρ ˙ u , j = j = 1 n   ρ ˙ u , j * Π ˙ j u ρ ˙ u , j + n 1 / 2 j = 1 n   ρ ˙ u , j * k = 1 q Π ˙ k , j u w k ρ ˙ u , j 2 n 1 j = 1 n   ρ ˙ u , j * k , l = 1 q Π ˙ k , l , j u w k w l ρ ˙ u , j + β n ρ ¯ ˙ u , n ;   u , w , ρ ¯ ˙ u , n = ρ ˙ u , j   j = 1 , n ¯ , w < C
where n 1 / 2 Π ˙ k , j u = w k Π ˙ j u + n 1 / 2 w w = 0 ; n 1 Π ˙ k , l , j u = 2 w k w l Π ˙ j u + n 1 / 2 w w = 0 ;
ρ ˙ u , j = d ˙ j u s ˙ j + η ˙ j .
The remainder term β n ρ ¯ ˙ u , n ;   w on the right side of Equation (19) has the form:
β n x ¯ ˙ u , n ;   w = n 3 / 2 k , l , r q w k w l w r j = 1 n   ρ ˙ u , j * Π ˙ k , l , r , j u + n 1 / 2 θ k , l , r , j w ρ ˙ u , j ,
where w < C , θ k , l , r , j 0 , 1 .
We use the Lagrange form of the remainder term β n x ¯ ˙ u , n ;   w above because it is the simplest form of the Tailor expansion of the objective function Q n x ¯ ˙ u , n ; v by Equation (13) given on the closed set v = U , but this form requires the existence of third bounded partial derivatives of the objective function Q n x ¯ ˙ u , n ; v . Let us assume that this is true for the sake of simplicity. Due to this assumption, there exists a matrix function:
Π ˙ k , l , r , j v = F ˙ 1 / 2 λ j Λ k , l , r , j v F ˙ 1 / 2 λ j ,
where Λ ˙ j v = Λ ˙ λ j ; v ; λ j = 2 π j n 1 ; Λ ˙ λ ; v = F ˙ ξ 1 λ h ˙ λ ; v h ˙ λ ; v F ˙ ξ 1 λ h ˙ j * λ ; v F ˙ ξ 1 λ h ˙ λ ; v , v U , which is continuous on v U and is bounded for all valid values of its arguments.
In accordance with Equation (7), we have:
ρ ˙ u , j = F ˙ ξ , j 1 / 2 x ˙ u , j = d ˙ j u s ˙ j + η ˙ j ,   where   d ˙ j v = F ˙ ξ , j 1 / 2 h ˙ j v .
The term d ˙ j u s ˙ j is bounded for all j 1 , n ¯ , u U ˜ U . The pairs of random vectors η ˙ j ,     η ˙ k , j k 1 , n ¯ are mutually asymptotically independent for n ; the Gaussian random vectors η ˙ j have asymptotic covariance matrices equal to I for n . That is, we have (see Appendix A):
E η ˙ j = 0 ,   j 1 , n ¯ ;   E η ˙ j η ˙ k = F ˙ j 1 / 2 E ξ ˙ j ξ ˙ k F ˙ k 1 / 2 = O ˙ j , k η ,   j k 1 , n ¯ ;
E η ˙ j η ˙ j = F ˙ j 1 / 2 E ξ ˙ j ξ ˙ j F ˙ j 1 / 2 = F ˙ j 1 / 2 F ˙ j + O ˙ j ξ F ˙ j 1 / 2 = I + O ˙ j η ,
where O ˙ j η C n 1 β , β 0 , 1 for all j 1 , n ¯ .
Taking into account Equation (21) and the boundedness of the terms w k w l w r in Equation (20), we conclude that the sequence of random variables β n ρ ¯ ˙ u , n ;   w tends to zero in probability as n , in accordance with the Law of Large Numbers [18] (Lemma 2.4):
P lim n   n 3 / 2 j = 1 n   k , l , r q   w k w l w r ρ ˙ x ˙ u , j * Π ˙ k , l , r , j u + n 1 / 2 θ k , l , r , j w ρ ˙ x ˙ u , j = 0 .
Let us assume that the vector statistic
n 1 / 2 δ n x ¯ ˙ u , n ;   u = n 1 / 2 δ k , n x ¯ ˙ u , n ;   u ,   k 1 , q ¯ ,
where n 1 / 2 δ k , n x ¯ ˙ u , n ;   u = n 1 / 2 j = 1 n   ρ ˙ u , j * Π ˙ k , j v v = u ρ ˙ u , j ,   k = 1 , q ¯ =
= n 1 / 2 j = 1 n d ˙ j u s ˙ j + η ˙ j Π ˙ k , j u d ˙ j u s ˙ j + η ˙ j ,
particularly satisfies conditions B of Theorem 1. That is, n 1 / 2 δ n x ¯ ˙ u , n ;   u has a Gaussian limit distribution. Let us prove that the moments of this distribution are
lim n   E n 1 / 2 δ n x ¯ ˙ u , n ; u = 0 ;   lim n   E n 1 δ n x ¯ ˙ u , n ; u δ n T x ¯ ˙ u , n ; u = Ψ u ;   Ψ u < C .
Lemma 2.
The statistics given by (24):
n 1 / 2 δ k , n x ¯ ˙ u , n ; u = n 1 / 2 j = 1 n d ˙ j * u Π ˙ k , j u d ˙ j u s j 2 + n 1 / 2 j = 1 n η ˙ j Π ˙ k , j u η ˙ j + + 2 n 1 / 2 j = 1 n Re η ˙ j Π ˙ k , j u d ˙ j u s ˙ j ,   k = 1 , q ¯
have zero mean in the asymptotic  n :
lim n   E n 1 / 2 δ k x ¯ ˙ u , n ;   u = 0 ,   k 1 , q ¯ .
Proof of Lemma 2.
The term Π ˙ k , j u has the following form:
Π ˙ k , j u = v k d ˙ v d ˙ v d v 2 = d v 4 d ˙ v d ˙ v v k d v d v d ˙ v d ˙ v d v d v v k v = u = = d u 4 d ˙ k , u d ˙ u d ˙ u 2 + d u d k , u d ˙ u 2 d ˙ u d ˙ u d k , u d u d ˙ u d ˙ u d u d k , u ,
where u U ˜ ,     j 1 , n ¯ ,     k 1 , q ¯ .
Statement 1.
The first term on the right side of Equation (25) is equal to zero:
n 1 / 2 j = 1 n d ˙ j * u Π ˙ k , j v d ˙ j u s j 2 = 0   for   all   k 1 , q ¯ .
Proof of Statement 1.
Let us rewrite the terms d ˙ j * u Π ˙ k , j u d ˙ j u for any j 1 , n ¯ ,     n > q , k 1 , q ¯ using the associative law for products of several vectors:
d ˙ j * u Π ˙ k , j u d ˙ j u = d u 4 d ˙ u 2 d ˙ u d ˙ k , u d ˙ u d ˙ u + d ˙ u d u d ˙ k , u d ˙ u d ˙ u 2 + + d u 4 d ˙ u d ˙ u d ˙ u d ˙ u d u d ˙ k , u + d ˙ u d ˙ u d ˙ u d ˙ u d u d ˙ k , u = = d u 4 d u 2 d u d ˙ k , u d u 2 + d u 2 d ˙ k , u d u d u 2 + + d u 4 d u 2 d u 2 d u d k , u + d u 2 d u 2 d u d ˙ k , u 0   for   any   j 1 , n ¯ ,   n > q ,   k 1 , q ¯ .
The mathematical expectation of the last term on the right side of Equation (25) is also equal to zero due to Equation (22): 2 n 1 / 2 j = 1 n Re E η ˙ j Π ˙ k , j u d ˙ j u s ˙ j = 0 .
Statement 2.
The limit of the mathematical expectation of the second term on the right side of Equation (25) is equal to zero:
n 1 / 2 lim n   j = 1 n E η ˙ j Π ˙ k , j u η ˙ j = n 1 / 2 lim n   j = 1 n t r Π ˙ k , j u E η ˙ j η ˙ j = 0 .
Proof of Statement 2.
As follows from Equation (22):
t r Π ˙ k , j u E η ˙ j η ˙ j = t r Π ˙ k , j u + t r Π ˙ k , j u O ˙ j η ,
where O ˙ j η C n 1 β , β 0 , 1 for all j 1 , n ¯ .
Let us first prove that t r Π ˙ k , j u = 0 for any u U ˜ ,     k 1 , q ¯ ;     j 1 , n ¯ ,     n > q . The following equalities are held:
tr Π ˙ j u = t r I d ˙ j u d ˙ j u 2 d ˙ j * u = m d ˙ j u 2 tr d j u d j * u = = m d ˙ j u 2 l = 1 m d ˙ l , j u d ˙ l , j * u = m d ˙ j u 2 d ˙ j u 2 = m 1 .
Therefore, t r Π ˙ k , j u = v k t r Π ˙ j v v = u 0 for any u U ˜ ,     k 1 , q ¯ ;     j 1 , n ¯ ,     n > q . Consequently:
n 1 / 2 j = 1 n t r Π ˙ k , j u E η ˙ j η ˙ j = n 1 / 2 j = 1 n v k tr Π ˙ j v v = u O ˙ j η ,
where O ˙ j η C n 1 β , β 0 , 1 .
The matrix functions Π ˙ j v have continuous partial derivatives with respect to v k for any v U , j 1 , n ¯ . Consequently, Π ˙ k , j v C 1 for any v U , j 1 , n ¯ . As follows from Equation (28):
lim n   n 1 / 2 s u p v U j = 1 n tr Π ˙ k , j v O ˙ j η n 1 β C n 1 / 2 β = 0 .
Finally, we deduce from (26)–(29) that
lim n   n 1 / 2 j = 1 n E η ˙ j   Π ˙ k , j v η ˙ j = 0 .
□ □
It follows from Lemmas 1 and 2 that the components of the family of statistics n 1 / 2 δ x ¯ ˙ u , n ; v ,     v U in the limit n have a zero mathematical expectation:
lim n   E n 1 / 2 δ k x ¯ ˙ u , n ; u = lim n   E n 1 / 2 j = 1 n η ˙ u , j Π ˙ k , j u η ˙ u , j + 2 n 1 / 2 j = 1 n Re η ˙ u , j Π ˙ k , j u d ˙ j u s ˙ j = 0 .
Lemma 3.
The covariance matrix Ψ u of the statistic n 1 / 2 δ x ¯ ˙ u , n ; u has the limit
l i m n E n 1 δ x ¯ ˙ u , n ; u δ T x ¯ ˙ u , n ; u = Ψ u = Ψ k , l u ,   k , l 1 , q ¯ ,   Ψ u < C < ,
where  Ψ k , l u = 2 π π tr Λ ˙ k λ ; u F ˙ ξ λ Λ ˙ l λ ; u F ˙ ξ λ d λ  +  4 π π h ˙ * λ ; u Λ ˙ k λ ; u F ˙ ξ λ Λ ˙ l λ ; u   h ˙ λ ; u d w s λ ; k , l 0 , q ¯ , w s λ = lim n   n 1 j = 1 ν λ s ˙ 2 π j n 1 2 ; ν λ = λ  is the largest integer n for which  2 π j n 1 λ ; Λ ˙ λ ; v = F ˙ ξ 1 λ h ˙ λ ; v h ˙ λ ; v F ˙ ξ 1 λ h ˙ j * λ ; v F ˙ ξ 1 λ h ˙ λ ; v ; Λ ˙ k λ ; u = v k Λ λ ; v v = u .
Proof of Lemma 3.
According to Lemma 2, lim n   E n 1 / δ k x ¯ ˙ u , n ; u = 0 , k 1 , q ¯ . Consequently,
Ψ k , l u = lim n   E n 1 δ n , k x ¯ ˙ u , n ; u   δ n , l x ¯ ˙ u , n ; u .
According to Equation (25) and Statement S.1, we have:
n 1 / 2 δ x ¯ ˙ u , n ; u = n 1 / 2 j = 1 n η ˙ j   Π ˙ k , j u η ˙ j + 2 Re η ˙ j Π ˙ k , j u d ˙ j u s ˙ j .
Therefore, Ψ k , l u can be written as:
Ψ k , l u = lim n   n 1 i = 1 n j = 1 n E Q k , i Q l , j + Q k , i L l , j + L k , i Q l , j + L k , i L l , j ,
where Q k , j = η ˙ j   Π ˙ k , j u η ˙ j ; L k , j = 2 Re η ˙ i Π ˙ k , j u d ˙ j u s ˙ j .
The double sum on the right side of Equation (32) can be calculated as:
Ψ k , l u = l i m n n 1 j = 1 n E Q k , j Q l , j + Q k , j L l , j + L k , j Q l , j + L k , j L l , j + + lim n   n 1 i = 1 n j i = 1 n E Q k , i Q l , j + Q k , i L l , j + L k , i Q l , j + L k , i L l , j = A k , l + B k , l .
According to (23), the complex vectors η ˙ j ,     j 1 , n ¯ become mutually independent for large values of n:
E η ˙ i η ˙ j = O ˙ i , j η n 1 β , where O ˙ i , j η n 1 β C ¯ n 1 β , β 0 , 1 . It is easy to show that the term B k , l in Equation (33) has the form:
B k , l = lim n   n 1 i = 1 n E Q k , i j i = 1 n E Q l , j + + lim n   n 1 i = 1 n j i = 1 n E Q k , i E L l , j + E L k , i E Q l , j + E L k , i E L l , j .
The first limit on the right side of (34) is equal to zero due to Statement S1:
lim n   n 1 i = 1 n E Q k , i j i 1 , n ¯ E Q l , j l i m n n 1 / 2 i = 1 n E Q k , i l i m n n 1 / 2 j = 1 n E Q l , j = 0 .
The second limit on the right side of (34) is also equal to zero since E η ˙ j = 0 . Hence,
E L k , j = 2 Re E η ˙ j Π ˙ k , j u d ˙ j u s ˙ j = 0 ,   k 1 , q ¯ ,   j .
As a result, we obtain:
Ψ k , l u = l i m n n 1 j = 1 n E Q k , j L l , j + E Q l , j L k , j + l i m n n 1 j = 1 n E Q k , j Q l , j + E L k , j L l , j .
The quantities E Q k , j L l , j = E η ˙ j   Π ˙ k , j u η ˙ j × 2 Re η ˙ j Π ˙ l , j u d ˙ j u s ˙ j , k , l 1 , q ¯ are the sums of terms having the forms E η s , j η r , j η w , j a s , r , j b w , j . That is, they are the sums of the products of three random variables having Gaussian distributions with zero mean values. The factors in these products belong to the sets η k , j re ,   k 1 , q ¯ , η k , j im ,   k 1 , q ¯ . Due to the properties of Gaussian distributions, the mathematical expectations of these products are equal to zero.
Consequently,
l i m n n 1 j = 1 n E Q k , j L l , j = 0 ,   j 1 , n ¯ ,   k , l 1 , q ¯ .
Thus, we finally obtain
Ψ k , l u = l i m n n 1 j = 1 n E Q k , j Q l , j + E L k , j L l , j = = lim n   n 1 j = 1 n E η ˙ j   Π ˙ k , j u η ˙ j η ˙ j   Π ˙ l , j u η ˙ j + + 4 lim n   n 1 j = 1 n E Re η ˙ j Π ˙ k , j u d ˙ j u s ˙ j Re η ˙ j Π ˙ l , j u d ˙ j u s ˙ j .
To find the expression for the mathematical expectation E η ˙ j   Π ˙ k , j u η ˙ j η ˙ j   Π ˙ l , j u η ˙ j , we can use the following theorem proven in [20] (Theorem 5.2c, p. 109):
Theorem 2.
If the random vector  y ˙  has the complex Gaussian distribution  N C 0 , Κ ˙ , then the covariance of the product of quadratic forms  y ˙ T A ˙ y ˙ y ˙ T B ˙ y ˙ , where A ˙  u  B ˙  are Hermitian matrices, is determined by the following equation: E ( y T A ˙ y ) ( y T B ˙ y ) = 2 tr [ A ˙ Κ ˙ B ˙ Κ ˙ ] .
Using this theorem when y ˙ = η ˙ N C 0 , Κ ˙ = N 0 , I , A ˙ = Π ˙ k , j , B ˙ = Π ˙ l , j , we obtain:
E η ˙ j   Π ˙ k , j u η ˙ j η ˙ j   Π ˙ l , j u η ˙ j = 2 tr Π ˙ k , j u Π ˙ l , j u .
Since Π ˙ k , j u = F ˙   1 / 2 λ j Λ ˙ k , j λ j ; u F ˙ 1 / 2 λ j , the first term on the right side of (37) takes the following form:
lim n   n 1 j = 1 n E η ˙ j   Π ˙ k , j u η ˙ j η ˙ j   Π ˙ l , j u η ˙ j = = lim n   2 n 1 j = 1 n tr F ˙ 1 / 2 λ j Λ ˙ k , j λ j ; u F ˙ 1 / 2 λ j F ˙ 1 / 2 λ j Λ ˙ l , j λ j ; u F ˙ 1 / 2 λ j = = 2 0 2 π tr Λ ˙ k λ ; u F ˙ λ Λ ˙ l λ ; u F ˙ λ d λ .
Let us find an expression for the last term on the right side of Equation (36):
4 lim n   n 1 j = 1 n E Re η ˙ j Π ˙ k , j u d ˙ j u s ˙ j Re η ˙ j Π ˙ l , j u d ˙ j u s ˙ j .
Since Π ˙ k , j u ,     k 1 , q ¯ are Hermitian matrices, then the terms of sum (39) can be rewritten as E Re a ˙ C ˙ b ˙ Re a ˙ B b ˙ , where a ˙ = η ˙ j ; b ˙ = d ˙ j u s ˙ j ; C ˙ = Π ˙ k , j u ; B ˙ = Π ˙ l , j u . The following equalities are valid for the Hermitian matrices C ˙ , B ˙ and for the vectors a ˙ ,   b ˙ :
2 Re a ˙ C ˙ b ˙ 2 Re a ˙ B b ˙ = a ˙ C ˙ b ˙ + a ˙ C ˙ b ˙ a ˙ B ˙ b ˙ + a ˙ B ˙ b ˙ = a ˙ C ˙ b ˙ + b ˙ C ˙ a ˙ a ˙ B b ˙ + b ˙ B a ˙ = = a ˙ p + p ˙ a ˙ a ˙ q ˙ + q ˙ a ˙ = 2 a ˙ p 2 q ˙ a ˙ = 4 a ˙ C ˙ b ˙ b ˙ B ˙ a ˙ ,
where p ˙ = C ˙ b ˙ ,     q ˙ = B ˙ b ˙ and a ˙ p = p ˙ a ˙ due to the property of the scalar product, and C ˙ = C ˙ ,     B ˙ = B ˙ due to the property of Hermitian matrices. Since η ˙ j = F ˙ ξ , j 1 / 2 ξ ˙ j ;       d ˙ j u = F ˙ ξ , j 1 / 2 h ˙ j λ j ; u , we can write:
E 2 R e s ˙ j + h ˙ j * u Λ ˙ k , j u ξ ˙ j 2 R e s ˙ j + h ˙ j * u Λ ˙ l , j u ξ ˙ j = = 4 h ˙ j * u Λ ˙ l , j u E ξ ˙ j ξ ˙ j * Λ ˙ k , j u h ˙ j u s ˙ j 2 = = 4 h ˙ j * u Λ ˙ l , j u F ˙   j + O ˙ j n 1 β Λ ˙ k , j u h ˙ j u s ˙ j 2 ,
where O ˙ j n 1 β C n 1 β , β 0 , 1 , and C is some constant.
Finally, we obtain:
Ψ k , l u = lim n   2 n 1 j = 1 n tr Λ ˙ k , j u F ˙ ξ , j Λ ˙ l , j u F ˙ ξ , j   + +   lim n   4 n 1 j = 1 n h ˙ j * u Λ ˙ k , j u F ˙ ξ , j Λ ˙ l , j u h ˙ j u s j 2 + lim n   n 1 j = 1 n O j n 1 β ,
where β 0 , 1 and lim n   n 1 j = 1 n O j n 1 β = 0 .
Equation (36) can then be written in integral form:
Ψ k , l u = 2 0 2 π tr Λ ˙ k λ ; u F ˙ ξ λ Λ l λ ; u F ˙ ξ λ d λ + + 4 0 2 π h ˙ * λ ; u Λ ˙ k λ ; u F ˙ ξ λ Λ ˙ l λ ; u   h ˙ λ ; u d w s λ ,
where Λ r λ ; u = v r F ˙   ξ 1 λ h ˙ λ ; v h ˙ λ ; v F ˙ ξ 1 λ h ˙ λ ; v F ˙   ξ 1 λ h ˙ λ ; v v = u ; w s λ = lim n   n 1 j = 1 ν λ s 2 π j / n 2 ; ν λ = λ is the largest integer less than λ . □
Now it is not difficult to write an analytical expression for the elements of matrix Φ u = lim n   Φ n u . According to expression (17), we have
Φ k , l , n u = n 1 E v l δ k x ¯ ˙ u , n ; u + w n 1 / 2 w = 0 = n 1 j = 1 n E η ˙ j   Λ k , l , j u η ˙ j = = n 1 j = 1 n tr Λ k , l , j u E n ˙ j n ˙ j * + 2 Re E n ˙ j Λ k , l , j u d ˙ j u s ˙ j = n 1 j = 1 n tr Λ k , l , j u E n ˙ j n ˙ j * .
lim n   Φ n , k , l u = lim n n 1 j = 1 n E x ˙ j Λ k , l , j u x ˙ j = 0 2 π tr Λ k , l λ ; u F λ d λ = Φ k , l u
Let us prove that condition C of Theorem 1 is satisfied.
Lemma 4.
([18]). Let the objective function n 1 Q n x ¯ ˙ u , n ; v , v U , generating the estimate by the equation
u ^ n Q x ¯ ˙ u , n = arg max v U n 1 Q n x ¯ ˙ u , n ; v ,
have the following properties:
  • For any  v U , the random function  n 1 Q n x ¯ ˙ u , n , v  converges in probability to the deterministic function  Q v uniformly on  v U :
    P - lim n   n 1 Q n x ¯ ˙ u , n ; v = Q v .
  • The function  Q v  is continuous on  v U .
  • The function  Q v  has a unique maximum  u  at the set  U .
Then, the estimate  u ^ n x ¯ ˙ u , n = arg max v U n 1 Q x ¯ ˙ u , n ; v   is the consistent one:
P - lim n   u ^ n Q x ¯ ˙ u , n = u .
Lemma 4 can be formulated in the following equivalent form:
Lemma 5.
Let the vector random function
n 1 / 2 δ n x ¯ ˙ u , n , v = n 1 / 2 δ k , n x ¯ ˙ u , n ;   v = v k n 1 Q n x ¯ ˙ u , n ;   v ,   k 1 , q ¯ ,   v U ,
generating the estimate  u ^ n δ x ¯ ˙ u , n as the root of the equation
n 1 / 2 δ n x ¯ ˙ u , n ; v 0   for   each   n > q ,
have the following properties:
  • For any  v U , the random vector function  n 1 / 2 δ n x ¯ ˙ u , n ;   v converges in probability to the deterministic vector function  δ v uniformly on  v U :
    P u , n - lim n   n 1 / 2 δ n x ¯ ˙ u , n ;   v = δ v .
  • The vector function  δ v  is continuous on  v U .
  • The vector function  δ v  has a unique root  u at the set U .
Then, the estimate  u ˜ n δ x ¯ ˙ u , n , which is the root of the equation  n 1 / 2 δ n x ¯ ˙ u , n ;   v = 0 , is the consistent one P - lim n   u ^ n δ x ¯ ˙ u , n = u .
Let us prove that the random function  δ n x ¯ ˙ u , n ; v  from the left side of Equation (12), normalized to  n 1 / 2  as
n 1 / 2 δ n x ¯ ˙ u , n ; v = n 1 j = 1 n   ρ ˙ x u , j * Π ˙ k , j v ρ ˙ x u , j ,   k 1 , q ¯ ,     v U ,
satisfies the conditions of Lemma 4. To do this, we analyze the limits in probability of the random functions  n 1 / 2 δ k x ¯ ˙ u , n ; v k 1 , q ¯ v U  when  n . These functions have the form:
n 1 / 2 δ k x ¯ ˙ u , n ; v = n 1 j = 1 n d ˙ j * u Π ˙ k , j v d ˙ j u s j 2 + n 1 j = 1 n η ˙ u , j Π ˙ k , j v η ˙ u , j + + n 1 j = 1 n 2 Re η ˙ u , j Π ˙ k , j v d j u s j ,   k 1 , q ¯ .
The following statement is proven similarly to Theorem 1 in [18]:
Since the random values η ˙ j ,     j 1 , n ¯ have their properties defined by Equation (22), and functions Π ˙ k , j v , v U are continuous on the compact set U , the random functions (44) converge in probability to the limits of their mathematical expectations when n uniformly in U due to the Law of Large Numbers:
P - l i m n n 1 / 2 δ k x ¯ ˙ u , n ; v = l i m n n 1 / 2 j = 1 n E δ k x ¯ ˙ u , n ; v = δ k v ,   k 1 , q ¯ .
Let us find expressions for the values δ k v . Since
2 Re E η ˙ u , j Π ˙ k , j v d j u s j = 0   for   all   k 1 , q ¯ , j 1 , n ¯ ,   n > q ,
then
δ k v = l i m n n 1 j = 1 n E η ˙ u , j Π ˙ k , j v η ˙ u , j + lim n   n 1 j = 1 n d ˙ j * u Π ˙ k , j v d ˙ j u s ˙ j 2 .
According to (27), the first term in (46) is equal to zero for all v U ,     k 1 , q ¯ . The second term in (46) converges as n to the following integrals uniformly in v U :
δ k v = 2 π 1 0 2 π d ˙ λ ; u Π ˙ k λ ; v d ˙ λ ; u w λ 2 d λ ,   k 1 , q ¯ ,
where w s λ = lim n   n 1 j = 1 ν λ s 2 π j / n 2 , ν λ = λ is the largest integer n for which 2 π j n 1 λ .
Due to the properties of Π ˙ k λ ; v , the integrals (47) are continuous in v U , and the functions δ k v , k 1 , q ¯ have the unique roots u k on the compact set U .
Then, according to the statement of Lemma 5, the estimate u ˜ δ x ¯ ˙ n , which is the root of the equation δ n x ¯ ˙ n ; u = 0 , i.e.,
P u , n δ n x ¯ ˙ n ; u ˜ δ x ¯ ˙ n 0 = 1   for   n > q ,
is the consistent estimate of the value u of the MLS parameter.
Lemma 6.
The consistent estimate u ˜ n δ x ¯ ˙ n , satisfying Equation (48), is the n -consistent estimate, i.e., it has the following property: for any ε > 0 , there exists a C ε > 0 for which P n u ˜ δ x ¯ ˙ u , n u > C ε ε for every  n > q .
Proof 
Let us formulate the definitions of consistency and n -consistency in the equivalent forms:
(a)
For any ε n 0 , n , there exists a C 1 , n ε n for which
P u ˜ n δ x ¯ ˙ n u < C 1 , n 1 = 1 ε n .
(b)
For any ε n 0 , n , there exists a C ε n for which P n u ˜ n δ x ¯ ˙ n u < C ε n = 1 ε n .
Let us consider the random event
A n = u ˜ n δ x ¯ ˙ u , n u < C n 1 .
Then, condition (a) means that:
P A n = 1 ε n ,   where   ε n 0   n .
Let us consider the value θ ˜ n x ¯ ˙ n ;   u = n u ˜ n δ x ¯ ˙ n u and the event B n = θ ˜ n x ¯ ˙ n ;   u < C 2 , n ε n .
Then, condition (b) means that: P B n = 1 ε n   n .
The estimate u ˜ n δ x ¯ ˙ n , as the root of the equation δ n x ¯ ˙ n ; u = 0 with a probability equal to 1, is determined by the following relation:
P δ n x ¯ ˙ n ; u ˜ n δ x ¯ ˙ n 0 = 1   for   every   n > q .
Using the above designations, the statement of Lemma 5 can be written in the following form:
P δ n x ¯ ˙ n ; u ˜ n δ x ¯ ˙ n 0 A n P A n = = P δ n x ¯ ˙ n ; u ˜ n δ x ¯ ˙ n 0 B n P B n = 1 ε n ,   ε n 0   n .
Hence, the consistent estimator u ˜ n δ x ¯ ˙ u , n is also the n -consistent one. □
It follows from Lemmas 1–6 that all conditions of Theorem 1 are satisfied, and hence the random variable ζ n x ¯ ˙ u , n = n u ˜ δ x ¯ ˙ u , n u has, in asymptotic n , a zero mean and bounded covariance:
lim n   E z ¯ n ζ n z ¯ n = 0 ;   l i m n E z ¯ n ζ n z ¯ n ζ n z ¯ n T = Φ 1 u Ψ u Φ 1 u ,
where the elements of the matrices Φ u and Ψ u are represented in the form of integral expressions (42) and (43).
In order to understand the meaning of Formula (50), let us return to Section 2 and consider a case where q = 2 ,   u U ˜ 2 and m = 150 . Then, Formula (50) represents the covariance matrix of a Gaussian two-dimensional distribution that, for certain h τ u and F ˙ ξ λ 150 × 150 (used in Monte-Carlo modeling), will be very close to the empirical distribution shown in Figure 1b.
Discussion on the asymptotic normality of n 1 / 2 δ k x ¯ ˙ u , n ; u .
As it was noted before, the fact of the asymptotic normality of statistics (24) is not obvious and poses a separate problem that is more difficult than the Law of Large Numbers. It is required to establish central limit theorem (CLT) conditions for the weak dependent random variables that are represented as quadratic forms of functions of the discrete Fourier transform of stationary time series.
The CLT formulated for cases of dependent random variables are presented, for instance, in [21,22]. However, the conditions under which these theorems are stated are very restrictive, which makes them difficult to apply to statistics (24). In addition to the existence of finite first- and high-order moments ([21], theorem 27.4), there is a strong mixing condition for the measure of dependency between random variables, which cannot be applied to the terms of sum (24). The author believes that there is no ready solution to this problem in the form of a suitable theorem. At the moment, the author leaves this problem open for investigation in the future.

6. Conclusions

In this paper, we considered an important case of the vector parameter estimation problem for an MLS model with one input and several outputs, where the number of unknown parameters tends to infinity along with the number of observations. This case has never been studied before, and the explicit analytic form for the covariance matrix (50) of estimate (14) defined by (42) and (43) is the main theoretical result of this paper.
In practice, output processes y u , t are always heavily distorted by additive noise ξ t . That is why specialists in signal processing always look for the best estimate that can improve the accuracy of the estimation of unknown parameters. For that purpose, every new suggested estimate should be compared with the existing actual one, as described in Section 4. Using Formula (50) allows one to calculate the variance of estimator (14) for a given impulse response vector h τ u and F ˙ ξ λ m × m directly instead of performing the Monte-Carlo procedure, which assumes a mixture simulation of additive noise ξ t and outputs y u , t multiple times for the calculation of the empirical covariance matrix of estimator (14).
There are two unsolved theoretical problems regarding statistics n 1 / 2 δ k x ¯ ˙ u , n ; u that can be considered for future work. The first is establishing the validity of the CLT under the proven conditions of Theorem 1. The second is related to the question of boundary existence in a nonparametric model in terms of the lowest covariance matrix (50) in a class of regular estimators. This boundary can be achievable through estimate (14) as a likelihood-based estimate, but then it should be proven.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The author expresses gratitude to Alexander Kushnir for intensive help with proper literature, explanation of its contents and useful discussions that were important for the proofs of some theorems.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A. Probabilistic Properties of DFFT of a Vector Gaussian Stationary Time Series Satisfying Strong Mixing Conditions

Let ξ ˙ j , j 1 , n ¯ be the discrete finite Fourier transform (DFFT) of values ξ t ,     t 1 , n ¯ of m-dimensional stationary Gaussian time series ξ t m ,     t with zero mean and matrix power spectral density (MPSD) F ˙ λ = F re λ + i F im λ , λ [ 0 , 2 π ] :
ξ ˙ j = μ j + i ν j D F F T ξ t ,   t , j 1 , n ¯ ,   μ j = n 1 / 2 t = 1 n ξ t cos λ j t ,   ν j = n 1 / 2 t = 1 n ξ t sin λ j t ,   λ j = 2 π j   n 1 .
Theorem A1.
([23], chapter 4.2). The real and imaginary parts of   μ j ,   v j , ξ ˙ j = μ j + i ν j , and j 1 , n ¯ have the following probabilistic characteristics:
(a)
For every  j 1 , n ¯ , the following equations are correct:
E μ j = E v j = 0 ;   E μ j μ j T = 0.5 F ˙ j Re + O ˙ μ j n 1 β
E v j v j T = 0.5 F j Re + O v j n 1 β
E μ j ν j T = 0.5 F j Im + O μ j , ν j n 1 β ;   E v j μ j T = 0.5 F j Im + O v j , μ j n 1 β
sup j 1 , n ¯ O μ j n 1 β = sup j 1 , n ¯ O v j n 1 β C ¯ n 1 β
sup j 1 , n ¯ O μ j , v j n 1 β = sup j 1 , n ¯ O v j , μ j n 1 β C ¯ n 1 β
(b)
For all  j k 1 , n ¯ , the following inequalities are correct:
sup j k 1 , n ¯ E μ j μ k T = sup j k 1 , n ¯ E v j v k T C ¯ n 1 β
sup j k 1 , n ¯ E μ j v k T = sup j k 1 , n ¯ E v j μ k T C ¯ n 1 β
where  C ¯  and  β 0 , 1  are constants.

References

  1. Hannan, E.J. Multiple Time Series; John Willey and Sons, Inc.: New York, NY, USA; London, UK, 1970. [Google Scholar]
  2. Ibragimov, I.A.; Rozanov, Y.A. Gaussian Random Processes. Applications of Mathematics; Springer: New York, NY, USA, 1978; 277p. [Google Scholar]
  3. Diaz-Guerra, D.; Miguel, A.; Beltran, J.R. Robust sound source tracking using SRP-PHAT and 3D convolutional neural networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 300–311. [Google Scholar] [CrossRef]
  4. Kushnir, A.; Varypaev, A.; Dricker, I.; Rozhkov, M.; Rozhkov, N. Passive surface micro seismic monitoring as a statistical problem: Location of weak micro seismic signals in the presence of strongly correlated noise. Geophys. Prospect. 2014, 62, 819–833. [Google Scholar] [CrossRef]
  5. Rost, S.; Thomas, C. Array seismology: Methods and applications. Rev. Geophys. 2002, 40, 2.1–2.27. [Google Scholar] [CrossRef]
  6. Zhang, C.; Florêncio, D.; Ba, D.E.; Zhang, Z. Maximum likelihood sound source localization and beam forming for directional microphone arrays in distributed meetings. IEEE Trans. 2008, 10, 538–548. [Google Scholar]
  7. Nussbaum, M. An asymptotic minimal risk for estimation of a linear functional relationship. J. Multivar. Anal. 1984, 4, 300–314. [Google Scholar] [CrossRef]
  8. Ibragimov, I.A.; Has, R.Z. Statistical Estimation, Asymptotic Theory. Applications of Mathematics; Springer: New York, NY, USA, 1981; Volume 16, 403p. [Google Scholar]
  9. Berger, J.O. Statistical Decision Theory and Bayesian Analysis, 2nd ed.; Springer: New York, NY, USA, 1985. [Google Scholar]
  10. Borovkov, A.A. Mathematical Statistics; Gordon and Breach Science Publishers: Amsterdam, The Netherlands, 1998. [Google Scholar]
  11. Kushnir, A.F.; Pinskii, V.I. Asymptotically Efficient Estimation of Parameters of Linear Systems. I. Local Asymptotic Normality. Computational Seismology; Allerton Press Inc.: New York, NY, USA, 1987; Volume 19, pp. 98–116. [Google Scholar]
  12. Kushnir, A.F.; Rozhkov, N.M.; Varypaev, A.V. Statistically Based Approach to Monitoring of Microseismic Events. GEM Int. J. Geomath. 2013, 4, 201–225. [Google Scholar] [CrossRef]
  13. Linnik, J.V. Statistical Problems with Nuisance Parameters; Translations of Mathematical Monographs, American Mathematical Society: New York, NY, USA, 1968; Volume 20, 258p. [Google Scholar]
  14. Cramer, H.; Leadbetter, M.R. Stationary and Related Stochastic Processes. Sample Function. Properties and Their Applications; John Wiley & Sons: New York, NY, USA, 1967; 398p. [Google Scholar]
  15. Taniguchi, M.; Kakizawa, Y. Asymptotic Theory of Statistical Inference for Time Series; Springer Series in Statistics; Springer: New York, NY, USA, 2000; 661p. [Google Scholar]
  16. Capon, J. Application of the Theory of Detection and Estimation of Signals to the Seismology of Large Arrays. Proc. IEEE 1970, 58, 170–180. [Google Scholar] [CrossRef]
  17. Huber, P.J. Robust Statistics; John Wiley and Sons: New York, NY, USA, 1981. [Google Scholar]
  18. Newey, W.K.; McFadden, D. Large Sample Estimation and Hypothesis Testing. In Handbook of Econometrics; Engle, R.F., McFadden, D.L., Eds.; Elsevier: Amsterdam, The Netherlands, 1994; Volume 4, Chapter 36. [Google Scholar]
  19. Kushnir, A.F.; Varypaev, A.V. Asymptotical Distributions of M-estimates for Parameters of Multivariate Time Series with Strong Mixing Property. Eng. Proc. 2021, 5, 19. [Google Scholar]
  20. Rencher, A.C.; Schaalje, G.B. Linear Models in Statistics; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2008; 672p. [Google Scholar]
  21. Billingsley, P. Probability and Measure, 3rd ed.; John Wiley & Sons: New York, NY, USA, 1995; 593p. [Google Scholar]
  22. Bradley, R. Central limit theorems under weak dependence. J. Multivar. Anal. 1981, 11, 1–16. [Google Scholar] [CrossRef]
  23. Brillinger, D. Data Analysis and Theory; Time Series; SIAM: Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2001; 570p. [Google Scholar]
Figure 1. Empirical two-dimensional probability density functions built by two different sets of 110 estimates: (a) set is given by SRP-PHAT, ρ R M S E = 0.009 ; (b) set is given by u ^ n δ x ¯ ˙ u , n , ρ R M S E = 0.005 . Case of real noise.
Figure 1. Empirical two-dimensional probability density functions built by two different sets of 110 estimates: (a) set is given by SRP-PHAT, ρ R M S E = 0.009 ; (b) set is given by u ^ n δ x ¯ ˙ u , n , ρ R M S E = 0.005 . Case of real noise.
Mathematics 12 00473 g001
Figure 2. Empirical two-dimensional probability density functions built by two different sets of 110 estimates: (a) set is given by SRP-PHAT, ρ R M S E = 0.0532 ; (b) set is given by u ^ n δ x ¯ ˙ u , n , ρ R M S E = 0.0298 . Case of noise with F ˙ ξ λ = c o n s t * I .
Figure 2. Empirical two-dimensional probability density functions built by two different sets of 110 estimates: (a) set is given by SRP-PHAT, ρ R M S E = 0.0532 ; (b) set is given by u ^ n δ x ¯ ˙ u , n , ρ R M S E = 0.0298 . Case of noise with F ˙ ξ λ = c o n s t * I .
Mathematics 12 00473 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Varypaev, A. Asymptotic Form of the Covariance Matrix of Likelihood-Based Estimator in Multidimensional Linear System Model for the Case of Infinity Number of Nuisance Parameters. Mathematics 2024, 12, 473. https://doi.org/10.3390/math12030473

AMA Style

Varypaev A. Asymptotic Form of the Covariance Matrix of Likelihood-Based Estimator in Multidimensional Linear System Model for the Case of Infinity Number of Nuisance Parameters. Mathematics. 2024; 12(3):473. https://doi.org/10.3390/math12030473

Chicago/Turabian Style

Varypaev, Alexander. 2024. "Asymptotic Form of the Covariance Matrix of Likelihood-Based Estimator in Multidimensional Linear System Model for the Case of Infinity Number of Nuisance Parameters" Mathematics 12, no. 3: 473. https://doi.org/10.3390/math12030473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop