Next Article in Journal
Counting Rules for Computing the Number of Independent Sets of a Grid Graph
Previous Article in Journal
Federated Learning with Efficient Aggregation via Markov Decision Process in Edge Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Chi-Square Approximation for the Distribution of Individual Eigenvalues of a Singular Wishart Matrix

by
Koki Shimizu
* and
Hiroki Hashiguchi
Department of Applied Mathematics, Tokyo University of Science, 1-3 Kagurazaka, Shinjuku-ku, Tokyo 162-8601, Japan
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(6), 921; https://doi.org/10.3390/math12060921
Submission received: 27 February 2024 / Revised: 15 March 2024 / Accepted: 19 March 2024 / Published: 20 March 2024
(This article belongs to the Special Issue Theory and Applications of Random Matrix)

Abstract

:
This paper discusses the approximate distributions of eigenvalues of a singular Wishart matrix. We give the approximate joint density of eigenvalues by Laplace approximation for the hypergeometric functions of matrix arguments. Furthermore, we show that the distribution of each eigenvalue can be approximated by the chi-square distribution with varying degrees of freedom when the population eigenvalues are infinitely dispersed. The derived result is applied to testing the equality of eigenvalues in two populations.

1. Introduction

The Wishart matrix is a symmetric random matrix defined by the sum of squares and cross-products of samples from a multivariate normal distribution. It becomes non-singular when the dimension is smaller than or equal to the number of observations; otherwise, it is singular. The distributions for the Wishart matrix and its eigenvalues have been used in many areas of science and technology, including multivariate analysis, Bayesian statistics, random matrix theory, and wireless communications. Some exact distributions of eigenvalues for a Wishart matrix are represented by the hypergeometric functions of matrix arguments. James [1] classified multivariate statistics problems into five categories based on hypergeometric functions. However, the convergence of these functions is slow, and their numerical computation is cumbersome when sample sizes or dimensions are large. Consequently, the derivation of approximate distributions of eigenvalues has received a great deal of attention. Sugiyama [2] derived the approximate distribution for the largest eigenvalue through the integral representation of the confluent hypergeometric function. Sugiura [3] showed that the asymptotic distribution of the individual eigenvalues is expressed by a normal distribution for a large sample size. The chi-square approximation was discussed when the population eigenvalues are infinitely dispersed in Kato and Hashiguchi [4] and Takemura and Sheena [5]. Approximations for hypergeometric functions have been developed and applied to the multivariate distribution theory in Butler and Wood [6,7,8]. Butler and Wood [6] provided the Laplace approximation for the hypergeometric functions of a single matrix argument. The numerical accuracies for that approximation were shown in the computation of noncentral moments of Wilk’s lambda statistic and the likelihood ratio statistic for testing block independence. This approximation was extended to the case of two matrix arguments in Butler and Wood [7]. All the results addressed above were carried out for eigenvalue distributions for a non-singular Wishart matrix.
Recently, the distribution of eigenvalues for the non-singular case has been extended to the singular case; see Shimizu and Hashiguchi [9] and Shinozaki et al. [10]. Shimizu and Hashiguchi [9] showed the exact distribution of the largest eigenvalue for a singular case is represented in terms of the confluent hypergeometric function as well as the non-singular case. The generalized representation for the non-singular and singular cases under the elliptical model was provided by Shinozaki et al. [10].
This paper is organized as follows. In Section 2, we apply the Laplace approximation introduced by Butler and Wood [7] to the joint density of eigenvalues of a singular Wishart matrix. Furthermore, we show that the approximation for the distribution of the individual eigenvalues can be expressed by the chi-square distribution with varying degrees of freedom when the population covariance matrix has spiked eigenvalues. Section 3 discusses the equality of the individual eigenvalues in two populations. Finally, we evaluate the precision of the chi-square approximation by comparing it to the empirical distribution through Monte Carlo simulation in Section 4.

2. Approximate Distributions of Eigenvalues of a Singular Wishart Matrix

Suppose that an m × n real Gaussian random matrix X is distributed as X N m , n ( O , Σ I n ) , where O is the m × n zero matrix, Σ is a m × m positive symmetric matrix, and ⊗ is the Kronecker product. This means that the column vectors of X are independently and identically distributed (i.i.d.) from N m ( 0 , Σ ) with sample size n, where 0 is the m-dimensional zero vector. The eigenvalues of Σ are denoted by λ 1 , λ 2 , , λ m , and λ 1 λ 2 λ m > 0 . Subsequently, we define the singular Wishart matrix as W = X X , where m > n and its distribution is denoted by W ( n , Σ ) . The spectral decomposition of W is represented as W = H 1 L 1 H 1 , where L 1 = diag ( 1 , , n ) with 1 > 2 > > n > 0 , and the m × n matrix H 1 is satisfied by H 1 H 1 = I n . The set of all m × n matrices H 1 with orthonormal columns is called the Stiefel manifold, denoted by V n , m = { H 1 H 1 H 1 = I n } , where m n . The volume of V n , m is represented by
Vol ( V n , m ) = H 1 V n , m ( H 1 d H 1 ) = 2 n π m n / 2 Γ n ( m / 2 ) .
For the definition of the above exterior product ( H 1 d H 1 ) , see page 63 of Muirhead [11]. If m = n , Stiefel manifold V m , m coincides with the orthogonal groups O ( m ) . Uhlig [12] gave the density of W as
f ( W ) = π ( m n + n 2 ) / 2 2 m n / 2 Γ n ( n / 2 ) | Σ | n / 2 | L 1 | ( n m 1 ) / 2 etr ( Σ 1 W / 2 ) ,
where Γ m ( a ) = π m ( m 1 ) / 4 i = 1 m Γ { a ( i 1 ) / 2 } and etr ( · ) = exp ( tr ( · ) ) . Srivastava [13] represented the joint density of eigenvalues of W in a form that includes an integral over the Stiefel manifold;
f ( 1 , , n ) = 2 n m / 2 π n 2 / 2 | Σ | n / 2 Γ n ( n / 2 ) Γ n ( m / 2 ) i = 1 n i ( m n 1 ) / 2 i < j n ( i j ) × H 1 V n , m etr ( 1 2 Σ 1 H 1 L 1 H 1 ) ( d H 1 ) ,
where ( d H 1 ) = ( H 1 d H 1 ) Vol ( V n , m ) and H 1 V n , m ( d H 1 ) = 1 .
The above integral over the Stiefel manifold was evaluated by Shimizu and Hashiguchi [9] as the hypergeometric functions of the matrix arguments. We approximate (1) by Laplace approximation for the hypergeometric functions of two matrix arguments provided by Butler and Wood [7].
For a positive integer k, let κ = ( κ 1 , κ 2 , , κ m ) denote a partition of k with κ 1 κ 2 κ m 0 and κ 1 + + κ m = k . The set of all partitions with less than or equal to m is denoted by P m k = { κ = ( κ 1 , , κ m ) κ 1 + + κ m = k , κ 1 κ 2 κ m 0 } . The Pochhammer symbol for a partition κ is defined as ( α ) κ = i = 1 m { α ( i 1 ) / 2 } κ i , where ( α ) k = α ( α + 1 ) ( α + k 1 ) and ( α ) 0 = 1 . For integers, p , q 0 and m × m real symmetric matrices A and B, we define the hypergeometric function of two matrix arguments as
F q p ( m ) ( α ; β ; A , B ) = k = 0 κ P m k ( α 1 ) κ ( α p ) ( β 1 ) κ ( β q ) C κ ( A ) C κ ( B ) k ! C κ ( I m ) ,
where α = ( α 1 , , α p ) , β = ( β 1 , , β q ) and C κ ( A ) is the zonal polynomial indexed by κ with the symmetric matrix A; see the details provided in Chapter 7 of Muirhead [11]. The hypergeometric functions with a single matrix are defined as
F q p ( α ; β ; A ) = F q p ( α ; β ; A , I m ) .
The special cases F 1 1 and F 1 2 of (3) are called the confluent and Gauss hypergeometric functions, respectively. Butler and Wood [6]. proposed a Laplace approximation of F 1 1 and F 1 2 through their integral expressions. They showed that the accuracy of that approximation is greater than the previous results. This approximation was extended to the complex case in Butler and Wood [8]. The important property of (2) is the integral representation over the orthogonal group.
F q p ( m ) ( α ; β ; A , B ) = H O ( m ) F q p ( α ; β ; A H B H ) ( d H ) ,
where ( d H ) is the invariant measure on the m × m orthogonal group O ( m ) . Integral representations (4) are a useful tool for obtaining an approximation of F q p ( m ) . Asymptotic expansions of F 0 0 ( m ) are given in Anderson [14] when both two positive definite matrix arguments are widely spaced. Constantine and Muirhead [15] gave the asymptotic behavior of F 0 0 ( m ) when the population eigenvalues are multiple. From the integral expression (4), Butler and Wood [7] provided Laplace approximations for F q p ( m ) .
Lemma 1.
Let the two diagonal matrices be A = diag ( a 1 , , a m ) and B = diag ( b 1 , , b 1 , b 2 , , b r , , b r ) , where a 1 > a 2 > > a m > 0 , b 1 > b 2 > > b r 0 , and b j have multiplicity m j , in which m = j = 1 r m j . Let Ω ( m 1 , , m r ) = Vol ( O ( m ) ) 1 j = 1 r Vol ( O ( m j ) ) . Then the Laplace approximation of F q p ( m ) is given as
F ^ q p ( m ) ( α ; β ; A , B ) = ( 2 π ) s 2 Ω ( m 1 , , m r ) J 1 2 F q p ( α ; β ; A B ) ,
where s = i = 1 r 1 j = i + 1 r m i m j and Hessian J is defined in Butler and Wood [7].
Shimizu and Hashiguchi [9] showed the following relationship
H 1 V n , m F q p ( A H 1 B 1 H 1 ) ( d H 1 ) = H O ( m ) F q p ( A H B H ) ( d H )
for an m × m matrix B = B 1 O O O , where B 1 is an n × n symmetric matrix and O is the zero matrix. From (5), the joint density (1) can be rewritten by
f ( 1 , , n ) H O ( m ) etr ( 1 2 Σ 1 H L H ) ( d H ) = F 0 0 ( m ) ( 1 2 Σ 1 , L ) ,
where L = diag ( 1 , , n , 0 , , 0 ) is the m × m matrix and the symbol “∝” means that a constant required for scaling is removed. Applying Laplace’s method to the above joint density, we have an approximation for the joint density of eigenvalues.
Proposition 1.
The joint density of eigenvalues of a singular Wishart matrix by Laplace approximation is expressed by
π n ( n m ) / 2 2 n m / 2 | Σ | n / 2 Γ n ( n / 2 ) i = 1 n i ( m n 1 ) / 2 i < j n ( i j ) exp 1 2 i = 1 n i λ i i < j n 2 π c i j 1 / 2 i = 1 n j = n + 1 m 2 π d i j 1 / 2 ,
where c i j = ( i j ) ( λ i λ j ) λ i λ j , d i j = i ( λ i λ j ) λ i λ j .
Proof. 
Applying Lemma 1 to the hypergeometric functions in (6), the integral over the Stiefel manifold in (1) is approximated by
2 n Vol ( V n , m ) exp 1 2 i = 1 n i λ i i < j n 2 π c i j 1 / 2 i = 1 n j = n + 1 m 2 π d i j 1 / 2 .
Substituting (8) to (1), we have the desired result. □
In order to derive the approximate distributions of individual eigenvalues, we define the spiked covariance model ρ k that implies the first k-th eigenvalues of Σ > 0 are infinitely dispersed, namely
ρ k = max ( λ 2 λ 1 , λ 3 λ 2 , , λ k + 1 λ k ) 0 ,
where k n . Under the condition of (9) when k = n , Takemura and Sheena [5] proved that the distribution of individual eigenvalues for a non-singular Wishart matrix is approximated by a chi-square distribution. The improvement for that approximation, that is, when the condition listed in (9) cannot be assumed, was discussed in Tsukada and Sugiyama [16]. The following lemma was provided by Nasuda et al. [17] and Takemura and Sheena [5] in the non-singular case and could be easily extended to the singular case.
Lemma 2.
Let W W m ( n , Σ ) , where m > n and 1 , 2 , , n be the eigenvalues of W. If ρ k 0 , we have
r k = max ( 2 1 , 3 λ 2 , , k + 1 k ) p 0 ,
in the sense that ϵ > 0 , δ > 0 ,
ρ k < δ Pr ( r k > ϵ ) < ϵ ,
where k n .
From Proposition 1 and Lemma 2, we obtain the chi-square approximation that is the main result of this paper.
Theorem 1.
Let W W m ( n , Σ ) , where m > n and 1 , 2 , , n be the eigenvalues of W. If ρ k 0 , it holds that
i / λ i d χ n i + 1 2 , 1 i k ,
where χ 2 is a chi-square distribution with n i + 1 degrees of freedom and the symbol “ d ” means convergence in the distribution.
Proof. 
First, we rewrite the approximate distribution (7) as
f ( 1 , , n ) = 1 2 n ( n + 1 ) / 4 | Σ | n / 2 i = 1 n Γ ( n i + 1 2 ) i = 1 n i ( m n 1 ) / 2 exp 1 2 i = 1 n i λ i × i < j n ( i j ) λ i λ j λ i λ j 1 / 2 i = 1 n j = n + 1 m 1 d i j 1 / 2 .
From Lemma 2, we have
i < j n ( i j ) 1 / 2 = 1 i k , i < j n i 1 / 2 ( 1 j i ) 1 / 2 i = k + 1 < j n ( i j ) 1 / 2 i = 1 k i ( n i ) / 2 i = k + 1 < j n ( i j ) 1 / 2 .
Then the approximate joint density is expressed by
f ( 1 , , n ) 1 2 n ( n + 1 ) / 4 | Σ | n / 2 i = 1 n Γ ( n i + 1 2 ) i = 1 k i ( m i + 1 ) / 2 i = k + 1 n i ( m n 1 ) / 2 i = k + 1 < j n ( i j ) 1 / 2 × exp 1 2 i = 1 n i λ i i < j n λ i λ j λ i λ j 1 / 2 i = 1 n j = n + 1 m 1 d i j 1 / 2 = 1 2 n ( n + 1 ) / 4 | Σ | n / 2 i = 1 n Γ ( n i + 1 2 ) i = 1 k i ( n i + 1 ) / 2 i = k + 1 n i ( m n 1 ) / 2 i = k + 1 < j n ( i j ) 1 / 2 × exp 1 2 i = 1 n i λ i i < j n λ i λ j λ i λ j 1 / 2 i = 1 k j = n + 1 m λ j 1 λ j / λ i 1 / 2 i = k + 1 n j = n + 1 m 1 d i j 1 / 2 .
If ρ k 0 , we have
i < j n λ i λ j λ i λ j 1 / 2 = 1 i k , i < j n ( λ j 1 λ j / λ i ) 1 / 2 i = k + 1 < j n ( λ i λ j λ i λ j ) 1 / 2 i = 1 k λ i ( n i ) / 2 i = k + 1 < j n ( λ i λ j λ i λ j ) 1 / 2 .
We note | Σ | n / 2 = i = 1 k λ i n / 2 i = k + 1 n λ i n / 2 j = n + 1 m λ j n / 2 ; the joint density (10) is approximated as
f ( 1 , , n ) i = 1 k i ( n i 1 ) / 2 ( 2 λ i ) ( n i + 1 ) / 2 Γ ( n i + 1 2 ) exp i 2 λ i · f ( k + 1 , , n ) i = 1 k λ i ( n 2 i + 1 ) / 2 j = n + 1 m λ j ( k n ) / 2 = i = 1 k g n i + 1 ( i / λ i ) · f ( k + 1 , , n ) i = 1 k λ i ( n 2 i + 1 ) / 2 j = n + 1 m λ j ( k n ) / 2 ,
where g n i + 1 ( · ) is the density function of the chi-square distribution, and degree of freedom n i + 1 and f ( k + 1 , , n ) is represented by
f ( k + 1 , , n ) = i = k + 1 n 1 2 ( n i + 1 ) / 2 λ i n / 2 Γ ( n i + 1 2 ) i = k + 1 n i ( m n 1 ) / 2 i = k + 1 < j n ( i j ) × exp 1 2 i = k + 1 n i λ i i = k + 1 < j n 1 c i j 1 / 2 i = k + 1 n j = n + 1 m 1 d i j 1 / 2 .
Corollary 1 shows the chi-square approximation when all population eigenvalues are infinitely dispersed.
Corollary 1.
Let W W m ( n , Σ ) , where m > n and 1 , 2 , , n are the eigenvalues of W. If ρ n 0 , it holds that
i / λ i d χ n i + 1 2 , 1 i n .
Proof. 
The proof is provided in the Appendix A. □
In the context of the High Dimension-Low Sample Size (HDLSS) setting, the asymptotic behavior of the eigenvalue distribution of a sample covariance matrix was discussed in Ahn et al. [18], Bolivar-Cime and Perez-Abreu [19], Jung and Marron [20]. Jung and Marron [20] showed that the spiked sample eigenvalues are approximated by the chi-square distribution with a degree of freedom of n. In contrast, Theorem 1 provides the approximation of the distribution of individual eigenvalues by a chi-square distribution with varying degrees of freedom.

3. Application to Test for Equality of the Individual Eigenvalues

This section discusses testing for equality of individual eigenvalues of the covariance matrix in two populations. For testing problems, we give the approximate distribution of the statistic based on the derived results from the previous section.
Let an m × n i Gaussian random matrix X ( i ) be distributed as X ( i ) N m , n ( O , Σ ( i ) I n i ) , where Σ ( i ) > 0 and i = 1 , 2 . The eigenvalues of Σ ( i ) are denoted by λ 1 ( i ) , λ 2 ( i ) , , λ m ( i ) , where λ 1 ( i ) λ 2 ( i ) λ m ( i ) > 0 . We denote the eigenvalues of W ( i ) = X ( i ) X ( i ) by 1 ( i ) , 2 ( i ) , , m ( i ) , where 1 ( i ) > 2 ( i ) > > m ( i ) 0 . For fixed j, we consider the test of the equality of the individual eigenvalues in two populations as
H 0 : λ j ( 1 ) = λ j ( 2 ) , vs . H 1 : λ j ( 1 ) λ j ( 2 ) .
Sugiyama and Ushizawa [21] reduced (11) to the equality of variance test for the principal components and proposed a testing procedure using the Ansari-Bradley test. Takeda [22] proposed the test statistic j ( 1 ) / j ( 2 ) with n m for (11) and derived the exact distribution of 1 ( 1 ) / 1 ( 2 ) . Since Johnstone [23] indicated that the first few eigenvalues are very large compared to the others in the large dimensional setting, it is essential to understand how the distribution for the first few eigenvalues is constructed. We provide the exact density function of 1 ( 1 ) / 1 ( 2 ) with n < m in the same way as Takeda [22].
Theorem 2.
Let W ( 1 ) and W ( 2 ) be two independent Wishart matrices with distribution W m ( n 1 , Σ ( 1 ) ) and W m ( n 2 , Σ ( 2 ) ) , respectively, where m > n i ( i = 1 , 2 ) . Then we have the density of q = 1 ( 1 ) / 1 ( 2 ) as
f ( q ) = C k = 0 κ P m k t = 0 τ P m t { ( m + 1 ) / 2 } κ C κ ( Σ ( 1 ) 1 / 2 ) { ( n 1 + m + 1 ) / 2 } κ k ! { ( m + 1 ) / 2 } τ C τ ( Σ ( 2 ) 1 / 2 ) { ( n 2 + m + 1 ) / 2 } τ t ! × { ( m n 1 / 2 + k ) ( m n 2 / 2 + t ) q m n 2 / 2 + t 1 Γ ( u ) / v u ( m n 1 / 2 + k ) ( tr Σ ( 2 ) 1 / 2 ) q m n 2 / 2 + t Γ ( u + 1 ) / v u + 1 ( m n 2 / 2 + t ) ( tr Σ ( 1 ) 1 / 2 ) q m n 2 / 2 + t 1 Γ ( u + 1 ) / v u + 1 + ( tr Σ ( 1 ) 1 / 2 ) ( tr Σ ( 2 ) 1 / 2 ) q m n 2 / 2 + t Γ ( u + 2 ) / v u + 2 } ,
where u = m ( n 1 + n 2 ) / 2 + k + t , v = tr Σ ( 1 ) 1 q tr Σ ( 2 ) 1 and
C = Γ n 1 { ( n 1 + 1 ) / 2 } Γ n 2 { ( n 2 + 1 ) / 2 } 2 m ( n 1 + n 2 ) / 2 Γ n 1 { ( n 1 + m + 1 ) / 2 ) Γ n 2 { ( n 2 + m + 1 ) / 2 ) } | Σ | n 1 / 2 | Σ | n 2 / 2 .
Proof. 
The exact expression of 1 ( i ) was provided by Shimizu and Hashiguchi [9] as
Pr ( 1 ( i ) < x ) = Γ n i ( n i + 1 2 ) ( x 2 ) m n i / 2 Γ n i ( n i + m + 1 2 ) | Σ ( i ) | n i / 2 exp ( x 2 tr Σ ( i ) 1 ) F 1 1 m + 1 2 ; n i + m + 1 2 ; x 2 Σ ( i ) 1 .
The derivative of (13) is represented by
f ( x ) = Γ n i { ( n i + 1 ) / 2 } 2 m n i / 2 Γ n i { ( n i + m + 1 ) / 2 ) } | Σ ( i ) | n i / 2 k = 0 κ P m k { ( m + 1 ) / 2 } κ C κ ( Σ ( i ) 1 / 2 ) { ( n i + m + 1 ) / 2 } κ k ! × exp ( x 2 tr Σ ( i ) 1 ) { ( n i m / 2 + k ) x m n i / 2 + k 1 ( tr Σ ( i ) 1 / 2 ) x m n i / 2 + k } .
From (14), we have the joint density of 1 ( 1 ) and 1 ( 2 ) as
f ( x , y ) = C k = 0 κ P m k t = 0 τ P m t { ( m + 1 ) / 2 } κ C κ ( Σ ( 1 ) 1 / 2 ) { ( n 1 + m + 1 ) / 2 } κ k ! { ( m + 1 ) / 2 } τ C τ ( Σ ( 2 ) 1 / 2 ) { ( n 2 + m + 1 ) / 2 } τ t ! × { ( m n 1 / 2 + k ) x m n 1 / 2 + k 1 ( tr Σ ( 1 ) 1 / 2 ) x m n 1 / 2 + k } × { ( m n 2 / 2 + t ) y m n 2 / 2 + t 1 ( tr Σ ( 2 ) 1 / 2 ) y m n 2 / 2 + t } exp ( x 2 tr Σ ( 1 ) 1 ) exp ( y 2 tr Σ ( 2 ) 1 ) .
Translating x and y to q = y / x and r = x , we have
f ( q , r ) = C k = 0 κ P m k t = 0 τ P m t { ( m + 1 ) / 2 } κ C κ ( Σ ( 1 ) 1 / 2 ) { ( n 1 + m + 1 ) / 2 } κ k ! { ( m + 1 ) / 2 } τ C τ ( Σ ( 2 ) 1 / 2 ) { ( n 2 + m + 1 ) / 2 } τ t ! × { ( m n 1 / 2 + k ) ( m n 2 / 2 + t ) q m n 2 / 2 + t 1 r m ( n 1 + n 2 ) / 2 + k + t 1 ( m n 1 / 2 + k ) ( tr Σ ( 2 ) 1 / 2 ) q m n 2 / 2 + t r m ( n 1 + n 2 ) / 2 + k + t ( m n 2 / 2 + t ) ( tr Σ ( 1 ) 1 / 2 ) q m n 2 / 2 + t 1 r m ( n 1 + n 2 ) / 2 + k + t + ( tr Σ ( 1 ) 1 / 2 ) ( tr Σ ( 2 ) 1 / 2 ) q m n 2 / 2 + t r m ( n 1 + n 2 ) / 2 + k + t + 1 } × exp ( ( tr Σ ( 1 ) 1 q tr Σ ( 2 ) 1 ) r ) .
Noting that 0 x α 1 e β x d x = Γ ( α ) / β α , where α , β > 0 , and integrating r with respect to f ( q , r ) , we have the desired result. □
As the dimension increases, it is difficult to perform the numerical computation of (12) due to the high computational complexity. From Theorem 1, we provide the approximate distribution for (12) by F-distribution.
Corollary 2.
Let W ( 1 ) and W ( 2 ) be two independent Wishart matrices with distribution W m ( n 1 , Σ ( 1 ) ) and W m ( n 2 , Σ ( 2 ) ) , respectively, where m > n i ( i = 1 , 2 ) and 1 ( i ) , 2 ( i ) , , n ( i ) are the eigenvalues of W ( i ) . If the first k-th eigenvalues of Σ ( i ) are spiked, then we have
j ( 1 ) / { ( n 1 j + 1 ) λ n 1 j + 1 ( 1 ) } j ( 2 ) / { ( n 2 j + 1 ) λ n 2 j + 1 ( 2 ) } d F ( n 1 j + 1 , n 2 j + 1 ) , 1 j k ,
where F is an F distribution with n 1 and n 2 degrees of freedom.

4. Simulation Study

We investigate the accuracy of the approximation for the derived distributions. In the simulation study, we consider the following population covariance matrix:
Σ = diag ( a b , a b / 2 , , a b / m ) ,
where a , b > 0 . In the large-dimensional setting, mainly the accuracy of the approximate distributions for the largest and second eigenvalues was investigated; see Iimori et al. [24]. In (15), we set ( a , b ) = ( 200 , 3 ) as Case 1 and ( a , b ) = ( 50 , 3 ) as Case 2. These two cases imply that the population covariance matrix has two spiked eigenvalues. Parameter ρ k in (9) is smaller in Case 1 than in Case 2. We denote F 1 ( x ) and F 2 ( x ) as the chi-square distributions with n and n 1 degrees of freedom, which are the approximate distributions of the largest and second eigenvalues, respectively. The empirical distribution based on 10 6 Monte Carlo simulations is denoted by F sim . Table 1 and Table 2 show the 100 α -percentile points of the distributions of 1 and 2 for m = 50 and n = 10 , respectively. From the simulation study, we know that sufficient accuracy of approximation for the largest eigenvalue has already been obtained in Case 2. Case 1 is more accurate than Case 2 for the second eigenvalue. It is seen that the desired accuracy can be achieved when the parameter ρ k is small.
Table 3 and Table 4 present the chi-square probabilities for Case 1 in 90% 95% 99% percentile points denoted as x 0.90 , x 0.95 , and x 0.99 from the empirical distribution. We denote G 1 ( x ) and G 2 ( x ) as the chi-square approximation of the distributions for the largest and second largest eigenvalues, respectively. In this simulation study, we set m = 20 , 30 , 40 , 100 , and n = 5 , 15 . It can be observed that all probabilities are close to the true theoretical probabilities.
Finally, we provide the graph of the density of F distribution in Corollary 2 compared to the empirical distribution function. In Figure 1, we superimpose the graph of the F approximation with the histogram of 1 ( 1 ) / 1 ( 2 ) for n i = 10 ( i = 1 , 2 ) and m = 30 in Case 2. The vertical line and histograms show the empirical distribution of the 1 ( 1 ) / 1 ( 2 ) based on 10 6 iteration, respectively. The solid line is the density function of the F distribution. From the 95 % points of F sim , we can confirm that the approximate probability is 0.950.

5. Concluding Remarks

In this study, we provided the approximate distribution of eigenvalues of the singular Wishart matrix, which is similar to the result of Takemura and Sheena [5] for the non-singular case. Through numerical experiments, we confirmed that the approximation accuracy is sufficient when the parameter ρ k is small. The distribution approximation proposed by Tsukada and Sugiyama [16] might be useful to improve the derived approximate results when ρ k 0 is not assumed. As a part of future work, it would be desirable to examine the robustness of the chi-square approximation to normality assumption.

Funding

This work was supported by JSPS KAKENHI Grant Numbers 23K19015 and 23K11016.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Corollary 1

From Lemma 2, we have
i < j n ( i j ) 1 / 2 = i < j n i 1 / 2 ( 1 j i ) 1 / 2 i = 1 n i ( n i ) / 2 , i < j n λ i λ j λ i λ j 1 / 2 = i < j n ( λ j 1 λ i / λ j ) 1 / 2 i = 1 n λ i ( i 1 ) / 2 .
Furthermore, we note | Σ | n / 2 = i = 1 n λ i n / 2 i = 1 n j = n + 1 m λ j 1 / 2 ; the joint density (10) is represented by
f ( 1 , , n ) i = 1 n i ( n i 1 ) / 2 ( 2 λ i ) ( n i + 1 ) / 2 Γ ( n i + 1 2 ) exp i 2 λ i i = 1 n i ( m n ) / 2 i = 1 n j = n + 1 m ( 1 λ j ) 1 / 2 i = 1 n j = n + 1 m 1 d i j 1 / 2 = i = 1 n i ( n i 1 ) / 2 ( 2 λ i ) ( n i + 1 ) / 2 Γ ( n i + 1 2 ) exp i 2 λ i i = 1 n j = n + 1 m i 1 2 i = 1 n i = j + 1 m ( 1 λ j ) 1 / 2 i = 1 n j = n + 1 m 1 d i j 1 / 2 = i = 1 n i ( n i 1 ) / 2 ( 2 λ i ) ( n i + 1 ) / 2 Γ ( n i + 1 2 ) exp i 2 λ i = i = 1 n g n i + 1 ( i / λ i ) .

References

  1. James, A.T. Distribution of matrix variates and latent roots derived from normal samples. Ann. Math. Stat. 1964, 35, 475–501. [Google Scholar] [CrossRef]
  2. Sugiyama, T. Approximation for the distribution of the largest latent root of a Wishart matrix. Aust. J. Stat. 1972, 14, 17–24. [Google Scholar] [CrossRef]
  3. Sugiura, N. Derivatives of the characteristic root of a symmetric or a hermitian matrix with two applications in multivariate analysis. Commun. Stat. Theory Methods 1973, 1, 393–417. [Google Scholar] [CrossRef]
  4. Kato, H.; Hashiguchi, H. Chi-square approximations for eigenvalue distributions and confidential interval construction on population eigenvalues. Bull. Comput. Stat. Jpn. 2014, 27, 11–28. [Google Scholar]
  5. Takemura, A.; Sheena, Y. Distribution of eigenvalues and eigenvectors of Wishart matrix when the population eigenvalues are infinitely dispersed and its application to minimax estimation of covariance matrix. J. Multivar. Anal. 2005, 94, 271–299. [Google Scholar] [CrossRef]
  6. Butler, R.W.; Wood, T.A. Laplace approximations for hypergeometric functions with matrix argument. Ann. Stat. 2002, 30, 1155–1177. [Google Scholar] [CrossRef]
  7. Butler, R.W.; Wood, T.A. Laplace approximations to hypergeometric functions of two matrix arguments. J. Multivar. Anal. 2005, 94, 1–18. [Google Scholar] [CrossRef]
  8. Butler, R.W.; Wood, T.A. Laplace approximations for hypergeometric functions with Hermitian matrix argument. J. Multivar. Anal. 2022, 192, 105087. [Google Scholar] [CrossRef]
  9. Shimizu, K.; Hashiguchi, H. Heterogeneous hypergeometric functions with two matrix arguments and the exact distribution of the largest eigenvalue of a singular beta-Wishart matrix. J. Multivar. Anal. 2021, 183, 104714. [Google Scholar] [CrossRef]
  10. Shinozaki, A.; Shimizu, K.; Hashiguchi, H. Generalized heterogeneous hypergeometric functions and the distribution of the largest eigenvalue of an elliptical Wishart matrix. Random Matrices Theory Appl. 2022, 11, 2250034. [Google Scholar] [CrossRef]
  11. Muirhead, R.J. Aspects of Multivariate Statistical Theory; Wiley: New York, NY, USA, 1982. [Google Scholar]
  12. Uhlig, H. On Singular Wishart and singular multivariate beta distributions. Ann. Stat. 1994, 22, 395–405. [Google Scholar] [CrossRef]
  13. Srivastava, M.S. Singular Wishart and multivariate beta distributions. Ann. Stat. 2003, 31, 1537–1560. [Google Scholar] [CrossRef]
  14. Anderson, G.A. An asymptotic expansion for the distribution of the latent roots of the estimated covariance matrix. Ann. Math. Stat. 1965, 36, 1153–1173. [Google Scholar] [CrossRef]
  15. Constantine, A.G.; Muirhead, R.T. Asymptotic expansions for distributions of latent roots in multivariate analysis. J. Multivar. Anal. 1976, 3, 369–391. [Google Scholar] [CrossRef]
  16. Tsukada, S.; Sugiyama, T. Distribution approximation of covariance matrix eigenvalues. Comm. Statist. Simul. Comput. 2023, 52, 4313–4325. [Google Scholar] [CrossRef]
  17. Nasuda, R.; Shimizu, K.; Hashiguchi, H. Asymptotic behavior of the distributions of eigenvalues for beta-Wishart ensemble under the dispersed population eigenvalues. Commun. Stat. Theory Methods 2023, 52, 7840–7860. [Google Scholar] [CrossRef]
  18. Ahn, J.; Marron, J.S.; Muller, K.M.; Chi, Y. The high-dimension, low-sample-size geometric representation holds under mild conditions. Biometrika 2007, 94, 760–766. [Google Scholar] [CrossRef]
  19. Bolivar-Cime, A.; Perez-Abreu, V. PCA and eigen-inference for a spiked covariance model with largest eigenvalues of same asymptotic order. Braz. J. Probab. Stat. 2014, 28, 255–274. [Google Scholar] [CrossRef]
  20. Jung, S.; Marron, J.S. PCA consistency in high dimension, low sample size context. Ann. Stat. 2009, 37, 4104–4130. [Google Scholar] [CrossRef]
  21. Sugiyama, T.; Ushizawa, K. A non-parametric method to test equality of intermediate latent roots of two populations in a principal component analysis. J. Jpn. Stat. Soc. 1998, 28, 227–235. [Google Scholar] [CrossRef]
  22. Takeda, Y. Permutation test for equality of each characteristic root in two populations. J. Jpn. Soc. Comput. Stat. 2001, 14, 1–10. [Google Scholar] [CrossRef]
  23. Johnstone, I.M. On the distribution of the largest eigenvalue in principal components analysis. Ann. Stat. 2001, 29, 295–327. [Google Scholar] [CrossRef]
  24. Iimori, T.; Ogura, T.; Sugiyama, T. On the distribution of the second-largest latent root for certain high dimensional Wishart matrices. Int. J. Knowl. Eng. Soft Data Paradig. 2013, 4, 187–197. [Google Scholar] [CrossRef]
Figure 1. n i = 10 ( i = 1 , 2 ) and m = 30 .
Figure 1. n i = 10 ( i = 1 , 2 ) and m = 30 .
Mathematics 12 00921 g001
Table 1. Percentile points of the distributions of 1 and 2 of W 50 ( 10 , Σ ) (Case 1).
Table 1. Percentile points of the distributions of 1 and 2 of W 50 ( 10 , Σ ) (Case 1).
α F sim 1 ( α ) F 1 1 ( α ) α F sim 1 ( α ) F 2 1 ( α )
0.9923.235923.20930.9921.79121.666
0.9518.302618.3070.9517.060116.919
0.9015.982515.98720.9014.837714.6837
0.509.344669.341820.508.486768.34283
0.053.943893.94030.053.477963.32511
Table 2. Percentile points of the distributions of 1 and 2 of W 50 ( 10 , Σ ) (Case 2).
Table 2. Percentile points of the distributions of 1 and 2 of W 50 ( 10 , Σ ) (Case 2).
α F sim 1 ( α ) F 1 1 ( α ) α F sim 1 ( α ) F 2 1 ( α )
0.9923.23923.20930.9922.128521.666
0.9518.30618.3070.9517.407916.919
0.9015.985715.98720.9015.185614.6837
0.509.348449.341820.508.843948.34283
0.053.947443.94030.053.865663.32511
Table 3. Approximate probabilities of 1 based on the empirical percentile points (Case 1).
Table 3. Approximate probabilities of 1 based on the empirical percentile points (Case 1).
nm G 1 ( x 0.90 ) G 1 ( x 0.95 ) G 1 ( x 0.99 )
5200.9000470.9500110.990165
300.9002270.9500020.990018
400.9001730.9501850.990072
1000.8999520.9498740.990048
15200.9004090.9503750.990088
300.9002580.9502320.990124
400.9007290.9504640.990213
1000.9003310.9500390.990083
Table 4. Approximate probabilities of 2 based on the empirical percentile points (Case 1).
Table 4. Approximate probabilities of 2 based on the empirical percentile points (Case 1).
nm G 2 ( x 0.90 ) G 2 ( x 0.95 ) G 2 ( x 0.99 )
5200.9048680.9522950.990482
300.9055120.9530160.990761
400.9058380.9529240.990609
1000.9063930.953150.990613
15200.9031180.9515870.990331
300.9032750.9518560.990335
400.9039430.9520620.990315
1000.9040270.9521630.990498
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shimizu, K.; Hashiguchi, H. Chi-Square Approximation for the Distribution of Individual Eigenvalues of a Singular Wishart Matrix. Mathematics 2024, 12, 921. https://doi.org/10.3390/math12060921

AMA Style

Shimizu K, Hashiguchi H. Chi-Square Approximation for the Distribution of Individual Eigenvalues of a Singular Wishart Matrix. Mathematics. 2024; 12(6):921. https://doi.org/10.3390/math12060921

Chicago/Turabian Style

Shimizu, Koki, and Hiroki Hashiguchi. 2024. "Chi-Square Approximation for the Distribution of Individual Eigenvalues of a Singular Wishart Matrix" Mathematics 12, no. 6: 921. https://doi.org/10.3390/math12060921

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop