Next Article in Journal
Radio Mean Labeling Algorithm, Its Complexity and Existence Results
Previous Article in Journal
Nonlocal Nonlinear Fractional-Order Sequential Hilfer–Caputo Multivalued Boundary-Value Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Limiting Distribution of the Spectra of Random Block Matrices

by
Alexander N. Tikhomirov
Institute of Physics and Mathematics of FRC “Komi Science Center of Ural Branch of RAS”, Syktyvkar 167982, Russia
Mathematics 2025, 13(13), 2056; https://doi.org/10.3390/math13132056
Submission received: 31 March 2025 / Revised: 14 May 2025 / Accepted: 2 June 2025 / Published: 20 June 2025

Abstract

The behavior of the spectra of symmetric block-type random matrices with symmetric blocks of high dimensionality is considered in this paper. Under minimal conditions regarding the distributions of matrix block elements (Lindeberg conditions), the universality of the limiting empirical distribution function of block-type random matrices is shown.

1. Introduction

In the theory of high-dimensional random matrices, one of the main objects to study is the empirical spectral distribution function of high-dimensional random matrices and its approximation. The most studied matrices are ensembles of symmetric real or Hermite matrices with independent elements (semicircular law) and sample covariance matrices (Marchenko–Pastur distribution law). We consider ensembles of random matrices of a rather wide class—matrices with a block structure. The study of the spectra of the block-type random matrices began at the end of the last century and the beginning of the 2000s. One of the first works in this field was Girko’s paper [1]. He considered matrices of increasing dimensions with block elements of fixed dimensions and proved various laws for the limiting spectral distribution of such matrices. In the first decade of the 21st century, a number of papers appeared focused on the study of the spectra of block matrices with blocks of increasing dimensions. See, for example, the works [2,3,4,5,6]. The study of such matrix models is gaining particular popularity in connection with various applications in machine learning and neural networks. See, for instance, [7,8].
One such example is communication channels with multiple transmitting and receiving antennas, the so-called MIMO (Multi-Input, Multi-Output) channels. The channel matrix has the form
A = A 1 A 2 A L O O O O A 1 A 2 A L O O O O A 1 A 2 A L O O O O A 1 A L
where A 1 , A 2 , , A L are matrices of size n × n and A ν = ( h i j ν ) i , j = 1 n , where h i j ( ν ) reflects the channel’s effect on the signal transmitted from antenna j of the transmitter and received at antenna i of the receiver.
In what follows, we denote matrices and vectors in bold. To calculate the capacity of a channel, one needs to know the distribution of the eigenvalues of the matrix, as follows:
H = O A A * O ,
where O denotes a matrix with zero entries and A * denotes the complex conjugate of A . For details, see, e.g., [4]. Some more examples of block matrix applications can be found in spectral graph theory. Adjacency matrices and Laplace matrices of random multipartite graphs are examples of block-type random matrices; see [9,10,11]. Block random matrix models are applicable to the study of products of random matrices due to so-called linearization. The simplest example is sample covariance matrices, where instead of matrices of the form X X * , matrices of the form H = O X X * O are studied; see [12] and the references therein for more details. The asymptotics of the spectra of the products of random matrices are studied, for example, when analyzing the asymptotics of the information capacity of communication channels; see e.g., [13,14]. Block matrices play a fundamental role in neural networks, particularly in large-scale deep learning, optimization, and structured architectures. Their applications span distributed training, sparse computation, attention mechanisms, and optimization theory, making them essential for modern machine learning systems. Recent work, such as that by Xiao-Kai An et al. [15], has explored nonconvex optimization techniques for analyzing spectral distributions in random matrices, which have implications for neural network training dynamics.
The development of methods for studying the spectra of random classical matrix ensembles (Wigner random matrices, sampled covariance matrices) in recent years has led to an increased interest in studying the spectra of random matrices with block structures. See, for example, [6,16,17,18,19,20,21,22,23,24,25,26] and references therein.

1.1. List of Notations

We denote matrices by capital letters in bold, (column) vectors by lowercase letters in bold, and matrix elements by the same letters as matrices (as a rule) but in regular font.
The symbol I denotes the unit matrix; when it is necessary to specify the dimensions, we provide a lower index; e.g., I n 1 means a unit matrix of dimensions n 1 .
With R A , we denote the resolvent matrix for matrix A , i.e., R A = ( A z I ) .
With e j , j = 1 , , k , we denote the standard base vectors in R k (j-th component equals one, all the others are zeros).
With E ( i j ) , we denote the matrix defined as E ( i j ) = e i e j T , i.e., all the elements of E ( i j ) are zeros, except for E i j ( i j ) = 1 .
For any vector a and any matrix A , we denote the transposed vector a and the transposed matrix A by a T and A T , respectively. For a complex matrix A , we denote the Hermitian conjugate of A by A * , i.e., A * = ( A ¯ ) T .
We denote the indicator function by I { · } .

1.2. Model Representation

Consider the following model of block random matrices. Let X ( i j ) , i , j = 1 , , k , be a family of n × n random matrices. We will assume that for any fixed pair of indices (i, j), i , j = 1 , , k , the elements of matrix X ( i , j ) are independent (with respect to symmetry) and have a mean of zero and finite variance, i.e.,
E X p q ( i , j ) = 0 and E ( X p q ( i , j ) ) 2 = ( σ p q ( i , j ) ) 2 < .
We will also assume that the matrices X ( i , j ) either coincide at ( i , j ) ( i 1 , j 1 ) or are independent.
Let the symbol ⊗ denote the Kronecker product. Consider a matrix of the form
W X = i , j = 1 k E ( i , j ) X ( i , j ) .
We find it more convenient to change the representation of (2). Since we have assumed that some of the matrices X ( i , j ) may coincide, we define the so-called “generating matrices” Z ( 1 ) , , Z ( m ) (with m k 2 ), which are random and independent, such that for any pair of indices ( i , j ) : i , j = 1 , , k , there exists some l = l ( i , j ) : 1 l m with X ( i , j ) = Z ( l ) . Suppose, for l = 1 , , m ,
A l = { ( i , j ) i , j = 1 , , k : X ( i , j ) = Z ( l ) } .
Then the matrix W can be rewritten as
W = l = 1 m ( i , j ) A l E ( i , j ) Z ( l ) .

1.3. Toeplitz and Hankel Random Matrices

The Toeplitz and Hankel random matrices occupy a special place in the study of structured matrices. In [3], the existence of a limiting spectral distribution for random heats and Hankel random matrices was proven. In particular, it was proven that the limiting distribution of the Toeplitz and Hankel matrices has infinite support and differs from the normal distribution (the variance is 1, but the fourth moment is 8 / 3 ). Oraby in [5] showed that for circulant block-type Toeplitz matrices with random blocks of high dimensionality, the limiting spectral distribution is a mixture of semicircular laws with variances depending on the dimensionality of the original Töplitz matrix. In [27], Oraby’s result was obtained under weaker conditions for the distribution of block elements. In [28], it was shown that for a palindromic Toeplitz matrix with Wigner blocks, the limiting spectral distribution is normal. In [18,29], symmetric block Toeplitz and Hankel matrices were considered. Let the generating matrices be Z ( 1 ) , , Z ( k ) . Define the multiplicity sets as A l = { ( i , j ) : | i j | = l + 1 } , for l = 1 , , k . The symmetric Toeplitz matrix has the form
W = l = 1 k Z ( l ) i , j A l E ( i j ) = Z ( 1 ) , Z ( 2 ) Z ( k 1 ) Z ( k ) Z ( 2 ) Z ( 1 ) Z ( k 2 ) Z ( k 1 ) Z ( k ) Z ( k 1 ) Z ( 2 ) Z ( 1 ) .
Hankel block matrices and circulant block matrices can be represented in a similar way.

1.3.1. Hankel Matrices

In a Hankel matrix, the number of distinct blocks is m = 2 k 1 , and there are equal blocks on all diagonals perpendicular to the main diagonal, A l = { ( i , j ) : i + j = l + 1 } , l = 1 , , 2 k 1 :
W X = Z ( 1 ) 1 Z ( 2 ) 1 Z ( 3 ) 1 1 Z ( k 1 ) Z ( k ) 1 Z ( 2 ) 1 Z ( 3 ) 1 Z ( 4 ) 1 1 Z ( k ) 1 Z ( k + 1 ) Z ( 3 ) 1 Z ( 4 ) 1 1 1 1 1 1 1 1 1 Z ( k 1 ) Z ( k ) 1 1 1 Z ( 2 k 3 ) Z ( 2 k 2 ) Z ( k ) 1 Z ( k + 1 ) 1 1 Z ( 2 k 2 ) Z ( 2 k 1 ) .

1.3.2. Circulant Matrices

In a symmetric circulant matrix, the number of distinct blocks is
m = s + 1 , if k = 2 s , s , if k = 2 s 1 .
This is a special case of Toeplitz matrices with equal blocks on the diagonals, A l = { ( i , j ) : | i j | = l 1 k l + 1 } , l = 1 , , m :
W X = Z ( 1 ) Z ( 2 ) Z ( l ) 1 Z ( 3 ) Z ( 2 ) Z ( 2 ) Z ( 1 ) Z ( l 1 ) Z ( 4 ) Z ( 3 ) Z ( 3 ) Z ( 4 ) Z ( m 1 ) Z ( 1 ) Z ( 2 ) Z ( 2 ) Z ( 3 ) Z ( m ) Z ( 2 ) Z ( 1 ) .
The spectral analysis of random block matrices like Toeplitz or Hankel matrices helps to understand convergence for deep learning optimization; see, e.g., Xia et al. [8].

2. Main Result

Let V ( 1 ) , V ( 2 ) , , V ( m ) be Hermitian random matrices of size n × n whose elements are independent. Using the index sets introduced above (3), we define the following matrices:
Y ( i j ) = V ( l ) , for ( i , j ) A l , l = 0 , , m , W Y = i , j = 1 k E ( i j ) Y ( i j ) = l = 0 m ( i , j ) A l E ( i j ) V ( l ) .
Assume that the random variables Y p q ( i j ) with any fixed i , j = 1 < k , i j , are independent for p , q = 1 , , n , p q , and have the same first two moments as the random variables X p q ( i j ) ; that is,
E Y p q ( i j ) = 0 , and E ( Y p q ( i j ) ) 2 = ( σ p q ( i j ) ) 2 .
Denote the eigenvalues of matrix 1 n k W X as λ 1 , , λ n k and the eigenvalues of 1 n k W Y as μ 1 , , μ n k . We define the empirical spectral distribution functions
F n X ( x ) = 1 n k j = 1 n k I { λ j < x }
and
F n Y ( x ) = 1 n k j = 1 n k I { μ j < x } ,
where I { · } stands for the event indicator.
The empirical spectral distributions represent one of the main subjects of investigation in random matrix theory. We are interested in conditions regarding the distributions of X r s ( i j ) ( X r s ( i j ) ) that would guarantee that the distributions F n X ( x ) and F n Y ( x ) become infinitely close in the limit of large n. We prove this closeness under Lindeberg’s condition (see (7)). Lindeberg’s condition shows that the contribution to the variance of the sum of n 2 random variables and the values of individual random variables exceeding the n level is negligibly small compared to n 2 .
This condition is the optimal moment condition in many limit theorems, both for sums of independent random variables (the central limit theorem for instance) and in random matrix theory (Wigner’s semicircular law, as shown in [30], and the Marchenko–Pastur law for the empirical spectral distribution of sample covariance matrices; see [31]).
In what follows, for ( i , j ) A l , we shall write
σ p q ( i j ) = σ p q ( l ) .
Theorem 1. 
We assume that there exists a constant C 0 such that for all n 1 ,
1 ( n k ) 2 l = 1 m p , q = 1 n ( σ p q ( l ) ) 2 C 0 .
Suppose that for all i , j = 1 , , k for random matrices X ( i j ) and Y ( i j ) , Lindeberg’s condition is satisfied, i.e., for any τ > 0 ,
L n ( τ ) : = max { 1 n 2 r , s = 1 n E X r s ( i j ) 2 I { | X r s ( i j ) | > τ n } , 1 n 2 r , s = 1 n E Y r s ( i j ) 2 I { | Y r s i j | > τ n } , } n 0 .
Then
F n X ( x ) F n Y ( x ) n 0 in probability .
Remark 1. 
Note that if for some δ > 0 , there exists a constant μ 2 + δ > 0 such that
sup i , j , p , q 1 max { E | Y p q ( i , j ) | 2 + δ , E | X p q ( i , j ) | 2 + δ } μ 2 + δ ,
then Lindeberg’s condition is satisfied. This follows from the Markov inequality. To find the limit distribution of F n X ( x ) , we can compute it for F n Y ( x ) , where Y p q ( i j ) are Gaussian, for instance.
Remark 2. 
Figure 1, Figure 2, Figure 3 and Figure 4 show histograms of the distribution of the eigenvalues of the Hankel (Figure 1) and Teuplitz (Figure 2) matrices with different distributions of block elements. We consider distributions with polynomially decreasing tails (stepped tails with decreasing order | x | k , where k = 100 b in Figure 1A and Figure 2A and k = 7 / 2 in Figure 1B and Figure 2B, respectively. Figure 3 and Figure 4 show the histograms of the Teuplitz (Figure 3) and Hankel (Figure 4) matrices with block elements distributed according to Student’s law with five degrees of freedom (Figure 3A and Figure 4A) and with block elements distributed according to the normal law.
It was proven in [5] that there exists a limiting distribution G ( x ) = lim n F n X ( x ) for the i.i.d. matrix elements Z ( l ) , for l = 1 , , m , and E | Z j k ( l ) | 4 < . Since we can choose matrices with identically distributed Gaussian elements as a sequence of matrices Y ( l ) , we immediately obtain the following result as a consequence of the Theorem 1.
Theorem 2. 
Let ( σ p q ( i , j ) ) 2 = σ 2 . Suppose that for all i , j = 1 , , k , Lindeberg’s condition for random variables X p q ( i j ) and Y p q ( i j ) is satisfied, i.e., for any τ > 0 ,
L n ( τ ) : = 1 n 2 r , s = 1 n E X r s i j 2 I { | X r s i j | 2 > τ n } n 0 .
Then there exists a distribution function, G ( x ) , depending only on the structure of the block matrix, such that
lim n F n X ( x ) = G ( x ) .
The proof follows obviously from Theorem 1 above and Theorem 1 in [5].

Rectangular Blocks

Consider a block random matrix model with rectangular blocks. In this case, matrices X ( i j ) satisfy the relation X ( i j ) = ( X ( j i ) ) * (where, for any matrix A , its Hermite conjugate A ¯ T is denoted by A * ). The dimensions of the matrix X ( i j ) are n i × n j , respectively. Consider the class of generating matrices H 1 , , H k , Z 1 , Z 2 , , Z m , Z 1 * , Z 2 * , , Z m * , where the diagonal blocks H 1 , , H k are Hermitian matrices. The sets A l and A l * are defined by the equations
A l = { ( i , j ) i , j = 1 , , k : X ( i j ) = Z l , A l * = { ( i , j ) i , j = 1 , , k : X ( i j ) = Z l * } .
Clearly, ( i , j ) A l if and only if ( j , i ) A l * . We then have the following representation:
W X = i = 1 k e i e i T H i + l = 1 m ( i , j ) A l e i e j T Z ( l ) + ( i , j ) A l * e i e j T Z ( l ) * .
Since we consider the case where the matrix W is Hermitian, i.e., W X = W X * , we have a number of restrictions on the matrix block dimensions. For example, the matrix
W X = H 1 X ( 12 ) X ( 13 ) X ( 12 ) * H 2 X ( 23 ) X ( 13 ) * X ( 23 ) * H 3
can be represented as
W X = ( e 1 e 1 T H 1 + e 2 e 2 T H 2 + e 3 e 3 T H 3 ) + ( e 1 e 2 T ) X ( 12 ) + ( e 2 e 1 T ) X 12 * + ( e 1 e 3 T ) X 13 + ( e 3 e 1 T ) X ( 13 * + ( e 2 e 3 T ) X ( 23 ) + ( e 3 e 2 T ) X ( 23 ) * .
Matrices H ( i ) are independent Hermitian matrices of order N i × N i . Assume that the elements of H ( i ) = ( h p q ( i ) ) p , q = 1 N i are independent (except for H ( i ) being Hermitian) and
E h p q ( i ) = 0 , E | h p q ( i ) | 2 = [ σ p q ( i , 0 ) ] 2 < .
For matrices Z ( 1 ) , Z ( 2 ) , , Z ( m ) , we also assume that the elements are independent and
E [ Z ( l ) ] p q = 0 , E | [ Z ( l ) ] p q | 2 = [ σ p q ( l ) ] 2 < .
We denote
n = N 1 + N 2 + + N k .
Figure 5 and Figure 6 show the histograms of random matrices in block form with rectangular blocks with different distributions of block elements. The figures illustrate the independence of the limiting spectral distribution of block random matrices with rectangular blocks from the block element distributions. In these notations, the matrix W X has dimension n × n .
Now, let D ( 1 ) , , D ( l ) and Y ( 1 ) , , Y ( m ) be random matrices whose dimensions coincide with those of H ( 1 ) , , H ( k ) and Z ( 1 ) , Z ( 2 ) , , Z ( m ) and whose elements are independent and have moments of the first two orders, coinciding with those of the elements of the above-mentioned matrices. Denote the matrix obtained from W X by replacing blocks H 1 , , H k , Z ( 1 ) , Z ( 2 ) , , Z ( m ) with blocks D ( 1 ) , , D ( l ) , Y ( 1 ) , , Y ( m ) , respectively, by W Y . We denote the eigenvalues of W X by λ 1 , λ 2 λ n in decreasing order. We then have the empirical spectral distribution function
F n X ( x ) = 1 n j = 1 n I { λ j x } ,
We denote the eigenvalues of matrix W Y by μ 1 μ 2 μ n , and again, we have the empirical spectral distribution function
F n Y ( x ) = 1 n j = 1 n I { μ j x } .
Theorem 3. 
Assume that there exists a constant C 0 such that for all n 1 ,
1 ( n k ) 2 l = 1 m ( i , j ) A l p = 1 N i q = 1 N j ) ( σ p q ( i , j ) ) 2 + 1 n 2 l = 1 k p , q = 1 N l σ p q ( l ) C 0 .
Suppose that for all i , j = 1 , , k Lindeberg’s condition is satisfied, i.e., for any τ > 0 ,
L n ( τ ) : = 1 n 2 l = 1 m ( i , j ) A l r = 1 n i s = 1 n j E X r s i j 2 I { | X r s i j | 2 > τ n } + 1 n 2 l = 1 k p , q = 1 n l E [ h p q ( l ) ] 2 I { | h p q ( l ) | > τ n } n 0 .
Then
F n X ( x ) F n Y ( x ) n 0 in probability .
Remark 3. 
Figure 5 and Figure 6 show the histograms of random matrices in block form with rectangular blocks with different distributions of block elements. The figures illustrate the independence of the limiting spectral distribution of block random matrices with rectangular blocks from the block element distributions.
For many block-type random matrices (both with square blocks and rectangular blocks), we know the limiting empirical spectral distributions in the case when the block elements are Gaussian values and the variances [ σ p q ( i j ) ] 2 satisfy certain conditions (e.g., they are all equal). Theorems 1 and 3 show that the same limit distributions will hold for blocks with arbitrary element distributions, as long as Lindeberg’s conditions are satisfied. We can consider several examples.
We first consider two simple ones, the Hermitian matrix X = 1 n ( X p q ) p , q = 1 n and a rectangular matrix X ^ of dimensions n × m , whose elements are independent. Consider the following block matrices:
W = O X O X O X O X O , W ^ = O X ^ O X ^ * O X ^ O X ^ * O ,
It is well known that if the random variables X i j are standard Gaussian and identically distributed, then the empirical spectral distribution function of matrix X is a standard semicircular law, i.e., the distribution density is
p ( x ) = 1 2 π 4 x 2 I { | x | 2 } .
It is easy to see that
det { ( W X z I } = z n det ( 2 X 2 z 2 I ) .
It follows that the limiting empirical spectral distribution is a mixture of a 1 3 atom at zero and a distribution with density
p 1 ( x ) = 1 4 π 8 x 2 I { | x | 2 2 }
and mass 2 3 , so that the density of the distribution can be written as
p ˜ ( x ) = 1 3 δ 0 ( x ) + 2 3 p 1 ( x ) ,
where δ 0 ( x ) is the Dirac delta function. From Theorem 1, if for random variables X p q , p , q = 1 , , n , we have E [ X p q | 2 = 1 and Lindeberg’s condition is satisfied, then the empirical spectral distribution function of W X has a density p ˜ ( x ) . (Matrices of this type have also been considered in [12].) In a similar way, we can consider the case of a rectangular block. Assume that lim n n m = y ( 0 , 1 ) . Applying the formula for the determinant of a block matrix, we obtain
det { W ^ z I } = ( z ) n det { 2 X ^ * X ^ z 2 I } 1
From here, one easily sees that the eigenvalues of W X ^ are zero and of multiplicity 2 m n and ± s j , j = 1 , , n , where s j are singular numbers of X ^ . For example, in the case where X i j is Gaussian with variance, the relation
lim n 1 2 n j = 1 n ( I { s j x } + I { s j x } ) G y ( x ) , at n ,
where G y ( x ) is the function of the symmetrized Marchenko–Pastur distribution with the parameter y, i.e., a distribution with density
g y ( x ) = 1 2 π y x ( x 2 a 2 ) ( b 2 a 2 I { a | x | b } ,
where a = ( 1 y ) 2 , b = ( 1 + y ) 2 . From Theorem 3, we obtain the following: if E [ X ^ i j ] 2 = 1 and Lindeberg’s condition is satisfied, then the limiting distribution of the spectral distribution function of the matrix W ^ will have the form
p ^ ( x ) = 2 y 2 + y δ 0 ( x ) + 2 y 2 + y 1 2 g y ( x / 2 ) .
Results similar to Theorem 3 were obtained in [32]. In contrast to our paper, in [32], the proof was carried out using the method of moments, and instead of condition (6), the condition
sup j , l , p , q ( σ p q ( j l ) ) 2 C 0
was considered.
Furthermore, in [32], it was assumed that all blocks X ( i j ) , i , j = 1 , , k , were independent, which rules out the case of coincident blocks, such as in the case of Hankel or Töplitz matrices. Finally, in [32], it was assumed that σ p q ( i j ) = σ p q , i.e., the variance within blocks is the same. One cannot, for example, apply the results of [32] to the matrix W = O X X O , where X is a Wigner matrix (Hermitian matrices with independent elements (to Hermitian accuracy) with equal variance) whose limiting spectral distribution is a semicircular law (cf. Corollary 2 in [32]).
Based on Theorems 1 and 3 and the results of Theorems 2 and 4 in [12], we can formulate some theorems on the convergence of the empirical spectral distribution functions of block-type random matrices. For this purpose, we need one more notation. We denote the covariance of the random elements of the matrices X ( i j ) and X ( p q ) by σ i , j ; p . q . By virtue of our agreement that the matrices X ( i j ) and X ( p q ) are either independent or coincident, the values of the function σ i , j ; p . q are either 0 or 1. It is obvious that
σ i , j ; p , q = σ p , q ; i , j .
Let M k ( C ) be a set of k × k matrices with complex elements. We define the mapping η : M k ( C ) M k ( C ) as follows. For a matrix D = ( d i j ) i , j = 1 k M k ( C ) , we have η ( D ) = ( η ( D ) i j ) i , j = 1 k :
[ η ( D ) ] i j = 1 k p , q = 1 k σ ( i , j ; p , q ) d p q .
Theorem 4. 
Suppose that the random variables X p q ( i j ) , where i , j = 1 , , k and p , q = 1 , , n , satisfy the following conditions:
(1) E X p q ( i j ) = 0 and E | X p q ( i j ) | 2 = σ 2 ;
(2) Lindeberg’s condition: for any τ > 0 , we have
L n ( τ ) : = 1 n 2 i , j = 1 k p , q = 1 n E X p q ( i j ) | 2 I { | X p q ( i j ) | > τ n } 0 when n .
Then there exists a distribution function F ( x ) with a Stieltjes transform S ( z ) that satisfies the equality S ( z ) = G ( z ) / k , where the function G ( z ) is an analytic function taking values in the space of square matrices of dimensions k × k , defined in the upper complex half-plane and uniquely defined by the properties
lim | z | , Im z > 0 z G ( z ) = I k
and
z G ( z ) = I k + η ( G ( z ) ) G ( z ) ) .

3. Proof of the Main Result

We only need to show the convergence of the corresponding Stieltjes transforms in any subset of the upper complex half-plane with a non-empty interior. Consider the resolvent R A ( z ) of a symmetric matrix A . This is defined for all z = u + i v with v > 0 . We only need to prove that in some region G C + with a non-empty interior, there is the convergence
1 n k Tr R W X ( z ) 1 n k Tr R W Y 0 for n
in probability.
We shall divide the proof into three parts. First, we show that we can replace random variables with so-called truncated random variables. Then, we prove that for truncated quantities, the difference between the expectations of the Stiltjes transformations of the matrices W X and W Y tends to 0. Finally, we show that the variance of the Stiltjes transformations tends to 0 (Girko’s lemma). The convergence of expectations and the convergence of variances to zero entail the convergence of probabilities.

3.1. Truncation

As a first step in the proof, we reduce the problem to truncated random variables. For some τ > 0 , we introduce the quantities
X ˜ p q ( l ) = X p q ( l ) I { | X p q ( l ) | τ n } .
Next, consider their centered version
X ^ p q ( l ) = X ˜ p q ( l ) E X ˜ p q ( l )
and the normalized version
X ˘ p q ( l ) = σ p q ( l ) σ ^ p q ( l ) X ^ p q ( l ) ,
where [ σ ^ p q ( l ) ] 2 = E [ X ^ p q ( l ) ] 2 . Note that
E X ˘ p q ( l ) = 0 , E [ X ˘ p q ( l ) ] 2 = [ σ p q ( l ) ] 2 .
Consider the matrices W ˜ X , W ^ X , and W ˘ X obtained from W X by replacing X p q ( l ) with X ˜ p q ( l ) , X ^ p q ( l ) , and X ˘ p q ( l ) , respectively. Also, consider the resolvent matrices R ˜ W X ( z ) , R ^ W X ( z ) , and R ˘ W X ( z ) .
Lemma 1. 
Under the conditions of Theorem 1, we have
lim n 1 n k Tr R ˘ W X 1 n k Tr R W X = 0 i n p r o b a b i l i t y .
Proof. 
We start by estimating the quantity
B 1 : = 1 n k Tr R ˜ W X ( z ) 1 n k Tr R W X ( z ) .
The following inequality holds:
E W X W ˜ X 2 2 1 n k l = 1 m ( i , j ) A l p , q = 1 n E | X p q ( i j ) | 2 I { | X p q ( i j ) | > τ n } .
Then, applying Cauchy’s inequality, it is easy to obtain that
E | B 1 | 1 n k E l , r = 1 n k | j = 1 n k [ R ( z ) ] j l [ R ˜ ( z ) ] j r | 2 1 2 l , r = 1 k n E | W l r W ˜ l r | 2 1 2 .
It is then not hard to see that
l , r = 1 n k | j = 1 n k [ R ( z ) ] j l [ R ˜ ( z ) ] j r | 2 = l , r = 1 n k | [ R ( z ) R ˜ ( z ) ] l r | 2 = R ( z ) R ˜ ( z ) 2 2 n k v 4 .
Combining inequalities (24) and (25), we obtain the following:
E | B 1 | C n 2 l = 1 m p , q = 1 n E ( X p q ( l ) ) 2 I { | X p q ( l ) | > τ n } 1 2 = 1 v 2 L n 1 2 ( τ ) .
We obtain a similar estimate for the quantity
E | B 2 | : = E | 1 n k Tr R ˜ ( z ) ] X 1 n k Tr R ^ X ( z ) | 1 v 2 L n 1 2 ( τ ) .
Let us now estimate the value
B 3 : = 1 n k Tr R ˘ X ( z ) 1 n k Tr R ^ X ( z ) .
We have
E | B 3 | 1 n k E R ^ X ( z ) R ˘ X ( z ) R ˘ X ( z ) 2 , E W ˘ W ^ 2 1 n k v 2 E W ˘ W ^ 2 = 1 n k v 2 E 1 2 l = 1 m ( i j ) A l p , q = 1 n | X ^ p q ( i j ) X ˘ p q ( i j ) | 2 .
Obviously,
l = 1 m ( i j ) A l p , q = 1 m E | X ^ p q ( i j ) X ˘ p q ( i j ) | 2 l = 1 m ( i j ) A l p , q = 1 m | σ ^ p q ( i j ) ) 2 ( σ p q ( i j ) ) 2 | C ( n k ) 2 L n ( τ ) .
Substituting the last estimate into (29), we obtain
E | B 3 | C v 2 L n 1 2 ( τ ) ,
and therefore,
E | 1 n k Tr R ˘ W X ( z ) 1 n k Tr R W X ( z ) | C v 2 L n 1 2 ( τ ) .
Lemma 1 has been proven. □
Remark 4. 
Since the function L n ( τ ) at fixed τ is monotone on τ and L n ( τ ) 0 at n for any fixed τ, there exists a sequence τ n such that τ n 0 and L n ( τ n ) 0 at n . In what follows, we write τ = τ n 0 such that L n ( τ n ) 0 as n . Without loss of generality, we can assume that the random variables X p q ( i j ) , i , j = 1 , , k ; p , q = 1 , n , and Y p q ( i j ) , i , j = 1 , , k ; p , q = 1 , n , satisfy the condition
max { | Y p q ( i j ) | , | X p q ( i j ) | } τ n .

3.2. Special Representation for the Difference Between Two Resolvents

For any α [ 0 , π / 2 ] , consider the random matrix
Z ( l ) ( α ) = X l cos α + Y l sin α .
We denote
W X ( α ) : = 1 n k l = 1 m ( i , j ) A l E ( i , j ) Z ( l ) ( α ) , R X ( z , α ) = ( W ( α ) z I ) 1 .
In what follows, we denote the elements of W X ( α ) = ( W j k ) and Z ( l ) ( α ) = ( Z j k ( l ) ) simply by W and Z ( l ) , omitting α and X from notations (unless this is ambiguous). Obviously, we have
1 n k Tr R W Y ( z ) 1 n k Tr R W X ( z ) = 1 n k Tr R W X ( z , π 2 ) 1 n k Tr R W X ( z , 0 ) = 1 n k 0 π 2 Tr R W X ( z , α ) α d α = 1 n k j = 1 n k 0 π 2 [ R W X ] j j ( z , α ) α d α .
We can now write
[ R W X ] j j ( z , α ) α = l = 1 m p , q = 1 n [ R W X ] j j ( z , α ) Z p q ( l ) Z p q ( l ) α .
Note that for any invertible matrix A , the following differentiation formula is valid:
A j k 1 A p q = [ A 1 ] j p [ A 1 ] q k
(see, for example, [33]). Applying (32) to the resolvent R W X ( z , α ) , we obtain
[ R W X ] j j ( z , α ) Z p q ( l ) = ( r , s ) A l ( 2 δ p q ) [ R W X ( z , α ) ] j , ( r 1 ) n + p [ R W X ( z , α ) ] j , ( s 1 ) n + q ,
where δ p q stands for the Kronecker symbol. Also, note that
Z p q ( l ) α = X p q ( l ) sin α + Y p q ( l ) cos α = : Z ˜ p q ( l ) .
Summing up the above formulas, we can write
E ( 1 n k Tr R W Y ( z ) 1 n k Tr R W X ( z ) ) = 1 n k n k j = 1 n k l = 1 m p , q = 1 n ( 2 δ p q ) 0 π 2 E Z ˜ p q ( l ) ( α ) × ( r , s ) A l ( 2 δ p q ) [ R W X ( z , α ) ] j , ( r 1 ) n + p [ R W X ( z , α ) ] j , ( s 1 ) n + q d α .
In what follows, we are going to estimate the values
D p q ( l , r , s ) = E Z ˜ p q ( l ) [ R W X ( z , α ) ] j , ( r 1 ) n + p [ R W X ( z , α ) ] j , ( s 1 ) n + q
for l = 1 , , m ; r , s A l and p , q = 1 , , n . We introduce the matrices Z ( l , p , q ) = Z ( l ) 1 n k Z p q ( l ) ( E ( p , q ) + E ( q , p ) ) (i.e., the matrix Z ( l , p , q ) has zeros instead of each Z p q ( l ) ). In the same way, we define the matrices W ( l , p , q ) ( z , α ) and R W X ( l , p , q ) ( z , α ) . The simple resolvent equality R A 1 R A 2 = R A 1 ( A 2 A 1 ) R A 2 shows that
R W X ( z , α ) = R W X ( l , p , q ) ( z , α ) R W X ( z , α ) ( W W ( l , p , q ) ( z , α ) ) R W X ( l , p , q ) ( z , α ) .
Let u j , j = 1 , , n , denote the standard base vectors (column vectors) of the space R n . It is easy to see that we can then write
W ( α ) W ( l , p , q ) ( α ) = 1 n k Z p q ( l ) r , s A l E ( r , s ) ( u p u q T + u q u p T )
Now, let v j , j = 1 , , k n , be the corresponding basis vectors for R k n . Then we can write
E ( r , s ) ( u p u q T + u q u p T ) = v ( r 1 ) n + p v ( s 1 ) n + q T + v ( r 1 ) n + q v ( s 1 ) n + p T .
And, further,
[ R W X ( z , α ) ] j , ( r 1 ) n + p = [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 ) n + p 1 n k Z p q ( l ) r 1 , s 1 A l v j T R W X ( z , α ) × ( v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v ( s 1 1 ) n + p T ) R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p , R W X ( z , α ) ] j , ( s 1 ) n + q = R W X ( l , p , q ) ( z , α ) j , ( s 1 ) n + q 1 n k Z p q ( l ) r , s A l v j T R W X ( z , α ) × v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v s 1 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( s 1 ) n + q .
Now applying (34) to the matrix R W X on the right-hand side of (36), we arrive at the following equality:
[ R W X ( z , α ) ] j , ( r 1 ) n + p = [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 ) n + p 1 n k Z p q ( l ) r 1 , s 1 A l ( v j T R W X ( l , p , q ) ( z , α ) ( v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 q ) k + q v ( s 1 q ) k + p T ) × R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p ) + 1 n k ( Z p q ( l ) ) 2 [ T 1 ( l , p , q ) ] j , r , s , R W X ( l , p , q ) ( z , α ) ] j , ( s 1 ) n + q = [ R W X ( l , p , q ) ( z , α ) ] j , ( s 1 ) n + q 1 n k Z p q ( l ) r 1 , s 1 A l ( v j T R W X ( z , α ) ( v ( r 1 ) n + p v ( s 1 ) n + q T + v ( r q ) k + q v s q ) k + p T ) × R W X ( l , p , q ) ( z , α ) v ( s 1 ) n + q ) + 1 n k ( Z p q ( l ) ) 2 [ T 2 ( l , p , q ) ] j , r , s ,
where
[ T 1 ( l , p , q ) ] j , r , s = r 1 , s 1 A l r 2 , s 2 A l v j T R W X ( z , α ) v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v s 1 1 ) k + p T × R W X ( l , p , q ) ( z , α ) v ( r 2 1 ) n + p v ( s 2 1 ) n + q T + v ( r 2 1 ) n + q v s 2 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( s 1 ) n + q , ] j , r , s 2 ( l , p , q ) = r 1 , s 1 A l v j T R W X ( z , α ) v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v s 1 1 ) k + p T × R W X ( l , p , q ) ( z , α ) v ( r 2 1 ) n + p v ( s 2 1 ) n + q T + v ( r 2 1 ) n + q v s 2 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p .
Note that
E Z ˜ ( l ) ( α ) Z p q ( l ) ( α ) = E ( ( X p q ( l ) ) 2 ( Y p q ( l ) ) 2 ) sin α cos α = 0 .
Through the independence of the quantities X p q ( l ) and Y p q ( l ) and the matrices R W X ( l , p , q ) ( z , α ) , we obtain
E Z ˜ ( l ) ( α ) [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] j , ( s 1 ) n + q = 0 , E Z ˜ ( l ) ( α ) Z p q ( l ) ( α ) v j T R W X ( l , p , q ) ( z , α ) × v ( r 1 ) n + p v ( s 1 ) n + q T + v ( r 1 ) n + q v s 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p = 0 .
Equalities (33), (36), and (38) imply that
E ( 1 n k Tr R W Y ( z ) 1 n k Tr R W X ( z ) ) = 0 π 2 ( Q 1 + + Q 6 ) d α ,
where
Q 1 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 2 × v j T R W X ( l , p , q ) ( z , α ) v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v ( s 1 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p × v j T R W X ( l , p , q ) ( z , α ) v ( r 2 1 ) n + p v ( s 2 1 ) n + q T + v ( r 2 1 ) n + q v ( s 2 1 ) n + p T ) R W X ( l , p , q ) ( z , α ) v ( s 1 ) n + q ,
Q 2 = 1 ( n k ) 3 j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 3 × v j T R W X ( l , p , q ) ( z , α ) v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v s 1 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( r 1 ) n + p [ T 2 ( l , p , q ) ] j , r , s , Q 3 = 1 ( n k ) 3 j = 1 n k l = 1 m p , q = 1 n ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 3 × r , s A l ( 2 δ p q ) [ T 1 ( l , p , q ) ] j , r , s × v j T R W X ( l , p , q ) ( z , α ) v ( r 1 1 ) n + p v ( s 1 1 ) n + q T + v ( r 1 1 ) n + q v s 1 1 ) k + p T R W X ( l , p , q ) ( z , α ) v ( s 1 ) n + q , Q 4 = 1 ( n k ) 3 n k j = 1 n k l = 1 m p . q = 1 n ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 4 × ( r , s ) A l ( 2 δ p q ) [ T 1 ( l , p , q ) ] j , r , s [ T 2 ( l , p , q ) ] j , r , s , Q 5 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p . q = 1 n ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 ) n + p [ T 2 ( l , p , q ) ] j , r , s , Q 6 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p . q = 1 n ( 2 δ p q ) E Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( s 1 ) n + q [ T 1 ( l , p , q ) ] j , r , s .

3.3. Estimations of Quantities Q 1 Q 6

3.3.1. Estimation of Q 1

We represent Q 1 in the form
Q 1 = Q 11 + + Q 14 ,
where
Q 11 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ p q ( l ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + q , ( s 1 ) n + q , Q 12 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ p q ( l ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + p , ( s 1 ) n + q ,
Q 13 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ p q ( l ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + p , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + q , ( s 1 ) n + q , Q 14 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ p q ( l ) ( Z p q ( l ) ) 2 × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + p , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + p , ( s 1 ) n + q
There is an obvious inequality for ν = 2 , 3 , 4 :
1 ( n ) ν 1 E | Z ˜ p q ( l ) ( Z p q ( l ) ) ν | C τ ν 1 ( σ p q ( l ) ) 2 .
This follows from condition (30).
All values of Q 1 ν , ν = 1 , , 4 , are estimated similarly. Let us first consider the estimation of Q 11 . Note that the number of elements in sets A l for l = 1 , , m is independent of n and does not exceed k 2 . Given the independence of X p q ( l ) , Y p q ( l ) , and the matrices R W X ( l , p , q ) , we can write
Q 11 = 1 ( n k ) 2 n k j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( 2 δ p q ) E Z ˜ p q ( l ) ( Z p q ( l ) ) 2 × E [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + q , ( s 1 ) n + q .
Applying inequality (40), we get
| Q 11 | C τ 1 ( n k ) 2 j = 1 n k l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l r 2 , s 2 A l ( σ p q ( l ) ) 2 × E | [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + q , ( s 1 ) n + q |
The following inequality is valid
j = 1 n k | [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p | j = 1 n k | [ R W X ( l , p , q ) ( z , α ) ] j , ( r 1 1 ) n + p | 2 1 2 j = 1 n k | R W X ( l , p , q ) ( z , α ) ] j , ( r 2 1 ) n + p | 2 1 2 1 v 2
Here, we apply Cauchy’s inequality and use the fact that for any resolvent matrix R A = ( A z I ) 1 constructed by a symmetric real or Hermite matrix A = ( A i j ) i , j = 1 n , the operator norm at the point z = u + i v does not exceed R A 1 v , and hence the sum of squares of the elements of any row does not exceed 1 v 2 , i.e., for any i = 1 , , n ,
j = 1 n | [ R A ] i j | 2 1 v 2 .
In addition,
max { | [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p | , [ R W X ( l , p , q ) ( z , α ) ] ( s 2 1 ) n + q , ( s 1 ) n + q | } 1 v .
Given these inequalities, we arrive at the following estimation:
| Q 11 | C τ v 4 1 ( n k ) 2 l = 1 m p , q = 1 n ( σ p q ( l ) ) 2 ,
Similarly, we obtain estimates for ν = 2 , 3 , 4 :
| Q 1 ν | C τ v 4 1 ( n k ) 2 l = 1 m p , q = 1 n ( σ p q ( l ) ) 2 .
Hence, from condition (6),
| Q 1 | C τ v 4 .
The estimates of Q 2 , Q 3 , and Q 4 are slightly different due to the dependence of the values X p q ( l ) and Y p q ( l ) and the matrices R W X . But this problem is easily circumvented. All estimates associated with the elements of the resolvent are fulfilled uniformly over the entire space where the random variables X p q ( l ) and Y p q ( l ) are defined.

3.3.2. Estimation of Q 2

To estimate Q 2 , we note that there exists a constant C such that
| [ T 2 ( l , p , q ) ] j , r , s | C v 3 r 1 A l | [ R W X ( z , α ) ] j , ( r 1 1 ) n + p | .
Applying this inequality, we obtain
| Q 2 | C 1 v 4 ( n k ) 3 l = 1 m p , q = 1 n ( r , s ) A l r 1 , s 1 A l E | Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 3 | × j = 1 n k | R W X ( l , p , q ) ( z , α ) j , ( r 1 1 ) n + p | [ R W X ( z , α ) ] j , ( r 1 ) n + p | ,
Using inequalities (42) and (43), we obtain
| Q 2 | C 1 v 6 ( n k ) 3 l = 1 m p , q = 1 n E | Z ˜ ( l ) ( α ) ( Z p q ( l ) ) 3 | .
Inequalities (40) and (16) and the last inequality together imply that
| Q 2 | C τ 2 v 5 .

3.3.3. Estimation of Q 3

Let us rewrite the definition of the quantity [ T 1 l , p , q ] j , r , s taking into account equality (35):
[ T 1 ( l , p , q ) ] j r s = [ T 11 ( l , p , q ) ] j r s + + [ T 14 ( l , p , q ) ] j r s ,
where
[ T 11 ( l , p , q ) ] j , r , s = r 1 , s 1 A l [ R W X ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] ( s 1 ) n + q , ( s 1 ) n + q , [ T 12 ( l , p , q ) ] j , r , s = r 1 , s 1 A l [ R W X ( z , α ) ] j , ( r 1 1 ) n + p [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + q , ( r 1 ) n + q × [ R W X ( l , p , q ) ( z , α ) ] ( s 1 ) n + p , ( s 1 ) n + q , [ T 13 ( l , p , q ) ] j , r , s = r 1 , s 1 A l [ R W X ( z , α ) ] j , ( r 1 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + p , ( r 1 ) n + p × [ R W X ( l , p , q ) ( z , α ) ] ( s 1 ) n + q , ( s 1 ) n + q , [ T 14 ( l , p , q ) ] j , r , s = r 1 , s 1 A l [ R W X ( z , α ) ] j , ( r 1 1 ) n + q [ R W X ( l , p , q ) ( z , α ) ] ( s 1 1 ) n + p , ( r 1 ) n + q × [ R W X ( l , p , q ) ( z , α ) ] ( s 1 ) n + p , ( s 1 ) n + q .
Similarly to inequality (49), we obtain the estimate
max { | [ T 11 ( l , p , q ) ] j , r , s | , | [ T 12 ( l , p , q ) ] j , r , s | , | [ T 13 ( l , p , q ) ] j , r , s | , | [ T 14 ( l , p , q ) ] j , r , s | } C v 2 r 1 , s 1 A l ( [ | R W X ( z , α ) ] j , ( r 1 1 ) n + p | + | [ R W X ( z , α ) ] j , ( r 1 1 ) n + q | ) .
Inequalities (50), (41), and (42) together imply that
| Q 3 | C τ 2 ( n k ) 2 v 5 l = 1 m p , q = 1 n σ p q ( l ) C τ 2 v 5 .

3.3.4. Estimation of Q 4

Applying inequalities (50), (41), (42), and (46), we obtain the estimate
| Q 4 | C τ 3 v 6 .

3.3.5. Estimation of Q 5

We use inequalities (41), (42), and (46) again. We obtain
| Q 5 | C τ v 4 .

3.3.6. Estimation of Q 6

Similarly to Q 5 , we apply inequalities (41), (42), and (45). We obtain
| Q 6 | C τ v 4 .
Estimations (45), (49), and (51)–(54) together with the representation in (39) imply that
| 1 n k E Tr R X ( z , α ) 1 n k E Tr R Y | C τ v 4 + C τ 3 v 6 + C τ 2 v 5 .
Without loss of generality, we can assume that τ < 1 . The final estimate can be written as
| 1 n k E Tr R X ( z , α ) 1 n k E Tr R Y | C τ ,
where the constant C depends on v, k, and m. Considering the transition to truncated values, the final estimate is as follows:
| 1 n k E Tr R X ( z , α ) 1 n k E Tr R Y | C ( τ + ( L n ( τ ) 1 2 ) .
According to Remark 4, we can choose a sequence τ n such that τ n 0 and L n ( τ n ) 0 as n . It follows that
lim n E 1 n k Tr R W X ( z , α ) E 1 n k Tr R W Y = 0 .
To complete the proof, we give the following lemma.

3.4. Girko’s Lemma

Lemma 2. 
Under conditions of Theorem 1, the following inequality holds:
E | 1 n k ( Tr R W X E Tr R W X ) | 2 k 2 n v 2 .
Proof. 
We prove this lemma using the method proposed by Girko (with a slight modification). For l = 1 , , m and j = 1 , , n , we define σ -algebras M ( l , j ) = σ { X p q ( s ) , s < l ; p , q j } . Note that M ( 1 , 1 ) = { , Ω } is a trivial σ -algebra and M m , n is a σ -algebra with respect to which all random variables X p q ( l ) for l = 1 , ˙ m , 1 p , and q n are measurable. It is obvious that for any fixed l = 1 , , m , the σ -algebras of M ( l , j ) are non-decreasing for j = 1 , , n . Suppose that
η l j = 1 n k ( E { Tr R W X ( z ) | M ( l , j ) } E { Tr R W X ( z ) | M ( l , j 1 ) } .
When l = 1 , , m is fixed, the random variables are uncorrelated. It is easy to see that
1 n k Tr R W X ( z ) E Tr R W X ( z ) = l = 1 n j = 1 n η l j .
The Cauchy inequality implies that
E | 1 n k Tr R W X ( z ) E Tr R W X ( z ) | 2 m ( n k ) 2 l = 1 m E | j = 1 n η l j | 2 = m ( n k ) 2 l = 1 m j = 1 n E | η l j | 2 .
For any fixed l = 1 , , , m and j = 1 , , , n , define the matrix W ( l , j ) obtained from the matrix W by removing, for any p , q A l , rows with numbers ( p 1 ) n + j and columns with numbers ( q 1 ) n + j . Note that
E { Tr R W X ( l , j ) | M l , j ) } E { Tr R W X | M ( l , j 1 ) } = 0 .
Hence, we have the representation
η l j = 1 n k ( E { Tr R W X ( z ) Tr R W X ( l , j ) | M ( l , j ) } E { Tr R W X ( z ) Tr R W X ( l , j ) | M ( l , j 1 ) } .
Now, we use the inequality
| Tr R W X ( z ) Tr R W X l , j ) | | A l | v k 2 v ,
which, together with (59), gives the estimate
E | 1 n k Tr R W X ( z ) E Tr R W X ( z ) | 2 k 2 n v 2 .
The lemma has been proven. □
Lemma 2, Remark 4, and relation (58) now complete the proof of the theorem.

4. Conclusions

Block-type random matrices play an important role in many studies. In this case, special attention is paid to the limiting distribution of the spectrum of such types of matrices. This is a difficult problem, but for special distributions of matrix elements, it is often possible to solve this problem. The universality of the limiting spectral distribution (its independence from the distribution of matrix elements) allows us to choose matrices with elements that have a distribution with certain good properties, for which special methods can be used, for example, for Gaussian distributions. However, this is not always possible. For example, even in the Gaussian case, there is no exact description of the limiting spectral distribution for the Teuplice matrix. It is known that it has an infinite carrier and is non-Gaussian (the second moment is 1 and the fourth moment is 8 / 3 ; see, for instance, [3]).
Many important problems in random matrix theory remain unsolved for block-type random matrices. First of all is the study of the distribution of eigennumbers in the local regime (convergence rate, rigidity, delocalisation of eigenvectors, and so on), and then the distribution of extreme values of eigennumbers, and so on.

Funding

This work was carried out within the framework of the project “Mathematical problems of the theory of stochastic and deterministic systems, including high-dimensional systems” (project no. 122040600066-5).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the author.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Girko, V. Random block matrix density and SS-Law. Random Oper. Stock. Equ. 2000, 8, 189–194. [Google Scholar] [CrossRef]
  2. Bolla, M. Distribution of the eigenvalues of random block-matrices. Linear Algebra Its Appl. 2004, 377, 219–240. [Google Scholar] [CrossRef]
  3. Bruc, W.; Dembo, A.; Iang, T.J. Spectral Measure of Large Random Hankel, Markov and Toeplitz Matrices. Ann. Probab. 2006, 34, 1–38. [Google Scholar] [CrossRef]
  4. Far, R.; Oraby, T.; Bruc, W.; Speicher, R. Spectra of large block matrices. Eur. J. Pure Appl. Math. 2024, 17, 2550–2561. [Google Scholar] [CrossRef]
  5. Oraby, T. The Spectral Laws og Hermitian Block-Matrices with Large Blocks. Elect. Comm. Probab. 2007, 12, 465–476. [Google Scholar]
  6. Kologlu, M.; Kopp, G.S.; Miller, S.J. The limiting spectral measure for ensemble of symmetric block circulant matrices. arXiv 2010, arXiv:1008.4812v5. [Google Scholar] [CrossRef][Green Version]
  7. Granzio, D.; Zohren, S.; Roberts, S. Learning Rates as a Function of Batch Size: A Random Matrix Theory Approach to Neural Network Training. J. Mach. Learn. Res. 2022, 23, 1–65. [Google Scholar]
  8. Xia, J.; Li, S.; Yang, Z.; Jaimoukha, I.; Gunduz, D. Meta-learning based Altarnating Minimization Algorithm for Non-convex Optimization. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 5366–5380. [Google Scholar] [CrossRef]
  9. El Karoui, N. Graph connection Laplacian and random matrices with random blocks. Inf. Inference 2015, 4, 1–42. [Google Scholar] [CrossRef][Green Version]
  10. Nadutkina, A.V.; Tikhomirov, A.N.; Timushev, D.A. Marchenko–Pastur law for the spectrum of a random weighted bipartite graph. Sib. Adv. Math. 2024, 34, 146–153. [Google Scholar] [CrossRef]
  11. Avrachenkov, K.; Cottatellucci, L.; Kadavankandy, A. Spectral Properties of Random Matrices for Stochastic Block Model. In Proceedings of the 13th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt), Mumbai, India, 25–29 May 2015; pp. 537–544. [Google Scholar] [CrossRef]
  12. Speicher, R. Free probability and random matrices. arXiv 2014, arXiv:14043393v1. [Google Scholar]
  13. Fawaz, N.; Zarifi, K.; Debbah, M.; Gesbert, D. Asymptotic Capacity and Optimal Precoding in MIMO Multi-Hop Relay Networks. IEEE Trans. Inf. Theory 2011, 57, 2050–2069. [Google Scholar] [CrossRef]
  14. Müller, R. On the Asymptotic Eigenvalue Distribution of Concatenated Vector-Valued Fading Channels. IEEE Trans. Inf. Theory 2002, 48, 2086–2091. [Google Scholar] [CrossRef]
  15. An, X.; Du, L.; Jiang, F.; Zhang, Y.; Deng, Z.; Kurths, J. A few-shot identification method for stochastic dynamical systems on residual multieaks adaptive sampling. Caos 2024, 34, 073118. [Google Scholar] [CrossRef]
  16. Pfaffel, O.; Schlemm, E. Limiting Spectral Distribution of a New Random Matrix Model with Dependece across Rows and Columns. Linear Algebra Its Appl. 2012, 436, 2966–2979. [Google Scholar] [CrossRef]
  17. Aljadeff, J.; Renfrew, D.; Stern, M. Eigenvalues of block structured asymmetric random matrices. J. Math. Phys. 2015, 56, 103502. [Google Scholar] [CrossRef]
  18. Basu, R.; Bose, A.; Ganguly, S.; Hazra, R.S. Limiting Spectral Distribution of Block matrices with Töplitz Block Structure. arXiv 2011, arXiv:1111.1901v1. [Google Scholar]
  19. Beckwith, O.; Luo, V.; Miller, S.J.; Shen, K.; Triantofillou, N. Distribution of eigenvalues of weighted structured matrix ensembles. arXiv 2015, arXiv:1112.3719v2. [Google Scholar]
  20. Blackwell, K.; Borade, N.; Vi, C.D.; Luntzlara, N.; Ma, R.; Miller, S.J.; Wang, M.; Xu, W. Distribution of eigenvalues of random real symmetric block matrices. arXiv 2019, arXiv:1908.03834v4. [Google Scholar]
  21. Cicuta, G.M.; Pernici, M. Sparse Random Block Matrices. J. Phys. A Math. Theor. 2021, 55, 175202. [Google Scholar] [CrossRef]
  22. Dette, H.; Reuther, B. Random Block Matrices and Matrix Orthogonal Polynomials. J. Theor. Probab. 2010, 23, 378–400. [Google Scholar] [CrossRef]
  23. Guhlich, M.; Nagel, J.; Dette, H. Random block matrices generalizing the classical Jacobi and Laguerre ensembles. J. Multivar. Anal. 2010, 101, 1884–1897. [Google Scholar] [CrossRef]
  24. Dunn, T.; Fleishmann, H.L.; Jackson, F.; Khunger, S.; Miller, S.J.; Reifenberg, L.; Shashkov, A.; Willis, S. Limiting Spectral Distirbutions of Families of Block Matrix Ensembles. arXiv 2022, arXiv:2109.01464v1. [Google Scholar]
  25. Bogomolny, E.; Giraud, O. Statistical properties of structured random matrices. arXiv 2021, arXiv:2012.14322v1. [Google Scholar] [CrossRef] [PubMed]
  26. Krueger, T.; Renfrew, D. Singularity degree of structured random matrices. In Annales de l’Institut Henri Poincare (B) Probabilites et Statistiques; Institut Henri Poincaré: Paris, France, 2025; Volume 61, pp. 1416–1442. [Google Scholar] [CrossRef]
  27. Tikhomirov, A.N.; Timushev, D.A.; Gulyaeva, S.T. Limit Theorems for Spectra of Circulant Block Matrices with Large Random Blocks. Mathematics 2024, 12, 2291. [Google Scholar] [CrossRef]
  28. Blackwell, K.; Borade, N.; Bose, A.; Vi, C.D.; Luntzlara, N.; Ma, R.; Miller, S.J.; Mukherjee, S.S.; Wang, M.; Xu, W. Distribution of eigenvalues of matrix ensembles arising from Wigner and palindromic Toeplitz blocks. arXiv 2021, arXiv:2102.05839v1. [Google Scholar]
  29. Li, Y.; Liu, D.; Wang, Z. Limit Distributions of Eigenvalues for Random Block Toeplitz and Hankel Matrices. J. Theor. Probab. 2011, 24, 1063–1086. [Google Scholar] [CrossRef]
  30. Chin, C.W. Necessary and Sufficient Conditions for Convergence to the Semicircle Distribution. Random Matrices Theory Appl. 2023, 12, 2250045. [Google Scholar] [CrossRef]
  31. Dong, Z.; Yao, J. Necessary and sufficient conditions for the Marcĕnko–Pastur law for sample correlation matrices. Stat. Probab. Lett. 2025, 221, 110377. [Google Scholar] [CrossRef]
  32. Ding, X. Spectral analysis of large block random matrices with rectangular blocks. Lith. Math. J. 2014, 54, 115–126. [Google Scholar] [CrossRef]
  33. Khorunzhy, A.M.; Khoruzhenko, B.A.; Pastur, L.A. On asymptotic propertiies of large random matrices with independnet entries. J. Math. Phys. 1996, 37, 5033–5060. [Google Scholar] [CrossRef]
Figure 1. The histograms of 5 × 5 Hankel matrices with block entries distributed according to the polynomial densities p ( x ) = C ( 1 + | x | ) 100 (A) and p ( x ) = C ( 1 + | x | ) 7 2 (B).
Figure 1. The histograms of 5 × 5 Hankel matrices with block entries distributed according to the polynomial densities p ( x ) = C ( 1 + | x | ) 100 (A) and p ( x ) = C ( 1 + | x | ) 7 2 (B).
Mathematics 13 02056 g001
Figure 2. The histograms of 5 × 5 Toeplitz matrices with block entries distributed according to the polynomial densities p ( x ) = C ( 1 + | x | ) 100 (A) and p ( x ) = C ( 1 + | x | ) 7 2 (B).
Figure 2. The histograms of 5 × 5 Toeplitz matrices with block entries distributed according to the polynomial densities p ( x ) = C ( 1 + | x | ) 100 (A) and p ( x ) = C ( 1 + | x | ) 7 2 (B).
Mathematics 13 02056 g002
Figure 3. The histograms of 5 × 5 Töplitz matrices with block entries distributed according to Student’s distribution with d f = 5 degrees of freedom (A) and a standard normal distribution (B).
Figure 3. The histograms of 5 × 5 Töplitz matrices with block entries distributed according to Student’s distribution with d f = 5 degrees of freedom (A) and a standard normal distribution (B).
Mathematics 13 02056 g003
Figure 4. The histograms of 5 × 5 Hankel matrices with block entries distributed according to Student’s distribution with d f = 5 degrees of freedom (A) and a standard normal distribution (B).
Figure 4. The histograms of 5 × 5 Hankel matrices with block entries distributed according to Student’s distribution with d f = 5 degrees of freedom (A) and a standard normal distribution (B).
Mathematics 13 02056 g004
Figure 5. The histogram of 3 × 3 matrices with rectangular blocks N 1 = 5000 , N 2 = 3333 , N 3 = 1666 . Student distribution with d f = 3 degrees of freedom with different variances for different blocks.
Figure 5. The histogram of 3 × 3 matrices with rectangular blocks N 1 = 5000 , N 2 = 3333 , N 3 = 1666 . Student distribution with d f = 3 degrees of freedom with different variances for different blocks.
Mathematics 13 02056 g005
Figure 6. The histograms of 3 × 3 matrices with rectangular blocks: N 1 = 5000 , N 2 = 3333 , and N 3 = 1666 . Student’s distribution with d f = 5 degrees of freedom (A) and normal distribution (B) with different variances for different blocks.
Figure 6. The histograms of 3 × 3 matrices with rectangular blocks: N 1 = 5000 , N 2 = 3333 , and N 3 = 1666 . Student’s distribution with d f = 5 degrees of freedom (A) and normal distribution (B) with different variances for different blocks.
Mathematics 13 02056 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tikhomirov, A.N. On the Limiting Distribution of the Spectra of Random Block Matrices. Mathematics 2025, 13, 2056. https://doi.org/10.3390/math13132056

AMA Style

Tikhomirov AN. On the Limiting Distribution of the Spectra of Random Block Matrices. Mathematics. 2025; 13(13):2056. https://doi.org/10.3390/math13132056

Chicago/Turabian Style

Tikhomirov, Alexander N. 2025. "On the Limiting Distribution of the Spectra of Random Block Matrices" Mathematics 13, no. 13: 2056. https://doi.org/10.3390/math13132056

APA Style

Tikhomirov, A. N. (2025). On the Limiting Distribution of the Spectra of Random Block Matrices. Mathematics, 13(13), 2056. https://doi.org/10.3390/math13132056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop