Next Article in Journal
On the Classification of Entire Solutions to the Critical Lane–Emden Equation
Previous Article in Journal
Approximations of the Euler–Maruyama Method of Stochastic Differential Equations with Regime Switching
Previous Article in Special Issue
Using a Node–Child Matrix to Address the Quickest Path Problem in Multistate Flow Networks under Transmission Cost Constraints
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geary’s c for Multivariate Spatial Data

Graduate School of Humanities and Social Sciences, Hiroshima University, 1-2-1 Kagamiyama, Higashi-Hiroshima 739-8525, Japan
Mathematics 2024, 12(12), 1820; https://doi.org/10.3390/math12121820
Submission received: 11 May 2024 / Revised: 3 June 2024 / Accepted: 6 June 2024 / Published: 12 June 2024
(This article belongs to the Special Issue Graph Theory and Network Theory)

Abstract

:
Geary’s c is a prominent measure of spatial autocorrelation in univariate spatial data. It uses a weighted sum of squared differences. This paper develops Geary’s c for multivariate spatial data. It can describe the similarity/discrepancy between vectors of observations at different vertices/spatial units by a weighted sum of the squared Euclidean norm of the vector differences. It is thus a natural extension of the univariate Geary’s c. This paper also develops a local version of it. We then establish their properties.

1. Introduction

Let y i denote the realization of a single variable y at the vertex/spatial unit v i for i = 1 , , n . Spatial autocorrelation is a notion that describes the similarity/discrepancy between observations at different vertices. One way to measure it between y i and y j is to use the squared difference:
( y i y j ) 2 .
This approach was adopted by Geary [1]. He proposed a spatial autocorrelation measure given by a binary weighted sum of ( y i y j ) 2 for i , j = 1 , , n , which was modified by Cliff and Ord [2,3,4,5]. It is now called Geary’s c and is one of the leading measures of spatial autocorrelation (Geary’s c is a spatial extension of von Neumann [6] ratio. See, e.g., [7]). Subsequently, Anselin [8] developed a local version of Geary’s c, local Geary’s c. To distinguish them, Geary’s c is sometimes referred to as global Geary’s c.
Let y i = [ y ( 1 ) , i , , y ( p ) , i ] denote the realization of multiple variables y = [ y ( 1 ) , , y ( p ) ] at the vertex/spatial unit v i for i = 1 , , n . How can we measure the similarity/discrepancy between y i and y j , which are both p-dimensional column vectors? A natural approach is to extend (1). That is, it can be measured with the following squared Euclidean norm of the vector difference:
y i y j 2 .
In this paper, we develop a spatial autocorrelation measure for multivariate spatial data, y 1 , , y n R p . It uses y i y j 2 . More specifically, as a global Geary’s c, we define it as a weighted sum of y i y j 2 for i , j = 1 , , n . This measure allows us to describe the similarity/discrepancy between vectors of observations at different vertices. In other words, we can use it to measure the spatial autocorrelation in multivariate spatial data. In addition, from it, we develop a local version of it, the multivariate local Geary’s c. Subsequently, we establish their properties. We show how our two multivariate autocorrelation measures relate to the two existing univariate autocorrelation measures, (univariate) global and local Geary’s c. We also note that our multivariate global Geary’s c is related to the recently popular graph learning.
Wartenberg [9], Jombart et al. [10], Anselin [11], Lin [12], and Yamada [13,14,15] are relevant to this paper. Among them, (i) Yamada [13,14] presented several results of univariate global Geary’s c using notions of spectral graph theory. This paper depends on them, and it can be regarded as an extension of them. In addition, (ii) Yamada [15] dealt with multivariate extension of Moran’s I, which was developed by Moran [16] and Cliff and Ord [2,3,4,5] and is another prominent measure of spatial autocorrelation (Recent studies on Moran’s I and its extension include [17,18,19,20,21,22,23,24,25]. Thus, this paper can be regarded as a companion paper to it. Finally, (iii) the relevance of [9,10,11,12] to this paper is discussed in the final section. This is mainly because we would like to use our notations below to describe their studies.
This paper is organized as follows: In Section 2, we provide some preliminaries for the following two sections. In Section 3, after defining global and local Geary’s c for multivariate spatial data, we establish their properties. In Section 4, we make additional remarks to the results of the previous section. In the section, we also mention how our multivariate global Geary’s c is related to the recently popular graph learning. Section 5 concludes. In the Appendix A and Appendix B, we provide some of the proofs and two MATLAB/GNU Octave user-defined functions.

2. Preliminaries

Some of the notations. Let I n be the identity matrix of order n and e n , i be the i-th column of I n , i.e., I n = [ e n , 1 , , e n , n ] . Let ι n be the n-dimensional vector of ones, i.e., ι n = [ 1 , , 1 ] R n and Q ι = I n ι n ( ι n ι n ) 1 ι n = I n 1 n ι n ι n . For a matrix Γ = [ γ i , j ] R r × s , Γ F denotes the Frobenius norm of Γ , i.e., Γ F = ( i = 1 r j = 1 s γ i , j 2 ) 1 2 . For a full column rank matrix Γ R r × s , denote the column space of Γ and its orthogonal complement by S ( Γ ) and S ( Γ ) , respectively.
Undirected graph. Let G = ( V , E ) be a simple undirected graph whose vertex set is V = { v 1 , , v n } , where n 2 . In addition, let W = [ w i , j ] R n × n , where
w i , j > 0 , { v i , v j } E , w i , j = 0 , { v i , v j } E .
Given that G = ( V , E ) is a simple undirected graph, W is a symmetric matrix such that w i , i = 0 for i = 1 , , n . We assume that | E | > 0 , i.e., G = ( V , E ) is not an empty graph. Accordingly, q = i = 1 n j = 1 n w i , j > 0 .
Multivariate/multiple spatial data. Let
Y = y ( 1 ) , 1 y ( p ) , 1 y ( 1 ) , n y ( p ) , n = y 1 y n = [ y ( 1 ) , , y ( p ) ] R n × p ,
where y ( h ) , i denotes the realization of variable y ( h ) at vertex v i . We refer to y 1 , , y n R p as multivariate spatial data and y ( 1 ) , , y ( p ) R n as multiple spatial data. We assume that i = 1 n ( y ( h ) , i y ¯ ( h ) ) 2 > 0 for h = 1 , , p , where y ¯ ( h ) = 1 n i = 1 n y ( h ) , i . That is, by assumption, y ( h ) does not belong to S ( ι n ) .
Standardized multiple/multivariate spatial data. Let z ( h ) , i = y ( h ) , i y ¯ ( h ) s ( h ) for i = 1 , , n and h = 1 , , p , where s ( h ) is the positive square root of 1 n 1 i = 1 n ( y ( h ) , i y ¯ ( h ) ) 2 . Then, let
Z = z ( 1 ) , 1 z ( p ) , 1 z ( 1 ) , n z ( p ) , n = z 1 z n = [ z ( 1 ) , , z ( p ) ] R n × p .
By construction, z ( h ) is related to y ( h ) as
z ( h ) = 1 s ( h ) Q ι y ( h ) = 1 s ( h ) ( y ( h ) y ¯ ( h ) ι n ) , h = 1 , , p .
Then, given that ι n Q ι = 0 , it follows that ι n z ( h ) = 0 , which implies that z ( h ) belongs to S ( ι n ) .
Geary’s c. Let c ( h ) denote the global Geary’s c of y ( h ) = [ y ( h ) , 1 , , y ( h ) , n ] :
c ( h ) = n 1 2 q i = 1 n j = 1 n w i , j ( y ( h ) , i y ( h ) , j ) 2 i = 1 n ( y ( h ) , i y ¯ ( h ) ) 2 , h = 1 , , p .
In addition, denote the local Geary’s c of y ( h ) at vertex v i by c ( h ) , i :
c ( h ) , i = n 1 2 q j = 1 n w i , j ( y ( h ) , i y ( h ) , j ) 2 i = 1 n ( y ( h ) , i y ¯ ( h ) ) 2 , i = 1 , , n , h = 1 , , p .
Then, c ( h ) and c ( h ) , i can be represented using z ( h ) , i as
c ( h ) = 1 2 q i = 1 n j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 , h = 1 , , p ,
c ( h ) , i = 1 2 q j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 , i = 1 , , n , h = 1 , , p .
By construction, it follows that
c ( h ) = i = 1 n c ( h ) , i .
Graph Laplacian. Let D = diag d 1 , , d n , where d i = j = 1 n w i , j for i = 1 , , n . Denote the graph Laplacian corresponding to W by L . Then, from, e.g., [26,27],
L = D W .
Global Geary’s c in matrix notation. From [13] (Proposition 1), c ( h ) in (5) can be represented using L as
c ( h ) = 1 q z ( h ) L z ( h ) , h = 1 , , p .
Spectral decomposition of graph Laplacian. Given that L in (8) is a real symmetric matrix, it can be spectrally decomposed as
L = U Λ U ,
where Λ = diag ( λ 1 , , λ n ) such that λ 1 λ n and U = [ u 1 , , u n ] is an orthogonal matrix. Given that L is a positive semidefinite matrix such that L ι n = 0 = 0 · ι n , we can let ( λ 1 , u 1 ) = 0 , 1 n ι n . Then, given that U is an orthogonal matrix, it follows that u 1 S ( ι n ) and u i S ( ι n ) for i = 2 , , n .
Graph Fourier transform. Let
F = U Z = [ U z ( 1 ) , , U z ( p ) ] .
U z ( h ) in (11) is referred to as graph Fourier transform of z ( h ) [28,29]. Let
F = f ( 1 ) , 1 f ( p ) , 1 f ( 1 ) , n f ( p ) , n = f 1 f n = [ f ( 1 ) , , f ( p ) ] R n × p .
Then, from (11), it follows that f ( h ) = U z ( h ) for h = 1 , , p . In addition, given that f ( h ) , i is the i-th entry of f ( h ) , it follows that f ( h ) , i = u i z ( h ) for i = 1 , , n . Since u 1 S ( ι n ) and z ( h ) S ( ι n ) , it follows that f ( h ) , 1 = u 1 z ( h ) = 0 for h = 1 , , p . Accordingly, f 1 , which denotes the first row of F in (11), is equal to 0 .
Weighted average representation of global Geary’s c. Let
c ( u i ) = n 1 q u i L u i u i Q ι u i , i = 2 , , n ,
which denotes the global Geary’s c of u i for i = 2 , , n . Then, given that u i L u i = e n , i Λ e n , i = λ i and u i Q ι u i = u i u i = 1 for i = 2 , , n , it follows that
c ( u i ) = n 1 q λ i , i = 2 , , n .
Accordingly, c ( h ) can be represented as
c ( h ) = 1 q z ( h ) L z ( h ) = 1 q z ( h ) U Λ U z ( h ) = 1 q f ( h ) Λ f ( h ) = 1 q i = 2 n λ i f ( h ) , i 2 , h = 1 , , p .
Then, substituting (13) into (14) yields
c ( h ) = i = 2 n ψ ( h ) , i c ( u i ) ,
where ψ ( h ) , i = f ( h ) , i 2 n 1 for i = 2 , , n . Here, since ψ ( h ) , i 0 and
i = 2 n ψ ( h ) , i = 1 ,
c ( h ) is an weighted average of c ( u 2 ) , , c ( u n ) . For a proof of (16), see Appendix A.1. Note that (15) is a restatement of [14] (Proposition 1(a)).
Bounds of global Geary’s c. From [30], it follows that
c ( h ) n 1 q λ 2 , n 1 q λ n , h = 1 , , p .

3. Multivariate Global and Local Geary’s c

We define the following measure as the global Geary’s c for multivariate spatial data, y 1 , , y n R p :
c = 1 2 p q i = 1 n j = 1 n w i , j z i z j 2 .
We refer to c as multivariate global Geary’s c. Similarly, we define the following measure as the local Geary’s c for multivariate spatial data at vertex v i , y i R p :
c i = 1 2 p q j = 1 n w i , j z i z j 2 , i = 1 , , n .
which we refer to as multivariate local Geary’s c. Then, by construction, it follows that
i = 1 n c i = i = 1 n 1 2 p q j = 1 n w i , j z i z j 2 = 1 2 p q i = 1 n j = 1 n w i , j z i z j 2 = c .
Recall that p denotes the number of variables and z i is the i-th row of Z , i.e., z i = [ z ( 1 ) , i , , z ( p ) , i ] . Accordingly, when p = 1 , given that z i reduces to z ( 1 ) , i for i = 1 , , n , it follows that
c = 1 2 q i = 1 n j = 1 n w i , j ( z ( 1 ) , i z ( 1 ) , j ) 2 = c ( 1 ) , c i = 1 2 q j = 1 n w i , j ( z ( 1 ) , i z ( 1 ) , j ) 2 = c ( 1 ) , i .
Thus, the multivariate global (resp. local) Geary’s c is a generalization of the univariate global (resp. local) Geary’s c.
The next two results document how the multivariate local (resp. global) Geary’s c relates to the univariate local (resp. global) Geary’s c’s.
Proposition 1. 
For i = 1 , , n , c i in (19) is equal to the simple average of c ( 1 ) , i , , c ( p ) , i , i.e.,
c i = 1 p h = 1 p c ( h ) , i .
Proof. 
See Appendix A.2. □
Proposition 2. 
c in (18) is equal to the simple average of c ( 1 ) , , c ( p ) , i.e.,
c = 1 p h = 1 p c ( h ) .
Proof. 
From (7), (20), and Proposition 1, we have
c = i = 1 n c i = i = 1 n 1 p h = 1 p c ( h ) , i = 1 p h = 1 p i = 1 n c ( h ) , i = 1 p h = 1 p c ( h ) .
Remark 1. 
By combining Proposition 2 and (7), c in (18) can be represented by c ( h ) , i as follows.
c = 1 p h = 1 p i = 1 n c ( h ) , i .
Next, we will document matrix form representations of the multivariate Geary’s c’s. In preparation for this, we show a matrix form representation of c ( h ) , i in (4). Let
L i = 1 2 d i e n , i e n , i e n , i e n , i W + 1 2 diag ( e n , i W ) .
Proposition 3. 
c ( h ) , i in (6) can be represented in matrix notation as
c ( h ) , i = 1 q z ( h ) L i z ( h ) .
Proof. 
See Appendix A.3. □
Now, we are ready to document matrix form representations of the multivariate Geary’s c’s.
Proposition 4. 
c i in (19) can be represented in matrix notation as
c i = 1 p q tr ( Z L i Z ) .
Proof. 
z ( h ) L i z ( h ) in (25) is the ( h , h ) -th entry of Z L i Z R p × p . In addition, from Proposition 1, it follows that c i = 1 p h = 1 p c ( h ) , i . Combining these results yields
c i = 1 p h = 1 p c ( h ) , i = 1 p q h = 1 p z ( h ) L i z ( h ) = 1 p q tr Z L i Z .
Proposition 5. 
c in (18) can be represented in matrix notation as
c = 1 p q tr Z L Z .
Proof. 
z ( h ) L z ( h ) in (9) is the ( h , h ) -th entry of Z L Z R p × p . In addition, from Proposition 2, it follows that c = 1 p h = 1 p c ( h ) . Combining these results yields
c = 1 p h = 1 p c ( h ) = 1 p q h = 1 p z ( h ) L z ( h ) = 1 p q tr Z L Z .
Remark 2. 
We make four remarks on Propositions 4 and 5.
(i) 
From (20), the next equation must follow:
i = 1 n L i = L .
For a proof of (28), see Appendix A.4.
(ii) 
Given that tr Z L i Z = vec ( Z ) ( I p L i ) vec ( Z ) (e.g., [31], p. 241), from Proposition 4, c i in (19) can be represented as a quadratic form:
c i = 1 p q vec ( Z ) ( I p L i ) vec ( Z ) .
Similarly, from Proposition 5, c in (18) can be represented as
c = 1 p q vec ( Z ) ( I p L ) vec ( Z ) .
(iii) 
Given that vec ( Z ) = [ z ( 1 ) , , z ( p ) ] , from (29) and (30), it immediately follows that
c i = 1 p q vec ( Z ) ( I p L i ) vec ( Z ) = 1 p h = 1 n 1 q z ( h ) L i z ( h ) = 1 p h = 1 n c ( h ) , i ,
c = 1 p q vec ( Z ) ( I p L ) vec ( Z ) = 1 p h = 1 n 1 q z ( h ) L z ( h ) = 1 p h = 1 n c ( h ) .
(31) (resp. (32)) can be regarded as another proof of Proposition 1 (resp. Proposition 2).
(iv)
User-defined MATLAB/GNU Octave function for calculating c i (resp. c) based on (26) (resp. (27)) is provided in Appendix B.1 (resp. Appendix B.2).
Denote the eigenvalues of 1 q Z L Z by θ 1 , , θ p . Then, since L is a positive semidefinite matrix, 1 q Z L Z is also a positive semidefinite matrix, from which we obtain θ i 0 for i = 1 , , p . From Proposition 5, c can be represented using these eigenvalues as follows:
Proposition 6. 
c in (18) is equal to the simple average of θ 1 , , θ p , i.e.,
c = 1 p i = 1 p θ i .
Proof. 
Given that 1 q Z L Z and diag ( θ 1 , , θ p ) are similar, it follows that 1 q tr Z L Z = i = 1 p θ i . Then, (33) immediately follows from Proposition 5. □
Again, from Proposition 5, we obtain the following result:
Proposition 7. 
c in (18) can be represented as
c = 1 p q B Z F 2 ,
where B R n × | E | is an incidence matrix corresponding to G = ( V , E ) .
Proof. 
Given that L = B B (e.g., [27], Proposition 2.3), from Proposition 5, we have tr ( Z L Z ) = tr Z B B Z = B Z F 2 . □
Example 1. 
Consider the case where E = { { v 1 , v 2 } , , { v n 1 , v n } } in G = ( V , E ) . In the case, G = ( V , E ) is a path graph, and thus q = 2 ( n 1 ) and we can let
B = 1 1 0 0 0 1 1 0 0 0 1 1 R ( n 1 ) × n .
L in the case is explicitly given, e.g., in [32] (p. 68) and [33] (p. 3233). Accordingly, given that B Z = [ z 2 z 1 , , z n z n 1 ] , from (34), it follows that
c = 1 2 p ( n 1 ) h = 1 p i = 2 n ( z ( h ) , i z ( h ) , i 1 ) 2 = 1 2 p ( n 1 ) i = 2 n z i z i 1 2 ,
which can be regarded as a multivariate von Neumann [6] ratio.
The next result documents that c is also a weighted average of c ( u 2 ) , , c ( u n ) .
Proposition 8. 
c in (18) can be represented as a weighted average of c ( u 2 ) , , c ( u n ) as follows:
c = i = 2 n ψ ¯ i c ( u i ) ,
where ψ ¯ i = 1 p h = 1 p ψ ( h ) , i for i = 2 , , n .
Proof. 
See Appendix A.5. □
Proposition 9. 
c in (18) belongs to the closed interval given by
n 1 q λ 2 , n 1 q λ n .
Proof. 
It immediately follows by combining the results in (17) and Proposition 2. □

4. Additional Remarks

In this section, we make additional remarks to the results obtained in the previous section.

4.1. Local Geary Matrix

Proposition 1, (7), and (23) motivate us to consider the following matrix:
C = c ( 1 ) , 1 c ( p ) , 1 c ( 1 ) , n c ( p ) , n R n × p .
Recall that c ( h ) , i denotes the local Geary’s c of y ( h ) at vertex v i . We refer to C as local Geary matrix. From (7), Proposition 1, and (23), c ( h ) , c i , and c can be represented using C as
Equation ( 7 ) : c ( h ) = ι n C e p , h , h = 1 , , p , Proposition 1 : c i = 1 p e n , i C ι p , i = 1 , , n , Equation ( 23 ) : c = 1 p ι n C ι p .

4.2. Relationship to Graph Signal Processing

As shown in Proposition 5, our multivariate global Geary’s c in (18) can be represented as 1 p q tr Z L Z . We note here that a similar expression has appeared in graph signal processing. Specifically, Dong et al. [34] proposed the following filtering method in our notation:
min X R n × p , L R n × n Y X F 2 + α tr ( X L X ) + β L F 2 ,
where L is a symmetric matrix such that tr ( L ) = n , L ι n = 0 , and its off-diagonal entries are nonpositive. α and β are positive regularization parameters. It is very interesting that tr ( X L X ) is included as a penalty term in the above constrained minimization problem. It is notable that X and L are simultaneously estimated by solving (40). See also [35,36].

5. Concluding Remarks

In this paper, we developed global Geary’s c for multivariate spatial data. It describes the similarity/discrepancy between vectors of observations at different vertices/spatial units by a weighted sum of the squared Euclidean norm of the vector differences, and is thus a natural extension of the global Geary’s c for univariate spatial data. In addition, from it, we developed local Geary’s c for multivariate spatial data. Subsequently, we established their properties. The results obtained are summarized in Propositions 1–9. We also mentioned how our multivariate global Geary’s c is related to the recently popular graph learning.
We make a remark. As mentioned in Section 1, Wartenberg [9], Jombart et al. [10], Anselin [11], and Lin [12] are relevant to this paper. Here, we would like to clarify the relationship between those papers and this paper.
First, Anselin [11] defined multivariate local Geary’s c as
h = 1 p c ( h ) , i or 1 p h = 1 p c ( h ) , i
in our notation. As shown in Proposition 1, the latter of (41) is identical to c i in (19). In this sense, it would be accurate to say that Anselin [11] developed the multivariate local Geary’s c, c i in (19), and we established its properties. Nevertheless, we emphasize here that we defined the multivariate local Geary’s c by (19) and derived it from the multivariate global Geary’s c in (18).
Second, as an extension of global Moran’s I, Wartenberg [9] developed the following spatial correlation matrix in our notations:
M = 1 q n n 1 Z W Z .
When p = 1 , given that z ( 1 ) = 1 s ( 1 ) Q ι y ( 1 ) and s ( 1 ) 2 = 1 n 1 y ( 1 ) Q ι y ( 1 ) , M reduces to the following scalar:
1 q n n 1 z ( 1 ) W z ( 1 ) = 1 q 1 s ( 1 ) 2 n n 1 y ( 1 ) Q ι W Q ι y ( 1 ) = n q y ( 1 ) Q ι W Q ι y ( 1 ) y ( 1 ) Q ι y ( 1 ) ,
which is the global Moran’s I for y ( 1 ) . Moreover, the multivariate global Moran’s I developed by [15] can be represented as
1 p tr ( M ) .
Following [9], Jombart et al. [10] used the spatial correlation matrix given by
1 2 n Y Q ι ( D 1 W + W D 1 ) Q ι Y .
Recalling that z ( h ) = 1 s ( h ) Q ι y ( h ) , where z ( h ) is the h-th column of Z , the main difference is the use of the sum of D 1 W and its transpose instead of W . Here, since D 1 W ι n = ι n , D 1 W is a row-standardized matrix. In addition, another extension of [9] was recently completed by Lin [12]. He developed the following spatial correlation matrix:
G = 1 q Z L Z .
Of these three spatial correlation matrices, G in (44) is the most relevant for this paper. This is because, from Proposition 5, the multivariate global Geary’s c, c in (18), can be represented using G as
c = 1 p tr ( G ) .
In other words, c in (18) is equal to the simple average of the eigenvalues of G (Proposition 6). Thus, our multivariate global Geary’s c can be regarded as a value obtained from the spatial correlation matrix, G .

Funding

The Japan Society for the Promotion of Science supported this work through KAKENHI (grant number: 23K013770A).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The author thanks three anonymous referees for their valuable comments. The usual caveat applies.

Conflicts of Interest

The author declare no conflict of interest.

Appendix A. Some of the Proofs

Appendix A.1. Proof of (16)

It follows that
i = 1 n ψ ( h ) , i = i = 1 n f ( h ) , i 2 n 1 = 1 n 1 i = 1 n ( u i z ( h ) ) 2 = 1 n 1 tr ( U z ( h ) z ( h ) U ) = 1 n 1 tr ( z ( h ) z ( h ) ) = z ( h ) z ( h ) n 1 = 1 .
In addition, given that f ( h ) , 1 = 0 , it follows that i = 1 n f ( h ) , i 2 n 1 = i = 2 n f ( h ) , i 2 n 1 . Therefore, we obtain (16).

Appendix A.2. Proof of Proposition 1

Given that z i = [ z ( 1 ) , i , , z ( p ) , i ] , it follows that z i z j = [ z ( 1 ) , i z ( 1 ) , j , , z ( p ) , i z ( p ) , j ] , from which we have z i z j 2 = h = 1 p ( z ( h ) , i z ( h ) , j ) 2 . Accordingly, we obtain
c i = 1 2 p q j = 1 n w i , j z i z j 2 = 1 2 p q j = 1 n w i , j h = 1 p ( z ( h ) , i z ( h ) , j ) 2 = 1 p h = 1 p 1 2 q j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 = 1 p h = 1 p c ( h ) , i .

Appendix A.3. Proof of Proposition 3

j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 in (6) can be represented as a quadratic form as follows:
j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 = j = 1 n w i , j ( z ( h ) , i 2 2 z ( h ) , i z ( h ) , j + z ( h ) , j 2 ) = j = 1 n w i , j z ( h ) , i 2 2 j = 1 n w i , j z ( h ) , i z ( h ) , j + j = 1 n w i , j z ( h ) , j 2 = d i z ( h ) , i 2 2 z ( h ) , i j = 1 n w i , j z ( h ) , j + j = 1 n w i , j z ( h ) , j 2 = d i z ( h ) e n , i e n , i z ( h ) 2 z ( h ) e n , i e n , i W z ( h ) + z ( h ) ( w i , 1 e n , 1 e n , 1 + + w i , n e n , n e n , n ) z ( h ) = z ( h ) d i e n , i e n , i 2 e n , i e n , i W + diag ( e n , i W ) z ( h ) = z ( h ) ( 2 L i ) z ( h ) .
Finally, given c ( h ) , i = 1 2 q j = 1 n w i , j ( z ( h ) , i z ( h ) , j ) 2 , dividing the above equation by 2 q yields (25).

Appendix A.4. Proof of (28)

Given that i = 1 n e n , i e n , i = I n and i = 1 n e n , i = ι n , we can prove (28) as
i = 1 n L i = i = 1 n 1 2 d i e n , i e n , i e n , i e n , i W + 1 2 diag ( e n , i W ) = 1 2 i = 1 n d i e n , i e n , i e n , 1 e n , 1 + + e n , n e n , n W + 1 2 diag e n , 1 + + e n , n W = 1 2 D W + 1 2 D = D W = L .

Appendix A.5. Proof of Proposition 8

From (15) and Proposition 2, it follows that
c = 1 p h = 1 p c ( h ) = 1 p h = 1 p i = 2 n ψ ( h ) , i c ( u i ) = i = 2 n 1 p h = 1 p ψ ( h ) , i c ( u i ) = i = 2 n ψ ¯ i c ( u i ) .
Here, given that ψ ( h ) , i 0 for h = 1 , , p , it follows that ψ ¯ i = 1 p h = 1 p ψ ( h ) , i 0 . In addition, since i = 2 n ψ ( h ) , i = 1 from (16), it follows that
i = 2 n ψ ¯ i = i = 2 n 1 p h = 1 p ψ ( h ) , i = 1 p h = 1 p i = 2 n ψ ( h ) , i = 1 .

Appendix B. User-Defined Functions

In this section, we provide two MATLAB/GNU Octave user-defined functions. One is a function for calculating c i based on (26). The other is a function for calculating c based on (27). We used MATLAB R2018b and GNU Octave version 6.2.0 to verify these user-defined functions.

Appendix B.1. A Function for Calculating ci Based on (26)

  • function c_i=calc_c_i (Y, W, i)
[n, p]= size (Y); Z=zscore (Y); I=eye (n); e_i=I (: , i);
D=diag (sum(W)); q=trace (D);
L_i=D(i , i)*e_i*e_i’/2 − e_i*W(i , :) + diag (W(i , :))/2;
c_i=trace (Z’*L_i *Z)/(p*q );
  • end

Appendix B.2. A Function for Calculating c Based on (27)

  • function c=calc _ c (Y,W)
p=size (Y, 2) ; Z=zscore (Y); D=diag (sum(W));
q=trace (D); L=D−W; c=trace (Z’*L*Z)/(p*q);
  • end

References

  1. Geary, R.C. The contiguity ratio and statistical mapping. Inc. Stat. 1954, 5, 115–145. [Google Scholar] [CrossRef]
  2. Cliff, A.D.; Ord, J.K. The problem of spatial autocorrelation. In Studies in Regional Science; Scott, A.J., Ed.; Pion: London, UK, 1969; pp. 25–55. [Google Scholar]
  3. Cliff, A.D.; Ord, J.K. Spatial autocorrelation: A review of existing and new measures with applications. Econ. Geogr. 1970, 46, 269–292. [Google Scholar] [CrossRef]
  4. Cliff, A.D.; Ord, J.K. Spatial Autocorrelation; Pion: London, UK, 1973. [Google Scholar]
  5. Cliff, A.D.; Ord, J.K. Spatial Processes: Models and Applications; Pion: London, UK, 1981. [Google Scholar]
  6. Von Neumann, J. Distribution of the ratio of the mean square successive difference to the variance. Ann. Math. Stat. 1941, 12, 367–395. [Google Scholar] [CrossRef]
  7. Yamada, H. Spatial smoothing using graph Laplacian penalized filter. Spat. Stat. 2024, 60, 100799. [Google Scholar] [CrossRef]
  8. Anselin, L. Local indicators of spatial association—LISA. Geogr. Anal. 1995, 27, 93–115. [Google Scholar] [CrossRef]
  9. Wartenberg, D. Multivariate spatial correlation: A method for exploratory geographical analysis. Geogr. Anal. 1985, 17, 263–283. [Google Scholar] [CrossRef]
  10. Jombart, T.; Devillard, S.; Dufour, A.-B.; Pontier, D. Revealing cryptic spatial patterns in genetic variability by a new multivariate method. Heredity 2008, 101, 92–103. [Google Scholar] [CrossRef]
  11. Anselin, L. A local Indicator of multivariate spatial association: Extending Geary’s c. Geogr. Anal. 2019, 51, 131–248. [Google Scholar] [CrossRef]
  12. Lin, J. Comparison of Moran’s I and Geary’s c in multivariate spatial pattern analysis. Geogr. Anal. 2023, 55, 685–702. [Google Scholar] [CrossRef]
  13. Yamada, H. Geary’s c and spectral graph theory. Mathematics 2021, 9, 2465. [Google Scholar] [CrossRef]
  14. Yamada, H. Geary’s c and spectral graph theory: A complement. Mathematics 2023, 11, 4228. [Google Scholar] [CrossRef]
  15. Yamada, H. Moran’s I for Multivariate Spatial Data. ResearchGate 2024. Available online: https://www.researchgate.net/publication/381256732_Moran’s_I_for_Multivariate_Spatial_Data (accessed on 5 June 2024).
  16. Moran, P.A.P. Notes on continuous stochastic phenomena. Biometrika 1950, 37, 17–23. [Google Scholar] [CrossRef]
  17. Maruyama, Y. An alternative to Moran’s I for spatial autocorrelation. arXiv 2015, arXiv:1501.06260. [Google Scholar] [CrossRef]
  18. Murakami, D.; Griffith, D.A. Random effects specifications in eigenvector spatial filtering: A simulation study. J. Geogr. Syst. 2015, 17, 311–331. [Google Scholar] [CrossRef]
  19. Seya, H.; Murakami, D.; Tsutsumi, M.; Yamagata, Y. Application of LASSO to the eigenvector selection problem in eigenvector-based spatial filtering. Geogr. Anal. 2015, 47, 284–299. [Google Scholar] [CrossRef]
  20. Murakami, D.; Griffith, D.A. Eigenvector spatial filtering for large data sets: Fixed and random effects approaches. Geogr. Anal. 2019, 51, 23–49. [Google Scholar] [CrossRef]
  21. Murakami, D.; Griffith, D.A. Spatially varying coefficient modeling for large datasets: Eliminating N from spatial regressions. Spat. Stat. 2019, 30, 39–64. [Google Scholar] [CrossRef]
  22. Murakami, D.; Yoshida, T.; Seya, H.; Griffith, D.A.; Yamagata, Y. A Moran coefficient-based mixed effects approach to investigate spatially varying relationships. Spat. Stat. 2017, 19, 68–89. [Google Scholar] [CrossRef]
  23. Nishi, H.; Asami, Y.; Baba, H.; Shimizu, C. Scalable spatiotemporal regression model based on Moran’s eigenvectors. Int. J. Geogr. Inf. Sci. 2023, 37, 162–188. [Google Scholar] [CrossRef]
  24. Yamada, H. A unified perspective on some autocorrelation measures in different fields: A note. Open Math. 2023, 21, 20220574. [Google Scholar] [CrossRef]
  25. Yamada, H. A new perspective about Moran’s coefficient: Revisited. Mathematics 2024, 12, 253. [Google Scholar] [CrossRef]
  26. Estrada, E.; Knight, P. A First Course in Network Theory; Oxford University Press: Oxford, UK, 2015. [Google Scholar]
  27. Gallier, J. Spectral theory of unsigned and signed graphs. Applications to graph clustering: A survey. arXiv 2016, arXiv:1601.04692. [Google Scholar] [CrossRef]
  28. Hammond, D.K.; Vandergheynst, P.; Gribonval, R. Wavelets on graphs via spectral graph theory. Appl. Comput. Harmon. Anal. 2011, 30, 129–150. [Google Scholar] [CrossRef]
  29. Shuman, D.I.; Narang, S.K.; Frossard, P.; Ortega, A.; Vandergheynst, P. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process. Mag. 2013, 30, 83–98. [Google Scholar] [CrossRef]
  30. De Jong, P.; Sprenger, C.; van Veen, F. On extreme values of Moran’s I and Geary’s c. Geogr. Anal. 1984, 16, 17–24. [Google Scholar] [CrossRef]
  31. Seber, G.A.F. A Matrix Handbook for Statisticians; Wiley: Hoboken, NJ, USA, 2008. [Google Scholar]
  32. Anderson, T.W. The Statistical Analysis of Time Series; John Wiley and Sons: New York, NY, USA, 1971. [Google Scholar]
  33. Nakatsukasa, Y.; Saito, N.; Woei, E. Mysteries around the graph Laplacian eigenvalue 4. Linear Algebra Its Appl. 2013, 438, 3231–3246. [Google Scholar] [CrossRef]
  34. Dong, X.; Thanou, D.; Frossard, P.; Vandergheynst, P. Learning Laplacian matrix in smooth graph signal representations. IEEE Trans. Signal Process. 2016, 64, 6160–6173. [Google Scholar] [CrossRef]
  35. Dong, X.; Thanou, D.; Rabbat, M.; Frossard, P. Learning graphs from data: A signal representation perspective. IEEE Signal Process. Mag. 2019, 36, 44–63. [Google Scholar] [CrossRef]
  36. Humbert, P.; Bars, B.L.; Oudre, L.; Kalogeratos, A.; Vayatis, N. Learning Laplacian matrix from graph signals with sparse spectral representation. J. Mach. Learn. Res. 2021, 22, 1–47. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yamada, H. Geary’s c for Multivariate Spatial Data. Mathematics 2024, 12, 1820. https://doi.org/10.3390/math12121820

AMA Style

Yamada H. Geary’s c for Multivariate Spatial Data. Mathematics. 2024; 12(12):1820. https://doi.org/10.3390/math12121820

Chicago/Turabian Style

Yamada, Hiroshi. 2024. "Geary’s c for Multivariate Spatial Data" Mathematics 12, no. 12: 1820. https://doi.org/10.3390/math12121820

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop