Next Article in Journal
Bayesian Methods for Step-Stress Accelerated Test under Gamma Distribution with a Useful Reparametrization and an Industrial Data Application
Previous Article in Journal
Weighted Graph-Based Two-Sample Test via Empirical Likelihood
Previous Article in Special Issue
Solving Mazes: A New Approach Based on Spectral Graph Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Moran’s I for Multivariate Spatial Data

Graduate School of Humanities and Social Sciences, Hiroshima University, 1-2-1 Kagamiyama, Hiroshima 739-8524, Japan
Mathematics 2024, 12(17), 2746; https://doi.org/10.3390/math12172746
Submission received: 8 July 2024 / Revised: 22 August 2024 / Accepted: 2 September 2024 / Published: 4 September 2024
(This article belongs to the Special Issue Graph Theory and Applications, 2nd Edition)

Abstract

:
Moran’s I is a spatial autocorrelation measure of univariate spatial data. Therefore, even if p spatial data exist, we can only obtain p values for Moran’s I. In other words, Moran’s I cannot measure the degree of spatial autocorrelation of multivariate spatial data as a single value. This paper addresses this issue. That is, we extend Moran’s I so that it can measure the degree of spatial autocorrelation of multivariate spatial data as a single value. In addition, since the local version of Moran’s I has the same problem, we extend it as well. Then, we establish their properties, which are fundamental for applied work. Numerical illustrations of the theoretical results obtained in the paper are also provided.

1. Introduction

Spatial autocorrelation is a notion that describes the similarities/discrepancies between data at different vertices/spatial units. It is fundamental to spatial science, which includes spatial statistics, spatial econometrics, geographical analysis, and so on. Many measures of it have been proposed. For a historical overview of spatial autocorrelation, see, e.g., Getis [1]. Among them, Moran’s I is the most prominent spatial autocorrelation measure and was developed by Moran [2] and Cliff and Ord [3,4,5,6]. Roughly speaking, like Pearson’s sample correlation coefficient, a positive (respectively, negative) Moran’s I indicates the presence of positive (respectively, negative) spatial autocorrelation. (However, unlike Pearson’s sample correlation coefficient, its range is not necessarily [ 1 , 1 ] . As will be shown later, it depends on the spatial weight matrix. See also de Jong et al. [7] and Maruyama [8].) Later, Anselin [9] developed a local version of Moran’s I: local Moran’s I. To distinguish them, Moran’s I is sometimes referred to as global Moran’s I.
Moran’s I is designed to measure the spatial autocorrelation of univariate spatial data. Therefore, even if p spatial data exist, we can only obtain p values of Moran’s I. In other words, Moran’s I cannot measure the degree of spatial autocorrelation of multivariate spatial data as a single value. In this paper: (i) We address this issue. That is, we extend Moran’s I so that it can measure the degree of spatial autocorrelation of multivariate spatial data as a single value. (ii) In addition, since the local version of Moran’s I has the same problem, we extend it as well. (iii) Subsequently, we establish their properties, which are fundamental for applied work. (iv) Numerical illustrations of the theoretical results are also provided.
Here, we discuss existing research related to this study. In addition to the papers listed above, the following papers are closely related to this paper: Wartenberg [10], Anselin [11], Lin [12], and Yamada [13,14]. First, Yamada [13] presented several results on univariate global Moran’s I. This paper depends on them. Second, Yamada [14] dealt with the multivariate extension of Geary’s c, which was developed by Geary [15] and modified by Cliff and Ord [3,4,5,6]. Thus, the present paper can be seen as a companion paper to it. It should be noted that the multivariate local Geary’s c was developed by Anselin [11]. Third, the relevance of Wartenberg [10] as well as Lin [12] to this paper is discussed in the second section from the end. This is mainly because we want to use our notations as following to describe their studies.
The paper is organized as follows. In Section 2, we sketch how Moran’s I is extended in this paper. In Section 3, we provide some preliminaries. More specifically, after stating the multivariate spatial data that will be considered in the paper, we review global and local Moran’s I for univariate spatial data. In Section 4, we define two new measures, i.e., multivariate global and local Moran’s I, and establish their properties. Section 5 provides numerical illustrations of the theoretical results obtained in Section 4. Section 6 clarifies the relationship between our multivariate global Moran’s I and Wartenberg’s [10] spatial correlation matrix. Section 7 concludes the paper.

2. A Sketch of How Moran’s I is Extended

In this section, we sketch how Moran’s I is extended in this paper.
Let y i denote a realization of a single variable y at the vertex/spatial unit v i for i = 1 , , n . Moran’s I uses the product given by
y i y ¯ s y j y ¯ s ,
for i , j = 1 , , n , where y ¯ = 1 n i = 1 n y i , and s is the positive square root of 1 n i = 1 n ( y i y ¯ ) 2 . Roughly speaking, like the sample correlation coefficient, a positive (respectively, negative) global Moran’s I indicates the presence of positive (respectively, negative) spatial autocorrelation.
Let y i = [ y ( 1 ) , i , , y ( p ) , i ] denote a realization of a multivariate vector y = [ y ( 1 ) , , y ( p ) ] at the vertex/spatial unit v i for i = 1 , , n . We ask how we can measure the similarity/discrepancy between y i and y j , which are both p-dimensional column vectors. A natural approach is to extend (1). That is, it can be measured with the inner product given by
( y i y ¯ ) S 1 ( y j y ¯ ) ,
for i , j = 1 , , n , where y ¯ = [ y ¯ ( 1 ) , , y ¯ ( p ) ] , and S 1 2 = diag ( s ( 1 ) 1 , , s ( p ) 1 ) . Here, y ¯ ( h ) = 1 n i = 1 n y ( h ) , i , and s ( h ) is the positive square root of 1 n i = 1 n ( y ( h ) , i y ¯ ( h ) ) 2 for h = 1 , , p . We develop a spatial autocorrelation measure that uses (2).

3. Preliminaries

In this section, after clarifying the multivariate spatial data that will be considered in the paper, we review global and local Moran’s I for univariate spatial data.
Before we do this, we introduce some notation. Let I n be the identity matrix of order n, and let e i be the i-th column of I n . Let ι be the n-dimensional vector of ones, and let Q ι = I n ι ( ι ι ) 1 ι . Note that Q ι is a symmetric and idempotent matrix, i.e., Q ι = Q ι and Q ι 2 = Q ι .

3.1. Multivariate Spatial Data

Following de Jong et al. [7], we treat the problem of spatial autocorrelation in terms of a graph. Let G = ( V , E ) denote a directed/undirected graph with n vertices. In addition, denote its vertex set by V = { v 1 , , v n } , where n 2 . For i , j = 1 , , n , let
w i , j > 0 if ( v i , v j ) E , w i , j = 0 otherwise ,
and W = [ w i , j ] R n × n . We assume that w i , i = 0 for i = 1 , , n , and accordingly, W is a hollow matrix by assumption. In addition, we assume that i = 1 n j = 1 n w i , j > 0 . Then W is a nonzero matrix. For example, when V = { v 1 , v 2 , v 3 , v 4 } and E = { ( v 1 , v 2 ) , ( v 1 , v 4 ) , ( v 2 , v 1 ) , ( v 2 , v 3 ) , ( v 3 , v 4 ) , ( v 4 , v 2 ) } , its binary weight matrix (adjacency matrix) is
W = 0 1 0 1 1 0 1 0 0 0 0 1 0 1 0 0 .
(Note that the corresponding graph is shown in Figure 1. The edge between v 1 and v 2 is undirected because both ( v 1 , v 2 ) and ( v 2 , v 1 ) belong to E.)
As illustrated above, W is not necessarily symmetric. However, for any x R n , given that x W x = ( x W x ) = x W x , it follows that x W x = x W x , where W = W + W 2 . Note that W is symmetric even though W is not symmetric. (If G = ( V , E ) is an undirected graph, then W is identical to W . We provide such a W in Section 5.2) Accordingly, given that ι W ι = i = 1 n j = 1 n w i , j , it follows that ι W ι = i = 1 n j = 1 n w i , j . Moreover, as w i , i = 0 for i = 1 , , n by assumption, tr W = tr W + tr W 2 = tr W = 0 .
Let
Y = y ( 1 ) , 1 y ( p ) , 1 y ( 1 ) , n y ( p ) , n = y 1 y n = [ y ( 1 ) , , y ( p ) ] R n × p .
Recall that y i = [ y ( 1 ) , i , , y ( p ) , i ] in (3) denotes a realization of a multivariate vector y = [ y ( 1 ) , , y ( p ) ] at the vertex/spatial unit v i for i = 1 , , n .
Let
Z = z ( 1 ) , 1 z ( p ) , 1 z ( 1 ) , n z ( p ) , n = z 1 z n = [ z ( 1 ) , , z ( p ) ] R n × p ,
where
z ( h ) , i = y ( h ) , i y ¯ ( h ) s ( h ) , i = 1 , , n , h = 1 , , p .
Then, by construction, z i is related to y i as
z i = z ( 1 ) , i z ( p ) , i = y ( 1 ) , i y ¯ ( 1 ) s ( 1 ) y ( p ) , i y ¯ ( p ) s ( p ) = S 1 2 ( y i y ¯ ) , i = 1 , , n ,
which appears in (2). Accordingly, it follows that
z i z j = ( y i y ¯ ) S 1 ( y j y ¯ ) .
In addition, z ( h ) is related to y ( h ) as
z ( h ) = z ( h ) , 1 z ( h ) , n = y ( h ) , 1 y ¯ ( h ) s ( h ) y ( h ) , n y ¯ ( h ) s ( h ) = 1 s ( h ) ( y ( h ) y ¯ ( h ) ι ) = 1 s ( h ) Q ι y ( h ) , h = 1 , , p .

3.2. Global Moran’s I for Univariate Spatial Data

Denote the global Moran’s I for a univariate spatial data, y ( h ) = [ y ( h ) , 1 , , y ( h ) , n ] , by I ( h ) :
I ( h ) = n q i = 1 n j = 1 n w i , j ( y ( h ) , i y ¯ ( h ) ) ( y ( h ) , j y ¯ ( h ) ) k = 1 n ( y ( h ) , k y ¯ ( h ) ) 2 , h = 1 , , p ,
where q = i = 1 n j = 1 n w i , j .
Then, as shown in, e.g., de Jong et al. [7], Dray [16], Maruyama [8], Murakami and Griffith [17], and Nishi et al. [18], I ( h ) can be expressed in matrix notation as
I ( h ) = n q y ( h ) Q ι W Q ι y ( h ) y ( h ) Q ι Q ι y ( h ) = n q y ( h ) Q ι W Q ι y ( h ) y ( h ) Q ι Q ι y ( h ) , h = 1 , , p .
I ( h ) in (9) can also be represented by using z ( h ) , 1 , , z ( h ) , n as
I ( h ) = 1 q i = 1 n j = 1 n w i , j y ( h ) , i y ¯ ( h ) s ( h ) y ( h ) , j y ¯ ( h ) s ( h ) = 1 q i = 1 n j = 1 n w i , j z ( h ) , i z ( h ) , j , h = 1 , , p .
Accordingly, given that i = 1 n j = 1 n w i , j z ( h ) , i z ( h ) , j = z ( h ) W z ( h ) = z ( h ) W z ( h ) , I ( h ) in (11) can be expressed in matrix notation as follows.
I ( h ) = 1 q z ( h ) W z ( h ) = 1 q z ( h ) W z ( h ) , h = 1 , , p .
Incidentally, given that s ( h ) 2 = 1 n y ( h ) Q ι Q ι y ( h ) and z ( h ) = Q ι y ( h ) s ( h ) , (12) can be obtained directly from (10).
Denote a spectral decomposition of a real symmetric matrix Q ι W Q ι by
Q ι W Q ι = U Λ U ,
where U = [ u 1 , , u n ] , and Λ = diag ( λ 1 , , λ n ) . Then, given that Q ι W Q ι ι = 0 , we can let ( λ n , u n ) = 0 , 1 n ι . With respect to the other eigenvalues, let λ 1 , , λ n 1 be in ascending order.
We document two known results with respect to I ( h ) .
Proposition 1.
(Equation (19) of de Jong et al. [7], Theorem 2.1 of Maruyama [8]). It follows that I ( h ) n q λ 1 , n q λ n 1 . I ( h ) = n q λ 1 when y ( h ) = μ ι + γ ( h ) u 1 such that γ ( h ) 0 . Likewise, I ( h ) = n q λ n 1 when y ( h ) = μ ι + γ ( h ) u n 1 such that γ ( h ) 0 .
Proof. 
Omitted. □
Remark 1.
(i) Although the range of Pearson’s sample correlation coefficient is [ 1 , 1 ] , that of Moran’s I is [ n q λ 1 , n q λ n 1 ] , which depends on W = W + W 2 . (ii) From Poincaré’s separation theorem Rao [19] (p. 64), Scott and Styan [20] (Theorem 1), Abadir and Magnus [21] (p. 347), Seber [22] (p. 113), it follows that
ζ 1 λ 1 ζ 2 , ζ n 1 λ n 1 ζ n ,
where ζ 1 , , ζ n are the eigenvalues of W in ascending order. Accordingly, it follows that
I ( h ) n q λ 1 , n q λ n 1 n q ζ 1 , n q ζ n , h = 1 , , p .
Incidentally, if G = ( V , E ) is an undirected bipartite graph whose weight matrix is binary, then ζ n = ζ 1 . See, e.g., Bapat [23] (Lemma 3.13) and Estrada and Knight [24] (p. 68).
Denote I ( h ) in (10) by I ( y ( h ) ) . Then, from Yamada [13] (Proposition 1(b)), it follows that
I ( u 1 ) I ( u n 1 ) .
In addition, let ψ ( h ) , i = a ( h ) , i 2 j = 1 n 1 a ( h ) , j 2 , where a ( h ) , i = u i y ( h ) for i = 1 , , n 1 . Then, ψ ( h ) , i 0 for i = 1 , , n 1 and i = 1 n 1 ψ ( h ) , i = 1 .
Proposition 2.
(Proposition 1(a) of Yamada [13]). I ( h ) in (9) is a weighted average of I ( u 1 ) , , I ( u n 1 ) as follows.
I ( h ) = i = 1 n 1 ψ ( h ) , i I ( u i ) .
Proof. 
Omitted. □

3.3. Local Moran’s I for Univariate Spatial Data

Denote the local Moran’s I of a univariate spatial data y ( h ) = [ y ( h ) , 1 , , y ( h ) , n ] by I ( h ) , i :
I ( h ) , i = n q ( y ( h ) , i y ¯ ( h ) ) j = 1 n w i , j ( y ( h ) , j y ¯ ( h ) ) k = 1 n ( y ( h ) , k y ¯ ( h ) ) 2 , i = 1 , , n .
Then, I ( h ) , i in (18) can also be represented using z ( h ) , 1 , , z ( h ) , n as
I ( h ) , i = 1 q z ( h ) , i j = 1 n w i , j z ( h ) , j , i = 1 , , n .
Note that I ( h ) , i is constructed so that
I ( h ) = i = 1 n I ( h ) , i .
We can confirm (20) as follows.
i = 1 n I ( h ) , i = i = 1 n 1 q z ( h ) , i j = 1 n w i , j z ( h ) , j = 1 q i = 1 n j = 1 n w i , j z ( h ) , i z ( h ) , j = I ( h ) .
Given that z ( h ) , i = z ( h ) e i and j = 1 n w i , j z ( h ) , j = w i , 1 z ( h ) , 1 + + w i , n z ( h ) , n = e i W z ( h ) , I ( h ) , i in (19) can be represented in matrix notation as
I ( h ) , i = 1 q z ( h ) , i j = 1 n w i , j z ( h ) , j = 1 q e i W z ( h ) z ( h ) e i = e i 1 q W z ( h ) z ( h ) e i , i = 1 , , n .
Incidentally, we can give another proof of I ( h ) = i = 1 n I ( h ) , i based on (21) as follows. From (21), since I ( h ) , i is the ( i , i ) -entry of 1 q W z ( h ) z ( h ) , it follows that
i = 1 n I ( h ) , i = tr 1 q W z ( h ) z ( h ) = 1 q tr z ( h ) W z ( h ) = 1 q z ( h ) W z ( h ) = I ( h ) .

4. Moran’s I’s for Multivariate Spatial Data

In this section, we newly introduce multivariate global and local Moran’s I and establish their properties.

4.1. Global Moran’s I for Multivariate Spatial Data

We define the following measure as the global Moran’s I for multivariate spatial data y 1 , , y n , which are p-dimensional column vectors:
I = 1 p q i = 1 n j = 1 n w i , j ( y i y ¯ ) S 1 ( y j y ¯ ) = 1 p q i = 1 n j = 1 n w i , j z i z j ,
which we refer to as “multivariate global Moran’s I” or simply “multivariate Moran’s I”. Note that the second equality in (23) follows from (7). When p = 1 , given that z i reduces to z ( 1 ) , i for i = 1 , , n , it follows that I = 1 q i = 1 n j = 1 n w i , j z ( 1 ) , i z ( 1 ) , j = I ( 1 ) . See (11). Thus, the multivariate Moran’s I given by (23) is a generalization of the conventional Moran’s I for univariate spatial data.
Let I ¯ = 1 p h = 1 p I ( h ) , which is the simple average of the univariate global Moran’s I’s, I ( 1 ) , , I ( p ) . Then, I has the following property:
Proposition 3.
I in (23) is equal to I ¯ .
Proof. 
Given that z i = [ z ( 1 ) , i , , z ( p ) , i ] for i = 1 , , n , z i z j = h = 1 p z ( h ) , i z ( h ) , j . Accordingly, it follows that
I = 1 p q i = 1 n j = 1 n w i , j z i z j = 1 p q i = 1 n j = 1 n w i , j h = 1 p z ( h ) , i z ( h ) , j = 1 p h = 1 p 1 q i = 1 n j = 1 n w i , j z ( h ) , i z ( h ) , j = 1 p h = 1 p I ( h ) = I ¯ .
For any A = [ a 1 , , a p ] R n × p , given that a k W a k = a k W a k , it follows that
tr A W A = vec ( A ) ( I p W ) vec ( A ) = k = 1 p a k W a k = k = 1 p a k W a k = vec ( A ) ( I p W ) vec ( A ) = tr A W A .
We use (24) to derive the following results.
Proposition 4.
I in (23) can be represented compactly in matrix form as
I = 1 p q tr Z W Z = 1 p q tr Z W Z .
Proof. 
Given that z j is the j-th row of Z , z j can be represented as e j Z . Accordingly, it follows that z i z j = z j z i = tr z j z i = tr z i z j = tr Z e i e j Z , from which we have
i = 1 n j = 1 n w i , j z i z j = i = 1 n j = 1 n w i , j tr Z e i e j Z = tr Z i = 1 n j = 1 n w i , j e i e j Z = tr Z W Z .
Thus, substituting this result into (23), we have I = 1 p q tr Z W Z . Finally, the second equality in (25) follows from (24). □
Remark 2.
Regarding Proposition 4, we make three remarks: (i) (25) is useful to calculate I. (ii) Given (24), it immediately follows that
I = 1 p q vec ( Z ) ( I p W ) vec ( Z ) = 1 p q vec ( Z ) ( I p W ) vec ( Z ) ,
which are representations corresponding to those in (12). Using (26), we can give another proof of Proposition 3 as follows.
I = 1 p q vec ( Z ) ( I p W ) vec ( Z ) = 1 p h = 1 p 1 q z ( h ) W z ( h ) = 1 p h = 1 p I ( h ) = I ¯ .
(iii) A MATLAB/GNU Octave user-defined function for calculating I in (25) is as follows:
  • function I = calc_I(Y,W)
  •      p = size(Y,2); Z = zscore(Y,1); q = sum(sum(W));
  •      I = trace(Z′ ∗ W ∗ Z)/(p ∗ q);
  • end
Note that as I ( 1 ) is a special case of I, the function can also be used for obtaining I ( 1 ) .
The following result shows the bounds of I.
Proposition 5.
It follows that I n q λ 1 , n q λ n 1 . I = n q λ 1 when y ( h ) = μ ι + γ ( h ) u 1 such that γ ( h ) 0 . Likewise, I = n q λ n 1 when y ( h ) = μ ι + γ ( h ) u n 1 such that γ ( h ) 0 .
Proof. 
The proposition immediately follows from Propositions 1 and 3. □
Remark 3.
Regarding Proposition 5, we make two remarks: (i) As in the case of global Moran’s I for univariate spatial data, the bounds of global Moran’s I for multivariate spatial data also depend only on the structure of the graph represented by W . (ii) From Proposition 5 and Equation (15), it follows that
I n q λ 1 , n q λ n 1 n q ζ 1 , n q ζ n .
Proposition 6.
I in (23) is a weighted average of I ( u 1 ) , , I ( u n 1 ) as follows.
I = i = 1 n 1 ψ ¯ i I ( u i ) ,
where ψ ¯ i = 1 p h = 1 p ψ ( h ) , i for i = 1 , , n 1 .
Proof. 
From Propositions 2 and 3, it follows that
I = 1 p h = 1 p I ( h ) = 1 p h = 1 p i = 1 n 1 ψ ( h ) , i I ( u i ) = i = 1 n 1 1 p h = 1 p ψ ( h ) , i I ( u i ) , = i = 1 n 1 ψ ¯ i I ( u i ) .
Here, ψ ¯ i 0 because ψ ( h ) , i 0 for h = 1 , , p . In addition, since i = 1 n 1 ψ ( h ) , i = 1 , it follows that
i = 1 n 1 ψ ¯ i = i = 1 n 1 1 p h = 1 p ψ ( h ) , i = 1 p h = 1 p i = 1 n 1 ψ ( h ) , i = 1 .
Remark 4.
Regarding Proposition 6, we make two remarks. (i) Proposition 6 is a generalization of Yamada [13] (Proposition 1(a)). (ii) Given (16), the distribution of ψ ¯ 1 , , ψ ¯ n 1 represents the spatial autocorrelation structure of the multivariate spatial data y 1 , , y n .

4.2. Local Moran’s I for Multivariate Spatial Data

Then, as we defined I from I ( h ) , we define the multivariate local Moran’s I from I ( h ) , i as follows.
I i = 1 p q z i j = 1 n w i , j z j , i = 1 , , n ,
which we refer to as the “multivariate local Moran’s I”. When p = 1 , given that z i = z ( 1 ) , i for i = 1 , , n , it follows that
I i = 1 p q z i j = 1 n w i , j z j = 1 q z ( 1 ) , i j = 1 n w i , j z ( 1 ) , j = I ( 1 ) , i , i = 1 , , n .
Thus, the multivariate local Moran’s I given by (30) is a generalization of the local Moran’s I for univariate spatial data. In addition, it follows that
i = 1 n I i = i = 1 n 1 p q z i j = 1 n w i , j z j = 1 p q i = 1 n j = 1 n w i , j z i z j = I .
Let I ¯ i = 1 p h = 1 p I ( h ) , i , which is the average of the univariate local Moran’s I’s, I ( 1 ) , i , , I ( p ) , i . Then, we have the following results.
Proposition 7.
I i in (31) is equal to I ¯ i for i = 1 , , n .
Proof. 
Given z i = [ z ( 1 ) , i , , z ( p ) , i ] , it follows that
I i = 1 p q z i j = 1 n w i , j z j = 1 p q j = 1 n w i , j z i z j = 1 p q j = 1 n w i , j h = 1 p z ( h ) , i z ( h ) , j = 1 p h = 1 p 1 q z ( h ) , i j = 1 n w i , j z ( h ) , j = 1 p h = 1 p I ( h ) , i = I ¯ i , i = 1 , , n .
The following result is useful to calculate I i .
Proposition 8.
I i in (31) can be represented compactly in matrix form as
I i = e i 1 p q W Z Z e i , i = 1 , , n .
Proof. 
From Proposition 7, Equation (21), and h = 1 p z ( h ) z ( h ) = Z Z , it follows that
I i = 1 p h = 1 p I ( h ) , i = 1 p h = 1 p 1 q e i W z ( h ) z ( h ) e i = 1 p q e i W h = 1 p z ( h ) z ( h ) e i = 1 p q e i W Z Z e i = e i 1 p q W Z Z e i , i = 1 , , n .
Remark 5.
Regarding Proposition 8, we make two remarks: (i) We can give another proof of i = 1 n I i = I based on (33) as follows. From (33), since I i is the ( i , i ) -entry of 1 p q W Z Z , it follows that
i = 1 n I i = tr 1 p q W Z Z = 1 p q tr Z W Z = I .
(ii) A MATLAB/GNU Octave user-defined function for calculating I i in (33) is as follows:
  • function Ii = calc_Ii(Y,W,i)
  •      p = size(Y,2); Z = zscore(Y,1); q = sum(sum(W));
  •      Ii = (W(i,:) * Z * Z(i,:)’)/(p * q);
  • end
Note that as I ( 1 ) , i is a special case of I i , the function can also be used for obtaining I ( 1 ) , i .

5. Numerical Illustrations

In this section, we provide numerical illustrations of the theoretical results obtained in the previous section. Before showing these, we introduce a table of Moran’s I’s.

5.1. Table of Moran’s I’s

Table 1 tabulates Moran’s I’s. In the next subsection, we make a table from two sets of generated multivariate spatial data. I 1 , , I n and I in the table are the measures that are newly defined in the paper. Recall that I i is the multivariate local Moran’s I at vertex v i and I is the multivariate global Moran’s I. They are located in the last column of the table. In this sense, the contribution of this paper can be expressed as the addition of a final column to this table. Again, recall that I ( h ) , i denotes the local Moran’s I for variable h at vertex v i and I ( h ) denotes the Moran’s I for variable h for h = 1 , , p and i = 1 , , n . As shown, among the measures, there are the following relations:
Proposition   7 : I i = 1 p ( I ( 1 ) , i + + I ( p ) , i ) , i = 1 , , n ,
Proposition   3 : I = 1 p ( I ( 1 ) + + I ( p ) ) ,
Equation   ( 20 ) : I ( h ) = I ( h ) , 1 + + I ( h ) , n , h = 1 , , p ,
Equation   ( 32 ) : I = I 1 + + I n .

5.2. Numerical Illustrations

For numerical illustrations of the theoretical results obtained in Section 4, consider the undirected graph shown in Figure 2. It is a two-dimensional square lattice graph, which is a Cartesian product of two path graphs. We suppose that it has a binary weight matrix, and accordingly, its W is given by
W = 0 1 0 1 0 0 0 0 0 1 0 1 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0 1 0 0 0 1 0 1 0 1 0 1 0 0 0 1 0 1 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 0 0 1 0 1 0 1 0 0 0 0 0 1 0 1 0 R 9 × 9 .
In this case, q = i = 1 n j = 1 n w i , j is equal to 24. Note that since W in (39) is symmetric, it is identical to W = W + W 2 . Incidentally, Moran’s [2] r 11 is a univariate global Moran’s I for a two-dimensional square lattice graph. For details, see Yamada [25].
Figure 2. A two-dimensional square lattice graph.
Figure 2. A two-dimensional square lattice graph.
Mathematics 12 02746 g002
We generate Y = [ y ( 1 ) , , y ( p ) ] R n × p by
y ( h ) = μ ι + γ ( h ) x + ε , h = 1 , , p ,
x = u n 1 , u 1 , ε i . i . d . N ( 0 , σ 2 I n ) ,
where n = 9 , p = 4 , μ = 0 , σ = 0.1 , and [ γ ( 1 ) , γ ( 2 ) , γ ( 3 ) , γ ( 4 ) ] = [ 0.5 , 1 , 2 , 4 ] . Recall that u n 1 and u 1 are eigenvectors of Q ι W Q ι . Note that when W is given by (39), both I ( h ) for h = 1 , , p and I belong to [ 1.0590 , 0.5303 ] . In addition, given that Q ι ι = 0 , μ in (40) does not affect Moran’s I’s even though it does affect y ( h ) .
Figure 3 (respectively, Figure 4) depicts the heatmaps of y ( h ) for h = 1 , , 4 , which are generated by (40) and (41) when x = u n 1 (respectively, x = u 1 ). In both figures, Panel A plots y ( 1 ) . Likewise, Panels B, C, and D respectively plot y ( 2 ) , y ( 3 ) , and y ( 4 ) . As expected, in both cases, the spatial autocorrelation structure of y ( 4 ) (respectively, y ( 1 ) ) is (respectively, not) the clearest.
Table 2 (respectively, Table 3) is the table of Moran’s I’s corresponding to the data depicted in Figure 3 (respectively, Figure 4). From these tables, we can confirm that the relations in (35)–(38) hold. For example, in Table 2, I 1 = 0.0373 is the average of I ( 1 ) , 1 = 0.0568 , I ( 2 ) , 1 = 0.0197 , I ( 3 ) , 1 = 0.0316 , and I ( 4 ) , 1 = 0.0413 , and I = 0.3948 is the average of I ( 1 ) = 0.1001 , I ( 2 ) = 0.4324 , I ( 3 ) = 0.5225 , and I ( 4 ) = 0.5241 . In addition, from the tables, it is observable that I ( h ) for h = 1 , , 4 and I certainly belong to [ 1.0590 , 0.5303 ] .
Finally, we make a remark. From Propositions 1 and 5, I ( h ) for h = 1 , , 4 and I of γ ( h ) u n 1 (respectively, γ ( h ) u 1 ) equal 0.5303 (respectively, 1.0590 ), which is the upper (respectively, lower) bound, regardless of γ ( h ) if γ ( h ) 0 . Nevertheless, the values corresponding to y ( 1 ) are far from these values, whereas those of y ( 4 ) are close to these values. This is due to the fact that y ( 1 ) is more highly contaminated by noise than y ( 4 ) is.

6. Discussion

In this section, we clarify the relationship between our multivariate global Moran’s I in (23) and Wartenberg’s [10] spatial correlation matrix M . Wartenberg [10] defined the following p × p matrix in our notation:
M = 1 q Z W Z
and the analysis based on its spectral decomposition is called Moran component analysis (MCA) (Lin [12]). When p = 1 , given that Z = z ( 1 ) , M reduces to I ( 1 ) in (12) as
M = 1 q Z W Z = 1 q z ( 1 ) W z ( 1 ) = I ( 1 ) .
From Proposition 4, our multivariate global Moran’s I in (23) is related to the eigenvalues of M as follows:
I = 1 p q tr Z W Z = 1 p tr M = 1 p k = 1 p θ k ,
where θ k for k = 1 , , p are the eigenvalues of M . Hence, I is identical to the average of the eigenvalues of M . In this sense, our multivariate global Moran’s I can be regarded as a value obtained from the spatial correlation matrix, M . Nevertheless, it should be emphasized that M is a matrix and not a single measure of spatial autocorrelation of multivariate spatial data.

7. Concluding Remarks

Conventional Moran’s I cannot measure the degree of spatial autocorrelation of multivariate spatial data as a single value. To address this issue, we developed Moran’s I for multivariate spatial data. It can describe the similarity/discrepancy between vectors of data at different vertices/spatial units. In addition, since the local version of Moran’s I has the same problem, we extended it as well. Subsequently, we established their properties, which are fundamental for applied work. They are summarized in Propositions 3–8. We have also illustrated them numerically.
Finally, we make a remark. In this paper, we developed Moran’s I’s for multivariate spatial data y 1 , , y n . Although we did not impose a stochastic model on y 1 , , y n , it is of course an interesting research topic to investigate the distribution of multivariate Moran’s I’s when such an assumption is made. Such investigations could be done using Imhof’s [26] method. However, this is beyond the scope of this paper.

Funding

The Japan Society for the Promotion of Science supported this work through KAKENHI (grant number: 23K01377).

Data Availability Statement

Data are contained within the article (Figure 3 and Figure 4).

Acknowledgments

The author thanks four anonymous referees for their valuable comments. The usual caveat applies.

Conflicts of Interest

The author declare no conflicts of interest.

References

  1. Getis, A. A history of the concept of spatial autocorrelation: A geographer’s perspective. Geogr. Anal. 2008, 40, 297–309. [Google Scholar] [CrossRef]
  2. Moran, P.A.P. Notes on continuous stochastic phenomena. Biometrika 1950, 37, 17–23. [Google Scholar] [CrossRef] [PubMed]
  3. Cliff, A.D.; Ord, J.K. The problem of spatial autocorrelation. In Studies in Regional Science; Scott, A.J., Ed.; Pion: London, UK, 1969; pp. 25–55. [Google Scholar]
  4. Cliff, A.D.; Ord, J.K. Spatial autocorrelation: A review of existing and new measures with applications. Econ. Geogr. 1970, 46, 269–292. [Google Scholar] [CrossRef]
  5. Cliff, A.D.; Ord, J.K. Spatial Autocorrelation; Pion: London, UK, 1973. [Google Scholar]
  6. Cliff, A.D.; Ord, J.K. Spatial Processes: Models and Applications; Pion: London, UK, 1981. [Google Scholar]
  7. De Jong, P.; Sprenger, C.; van Veen, F. On extreme values of Moran’s I and Geary’s c. Geogr. Anal. 1984, 16, 17–24. [Google Scholar] [CrossRef]
  8. Maruyama, Y. An alternative to Moran’s I for spatial autocorrelation. arXiv 2015, arXiv:1501.06260v1. [Google Scholar]
  9. Anselin, L. Local indicators of spatial association—LISA. Geogr. Anal. 1995, 27, 93–115. [Google Scholar] [CrossRef]
  10. Wartenberg, D. Multivariate spatial correlation: A method for exploratory geographical analysis. Geogr. Anal. 1985, 17, 263–283. [Google Scholar] [CrossRef]
  11. Anselin, L. A local Indicator of multivariate spatial association: Extending Geary’s c. Geogr. Anal. 2019, 51, 131–248. [Google Scholar] [CrossRef]
  12. Lin, J. Comparison of Moran’s I and Geary’s c in multivariate spatial pattern analysis. Geogr. Anal. 2023, 55, 685–702. [Google Scholar] [CrossRef]
  13. Yamada, H. A new perspective about Moran’s coefficient: Revisited. Mathematics 2024, 12, 253. [Google Scholar] [CrossRef]
  14. Yamada, H. Geary’s c for multivariate spatial data. Mathematics 2024, 12, 1820. [Google Scholar] [CrossRef]
  15. Geary, R.C. The contiguity ratio and statistical mapping. Inc. Stat. 1954, 5, 115–145. [Google Scholar] [CrossRef]
  16. Dray, S. A new perspective about Moran’s coefficient: Spatial autocorrelation as a linear regression problem. Geogr. Anal. 2011, 43, 127–141. [Google Scholar] [CrossRef]
  17. Murakami, D.; Griffith, D.A. Eigenvector spatial filtering for large data sets: Fixed and random effects approaches. Geogr. Anal. 2019, 51, 23–49. [Google Scholar] [CrossRef]
  18. Nishi, H.; Asami, Y.; Baba, H.; Shimizu, C. Scalable spatiotemporal regression model based on Moran’s eigenvectors. Int. J. Geogr. Inf. Sci. 2023, 37, 162–188. [Google Scholar] [CrossRef]
  19. Rao, C.R. Linear Statistical Inference and Its Applications, 2nd ed.; Wiley: New York, NY, USA, 1965. [Google Scholar]
  20. Scott, A.J.; Styan, G.P.H. On a separation theorem for generalized eigenvalues and a problem in the analysis of sample surveys. Linear Algebra Its Appl. 1985, 70, 209–224. [Google Scholar] [CrossRef]
  21. Abadir, M.K.; Magnus, J.R. Matrix Algebra; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  22. Seber, G.A.F. A Matrix Handbook for Statisticians; Wiley: Hoboken, NJ, USA, 2008. [Google Scholar]
  23. Bapat, R.B. Graphs and Matrices, 2nd ed.; Springer: London, UK, 2014. [Google Scholar]
  24. Estrada, E.; Knight, P. A First Course in Network Theory; Oxford University Press: Oxford, UK, 2015. [Google Scholar]
  25. Yamada, H. A unified perspective on some autocorrelation measures in different fields: A note. Open Math. 2023, 21, 20220574. [Google Scholar] [CrossRef]
  26. Imhof, J.P. Computing the distribution of quadratic forms in normal variables. Biometrika 1961, 48, 419–426. [Google Scholar] [CrossRef]
Figure 1. A graph consists of 4 vertices.
Figure 1. A graph consists of 4 vertices.
Mathematics 12 02746 g001
Figure 3. Heatmaps of y ( h ) generated by (40) and (41) for h = 1 , , 4 when x = u n 1 . Panel A plots y ( 1 ) . Likewise, Panels B, C, and D respectively plot y ( 2 ) , y ( 3 ) , and y ( 4 ) .
Figure 3. Heatmaps of y ( h ) generated by (40) and (41) for h = 1 , , 4 when x = u n 1 . Panel A plots y ( 1 ) . Likewise, Panels B, C, and D respectively plot y ( 2 ) , y ( 3 ) , and y ( 4 ) .
Mathematics 12 02746 g003
Figure 4. Heatmaps of y ( h ) generated by (40)–(41) for h = 1 , , 4 when x = u 1 . Panel A plots y ( 1 ) . Likewise, Panels B, C, and D respectively plot y ( 2 ) , y ( 3 ) , and y ( 4 ) .
Figure 4. Heatmaps of y ( h ) generated by (40)–(41) for h = 1 , , 4 when x = u 1 . Panel A plots y ( 1 ) . Likewise, Panels B, C, and D respectively plot y ( 2 ) , y ( 3 ) , and y ( 4 ) .
Mathematics 12 02746 g004
Table 1. Table of Moran’s I’s.
Table 1. Table of Moran’s I’s.
h = 1 h = 2 h = p
v 1 I ( 1 ) , 1 I ( 2 ) , 1 I ( p ) , 1 I 1
v 2 I ( 1 ) , 2 I ( 2 ) , 2 I ( p ) , 2 I 2
v n I ( 1 ) , n I ( 2 ) , n I ( p ) , n I n
I ( 1 ) I ( 2 ) I ( p ) I
Note: Among the values in this table, the relations in (35)–(38) hold.
Table 2. Table of Moran’s I’s: x = u n 1 in (40) and (41).
Table 2. Table of Moran’s I’s: x = u n 1 in (40) and (41).
h = 1 h = 2 h = 3 h = 4
v 1 0.0568 0.0197 0.0316 0.0413 0.0373
v 2 0.0347 0.0764 0.1277 0.1209 0.0899
v 3 0.0146 0.0668 0.0973 0.0896 0.0598
v 4 0.0013 0.0045 0.0103 0.0049 0.0053
v 5 0.0001 0.0071 0.0000 0.0001 0.0018
v 6 0.0146 0.0106 0.0080 0.0064 0.0026
v 7 0.0761 0.0979 0.1030 0.0962 0.0933
v 8 0.0286 0.1247 0.1152 0.1298 0.0996
v 9 0.0681 0.0247 0.0293 0.0348 0.0052
0.1001 0.4324 0.5225 0.5241 0.3948
Note: See Table 1 for what is tabulated in this table.
Table 3. Table of Moran’s I’s: x = u 1 in (40) and (41).
Table 3. Table of Moran’s I’s: x = u 1 in (40) and (41).
h = 1 h = 2 h = 3 h = 4
v 1 0.0439 0.0091 0.0509 0.0691 0.0432
v 2 0.1839 0.0428 0.1400 0.1436 0.1276
v 3 0.0986 0.0030 0.0729 0.0681 0.0606
v 4 0.0974 0.1543 0.0944 0.1492 0.1238
v 5 0.1439 0.2973 0.2681 0.2770 0.2466
v 6 0.0040 0.1223 0.1489 0.1251 0.1001
v 7 0.0621 0.0958 0.0604 0.0652 0.0709
v 8 0.0344 0.1334 0.1393 0.1083 0.1038
v 9 0.0287 0.0475 0.0703 0.0468 0.0340
0.6395 0.9054 1.0452 1.0523 0.9106
Note: See Table 1 for what is tabulated in this table.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yamada, H. Moran’s I for Multivariate Spatial Data. Mathematics 2024, 12, 2746. https://doi.org/10.3390/math12172746

AMA Style

Yamada H. Moran’s I for Multivariate Spatial Data. Mathematics. 2024; 12(17):2746. https://doi.org/10.3390/math12172746

Chicago/Turabian Style

Yamada, Hiroshi. 2024. "Moran’s I for Multivariate Spatial Data" Mathematics 12, no. 17: 2746. https://doi.org/10.3390/math12172746

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop