Next Article in Journal
Data-Centric Benchmarking of Neural Network Architectures for the Univariate Time Series Forecasting Task
Previous Article in Journal
Evaluating Wind Speed Forecasting Models: A Comparative Study of CNN, DAN2, Random Forest and XGBOOST in Diverse South African Weather Conditions
Previous Article in Special Issue
Impact of PV and EV Forecasting in the Operation of a Microgrid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Lattice and Point Spatial Data: Comparison of Unilateral and Multilateral SAR Models

by
Carlo Grillenzoni
Dipartimento Di Culture Del Progetto, Universitá IUAV di Venezia, St Croce, n. 1957, 30135 Venezia, Italy
Forecasting 2024, 6(3), 700-717; https://doi.org/10.3390/forecast6030036
Submission received: 24 June 2024 / Revised: 16 August 2024 / Accepted: 19 August 2024 / Published: 23 August 2024
(This article belongs to the Special Issue Feature Papers of Forecasting 2024)

Abstract

:
Spatial auto-regressive (SAR) models are widely used in geosciences for data analysis; their main feature is the presence of weight (W) matrices, which define the neighboring relationships between the spatial units. The statistical properties of parameter and forecast estimates strongly depend on the structure of such matrices. The least squares (LS) method is the most flexible and can estimate systems of large dimensions; however, it is biased in the presence of multilateral (sparse) matrices. Instead, the unilateral specification of SAR models provides triangular weight matrices that allow consistent LS estimates and sequential prediction functions. These two properties are strictly related and depend on the linear and recursive nature of the system. In this paper, we show the better performance in out-of-sample forecasting of unilateral SAR (estimated with LS), compared to multilateral SAR (estimated with maximum likelihood, ML). This conclusion is supported by numerical simulations and applications to real geological data, both on regular lattices and irregularly distributed points.

1. Introduction

Spatial data are nowadays present in many social and natural phenomena. They are mainly available in two formats: on regular lattices (as digital images and remotely-sensed data), and irregular polygons and sparse points (as in territorial zoning and environmental surveys); see [1] for a detailed description. Spatial data are representable within geographic information systems (GIS) and statistical models may be built for practical purposes, such as filtering and prediction. In digital images, the main goals are denoising and sharpening (e.g., [2]), to perform image segmentation and classification. In polygonal data, the aims are structural analyses and spatial forecasting (e.g., [3]) to support planning decisions.
One of the main features of spatial data is auto-correlation (ACR), which arises from the interaction between random variables located ad different spatial units (pixels in images, and centroids in polygons). The general rule is that the minor is the distance of units, and the greater is the size of the ACR. This dependence must be represented in regression models in order to satisfy the basic assumption of uncorrelated residuals. This, in turn, is necessary to have unbiased and efficient estimates of the parameters and their standard errors; that is, for performing statistical inference and prediction without bias.
On the other hand, ACR is useful for forecasting the character under study in areas where it is not measured, especially when exogenous (X) covariates are not available. It follows that representing the ACR in spatial systems is one of the major concerns of statistical modeling. As in time series analysis, it may be accomplished by introducing into the equations suitable “lagged” terms, i.e., the value of the dependent variable in nearby units. This leads to the spatial auto-regressive (SAR) models, which resemble the AR schemes of time series and dynamical systems. However, while in time there is only a single direction (from past to present), in the plane there are almost infinite directions.
The specification of the models is twofold: unilateral or multilateral, depending on whether the linkage between the nearby units reflects or not a sequential ordering. As an example, in lattice data, one has the triangular and rook schemes in Figure 1a,b [1]; in the first, the dynamic is unidirectional since starting from the upper-left corner it follows the sequence of writing a text [4]. Instead, in Figure 1b the dynamic is multi-directional as each cell involves parts of the text that are not yet written [2]. Figure 1b,c show the analogous situations in polygonal data, where the position of each unit is defined by its center (geometric or political). In Figure 1d the link is multi-directional with four nearest neighbors (NN), whereas in Figure 1c it is in the north direction only.
The statistical consequences of the two specifications of SAR models are important. Multidirection linkages violate the condition of independence between regressors and residuals, therefore, making least squares (LS) estimates biased and inconsistent. To solve this problem, many alternative estimators have been proposed, such as maximum likelihood (ML [1,5,6]), generalized method of moments (GM [7]), two-stage least squares [8] and, recently, indirect inference (II [9]). However, these methods are computationally demanding and use iterative algorithms of maximization which may not converge; involve the matrices of spatial contiguity, that must be inverted; require a condition of spatial stationarity (AR parameter | ϕ | < 1 ) that may not be fulfilled.
There are also theoretical works that analyze the LS method in multilateral SAR: [10] showed that when the distance among random variables goes to infinity as the spatial dimension increases or the weight matrix converges to zero, then LS is consistent. Ref. [11] investigate the sensitivity of various LS estimators of the AR coefficient using Taylor expansions and find that it is moderate for moderate ACR. Ref. [12] under the null hypothesis, develop refined tests for first-order ACR based on Edgeworth expansion of the LS distribution. Ref. [13] developed an II estimator that implicitly corrects the bias of LS with a mechanism that involves data simulations from a related model; they find that finite sample performance is similar to that of ML. These works show the potential use of the LS method even in SAR models with Toepliz-type contiguity matrices, which are used in social and economic studies.
In this paper, we focus on unidirectional SAR for lattice and polygonal data and we show their ability to preserve the optimal properties of LS estimates even compared to the efficient ML and GM methods. This feature is important because the LS method is linear and can manage datasets of large dimensions, as it may avoid the direct use of the contiguity matrices. Further, unilateral SAR models have a recursive structure, which enables the application of the chain rule of forecasting also outside the observation area. Instead, multilateral SAR models have prediction functions that involve “forward” values and require iterations.
The spatial prediction of missing data (inside the perimeter of the observed area) has been treated by [14,15,16,17] for multilateral SAR models, and [18] in the lattice case. They show that the naive predictor based on the reduced form of SAR systems must be corrected with the best linear unbiased predictor (BLUP) of classical statistics. However, the recursive forecasts of unilateral SAR models do not need this BLUP correction and outperform the others. We show these results with Monte Carlo experiments on synthetic data and out-of-sample forecasting on real data of geosciences, such as digital elevation models (e.g., [19]) and the spatial diffusion of water isotopes (e.g., [20]).
The paper is organized as follows: Section 2 deals with lattice models, it reviews the conditions of identification and consistency and evaluates their forecasts on random surfaces and digital images; Section 3 deals with SAR models for polygonal and sparse point data, it compares the forecasts of unilateral and multilateral models with various algorithms.

2. Regular Lattice Data

Regular lattice data are mostly present in remote sensing and digital images. These data are in the form of rectangular arrays of the type Y = { y i j } , with i = 1 , 2 n , j = 1 , 2 m the indexes of position, which may be transformed into latitude and longitude. The values y i j are usually autocorrelated (e.g., clustered) and one of the main goals is to filter the array Y with its own values, to obtain interpolates and forecasts: y ^ n + h , m + k . To accomplish this task, SAR modeling puts each cell in relation to the contiguous ones [1,21], such as y i j = g ( y i ± h , j ± k ) + e i j , where h , k = 1 , 2 p are spatial lags, p is the order of dependence and { e i j } is an unpredictable sequence.
In time series, the unidirectional (past-present) ordering of data is also called causal and found the analysis of causality between stochastic processes. In lattice data, the causal ordering can be established by following the lexicographic way of reading/writing a text, i.e., processing the cells of Y starting from the upper-left corner. Unilateral dynamics which satisfy this feature are the row-wise y i , j k , the half-plane y i h , j ± k , h > 0 and the one-quadrant (or triangular) with y i h , j k , ( h + k ) > 0 in Figure 1a [2,4]. Although causality is naturally related to the dynamic of events and aspects of human learning, digital filters of denoising and sharpening may follow non-causal models, such as the rook scheme in Figure 1b (e.g., [22]). However, the unidirectional approach remains the favorite solution, because it enjoys properties of sequentiality.
Under linear and unilateral constraints, the triangular SAR(p) model for y i j is defined as follows
y i j = h = 0 p k = 0 ( h + k ) > 0 p ϕ h k y i h , j k + e i j , e i j IN ( 0 , σ e 2 ) ,
with independent normal (IN) residuals. In statistics, the system (1) is usually estimated with the maximum likelihood (ML); this requires the vectorization of the data matrix y = vec j ( Y ) , of length N = n × m and the inversion of auto-covariance matrices Γ y of size N × N , see [3,23]. This solution is statistically efficient but computationally expensive and can be implemented only for moderate dimensions of the lattice field.
Instead, the LS method can be applied even for large values of n , m ; rewriting the model (1) in regression form as
y i j = ϕ x i j + e i j , x i j = y i , j 1 y i h , j k y i p , j p , ϕ = ϕ 01 ϕ h k ϕ p p ,
the LS estimator becomes
ϕ ^ N = i = p + 1 n j = p + 1 m x i j x i j 1 i = p + 1 n j = p + 1 m x i j y i j , = ϕ + i = p + 1 n j = p + 1 m x i j x i j 1 i = p + 1 n j = p + 1 m x i j e i j .
The second equation follows from (2) by inserting y = ϕ x + e and is useful for statistical analysis. The linearity of the algorithm (2) can be appreciated with respect to both y i j | x i j and the recursive calculation of the sums, see [24]. Unlike the ML approach, it only involves the inversion of a matrix of size ( p + 1 ) 2 1 , for any value of n , m . The LS estimator is unbiased and consistent for the unilateral model (1) because E ( x i j e i j ) = 0 .
SAR(1). As an example, let p = 1, then the model (1) is y i j = ( ϕ 1 y i , j 1 + ϕ 2 y i 1 , j + ϕ 3 y i 1 , j 1 ) + e i j . Applying the formula of [25] to the term y i , j 1 , one obtains the moving average (MA) representation
y i , j 1 = e i , j 1 + h = 1 ( h + l ) < i i k = 1 ( k + l ) < j j 1 l = 0 max ( i , j 1 ) ψ h k l e i h l , j 1 k l ,
from which E ( y i , j 1 e i j ) = 0. Similarly, E ( x i j e i j ) = 0 because all entries of x i j = [ y i , j 1 , y i 1 , j , y i 1 , j 1 ] do not depend on e i j ; hence, the LS estimator (2) is unbiased. This result does not apply to multilateral SAR models, because the MA decomposition would involve e i + h , j + k ; for the rook scheme in Figure 1b the ML estimator is necessary. Finally, as in a time series, the consistency of LS does not need the stability of the model (1), e.g., | ϕ h k / 3 | < 1 ; rather, it improves with the signal-to-noise ratio σ y 2 / σ e 2 (e.g., [26]).
SAR(p,q,r). Empirical models often have a subset structure, i.e., have missing lagged terms or they aggregate y i k , j h under a common parameter. The proper definition of such models is SAR( p , q , r ), where p = maximum lag of regressors, q = the number of spatial units and r = the number of parameters. Two parsimonious models that will be extensively applied in the paper are the triangular SAR(1,3,1) and the rook SAR(1,4,1) with drift α , defined as
y i j = α + ϕ y i 1 , j + y i , j 1 + y i 1 , j 1 / 3 + e i j , y i 0 = y 0 j = 0 ,
y i j = α + ϕ y i , j 1 + y i , j + 1 + y i 1 , j + y i + 1 , j / 4 + e i j , y i , m + 1 = y n + 1 , j = 0 ,
see Figure 1a,b and [1]. These models have almost the same number of regressors and can be written as y i j = α + ϕ y ¯ i j + e i j , where y ¯ is a local average. The condition of stability in (3) is | ϕ / 3 | < 1 , because it admits a MA decomposition with weights ψ h k l ( ϕ / 3 ) ( h + k + l ) which converge to 0, see [2]. By analogy, the model (4) is stable if | ϕ / 4 | < 1 , and this is a necessary condition for the convergence of its ML estimator (see Section 2.1).

2.1. The Vector Form

Previous representations are termed raster in GIS software. When the lattice dimension is moderate (e.g., n × m < 1000 ), it may be useful to pass at the vector form. Refs. [18,23,25] consider complex forms of vectorization for ML estimation. We follow the spatial econometric approach which vectorizes the data matrix by columns and builds the contiguity matrix of all cells, see [6]. Thus, let y = vec j ( Y ) of size N = n × m ; the first-order vector model with drift becomes
y = α 1 N + ϕ W y + e , e N ( 0 , σ e 2 I N ) ,
where 1 N is a unit vector of length N and W is an N × N contiguity matrix. Its structure depends on the order p and the marginal values: the models (3), (4) have matrices W with sub-diagonals < 1 and sparse 1 on the main diagonal, in correspondence of the border values y i 1 , y 1 j , y i m , y n j . For n = 4, m = 7 one has the arrays in Figure 2, see [24]; the distinctive feature of unilateral models is that W is always triangular, this provides recursivity to the system.
The arrays in Figure 2 are obtained by solving the static system y = W y , where m determines the number of blocks of size n × n , with matrices I n at the corners. However, when W is inserted in the dynamic model (5), its diagonal elements w l l = 1 must be replaced by 0, obtaining W 0 . Each sub-diagonal corresponds to a specific regressor y i h , j k of models (3), (4); thus, the array in Figure 2a can be decomposed as W 0 = ( W 1 + W 2 + W 3 ) , providing an SAR(3) model. In simulations, the data-generation of { y l } proceeds by rows, starting from an initial y = y 0 , as
y l = α + ϕ w l y + e l , l = 1 , 2 N ,
where w l is the l-th row of W 0 . Unilateral processes are insensitive to y 0 , since W 0 triangular fills y recursively; instead, the model (4), with the forward terms y i , j + 1 , y i + 1 , j , needs a non-null initial vector, e.g., y 0 = α 1 N .
Using matrix algebra, one may obtain the reduced (MA) form of the system (5)
I N ϕ W y = α 1 N + e , y = I N ϕ W 1 α 1 N + e ,
this provides an automatic way to generate SAR data, independent of the initial/border condition y 0 . The reduced form (6) is fundamental for ML estimation because, under Gaussianity, the log-likelihood function takes the form
( α , ϕ , σ ) N log ( σ ) + log det I N ϕ W e e / 2 σ e 2 , e ( α , ϕ ) = I N ϕ W y α 1 N ,
The maximization of Equation (7) proceeds iteratively and requires det ( · ) > 0 , i.e., the stability condition | ϕ | < 1 , to converge, see [6]. The ML method is compulsory for the rook model (4), but may also be applied to the triangular (3); although LS is preferable (see Table 1).
In the vector representation (5), the LS estimator of parameters δ = [ α , ϕ ] of the models (3), (4) may use the entire sample Z = 1 N , W y as
δ ^ N = Z Z 1 Z y , Σ ^ N = Z Z 1 σ ^ e 2 ,
improving the estimator (2) based on N p ( n + m ) observations. The matrix Σ ^ N provides the dispersion of estimates and if W is triangular the method is consistent; a formal analysis is in [24].
Spatial Forecasting. Prediction is one of the central goals of SAR modeling; typical examples with lattice data are restoring parts of remote-sensing images hidden by clouds or extrapolating their value outside the observed range. However, existing literature has mostly concerned with in-sample filtering and interpolation, focusing on techniques of image sharpening [22], robust denoising [2] and trend estimation [18], also in conjunction with non-parametric smoothing.
This paper aims to forecast data that are external to the measured perimeter n × m , e.g., on the right-hand side as { y i , m + k } with i , k > 0 or, in vector form y N + l on units placed beyond N. By defining h , k , l = 1 , 2 H , K , L the forward indexes, the forecast function depends on the SAR representations, see [15]. Here, we have the lattice form (3), (4), the vector AR (5) and the reduced MA (6); for the triangular model (3), the 3 predictors are
y ^ n + h , m + k = E y n + h , m + k | y n i + 1 , m j + 1 , h , k = 1 , 2 H , K ,
y ^ n + h , m + k = α + ϕ y ^ n + h , m + k 1 + y ^ n + h 1 , m + k 1 + y ^ n + h 1 , m + k / 3 ,
y ^ N + l = α + ϕ w N + l y ^ , y ^ = y 11 , y 21 y n m , y ^ N + 1 y ^ N + l 1 0 ,
y ^ = I N + L ϕ W N + L 1 α 1 N + L , l = 1 , 2 L ,
where w N + l is the ( N + l ) -th row of the augmented matrix W N + L and the vector y ^ in (10) is updated with the forecasts. The predictor (11) is nearly automatic, but in the absence of exogenous variables provides constant forecasts; on the other hand, the functions (9), (10) need border values to start. These can be set as y ^ 1 , m + k ( 0 ) = α ^ for all k, and may be improved by r-iterating the forecasts with the nearby values as y ^ 1 , m + k ( r + 1 ) = y ^ 2 , m + k 1 ( r ) , etc. This approach can extensively be applied to the rook model (4), with regard to its forward elements y i , j + 1 , y i + 1 , j ; however, the convergence of y ^ n + h , m + k ( r ) , as r increases, is not guaranteed and may give biased forecasts.
The variance of predictors is useful for building confidence intervals and testing hypotheses. A general expression can be obtained for the representations (6) and (11), such as
V ( y ^ ) = I N + L ϕ W N + L 1 I N + L ϕ W N + L 1 σ e 2 ,
Finally, in order to compare the models’ performance, a portion of the observed data, e.g., y i , m k , k = 1 K is excluded from the parameter estimation; then the forecasts of each model are computed y ^ i , m K + k and are compared with the mean absolute percentage error (MAPE) statistic
MAPE = 1 n K i = 1 n j = m K + 1 m | y ^ i j y i j y i j | .

2.2. Simulations and Applications

We develop simulation experiments to test the performance of LS (8) and ML (7) estimators, applied to unilateral and multilateral SAR models. We consider the models (3), (4) with parameters α 0 = 1 , ϕ 0 = 0.5 , 0.75 , 1.0 , 1.025 and border values ( α + e i j ) , i.e., y 00 = 0 , etc. Data are generated with the reduced form (6), with matrices as in Figure 2 with zero diagonal; 200 replications on a 32 × 32 lattice (N = 1024 cells) are obtained with Normal disturbances and σ e 2 = 1 . Performance statistics are the relative bias | ϕ ^ ϕ 0 | / ϕ 0 , the relative root mean squared errors (the square root of MSE( ϕ ^ )/ ϕ 0 2 ) and the p-values of the Normality test of [27]. Since the relative statistics of α ^ , ϕ ^ do not depend on the parameter size, their value is averaged to provide a single indicator of performance.
The results are displayed in Table 1 and Figure 3; the main findings are as follows: LS (8) is uniformly better than ML (7) when the matrix W is triangular. Its efficiency improves as ϕ 0 1 , as a consequence of the super-consistency property of LS estimates in SAR models (e.g., [26]). LS estimates in multilateral SAR models are biased for ϕ < 1 but still benefit from the super-consistency. Further, unlike time series, the Normal distribution of LS estimates holds even in the presence of the unit root ϕ 0 = 1 (see Figure 3b).
Regarding the ML method (7), we used the Matlab implementation of [6], which is computationally demanding. However, it significantly improves the estimates of multilateral models, providing good levels of unbiasedness and efficiency when ϕ 0 1 . On the other hand, it requires conditions of stationarity and its performance in triangular models is disappointing, meaning that ML is mainly designed for sparse-weight matrices. In conclusion, we can state that LS is suitable for unilateral models, whereas ML must be used in multilateral SAR.
We also carry out two applications to test the performance of SAR models in out-of-sample forecasting of lattice data. In comparing predictions, one cannot proceed as in Table 1, because data generated by a certain SAR model will only be better predicted by that model. Hence, we consider data from external sources.
Random Surface. The first application considers rough random surfaces [28]; these models are used in physics to study electromagnetic, fluid and plasma phenomena. We follow an approach based on a fast Fourier transform, where a matrix of random numbers u i j is convolved with a Gaussian filter to achieve a certain spatial autocorrelation. The resulting surfaces g ( u i j ) are random but smooth and one may add further noise to make them similar to the real data of geostatistics y i j = g ( u i j ) + e i j .
Figure 4a,b show an example of g ( u i j ) and y i j obtained with two Gaussian noises on a 32 × 32 lattice; the computational details are in [24]. The goal is to forecast the last K = 7 columns (about 22%) of the surface Y with the algorithms (9)–(11). Table 2 reports the in-sample parameter estimates (with LS and ML methods) of the models (3), (4) and their MAPE statistics (12).
A triangular SAR(1,3,3) (with three AR parameters) is also fitted to the data, but it does not improve the forecasts of the constrained (3). However, the best performance is provided by (3) with ML estimates. The rook (4) performs poorly with the (biased) LS estimates, whereas it significantly improves with the ML ones; anyway, the path of the predicted surface in Figure 4d remains disappointing. Ref. [24] provides further results; on 10 replications of the experiment, the mean value of MAPE of (3)-LS is − 37% smaller than that of (4)-ML.
Remote Sensing. The second application regards a real case study; we consider high-resolution elevation data obtained with aerial laser scanners. The sample array Y comes from USGS [19] and has 340 × 455 pixels (see Figure 5); this would yield a vector SAR model with 154,700 rows. Further, the terrain morphology may require the inclusion of spatial trend components (e.g., [18]), such as bivariate polynomials g d ( i , j ) of degree d 1 . Given the numerical issues implied by computing W , we only perform analyses with the LS estimator (2); for a model (3) with a quadratic trend, the forecasting function (9) becomes
y ^ i + h , j + k = α + g 2 ( i + h , j + k ) + ϕ y ^ i + h 1 , j + k 1 + y ^ i + h 1 , j + k + y ^ i + h , j + k 1 / 3 , g 2 ( i + h , j + k ) = β 1 ( i + h ) + β 2 ( j + k ) + β 3 ( i + h ) 2 + β 4 ( j + k ) 2 + β 5 ( i + h ) ( j + k ) ,
with h , k = 1 , 2 H , K , which can be sequentially managed. The AR part of the rook model involves forward values in the lower-right side; however, these may be provided by the polynomial itself: y ˜ i + h + 1 , j + k + 1 = g ^ 2 ( i + h + 1 , j + k + 1 ) .
In the prediction analysis, we left the right portion of 55 columns (about 12% of the array) as out-of-sample data to forecast. LS estimates of the parameters are reported in Table 3 and graphical results are displayed in Figure 5. Despite the better in-sample fitting (with R 2 = 0.977), the rook model has a disappointing performance both in terms of MAPE statistic and visual appearance (see Figure 5b). The reason is partly due to LS estimates of its parameters, with α ^ < 0 , ϕ ^ > 1 ; however, adjusting their values reduces the MAPE but does not change the (uniform) path of the predicted surface.
In real-life applications, where small and inner portions of images must be restored, the spatial polynomials g ( i , j ) may be fitted locally (around the missing pixels) or may be replaced by nonparametric smoothers (e.g., [18]). In any event, the role of the AR component remains fundamental, as the coefficient ϕ ^ is the most significant in both models of Table 3, with a statistic T ϕ ^ > 100 .

3. Point and Polygonal Data

When the data are irregularly distributed in space (either as point processes or polygonal areas), the representation has to change. The spatial process is now defined as { y s } , s = 1 , 2 , N ; where the index s also applies to the planar coordinates x s = [ i s , j s ] (latitude and longitude) of the polygon centers (geometric or political), see Figure 1c,d. These variables are irregularly distributed in the plane but have a fixed (non-stochastic) nature; moreover, they may influence the level of the process y s , according to spatial trends. Thus, the SAR(1) model becomes
y s = α + ϕ y s , 1 + β x s + e s , s = 1 , 2 N , y = α 1 N + ϕ W y + X β + e , e N 0 , σ e 2 I N , = I N ϕ W 1 α 1 N + X β + e ,
where y s , 1 belongs to the unit which is closest to the s-th term, and the vector of regressors x s = [ i s , j s ] may include other covariates.
Unlike the previous section, the adjacency matrix W has an irregular structure, which depends on the rule of contiguity. The most common rule is to put each observation y s in relation to its nearest neighbor (NN) term y r , according to the Euclidean distance: min r { d s r = [ ( i s i r ) 2 + ( j s j r ) 2 ] 1 / 2 } . Furthermore, under the unilateral constraint, the north/west (NW) → south/east (SE) direction may be followed as in the lattice case. However, polygonal and point data have not lexicographic order; hence, the unilateral constraint may simply be defined along the N-S direction. In this setting, it is useful to order the observations { y s , x s } according to the shortest distance from the northern border, i.e., according to the inverse of the latitude 1 / i s . Such ordering may be denoted as ( i s ) or simply ( s ) , and the model (13) can be written in a sequential form as
y ( s ) = α + ϕ y ( s , 1 ) + β x ( s ) + e ( s ) , y ( 0 , 1 ) = 0 ,
where y ( s , 1 ) is the north NN of y ( s ) , while the simple NN may be denoted as y ( s , 1 ) . As in the lattice case, the property E [ y ( s , 1 ) e ( s ) ] = 0 is the basis for unbiased LS estimation and forecasting.
As an illustration of NN matrices, we simulate N = 30 random points with [ i s , j s ] IU ( 0 , 1 ) 2 and consider q = 3 contiguous terms, see Figure 6. Under the north-south ordering of data, the entries of matrices are concentrated around the (null) main diagonal, and with the unilateral (north) constraint the array is lower triangular. Similarly, to obtain concentrated matrices, ref. [16] ordered the data according to the sum ( i s + j s ) , but they used the simple NN rule. The matrices W k in Figure 6 provide SAR(3) models but averaged as W ¯ = 1 3 k = 1 3 W k , they yield constrained SAR(3,1). The analogous model to the lattice model (3) is then
y ( s ) = α + ϕ q 1 k = 1 q y ( s , k ) + β x ( s ) + e ( s ) ,
where y ( s , k ) is the k-north NN of y ( s ) . Instead, in the multilateral model of Figure 1d one has q = 4 and the simple NN terms y ( s , k ) .
Unlike regular lattices, polygonal and point data are not equidistant; therefore, it is useful to consider spatially weighted averages in the model (15). A simple approach is based the inverse distance weighting (IDW)
y ( s ) = α + ϕ k = 1 q w ( s k ) y ( s , k ) + β x ( s ) + e ( s ) , w ( s k ) = d ( s , k ) 1 / k = 1 q d ( s , k ) 1 , d ( s , k ) = i ( s ) i ( s , k ) 2 + j ( s ) j ( s , k ) 2 1 / 2 ,
where [ i ( s , k ) , j ( s , k ) ] is the k-NN of the point [ i ( s ) , j ( s ) ] . Alternatively, since the inverse distances decay too fast, one can use the exponential weights w ( s k ) λ d ( s , k ) , with coefficient 0 < λ < 1. The resulting model is nonlinear and requires complex estimators; however, the value of λ may be selected a priori and its effect may be evaluated on the LS-ML estimates of ϕ .
Models (15) and (16) admit a vector representation (13), with W triangular or sparse as in Figure 6. One may also include a lagged term on the exogenous part
y = α 1 N + ϕ W y y + X β + W x X γ + e ,
this is the Durbin model [6], which is useful when x ( s ) are autocorrelated, i.e., the points ( i s , j s ) have a non-random pattern. The contiguity matrices in Equation (17) may be different, e.g., W y is preferably triangular, while W x may be sparse, without affecting the consistency of LS estimates.

3.1. Estimation and Forecasting

The LS estimator of the models (13)–(17) can be obtained from Equation (2) by writing the models in regression form, as y ( s ) = δ z ( s ) + e ( s ) , with δ = α , ϕ , β and z ( s ) = 1 , y ¯ ( s ) , x ( s ) , where y ¯ ( s ) is an average of q NN terms. In the matrix form (8), the LS estimator of δ is given by
δ ^ N = Z Z 1 Z y , Z = 1 N , W y , X ,
which can be applied to the Durbin model (17) by including in Z the term W X . In the case of triangular W , the estimator (18) is unbiased because E y ¯ ( s ) e ( s ) k E [ y ( s , k ) e ( s ) ] = 0 by the recursive computation of residuals. In general, under stationarity one can also prove the general result
N δ ^ N δ N D N 0 , E z s z s 1 σ e 2 ,
see [24], where the convergence is in distribution (D).
For non-triangular W , the LS is generally biased and one must use ML or generalized moments (GM [7]). The GM solution arises by applying the instrumental variables (IV) method to the LS estimator (18); by using the matrix of instruments M = X , W X , W 2 X W k X , the projection matrix P = M M M 1 M and the modified regressors V = P Z , the GM estimator is given by δ ˜ N = V Z 1 V y . In the numerical section, we compare ML and GM estimators with simulation experiments.
In forecasting, we have L out-of-sample units with value y L = 0 , whose coordinates and regressors x l = 1 , i l , j l , x 1 l x m l are known for all l = 1 , 2 L . Their locations ( i l , j l ) may be inside or outside the observed region; in both cases, they are placed at the end of the data matrix X N + L . If the data are ordered in a certain direction (e.g., north-south with i ( s ) i ( s + 1 ) ), and all L units are outside the observed region, and in the same direction, then the weight matrix is nearly block-diagonal: W N + L = diag W N , W L , where W L is overlapped to W N for the contiguity of L target units with the observed ones. In the other cases, it has a more complex structure, with triangular sub-matrices under the unilateral constraint
W N + L = W N N O N L W L N W L L
The forecasting function of y L depends on the SAR representation used for y ; in the reduced (MA) form of (13), the fitted values and the forecasts are jointly computed as
y ˜ L = E y L | X N + L , W N + L , y ˜ N + L = I N + L ϕ W N + L 1 α 1 N + L + X N + L β , = y ˜ N , y ˜ L ,
where the joint matrix W N + L has q-entries per row and may use non-uniform (IDW) weights. The solution (20) is nearly automatic, but in the absence of exogenous variables, it provides constant forecasts.
The second predictor comes from the structural (AR) representation (13); it is not automatic and must be managed sequentially by rows as
y ^ L = E y L | y N , X N + L , W N + L , y ^ N + l = α + ϕ w N + l y ^ N + l 1 + x N + l β , l = 1 , 2 L , y ^ N + l 1 = y 1 , y 2 , , y N , y ^ N + 1 y ^ N + l 1 , 0 , , 0 ,
where w l , x l are the l-th rows of the matrices W N + L , X N + L . Note, that for non-triangular W , Equation (21) involves missing values in the running vector y ^ N + l 1 . If these values are provided by Equation (20), then the vector of observations and forecasts at the l-th step becomes y ^ N + l 1 = y 1 y N , y ^ N + 1 y ^ N + l 1 , y ˜ N + l y ˜ N + L and in the end l = L , and will only contain the improved forecasts (21).
Refs. [14,17] also discussed solutions based on the best linear unbiased predictor (BLUP) of Goldberger. This approach arises from the conditional mean and variance of Gaussian random vectors:
E y L | y N = μ L + Σ L N Σ N N 1 y N μ N , y ˜ ^ L = y ˜ L + Σ L N Σ N N 1 y N y ˜ N ,
where y ˜ L is a subvector of the predictor in Equation (20) and Σ L N , Σ N N come from the partition of the joint covariance matrix of y N + L , given by
Σ N + L = I N + L ϕ W N + L 1 I N + L ϕ W N + L 1 σ e 2 = Σ N N Σ N L Σ L N Σ L L .
The conditional variance V y L | y N of the partition of Gaussian vectors also provides the dispersion of the forecasts (22), such as
V y ˜ ^ L = Σ L L Σ L N Σ N N 1 Σ N L ,
this matrix requires the error variance σ e 2 , which can be estimated with the in-sample residuals e ^ N = y N y ^ N of Equation (20).

3.2. Simulations and Applications

To compare the various estimators, we perform simulation experiments on SAR models (13)–(16) defined on a random grid of N = 150 points, uniformly distributed in the unit square: [ i s , j s ] IU ( 0 , 1 ) 2 , as in Figure 6c. We generate the process { y s } with an exogenous input x s IU ( 0 , 2 ) , Normal residuals e s IN ( 0 , 1 ) , parameters α = 1 , ϕ = 0.65 , 0.95 , β = 1 , contiguity matrices W with q = 1, 3 lags and with north NN and simple NN links. Subsequently, M = 500 replications are fitted with LS, ML, GM estimators and relative biases, relative RMSE and p-values of the Normality test were computed for the coefficients α ^ , ϕ ^ , β ^ and then averaged. Note, that relativization (e.g., bias ( ϕ ^ N ) / | ϕ 0 | ) allows to combine the statistics of three parameters to have a single indicator of performance.
The results are reported in Table 4, where “variable grid” means that at each replication the centroids change, and “IDW weights” refers to the model (16). The Durbin model includes a lagged term on the exogenous part, as β 1 x + β 2 W x , with coefficients [ β 1 , β 2 ] = [ 1 , 0.65 ] . The software used for ML and GM estimates is the Matlab package of [6], but it does not fit the Durbin model with the GM method. The main conclusions from Table 4 are that ML and GM methods do not improve the LS estimates in the case of unilateral (triangular) W matrices. However, they significantly outperform LS in the multidirectional (simple NN) case; in particular, the ML method is the best in terms of MSE. However, these results are not homogeneous regarding unbiasedness and efficiency, and the GM estimator may locally have a smaller bias.
Forecasting. As in the lattice case, when comparing the predictions of SAR models one must consider data generated from autonomous sources. An idea is to sample the surfaces in Figure 4 at random points i s IU [ 1 , n ] , j s IU [ 1 , m ] , and define y s as the terrain height; however, the forecasting results were too favorable to the unilateral SAR model. We then use the real USGS image in Figure 5, with n = 340, m = 455; we sampled N = 150 pixels and withheld for forecasting L = 30 on the right side, see Figure 7a. We fitted the constrained SAR (15) with q = 3, x s = [ i s , j s ] , W triangular and sparse, and forecast functions (20)–(22). The unilateral (west NN) is estimated with LS and the multilateral (simple NN) with ML.
The experiment is replicated M = 100 times (25 in each direction NS, SN, WE, EW, obtained by simply rotating the image Y ) and the average estimates are reported in Table 5. Both ML and LS estimates detect non-significant spatial coordinates ( β 1 , β 2 ), which means that the terrain level y s does not have a (monotone) spatial trend. As in the lattice case, the fitting statistics σ e , R 2 of the multilateral model are the best, but they are not confirmed out-of-sample. The predictor (21) moderately outperforms the naive (20), with a mild superiority (about 10% on average) of the unilateral model. Notice that the BLUP solution (22) mainly improves the multilateral model, whereas slightly worsens the unilateral one. Figure 7b shows the results of the predictor (21) in a single replication.
As a final application, we consider original point data concerning the measurement of stable isotopes of oxygen ( δ 18 O) and hydrogen ( δ 2 H) in the groundwater. Mapping isoscapes is useful for physical monitoring of the hydrological cycle and for anthropological and forensic investigations, e.g., regarding the path of people’s movement. We consider the datasets of [29], recorded in South Korea in 2010, with N = 130 and [20], recorded in Mexico in 2007, with N = 234. We estimate the SAR model (15) with q = 3 for y s = δ 18 O , δ 2 H and x s = latitude, longitude, to forecast about 20% of observations withheld from estimations. The unilateral constraints of the W matrix are south NN for Korea and west NN for Mexico; the results are in Table 6 and Figure 8. They confirm the better performance of unilateral SAR models with the forecast Function (21), especially when the data have significant autocorrelation (i.e., large | ϕ | ) and marked spatial trends (significant β 1 , 2 ). The reduction in the MAPE statistics in Table 6 ranges from −18% in Korea, to −38% in Mexico.

4. Conclusions

In this paper, we have compared unilateral and multilateral SAR models in forecasting spatial data of both lattice and point types. SAR systems are natural extensions of classical autoregressions, their difference is in the treatment of the spatial dependence: while unilateral models choose a single direction in space, as in time series, multilateral models consider multiple directions. These approaches lead to different contiguity matrices: triangular and sparse, usually computed with the nearest-neighbor approach. While triangular SAR can be consistently estimated with least squares, multilateral SARs require maximum likelihood or method of moments; in the latter cases, numerical complexity increases with the dimension of the contiguity matrices. Instead, LS is not sensitive to the size of the lattice or the number of spatial units and can manage even large-scale systems; see the estimator (2). Simulation experiments on small-medium scale systems have shown that ML and GM are suitable for multilateral models, but LS is preferable for unilateral ones, especially in the presence of unstable roots ( | ϕ | 1 ).
These features can be summarized in Table 7, which are vertically dependent; in fact, the concepts of triangularity, sequentiality, consistency of LS, recursive calculation, chain rule of forecasting, etc., are closely intertwined in the unilateral case.
For structural analyses, multilateral models are preferable because the interaction between spatial units does not occur in a single direction. However, structural analyses are mainly concerned with the dependence between the variables y s , x s ; now, in the Durbin model (17), the matrix W x may be sparse (for detecting simultaneous relationships), without affecting the properties of unilateral SAR. The role of W y in structural models is nearly ancillary: it has to remove the ACR of residuals e s , to have consistent estimates of the standard errors of β ^ . It is unlikely that a specific direction, i.e., a particular W y triangular, may hinder this goal.
Things are different in forecasting, unilateral models with ordered data enable recursive calculations, which allow linear predictions (the chain rule of forecasting). Instead, multilateral models involve forward values which may be pre-estimated only with the (inefficient) reduced form (20) and require iterations. In the applications we have seen that these models benefit from the BLUP improvement (22), but the performance of the unilateral models with the AR predictor (21) remains better, (see Table 5 and Table 6). Regarding the practical usage of SAR point forecasts, we mention the possibility of using them as alternatives to non-parametric smoothers (such as Kriging and Kernels), which are unstable outside the observation perimeter. Further, to overcome the limits of unidirectionality, one can estimate unilateral models in various directions (e.g., NS, SN, WE, EW), and then combine their forecasts with weights proportional to their fitting statistics (e.g., R 2 or 1 / σ ^ e 2 ). This solution is suitable for points within the investigated area, but can also be applied to external points, by using diagonal paths (e.g., NW-SE).

Funding

This research has not received fundings.

Data Availability Statement

The data are published in the references [19,20,29].

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Besag, J. Spatial interaction and the statistical analysis of lattice systems. J. R. Stat. Soc. Ser. B 1974, 36, 192–236. [Google Scholar] [CrossRef]
  2. Bustos, O.; Ojeda, S.; Vallejos, R. Spatial ARMA models and its applications to image filtering. Braz. J. Probab. Stat. 2009, 23, 141–165. [Google Scholar] [CrossRef]
  3. Barbiero, T.; Grillenzoni, C. Planning a novel regional methane network: Demand forecasting and economic evaluation. Energy Convers. Manag. X 2022, 16, 100294. [Google Scholar] [CrossRef]
  4. Tjøstheim, D. Statistical Spatial Series Modelling II: Some further Results in Unilateral Processes. Adv. Appl. Probab. 1983, 15, 562–684. [Google Scholar] [CrossRef]
  5. Anselin, L. Spatial Econometrics. In A Companion to Theoretical Econometrics; Baltagi, B.H., Ed.; Blackwell Publishing: Oxford, UK, 2003; pp. 310–330. [Google Scholar]
  6. LeSage, J.P.; Pace, R.K. Introduction to Spatial Econometrics; CRC Press, Taylor & Francis Group: Boca Raton, FL, USA, 2009. [Google Scholar]
  7. Kelejian, H.H.; Prucha, I.R. A generalized moments estimator for the autoregressive parameter in a spatial model. Internat. Econom. Rev. 1999, 40, 509–533. [Google Scholar] [CrossRef]
  8. Lee, L.F. Best spatial two-stage least squares estimators for a spatial autoregressive model with autoregressive disturbances. Econom. Rev. 2003, 22, 307–335. [Google Scholar] [CrossRef]
  9. Bao, Y.; Liu, X.; Yang, L. Indirect inference estimation for spatial autoregressions. Econometrics 2020, 8, 34. [Google Scholar] [CrossRef]
  10. Lee, L.F. Consistency and efficiency of least squares estimation for mixed regressive, spatial autoregressive models. Econom. Theory 2002, 18, 252–277. [Google Scholar] [CrossRef]
  11. Liu, S.; Polasek, W.; Sellner, R. Sensitivity analysis of SAR estimators: A numerical approximation. J. Stat. Comput. Simul. 2012, 82, 325–342. [Google Scholar] [CrossRef]
  12. Robinson, P.M.; Rossi, F. Refined tests for spatial correlation. Econom. Theory 2015, 31, 1249–1280. [Google Scholar] [CrossRef]
  13. Kyriacou, M.; Phillips, P.C.B.; Rossi, F. Indirect inference in spatial autoregression. Econom. J. 2017, 20, 168–189. [Google Scholar] [CrossRef]
  14. Goulard, M.; Laurent, T.; Thomas-Agnan, C. About predictions in spatial autoregressive models: Optimal and almost optimal strategies. Spat. Econ. Anal. 2017, 12, 304–325. [Google Scholar] [CrossRef]
  15. Kelejian, H.H.; Prucha, I.R. The efficiencies of various predictors in spatial econometric models containing lags. Reg. Sci. Urban Econ. 2007, 37, 363–374. [Google Scholar] [CrossRef]
  16. LeSage, J.P.; Pace, R.K. Models for spatially dependent missing data. J. Real Estate Financ. Econ. 2004, 29, 233–254. [Google Scholar] [CrossRef]
  17. LeSage, J.P.; Pace, R.K. Spatial econometric models, prediction. In Encyclopedia of Geographic Information Science; Shekhar, S., Xiong, H., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 1095–1100. [Google Scholar]
  18. Mojiri, A.; Waghei, Y.; Nili-Sani, H.R.; Mohtashami Borzadaran, H.R. Non-stationary spatial autoregressive modeling for the prediction of lattice data. Commun. Stat.-Simul. Comput. 2023, 52, 5714–5726. [Google Scholar] [CrossRef]
  19. USGS (2017), U.S. Geological Survey, Digital Elevation Models (DEMs). Available online: https://www.sciencebase.gov/catalog/item/543e6b86e4b0fd76af69cf4c (accessed on 1 January 2023).
  20. Reynaga, D.K.; Millaire, J.F.; Balderas, X.C.; Berrelleza, J.A.; Luján, L.L.; Longstaffe, F.J. Building Mexican isoscapes: Oxygen and hydrogen isotope data of meteoric water. Data Brief 2021, 36, 107084. [Google Scholar] [CrossRef] [PubMed]
  21. Fingleton, B. Spatial Autoregression. Geograph. Anal. 2009, 41, 385–391. [Google Scholar] [CrossRef]
  22. Grillenzoni, C. Statistics for image sharpening. Stat. Neerl. 2008, 62, 173–192. [Google Scholar] [CrossRef]
  23. Awang, N.; Shitan, M. Estimating the parameters of the second order spatial unilateral autoregressive model. Int. J. Stat. Sci. 2006, 5, 37–58. [Google Scholar]
  24. Grillenzoni, C. Forecasting lattice and point spatial data: Comparison of unilateral and multilateral SAR models. Preprints 2024, 1, 2024080947. [Google Scholar] [CrossRef]
  25. Basu, S.; Reinsel, G.C. Properties of the spatial nilateral first-order ARMA model. Adv. Appl. Probab. 1993, 25, 631–648. [Google Scholar] [CrossRef]
  26. Bhattacharyya, B.B.; Richardson, G.D.; Franklin, L.A. Asymptotic inference for near unit roots in spatial autoregression. Ann. Stat. 1997, 28, 173–179. [Google Scholar] [CrossRef]
  27. Jarque, C.M.; Bera, A.K. A test for normality of observations and regression residuals. Int. Stat. Rev. 1987, 55, 163–172. [Google Scholar] [CrossRef]
  28. Sheffield, S. What is a random surface? In Proceedings of the International Congress of Mathematicians 2022, Virtual, 6–14 July 2022; International Mathematical Union, EMS Press: Philadelphia, PA, USA, 2022. [Google Scholar]
  29. Gautam, M.K.; Song, B.-Y.; Shin, W.-J.; Bong, Y.-S.; Lee, K.-S. Datasets for spatial variation of O and H isotopes in waters and hair across South Korea. Data Brief 2020, 30, 105666. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Examples of lattice (a,b) and polygonal (c,d) data, with unilateral (a,c) and multilateral (b,d) spatial dependence schemes, with respect to their nearest neighbors.
Figure 1. Examples of lattice (a,b) and polygonal (c,d) data, with unilateral (a,c) and multilateral (b,d) spatial dependence schemes, with respect to their nearest neighbors.
Forecasting 06 00036 g001
Figure 2. Contiguity matrices W of the models (3), (4) on a lattice of size n = 4, m = 7. Red dots have value 1 and correspond to the border values; while the blue have value 1/3 in Panel (a) and 1/4 in Panel (b); see [24].
Figure 2. Contiguity matrices W of the models (3), (4) on a lattice of size n = 4, m = 7. Red dots have value 1 and correspond to the border values; while the blue have value 1/3 in Panel (a) and 1/4 in Panel (b); see [24].
Forecasting 06 00036 g002
Figure 3. Distribution of LS and ML estimates in 200 replications of the SAR models (3), (4) with α 0 = 1, ϕ 0 = 1 (green), and mean values of α ^ , ϕ ^ (red): Panels (a,b) LS (8) of model (3); Panels (c,d) ML (7) of model (4).
Figure 3. Distribution of LS and ML estimates in 200 replications of the SAR models (3), (4) with α 0 = 1, ϕ 0 = 1 (green), and mean values of α ^ , ϕ ^ (red): Panels (a,b) LS (8) of model (3); Panels (c,d) ML (7) of model (4).
Forecasting 06 00036 g003
Figure 4. Panels (a,b): Simulation of a rough random surface Y of size n = m = 32. Panels (c,d): Out-of-sample forecasts of Y (:,26:32) with models (3), (4) and predictors (9), (10).
Figure 4. Panels (a,b): Simulation of a rough random surface Y of size n = m = 32. Panels (c,d): Out-of-sample forecasts of Y (:,26:32) with models (3), (4) and predictors (9), (10).
Forecasting 06 00036 g004
Figure 5. Application of SAR models (3), (4) with quadratic trend to [19] data: Out-of-sample forecasts of the image portion Y (:,401:455), displayed in pseudocolor.
Figure 5. Application of SAR models (3), (4) with quadratic trend to [19] data: Out-of-sample forecasts of the image portion Y (:,401:455), displayed in pseudocolor.
Forecasting 06 00036 g005
Figure 6. Contiguity matrices and planar graphs of N = 30 random centroids on the unit square [0, 1], based on nearest neighbors (NN). Panels: (a,c) North NN; (b,d) Simple NN. Matrices: W 1 first NN (blue); W 2 second NN (red); W 3 third NN (green).
Figure 6. Contiguity matrices and planar graphs of N = 30 random centroids on the unit square [0, 1], based on nearest neighbors (NN). Panels: (a,c) North NN; (b,d) Simple NN. Matrices: W 1 first NN (blue); W 2 second NN (red); W 3 third NN (green).
Forecasting 06 00036 g006
Figure 7. Graphical results of a single replication of the estimates in Table 5: (a) Random sampling of USGS surface (the points to forecast are in red); (b) Longitudinal display of data and forecasts (21) (unilateral = green, multilateral = black); (c) Contiguity matrix of the unilateral model with q = 3; (d) W matrix of the multilateral model (simple NN).
Figure 7. Graphical results of a single replication of the estimates in Table 5: (a) Random sampling of USGS surface (the points to forecast are in red); (b) Longitudinal display of data and forecasts (21) (unilateral = green, multilateral = black); (c) Contiguity matrix of the unilateral model with q = 3; (d) W matrix of the multilateral model (simple NN).
Forecasting 06 00036 g007
Figure 8. Water isotope data and SAR forecasts (22): Blue = in-sample data, Red = out-of-sample data, Green = forecasts with W triangular; Black = forecasts with W multilateral. Panels: (a) South-Korea sample locations; (b) Latitudinal view of Hydrogen isotope; (c) Mexico sample locations; (d) Longitudinal view of Oxygen isotope.
Figure 8. Water isotope data and SAR forecasts (22): Blue = in-sample data, Red = out-of-sample data, Green = forecasts with W triangular; Black = forecasts with W multilateral. Panels: (a) South-Korea sample locations; (b) Latitudinal view of Hydrogen isotope; (c) Mexico sample locations; (d) Longitudinal view of Oxygen isotope.
Forecasting 06 00036 g008
Table 1. Performance of LS (8) and ML (7) estimators applied to triangular and rook SAR models (3), (4) with n = m = 32: R-bias = relative bias, R-RMSE = relative root mean squared error, N-test = p-values of the normality test. The bold character indicates the best methods (LS or ML) over model classes (Triang and Rook).
Table 1. Performance of LS (8) and ML (7) estimators applied to triangular and rook SAR models (3), (4) with n = m = 32: R-bias = relative bias, R-RMSE = relative root mean squared error, N-test = p-values of the normality test. The bold character indicates the best methods (LS or ML) over model classes (Triang and Rook).
MethodLSTriang.LSRookMLTriang.MLRook α 0 = 1
Stat. α ^ ϕ ^ α ^ ϕ ^ α ^ ϕ ^ α ^ ϕ ^ ϕ 0
R-bias0.0150.0130.2790.3250.2950.3240.0140.014
R-RMSE0.0780.0780.2890.3340.3020.3290.0620.0660.5
N-test>0.50.4150.4560.2670.3190.0650.2210.138
R-bias0.0210.0070.3300.1340.3720.1480.0230.008
R-RMSE0.0880.0320.3390.1360.3840.1530.0740.0280.75
N-test>0.5>0.50.4130.135>0.5>0.50.1920.052
R-bias0.0070.00030.0040.00010.0490.0030.0230.0002
R-RMSE0.0530.0030.0540.00040.0670.0040.0430.00021
N-test>0.5>0.50.4120.1190.1260.0010.0920.001
R-bias0.00570.00010.0030.00000.5100.0242.2700.023
R-RMSE0.0470.0020.0350.00010.5120.0242.2860.0241.025
N-test>0.5>0.5>0.50.001>0.50.0010.4500.001
Table 2. Parameter estimates (with T-statistics in parentheses) of SAR models (3), (4) on the surface Y (:,1:25) in Figure 4b, and forecast statistics on the data Y (:,26:32).
Table 2. Parameter estimates (with T-statistics in parentheses) of SAR models (3), (4) on the surface Y (:,1:25) in Figure 4b, and forecast statistics on the data Y (:,26:32).
ModelEstim. α ^ ϕ ^ 1 ϕ ^ 2 ϕ ^ 3 R 2 MAPE
Triang. (3)LS0.096 (4.45)0.893 (40.3)..0.6860.143
Triang. (3)ML0.247 (15.2)0.739 (45.1)..0.665 0.125
SAR(1,3,3)LS0.094 (4.42)0.448 (13.3)0.097 (2.69)0.352 (10.1)0.7010.140
Rook (4)LS−0.008 (−0.39)1.008 (50.6)..0.7880.864
Rook (4)ML0.198 (16.3)0.795 (64.9)..0.7530.224
Table 3. LS estimates of SAR models (3), (4) with bivariate polynomials (T-statistics are in parentheses) applied to USGS data [19]: Y (:,1:400) in-sample, Y (:,401:455) forecasts.
Table 3. LS estimates of SAR models (3), (4) with bivariate polynomials (T-statistics are in parentheses) applied to USGS data [19]: Y (:,1:400) in-sample, Y (:,401:455) forecasts.
Model α ^ ϕ ^ β ^ 1 β ^ 2 β ^ 3 β ^ 4 β ^ 5 σ ^ e 2 , R 2 MAPE
Triang. (3)8.500.961−0.024−0.01111 × 10−543 × 10−7−88 × 10−7338.4 0.848
(25.6)(105)(−10.6)(−5.5)(22.5)(0.73)(−2.1)0.913.
Rook (4)−6.121.0260.0070.016−58 × 10−617 × 10−6−17 × 10−688.151.362
(−35.8)(214)(6.3)(15.9)(−22.8)(5.5)(−7.8)0.977.
Table 4. Results of the simulation experiments on the SAR models (13)–(16) with spatial lags q = 1–3 and single ϕ . The statistics are average values (over α ^ , ϕ ^ , β ^ ) of relative biases, relative RMSE and p-value of N-test. Bold characters indicate the best results.
Table 4. Results of the simulation experiments on the SAR models (13)–(16) with spatial lags q = 1–3 and single ϕ . The statistics are average values (over α ^ , ϕ ^ , β ^ ) of relative biases, relative RMSE and p-value of N-test. Bold characters indicate the best results.
WNorth NNSimple NN
Estimator LS ML GM LS ML GM ϕ Lags
Ave. Rbias0.02050.13380.02350.13630.02870.02160.651
Ave. RRMSE0.13360.24020.18310.18870.13430.1581
Ave. Ntest0.32180.29830.16750.15610.34010.3337
Ave. Rbias0.00770.12080.00980.12320.04000.00750.651
Ave. RRMSE0.13960.24450.19240.18730.14300.1609(Variable
Ave. Ntest0.32740.35540.22910.16370.24990.1626grid)
Ave. Rbias0.03400.13300.02170.13240.02590.03760.653
Ave. RRMSE0.15860.24160.23650.19980.13900.2035
Ave. Ntest0.24420.26120.09360.18110.25590.0960
Ave. Rbias0.02980.12890.02280.14870.02080.03430.653
Ave. RRMSE0.15180.23600.22280.20880.13260.1963(IDW
Ave. Ntest0.25300.25610.16730.18050.29000.1736weights)
Ave. Rbias0.03120.09800.02480.05810.01090.01640.953
Ave. RRMSE0.14750.25140.31900.12080.10860.1648
Ave. Ntest0.15160.16780.00100.26110.30330.1601
Ave. Rbias0.06630.3405.0.33330.0704.0.653
Ave. RRMSE0.28490.4659.0.41410.2764.(Durbin
Ave. Ntest0.25840.3891.0.12860.1313.model)
Table 5. Average values, over 100 replications, of parameter estimates (T-statistics are in parentheses), and MAPE forecast statistics of the SAR model (15) on random samples of pixels of the USGS image. Bold character indicates the best results.
Table 5. Average values, over 100 replications, of parameter estimates (T-statistics are in parentheses), and MAPE forecast statistics of the SAR model (15) on random samples of pixels of the USGS image. Bold character indicates the best results.
ModelEstim. α ¯ ϕ ¯ β ¯ 1 β ¯ 2 σ ¯ e , R ¯ 2 MAPE ( 20 ) MAPE ( 21 ) MAPE ( 22 )
UnilateralLS (18)68.40.5430.012−0.01854.30.9480.8150.830
(west NN) (3.0)(5.1)(0.21)(−0.31)0.213...
MultilateralML (7)76.50.4920.005−0.00950.20.9960.9510.899
(simple NN) (4.2)(6.3)(0.06)(−0.19)0.328...
Table 6. Parameter estimates (with T-statistics in parentheses) and MAPE forecast statistics of the SAR model (15) with q = 3 applied to the water isotope data of South Korea δ 2 H (first block) and Mexico δ 18 O (second block). Bold character indicates the best results.
Table 6. Parameter estimates (with T-statistics in parentheses) and MAPE forecast statistics of the SAR model (15) with q = 3 applied to the water isotope data of South Korea δ 2 H (first block) and Mexico δ 18 O (second block). Bold character indicates the best results.
Model α ^ ϕ ^ β ^ 1 β ^ 2 σ ^ e , R 2 MAPE ( 20 ) MAPE ( 21 ) MAPE ( 22 )
UnilateralLS (18)123.70.337−0.557−2.514.280.0690.0650.065
(south NN) (1.70)(4.72)(−1.04)(−3.37)0.478...
MultilateralML (7)97.90.622−0.376−1.983.530.0800.0790.079
(simple NN) (1.69)(9.46)(−0.86)(−3.41)0.645...
UnilateralLS (18)4.300.6360.0890.0931.280.3770.3690.370
(west NN) (1.69)(8.17)(3.19)(2.72)0.336...
MultilateralML (7)−0.850.4790.0530.1011.260.6000.5970.588
(simple NN) (−0.36)(7.41)(1.88)(2.89)0.351...
Table 7. Summary of the main features of SAR models.
Table 7. Summary of the main features of SAR models.
ItemUnilateralMultilateral
Matrix W y TriangularSparse
SystemSequentialSimultaneous
EstimatorLSML, GM
Sample sizeLargeModerate
AR root | ϕ | 1 | ϕ | 1
ComputationRecursiveIterative
ApplicationForecastingAnalysis
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Grillenzoni, C. Forecasting Lattice and Point Spatial Data: Comparison of Unilateral and Multilateral SAR Models. Forecasting 2024, 6, 700-717. https://doi.org/10.3390/forecast6030036

AMA Style

Grillenzoni C. Forecasting Lattice and Point Spatial Data: Comparison of Unilateral and Multilateral SAR Models. Forecasting. 2024; 6(3):700-717. https://doi.org/10.3390/forecast6030036

Chicago/Turabian Style

Grillenzoni, Carlo. 2024. "Forecasting Lattice and Point Spatial Data: Comparison of Unilateral and Multilateral SAR Models" Forecasting 6, no. 3: 700-717. https://doi.org/10.3390/forecast6030036

APA Style

Grillenzoni, C. (2024). Forecasting Lattice and Point Spatial Data: Comparison of Unilateral and Multilateral SAR Models. Forecasting, 6(3), 700-717. https://doi.org/10.3390/forecast6030036

Article Metrics

Back to TopTop