Next Article in Journal
Biosynthesis of Silver Nanoparticles Using Barleria albostellata C.B. Clarke Leaves and Stems: Antibacterial and Cytotoxic Activity
Previous Article in Journal
DBCAN: DFormer-Based Cross-Attention Network for RGB Depth Semantic Segmentation
Previous Article in Special Issue
Refinement and Computation Method for Line/Body Topological Relationships
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatio-Temporal Heterogeneous Ensemble Learning Method for Predicting Land Subsidence

School of Geosciences and Info-Physics, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(18), 8330; https://doi.org/10.3390/app14188330
Submission received: 6 July 2024 / Revised: 30 August 2024 / Accepted: 12 September 2024 / Published: 16 September 2024
(This article belongs to the Special Issue Recent Advances in Geospatial Big Data Mining)

Abstract

:
The prediction of land subsidence is of significant value for the early warning and prevention of geological disasters. Although numerous land subsidence prediction methods are currently available, two obstacles still exist: (i) spatio-temporal heterogeneity of land subsidence is not well considered, and (ii) the prediction performance of individual models is unsatisfactory when the data do not meet their assumptions. To address these issues, we developed a spatio-temporal heterogeneous ensemble learning method for predicting land subsidence. Firstly, a two-stage hybrid spatio-temporal clustering method was proposed to divide the dataset into internally homogeneous spatio-temporal clusters. Secondly, within each spatio-temporal cluster, an ensemble learning strategy was employed to combine one time series prediction model and three spatio-temporal prediction models to reduce the prediction uncertainty of an individual model. Experiments on a land subsidence dataset from Cangzhou, China, show that the prediction accuracy of the proposed method is significantly higher than that of four individual prediction models.

1. Introduction

Land subsidence is a geological phenomenon characterized by a slow loss of ground elevation due to natural processes or human activities [1,2,3]. Land subsidence commonly affects large and medium-sized cities, with significant impacts on human activities (e.g., casualties), life, and ecology (e.g., damage to historical heritage, drainage system failures). Over 100 cities worldwide, including China [4,5], Indonesia [6], and Italy [7,8], suffer from tens of centimeters of land subsidence annually [3]. The primary challenge in studying land subsidence is accurately characterizing it and implementing effective prevention and control measures. The accurate prediction of land subsidence is crucial for the early monitoring of geological hazards.
Existing methods for predicting land subsidence can be categorized into three types: physical mechanisms-based methods [4,9,10], statistics-based methods, and machine learning-based methods. Physical mechanisms-based methods utilize field surveys and experiments to characterize complex geological and hydrophysical parameters for predicting land subsidence [11,12]. Bajni, Apuani, and Beretta [9] developed a hydro-geotechnical numerical model with good interpretability for predicting land subsidence. However, these models rely on complex assumptions about data and parameters, which often fail to align with actual land subsidence data, leading to poor fitting results. Statistical methods predict land subsidence by establishing mathematical functions related to the spatial and temporal characteristics of the study area, based on historical observation data. For example, Tang, et al. [13] applied a grey model to predict land subsidence in Shanghai’s Lujiazui area. Deng et al. [14] used time series deformation to build a prediction model based on an improved Grey–Markov model (IGMM) for predicting future subsidence. Statistical methods are simple, easy to use, and low-cost. However, these methods require data to meet certain ideal assumptions, which often conflict with the complex nonlinear characteristics of land subsidence data, resulting in poor prediction results. Machine learning algorithms have been widely used in land subsidence prediction tasks due to their powerful ability to extract high-dimensional, nonlinear features from data, including backpropagation (BP) algorithms, CART, support vector regression (SVR), convolutional neural networks (CNNs), Bayesian models, fuzzy set theory, and long short-term memory (LSTM) networks [15,16,17,18,19,20,21,22]. However, the underlying assumptions of these algorithms may be not satisfied in various scenarios, leading to limited generalization performance. Consequently, ensemble learning methods, which combine multiple individual base learners to improve prediction accuracy, have been applied to land subsidence prediction [23,24,25,26]. Shi et al. [27] proposed a land subsidence prediction model based on XGBoost for the Beijing Plain. Li, Liu, Tao, and Zhai [24] combined the Gradient Boosting Decision Tree (GBDT) and Random Forest (RF) methods to predict land subsidence. Both boosting and bagging strategies combine the same base model multiple times, thus categorizing them as homogeneous ensemble learning methods.
Despite the development of various prediction models, a critical gap remains in addressing the spatio-temporal heterogeneity of subsidence data and the inconsistent performance of individual models. Specifically, two obstacles still exist:
(i)
The spatio-temporal heterogeneity of land subsidence is not adequately considered in existing prediction models. Spatio-temporal heterogeneity refers to the variation in observed values of a spatio-temporal object across time, space, or both. Land subsidence is a spatio-temporally dynamic process, and its spatio-temporal heterogeneity is difficult to characterize. Existing land subsidence prediction methods primarily focus on time series analysis of historical data, often neglecting the spatial dependence of land subsidence data [27]. Recently, some methods have incorporated spatial clustering or time series clustering to consider spatial or temporal heterogeneity in land subsidence prediction [12,21]. However, these methods still struggle to effectively model the spatio-temporal heterogeneity of land subsidence, directly impacting the accuracy and reliability of predictions.
(ii)
The prediction performance of individual models is unsatisfactory when the data deviate from the models’ assumptions. Although some studies have attempted to solve this problem by using homogeneous ensemble learning methods, when there is an inherent bias in the base model, the ensemble model generated through homogeneous ensemble learning methods will inevitably carry this bias [28]. Consequently, homogeneous ensemble models cannot overcome the inherent biases of individual models, resulting in limited improvements in prediction accuracy.
To address these issues, a heterogeneous ensemble learning-based method that considers spatio-temporal heterogeneity was proposed for land subsidence prediction. The main contributions of this study are twofold:
(i)
To address the challenge of adequately considering the spatio-temporal heterogeneity of land subsidence in existing prediction models, we propose a two-stage hybrid clustering method based on co-clustering. In the first stage, the Bregman Block Average Co-clustering with I-divergence (BBAC_I) algorithm [29] is used to compress the original spatio-temporal data, dividing the large-scale land subsidence dataset into a series of small, internally homogeneous spatio-temporal clusters. In the second stage, the REgionalization with Dynamically Constrained Agglomerative clustering and Partitioning (REDCAP) method [30] is employed to further cluster the small spatio-temporal clusters, revealing complex cluster structures. Prediction models are then constructed within each spatio-temporal cluster, allowing for accurate modeling of spatio-temporal heterogeneity of land subsidence.
(ii)
To improve prediction performance when data do not meet the assumptions of individual models, this study adopts a heterogeneous ensemble learning method to combine the results of multi-view and multi-type machine learning prediction models. We developed a heterogeneous ensemble method based on the Blending strategy. For time series prediction, we chose the Sequence-to-Sequence model. For spatio-temporal prediction, we chose the GCN-Sequence-to-Sequence model, which captures the complex spatio-temporal dependencies in the graph structure; the Diffusion Convolutional Recurrent Neural Network model, which can capture the dynamic changes in spatio-temporal relationships; and the Graph Multi-Attention Network model, which incorporates the attention mechanism. The Blending strategy is then used to combine the prediction results of these four individual models, improving the accuracy of land subsidence predictions.
Experiments on a land subsidence dataset from Cangzhou, one of the most affected areas in China, demonstrate that the proposed method outperforms four state-of-the-art individual prediction models in terms of prediction accuracy. The proposed method considers spatio-temporal heterogeneity and combines multiple individual models, providing a novel approach for spatio-temporal data prediction that may aid in the early monitoring of ground geological hazards.

2. Methods

An overall research flowchart of this study is presented in Figure 1. First, a two-stage hybrid spatio-temporal clustering method based on co-clustering was proposed to divide the dataset into internally homogeneous spatio-temporal clusters. Then, within each cluster, a heterogeneous ensemble learning approach was developed to combine four heterogeneous prediction models for land subsidence prediction. In practice, some land subsidence outliers are useful for the early monitoring of ground geological disasters. Indeed, errors and meaningful outliers cannot be well distinguished. Therefore, in this study, we did not remove outliers from the dataset.

2.1. Two-Stage Hybrid Spatio-Temporal Clustering Method

In this study, a two-stage hybrid clustering method based on co-clustering was proposed to consider the spatio-temporal heterogeneity of the land subsidence dataset. In the first stage, the BBAC_I algorithm was applied to compress the large-scale spatio-temporal dataset, dividing the land subsidence dataset into a series of small, internally homogeneous spatio-temporal clusters. However, the limitation of BBAC_I is that it has difficulty detecting spatio-temporal clusters with complex shapes. Therefore, the REDCAP algorithm was employed in the second-stage to further cluster the small spatio-temporal clusters obtained by the BBAC_I, revealing complex cluster structures.

2.1.1. Co-Clustering of Large-Scale Land Subsidence Data

In this study, the original subsidence spatio-temporal data were treated as a data matrix, with rows representing different spatial points and columns representing different time observations. This study adopted a co-clustering strategy, treating spatial and temporal dimensions equally and dividing them into k spatial clusters and l temporal clusters to obtain a co-clustering spatio-temporal matrix. Banerjee et al. [29] proposed the Bregman co-clustering algorithm, which uses information divergence combined with multiple intra-cluster summary statistics. The Bregman co-clustering algorithm (BBAC_I) was used in the first stage because it preserves the mean of co-clusters, effectively identifying co-clusters and revealing the spatio-temporal pattern and its dynamic evolution in the dataset [31]. The BBAC_I algorithm is highly effective and has been successfully applied in document analysis, temperature studies, and phenological spatio-temporal pattern analysis. Hence, in this study, the BBAC_I algorithm was used to compress the original subsidence spatio-temporal dataset, dividing it into a series of small, internally homogeneous spatio-temporal clusters.
BBAC_I is suitable for co-clustering a data matrix with real and positive elements, representing a co-occurrence matrix of joint probabilities between two random variables in the rows and columns. It is an information-theoretic clustering method that treats co-clustering as an optimization problem, calculating the difference between the co-clustered and original data matrices as a loss function [31]. The BBAC_I algorithm optimizes this loss function to minimize information loss during spatio-temporal cluster partitioning. Assuming the spatio-temporal data matrix is O ST , where S is the number of spatial sampling points and T is the number of time clusters, the objective of BBAC_I is to partition rows and columns (Formulas (1) and (2)) to minimize the mutual information loss between O S ^ T ^ and O ST :
S : s 1 , s 2 , s m S ^ : s ^ 1 , s ^ 2 , s ^ k
T : t 1 , t 2 , t n T ^ : t ^ 1 , t ^ 2 , t ^ l
The main steps of the BBAC_I algorithm are as follows:
(1)
Randomly assign spatial sampling points and time slices to k spatial clusters and l temporal clusters, generating the initial co-clustered matrix O S ^ T ^ .
(2)
Calculate the mutual information loss between the co-clustered matrix and the original data matrix:
L I = D I O ST O S ^ T ^ = s ^   t ^   s s ^   t t ^   o s , t log o s , t a s , t
(3)
Iteratively update spatial and temporal cluster partitions for each spatial point and time cluster to minimize the mutual information loss between the co-clustered matrix and the original data matrix.
i.
Update the spatial cluster partition i to find the partition that minimizes mutual information loss:
i   = arg min i { 1 , , k }   L I
ii.
Update the temporal cluster partition j to find the partition that minimizes mutual information loss:
j = arg min i { 1 , , l }   L I
The algorithm concludes when the mutual information loss between iterations falls below a predefined threshold, indicating stability. However, BBAC_I cannot guarantee convergence to a global optimum for complex data, so multiple executions of the algorithm are often performed, with the best result used as the final clustering outcome.
For medium-sized and smaller datasets, the BBAC_I algorithm efficiently generates co-clustering results. However, land subsidence data often have high spatio-temporal resolutions and extensive coverage, with the number of spatio-temporal data matrix feature points exceeding millions. This scale poses convergence and computational challenges for BBAC_I, limiting its use to the first stage for dividing the dataset into intermediate internally homogeneous clusters.

2.1.2. Discovery of Spatio-Temporal Patterns Based on REDCAP

The REDCAP algorithm, which considers both attribute similarity and adjacency relationships of spatial points, is more efficient in large data sets compared to other regionalization methods. Therefore, to identify spatio-temporal clusters of complex morphology, the REDCAP algorithm was employed in the second-stage to cluster the small spatio-temporal clusters obtained from BBAC_I. The REDCAP algorithm adopts a hierarchical agglomerative–partitioning strategy to discover complex cluster structures.
The REDCAP algorithm first constructs a minimum spanning tree (MST) connecting all points based on adjacency relationships, using first-order (direct connection between points) or full-order (indirect connection between points) strategies. In this study, the adjacency relationships among spatial points in different spatio-temporal clusters are computed based on the union of their K-Nearest Neighbor (KNN) adjacency relations. If two spatio-temporal clusters have adjacent spatial points, they are considered adjacent; otherwise, they are considered non-adjacent. This rule defines the adjacency relationships among the intermediate spatio-temporal clustering results.
To ensure strong spatio-temporal consistency, the full-order strategy is used to construct adjacency relationships among small spatio-temporal clusters. The REDCAP algorithm can be classified into six strategies based on the chosen adjacency relationship determination and inter-cluster distance metric: first-order-SLK, first-order-ALK, first-order-CLK, full-order-SLK, full-order-ALK, and full-order-CLK. Among these, full-order-CLK generally provides high clustering quality and is therefore chosen for agglomerative clustering in this study, generating a spatial connectivity tree using a bottom-up approach [30].
The REDCAP algorithm segments the edges of the spatial connectivity tree in a top-down manner to obtain the final clustering results. This study selects the optimal cut-edge result based on the principle of regional homogenization optimization, using the sum of squared errors (SSE) as the heterogeneity measure:
H R = j = 1 d   i = 1 n r   x ij x j ¯ 2
where R is the subtree containing the spatial objects, H R represents its heterogeneity, d is the attribute dimension of each spatial object, n r is the number of objects in R, x ij is the j-th attribute value of the i-th object, and x j ¯ is the mean value of the j-th attribute value of all objects in R.
When the number of clusters k is set for the REDCAP algorithm, it should include k subtrees obtained from the segmentation. The overall heterogeneity H k of the partitioning result is the sum of the heterogeneity of the k subtrees:
H k = j = 1 k   H R j
During each split, the REDCAP algorithm divides the region R into two sub-regions, R a and R b , identifying the edge with the maximum homogeneity gain h g * R (i.e., the reduction in heterogeneity), calculated as:
h g * R = max H R H R a H R b
By iteratively finding the optimal cut-edge based on the principle of maximum homogeneity gain, the MST is segmented until the number of subtrees reaches k. At this point, the REDCAP algorithm is complete, and the spatial objects in each subtree form a spatial cluster. The REDCAP algorithm can produce clustering results with varying numbers of clusters. The optimal number of clusters is determined using the silhouette coefficient [32]. The average silhouette coefficient of all points in the clustering result is taken as the silhouette coefficient. Typically, the k value corresponding to the maximum silhouette coefficient is chosen.

2.2. Heterogeneous Ensemble Learning-Based Prediction Method

2.2.1. Individual Models for Land Subsidence Prediction

To ensure the diversity of surface subsidence prediction, four models were selected from different perspectives to predict land subsidence. For time series prediction, the Sequence-to-Sequence model, which makes minimal assumptions on the sequence structure, was chosen to model the time dependencies [33]. For spatio-temporal series prediction, the Graph Convolutional Sequence-to-Sequence model (GCN-Seq2Seq) [34] was chosen to capture the local spatial structure representation of land subsidence; the Diffusion Convolutional Recurrent Neural Network model (DCRNN) [35] was chosen to model highly nonlinear spatio-temporal dependencies; and the Graph Multi-Attention Network model (GMAN) [36], which integrates an attention mechanism, was used to model the impact of the spatio-temporal factors on land subsidence.
(i)
Sequence-to-Sequence model
The Seq2Seq model is an encoder–decoder model that performs well in natural language processing and time series analysis. This study uses the Seq2Seq model to predict sedimentation from a temporal perspective. The Seq2Seq model consists of three main components: the encoder, intermediate vector, and decoder, all of which are built using RNN units [33].
The encoder transforms each element of the input sequence into a low-dimensional embedding vector to generate a hidden state, represented as follows:
h t = f R x t , h t 1
where f R represents the function of the RNN unit.
The intermediate vector connects the encoder and the decoder and can be obtained by the hidden state of the last RNN unit of the encoder or the weighted average of all hidden states.
The decoder gradually builds the output sequence based on the intermediate vector. Assuming that the predicted output at time step t is y t , the calculation is as follows:
y t = f a W h t + b
The basic mechanism behind Seq2Seq model prediction is to construct hidden states and calculate intermediate vectors through the encoder, and then initialize and predict future moments using the decoder. To alleviate the error accumulation problem in predictions, a scheduled sampling strategy is adopted, where the decoder uses a certain probability of using real label data or the prediction results of the previous time step as input during the training phase [33]. This study uses the gated recurrent unit (GRU) method as the core component of the Seq2Seq model to enhance both performance and stability.
(ii)
Graph Convolutional Sequence-to-Sequence model
The GCN-Seq2Seq model combines the Graph Convolutional Network (GCN) [34] with a sequence-to-sequence architecture, focusing on processing spatio-temporal sequences with graph structures. It simultaneously considers the spatial interdependence and temporal continuity, capturing the spatial propagation patterns and temporal evolution trends of land subsidence to enhance predictions’ accuracy and reliability. In the GCN-Seq2Seq model, GCN is responsible for capturing the structural characteristics of graph data. By performing convolution operations on graph nodes and identifying and utilizing the dependencies between nodes, the features of graph data can be effectively extracted and represented.
The main steps of GCN-Seq2Seq model prediction include graph feature extraction, sequence encoding, and sequence decoding [34]. During graph feature extraction, GCN updates the feature representation of each node, as follows:
h v l + 1 = σ u N v v 1 c vu W l h u l
where σ is a nonlinear activation function (such as ReLU), N(v) represents the set of neighboring nodes of node v , and c vu is a normalization constant. The commonly used symmetric normalization d v d u is deployed, where d v and d u are the degrees of nodes v and u, respectively, W l is the weight matrix of the l-th layer, and h u l is the feature representation of the neighboring node u in the l-th layer.
After graph feature extraction, the model proceeds to the sequence encoding, similar to the Seq2Seq encoding process. In the sequence decoding stage, the model gradually generates a predicted sequence based on the encoded information. By combining graph structure and sequence information, the GCN-Seq2Seq model fully integrates spatial and temporal dependencies and is suitable for complex graph data and sequence prediction tasks.
(iii)
Diffusion Convolutional Recurrent Neural Network model
The DCRNN model [35] is a deep learning framework for spatio-temporal series prediction that combines Graph Convolutional Networks (GCNs) and recurrent neural networks (RNNs). DCRNN captures land subsidence propagation characteristics through diffusion convolutional layers and uses RNNs to handle time series data, adapting to nonlinear and dynamic changes [35]. This study selects the DCRNN model as the primary prediction model to account for the dynamic changes in spatio-temporal relationships. The DCRNN model prediction includes two steps. First, in the spatial dependency modeling stage, DCRNN models spatial dependencies through random walks on the graph and the state transition matrix D O 1 , as follows:
P t | 1 = α × D O 1 × P t + 1 α × P 0
where P t is the flow distribution at time step t and P 0 is the initial distribution. In the diffusion convolution process, DCRNN applies a defined diffusion convolution operation to the graph signal and filter, combining the forward and backward diffusion processes to capture the upstream and downstream impacts, as follows:
H l + 1 = W l × H l + b l M l
where H l + 1 is the node representation of the (l + 1)-th layer, W l is the weight matrix, b l is the bias term, M l is the diffusion matrix of the l-th layer, and represents element-by-element multiplication. Then, in the temporal dynamic modeling stage, DCRNN uses gated recurrent units (GRUs) to model temporal dependencies and forms DCGRUs by replacing matrix multiplications in GRUs with diffusion convolutions. DCGRUs process time series data by resetting gates, updating gates, and utilizing candidate hidden state mechanisms, all combined with spatial diffusion information. Finally, the DCRNN uses a Seq2Seq structure for multi-step prediction, with the encoder processing historical time series and the decoder generating predictions based on the final encoder state.
(iv)
Graph Multi-Attention Network model
The GMAN model is an encoder–decoder architecture that integrates an attention mechanism, including the ST-Attention module (STAtt) with residual connections [37]. GMAN incorporates graph structure and temporal information into a multi-attention mechanism through spatio-temporal embeddings, enriching contextual information and enhancing the expression ability of spatio-temporal data [36].
The basic mechanism of the GMAN model prediction is as follows. First, the spatio-temporal embedding module learns vertex representations through the node2vec method and inputs them into a two-layer, fully connected neural network to obtain spatial embedding. Since spatial embeddings provide only static representation, GMAN also introduces temporal embedding, encoding each time step into a vector, which is then processed through a fully connected neural network. The spatial embedding and temporal embedding are fused into the spatio-temporal embedding module (STE), which is defined as follows:
e v i , t j = e v i S + e t j T
The temporal attention module includes the spatial attention module, the temporal attention module, and the gated fusion. The model represents the input of the l-th block as H ( l 1 ) and the outputs of the spatial and temporal attention mechanisms as H S ( l ) and H T ( l ) , respectively. After gated fusion, the output of the l-th block is obtained as H ( l ) .
GMAN employs an encoder–decoder architecture with a residual-connected ST-Attention block to effectively capture and integrate features in both spatial and temporal dimensions, making it suitable for complex spatio-temporal data and multi-dimensional sequence prediction tasks.

2.2.2. Blending-Based Ensemble Learning Strategy

Blending [38] is a heterogeneous ensemble learning algorithm and a variant of Stacking [39]. Similar to Stacking, Blending improves overall prediction performance by combining the prediction results of multiple independent models of different types. However, Blending uses a hold-out strategy for dataset splitting during training, in contrast to the cross-validation strategy used in Stacking. The core mechanism of Blending is that by combining individual models, the advantages of each model can be fully utilized to achieve a prediction effect that exceeds that of an individual model [40]. The Blending algorithm contains two important components: base learners and a meta-learner. The base learners are a diverse set of pre-selected models. The base learners are as diverse as possible to capture the advantages of different aspects of the data. First, the base learners are trained on the training set. The meta-learner uses the prediction results of the base learner on the test set as input to learn how to best combine the predictions of the base learners, so as to combine different base learners to obtain more accurate prediction results. The four individual models for land subsidence prediction introduced in Section 2.2.1 were combined using the Blending algorithm. The flowchart of the Blending ensemble algorithm is illustrated in Figure 2.
For the selection of meta-learners, existing ensemble learning research has shown that simple linear models can achieve good performance [41], so linear regression was selected in this study as the meta-learner. The basic mechanism involves using the prediction results of the base learners as independent variables and generating the final prediction value through linear combination, represented as follows:
Y = β 1 X 1 + β 2 X 2 + + β i X I + ε
where β i represents the contribution of an individual model, i.e., the weight; X i represents the predicted value of an individual model. During the training process, the weight of each base learner is determined by minimizing the prediction error. In this way, the Blending algorithm effectively combines the prediction results of different models achieving more accurate land subsidence prediction results. The four individual models are implemented in the pytorch library and the TensorFlow python library. The pseudo-code of the proposed method (Algorithm 1) is presented below.
Algorithm 1: Spatio-temporal ensemble predicting (LS_Mat, k, l, k_final, split_ratio)
// Two-stage hybrid spatio-temporal clustering
Co_clusters= BBAC_I (LS_Mat, k, l); //first stage co_clustering using BBAC_I algorithm
Aggregated_clusters = Aggregate (LS_Mat, Co_clusters) //Aggregate each co-cluster into single object
ST_clusters = REDCAP(Aggregated_clusters, k_final); // second stage clustering using REDCAP algorithm
// Blending-based ensemble learning for prediction
Base_models = [S2S, GCN_S2S,DCRNN, GMAN]
Testing_results_all = []
for cluster in ST_clusters do
   Train_set, Test_set, Valiadate_set = Split(cluster, split_ratio)
   for model in Base_models do
      Trained_model = Train(model)
      Testing_result = Trained_model(Test_set)
      Testing_results_all.add(Testing_result)
   end for
   // Train Linear regression model for Blending ensemble
     Trained_LR = LR.fit(Testing_results_all)
   Valiadating_result = Trained_LR(Valiadate_set)
end for

2.3. Model Evaluation

To comprehensively evaluate the prediction performance of each model, this study selected root mean square error (RMSE), mean absolute error (MAE), and the coefficient of determination ( R 2 ) as the main evaluation indicators. RMSE measures prediction error by calculating the square root of the average squared deviations between the predicted value and actual observed value. RMSE is highly sensitive to large errors, making the performance of the model in extreme cases particularly important. MAE provides a more intuitive error assessment by calculating the average of the absolute deviations between the predicted value and actual value. The advantage of MAE is its simple calculation and ease of understanding, making it a direct indicator for evaluating the prediction accuracy. The R 2 indicator measures the model’s ability to explain variability in the data. The R 2 value ranges from 0 to 1, with values closer to 1 indicating better model fit and prediction accuracy. The specific calculation formulas for each metric (RMSE: Equation (16); MAE: Equation (17); R 2 : Equation (18)) are as follows:
RMSE = 1 N S N E s = 1 N S t = 1 N E y s t y ^ s t 2
MAE = 1 N S N E s = 1 N S t = 1 N E y s t y ^ s t
R 2 = 1 s = 1 N S t = 1 N E ( y s t y ^ s t ) 2 s = 1 N S t = 1 N E ( y s t Y ̄ ) 2
where y s t and y ^ s t represent the observed and predicted subsidence values at monitoring station j at time step t, respectively; N E is the total number of subsidence monitoring point samples in the test set; N S is the total number of subsidence monitoring points; and Y ̄ is the overall average of the observed values.

3. Experiments

3.1. Study Area

The Cangzhou area is located in the southeast of Hebei Province, China. It is 165 km wide from north to south and 187 km long from east to west, with a total area of about 14,304.26 km2. Since the 1960s, to meet the needs of economic development, the region has conducted large-scale groundwater resource exploitation, resulting in a sharp drop in groundwater levels, which in turn has caused the most severe land subsidence problem in North China. Therefore, Cangzhou was selected as the study area for this research.

3.2. Dataset

Eighty Sentinel-1A images acquired in the Cangzhou area between January 2017 and December 2019 were used as the dataset. The multiple Master-image Coherent Target Small-baseline Interferometric SAR method was used [42] to generate the InSAR deformation time series. Figure 3 shows the spatio-temporal distribution of spatio-temporal land subsidence data in the Cangzhou area (for part of the timestamps). It indicates that high subsidence values are mainly distributed in the western part of Cangzhou, whereas the eastern part of the Cangzhou area mainly consists of points with low subsidence values and uplift points. In terms of temporal distribution, high subsidence points are mainly found in the early stages, while low subsidence points are predominantly found in the later stages. This study organizes the subsidence data into a data matrix of (69,852, 80) for spatio-temporal clustering. The three representative spatio-temporal clusters obtained by clustering are selected as input data for subsequent predictions.

3.3. Data Preprocessing

Since the BBAC_I algorithm requires each entry in the data matrix to be positive, the original data matrix is transformed into a positive matrix by adding the opposite number of the minimum value. After spatio-temporal clustering, all subsidence values are transformed to [0, 1] and input into the ensemble model for prediction.

3.4. Parameter Setting

In this study, the number of clusters in the BBAC_I algorithm and the REDCAP algorithm of the two-stage hybrid clustering needs to be manually specified. To determine the number of spatial and temporal clusters in the BBAC_I algorithm, this study tested spatial clusters ranging from 200 to 400 with an interval of 50 and temporal clusters ranging from 3 to 8 with an interval of 1. It was found that when the number of spatial clusters was 200 and the number of temporal clusters was 3, the BBAC_I algorithm could reach a convergence state, and the clustering results have the minimum mutual information loss with the original data matrix. To determine the number of clusters k in the REDCAP algorithm, the silhouette coefficient was used in this study to find the optimal clustering. As shown in Figure 4, when k is 4, the silhouette coefficient reaches an inflection point; thus, the final number of clusters is set to 4.
To achieve optimal results, the BBAC_I algorithm was set with a maximum of 3000 iterations, a convergence threshold of 1 × 10−5, and 200 repetitions. For the adjacency relationship and difference measurement method of the REDCAP algorithm, Guo [30] demonstrated that using the full-order configuration for constructing the adjacency relationship and CLK for the difference measurement method achieves the best clustering performance compared with other methods. Therefore, this study adopts the REDCAP algorithm under the full-order-CLK configuration.
This study sets a unified input sequence length (i.e., 10 steps) and sliding window step length (i.e., 1 step) for the selected prediction model. The prediction output sequence length is set to 1 step (each step corresponds to 1 monitoring interval). These settings ensure that the model effectively captures short-term dynamic changes in land subsidence, providing accurate data support for prediction analysis. This study uses the KNN algorithm to find the 10 nearest neighbors around each point to establish spatial connections and uses the spatial distance between points as the weight of the edges to construct a spatial graph.
The dataset was split into training and testing sets with an 80-20 ratio, meaning 80% of the data was used for training the model, while the remaining 20% was reserved for testing its performance. This 80-20 split is a common practice in machine learning, as it provides a balanced approach: the majority of the data is used to train the model, allowing it to learn effectively, while a sufficient portion is held back to evaluate how well the model generalizes to unseen data. Within the training set, 20% was further utilized to train the ensemble learner. This step ensures that the ensemble learner, which typically combines predictions from multiple models, is trained on diverse subsets of the data, enhancing its robustness. A potential bias of the 80-20 split is that the model may not encounter all data variations during training, particularly in small or unevenly distributed datasets. To mitigate this, techniques such as cross-validation can be employed, where the training set is split into multiple subsets, and the model is trained and validated on these different subsets iteratively. This approach ensures the model is trained on the maximum data variation, reducing the risk of overfitting or underfitting. Overall, the 80-20 split was chosen to balance model training efficiency and evaluation reliability, while potential biases were addressed through careful validation strategies.
To comprehensively evaluate and compare the performance of different individual models in predicting spatio-temporal land subsidence, this study selected three clusters with representative subsidence patterns based on clustering results and conducted a detailed evaluation of the prediction accuracy of four individual models. In this study, a trial-and-error approach was employed for parameter tuning. Specifically, different parameter combinations were tried, and their performance was evaluated until the highest-performing set was determined. While this process was time-consuming, it ensured that the model was in its optimal state for the experiment. The specific parameter configurations of the basic models are listed in Table 1.

3.5. Results

3.5.1. Spatio-Temporal Clustering Results

Following the two-stage hybrid spatio-temporal clustering, the spatio-temporal data matrix of land subsidence was divided into four clusters. Based on the mean value of subsidence in each spatio-temporal cluster, the four homogeneous spatio-temporal clusters are defined as slight uplift, slight subsidence, moderate subsidence, and severe subsidence. According to these definitions, the spatio-temporal clustering results are shown in Figure 5. The time clusters are arranged in chronological order, and the spatial clusters are arranged by the mean value of the rows within the cluster. The clustering result matrix shows that most spatio-temporal points fall into the slight uplift and slight subsidence categories. Statistical information of the spatio-temporal clustering results is presented in Table 2. The mean values for the four subsidence patterns are as follows: slight uplift (15 mm), slight subsidence (−17 mm), moderate subsidence (−65 mm), and severe subsidence (−124 mm). Judging from the proportion of the four spatio-temporal subsidence patterns, the spatio-temporal points of land subsidence in the entire study area are mainly categorized as slight uplift and slight subsidence, accounting for 85.7%, indicating that the overall subsidence in the study area is slight. The spatio-temporal points of land subsidence with moderate and severe subsidence patterns account for a small proportion, totaling 14.3%.
(i)
Spatial distribution pattern
The spatio-temporal distribution pattern of subsidence was initially analyzed from a spatial perspective. Columns corresponding to the same time cluster in the spatio-temporal clustering matrix were merged to derive three spatial patterns, as illustrated in Figure 6. The proportion of each type of spatio-temporal subsidence pattern points in each spatial pattern and the number of occupied time periods are presented in Table 3.
In spatial pattern 1, the slight subsidence and slight uplift patterns account for nearly all spatio-temporal points, with the slight subsidence pattern representing the largest proportion. Moderate subsidence is represented minimally, while severe subsidence is absent in spatial pattern 1. In spatial pattern 3, the slight subsidence and slight uplift patterns still account for most spatio-temporal points, with the slight uplift pattern representing the largest proportion.
(ii)
Temporal distribution pattern
Additionally, the subsidence pattern was analyzed from a temporal perspective. Spatial points exhibiting the same changes in the temporal dimension in the clustering result matrix were merged and classified into six temporal patterns. As illustrated in Figure 7, six temporal patterns are identified based on the temporal and spatial patterns of spatial points in the time series: stable slight uplift (temporal pattern 1, Figure 7a), stable slight subsidence (temporal pattern 2, Figure 7b), slight subsidence–slight uplift (temporal pattern 3, Figure 7c), slight subsidence–moderate subsidence (temporal pattern 4, Figure 7d), slight subsidence–moderate subsidence–severe subsidence (temporal pattern 5, Figure 7e), and moderate subsidence–severe subsidence (temporal pattern 6, Figure 7f). The spatial distribution of each temporal pattern is presented in Figure 8, with different color markings distinguishing the spatial points of various temporal patterns. The proportion of spatial points in various temporal patterns within the spatial extent of Cangzhou area is detailed in Table 4.
The overall temporal pattern of land subsidence in Cangzhou area falls into three categories: stable (stable slight uplift, stable slight subsidence), fluctuating (slight subsidence–slight uplift, slight subsidence–moderate subsidence, and moderate subsidence–severe subsidence), and violent fluctuation (slight subsidence–moderate subsidence–severe subsidence). Of these, the stable temporal pattern accounts for the largest proportion (more than 50%), followed by the fluctuating temporal pattern (accounting for more than 40%), while the violent fluctuation temporal pattern accounts for the lowest proportion (less than 5%). The spatial distribution of the six temporal pattern points in the Cangzhou area is illustrated in Figure 8. Slight subsidence–moderate subsidence is primarily found in the southwest part, while moderate subsidence–severe subsidence is primarily found in the westernmost part. This indicates that the land subsidence conditions in these areas have been continuously intensifying over the 80-period timeframe. Slight subsidence–moderate subsidence–severe subsidence is primarily found in the southwest part; moderate subsidence–severe subsidence is mainly distributed in the northwest part. This suggests that the land subsidence in these areas is worsening and necessitates intervention and control by relevant authorities.
This study employs a two-stage hybrid spatio-temporal clustering method to conduct co-clustering on the land subsidence data of the Cangzhou area. Although each spatio-temporal cluster is roughly continuous in space, there are still a considerable number of points mixed with other clusters. To facilitate subsequent subsidence value prediction of land subsidence spatio-temporal data, the spatio-temporal clustering results were post-processed to obtain spatio-temporally continuous and homogeneous subclusters. First, KNN is employed to identify the spatio-temporal neighbors of each spatio-temporal object. Then, the spatio-temporal cluster IDs of all neighbors in its spatio-temporal neighborhood are voted on, and the cluster ID with the highest vote is assigned as the final ID of the object. Finally, the isolated points from the KNN spatio-temporal neighbor voting results are manually assigned the IDs of the surrounding points to obtain the final clustering results. For subsequent land subsidence prediction, this study selected three representative subsidence pattern clusters: the slight subsidence pattern (cluster 1), the moderate subsidence pattern (cluster 2), and the severe subsidence pattern (cluster 3). The statistical information of these three spatio-temporal clusters is presented in Table 5. As the subsidence pattern becomes more severe, the size of the spatio-temporal cluster matrix increases correspondingly.

3.5.2. Spatio-Temporal Prediction Results

(i)
Comparison of overall prediction accuracy
The overall prediction accuracy of the ensemble model and the four base models when predicting subsidence values for cluster 1, cluster 2, and cluster 3 is displayed in Table 6. The experimental results demonstrate that the ensemble prediction model developed in this study exhibits the best prediction performance among the five models across various subsidence clusters. It is evident that the prediction accuracy of subsidence values significantly improves after the Blending ensemble is implemented in the model. Compared with the Seq2Seq model, the ensemble model reduces the RMSE by up to 86.2% and the MAE by up to 86.7% for cluster 3; compared with the GCN-Seq2Seq model, the RMSE is reduced by up to 86.9% and the MAE by up to 89.1% for cluster 3; compared with the DCRNN model, the RMSE is reduced by up to 83.4% and the MAE by up to 81.2% for cluster 3; and compared with the GMAN model, the RMSE is reduced by up to 79.2% and the MAE by up to 82.5% for cluster 3. Methods based on spatio-temporal graph neural networks (such as GCN-Seq2Seq, DCRNN, and GMAN) generally outperform Seq2Seq on the three clusters because these models consider the spatio-temporal correlation between different monitoring points. Among the methods based on spatio-temporal graph neural networks, GCN-Seq2Seq has the best prediction performance for cluster 1 (slight subsidence area), while GMAN performs best for cluster 2 (moderate subsidence area) and cluster 3 (severe subsidence area). Compared to the basic models, the prediction accuracy of the ensemble model is significantly improved for cluster 3, as the basic models achieve their best performance on cluster 3, leading to optimal ensembling results.
To intuitively compare the prediction effects of different models, this study visualized the prediction results of the five models across clusters. Given the short time series of the data, the time series in the prediction results are averaged, and the prediction results for each monitoring point are visualized, as shown in Figure 9, Figure 10 and Figure 11. The results indicate that the prediction outcomes of the ensemble model are highly consistent with the actual subsidence values and significantly outperform the other base prediction models.
This study further utilized a paired t-test to evaluate whether the prediction accuracy of the ensemble model is statistically significantly improved compared to the base model [43]. This test helps assess whether the observed differences in prediction performance are statistically significant or likely due to random variation. For each model M i , the error sequence is calculated:
E i = y A 1 y ^ A 1 , y A 2 y ^ A 2 , , y A t y ^ A t , y A N E y ^ A N E
where y A t and y ^ A t represent the observed and true values of the subsidence at each monitoring point under the average time step.
For two different models M i and M j , the null hypothesis (no significant difference in the prediction accuracy of the two models) and the alternative hypothesis (the prediction accuracy of the M i model is significantly better than the M j model) can be defined as follows:
H 0 : μ E i = μ E j
H 1 :   μ E i < μ E j
In this study, the significance level was set to 0.05. The statistical comparison of the prediction accuracy of different models is shown in Table 7, Table 8 and Table 9. If cell C ij is marked as “Yes”, it means that the prediction accuracy of model M i is statistically significantly better than model M j ; otherwise, there is no significant difference in the prediction accuracy of the two models. It can be observed that the prediction accuracy of the ensemble model is statistically significantly better than that of the base models in all data sets.
(ii)
Comparison of prediction accuracy at extreme monitoring points
This study further compared the prediction performance of different models at peak monitoring points. The sliding window method, with a window size set to 10, was used to identify the peak points of the subsidence value. For each cluster, the maximum and minimum values within the sliding window were identified as peak points. Figure 12 illustrates the peak points of the average time series of all stations for each cluster. As seen in Table 10, the performance of the ensemble model at peak monitoring points is also better than that of the base model; the prediction accuracy of models considering spatio-temporal dependence is generally better than that of models considering only time dependence.
This study also conducted a paired t-test on the prediction results of peak monitoring point data to determine whether the prediction accuracy of ensemble model at these points is statistically better than that of the basic model. As shown in Table 11, Table 12 and Table 13, compared to the basic model, the ensemble model achieves statistically significant improvement at extreme monitoring points.

3.6. Discussion

Experimental results indicate that considering the spatio-temporal heterogeneity of land subsidence can significantly improve the prediction accuracy of the model. Additionally, the Blending ensemble model significantly outperforms individual models in terms of overall prediction accuracy and peak monitoring point prediction performance. The main reasons for this are attributed to the following two aspects:
(i)
The proposed two-stage hybrid clustering method divides the large-scale land subsidence dataset into internally homogeneous spatio-temporal clusters. The heterogeneous ensemble model then shows excellent prediction performance on these clusters, demonstrating that considering the spatio-temporal heterogeneity improves the prediction performance.
(ii)
Ensemble learning enhances prediction accuracy and reduces errors by combining the results of multiple individual models. In this study, the Blending ensemble strategy effectively combines the strengths of different individual models to achieve higher accuracy in complex subsidence prediction tasks. Although the predictions of individual model may not meet the requirements in some cases, satisfactory land subsidence prediction results are obtained after ensembling, greatly improving accuracy.
The experimental results in Section 3.5 highlight the strengths and weaknesses of each individual model. For the slight subsidence pattern (Cluster 1), Seq2Seq, which only considers temporal autocorrelation performs worse than the methods considering spatio-temporal autocorrelation. This indicates that accounting for spatial autocorrelation can improve the prediction accuracy for slight subsidence patterns with relatively uniform spatial distribution. For moderate and severe subsidence patterns (Clusters 2 and 3), spatio-temporal prediction models do not always perform better than the time series prediction model. We found that GCN-Seq2Seq performs worse than Seq2Seq. This may be due to the fact that the user-specified spatial proximity relationship cannot effectively capture the spatial autocorrelation [43] due to the datasets exhibiting significant local spatial heterogeneity. In this situation, the diffusion convolution operation [35] and the attention mechanism [36] are more suitable for modeling spatio-temporal autocorrelation.
Additionally, this study’s results were compared with a previous study on land subsidence prediction in Cangzhou, which accounted for spatial heterogeneity [12]. We found that both of them consider the heterogeneity of land subsidence dataset, so the prediction performance can be improved compared with the global methods. Compared to [12], this study further addresses spatio-temporal heterogeneity in land subsidence prediction. Actually, the model constructed in [12] is similar to the Seq2Seq model tested in this study. According to the experimental results given in Section 3.5, we can infer that the model developed in this study outperforms that developed in [12]. In addition, the proposed method in this study is different from the related study in numerical performance. The reason is that the related study was proposed to predict differential deformation values, while this study predicts cumulative subsidence values.

4. Conclusions

This study proposes a spatio-temporal heterogeneous ensemble learning method for predicting land subsidence. The achievements of this study can be summarized as follows. (i) A hybrid spatio-temporal clustering method is developed to segment the dataset into a series of homogenous spatio-temporal clusters, thereby effectively considering the spatio-temporal heterogeneity of the land subsidence data. (ii) We construct a heterogeneous ensemble prediction model by combining four different individual prediction models, thus leveraging the advantages of multiple models through ensemble learning methods to adapt to the complex land subsidence data distribution and improve prediction accuracy. (iii) We conducted an experiment on a land subsidence dataset collected in Cangzhou, China. The results show that the proposed method performs better than four state-of-the-art models, providing a reference for effectively managing and controlling land subsidence. This study introduces a new approach to spatio-temporal prediction, particularly as a feasible solution for addressing the spatio-temporal heterogeneity of large-scale datasets.
Despite its strong performance in subsidence prediction, the ensemble model has some limitations. Ensemble learning performance is highly dependent on the chosen base models. Additionally, Blending is just one of many ensemble learning strategies, and other methods are worth considering. Their effectiveness in improving model performance and applicability requires further evaluation. Furthermore, the complexity of the ensemble model is typically higher than that of individual models, which may lead to higher computational costs and reduced interpretability. Future work will focus on comparing multiple ensemble learning strategies and developing methods with lower computational requirements.

Author Contributions

Conceptualization, B.Z. and G.W.; data curation, G.W. and Q.W.; methodology, B.Z., G.W., and J.L.; visualization, B.Z. and G.W.; writing—original draft, B.Z. and G.W.; writing—review and editing, Q.W. and M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded through support from the National Natural Science Foundation of China (NSFC) (No. 42430110); the Water conservancy science and technology project of Guizhou, China (No. KT202410); and the Research Foundation of the Department of Natural Resources of Hunan Province (Grant No. HBZ20240132).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author/s.

Acknowledgments

The authors gratefully acknowledge the comments from the reviewers.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Castellazzi, P.; Garfias, J.; Martel, R.; Brouard, C.; Rivera, A. InSAR to support sustainable urbanization over compacting aquifers: The case of Toluca Valley, Mexico. Int. J. Appl. Earth Obs. Geoinf. 2017, 63, 33–44. [Google Scholar] [CrossRef]
  2. Galloway, D.L.; Erkens, G.; Kuniansky, E.L.; Rowland, J.C. Preface: Land subsidence processes. Hydrogeol. J. 2016, 24, 547–550. [Google Scholar] [CrossRef]
  3. Zhou, C.; Gong, H.; Chen, B.; Li, X.; Li, J.; Wang, X.; Gao, M.; Si, Y.; Guo, L.; Shi, M. Quantifying the contribution of multiple factors to land subsidence in the Beijing Plain, China with machine learning technology. Geomorphology 2019, 335, 48–61. [Google Scholar] [CrossRef]
  4. Chen, B.; Gong, H.; Li, X.; Lei, K.; Zhu, L.; Gao, M.; Zhou, C. Characterization and causes of land subsidence in Beijing, China. Int. J. Remote Sens. 2017, 38, 808–826. [Google Scholar] [CrossRef]
  5. Gao, M.; Gong, H.; Chen, B.; Li, X.; Zhou, C.; Shi, M.; Si, Y.; Chen, Z.; Duan, G. Regional land subsidence analysis in eastern Beijing plain by insar time series and wavelet transforms. Remote Sens. 2018, 10, 365. [Google Scholar] [CrossRef]
  6. Du, Z.; Ge, L.; Ng, A.H.-M.; Zhu, Q.; Yang, X.; Li, L. Correlating the subsidence pattern and land use in Bandung, Indonesia with both Sentinel-1/2 and ALOS-2 satellite images. Int. J. Appl. Earth Obs. Geoinf. 2018, 67, 54–68. [Google Scholar] [CrossRef]
  7. Lanari, R.; Berardino, P.; Bonano, M.; Casu, F.; Manconi, A.; Manunta, M.; Manzo, M.; Pepe, A.; Pepe, S.; Sansosti, E. Surface displacements associated with the L’Aquila 2009 Mw 6.3 earthquake (central Italy): New evidence from SBAS-DInSAR time series analysis. Geophys. Res. Lett. 2010, 37. [Google Scholar] [CrossRef]
  8. Rosi, A.; Tofani, V.; Agostini, A.; Tanteri, L.; Stefanelli, C.T.; Catani, F.; Casagli, N. Subsidence mapping at regional scale using persistent scatters interferometry (PSI): The case of Tuscany region (Italy). Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 328–337. [Google Scholar] [CrossRef]
  9. Bajni, G.; Apuani, T.; Beretta, G.P. Hydro-geotechnical modelling of subsidence in the Como urban area. Eng. Geol. 2019, 257, 105144. [Google Scholar] [CrossRef]
  10. Chen, M.; Tomás, R.; Li, Z.; Motagh, M.; Li, T.; Hu, L.; Gong, H.; Li, X.; Yu, J.; Gong, X. Imaging land subsidence induced by groundwater extraction in Beijing (China) using satellite radar interferometry. Remote Sens. 2016, 8, 468. [Google Scholar] [CrossRef]
  11. Guo, H.; Yuan, Y.; Wang, J.; Cui, J.; Zhang, D.; Zhang, R.; Cao, Q.; Li, J.; Dai, W.; Bao, H. Large-scale land subsidence monitoring and prediction based on SBAS-InSAR technology with time-series sentinel-1A satellite data. Remote Sens. 2023, 15, 2843. [Google Scholar] [CrossRef]
  12. Liu, Q.; Zhang, Y.; Wei, J.; Wu, H.; Deng, M. HLSTM: Heterogeneous long short-term memory network for large-scale InSAR ground subsidence prediction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8679–8688. [Google Scholar] [CrossRef]
  13. Tang, Y.-Q.; Cui, Z.-D.; Wang, J.-X.; Yan, L.-P.; Yan, X.-X. Application of grey theory-based model to prediction of land subsidence due to engineering environment in Shanghai. Environ. Geol. 2008, 55, 583–593. [Google Scholar] [CrossRef]
  14. Deng, Z.; Ke, Y.; Gong, H.; Li, X.; Li, Z. Land subsidence prediction in Beijing based on PS-InSAR technique and improved Grey-Markov model. GISci. Remote Sens. 2017, 54, 797–818. [Google Scholar] [CrossRef]
  15. Ranjbar, A.; Ehteshami, M. Development of an uncertainty based model to predict land subsidence caused by groundwater extraction (case study: Tehran Basin). Geotech. Geol. Eng. 2019, 37, 3205–3219. [Google Scholar] [CrossRef]
  16. Hakim, W.L.; Fadhillah, M.F.; Park, S.; Pradhan, B.; Won, J.-S.; Lee, C.-W. InSAR time-series analysis and susceptibility mapping for land subsidence in Semarang, Indonesia using convolutional neural network and support vector regression. Remote Sens. Environ. 2023, 287, 113453. [Google Scholar] [CrossRef]
  17. Li, H.; Zhu, L.; Gong, H.; Sun, H.; Yu, J. Land subsidence modelling using a long short-term memory algorithm based on time-series datasets. Proc. Int. Assoc. Hydrol. Sci. 2020, 382, 505–510. [Google Scholar] [CrossRef]
  18. Li, H.; Zhu, L.; Guo, G.; Zhang, Y.; Dai, Z.; Li, X.; Chang, L.; Teatini, P. Land subsidence due to groundwater pumping: Hazard probability assessment through the combination of Bayesian model and fuzzy set theory. Nat. Hazards Earth Syst. Sci. 2021, 21, 823–835. [Google Scholar] [CrossRef]
  19. Zhu, L.; Gong, H.; Li, X.; Li, Y.; Su, X.; Guo, G. Comprehensive analysis and artificial intelligent simulation of land subsidence of Beijing, China. Chin. Geogr. Sci. 2013, 23, 237–248. [Google Scholar] [CrossRef]
  20. Kumar, S.; Kumar, D.; Donta, P.K.; Amgoth, T. Land subsidence prediction using recurrent neural networks. Stoch. Environ. Res. Risk Assess. 2022, 36, 373–388. [Google Scholar] [CrossRef]
  21. Zhou, D.; Zuo, X.; Zhao, Z. Constructing a large-scale urban land subsidence prediction method based on neural network algorithm from the perspective of multiple factors. Remote Sens. 2022, 14, 1803. [Google Scholar] [CrossRef]
  22. Yuan, Y.; Zhang, D.; Cui, J.; Zeng, T.; Zhang, G.; Zhou, W.; Wang, J.; Chen, F.; Guo, J.; Chen, Z. Land subsidence prediction in Zhengzhou’s main urban area using the GTWR and LSTM models combined with the Attention Mechanism. Sci. Total Environ. 2024, 907, 167482. [Google Scholar] [CrossRef] [PubMed]
  23. Arabameri, A.; Saha, S.; Roy, J.; Tiefenbacher, J.P.; Cerda, A.; Biggs, T.; Pradhan, B.; Ngo, P.T.T.; Collins, A.L. A novel ensemble computational intelligence approach for the spatial prediction of land subsidence susceptibility. Sci. Total Environ. 2020, 726, 138595. [Google Scholar] [CrossRef]
  24. Li, F.; Liu, G.; Tao, Q.; Zhai, M. Land subsidence prediction model based on its influencing factors and machine learning methods. Nat. Hazards 2023, 116, 3015–3041. [Google Scholar] [CrossRef]
  25. Zhao, D.; Sun, Z.; He, Y.; Chen, X.; Liu, R. Prediction of ground subsidence by shield tunneling using ensemble learning. Tunn. Undergr. Space Technol. 2023, 141, 105343. [Google Scholar] [CrossRef]
  26. Karimai, K.; Liu, W.; Maruyama, Y. Prediction and Factor Analysis of Liquefaction Ground Subsidence Based on Machine-Learning Techniques. Appl. Sci. 2024, 14, 2713. [Google Scholar] [CrossRef]
  27. Shi, L.; Gong, H.; Chen, B.; Zhou, C. Land subsidence prediction induced by multiple factors using machine learning method. Remote Sens. 2020, 12, 4044. [Google Scholar] [CrossRef]
  28. Fang, Z.; Wang, Y.; Peng, L.; Hong, H. A comparative study of heterogeneous ensemble-learning techniques for landslide susceptibility mapping. Int. J. Geogr. Inf. Sci. 2021, 35, 321–347. [Google Scholar] [CrossRef]
  29. Banerjee, A.; Dhillon, I.; Ghosh, J.; Merugu, S.; Modha, D.S. A generalized maximum entropy approach to bregman co-clustering and matrix approximation. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 509–514. [Google Scholar]
  30. Guo, D. Regionalization with dynamically constrained agglomerative clustering and partitioning (REDCAP). Int. J. Geogr. Inf. Sci. 2008, 22, 801–823. [Google Scholar] [CrossRef]
  31. Wu, X.; Zurita-Milla, R.; Kraak, M.-J. Co-clustering geo-referenced time series: Exploring spatio-temporal patterns in Dutch temperature data. Int. J. Geogr. Inf. Sci. 2015, 29, 624–642. [Google Scholar] [CrossRef]
  32. Rousseeuw, P.J. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 1987, 20, 53–65. [Google Scholar] [CrossRef]
  33. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2014; Volume 27. [Google Scholar]
  34. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  35. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. arXiv 2017, arXiv:1707.01926. [Google Scholar]
  36. Zheng, C.; Fan, X.; Wang, C.; Qi, J. Gman: A graph multi-attention network for traffic prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 1234–1241. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  38. Töscher, A.; Jahrer, M.; Bell, R.M. The bigchaos solution to the netflix grand prize. Netflix Prize Documentation 2009. pp. 1–52. Available online: https://www.researchgate.net/profile/Michael-Jahrer/publication/223460749_The_BigChaos_Solution_to_the_Netflix_Grand_Prize/links/0fcfd5108322324b0d000000/The-BigChaos-Solution-to-the-Netflix-Grand-Prize.pdf (accessed on 15 May 2024).
  39. Wolpert, D.H. Stacked generalization. Neural Netw. 1992, 5, 241–259. [Google Scholar] [CrossRef]
  40. Wang, J.; Zuo, R. Model averaging for identification of geochemical anomalies linked to mineralization. Ore Geol. Rev. 2022, 146, 104955. [Google Scholar] [CrossRef]
  41. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining Practical Machine Learning Tools and Techniques; Elsevier: Amsterdam, The Netherlands, 2017. [Google Scholar]
  42. Zhang, Y.; Liu, B.; Wu, H.; Cheng, X.; Kang, Y. Ground subsidence in Xiong’an new area from 2012 to 2016 monitored by InSAR technique. J. Earth Sci. Environ. 2018, 40, 452–462. [Google Scholar]
  43. Liu, Q.; Li, Y.; Yang, J.; Deng, M.; Li, J.; An, K. Physics-guided spatio–temporal neural network for predicting dissolved oxygen concentration in rivers. Int. J. Geogr. Inf. Sci. 2024, 38, 1207–1231. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Applsci 14 08330 g001
Figure 2. Flowchart of Blending ensemble algorithm.
Figure 2. Flowchart of Blending ensemble algorithm.
Applsci 14 08330 g002
Figure 3. Spatial distribution of land subsidence data of (a) timestamp 10, (b) timestamp 20, (c) timestamp 30, (d) timestamp 40, (e) timestamp 50, (f) timestamp 60, (g) timestamp 70.
Figure 3. Spatial distribution of land subsidence data of (a) timestamp 10, (b) timestamp 20, (c) timestamp 30, (d) timestamp 40, (e) timestamp 50, (f) timestamp 60, (g) timestamp 70.
Applsci 14 08330 g003
Figure 4. Silhouette coefficient of REDCAP algorithm clustering results.
Figure 4. Silhouette coefficient of REDCAP algorithm clustering results.
Applsci 14 08330 g004
Figure 5. Spatio-temporal clustering result matrix.
Figure 5. Spatio-temporal clustering result matrix.
Applsci 14 08330 g005
Figure 6. Spatial distribution patterns: (a) spatial pattern 1; (b) spatial pattern 2; (c) spatial pattern 3.
Figure 6. Spatial distribution patterns: (a) spatial pattern 1; (b) spatial pattern 2; (c) spatial pattern 3.
Applsci 14 08330 g006
Figure 7. Six types of time patterns: (a) Time pattern 1; (b) Time pattern 2; (c) Time pattern 3; (d) Time pattern 4; (e) Time pattern 5; (f) Time pattern 6.
Figure 7. Six types of time patterns: (a) Time pattern 1; (b) Time pattern 2; (c) Time pattern 3; (d) Time pattern 4; (e) Time pattern 5; (f) Time pattern 6.
Applsci 14 08330 g007aApplsci 14 08330 g007b
Figure 8. Time pattern distribution of land subsidence dataset.
Figure 8. Time pattern distribution of land subsidence dataset.
Applsci 14 08330 g008
Figure 9. Comparison of average time series prediction results for each site in cluster 1, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Figure 9. Comparison of average time series prediction results for each site in cluster 1, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Applsci 14 08330 g009aApplsci 14 08330 g009b
Figure 10. Comparison of average time series prediction results for each site in cluster 2, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Figure 10. Comparison of average time series prediction results for each site in cluster 2, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Applsci 14 08330 g010aApplsci 14 08330 g010b
Figure 11. Comparison of average time series prediction results for each site in cluster 3, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Figure 11. Comparison of average time series prediction results for each site in cluster 3, obtained by the ensemble model and the base model: (a) Seq2Seq; (b) GCN-Seq2Seq; (c) DCRNN; (d) GMAN; (e) Ensemble model.
Applsci 14 08330 g011aApplsci 14 08330 g011b
Figure 12. The peak monitoring points of the average time series of each cluster: (a) peak monitoring points of cluster 1; (b) peak monitoring points of cluster 2; (c) peak monitoring points of cluster 3.
Figure 12. The peak monitoring points of the average time series of each cluster: (a) peak monitoring points of cluster 1; (b) peak monitoring points of cluster 2; (c) peak monitoring points of cluster 3.
Applsci 14 08330 g012
Table 1. Parameters of the base models.
Table 1. Parameters of the base models.
ModelHyper Parameters
Seq2Seqbatch size = 64, epochs = 100, hidden units = 16, learning rate = 0.001, teacher forcing ratio = 0.2
GCN-Seq2Seqbatch size = 64, epochs = 100, hidden units = 16, learning rate = 0.001, teacher forcing ratio = 0.2
DCRNNbatch size = 64, epochs = 100, learning rate = 2.0 × 10−6
GMANbatch size = 64, epochs = 100, learning rate = 0.005
Table 2. Statistics of spatio-temporal clustering matrix of land subsidence data.
Table 2. Statistics of spatio-temporal clustering matrix of land subsidence data.
Cluster IDPatternMean (mm)Proportion
1slight uplift1543.28%
2slight subsidence−1742.42%
3moderate subsidence−6511.19%
4severe subsidence−1243.11%
Table 3. The proportion of each type of spatio-temporal subsidence pattern and the number of time periods they occupy in each spatial pattern.
Table 3. The proportion of each type of spatio-temporal subsidence pattern and the number of time periods they occupy in each spatial pattern.
Spatial PatternSlight UpliftSlight SubsidenceModerate SubsidenceSevere SubsidenceTime Period
Spatial pattern 133.78%66.10%0.1%027
Spatial pattern 241.20%37.75%18.33%2.72%29
Spatial pattern 356.47%21.40%15.04%7.08%24
Table 4. The proportion of each time pattern in the study area.
Table 4. The proportion of each time pattern in the study area.
Time PatternTrendProportion
Time pattern 1Stable slight uplift33.78%
Time pattern 2Stable slight subsidence21.40%
Time pattern 3Stable slight subsidence–uplift22.69%
Time pattern 4Slight subsidence–moderate subsidence15.03%
Time pattern 5Slight subsidence–moderate subsidence–severe subsidence2.61%
Time pattern 6Moderate subsidence–severe subsidence4.46%
Table 5. Spatio-temporal cluster information of three representative types of subsidence patterns.
Table 5. Spatio-temporal cluster information of three representative types of subsidence patterns.
Cluster IDPatternMeanTimestamp RangeData Matrix Size
1Slight subsidence−21Z27–Z5526,508 × 29
2Moderate subsidence−64Z27–Z556499 × 29
3Moderate subsidence−144Z56–Z791602 × 24
Table 6. Prediction accuracy of ensemble model and base models.
Table 6. Prediction accuracy of ensemble model and base models.
ModelMetricCluster 1Cluster 2Cluster 3
Seq2SeqRMSE3.56804.65096.9436
MAE2.86083.73895.1175
R20.95200.95910.9676
GCN-Seq2SeqRMSE2.80944.76647.3231
MAE2.10403.65686.2494
R20.97420.95700.9640
DCRNNRMSE2.99394.53495.7840
MAE2.16693.30073.6702
R20.96620.96100.9775
GMANRMSE3.12744.21754.6243
MAE2.48883.39093.8729
R20.96270.96620.9857
EnsembleRMSE1.39881.46960.9597
MAE1.03551.22480.6788
R20.99250.99590.9994
Table 7. Statistical comparison of prediction accuracy of different models (cluster 1).
Table 7. Statistical comparison of prediction accuracy of different models (cluster 1).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq NoNoNo
GCN-Seq2Seq YesYes
DCRNN Yes
Table 8. Statistical comparison of prediction accuracy of different models (cluster 2).
Table 8. Statistical comparison of prediction accuracy of different models (cluster 2).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq NoNoNo
GCN-Seq2Seq NoNo
DCRNN No
Table 9. Statistical comparison of prediction accuracy of different models (cluster 3).
Table 9. Statistical comparison of prediction accuracy of different models (cluster 3).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq YesNoNo
GCN-Seq2Seq NoNo
DCRNN No
Table 10. Prediction accuracy of the ensemble model and base models at peak monitoring points.
Table 10. Prediction accuracy of the ensemble model and base models at peak monitoring points.
ModelMetricCluster 1Cluster 2Cluster 3
Seq2SeqRMSE3.72654.64536.5843
MAE3.01593.80155.0155
R20.94940.95750.9659
GCN-Seq2SeqRMSE2.68634.7767.3002
MAE2.00183.65226.1603
R20.97370.95500.9624
DCRNNRMSE3.09554.71045.3870
MAE2.27443.44913.7927
R20.96510.95630.9796
GMANRMSE3.39132.72534.4028
MAE2.67072.11963.7022
R20.95810.98540.9864
EnsembleRMSE1.51671.61091.3213
MAE1.16971.32890.9396
R20.99160.99490.9988
Table 11. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 1).
Table 11. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 1).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq NoNoNo
GCN-Seq2Seq YesYes
DCRNN Yes
Table 12. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 2).
Table 12. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 2).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq NoNoNo
GCN-Seq2Seq NoNo
DCRNN No
Table 13. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 3).
Table 13. Statistical comparison of prediction accuracy of different models at peak monitoring points (cluster 3).
ModelSeq2SeqGCN-Seq2SeqDCRNNGMAN
EnsembleYesYesYesYes
Seq2Seq YesNoNo
GCN-Seq2Seq NoNo
DCRNN No
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, B.; Wu, G.; Li, J.; Wu, Q.; Deng, M. Spatio-Temporal Heterogeneous Ensemble Learning Method for Predicting Land Subsidence. Appl. Sci. 2024, 14, 8330. https://doi.org/10.3390/app14188330

AMA Style

Zhao B, Wu G, Li J, Wu Q, Deng M. Spatio-Temporal Heterogeneous Ensemble Learning Method for Predicting Land Subsidence. Applied Sciences. 2024; 14(18):8330. https://doi.org/10.3390/app14188330

Chicago/Turabian Style

Zhao, Bin, Gusheng Wu, Junjie Li, Qianhong Wu, and Min Deng. 2024. "Spatio-Temporal Heterogeneous Ensemble Learning Method for Predicting Land Subsidence" Applied Sciences 14, no. 18: 8330. https://doi.org/10.3390/app14188330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop