Next Article in Journal
Artificial Intelligence Integration in Sustainable Business Practices: A Text Mining Analysis of USA Firms
Previous Article in Journal
Determinants of Farmer Participation and Development of Shallot Farming in Search of Effective Farm Management Practices: Evidence Grounded in Structural Equation Modeling Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Slope Safety Factor Based on Attention Mechanism-Enhanced CNN-GRU

by
Qi Da
,
Ying Chen
*,
Bing Dai
,
Danli Li
and
Longqiang Fan
School of Resource Environment and Safety Engineering, University of South China, Hengyang 421001, China
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(15), 6333; https://doi.org/10.3390/su16156333
Submission received: 9 June 2024 / Revised: 14 July 2024 / Accepted: 22 July 2024 / Published: 24 July 2024

Abstract

:
This paper proposes a new method for predicting slope safety factors that combines convolutional neural networks (CNNs), gated recurrent units (GRUs), and attention mechanisms. This method can better capture long-term dependencies, enhance the ability to model sequential data, and reduce the dependence on noisy data, thereby reducing the risk of overfitting. The goal is to improve the accuracy of slope safety factor prediction, detect potential slope stability issues in a timely manner, and take corresponding preventive and control measures to ensure the long-term stability and safety of infrastructure and promote sustainable development. The Pearson correlation coefficient is used to analyze the relationship between the target safety factor and the collected parameters. A one-dimensional CNN layer is used to extract high-dimensional features from the input data, and then a GRU layer is used to capture the correlation between parameters in the sequence. Finally, an attention mechanism is introduced to optimize the weights of the GRU output, enhance the influence of key information, and optimize the overall prediction model. The performance of the proposed model is evaluated using metrics such as the mean absolute error (MAE), mean absolute percentage error (MAPE), mean squared error (MSE), root-mean-square error (RMSE), and R2. The results show that the CNN-GRU-SE model outperforms the GRU, CNN, and CNN-GRU models in terms of prediction accuracy for slope safety factors, with improvements of 4%, 2%, and 1%, respectively. Overall, the research in this paper makes valuable contributions to the field of slope safety factor prediction, and the proposed method also has the potential to be extended to other time-series prediction fields, providing support for a wide range of engineering applications and further promoting the realization of sustainable development.

1. Introduction

As the Chinese economy develops and material demand increases, the number of slope projects in China will continue to rise. The frequent occurrence of slope instability and damage poses significant potential safety hazards to workers and also causes serious economic losses [1]. Consequently, the establishment of a secure, dependable, and efficacious slope stability prediction model represents a pivotal challenge in geotechnical engineering. Nevertheless, the unpredictability of geological and geotechnical properties renders slope instability and failure a profoundly intricate geological process [2].
In recent years, researchers have conducted extensive research on slope stability prediction through theoretical calculation [3], numerical simulation [4], machine learning [5], and other methods, with notable outcomes. Both theoretical calculations and numerical simulations can be employed to consider the stress exerted on a slope, analyze deformation and stability, and elucidate the failure mode of a slope. Nevertheless, theoretical calculations cannot take into account the constitutive relationship of soil, especially in the case of multi-layer soil. Consequently, in slope stability analysis, this method is unable to accurately reveal the requisite safety and reliability. However, a numerical model necessitates a substantial amount of engineering geological investigation at the outset, which is costly and has limitations.
The advent of big data and machine learning has led to the widespread adoption of machine learning-based slope stability prediction techniques. M. G. Sakellariou et al. [6] applied an Artificial Neural Network (ANN) model to the analysis of slope stability, achieving preliminary results. Subsequently, H.B. Wang et al. [7] evaluated the data of the Yudonghe River landslide in Hubei Province of China through a BPNN and concluded that the landslide was in a critical stable state. Behrouz Gordan et al. [8] combined an ANN and PSO (Particle Swarm Optimization) to optimize the hyperparameters of the ANN model, thereby enhancing the model’s capacity to predict slope stability during an earthquake. Li et al. [9] studied the application of an RBFNN model in the prediction of open-pit slope stability and compared the optimization effects of BSO, GA, and MLP on it. Arsalan Mahmoodzadeh et al. [10] applied six kinds of machine learning algorithms to the analysis and prediction of slope stability. Huang et al. [11] proposed the application of a deep learning model, LSTM, to solve the problem of slope stability prediction. Table 1 provides incomplete statistics on the application of machine learning algorithms in slope stability prediction.
In other aspects, Tang et al. [12] performed a prediction study of ionospheric TEC based on a CNN-LSTM–attention mechanism and found that the model could maintain stability during different months and under different geomagnetic conditions. Liu et al. [13] used the earthworm algorithm to optimize support vector regression for predicting reservoir slopes. The experimental results showed that the model could accurately predict the displacement of reservoir landslides. Ma et al. [14] proposed a backend model for automated machine learning (AutoML), which is easy for field personnel to use and meets practical needs.
The limitations of traditional machine learning models include their inability to accurately capture complex patterns and relationships in data, their lack of capacity to deal with noise or missing data, and their inability to adapt to changes in the operating environment. In comparison with traditional machine learning, deep learning has a greater number of network layers and is better able to express the features of objects through the automatic extraction of feature information [11]. Consequently, an increasing number of scholars are utilizing deep learning techniques to address the limitations of traditional machine learning models. LSTM incorporates three gate functions: the input gate, forgetting gate, and output gate, which regulate the input value, memory value, and output value, respectively. However, it is challenging to train with numerous parameters. In the GRU (gated recurrent unit) model, there are only two gates: the update gate and the reset gate. When all hyperparameters are tuned, the performance of the two gates is comparable. The GRU structure is simpler, the number of training samples is smaller, and it is easier to implement. The gated recurrent unit network (GRU), as a recurrent neural network model, may encounter difficulties such as gradient disappearance, gradient explosion, and the loss of important information when processing ultra-long sequences. In order to address these issues, a convolutional neural network is employed to preprocess data and filter out irrelevant information [15]. Concurrently, the SE (Squeeze-and-Excitation) attention mechanism is incorporated into a CNN-GRU model, enabling it to dynamically adjust the importance of features, focus on key features, and enhance the performance of sequence data processing. The model is capable of more effectively capturing long-term dependencies, enhancing its ability to model sequence data, reducing its dependence on noisy data, and reducing the risk of overfitting [16].
The prediction of the slope safety factor is closely related to sustainable development. This study aims to promote the development of the slope stability prediction field and explore cutting-edge technologies to improve prediction accuracy. The article proposes a new method that combines an attention mechanism and CNN-GRU to enhance the predictive ability of the model. By helping the model focus on the most relevant input data features, attention mechanisms potentially enhance the model’s robustness to noise and data. This model is capable of capturing complex relationships and patterns in data and is applicable to a wide range of slope systems. The accurate prediction of slope stability can provide a foundation for engineering design, ensure construction safety, reduce geological hazard risks, and ensure the safety of life and property. This study makes valuable contributions to the accuracy and effectiveness of slope stability prediction and analysis by integrating advanced technologies and filling knowledge gaps.
Table 1. Machine learning methods applied to slope stability prediction.
Table 1. Machine learning methods applied to slope stability prediction.
ReferenceInput ParametersYearAlgorithm/MethodsData Number
[6]H, c, γ , β , φ , r u 2004ANN46
[7]H, c, γ , β , φ 2005BPNN27
[17]H, c, γ , β , φ , r u 2008SVM46
[18]H, c, γ , β , φ , r u 2011ANN46
[19]H, c, γ , β , φ 2013ANN675
[20]H, c, γ , β , φ , r u 2014ELM97
[21]H, c, γ , β , φ , r u 2015FA-LS-SVC168
[22] X 1 7 2015GA-BP-ANN120
[8]H, c, PGA, β , φ 2016PSO-ANN699
[23]H, c, γ , β , φ , r u 2016FNS, MARS, MGGP103
[24]c, β , φ , pp2016ANN100
[25]H, c, γ , β , φ , r u 2017NBC69
[26]H, c, γ , β , φ , r u 2017PSO-ANN83
[27]H, c, γ , β , φ , r u 2017PSO-LSSVM46
[28]H, c, γ , β , φ , r u 2018LR, DT, RF, GBM, SVM, MLPNN168
[29]H, c, γ , β , φ , r u 2018GSA, RF, SVM, Bayes107
[30]H, c, γ , β , φ , r u 2018GPC, QDA, SVM, ANN, ADB-DT, KNN168
[31]H, c, γ , β , φ , r u 2019GBM221
[32] c u , β , w, b/B2019MLP, GPR, MLR, SLR, SVR630
[33]w, c, γ , β , φ 2019HHO-ANN75
[34]H, c, γ , β , φ 2020M5Rules–GA450
[10]H, c, γ , β , φ , r u 2021GRP, SVM, DT, LSTM, DNN, KNN327
[9]H, c, γ , β , φ 2021BSO-RBFNN, GA-RBFNN, MLP-RBBBFNN495
[35] F 1 12 2022RF-XGBoost786
[36]H, c, β , φ , PGA2022DT, RF, AdaBoost700
[11]H, c, γ , β , φ 2023LSTM2640
[37]H, c, γ , β , φ , r u 2023DeepBoost444
[1]H, c, γ , β , φ 2023SVM, LR, DT, RF, KNN, NB, LDA77
[5]H, c, γ , β , φ , r u 2023SVM, RF, KNN, DT, GB117
Here, X 1 is the elastic modulus, X 2 is the rock mass classification, X 3 is the installation height of the instrument, X 4 is the excavation height of the slope, X 5 is the measurement start time, X 6 is the measurement time period, and X 7 is the actual excavation height after measurement; PGA is the peak ground acceleration; b/B is the retracement distance ratio; F 1 is the elevation of the front edge, F 2 is the elevation of the back edge, F 3 is the slope height, F 4 is the slope angle, F 5 is the lithological property, F 6 is the inclination angle, F 7 is the dip direction, F 8 is the structure type, F 9 is the plane morphology, F 10 is the profile shape, F 11 is the landslide volume, and F 12 is the influence degree of human activities.

2. Obtaining, Analyzing, and Processing Data

2.1. Factors Affecting Slope Stability

The selection of samples determines the upper limit of the model’s predictive ability. Given the intricate mechanism of slope stability and the multitude of influencing factors, scholars at home and abroad generally concur that slope instability is contingent upon the slope shape, rock and soil mass characteristics, and external influencing factors in machine learning or comprehensive evaluation methods. Among these factors, the most significant are the slope height and slope angle. The influence of rock and soil mass characteristics is also significant, with the bulk density, cohesion, internal friction angle, and pore pressure ratio of rock and soil masses being particularly relevant [38]. Figure 1 illustrates the factors that influence slope stability. The external factors mentioned, such as earthquakes or other human factors, were not considered in this study. The Factor of Safety (FOS) is a comprehensive index for evaluating slope stability. In the context of slope stability analysis, the FOS can be defined as the ratio of the slope sliding resistance to the slope sliding force. This ratio is directly related to the shear strength of the soil, as outlined in reference [39].

2.2. Establishment of Database

A total of 183 sets of slope stability samples were obtained from published public materials at home and abroad, with no duplicates or absences; see Appendix A [6,40,41]. All of the samples contain the six factors (H, α , c , φ , γ , and r u ) affecting slope stability that were previously discussed. In order to more effectively illustrate the distribution law and range of features, the slope stability prediction index system was analyzed in the form of a box diagram, a scatter diagram, and a half-violin diagram, as shown in Figure 2. In comparison to the conventional single violin or box plot, this integrated approach enables a more comprehensive illustration of the central tendency, dispersion, distribution density, and outliers of the data, thereby facilitating a more nuanced understanding of the data characteristics. At the same time, the Pearson correlation coefficient was employed to assess the six influencing factors. The resulting Pearson correlation coefficient and correlation scatter distribution diagram are presented in Figure 3. According to the literature, a Pearson correlation coefficient value between 0.4 and 0.6 is indicative of a moderate correlation between two factors. A correlation coefficient of 0.0 to 0.2 is indicative of a very weak or non-existent correlation. The analysis of the chart allows for a clear understanding of the degree of correlation between various factors. The observation of the scatter chart also permits the observation of a possible nonlinear relationship between factors. Consequently, in order to enhance the precision of the model, it is advisable to standardize and scale the data, with the objective of ensuring that they fall within the range of 0 to 1.

3. Establishment of Model

The CNN is employed to extract the overarching characteristics of the slope stability prediction model, whereas the GRU is utilized to discern the interrelationships between disparate sequences. The attention mechanism module is designed to extract key information from the GRU output, thereby assisting models in making more efficient use of the data.

3.1. CNN Model

The primary objective of a CNN is to identify and extract key features from the input data [42]. A typical convolutional neural network (CNN) architecture comprises multiple layers, including a convolutional layer, a pooling layer, a dropout layer, and a fully connected layer. In the process of feature extraction, the convolutional layer plays a pivotal role in capturing task-related feature information through the operation of a convolutional filter on input data [43]. As the number of convolutional cores increases, the abstraction level of extracted features will gradually increase, which will facilitate a more comprehensive understanding and analysis of the internal structure and underlying laws of the data, thereby enhancing the performance and generalization ability of the model [44].

3.2. GRU

A recurrent neural network (RNN) is a neural network architecture that has been specifically designed for the processing of sequential data. It can be scaled in order to efficiently utilize historical information. The Long Short-Term Memory Network (LSTM), an enhanced variant of the recurrent neural network (RNN), is designed to address the gradient disappearance issue and enhance the stability of the model [45]. The gated cycle unit (GRU) employs the gated mechanism to streamline the LSTM [46] process, as illustrated in Figure 4. Nevertheless, GRUs may be prone to the loss of crucial information when confronted with exceedingly lengthy sequences. To address this issue, convolutional neural networks can be employed to preprocess data and filter out irrelevant information, thereby enhancing the accuracy of slope stability prediction. Consequently, the gated structure of the GRU network, when combined with a convolutional neural network, represents an effective method for the processing of long-term dependencies and sequential data.

3.3. Squeeze-and-Excitation Attention

Attention mechanisms can extract valuable information from features and focus on local information without increasing computational complexity and are widely used to enhance the performance of network models. This approach does not increase the computational complexity, which has led to its widespread use in enhancing the performance of network models. In order to enhance the efficacy of the model, this study introduces Squeeze-and-Excitation attention, which facilitates the modeling of relationships between channels, thereby enhancing the capacity to extract efficacious features from constrained data. The enhancement of the response to important features and the suppression of the response to minor features result in the effective recalibration of features [47]. This approach to feature recalibration enables the network to focus more on useful information. Figure 5 depicts the structure of the attention mechanism, which is divided into four steps.
Step 1: Mapping. The essence of the implementation is convolution.
U = F t r ( X )
Here, X ϵ R H × W × C , U ϵ R H × W × C , H and H represent the heights of X and U , W   a n d   W denote the widths, and C and C indicate the number of channels [48].
Step 2: Compression. Global Average Pooling (GAP) is used to compress the height and width of each channel to 1, and the final dimension is 1 × 1 × C .
Z = F s q U = 1 h · w j = 1 w U ( i , j )
Step 3: Excitation. The Z obtained in step 2 is passed through two fully connected layers to obtain the weight value.
S = F e x Z = σ ( w 2 δ ( w 1 z ) )
where S is the generating weight, w 1 and w 2 represent two fully connected operations, δ is ReLU, and σ is sigmoid. The two fully connected layers of this step first reduce the dimension of Z and then increase the dimension, and the generalization ability of the model is enhanced [49].
Step 4: Dot product. The results of step 3 and step 1 are calculated from the dot product of the corresponding channel through the following formula:
X ~ = F s c a l c U , S = S × U

3.4. Model Frame

Figure 6 depicts the slope stability prediction model developed in this research, which is based on the CNN-GRU-SE method. This model comprises an input layer, a convolutional neural network (CNN), a gated recurrent unit (GRU), an attention layer, and an output layer. The input to the convolutional neural network (CNN) layer is the historical slope data, which are increased by the convolution operation, and the number of parameters is compressed. Furthermore, the feature dimension is reduced by a pooling process. The fully connected layer then converts the feature into a one-dimensional structure, thus completing the feature vector extraction. Concurrently, dropout layers are incorporated to offset the consequences of overfitting. The GRU and the attention layer learn the internal change rule from the extracted features, thereby enabling the prediction of the slope FOS. Furthermore, they fully extract relationships from historical data, expand important information about the acceptance domain, and reduce the feature dimension. The dynamic changes in CNN features are modeled and learned by the GRU network in order to extract the correlation between multiple features. The attention layer, in turn, employs the attention mechanism to assign different hidden-state probability weights to the GRU, thereby focusing on important information related to slope stability.
The proposed safety factor prediction model based on CNN-LSTM was optimized and implemented in Python 3.11.0 using the TensorFlow framework. The experimental hardware includes an Intel (R) Core(TM) i7-13700KF CPU and 32 GB RAM. The Adam optimization algorithm updates the network parameters with an initial learning rate of 0.01. The neural network training parameters are set to a maximum iteration count of 625 and a batch size of 16. Table 2 provides the specific structural parameters of the model.

4. Analysis of FOS Prediction Results of Slope

The training set (80%) and the test set (20%) should be split. In order to process inputs of different dimensions and sizes, it is necessary to normalize the sample data, as demonstrated in Equation (5):
X i j = x i j m i n ( x i j ) m a x ( x i j ) m i n ( x i j ) Y j = y j m i n ( y j ) m a x ( y j ) m i n ( y j )
where X i j is the JTH input sample value of the ith attribute after normalization, and x i j is the jth input sample value of the ith attribute. Y j is the jth output sample value after normalization, and y j is the jth output sample value.
In order to test the validity of the model and evaluate the prediction effect, a number of evaluation indexes were selected, including the mean absolute error (MAE), mean absolute percentage error (MAPE), mean square error (MSE), root-mean-square error (RMSE), and R2. The following is a description of the calculation formulae for each evaluation index:
M A E = 1 n i = 1 n y i ^ y i
M A P E = 100 % n i = 1 n y i ^ y i y i
M S E = 1 m i = 1 m ( y i y i ^ ) 2
R 2 = 1 i ( y i ^ y i ) 2 i ( y i ¯ y i ) 2
where y i is the actual value, y i ^ is the predicted value, and n is the number of data.

4.1. Performance Evaluation

This paper presents four models (CNN, GRU, CNN-GRU, CNN-GRU-SE) for the purpose of predicting the FOS of a slope to ascertain the superiority and prediction performance of the proposed model. The selection of the loss function and RMS error curves enables the assessment of the model’s performance during training. The change process is illustrated in Figure 7 and Figure 8. Figure 9 shows the prediction results of each model.
Figure 7 illustrates that the CNN model’s loss function value declines rapidly in the initial stage and then gradually stabilizes. As the number of iterations increases, the fluctuation of the loss value gradually decreases, indicating that the model becomes increasingly stable during the training process. The loss function value of the GRU model exhibits a relatively high degree of fluctuation following a period of decline, which may suggest that the model’s stability during training is inferior to that of the CNN model. The CNN-GRU model exhibits a similar decreasing trend in the loss function value to those of the CNN and CNN-GRU-SE models, although the amplitude of fluctuations during the iterative process is between those of the CNN and GRU models. The decline in the loss function value of the CNN-GRU-SE model is analogous to that observed in the CNN model. The loss function value is initially high and then decreases rapidly as the number of iterations increases, indicating that the model is learning and gradually improving its parameters to minimize the loss function. After approximately 500 iterations, the loss values begin to stabilize and fluctuate in a lower range, which may indicate that the model is approaching convergence and that further iterations will not significantly improve model performance. The overall loss function value of the CNN-GRU-SE model is lower than those of other models, indicating that it performs better than these models in predicting the FOS of the slope.
Figure 8 illustrates that in the initial stage of the CNN model, the error rate declines rapidly and then gradually stabilizes. Overall, although the error exhibited some fluctuations, it remained within the range of 0.5 to 1.5, and there were no discernible indications of overfitting or underfitting. In the initial iteration of the GRU model, the error exhibited a rapid decline, reaching a value of approximately 0.5, and remained at a low level throughout the training process. In comparison to the CNN model, the error fluctuations of the GRU model are more stable, although there are still small-amplitude fluctuations. The CNN-GRU model exhibits a pronounced decline in error at the outset of the iteration, accompanied by a subsequent pronounced fluctuation, particularly following the completion of 1000 iterations. The overall error level is comparable to that of the GRU model but exhibits greater volatility, which may indicate that the model adapts to the training data in a distinct manner. The CNN-GRU-SE model exhibits the most stable error variation during training. The error rapidly declines below 0.5 and remains at a low level throughout the training period. This suggests that the CNN-GRU-SE model may exhibit superior generalization and robustness.

4.2. Performance Comparison of Different Models

To further verify the effectiveness and applicability of the proposed model for predicting the FOSs of slopes, 37 test sets were used to compare the CNN-GRU-SE model with three widely used regression prediction models (ANN, DT, RF). MAE, MAPE, MSE, RMSE, and R2 were used as evaluation metrics, as shown in Figure 10.
The CNN-GRU-SE slope FOS prediction model presented in this paper demonstrated high accuracy through a comprehensive analysis of the comparison graphs of various indexes. In comparison to the GRU model, the CNN-GRU-SE model demonstrated a higher level of accuracy, with an RMSE measurement of 0.09 in the prediction of the FOS of the slope. With regard to the mean absolute percentage error (MAPE) measure, the CNN-GRU-SE model yielded an increase in the FOS of 0.04 in comparison to the GRU prediction. Furthermore, the RMSE of the CNN-GRU-SE model for the FOS was found to be 0.04 greater than that of the CNN-GRU model. The mean absolute error (MAE) measurements demonstrated that the CNN-GRU-SE model exhibited an increase in the FOS of 0.06 compared to the GRU prediction and 0.03 compared to the CNN-GRU prediction. For R2, the CNN-GRU-SE model demonstrated increases of 0.22 and 0.09, respectively, in comparison to the GRU and CNN-GRU models. Conversely, when the CNN-GRU-SE model is compared with the other three conventional regression models, it becomes evident that the CNN-GRU-SE model outperforms the conventional models in terms of the five aforementioned indicators. The comparison results demonstrate that the CNN-GRU-SE hybrid model exhibits superior prediction accuracy and stability in practical FOS prediction applications.
Figure 11 presents the error bar plots for the distinct models. It can be observed from the figure that the prediction effect of the CNN, GRU, DT, RF, and BP models applied independently is not satisfactory. This is because these models lack effective combinatorial optimization, which results in their prediction performance not being fully demonstrated. In contrast, the CNN-GRU-SE model demonstrates superior performance. With the exception of an outlier, the prediction results of this model are significantly superior to those of common regression models and are closer to the real values. Consequently, integrating the attention mechanism into the CNN-GRU model is a viable approach to enhancing the prediction precision.
The application of machine learning technology enables researchers to more fully and deeply mine the underlying information in a data set, thereby facilitating an understanding of the data, the discovery of the nature of the problem, and the identification of complex relationships. As engineering practice data sets continue to be enriched, it is anticipated that the accuracy and reliability of slope FOS prediction will be further enhanced. Researchers in the field of slope engineering can utilize the advantages of the attention mechanism to investigate complex phenomena associated with slope instability, with data serving as the primary focus.
To more intuitively display the contribution and relative importance of each feature to the model’s predictions, this paper introduces importance (Figure 12) and variable contribution (Figure 13) to quickly understand the ranking of feature importance.
Shap (Shapley Additive explanations) generates a predicted value for each sample model, and the Shap value is the numerical value assigned to each feature in the sample. Similar to the addition method of linear models, assuming the model’s baseline score (usually the mean of the target variable for all samples) is y b a s e , the ith sample is x i , the ith feature of the jth sample is x i , j , and the Shap value of this feature is f x i , j , then the model’s predicted value for sample x i is
y i = y b a s e + f x i , 1 + f x i , 2 + + f x i , j
When f x i , j > 0, the feature plays a positive role in predicting the target value; conversely, the feature has an opposite effect on the target prediction value. Therefore, Shap not only gives the magnitude of the influence of the feature but also reflects the positive and negative influences of the features in each sample.
Figure 12 takes the absolute value average of the SHAP values for each feature to obtain the distribution of feature importance, which is equivalent to blurring the positive and negative effects in the above figure. The vertical axis of Figure 13 ranks the features according to the sum of the SHAP values of all samples, and the horizontal axis is the SHAP value (the distribution of the impacts of features on the model output); each point represents a sample, and the sample size is accumulated vertically. The color represents the feature value (red corresponds to high values, and blue corresponds to low values).
As can be seen in Figure 12 and Figure 13, in the CNN model, the slope angle and unit weight have a significant impact on the model output, with a more dispersed distribution of SHAP values, and significant effects in both the positive and negative directions. The effects of cohesion and slope height are relatively small, with most SHAP values concentrated near zero. The effects of the angle of internal friction and the water ratio on the model output are moderate, with SHAP values distributed near zero, with a small positive and negative impact. In the GRU model, the unit weight and slope angle still have a significant impact on the model output, but the order of importance of the two has been reversed, and their impact has increased compared to the CNN model, with a wide distribution of SHAP values. The impacts of feature variables in the CNN-GRU model are different from those in the GRU model, with slightly increased effects of the angle of internal friction and the water ratio, and more SHAP values are distributed in both positive and negative directions. By combining the SE attention mechanism, the impact of the water ratio in the SNN-GRU-SE model is significantly increased, surpassing that of the angle of internal friction, which may be due to the attention mechanism improving the model’s ability to extract features.

5. Conclusions

This study proposes a novel method, the CNN-GRU-SE model, with the objective of enhancing the precision of the prediction of the slope’s FOS. This research method employs a combination of convolutional neural networks (CNNs), gated recurrent units (GRUs), and attention mechanisms to more effectively capture long-term dependencies, enhance the ability to model sequence data, reduce the dependence on noisy data, and reduce the risk of overfitting. The principal findings of this study are as follows:
  • The integration of an attention mechanism into the model has been demonstrated to significantly enhance the accuracy of weight allocation while simultaneously promoting rapid error convergence and a reduction in the error value. Concurrently, the performance and accuracy of the model are enhanced, and more favorable outcomes are achieved in the training and prediction processes of the model.
  • The CNN-GRU-SE model demonstrates superior performance in terms of accuracy, outperforming traditional deep learning models. This advantage is particularly evident in the FOS prediction accuracy of the slope, which significantly improves the accuracy and reliability of prediction results.
The findings indicate that the CNN-GRU-SE model exhibits superior accuracy and reliability in the prediction of the slope’s FOS, which is of paramount importance for the prevention of slope instability incidents. The introduction of an attention mechanism into the prediction of the slope’s FOS opens up new avenues for the application of artificial intelligence technology in this field. The prediction results of the CNN-GRU-SE model enable site workers to effectively prevent and control slope instability accidents, thereby enhancing the efficiency and safety of slope engineering. In conclusion, the CNN-GRU-SE model proposed in this study demonstrates considerable potential for enhancing the accuracy of slope safety factor prediction. The model combines convolutional neural networks (CNNs), gated recurrent units (GRUs), and an attention mechanism to construct a high-performance model with robust and generalizable capabilities. This study employs a quantitative approach to assess the factors influencing slope stability, identifying six evaluation indexes to construct a predictive model. This represents an initial attempt to predict the safety factor of a slope. Although it is challenging to convert qualitative factors into quantitative factors, factors that have a significant impact on slope stability, such as rainfall and existing joints, must be considered. In particular, a significant incident occurred on the Meida Expressway in Guangdong, China, which was primarily attributable to the increase in the proportion of pore water caused by rainfall, resulting in slope instability and, ultimately, pavement collapse. Consequently, it is imperative to reinforce the research and prediction of slope stability, enhance the precision and dependability of these predictions, and furnish more support and assurance for sustainable development. Concurrently, the selection of a more impartial and rational evaluation metric for the slope safety factor will emerge as a focal point and challenge of future research.

Author Contributions

Conceptualization, Q.D., B.D. and Y.C.; methodology, Q.D., D.L., and Y.C.; software, Q.D. and Y.C.; validation, Y.C.; investigation, L.F. and D.L.; resources, L.F. and D.L.; data curation, L.F. and D.L.; writing—original draft preparation, Q.D. and Y.C.; writing—review and editing, Q.D. and Y.C.; visualization, Q.D. and D.L.; supervision, Y.C. and B.D.; project administration, Y.C. and B.D. All authors have read and agreed to the published version of the manuscript.

Funding

This project was sponsored by the National Natural Science Foundation of China (No. 151374244), the Key Project of Education Department of Hunan Province (22A0293), and the Postgraduate Scientific Research Innovation Project of Hunan Province (QL20220213).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

NumberUnit Weight
γ / ( k N / m 3 )
Cohesion
c/kPa
Angle of Internal Friction
φ
Slope Angle
φ f
Slope Height
H/m
Water Ratio
r u
Slope Stability Factor
F
131.30 37.00 68.60 47.00 305.00 0.25 1.20
226.20 44.14 32.26 37.71 359.04 0.21 1.22
323.10 29.20 25.20 36.50 61.90 0.40 1.70
425.00 36.00 55.00 45.50 299.00 0.25 1.52
528.51 42.34 32.20 43.25 453.60 0.25 1.25
619.00 32.00 50.00 42.00 26.00 0.50 1.70
731.30 37.00 68.59 47.50 262.40 0.25 1.20
823.50 20.00 25.00 49.10 115.00 0.41 1.63
926.62 0.00 31.78 42.72 51.48 0.40 1.04
1027.00 35.00 37.50 37.80 320.00 0.29 1.24
1130.00 27.38 34.57 43.46 319.21 0.27 1.02
1220.00 30.00 35.00 40.00 25.00 0.29 1.90
1327.30 1.00 36.00 50.00 92.00 0.29 1.25
1431.30 68.00 37.00 49.00 200.50 0.32 1.20
1520.00 20.00 36.00 45.00 50.00 0.32 0.96
1627.00 40.00 35.00 47.10 292.00 0.32 1.15
1725.00 46.00 35.00 50.00 284.00 0.32 1.34
1831.30 68.00 37.00 46.00 366.00 0.32 1.20
1925.00 46.00 36.00 44.50 299.00 0.32 1.55
2027.30 10.00 39.00 40.00 480.00 0.32 1.45
2125.00 46.00 35.00 46.00 393.00 0.32 1.31
2225.00 48.00 40.00 49.00 330.00 0.32 1.49
2331.30 68.60 37.00 47.00 305.00 0.32 1.20
2422.40 10.00 35.00 45.00 10.00 0.40 0.90
2520.00 20.00 36.00 45.00 50.00 0.50 0.83
2620.00 0.10 36.00 45.00 50.00 0.25 0.79
2720.00 0.10 36.00 45.00 50.00 0.50 0.67
2822.00 0.00 40.00 33.00 8.00 0.35 1.45
2924.00 0.00 40.00 33.00 8.00 0.30 1.58
3020.00 0.00 24.50 20.00 8.00 0.35 1.37
3118.00 0.00 30.00 20.00 8.00 0.30 2.05
3227.00 40.00 35.00 43.00 420.00 0.25 1.15
3327.00 50.00 40.00 42.00 407.00 0.25 1.44
3427.00 35.00 35.00 42.00 359.00 0.25 1.27
3527.00 37.50 35.00 37.80 320.00 0.25 1.24
3627.00 32.00 33.00 42.60 301.00 0.25 1.16
3727.00 32.00 33.00 42.20 289.00 0.25 1.30
3827.30 14.00 31.00 41.00 110.00 0.25 1.25
3927.30 31.50 29.70 41.00 135.00 0.25 1.25
4027.30 16.80 28.00 50.00 90.50 0.25 1.25
4127.30 26.00 1.00 50.00 92.00 0.25 1.25
4227.30 10.00 39.00 41.00 511.00 0.25 1.47
4327.30 10.00 39.00 40.00 470.00 0.25 1.43
4425.00 46.00 35.00 47.00 443.00 0.25 1.28
4525.00 46.00 35.00 44.00 435.00 0.25 1.37
4625.00 46.00 35.00 46.00 432.00 0.25 1.23
4726.00 150.00 45.00 30.00 200.00 0.25 1.20
4818.50 25.00 0.00 30.00 6.00 0.25 1.09
4918.50 12.00 0.00 30.00 6.00 0.25 0.78
5022.40 10.00 35.00 30.00 10.00 0.25 2.00
5121.40 10.00 30.34 30.00 20.00 0.25 1.70
5222.00 10.00 36.00 45.00 50.00 0.25 1.02
5322.00 20.00 36.00 45.00 50.00 0.25 0.89
5412.00 0.00 30.00 35.00 4.00 0.25 1.46
5512.00 0.00 30.00 45.00 8.00 0.25 0.80
5622.00 10.00 35.00 45.00 10.00 0.40 0.90
5720.00 20.00 36.00 45.00 30.00 0.50 0.83
5820.00 0.10 36.00 45.00 50.00 0.29 0.79
5920.00 0.10 36.00 45.00 50.00 0.50 0.67
6022.00 0.00 40.00 33.00 8.00 0.39 1.45
6124.00 0.00 40.00 33.00 8.00 0.30 1.58
6220.00 0.00 24.50 20.00 8.00 0.35 1.37
6318.00 0.00 30.00 33.00 8.00 0.30 2.05
6427.00 43.00 35.00 43.00 420.00 0.29 1.15
6527.00 50.00 40.00 42.00 407.00 0.29 1.44
6627.00 35.00 35.00 42.00 359.00 0.29 1.27
6727.00 37.50 35.00 37.80 320.00 0.29 1.24
6827.00 32.00 33.00 42.60 301.00 0.29 1.16
6927.00 32.00 33.00 42.20 239.00 0.29 1.30
7027.30 14.00 31.00 41.00 110.00 0.29 1.25
7127.30 31.50 29.70 41.00 135.00 0.29 1.25
7227.30 16.20 28.00 50.00 90.50 0.29 1.25
7327.30 36.00 1.00 50.00 92.00 0.29 1.25
7427.30 10.00 39.00 41.00 511.00 0.29 1.47
7527.30 10.00 39.00 40.00 470.00 0.29 1.43
7625.00 46.00 35.00 47.00 443.00 0.29 1.28
7725.00 46.00 35.00 44.00 435.00 0.29 1.37
7825.00 46.00 35.00 46.00 432.00 0.29 1.23
7926.00 150.00 45.00 30.00 230.00 0.29 1.20
8018.50 25.00 0.00 30.00 6.00 0.29 1.09
8118.50 12.00 0.00 30.00 6.00 0.29 0.78
8222.00 10.00 35.00 30.00 10.00 0.29 2.00
8321.00 10.00 30.34 30.00 30.00 0.29 1.20
8422.00 10.00 36.00 45.00 50.00 0.29 1.02
8522.00 20.00 36.00 45.00 30.00 0.29 0.89
8612.00 0.03 30.00 35.00 4.00 0.29 0.46
8712.00 0.00 30.00 45.00 8.00 0.29 0.80
8812.00 0.00 30.00 35.00 4.00 0.29 1.44
8931.30 68.00 37.00 49.00 200.50 0.29 1.20
9020.00 30.00 36.00 45.00 50.00 0.29 0.96
9119.60 21.80 29.50 37.80 40.30 0.25 1.78
9223.10 25.20 29.20 36.50 61.90 0.40 1.70
9323.80 31.00 38.70 47.50 23.50 0.31 1.90
9422.30 20.10 31.00 40.20 88.00 0.19 1.47
9523.50 25.00 20.00 49.10 115.00 0.41 1.63
9623.00 20.00 20.30 46.20 40.30 0.25 1.4.8
9721.50 15.00 29.00 41.50 123.60 0.36 1.25
9823.40 15.00 38.50 30.30 45.20 0.28 1.17
9919.60 17.80 29.20 46.80 201.20 0.37 1.42
10022.10 24.20 39.70 45.80 49.50 0.21 1.58
10118.6826.3415358.2301.11
10216.511.490303.6601.00
10318.8414.36252030.501.88
10418.8457.46202030.502.05
10528.4429.42353510001.78
10628.4439.23383510001.99
10720.616.2826.5304001.25
10814.8017205001.13
1091411.9726308801.02
11025120455312001.30
11126150.05455020001.20
11218.525030601.09
11318.512030600.78
11422.41035301002.00
11521.41030.34302001.70
116222036455001.02
11722036455000.89
1181203035401.46
1191203045800.80
1201203035401.44
1211203045800.86
12223.470323721401.08
1231670204011501.11
12420.4124.9132210.670.351.40
12519.6311.97202212.190.4051.35
12621.828.62322812.80.491.03
12720.4133.52111645.720.21.28
12818.8415.32302510.670.381.63
12918.84020207.620.451.05
13021.4302020610.51.03
13119.0611.712835210.111.09
13218.8414.36252030.50.451.11
13321.516.94303176.810.381.01
1341411.972630880.450.63
135182430.1545200.121.12
13623020201000.31.20
13722.41004545150.251.80
13822.4103545100.40.90
13920203645500.250.96
14020203645500.50.83
1412003645500.250.79
1422003645500.50.67
143220403380.351.45
144240403380.31.58
14520024.52080.351.37
146185302080.32.05
14727.30 28.00 16.20 50.00 90.50 0.29 1.25
14827.30 31.00 26.00 50.00 92.00 0.25 1.25
14927.30 31.00 14.35 41.00 109.70 0.25 1.25
15025.00 35.00 46.00 46.00 393.00 0.25 1.31
15131.30 37.00 68.00 49.00 200.50 0.29 1.20
15227.00 33.00 31.99 42.40 290.00 0.25 1.30
15320.41 11.00 33.51 16.00 45.71 0.20 1.28
15420.20 22.30 16.70 42.40 25.00 0.25 1.39
15523.00 20.00 0.00 20.00 99.80 0.30 1.20
15627.30 31.00 14.00 41.00 110.00 0.29 1.25
15726.18 59.00 44.93 31.50 172.98 0.10 1.19
15822.40 27.00 20.00 30.00 54.00 0.29 1.48
15926.00 45.00 50.00 30.00 230.00 0.29 1.20
16022.30 31.00 20.10 40.20 88.00 0.19 1.47
16127.00 35.00 35.00 42.00 359.00 0.29 1.27
16227.00 33.00 32.00 42.60 301.00 0.25 1.16
16328.35 44.97 33.49 43.16 413.42 0.25 1.16
16427.30 39.00 10.00 40.00 470.00 0.29 1.43
16528.01 9.50 37.36 41.86 538.10 0.23 1.55
16620.40 20.40 25.00 35.00 35.00 0.32 1.77
16725.00 35.00 46.00 47.00 443.00 0.29 1.28
16827.30 31.00 14.00 41.00 511.00 0.25 1.25
16927.00 35.00 35.00 37.00 30.00 0.25 1.24
17031.25 25.73 27.97 48.23 91.55 0.21 1.11
17127.30 29.70 31.50 41.00 135.00 0.29 1.25
17227.00 40.00 50.00 42.00 407.00 0.29 1.44
17318.12 10.57 30.84 32.45 21.77 0.11 1.18
17427.30 29.70 31.50 41.00 135.00 0.25 1.25
17527.30 28.00 16.80 50.00 90.50 0.25 1.20
17626.78 26.79 30.66 43.66 249.70 0.25 1.26
17731.30 37.00 68.00 47.00 213.00 0.25 1.20
17819.60 29.20 17.80 46.20 20.20 0.37 0.96
17923.80 38.70 31.00 41.50 23.50 0.31 0.80
18021.50 19.30 14.00 38.90 35.00 0.27 1.42
18126.83 13.98 35.46 43.50 96.14 0.23 1.42
18225.00 40.00 48.00 49.00 330.00 0.25 1.49
18325.00 35.00 46.00 46.00 42.00 0.29 1.63

References

  1. Wang, G.; Zhao, B.; Wu, B.; Zhang, C.; Liu, W. Intelligent Prediction of Slope Stability Based on Visual Exploratory Data Analysis of 77 in Situ Cases. Int. J. Min. Sci. Technol. 2023, 33, 47–59. [Google Scholar] [CrossRef]
  2. Lin, S.; Zheng, H.; Han, B.; Li, Y.; Han, C.; Li, W. Comparative Performance of Eight Ensemble Learning Approaches for the Development of Models of Slope Stability Prediction. Acta Geotech. 2022, 17, 1477–1502. [Google Scholar] [CrossRef]
  3. Basahel, H.; Mitri, H. Probabilistic Assessment of Rock Slopes Stability Using the Response Surface Approach—A Case Study. Int. J. Min. Sci. Technol. 2019, 29, 357–370. [Google Scholar] [CrossRef]
  4. Rezaei, M.; Seyed Mousavi, S.Z. Slope Stability Analysis of an Open Pit Mine with Considering the Weathering Agent: Field, Laboratory and Numerical Studies. Eng. Geol. 2024, 333, 107503. [Google Scholar] [CrossRef]
  5. Yang, Y.; Zhou, W.; Jiskani, I.M.; Lu, X.; Wang, Z.; Luan, B. Slope Stability Prediction Method Based on Intelligent Optimization and Machine Learning Algorithms. Sustainability 2023, 15, 1169. [Google Scholar] [CrossRef]
  6. Sakellariou, M.G.; Ferentinou, M.D. A Study of Slope Stability Prediction Using Neural Networks. Geotech. Geol. Eng. 2005, 23, 419–445. [Google Scholar] [CrossRef]
  7. Wang, H.B.; Xu, W.Y.; Xu, R.C. Slope Stability Evaluation Using Back Propagation Neural Networks. Eng. Geol. 2005, 80, 302–315. [Google Scholar] [CrossRef]
  8. Gordan, B.; Jahed Armaghani, D.; Hajihassani, M.; Monjezi, M. Prediction of Seismic Slope Stability through Combination of Particle Swarm Optimization and Neural Network. Eng. Comput. 2016, 32, 85–97. [Google Scholar] [CrossRef]
  9. Shang, L.; Nguyen, H.; Bui, X.-N.; Vu, T.H.; Costache, R.; Hanh, L.T.M. Toward State-of-the-Art Techniques in Predicting and Controlling Slope Stability in Open-Pit Mines Based on Limit Equilibrium Analysis, Radial Basis Function Neural Network, and Brainstorm Optimization. Acta Geotech. 2022, 17, 1295–1314. [Google Scholar] [CrossRef]
  10. Mahmoodzadeh, A.; Mohammadi, M.; Farid Hama Ali, H.; Hashim Ibrahim, H.; Nariman Abdulhamid, S.; Nejati, H.R. Prediction of Safety Factors for Slope Stability: Comparison of Machine Learning Techniques. Nat. Hazards 2022, 111, 1771–1799. [Google Scholar] [CrossRef]
  11. Huang, F.; Xiong, H.; Chen, S.; Lv, Z.; Huang, J.; Chang, Z.; Catani, F. Slope Stability Prediction Based on a Long Short-Term Memory Neural Network: Comparisons with Convolutional Neural Networks, Support Vector Machines and Random Forest Models. Int. J. Coal Sci. Technol. 2023, 10, 18. [Google Scholar] [CrossRef]
  12. Kim, T.-Y.; Cho, S.-B. Predicting Residential Energy Consumption Using CNN-LSTM Neural Networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  13. Wang, J.X.; Tang, S.B.; Heap, M.J.; Tang, C.A.; Tang, L.X. An Auto-Detection Network to Provide an Automated Real-Time Early Warning of Rock Engineering Hazards Using Microseismic Monitoring. Int. J. Rock Mech. Min. Sci. 2021, 140, 104685. [Google Scholar] [CrossRef]
  14. Samui, P. Slope Stability Analysis: A Support Vector Machine Approach. Environ. Geol. 2008, 56, 255–267. [Google Scholar] [CrossRef]
  15. Das, S.K.; Biswal, R.K.; Sivakugan, N.; Das, B. Classification of Slopes and Prediction of Factor of Safety Using Differential Evolution Neural Networks. Environ. Earth Sci. 2011, 64, 201–210. [Google Scholar] [CrossRef]
  16. Erzin, Y.; Cetin, T. The Prediction of the Critical Factor of Safety of Homogeneous Finite Slopes Using Neural Networks and Multiple Regressions. Comput. Geosci. 2013, 51, 305–313. [Google Scholar] [CrossRef]
  17. Liu, Z.; Shao, J.; Xu, W.; Chen, H.; Zhang, Y. An Extreme Learning Machine Approach for Slope Stability Evaluation and Prediction. Nat. Hazards 2014, 73, 787–804. [Google Scholar] [CrossRef]
  18. Hoang, N.-D.; Pham, A.-D. Hybrid Artificial Intelligence Approach Based on Metaheuristic and Machine Learning for Slope Stability Assessment: A Multinational Data Analysis. Expert Syst. Appl. 2016, 46, 60–68. [Google Scholar] [CrossRef]
  19. Xue, X.; Li, Y.; Yang, X.; Chen, X.; Xiang, J. Prediction of Slope Stability Based on GA-BP Hybrid Algorithm. Neural Netw. World 2015, 25, 189–202. [Google Scholar] [CrossRef]
  20. Suman, S.; Khan, S.Z.; Das, S.K.; Chand, S.K. Slope Stability Analysis Using Artificial Intelligence Techniques. Nat. Hazards 2016, 84, 727–748. [Google Scholar] [CrossRef]
  21. Verma, A.K.; Singh, T.N.; Chauhan, N.K.; Sarkar, K. A Hybrid FEM–ANN Approach for Slope Instability Prediction. J. Inst. Eng. India Ser. A 2016, 97, 171–180. [Google Scholar] [CrossRef]
  22. Feng, X.; Li, S.; Yuan, C.; Zeng, P.; Sun, Y. Prediction of Slope Stability Using Naive Bayes Classifier. KSCE J. Civ. Eng. 2018, 22, 941–950. [Google Scholar] [CrossRef]
  23. Rukhaiyar, S.; Alam, M.N.; Samadhiya, N.K. A PSO-ANN Hybrid Model for Predicting Factor of Safety of Slope. Int. J. Geotech. Eng. 2017, 12, 556–566. [Google Scholar] [CrossRef]
  24. Xue, X. Prediction of Slope Stability Based on Hybrid PSO and LSSVM. J. Comput. Civ. Eng. 2017, 31, 04016041. [Google Scholar] [CrossRef]
  25. Qi, C.; Tang, X. Slope Stability Prediction Using Integrated Metaheuristic and Machine Learning Approaches: A Comparative Study. Comput. Ind. Eng. 2018, 118, 112–122. [Google Scholar] [CrossRef]
  26. Lin, Y.; Zhou, K.; Li, J. Prediction of Slope Stability Using Four Supervised Learning Methods. IEEE Access 2018, 6, 31169–31179. [Google Scholar] [CrossRef]
  27. Qi, C.; Tang, X. A Hybrid Ensemble Method for Improved Prediction of Slope Stability. Num. Anal. Meth. Geomech. 2018, 42, 1823–1839. [Google Scholar] [CrossRef]
  28. Zhou, J.; Li, E.; Yang, S.; Wang, M.; Shi, X.; Yao, S.; Mitri, H.S. Slope Stability Prediction for Circular Mode Failure Using Gradient Boosting Machine Approach Based on an Updated Database of Case Histories. Saf. Sci. 2019, 118, 505–518. [Google Scholar] [CrossRef]
  29. Bui, D.T.; Moayedi, H.; Gör, M.; Jaafari, A.; Foong, L.K. Predicting Slope Stability Failure through Machine Learning Paradigms. IJGI 2019, 8, 395. [Google Scholar] [CrossRef]
  30. Moayedi, H.; Osouli, A.; Nguyen, H.; Rashid, A.S.A. A Novel Harris Hawks’ Optimization and k-Fold Cross-Validation Predicting Slope Stability. Eng. Comput. 2021, 37, 369–379. [Google Scholar] [CrossRef]
  31. Bui, X.-N.; Nguyen, H.; Choi, Y.; Nguyen-Thoi, T.; Zhou, J.; Dou, J. Prediction of Slope Failure in Open-Pit Mines Using a Novel Hybrid Artificial Intelligence Model Based on Decision Tree and Evolution Algorithm. Sci. Rep. 2020, 10, 9939. [Google Scholar] [CrossRef] [PubMed]
  32. Zhang, W.; Li, H.; Han, L.; Chen, L.; Wang, L. Slope Stability Prediction Using Ensemble Learning Techniques: A Case Study in Yunyang County, Chongqing, China. J. Rock Mech. Geotech. Eng. 2022, 14, 1089–1099. [Google Scholar] [CrossRef]
  33. Asteris, P.G.; Rizal, F.I.M.; Koopialipoor, M.; Roussis, P.C.; Ferentinou, M.; Armaghani, D.J.; Gordan, B. Slope Stability Classification under Seismic Conditions Using Several Tree-Based Intelligent Techniques. Appl. Sci. 2022, 12, 1753. [Google Scholar] [CrossRef]
  34. Demir, S.; Sahin, E.K. Assessing the Predictive Capability of DeepBoost Machine Learning Algorithm Powered by Hyperparameter Tuning Methods for Slope Stability Prediction. Environ. Earth Sci. 2023, 82, 562. [Google Scholar] [CrossRef]
  35. Lu, R.; Wei, W.; Shang, K.; Jing, X. Stability Analysis of Jointed Rock Slope by Strength Reduction Technique Considering Ubiquitous Joint Model. Adv. Civ. Eng. 2020, 2020, 8862243. [Google Scholar] [CrossRef]
  36. Deng, D.; Li, L.; Zhao, L. Limit Equilibrium Method (LEM) of Slope Stability and Calculation of Comprehensive Factor of Safety with Double Strength-Reduction Technique. J. Mt. Sci. 2017, 14, 2311–2324. [Google Scholar] [CrossRef]
  37. Hu, S.; Lee, W.-H.; Shan, C.; Xue, X.; Yang, H. Research on slope stability based on improved PSO-BP neural network. J. Disaster Prev. Mitig. Eng. 2023, 43, 854–861. (In Chinese) [Google Scholar] [CrossRef]
  38. Fu, Y.; Liu, S.; Liu, D. RBF neural network in predicting the stability of rock slope. J. Wuhan Univ. Technol. (Traffic Sci. Eng. Ed.) 2003, 27, 170–173. (In Chinese) [Google Scholar]
  39. Wang, Z.-Z.; Goh, S.H. Novel Approach to Efficient Slope Reliability Analysis in Spatially Variable Soils. Eng. Geol. 2021, 281, 105989. [Google Scholar] [CrossRef]
  40. Hsiao, C.-H.; Chen, A.Y.; Ge, L.; Yeh, F.-H. Performance of Artificial Neural Network and Convolutional Neural Network on Slope Failure Prediction Using Data from the Random Finite Element Method. Acta Geotech. 2022, 17, 5801–5811. [Google Scholar] [CrossRef]
  41. Fu, Y.; Lin, M.; Zhang, Y.; Chen, G.; Liu, Y. Slope Stability Analysis Based on Big Data and Convolutional Neural Network. Front. Struct. Civ. Eng. 2022, 16, 882–895. [Google Scholar] [CrossRef]
  42. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef]
  43. Shewalkar, A.; Nyavanandi, D.; Ludwig, S.A. Performance Evaluation of Deep Neural Networks Applied to Speech Recognition: RNN, LSTM and GRU. J. Artif. Intell. Soft Comput. Res. 2019, 9, 235–245. [Google Scholar] [CrossRef]
  44. Park, J.; Woo, S.; Lee, J.-Y.; Kweon, I.S. A Simple and Light-Weight Attention Module for Convolutional Neural Networks. Int. J. Comput. Vis. 2020, 128, 783–798. [Google Scholar] [CrossRef]
  45. Chen, B.; Li, P.; Sun, C.; Wang, D.; Yang, G.; Lu, H. Multi Attention Module for Visual Tracking. Pattern Recognit. 2019, 87, 80–93. [Google Scholar] [CrossRef]
  46. Cun, X.; Pun, C.-M. Improving the Harmony of the Composite Image by Spatial-Separated Attention Module. IEEE Trans. Image Process. 2020, 29, 4759–4771. [Google Scholar] [CrossRef]
  47. Han, Y.; Liu, Y.; Huang, Q.; Zhang, Y. SOC Estimation for Lithium-Ion Batteries Based on BiGRU with SE Attention and Savitzky-Golay Filter. J. Energy Storage 2024, 90, 111930. [Google Scholar] [CrossRef]
  48. Feng, Y.; Chen, J.; Zhang, T.; He, S.; Xu, E.; Zhou, Z. Semi-Supervised Meta-Learning Networks with Squeeze-and-Excitation Attention for Few-Shot Fault Diagnosis. ISA Trans. 2022, 120, 383–401. [Google Scholar] [CrossRef]
  49. Saleem, N.; Elmannai, H.; Bourouis, S.; Trigui, A. Squeeze-and-Excitation 3D Convolutional Attention Recurrent Network for End-to-End Speech Emotion Recognition. Appl. Soft Comput. 2024, 161, 111735. [Google Scholar] [CrossRef]
Figure 1. Slope geometry and factors affecting slope stability.
Figure 1. Slope geometry and factors affecting slope stability.
Sustainability 16 06333 g001
Figure 2. Point box semi-violin diagram of slope stability prediction index system.
Figure 2. Point box semi-violin diagram of slope stability prediction index system.
Sustainability 16 06333 g002
Figure 3. Spearman multivariate graph.
Figure 3. Spearman multivariate graph.
Sustainability 16 06333 g003
Figure 4. GRU structure diagram.
Figure 4. GRU structure diagram.
Sustainability 16 06333 g004
Figure 5. Squeeze-and-Excitation attention structure diagram.
Figure 5. Squeeze-and-Excitation attention structure diagram.
Sustainability 16 06333 g005
Figure 6. Research model of this paper: CNN-GRU-SE.
Figure 6. Research model of this paper: CNN-GRU-SE.
Sustainability 16 06333 g006
Figure 7. Loss function curves.
Figure 7. Loss function curves.
Sustainability 16 06333 g007
Figure 8. Root-mean-square error curve.
Figure 8. Root-mean-square error curve.
Sustainability 16 06333 g008
Figure 9. Prediction of slope safety factor by various models.
Figure 9. Prediction of slope safety factor by various models.
Sustainability 16 06333 g009
Figure 10. Comparison of indicators: (a) MAE; (b) MAPE; (c) MSE; (d) RMSE; (e) R2.
Figure 10. Comparison of indicators: (a) MAE; (b) MAPE; (c) MSE; (d) RMSE; (e) R2.
Sustainability 16 06333 g010
Figure 11. Error bar plots for different prediction models.
Figure 11. Error bar plots for different prediction models.
Sustainability 16 06333 g011aSustainability 16 06333 g011b
Figure 12. Importance of each model.
Figure 12. Importance of each model.
Sustainability 16 06333 g012
Figure 13. Contributions of various model variables.
Figure 13. Contributions of various model variables.
Sustainability 16 06333 g013
Table 2. Specific structural parameters of the model.
Table 2. Specific structural parameters of the model.
Layer CategoryNeuronsRemark
Input layer6 × 1
Convolution layer5 × 32Weights 2 × 32
Bias 1 × 32
Batch normalization5 × 32Offset 1 × 32
Scale 1 × 32
Re LU5 × 32
Maximum pooling5 × 32
Full connection1 × 16Weights 16 × 64
Bias 16 × 1
Re LU8 × 16
Dimensional global averaging pooling8 × 32
Full connection1 × 16Weights 16 × 32
Bias 16 × 1
GRU1 × 16Input weights (48 × 16)
Recurrent weights (48 × 16)
Bias (48 × 1)
Self-attention1 × 16Query weights (2 × 16)
Key weights (2 × 16)
Value weights (2 × 16)
Output weights (16 × 2)
Query Bias (2 × 1)
Key Bias (2 × 1)
Value Bias (2 × 1)
Output Bias (16 × 1)
Full connection1 × 1Weights 1 × 16
Bias 1 × 1
Output layer1 × 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Da, Q.; Chen, Y.; Dai, B.; Li, D.; Fan, L. Prediction of Slope Safety Factor Based on Attention Mechanism-Enhanced CNN-GRU. Sustainability 2024, 16, 6333. https://doi.org/10.3390/su16156333

AMA Style

Da Q, Chen Y, Dai B, Li D, Fan L. Prediction of Slope Safety Factor Based on Attention Mechanism-Enhanced CNN-GRU. Sustainability. 2024; 16(15):6333. https://doi.org/10.3390/su16156333

Chicago/Turabian Style

Da, Qi, Ying Chen, Bing Dai, Danli Li, and Longqiang Fan. 2024. "Prediction of Slope Safety Factor Based on Attention Mechanism-Enhanced CNN-GRU" Sustainability 16, no. 15: 6333. https://doi.org/10.3390/su16156333

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop