Next Article in Journal
Influence of Low-Temperature Stress-Relieving Treatment in the Fatigue Life of Components Produced by Laser Powder Bed Fusion in AlSi10Mg
Previous Article in Journal
Optimizing Automated Hematoma Expansion Classification from Baseline and Follow-Up Head Computed Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of GNSS Velocity Accuracies Using Machine Learning Algorithms for Active Fault Slip Rate Determination and Earthquake Hazard Assessment

by
Halil İbrahim Solak
1,2
1
Distance Education Vocational School, Afyon Kocatepe University, Afyonkarahisar 03200, Türkiye
2
Earthquake Implementation and Research Center of Afyon Kocatepe University, Afyonkarahisar 03200, Türkiye
Appl. Sci. 2025, 15(1), 113; https://doi.org/10.3390/app15010113
Submission received: 26 November 2024 / Revised: 22 December 2024 / Accepted: 24 December 2024 / Published: 27 December 2024
(This article belongs to the Section Earth Sciences)

Abstract

:
GNSS technology utilizes satellite signals to determine the position of a point on Earth. Using this location information, the GNSS velocities of the points can be calculated. GNSS velocity accuracies are crucial for studies requiring high precision, as fault slip rates typically range within a few millimeters per year. This study employs machine learning (ML) algorithms to predict GNSS velocity accuracies for fault slip rate estimation and earthquake hazard analysis. GNSS data from four CORS stations collected over 1-, 2-, and 3-year intervals with observation durations of 2, 4, 6, 8, and 12 h, were analyzed to generate velocity estimates. Position accuracies, observation intervals, and corresponding velocity accuracies formed two datasets for the East and North components. ML models, including Support Vector Machine, Random Forest, K-Nearest Neighbors, and Multiple Linear Regression, were used to model the relationship between position and velocity accuracies. The findings reveal that the Random Forest, which makes more accurate and reliable predictions by evaluating many decision trees together, achieved over 90% accuracy for both components. Velocity accuracies of ±1.3 mm/year were obtained for 1-year interval data, while accuracies of ±0.6 mm/year were achieved for the 2- and 3-year intervals. Three campaigns were deemed sufficient for Holocene faults with higher slip rates. However, for Quaternary faults with lower slip rates, longer observation periods or additional campaigns are necessary to ensure reliable velocity estimates. This highlights the need for GNSS observation planning based on fault activity.

1. Introduction

Tectonic movements are fundamental processes that shape the dynamic structure of the Earth’s crust, involving the movement of tectonic plates [1]. In tectonically active regions, these plate movements can trigger catastrophic earthquakes [2]. As such, modeling tectonic motions is crucial for understanding the Earth’s crust dynamics and predicting natural disasters resulting from these movements [3,4,5]. Accurately determining the rate of energy accumulation along faults and forecasting how this accumulated energy will be released over time is essential for assessing earthquake risk.
The Global Navigation Satellite System (GNSS) technique that offers high location accuracy is widely used for modeling tectonic movements [3,4,6,7,8,9,10]. GNSS-based velocity measurements, geodetic observations, and artificial intelligence-assisted prediction methods have made great advances in modeling tectonic movements. Especially in regions such as Turkey, which is located at the intersection of multiple plate boundaries, such models both reveal new scientific knowledge and help the society to be prepared against natural disasters. The estimation of accurate geodetic velocity has, therefore, great importance in geosciences [11].
The modeling of tectonic movements with the GNSS technique consists of three stages. The first is to establish a GNSS network near the fault that is to be modeled with a geometry representing the fault. At this stage, factors such as the distribution of the stations to be included in the network, their location with respect to the fault and each other (proximity, distance, angle, etc.), and the type of ground on which the station will be established (e.g., rocky) are determined. The second one is where stations in the GNSS network are measured. At this stage, at least three campaigns of GNSS measures are performed (usually one year apart) at stations in the network. In the third and final, the GNSS data are evaluated with suitable software (GAMIT/GLBOK, Bernese, Gipsy X, etc.), and the location and velocity information of each station are obtained. By analyzing the velocity data that are obtained, the fault is modeled by using information that includes fault kinematics, slip rate, and strain rate, and an earthquake hazard analysis is performed with this information [2,10].
All stages mentioned above are important for modeling tectonic movements. To realize a better model, the velocities of the stations in the GNSS network must be determined with high accuracy. When it is considered that the slip velocities of the faults are within an accuracy of a few mm/year, it can be seen that the velocity accuracies to be obtained should be lower than the fault slip velocities. If this is not the case, it will not be possible to determine the fault slip rates with the velocity data to be obtained. On the other hand, the velocity accuracies are directly related to the position accuracies as a result of the measurements in each campaign, regardless of the displacement. High position accuracy is required for high velocity accuracy. To achieve high position accuracy, various models are employed during the data evaluation phase [12]. High accuracy in positioning with the GNSS depends on factors, including measurement time, facility type, and the GNSS data evaluation strategy. In addition, the time elapsed between campaigns is also important in determining velocity accuracy [13]. Thus, it is important to model the position accuracies in each campaign and the elapsed time between campaigns in determining GNSS velocities in order to produce a higher level of accuracy in deformation and tectonic modeling studies with GNSS.
It is possible to calculate the positions and velocities of continuous GNSS stations with high accuracy (provided sufficient data is available). This is because these stations do not include effects such as leveling error or antenna height measurement error, and they provide uninterrupted data. However, the installation of these stations is quite costly. Considering the need for GNSS stations for modeling tectonic movements, it is not possible to install, operate, and manage dozens of stations for each fault. Instead, periodic campaign-type GNSS measurements are preferred. In this way, larger areas can be modeled tectonically with less cost and effort thanks to GNSS measurements performed once or several times a year.
The determination of GNSS velocity has been investigated in various studies [9,10,13,14,15,16,17]. A velocity field obtained from GNSS observations offers essential insights into earthquake hazards, including the moment accumulation rate and strain rate [18]. Velocity data have been extensively utilized by numerous researchers in studies related to crustal motion, plate boundary dynamics, seismic site analysis, and deformation kinematics [3,4,5,9,10,19,20].
Ref. [11] analyzed two distinct Artificial Neural Network (ANN) models, namely, the Back-Propagation Artificial Neural Network and the Radial Basis Function Neural Network, to estimate the velocities of 125 GPS sites in Turkey, and they revealed that the Back-Propagation Artificial Neural Network is an alternative tool to conventional methods for geodetic station velocity estimation. Ref. [21] conducted a comparative assessment of three distinct Artificial Neural Network (ANN) models to estimate the geodetic velocities of 238 sites in Turkey, identifying the generalized regression neural network as the most suitable model. Ref. [22] investigated the applicability of four different deep learning (DL) methods for estimating GPS-derived geodetic velocities at 42 GNSS stations in northwestern Iran. The findings indicate that the CNN method exhibits a lower goodness of fit and higher root mean square error (RMSE). In the most recent study conducted by [17], ML algorithms were used to estimate horizontal GNSS velocities in active tectonic regions. In the study where cluster analysis was used, a velocity accuracy lower than 0.4 mm/year was estimated.
Machine learning constitutes a sub-branch of artificial intelligence that processes raw data by using mathematical and statistical methods and gains learning skills by detecting the relationship between data. With the success of machine learning techniques in revealing hidden patterns and complex relationships between data, they have been successfully applied in various fields such as pattern recognition, computer vision, and finance. The purpose of this discipline is to process large amounts of raw data and to create models that enable predictions to be made from this data. Machine learning techniques are grouped under three main headings: supervised learning, unsupervised learning, and reinforcement learning [23,24].
In this study, the GNSS data from four different CORS stations between 2011 and 2017 were divided into different measurement times and evaluated with GAMIT/GLOBK, the position accuracies of each campaign were obtained, and the velocity was produced by creating 375 combinations for each station. The results for the four stations (1500 rows of data) were evaluated with four different supervised machine learning estimation algorithms—(Support Vector Machine (SVM), Multiple Linear Regression (MLR), Random Forest (RF), and K-Nearest Neighbor (KNN))—and the effect of the time between campaigns and position accuracy in each campaign on velocity accuracy was modeled. In this way, we aimed to estimate the velocity accuracy by considering the accuracy achieved since the first campaign in studies where GNSS velocities were used.

2. Materials and Methods

2.1. Preparation of the Dataset

The dataset to be used in machine learning applications is of great importance for the success and accuracy of the model. Having a dataset that has been tested and approved in different studies in the literature saves time before the application. However, not having a ready dataset can complicate the pre-application process, but it also provides in-depth information during the data collection and preparation phase. Creating our own dataset for the application to be made using ML methods included collecting the specific data needed, cleaning these data, and labeling them appropriately. These stages are very important for obtaining accurate and reliable results, but they can also be time-consuming. The quality of the data directly affects the performance of the machine learning model; therefore, creating a good dataset by allocating enough time and resources is important for obtaining a more successful model in the long run.
In the literature, there is no ready dataset that includes the location, velocity accuracies, and year intervals of GNSS stations that can be used with machine learning algorithms. Therefore, the development of the machine learning (ML) model required the creation of a new dataset. To create this data set, four GNSS stations (BALK, BILE, ESKS, and HARC) belonging to the Turkish National Permanent GNSS Network Active (TUSAGA-Active, also known as CORS-TR) network in Turkey were selected considering their time series and whether each station had data on the same day of the years 2011–2017 (Figure 1 and Figure 2). During this period, no significant earthquakes occurred near the stations that would have caused any translational movement. The 24-h GNSS data from the selected stations were downloaded from the TUSAGA-Active website and were divided into 2-, 4-, 6-, 8-, and 12-h periods with Translation, Editing, and Quality Check (TEQC) software (February 2019 version). The precise Earth Orientation Parameters (EOPs), satellite orbits, and clocks [25] needed for the process were obtained from the IERS Bulletin EOP [26], while the antenna phase center models were obtained from the International GNSS Service (IGS). Noise stochasticity and offsets can significantly impact the accuracy of GNSS-derived velocity time series [27]. However, the GAMIT/GLOBK v10.71 software, employed for GNSS data processing, utilizes stochastic modeling, Kalman filtering, and statistical analysis of residuals to enhance the accuracy and reliability of velocity estimates [28]. Detailed information on this issue can be found in [28].
Following the processing of the GNSS data, a total of 1500 velocities were calculated in the Eurasian-fixed ITRF14 reference frame, with 375 velocities per station across four stations, to evaluate the effect of positional accuracy on velocity accuracy for each campaign. These velocities were calculated by considering all possible combinations of position data derived from GNSS observations conducted for 2, 4, 6, 8, and 12 h in each campaign, ensuring that only one observation duration was selected from each year. Following the processing of GNSS data and the velocity generation stage, the relevant data from the files, including the position and velocity accuracies produced by the GAMIT/GLOBK software [28], were extracted by writing small Python scripts, which were saved as a Comma Separated Values (CSV) file. In the dataset, there are seven independent variables (year interval, Se1, Sn1, Se2, Sn2, Se3, Sn3) and two dependent variables (Sve, Svn). Afterwards, velocity accuracies and position accuracies were separated on the basis of components, and two separate datasets were created. Detailed information about the datasets is given in Table 1 and Table 2.

2.2. Machine Learning

Machine learning is a sub-branch of artificial intelligence that works on numerical, text, and visual data to imitate the way humans learn, and it enables computers to learn with different algorithms and mathematical models with a focus on increasing accuracy. In other words, it is the use of statistical methods to determine the correlation between data in a certain data group, to detect complex patterns that humans cannot distinguish, and to give the computer the ability to decide [29]. In this context, regression algorithms from machine learning algorithms, including SVM, MLR, RF, and KNN, were used while analyzing the dataset.

2.2.1. Multiple Linear Regression

Multiple Linear Regression (MLR) is a statistical method used to model the linear relationship between a dependent variable (y) and multiple independent variables (x1, x2, x3, …). This technique enables the estimation of the overall impact on the dependent variable by simultaneously analyzing the contributions of multiple factors [30,31]. For example, the multiple linear regression model for p independent and 1 dependent variable is given in Equation (1).
y i = β o + β 1 x i + + β p x p i + ϵ i , i = 1 ,
From this point of view, the multivariate regression equation in Equation (2) is obtained.
y i = β o + j = 1 p X i j β i j + ϵ i
This algorithm models the linear relationship between the dependent variable, velocity, and the independent variables, which include coordinates and their corresponding sigma values. The model was trained using the Ordinary Least Squares (OLS) method.

2.2.2. Support Vector Machines

Support Vector Machines (SVMs) constitute a supervised machine learning algorithm that can be used for classification or regression problems. The basic idea of the SVM regression method is to find the linear discriminant function that reflects the characteristics of the available training data as closely as possible and that fits the statistical learning theory. Similar to classification, kernel functions are used in regression to handle nonlinear cases. Two situations that can be encountered in SVMs are that the data are in a structure that can be separated linearly or in a structure that cannot be separated linearly. In case the samples can be separated linearly, the aim is to find the most suitable separator plane dividing both classes equidistantly [32].
x i , y i , x i R n , y i 1 , 1 , i = 1 , , 1
The SVM model for Equation (3) is given in Equation (4).
m i n w , b 1 2 w t w + C i = 1 1 δ w , b ; x i , y i
δ w , b ; x i , y i is the loss function, and C 0 is the penalty parameter for the training error. Two loss functions that are widely used are given in Equations (5) and (6).
m a x ( 1 y i w T σ x i + b , 0 )
m a x ( 1 y i w T σ x i + b , 0 ) 2
Here, σ represents the function used to move the training data to a higher dimensional space. The decision function for each x test data is given in Equation (7).
f x = s g n w T σ x + b
For SVM training, the RBF kernel function was used, which gives better results than linear, polynomial, and RBF (Radial Basis Function) functions. The formula is given in Equation (8):
K x i , x j = exp γ x i x j 2 , γ > 0
γ = 1 2 σ 2
In the equation, x i x j 2 shows vectorial distance. The γ parameter is a free variable, and it determines the width of function. As its value increases, the width of the bell-shaped function becomes narrower, and, as its value decreases, the width increases [33].
The Support Vector Machine (SVM) algorithm predicts outcomes by identifying the optimal hyperplane that separates data points. During model training, k-fold cross-validation was employed to optimize the C and ϵ hyperparameters, ensuring improved model performance.

2.2.3. Random Forest

Random Forest (RF) is a regression technique that combines the performance of multiple Decision Tree (DT) algorithms to classify or predict the value of a variable [34,35,36,37]. That is, when RF receives an (x) input vector made up of the values of the different evidential features analyzed for a given training area, RF builds a number K of regression trees and averages the results. After K, such trees { T ( x ) } 1 K are grown, and the RF regression predictor is as seen in Formula (10) [38]:
f ^ r f K ( x ) = 1 K k = 1 K T x .
Even though it is expected to obtain a higher model performance when more trees are derived in the random forest regression algorithm, it can be stated that the increase in the number of trees does not always guarantee higher performance with regard to the type and size of the dataset. For this reason, it is recommended that model performances be compared by deriving different numbers of trees. One of the disadvantages of the model is that it is not possible to visually present the result with the tree structure via a single decision tree. Another one is that due to the complexity of the model, it does not allow the processing steps of the evaluation of many decision trees to be seen [38].
The RF algorithm is an ensemble method that combines multiple decision trees to make predictions. During model training, hyperparameters such as the number of trees (nestimators = 100) and the maximum depth (max_depth = 7) of the trees are optimized to enhance model accuracy.

2.2.4. K-Nearest Neighbor Regression (KNN)

K-nearest neighbor (K-NN) regression is a nonparametric regression method where the information derived from the observed data is applied to forecast the amount of predicted variables in real time without defining a predetermined parametric relation between the predictor and predicted variables. The basis of this method is calculating the similarity (neighborhood) of the real time amount of predictors X r = x 1 r , x 2 r , , x m r (with unknown forecast streamflow) with the amount of predictors for each of the historical observations X t = x 1 t , x 2 t , , x m t via the Euclidean distance function ( D r t ) as seen in Formula (11) below [31].
D r t = i = 1 m w i x i r x i t 2 , t = 1,2 , ,
Here, ω i ( i = 1,2 , , m ) are the weights of the predictors, and their sum is equal to 1. The estimated streamflow ( Y r ) is calculated using the following probabilistic function of the observed streamflows T j :
Y r = j = 1 K f ( D r j ) x T j
Here, f D r j is the kernel function of the K-nearest neighbors (K observed data with the lowest distance from the real time predictor), which is calculated based on distance amounts ( D r t ) as follows:
f D r j = 1 D r j j = 1 K 1 D r j
In the K-nearest neighbor algorithm, the amount of predictor weights ( ω in Equation (11)) and number of neighbors (K) affect the final results; therefore, their optimum amounts should be calculated to achieve the most appropriate results [39]. The K-Nearest Neighbors (KNNs) algorithm makes predictions by averaging the values of the (k) nearest neighbors to a given data point. During model training, the value of (k) (the number of neighbors) is optimized using k-fold cross-validation to improve model accuracy.
In this study, GNSS measurement data will be analyzed using the selected algorithms to estimate velocity accuracy. The dataset, divided into training and test subsets, will be utilized to assess model performance. This approach aims to predict the sigma value of velocity based on the sigma values of coordinates derived from each measurement.

2.3. Training, Testing, and Results of the ML Model

This study investigates the impact of position accuracy and the temporal intervals between campaigns on GNSS velocity accuracies for each component using machine learning techniques. The dataset is partitioned into two groups, corresponding to the East (E) and North (N) velocity components. For the E component, the independent variables (features) are Se1, Se2, and Se3, representing year intervals, while the dependent variable (target) is Sve. Similarly, for the N component, the independent variables are Sn1, Sn2, and Sn3, with Svn as the dependent variable. Both datasets are complete and contain no missing values (Table 1).
The data were partitioned into training (80%) and testing (20%) subsets. Four regression models—SVM, MLR, RF, and KNN—were implemented using the Scikit-Learn library in Python 3.10.0. The Elbow method was utilized to determine the optimal number of neighbors for the KNN algorithm, yielding values of 10 and 7 for the E and N components, respectively (Figure 3). During model training, k-fold (k = 5) cross-validation was employed [40]. Model performance, including training and testing scores, along with key evaluation metrics, is summarized in Table 3. Additional performance metrics that are commonly used in the literature were also evaluated, yielding comparable results [41].
The results presented in Table 2 indicate that the RF algorithm delivered the most accurate performance for both the E and N components. The training and testing scores for both components exceeded 95%, with an average RMSE of approximately ±0.1 mm/year. Additionally, k-fold cross-validation results for the RF algorithm demonstrated training and testing scores above 90%, suggesting robust generalization and reliable performance on real-world data. In comparison, the SVM and KNN algorithms achieved training scores above 91% for both components, with test scores ranging from 86% to 91%. However, the average RMSE values for these algorithms were higher at approximately ±0.3 mm/year. Notably, their k-fold cross-validation scores, which fall below 79%, reveal a discrepancy between the training/testing performance and generalization capability. This indicates that the optimistic results observed during training and testing might not hold consistently for unseen data, highlighting the potential for greater error in real-world applications. The MLR algorithm, which performed the weakest among all methods, achieved training and testing scores between 72% and 76% for both components, with its average RMSE values ranging from ±0.4 to ±0.5 mm/year. The k-fold cross-validation results for MLR align closely with its training and testing scores, falling within the 70–75% range. This consistency suggests that while MLR exhibited lower predictive accuracy, its performance was stable across datasets.
To evaluate the performance of all algorithms on the test data, the residuals (differences between the reference values and the predicted values) were calculated. The reference values (velocity accuracies) and residuals for each algorithm were visualized using dual vertical axes to enhance resolution and clarity (Figure 4, Figure 5, Figure 6 and Figure 7). The visualizations were created using Python’s Matplotlib library.
The RF algorithm demonstrated a balanced predictive behavior, underestimating 53.7% and overestimating 46.3% of the 300 test samples for the E component (Figure 4). For the N component, these percentages were 52.3% and 47.7%, respectively, indicating no systematic bias in the model’s predictions. For the E component, 3.7% of the residuals exceeded ±0.5 mm/year, with 1.3% categorized as underestimations (minimum residual: −1.1 mm/year) and 2.4% as overestimations (maximum residual: 1.25 mm/year). For the N component, 2% of the residuals exceeded ±0.5 mm/year, with 1% representing underpredictions (minimum residual: −0.67 mm/year) and 1% representing overpredictions (maximum residual: 0.83 mm/year). Outlier analysis did not reveal any evidence of systematic error within the dataset. In addition, no correlation was observed between the reference velocity accuracies and the predicted velocity accuracies (Figure 4).
The residual distribution showed that 89.3% of predictions for the E component and 93.7% for the N component fell within a ±0.2 mm/year residual margin. This high proportion of small residuals reflects the RF model’s strong predictive performance and low generalization error, indicating that the model effectively mapped the feature space to the target variable with minimal deviation.
According to the performance metrics, the KNN algorithm ranks as the second most successful model after the RF algorithm (Figure 5). Considering that the training performance for the KNN algorithm was 94% for both the E and N components, and the test performance was 91% and 89%, respectively, it can be concluded that the model provides very high accuracy (Figure 5). Residuals greater than ±0.5 mm/year account for 7.3% of the predictions for the E component and 5% for the N component. The residual values are predominantly concentrated around small negative and positive values, although larger positive residuals are observed at the extremes. For example, the highest residual is 1.73 mm/year, while the lowest is −0.99 mm/year. Despite the generally small residuals (82.7% for comp. E and 86.3% for comp. N at less than ±0.2 mm/year), a few outliers with high residuals suggest that the model occasionally makes larger errors. This aligns with the model’s 90% performance on the test data, indicating that while the model predicts correctly most of the time, it occasionally produces larger errors in some cases.
Similar to the RF and KNN models, the distribution of positive and negative residuals is balanced, with no indication of systematic bias. For the E and N components, the algorithm produced residuals larger than ±0.5 mm/year in 8% and 6.3% of the predictions, respectively. These residuals were predominantly positive (indicating overestimation), with the largest residuals being 1.40 mm/year for the E component and 1.58 mm/year for the N component. Additionally, the proportion of errors exceeding ±1 mm/year was 4% for the E component and 3% for the N component. These larger residuals have the potential to influence the results in the studies that require high-accuracy GNSS velocity estimates.
The MLR algorithm, which exhibited the lowest prediction accuracy in this study, achieved training and test scores of 72% and 71% for the E component and 76% and 72% for the N component, respectively (Table 3). These scores indicate that while MLR can reasonably explain the dependent variable, its predictive power is limited. The model produced residuals larger than ±0.5 mm/year in 21% of the predictions for the E component and 16% for the N component (Figure 7). Additionally, the largest residual observed across all models, 2.1 mm/year, was generated by MLR. In contrast, for residuals smaller than ±0.2 mm/year, the model performed relatively better, with 47.7% of predictions for the E component and 41.3% for the N component falling within this range. A notable limitation of the MLR model is its reduced performance when actual velocity accuracy values exceed ±5 mm/year. This suggests that MLR struggles to generalize effectively in scenarios with high-accuracy requirements or when the target variable exhibits significant variability.
Although the models trained with ML algorithms showed high performance in the test data, they can contain high errors in real data outside the dataset. For this reason, the four ML algorithms used in the study were tested with nine GNSS data from stations other than the dataset (Figure 8). These data had a campaign type measure and were measured at one-year intervals.
Table 4 and Figure 8 demonstrate that the RF, KNN, and SVM models delivered the most accurate predictions across all sites for both the E and N components. The residuals for these three models in all components remained below 0.3 mm/year. In contrast, the MLR model achieved high accuracy (residuals < 0.2 mm/year) when predicting reference values below 2 mm/year in both components but struggled to maintain accuracy for higher values, indicating its limitations in handling larger target ranges.
Using the developed machine learning model, velocity accuracies were predicted for varying position accuracies and temporal intervals between observations, which were tailored for tectonic-focused GNSS studies (Figure 9). Given that fault slip rates in many tectonic zones are on the order of a few millimeters per year, the prediction range for position accuracies was constrained between ±1.5 mm and ±5 mm. The lower limit corresponds to the highest position accuracy observed in the dataset, while the upper limit reflects a realistic threshold for tectonic applications. Notably, the model remains applicable for position accuracies exceeding ±5 mm, although such cases were not explicitly included in the training data.

3. Conclusions

Lower positional accuracy directly affects the accuracy of station velocity predictions, which in turn hampers the accurate calculation of fault slip velocities. This limitation is particularly critical for faults with low slip rates (e.g., 1–2 mm/year), as the calculated slip velocities may fall within or below the sigma value, obscuring or entirely masking the fault’s actual movement.
The key findings derived from the analysis in the presented study are summarized below:
  • East (E) Component: For 1-year interval GNSS data, achieving ±1.5 mm position accuracy per epoch resulted in a velocity accuracy greater than ±1 mm/year, with the best observed accuracy being ±1.3 mm/year. However, for 2- and 3-year interval datasets, submillimeter velocity accuracies could be achieved. Specifically, the best velocity accuracies were ±0.6 mm/year for 3-year intervals and ±0.7 mm/year for 2-year intervals.
  • North (N) Component: Similarly, for 1-year interval GNSS data with ±1.5 mm position accuracy per epoch, the maximum attainable velocity accuracy was ±1.4 mm/year. For 2- and 3-year interval data, the best achievable velocity accuracies improved to ±0.6 mm/year for 3-year intervals and ±0.8 mm/year for 2-year intervals.
  • Overall Observations: For GNSS campaigns conducted at 2- or 3-year intervals, velocity accuracies within ±1.5 mm/year are achievable for both components, provided the position accuracies remain below ±5 mm per epoch.
  • The position accuracies of campaigns 1 and 3 had a more pronounced impact on velocity accuracy. However, as the positional accuracy of campaign 1 deteriorated, the influence of campaign 2’s positional accuracy became increasingly significant.
The findings indicate that conducting three campaigns (one epoch per year) with 1-year interval GNSS data is sufficient to achieve a velocity accuracy of less than ±2 mm/year, which is considered the reference threshold for fault slip rate determination studies. However, under the tested conditions, it was not feasible to obtain a velocity accuracy lower than ±1 mm/year with only three campaigns of 1-year interval data. This highlights the importance of developing a tailored observation plan based on fault activity, particularly in newly established seismogeodetic networks. Specifically, three campaigns with a 1-year interval are adequate for Holocene faults with relatively higher slip velocities. In contrast, for Quaternary faults, which generally exhibit lower slip rates, GNSS observations at 2- to 3-year intervals or the inclusion of four or more GNSS observations are necessary to achieve the required accuracy.
The developed ML model enables the analysis of the relationship between position accuracies and velocity accuracies in a unique dataset of 1500 lines produced specifically for the study. With these results obtained for the first time, when establishing a new seismogeodetic network or planning GNSS observations around faults that have experienced destructive earthquakes and completed their post-seismic period, the velocity accuracy can be predicted, and GNSS observation intervals can be optimized based on the required position accuracies, as outlined by [13] or [42]. This allows for more efficient planning of field studies for tectonic-focused GNSS observations, such as extending the measurement period, adding a fourth campaign, or increasing the time between campaigns. Consequently, this approach can help achieve velocity accuracies that enhance the precision of earthquake hazard analysis, ultimately leading to more reliable risk assessments.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Uzel, T.; Eren, K.; Gulal, E.; Tiryakioglu, I.; Dindar, A.A.; Yilmaz, H. Monitoring the tectonic plate movements in Turkey based on the national continuous GNSS network. Arab. J. Geosci. 2013, 6, 3573–3580. [Google Scholar] [CrossRef]
  2. Özkan, A.; Solak, H.I.; Tiryakioğlu, İ.; Şentürk, M.D.; Aktuğ, B.; Gezgin, C.; Poyraz, F.; Duman, H.; Masson, F.; Uslular, G.; et al. Characterization of the co-seismic pattern and slip distribution of the February 06, 2023, Kahramanmaraş (Turkey) earthquakes (Mw 7.7 and Mw 7.6) with a dense GNSS network. Tectonophysics 2023, 866, 230041. [Google Scholar] [CrossRef]
  3. McClusky, S.; Balassanian, S.; Barka, A.; Demir, C.; Ergintav, S.; Georgiev, I.; Gurkan, O.; Hamburger, M.; Hurst, K.J.; Kahle, H.J.; et al. GPS constraints on crustal movements and deformations in the Eastern Mediterranean (1988–1997): Implications for plate dynamics. J. Geophys. Res. 2000, 105, 5695–5719. [Google Scholar] [CrossRef]
  4. Reilinger, R.; McClusky, S.; Vernant, P.; Lawrence, S.; Ergintav, S.; Cakmak, R.; Ozener, H.; Kadirov, F.; Guliev, I.; Stepanyan, R.; et al. GPS constraints on continental deformation in the Africa-Arabia-Eurasia continental collision zone and implications for the dynamics of plate interactions. J. Geophys. Res. Solid Earth 2006, 111, B05411. [Google Scholar] [CrossRef]
  5. Aktug, B.; Nocquet, J.M.; Cingöz, A.; Parsons, B.; Erkan, Y.; England, P.; Lenk, O.; Gürdal, M.A.; Kilicoglu, A.; Akdeniz, H.; et al. Deformation of western Turkey from a combination of permanent and campaign GPS data: Limits to block-like behavior. J. Geophys. Res. Solid Earth 2009, 114. [Google Scholar] [CrossRef]
  6. Yavaşoğlu, H.; Tarı, E.; Tüysüz, O.; Çakır, Z.; Ergintav, S. Determining and modeling tectonic movements along the central part of the North Anatolian Fault (Turkey) using geodetic measurements. J. Geodyn. 2011, 51, 339–343. [Google Scholar] [CrossRef]
  7. Ozener, H.; Dogru, A.; Acar, M. Determination of the displacements along the Tuzla fault (Aegean region-Turkey): Preliminary results from GPS and precise leveling techniques. J. Geodyn. 2013, 67, 13–20. [Google Scholar] [CrossRef]
  8. AKTUĞ, B.; Tiryakioğlu, I.; Sözbilir, H.; Özener, H.; Özkaymak, C.; Yiğit, C.O.; Solak, H.I.; Eyübagil, E.E.; Gelin, B.; Tatar, O.; et al. GPS derived finite source mechanism of the 30 October 2020 Samos earthquake, Mw = 6.9, in the Aegean extensional region. Turk. J. Earth Sci. 2021, 30, 718–737. [Google Scholar] [CrossRef]
  9. Eyübagil, E.E.; Solak, H.İ.; Kavak, U.S.; Tiryakioğlu, İ.; Sözbilir, H.; Aktuğ, B.; Özkaymak, Ç. Present-day strike-slip deformation within the southern part of İzmir Balıkesir Transfer Zone based on GNSS data and implications for seismic hazard assessment, western Anatolia. Turk. J. Earth Sci. 2021, 30, 143–160. [Google Scholar] [CrossRef]
  10. Solak, H.İ.; Tiryakioğlu, İ.; Özkaymak, Ç.; Sözbilir, H.; Aktuğ, B.; Yavaşoğlu, H.H.; Özkan, A. Recent tectonic features of Western Anatolia based on half-space modeling of GNSS Data. Tectonophysics 2024, 872, 230194. [Google Scholar] [CrossRef]
  11. Yilmaz, M.; Gullu, M. A comparative study for the estimation of geodetic point velocity by artificial neural networks. J. Earth Syst. Sci. 2014, 123, 791–808. [Google Scholar] [CrossRef]
  12. Langbein, J. Methods for rapidly estimating velocity precision from GNSS time series in the presence of temporal correlation: A new method and comparison of existing methods. J. Geophys. Res. Solid Earth 2020, 125, e2019JB019132. [Google Scholar] [CrossRef]
  13. Şafak, Ş.; Tiryakioğlu, İ.; Erdoğan, H.; Solak, H.İ.; Aktuğ, B. Determination of parameters affecting the accuracy of GNSS station velocities. Measurement 2020, 164, 108003. [Google Scholar] [CrossRef]
  14. Nocquet, J.M.; Calais, E. Crustal velocity field of western Europe from permanent GPS array solutions, 1996–2001. Geophys. J. Int. 2023, 154, 72–88. [Google Scholar] [CrossRef]
  15. Perez JA, S.; Monico JF, G.; Chaves, J.C. Velocity field estimation using GPS precise point positioning: The south American plate case. J. Glob. Position. Syst. 2023, 2, 90–99. [Google Scholar] [CrossRef]
  16. D’Anastasio, E.; De Martini, P.M.; Selvaggi, G.; Pantosti, D.; Marchioni, A.; Maseroli, R. Short-term vertical velocity field in the Apennines (Italy) revealed by geodetic levelling data. Tectonophysics 2006, 418, 219–234. [Google Scholar] [CrossRef]
  17. Özarpacı, S.; Kılıç, B.; Bayrak, O.C.; Taşkıran, M.; Doğan, U.; Floyd, M. Machine learning approach for GNSS geodetic velocity estimation. GPS Solut. 2024, 28, 65. [Google Scholar] [CrossRef]
  18. Kurt, A.İ.; Özbakir, A.D.; Cingöz, A.; Ergintav, S.; Doğan, U.; Özarpaci, S. Contemporary velocity field for Turkey inferred from combination of a dense network of long term GNSS observations. Turk. J. Earth Sci. 2023, 32, 275–293. [Google Scholar] [CrossRef]
  19. Reilinger, R.; McClusky, S. Nubia–Arabia–Eurasia plate motions and the dynamics of Mediterranean and Middle East tectonics. Geophys. J. Int. 2011, 186, 971–979. [Google Scholar] [CrossRef]
  20. Tiryakioğlu, İ.; Floyd, M.; Erdoğan, S.; Gülal, E.; Ergintav, S.; McClusky, S.; Reilinger, R. GPS constraints on active deformation in the Isparta Angle region of SW Turkey. Geophys. J. Int. 2013, 195, 1455–1463. [Google Scholar] [CrossRef]
  21. Konakoglu, B. Prediction of geodetic point velocity using MLPNN, GRNN, and RBFNN models: A comparative study. Acta Geod. Geophys. 2021, 56, 271–291. [Google Scholar] [CrossRef]
  22. Sorkhabi, O.M.; Alizadeh SM, S.; Shahdost, F.T.; Heravi, H.M. Deep learning of GPS geodetic velocity. J. Asian Earth Sci. X 2022, 7, 100095. [Google Scholar]
  23. Mitchell, T.M. Does machine learning really work? AI Mag. 1997, 18, 11. [Google Scholar]
  24. Flach, P. Machine Learning: The Art and Science of Algorithms That Make Sense of Data; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  25. Griffiths, J. Combined orbits and clocks from IGS second reprocessing. J. Geod. 2019, 93, 177–195. [Google Scholar] [CrossRef] [PubMed]
  26. Petit, G.; Luzum, B. The 2010 reference edition of the IERS conventions. In Reference Frames for Applications in Geosciences; Springer: Berlin/Heidelberg, Germany, 2013; pp. 57–61. [Google Scholar]
  27. Huang, J.; He, X.; Hu, S.; Ming, F. Impact of Offsets on GNSS Time Series Stochastic Noise Properties and Velocity Estimation. Adv. Space Res. 2024; in press. [Google Scholar] [CrossRef]
  28. Herring, T.A.; King, R.W.; Floyd, M.A.; McClusky, S.C. Introduction to GAMIT/GLOBK; Release 10.7; Massachusetts Institute of Technology: Cambridge, MA, USA, 2018; 54p, Available online: http://geoweb.mit.edu/gg/docs/Intro_GG.pdf (accessed on 20 November 2024).
  29. Crocetti, L.; Schartner, M.; Soja, B. Discontinuity detection in GNSS station coordinate time series using machine learning. Remote Sens. 2021, 13, 3906. [Google Scholar] [CrossRef]
  30. Sykes, A.O. An Introduction to Regression Analysis; Chicago Working Paper in Law & Economics; University of Chicago Law School: Chicago, IL, USA, 1993; p. 20. [Google Scholar]
  31. Araghinejad, S. Data-Driven Modeling: Using MATLAB® in Water Resources and Environmental Engineering; Water Science and Technology Library; Springer: Dordrecht, The Netherlands, 2014; Volume 67. [Google Scholar]
  32. Kaya, H.; Gündüz-Öğüdücü, Ş. A distance based time series classification framework. Inf. Syst. 2015, 51, 27–42. [Google Scholar] [CrossRef]
  33. Wang, L. Support Vector Machines: Theory and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2005; Volume 177. [Google Scholar]
  34. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Guo, L.; Chehata, N.; Mallet, C.; Boukir, S. Relevance of airborne lidar and multispectral image data for urban scene classification using random forests. ISPRS J. Photogramm. Remote Sens. 2011, 66, 56–66. [Google Scholar] [CrossRef]
  36. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sánchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  37. Belgiu, M.; Drăguţ, L. Random Forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
  38. Rodriguez-Galiano, V.; Sanchez-Castillo, M.; Chica-Olmo, M.; Chica-Rivas, M. Machine learning predictive models for mineral prospectivity: An evaluation of neural networks, random forest, regression trees and support vector machines. Ore Geol. Rev. 2015, 71, 804–818. [Google Scholar] [CrossRef]
  39. Modaresi, F.; Araghinejad, S.; Ebrahimi, K. A Comparative Assessment of Artificial Neural Network, Generalized Regression Neural Network, Least-Square Support Vector Regression, and K-Nearest Neighbor Regression for Monthly Streamflow Forecasting in Linear and Nonlinear Conditions. Water Resour Manag. 2018, 32, 243–258. [Google Scholar] [CrossRef]
  40. Kazemi, F.; Asgarkhani, N.; Jankowski, R. Optimization-based stacked machine-learning method for seismic probability and risk assessment of reinforced concrete shear walls. Expert Syst. Appl. 2024, 255, 124897. [Google Scholar] [CrossRef]
  41. Kazemi, F.; Asgarkhani, N.; Jankowski, R. Machine learning-based seismic fragility and seismic vulnerability assessment of reinforced concrete structures. Soil Dyn. Earthq. Eng. 2023, 166, 107761. [Google Scholar] [CrossRef]
  42. Eckl, M.C.; Snay, R.A.; Soler, T.; Cline, M.W.; Mader, G.L. Accuracy of GPS-derived relative positions as a function of interstation distance and observing-session duration. J. Geod. 2001, 75, 633–640. [Google Scholar] [CrossRef]
Figure 1. Locations of the stations.
Figure 1. Locations of the stations.
Applsci 15 00113 g001
Figure 2. Time series of ESKS station.
Figure 2. Time series of ESKS station.
Applsci 15 00113 g002
Figure 3. Number of neighbors determined for KNN according to elbow method.
Figure 3. Number of neighbors determined for KNN according to elbow method.
Applsci 15 00113 g003
Figure 4. Reference velocity accuracies and the residuals in RF for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Figure 4. Reference velocity accuracies and the residuals in RF for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Applsci 15 00113 g004
Figure 5. Reference velocity accuracies and the residuals in KNN for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Figure 5. Reference velocity accuracies and the residuals in KNN for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Applsci 15 00113 g005
Figure 6. Reference velocity accuracies and the residuals in SVM for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Figure 6. Reference velocity accuracies and the residuals in SVM for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Applsci 15 00113 g006
Figure 7. Reference velocity accuracies and the residuals in MLR for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Figure 7. Reference velocity accuracies and the residuals in MLR for test dataset (The red line represents the referenced GNSS velocity accuracies, while the blue line represents the residuals).
Applsci 15 00113 g007
Figure 8. Visualization of the test results for the developed models on campaign-based external GNSS datasets.
Figure 8. Visualization of the test results for the developed models on campaign-based external GNSS datasets.
Applsci 15 00113 g008
Figure 9. Prediction of velocity accuracies as a function of position accuracies and temporal intervals.
Figure 9. Prediction of velocity accuracies as a function of position accuracies and temporal intervals.
Applsci 15 00113 g009
Table 1. Properties of variables and dataset.
Table 1. Properties of variables and dataset.
Variable NameVariable TypeDataset NoValue TypeDetails
Year IntervalInput1 and 2IntegerNumber of years between measurements
Se1Input1DecimalPosition accuracy for the E component of the first measurement, independent of the measurement time
Se2Input1Position accuracy for the E component of the second measurement, independent of the measurement time
Se3Input1Position accuracy for the E component of the third measurement, independent of the measurement time
SveOutput1Velocity accuracy for the E component
Sn1Input2Position accuracy for the N component of the first measurement, independent of the measurement time
Sn2Input2Position accuracy for the N component of the second measurement, independent of the measurement time
Sn3Input2Position accuracy for the N component of the third measurement, independent of the measurement time
SvnOutput2Velocity accuracy for the N component
Table 2. Dataset summary.
Table 2. Dataset summary.
VariableCountMeanstdmin.25%50%75%max.
year interval150020.81676911233
Se115003.82.3475461.632.192.794.078.37
Sn115003.454381.4768441.732.433.074.066.49
Se215004.2573672.9652791.592.362.853.8611.2
Sn215003.86961.866191.72.513.214.458.55
Se315003.4113331.9823871.592.08752.53.52257.93
Sn315003.2791.3295251.72.28752.7653.75256.78
SVe15001.528560.9851740.450.84751.271.916.83
SVn15001.5148870.799910.510.91.31.964.83
Table 3. Training–test results for ML algorithms.
Table 3. Training–test results for ML algorithms.
ML AlgorithmsTrain and Test Results
MLRSVMRFKNN
ComponentsENENENEN
Train Score (%)7276929197989494
Test Score (%)7172908695979189
Avg. Train RMSE (mm/year)0.50.40.30.20.20.10.20.2
Avg. Test RMSE (mm/year)0.50.40.30.30.20.10.30.3
Table 4. Test results of the developed models on campaign-based external GNSS datasets.
Table 4. Test results of the developed models on campaign-based external GNSS datasets.
Position Accuracies (mm)Velocity Accuracy (mm/yr)Predicted Velocity Accuracy (mm/yr)
Station
Number
Se1Se2Se3SvMLRSVMRFKNN
East Component12.971.821.821.731.661.611.691.65
23.141.361.711.691.641.591.681.67
32.971.822.431.981.751.781.851.83
43.141.362.191.911.721.701.731.76
52.972.052.532.021.781.831.881.84
62.811.532.341.891.691.681.851.70
73.11.652.752.151.821.901.881.96
83.482.082.642.231.902.022.152.10
92.32.052.531.771.651.651.751.65
North Component13.722.182.052.061.971.972.082.10
23.981.651.912.011.971.932.072.11
33.722.183.022.452.122.322.422.29
43.981.652.652.362.092.172.312.22
53.252.432.792.222.002.102.112.08
63.571.762.882.352.042.162.372.18
74.051.983.062.582.192.412.512.36
84.362.433.22.742.302.642.542.46
92.512.432.791.931.841.841.941.90
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Solak, H.İ. Prediction of GNSS Velocity Accuracies Using Machine Learning Algorithms for Active Fault Slip Rate Determination and Earthquake Hazard Assessment. Appl. Sci. 2025, 15, 113. https://doi.org/10.3390/app15010113

AMA Style

Solak Hİ. Prediction of GNSS Velocity Accuracies Using Machine Learning Algorithms for Active Fault Slip Rate Determination and Earthquake Hazard Assessment. Applied Sciences. 2025; 15(1):113. https://doi.org/10.3390/app15010113

Chicago/Turabian Style

Solak, Halil İbrahim. 2025. "Prediction of GNSS Velocity Accuracies Using Machine Learning Algorithms for Active Fault Slip Rate Determination and Earthquake Hazard Assessment" Applied Sciences 15, no. 1: 113. https://doi.org/10.3390/app15010113

APA Style

Solak, H. İ. (2025). Prediction of GNSS Velocity Accuracies Using Machine Learning Algorithms for Active Fault Slip Rate Determination and Earthquake Hazard Assessment. Applied Sciences, 15(1), 113. https://doi.org/10.3390/app15010113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop