Next Article in Journal
Analysis of the Image Magnification Produced by Inline Holographic Systems Based on the Double-Sideband Filter
Previous Article in Journal
Derivation of the Ultimate Bearing Capacity Formula for Layered Foundations Based on Meyerhof’s Theory
Previous Article in Special Issue
Real-Time 3D Reconstruction of UAV Acquisition System for the Urban Pipe Based on RTAB-Map
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting the External Corrosion Rate of Buried Pipelines Using a Novel Soft Modeling Technique

1
College of Safety Engineering, Chongqing University of Science and Technology, Chongqing 401331, China
2
Chongqing Key Laboratory of Oil and Gas Production Safety and Risk Control, Chongqing 401331, China
3
Chongqing Gas District, PetroChina Southwest Oil and Gasfield Company, Chongqing 400021, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(12), 5120; https://doi.org/10.3390/app14125120
Submission received: 23 April 2024 / Revised: 31 May 2024 / Accepted: 8 June 2024 / Published: 12 June 2024
(This article belongs to the Special Issue Advances in Oil and Gas Storage, Transportation, and Safety)

Abstract

:
External corrosion poses a significant threat to the integrity and lifespan of buried pipelines. Accurate prediction of corrosion rates is important for the safe and efficient transportation of oil and natural gas. However, limited data availability often impacts the performance of conventional predictive models. This study proposes a novel composite modeling approach integrating kernel principal component analysis (KPCA), particle swarm optimization (PSO), and extreme learning machine (ELM). The key innovation lies in using KPCA for reducing the dimensionality of complex input data combined with PSO for optimizing the parameters of the ELM network. The model was rigorously trained on 12 different datasets and comprehensively evaluated using metrics such as the coefficient of determination (R2), standard deviation (SD), mean relative error (MRE), and root mean square error (RMSE). The results show that KPCA effectively extracted four primary components, accounting for 91.33% of the data variability. The KPCA-PSO-ELM composite model outperformed independent models with a higher accuracy, achieving an R2 of 99.59% and an RMSE of only 0.0029%. The model comprehensively considered various indicators under the conditions of limited data. The model significantly improved the prediction accuracy and provides a guarantee for the safety of oil and gas transport.

1. Introduction

The integrity and safety of pipelines play a crucial role in the oil and gas industry, especially given the massive exploitation of fossil fuels [1,2]. Most pipelines are laid underground and are susceptible to corrosion due to long-term contact with soil. Pipeline corrosion will lead to the continuous thinning of the pipeline wall, and perforation and leakage may occur. This leads to safety accidents and threatens the safety of life and property [3]. According to statistics, corrosion factors account for over 50% of pipeline failures [4]. Incidents of pipeline failures due to corrosion occur not only in China but also in other countries. For instance, corrosion damage occurred in 20% of pipelines in the US, resulting in economic losses of USD 203 million [5]. The data suggest pipeline corrosion as a significant safety hazard [6].
While effective pipeline protection methods have been found, ensuring integrity during laying is difficult. Factors like technology and materials make this challenging [7]. Corrosion rates rise with service life. This can cause perforations and leaks before the expected end of life. Such early issues can sharply reduce lifespans. Precisely predicting the extent of transmission pipeline corrosion can lower damage and maintenance costs. It can also minimize environmental impacts. For these reasons, studying pipeline corrosion failure prediction is critical for safe and effective operations [8]. Corrosion prediction models can be divided into mechanistic, statistical, and machine-learning-based intelligent models. Many scholars have extensively studied corrosion mechanisms [9]. Theoretical modeling is a popular method among them. However, the method relies on many mechanism parameters that also have poor replicability in terms of migration. This results in a costly and computationally inefficient model [10]. Machine learning is an important driver of industrial transformation and technological revolution [11]. It has the advantages of fast computation speeds, good model stability, and strong nonlinear data processing ability. Various new intelligent pipeline corrosion prediction methods have emerged, such as gray system theory (GM) [12], BP neural network (BPNN) [13], support vector machine (SVM) [14], and artificial neural network (ANN) [15]. While statistical and probabilistic methods can be used to build corrosion prediction models by taking corrosion data as inputs and outputting patterns of corrosion changes, they face some issues. These methods have problems such as poor generalization ability, narrow prediction scope, severe data dependence, and poor fitting capacity [16,17].
To address these issues, people have proposed various methods to improve the accuracy and generalization of corrosion prediction models. KPCA is used for data dimension reduction to eliminate redundant information between factors and reduce modeling complexity [18]. KPCA is a nonlinear dimensionality reduction technique that effectively extracts the main features of the data [19]. It is able to map them to a low-dimensional space, thus simplifying the model training process and improving generalization [20]. Unlike traditional principal component analysis (PCA), KPCA can handle nonlinear data and better preserve its nonlinear structure [21]. In addition, there are various models and algorithms in the field of corrosion prediction [22]. These include artificial neural network models (such as FF, MLP, and RBF) [23,24,25] and optimization algorithms (such as WOA, Bat, Firefly, and GWO) [26,27,28]. Compared to other optimization algorithms (e.g., WOA, Bat, and Firefly), PSO has faster convergence and global search capability [29,30,31]. It is more suitable for solving corrosion prediction for complex nonlinear data and local optimal problems [32]. ELM is a single-hidden-layer feedforward neural network with the advantages of fast training and a strong generalization ability [33]. Compared with other neural network models (e.g., FF and MLP), ELM is faster to train and has a stronger generalization ability [34,35,36]. It is more suitable for dealing with corrosion prediction scenarios with limited data and high requirements for prediction accuracy [37]. PSO optimizes ELM to establish an external corrosion prediction model for buried pipelines, verified by case analysis on model robustness and superiority [38]. While gray predictive models can be used with limited data, they ignore corrosion-related factors and only model the corrosion prediction process over time. This results in a relatively low prediction accuracy [39]. Intelligent algorithms are useful in modeling regression between corrosion influencing factors and failure outcomes but have some inherent limitations. For example, BPNN has a complex structure, and SVM is too dependent on parameter settings. Some optimization algorithms are slow to converge and are prone to local optimization.
In summary, the development of accurate and robust corrosion prediction models is essential to ensure the safe and efficient operation of pipelines. Different approaches have been proposed to improve the accuracy and generalization ability of the models, but there are still some challenges that need to be resolved.
This paper applies nonlinear kernel principal component analysis (KPCA) to reduce the dimensions of the data. It eliminates redundant information between factors and simplifies the modeling process. Additionally, particle swarm optimization (PSO) is utilized to optimize the extreme learning machine (ELM) and establish an external corrosion prediction model for buried pipelines. The robustness and superiority of the model are confirmed through a case study.

2. Materials and Methods

2.1. KPCA Algorithm

The kernel principal component analysis (KPCA) algorithm improves upon the principal component analysis (PCA) method. It is particularly suitable for dealing with nonlinear problems. The traditional PCA algorithm performs poorly in dealing with nonlinear data [40]. Therefore, scholars combined the PCA algorithm with the kernel function and proposed the KPCA algorithm. KPCA extends PCA’s application from linear domains to nonlinear domains. This measure allows for better handling of nonlinear data and improves the generalization of models [41]. It realizes this function by using nonlinear project functions.
This function inputs the original data into a feature space through a spatial project threshold. In the feature vector space, KPCA then realizes the linear dimensionality reduction in PCA [42]. The KPCA algorithm flow chart is shown in Figure 1.
Consider the input space dataset X = { x 1 , x 2 , , x m } R d and the high-dimensional feature space F , e a non-linear function of Φ ( x ) . Then,
Φ : R d F ; x η = Φ x
If Φ ( x ) = 0 , then the covariance matrix of F can be expressed as:
C F = 1 m j = 1 m Φ x j Φ x j T
Let be the eigenvalue of C F and be the eigenvector of C F , where λ 0 . Then, λ and η satisfy the following condition:
C F η = λ η
Multiplying both sides of Equation (3) by Φ x η , we have the following:
Φ x η · C F η = λ Φ x η · η
According to the kernel reproducing theory, the eigenvalue λ can be linearly represented by Φ x i as follows:
η = i = 1 m α i Φ x i
where α i is the weight of Φ x i .
Substituting Equations (3) and (5) into Equation (4), we obtain the following:
λ Φ x η · i = 1 m α i Φ x i = Φ x η · 1 m j = 1 m Φ x j Φ x j T i = 1 m α i Φ x i
Define an M × M matrix K as follows:
K i η = Φ x i Φ x η
Equation (7) can be simplified as follows:
m λ K α = K 2 α
It is obvious that:
m λ α = K α
By solving Equation (9), we obtain the corresponding eigenvalues and eigenvectors of the high-dimensional feature space F . Consequently, we can derive the k principal component projections of the input sample X in the feature space.
h k x = i = 1 m a i k Φ x i Φ x η
The k principal component vector is obtained by substituting the kernel function for the inner product in the expression.
η k · Φ x = i = 1 m a i k K x i , x
In practice, it is difficult to satisfy the condition Φ x = 0 . Therefore, it is necessary to modify the kernel matrix K to K ¯ , as shown below:
K i η ¯ = K i η 1 m w = 1 m K i w + τ = 1 m K τ η + 1 m 2 w , τ = 1 m w , τ
The implementation process of the KPCA algorithm is as follows:
  • Step 1: Process the original data samples X for the external corrosion prediction of a buried pipeline, and generate an M × N data matrix A .
    Step 2: Select the Gaussian function as the kernel function with preset parameters, and calculate the kernel matrix K according to K w l = K ( x w , x l ) = Φ ( x w ) T Φ ( x l ) .
    Step 3: Modify the kernel function to obtain a new kernel matrix K ¯ .
    Step 4: Calculate the eigenvalue λ 1 , λ 2 , , λ m and eigenvector η 1 , η 2 ,   η m of K ¯ .
    Step 5: Arrange the eigenvalues in descending order to obtain the sequence λ 1 * > λ 2 * > > λ m * and adjust the corresponding eigenvectors to obtain η 1 η 1 * , η 2 * , ,   η m * .
    Step 6: Appropriately orthogonalize the eigenvectors to obtain a set of orthogonal principal component vectors X .
    Step 7: Calculate the cumulative contribution of λ 1 λ 1 * , λ 2 * , , λ m * and set the extraction rate p = 85%. When p t 85 % , take the first t components as the principal components, whose corresponding principal component is α 1 , α 2 , , α t .
    Step 8: Calculate the projection Y of the sample X on α 1 , α 2 , , α t , Y = K ¯ α . Y is the reduced dimension data after KPCA feature extraction.

2.2. PSO Algorithm

The particle swarm optimization (PSO) algorithm is an intelligent optimization algorithm inspired by the group predatory nature of organisms [43]. The technique considers the mismatch between particles in the swarm and their previous error paths. Through iterative minimization of the cost function, PSO can solve the optimization problem. PSO exhibits flexibility due to its parameterization, allowing for the customization of velocity and acceleration update rules [44]. Different variants of this algorithm can be obtained by modifying the way the particle velocities and accelerations are described. This approach provides optimal solutions based on threshold conditions [45,46,47]. The PSO algorithm flow chart is shown in Figure 2.
Consider a swarm of particles (containing N particles) flying and searching simultaneously in a D-dimensional space. The attributes of each particle at different times are defined as follows, where 1 d D , 1 i N :
x i d t = x i 1 t , x i 2 t , , x i d t T , x i d t L d , U d
V i d t = V i 1 t , V i 2 t , , V i d t T , V i d t V d m i n , V d m a x
P i t = P i 1 t , P i 2 t , , P i D t T
P g t = P g 1 t , P g 2 t , , P g D t T
In this context, Equation (13) represents the position of the i particle at the t iteration, where L d and U d are the lower and upper bounds of the D-dimensional space, respectively, and t denotes the current iteration number. The particle’s velocity can be calculated using Equation (14), where V d m i n and V d m a x represent the lower and upper bounds of the velocity in the d dimension. Equation (15) represents the best position found by the i particle during its search process, while Equation (16) represents the global best position found by the entire swarm.
Equations (17) and (18) represent the position and velocity of the particle at the t + 1 iteration, respectively.
x i d t + 1 = x i d 1 + v i d t + 1
v i d t + 1 = v i d t + c 1 r 1 p i d t x i d t + c 2 r 2 p g d t x i d t
In these equations, v i d t represents the velocity of the i particle in the d dimension at the t iteration; x i d t represents the position of the i particle in the d dimension at the t iteration; r 1 and r 2 are random numbers uniformly distributed in the range 0 ,   1 ; c 1 and c 2 are acceleration coefficients, which are typically set to c 1 = c 2 = 2 based on empirical experience; p i d t represents the individual best position of the i particle in the d dimension at the t iteration; and p g d t represents the global best position in the d dimension.
The implementation process of the PSO algorithm is as follows:
  • Step 1: Set the parameters and population size and initialize the velocity and position of all particles in the population.
    Step 2: Calculate and evaluate the fitness of each particle in the population.
    Step 3: Update the current best position ( p i b ). Compare the fitness value of the particle in Step 2 with the best historical position. Update to the current optimal position if it is better; otherwise, leave it unchanged.
    Step 4: Update the best global position ( p g ). Compare the fitness value of the particle in Step 2 with the best historical global position. The global optimum is updated using the current value if it is better than the historical optimum; otherwise, it remains unchanged.
    Step 5: Update the velocity and position of individual particles.
    Step 6: Judge whether the algorithm is over or not. If the end condition is satisfied, the algorithm ends; in other cases, it returns to Step 2.

2.3. ELM Algorithm

The extreme learning machine (ELM) algorithm is a new algorithm proposed by researcher Guangbin Huang in 2004 to improve the single-hidden-layer feedforward neural network. The ELM algorithm has advantages for problems involving multi-classification, high dimensionality, and predictions [48].
The ELM model has the same network structure as the single-hidden-layer feedforward neural network (SLFN) [49]. ELM uses random input layer weights and biases only in the training phase. The output layer weights are then computed by the generalized inverse matrix theory [50]. In this model, m represents the number of nodes in the input layer, M represents the number of nodes in the hidden layer, and n represents the number of nodes in the output layer. The activation function required for the network connection during training in the hidden layer is represented by g(x), and b i represents the threshold value within a neuron [51]. With N different sets of samples ( x i , t i ) , where 1 i N , the ELM network structural model can be depicted as shown in Figure 3.
Where: x i = x i 1 , x i 2 , x i 3 , , x i m , T ϵ R m , t i = t i 1 , t i 2 , t i 3 , , t i n , T ϵ R n
Mathematical representation of ELM model structure:
i = 1 M β i g ( ω i x i + b i ) = o j , j = 1 , 2 , , N
where ω i = [ ω i 1 , ω i 2 , ω i 3 , , ω i m ]  represents the input weight vector of the input layer node and the i hidden layer node in the network model, β i = [ β i 1 , β i 2 , β i 3 , , β i n ] T  represents the output weight vector of the i hidden layer node and the output layer node in the network model, and o i = [ o i 1 , o i 2 , o i 3 , , o i n ] T represents the network model output value.
The cost function of the ELM model is given by Equation (20):
E ( S , β ) = j = 1 N o j t j
where S = ( ω i , b i , i = 1 , 2 , , M ) contains the network model input weights and thresholds for the hidden layer nodes. The purpose of ELM model learning is to find the optimal S , β , such that the error between the model’s predicted structure and the measured values is minimized and noted as m i n ( E ( S , β ) ) .
The expression for m i n ( E ( S , β ) ) can be further written as per Equation (21):
m i n ( E ( S , β ) ) = min ω i , b i , β H ( ω 1 , , ω M , b 1 , , b M , x 1 , , x N ) β T
where H is the output matrix of the hidden layer in the prediction model, β is the weight matrix of the output in the prediction model, and T is the target value matrix in the prediction model, and H , β , T are defined as follows:
H ( ω 1 , , ω M , b 1 , , b M , x 1 , , x N ) = [ g ( ω 1 x 1 + b 1 ) g ( ω M x 1 + b M ) g ( ω 1 x N + b 1 ) g ( ω m x N + b M ) ] N × M
β = [ β 1 T β M T ] M × N ,   T = [ t 1 T t N T ] N × N
The ELM predictive model is a process of solving for the ideal solution in nonlinear variation. The essence of the training process is to solve for the least squares solution, as shown in Equation (24):
β ^ = H + T
where H + is the inverse of the generalization of the matrix H .

2.4. KPCA-PSO-ELM Combination Modeling Process

This paper selects KPCA for reducing the feature dimensionality. It utilizes neural networks to solve nonlinear problems. PSO is applied to optimize the ELM calculation for predicting the external corrosion rate of buried pipelines. The detailed steps of the KPCA-PSO-ELM algorithm are presented in Figure 4.
  • Step 1: Pretreat the raw data and introduce it into the KPCA algorithm for data dimensionality reduction.
    Step 2: Conduct random division of the training and test sets in proportion to the data normalization process.
    Step 3: Construct the ELM prediction model network structure.
    Step 4: Set the particle population and the number of particles.
    Step 5: Limit the maximum number of iterations, and preset learning factors c 1 and c 2 .
    Step 6: Calculate the fitness value of the particles.
    Step 7: Determine the number of runs corresponding to when the required error is reached. Stop running when the optimal accuracy is reached, and go to Step 9 or return to Step 8. Continue running until the requirement is met.
    Step 8: Update the particle velocity and position to execute the iteration.
    Step 9: Bring the globally optimal weights and thresholds into the ELM prediction model for training.
    Step 10: Predict the corrosion rate test set.

2.5. Indicators for Model Evaluation

This research presents a comprehensive comparison of the predictive effectiveness of the KPCA-PSO-ELM model. To achieve a more accurate and comprehensive analysis, multiple error metrics are utilized, including MRE, RMSE, and R2.
The error metrics are formulated as follows:
MRE = 1 n i = 1 n | y i y ^ i | y i × 100 %
MSE = 1 n i = 1 n ( y i y ^ i ) 2
R 2 = 1 i = 1 n ( y ^ i y i ) 2 i = 1 n ( y i ¯ y i ) 2
In this study, the actual value is denoted as y i , while the predicted value is represented by y ^ i . The MRE and RMSE indicators take values in the range of [ 0 , 1 ] , where a value closer to 0 indicates a lesser error and better prediction performance. In addition, the value of R2 ranges from 0 to 1 , with a value closer to 1 indicating a better fit of the model to the data and superior performance.

2.6. Sample Selection and Data Analysis

This work analyzed the status of a buried pipeline in Chongqing, China. It collected relevant data to understand the environmental conditions around the pipeline. This included the physical properties of the soil (water content, soil density, mechanical composition of the soil), chemical properties (salt content, pH, Cl content, SO42− content), electrochemical properties (resistivity, redox potential), and stray currents. Ten main control factors related to the external corrosion of buried pipelines were selected and analyzed based on the survey results, as shown in Figure 5.
The original dataset was obtained from the field survey, and 60 datasets were selected for the small-sample predictive analysis. These data were selected for the small-sample prediction analysis. For the analysis, the data were divided into training and prediction sets at a ratio of 4:1. V / mm a represents the corrosion rate. More details are provided in Appendix A.
The distribution of the parameter samples was analyzed. The data turned out to be complex, with irregular factors and non-normal distributions. As shown in Figure 6, the corrosion rate histogram presented an inhomogeneous shape. In order to analyze the degree of correlation between the corrosion rate and corrosion factors, we used the Pearson correlation coefficient to measure their correlation [52].
The coefficient ranges from −1 to 1, with a larger absolute value indicating a stronger correlation. Negative values indicate an inverse relationship between the main index and the features; in contrast, a positive value indicates a direct relationship. The operational principle is illustrated in Equation (28).
r = i = 1 n X i X ¯ Y i Y ¯ i = 1 n X i X ¯ 2 Y i Y ¯ 2
As shown in Figure 7, corrosion rate was moderately correlated with the stray current, sulphate content, chloride content, and redox potential. In order to better predict the corrosion rate, it is necessary to extract the characteristics of 10 external corrosion influences in buried pipelines. On the one hand, this can effectively reduce the data dimension and reduce the spatial complexity. On the other hand, it can simplify the model and reduce the time complexity.

2.7. KPCA Algorithm for Feature Extraction

In this study, the feature extraction was based on actual survey data and employed the Gaussian Radial Basis Function (RBF) as the kernel function. MATLAB R2020a programming was used to determine the component contribution and calculate the cumulative contribution. Based on this analysis, factors accounting for 85% of the total variation were retained as principal components. The correlation results are shown in Figure 8.
The first 4 principal components were selected to replace the 10 corrosion factor evaluation indexes in the original data for analysis. This ensured the orthogonal independence of the feature vectors for each principal component. As shown in Table 1, the analysis of the KPCA algorithm shows that the cumulative contribution of the first 4 principal components was 91.33%. Its cumulative contribution exceeded the threshold of 85%. This was better than the 3 principal components (90.66%) in the journal article [53].
The approach reduced the original 60 10 dimensional data to 60 4 dimensional indicator data processed by KPCA. This ensured the orthogonal independence of the feature vectors for each principal component. This simplified the model and reduced the time complexity. It still accurately predicted the external corrosion rate of the buried pipelines.

2.8. PSO-ELM Model Simulation Parameter Environment Settings

The PSO-ELM model simulation environment in this work is presented in Table 2, and the PSO algorithm parameters were set as illustrated in Table 3.
To determine the optimal ELM model settings for maximum prediction accuracy, the activation function and number of nodes in the hidden layer were investigated. With the increasing number of hidden nodes, the prediction accuracy gradually increased. The prediction accuracy stabilized at 15 nodes with the sigmoid activation function. A comparison of the sigmoid and ReLU functions found that the sigmoid function achieved a higher accuracy. Based on these results, ELM used 15 hidden nodes and optimized the sigmoid activation by PSO.

3. Results

3.1. PSO-ELM Model Training Results

In the established prediction process, the four principal components obtained by the KPCA dimensionality reduction were applied as the training inputs in the PSO-ELM model. The network training data were normalized using the mapminmax() function of MATLAB. The inverse of this normalization was used, and the training was performed using the train() function. Multiple training iterations were performed while varying the number of neurons in the hidden layer of the ELM. By comparing the results, the optimal number of neurons in the hidden layer was determined to be 15. This configuration improved the performance of the PSO-ELM model for corrosion rate prediction.
For this case, the ELM network structure was 4-15-1. The training obtained the best fitness value of 0.03856, with an R2 value of 99.59%. The iterative error variation curve of the PSO-optimized ELM parameters is shown in Figure 9.
Figure 9 shows that 17 iterations were required to achieve convergence at the minimum best adaptation value, indicating the best prediction performance. The PSO-optimized results were passed to the ELM algorithm to determine the input weights and hidden layer thresholds. It was then trained to obtain the optimized ELM model.
The KPCA-PSO-ELM model used 12 randomly selected test data samples for the corrosion rate prediction. The results are shown in Table 4. As can be seen from the predictions in Figure 10, the KPCA-PSO-ELM model predictions were very close to the original data samples.
Figure 11 shows the relative error curves between the measured values and the predicted values from the KPCA-PSO-ELM model for the 12 test samples. The first prediction had the smallest relative error of 1.94%. The eighth prediction had the largest relative error of 16.07%, but it was still within the acceptable range of 20% error. Therefore, the KPCA-PSO-ELM model has good generalization ability, high prediction accuracy, and ease of use. It can be effectively used for the prediction of corrosion rates of buried pipelines.

3.2. Comparative Analysis of Forecast Results

To facilitate a more intuitive comparison of the ELM prediction performance, this article employed three models: a conventional ELM model trained on raw data, an ELM model trained after reducing the dimensionality of the raw data, and a KPCA-PSO-ELM model trained using an intelligent optimization algorithm. Each model was separately trained and predicted using the same simulation environment parameters. Their prediction results were then compared and analyzed. Figure 12 and Table 5 present the statistical comparisons of the prediction results for each model type, drawn from table data. Figure 13 compares the standard deviation of the predicted data with the actual values for the different models. Table 6 calculates the prediction error comparisons.
A correlation analysis was carried out to compare the predicted corrosion rates from the different ELM models (ELM, KPCA-ELM, PSO-ELM, KPCA-PSO-ELM) with the actual corrosion rates. The R2 value indicates the degree of linear correlation between the predicted values of the measured model and the actual values. The results are shown in Figure 14 The black line represents the formula Y = X. The red line represents the fitted prediction result line. The R2 values for ELM, KPCA-ELM, PSO-ELM, and KPCA-PSO-ELM were 0.8867, 0.9435, 0.9462, and 0.9959, respectively.
The KPCA-PSO-ELM model had the highest R2 value. This indicates that the model fits very well. However, since the R2 value exceeded 99%, there may be a risk of overfitting. But the R2 value does not reflect the accuracy of the model’s predicted values. Therefore, we also need to consider other indicators, such as the standard deviation (SD) and mean square error (MSE) or root mean square error (RMSE), to evaluate the generalization ability of the model. The generalization ability refers to the model’s ability to predict unknown data.
By comparing the standard deviations between the different models, KPCA-ELM was closest to the standard deviation of the actual values followed by KPCA-PSO-ELM. The comparison found that the MRE and RMSE values for the KPCA-ELM and PSO-ELM models were lower than the traditional ELM model, showing significant performance improvements.
Compared with the average absolute error (3.24%) of the PSO-SVM model in the literature [14], the average absolute error of the KPCA-PSO-ELM model was only 0.475%, and the R² value was also improved. These results fully demonstrate the effectiveness of the selected optimization method and indicate that the KPCA-PSO-ELM model has significant advantages in terms of prediction accuracy.
Figure 15 illustrates the Taylor plot, presenting the performance of the different models in predicting the test set data. It combines three metrics: the standard deviation (SD), R2, and RMSE. The closer the prediction model is to the actual internal purple box point, the better the prediction results. The dashed circle at the origin represents the standard deviation value. The solid line centered at the origin indicates the correlation coefficient.
From the above figure, the KPCA-PSO-ELM model was closest to the bottom center purple box point compared to the other models. It had an R2 value of 0.9959 and an SD of 0.745. These results indicate that the KPCA-PSO-ELM model was the most suitable for predicting the external corrosion rates in this dataset. It demonstrated a high level of accuracy in predicting the corrosion rates of the buried pipelines.

4. Discussion

The accurate prediction of the external corrosion rates of buried pipelines is important in ensuring pipeline integrity and safe management. Nevertheless, the models developed currently are often limited by small sample sizes, and the accuracy is not as satisfactory as expected. To address these issues, this paper proposes a kernel principal component analysis (KPCA)–particle swarm optimization (PSO)–extreme learning machine (ELM) combination model for the prediction of the external corrosion rate of buried pipelines under conditions of scarce data.
This study applies the KPCA algorithm to lower the dimensions of influencing factors and then extract principal components, ultimately making the model structure more concise. The PSO algorithm is utilized to optimize the parameters of the ELM model, including the number of hidden layer nodes. A total of 60 sets of field monitoring data containing 10 influencing factors were collected and divided into training and validation sets at a ratio of 4:1.
When predicting the validation set, the R2 of the proposed model reached 99.59%, and the standard deviation was close to the standard deviation of the actual values, with an average relative error of only 6.394% and a root mean square error of only 0.0029%. Compared with the ELM, KPCA-ELM, and PSO-ELM models, the proposed model performed well in all indicators. It demonstrated an assessment under the composite indicator. The proposed model had a stronger learning and generalization ability even in the case of small sample data. Despite the valuable findings, it is important to acknowledge several limitations in this article. Firstly, the external applicability of the model requires further verification due to the single source of sample data. Secondly, the selection of model parameters such as the number of principal components lacks theoretical guidance. Lastly, the risk assessment of the prediction results, such as the risk of overfitting of the model, was not considered. In the future, there is a need to continuously increase the sample size to improve the external applicability of the model on new data. More advanced automatic hyperparameter optimization algorithms should be developed to improve the generalized learning capability of the model. Quantitative analysis of the uncertainty in the prediction results should be conducted to improve the model’s reliability. Different methods to combat the overfitting problem should be explored and compared, such as data augmentation, regularization, model selection, early stopping, or transfer learning.

5. Conclusions

This research proposes a kernel principal component analysis (KPCA)–particle swarm optimization (PSO)–extreme learning machine (ELM) model to predict the external corrosion rates in buried pipelines, particularly under data-limited conditions. The results underscore several key advantages:
(1)
KPCA is effective in extracting principal components. Four principal components containing 91.33% of the original information were extracted, reducing the input dimensions and simplifying the model.
(2)
PSO optimized the parameters of the ELM model to improve the prediction accuracy, and the KPCA-PSO-ELM model achieved excellent prediction results on 60 datasets, with an MAE of 0.475% and an MRE of 6.394%, and the standard deviation of its prediction data was also close to the actual values. In contrast, the R2 of the ELM model was 88.67%, with relatively low prediction accuracy. This indicates data degradation and global parameter optimization can effectively improve the prediction accuracy.
(3)
The optimized ELM model outshined the other models in generalization on the validation set, underscoring its significant utility in pipeline corrosion monitoring and early warning.
In summary, this paper introduces an advanced, small-sample-appropriate model for external corrosion rate prediction. The synergistic integration of KPCA and PSO augments both its accuracy and generalization capabilities, promising substantial applications in pipeline safety management. Future efforts aim to further validate the model’s effectiveness across varied datasets and enhance its predictive robustness.

Author Contributions

Conceptualization, Z.R. and D.Y.; methodology, Z.R. and D.Y.; software, Z.R. and D.Y.; validation, D.Y. and Z.W. and K.C.; formal analysis, K.C. and D.Y. and W.Q.; investigation, K.C. and D.Y. and Z.R.; resources, K.C. and Z.W. and W.Q.; data curation, W.Q. and D.Y.; writing—original draft preparation, Z.R. and D.Y.; writing—review and editing, Z.R. and D.Y.; visualization, Z.R. and D.Y. and W.Q.; supervision, K.C. and Z.W. and W.Q.; project administration, K.C. and Z.W.; funding acquisition, K.C. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Chongqing (Grant numbers: CSTB2023NSCQ-LZX0036 and CSTB2022NSCQ-MSX0051) and the Scientific and Technology Research Program of Chongqing Municipal Education Commission (Award number: KJZD-K201901501).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study can be found in the main article and the Appendix A.

Conflicts of Interest

Author Wei Qin was employed by the company Chongqing Gas District, Petro China Southwest Oil and Gasfield Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

Table A1. External corrosion inspection data for buried pipelines.
Table A1. External corrosion inspection data for buried pipelines.
No.W1W2W3W4W5W6W7W8W9W10V
1121.41254.670.0150.0220.05619.08.980.230.38−396.00.136
224.13153.480.0450.0100.02415.87.940.250.62−596.90.209
3152.29112.770.0240.0060.03820.77.860.210.45−332.20.105
474.54126.650.0140.0520.04716.78.170.180.96−798.10.220
519.53348.980.0270.0250.04217.14.580.230.65−425.70.220
6147.5685.120.0530.0360.04120.45.930.080.73−549.80.099
730.80239.480.0250.0290.05930.04.160.141.19−248.80.082
849.21114.380.0450.0450.05520.57.310.190.46−716.70.093
956.48123.270.0470.0400.02616.76.770.130.62−341.40.177
1057.6098.400.0290.0120.05627.66.690.040.61−544.70.130
1147.03155.100.0100.0350.03725.46.520.200.96−479.10.205
12153.14155.820.0330.0130.03523.58.390.130.60−588.10.043
1353.24329.980.0540.0380.02817.56.440.210.71−552.40.066
14111.45318.970.0110.0030.05018.44.900.130.90−545.40.233
1538.67138.300.0180.0350.04621.84.120.221.39−688.20.225
16140.55338.300.0150.0030.03723.17.510.250.36−496.50.094
1762.14212.450.0240.0280.02623.15.420.291.12−289.40.232
18122.51336.470.0330.0270.03622.17.900.271.30−578.30.166
19177.60279.870.0200.0260.05116.28.510.241.09−320.10.171
20133.65220.120.0500.0150.04918.97.350.020.43−452.10.079
21178.46292.920.0130.0330.05018.36.750.210.44−348.40.098
22125.60163.190.0320.0150.02926.97.500.281.13−596.60.038
23139.01244.340.0070.0200.02516.26.420.090.83−588.10.045
24155.49276.540.0470.0190.03918.38.050.070.49−631.90.159
2539.73117.300.0350.0380.04116.58.850.050.86−332.40.035
2624.22225.510.0430.0380.02815.88.500.231.24−397.90.214
278.10208.960.0510.0470.04925.77.430.240.73−793.00.084
28163.49274.010.0130.0140.03722.04.640.050.94−702.00.131
29126.22230.150.0290.0090.05017.07.690.030.92−443.50.147
3021.49237.820.0090.0310.05716.35.450.140.28−642.80.176
3158.37341.420.0300.0220.03729.04.080.061.31−508.90.232
32196.34103.480.0460.0460.03918.98.650.030.39−250.60.026
33162.2391.840.0510.0460.05624.08.290.290.99−603.60.042
3494.8597.670.0280.0280.05722.76.470.051.13−340.70.067
3568.57342.890.0280.0270.03429.26.110.041.31−355.80.128
3619.31147.530.0200.0420.04528.48.290.100.56−638.80.038
37175.50156.560.0380.0440.03515.24.480.171.09−275.00.124
3848.43341.470.0340.0040.03217.66.000.181.31−604.20.151
39142.96141.450.0480.0290.02017.96.780.271.40−736.90.124
4052.5892.520.0160.0220.04324.95.240.091.37−705.60.080
4150.96300.800.0380.0280.02225.36.300.030.44−422.10.053
4298.24192.470.0120.0240.05016.74.280.031.36−470.90.186
436.11264.400.0070.0400.02527.95.200.271.33−657.50.162
44157.07295.580.0210.0250.04815.28.010.200.41−778.60.161
4591.51184.350.0290.0310.04522.24.530.181.39−745.20.021
4660.22185.100.0170.0080.02427.27.450.031.31−269.50.220
4714.29185.470.0300.0050.03428.67.370.041.03−655.10.193
48168.38103.710.0480.0030.05523.34.930.180.47−442.30.155
4911.18105.860.0210.0500.02119.85.250.280.68−204.30.037
50140.58191.660.0100.0220.02917.85.240.180.43−265.40.231
5196.64170.420.0360.0350.02616.34.990.180.22−506.40.088
5280.38127.000.0200.0120.03626.58.410.150.22−790.70.043
5362.10253.720.0270.0090.02225.77.600.231.04−520.70.184
54199.69256.180.0250.0310.04318.94.040.140.46−506.00.054
5597.27168.740.0330.0280.03718.64.790.260.53−407.20.039
56124.87149.060.0470.0220.04216.67.730.291.03−776.00.070
57187.19215.240.0300.0290.02618.56.470.151.39−500.80.215
58147.84186.290.0440.0500.04720.36.490.180.54−430.20.127
5995.44298.070.0300.0200.02616.44.260.130.48−709.70.088
6012.46129.440.0510.0290.02625.88.630.180.96−438.00.154

References

  1. Shaik, N.B.; Benjapolakul, W.; Pedapati, S.R.; Bingi, K.; Le, N.T.; Asdornwised, W.; Chaitusaney, S. Recurrent neural network-based model for estimating the life condition of a dry gas pipeline. Process Saf. Environ. Prot. 2022, 164, 639–650. [Google Scholar] [CrossRef]
  2. Thakur, A.K.; Arya, A.K.; Sharma, P. The science of alternating current-induced corrosion: A review of literature on pipeline corrosion induced due to high-voltage alternating current transmission pipelines. Corros. Rev. 2020, 38, 463–472. [Google Scholar] [CrossRef]
  3. Shin, S.; Lee, G.; Ahmed, U.; Lee, Y.; Na, J.; Han, C. Risk-based underground pipeline safety management considering corrosion effect. J. Hazard. Mater. 2018, 342, 279–289. [Google Scholar] [CrossRef] [PubMed]
  4. Askari, M.; Aliofkhazraei, M.; Afroukhteh, S. A comprehensive review on internal corrosion and cracking of oil and gas pipelines. J. Nat. Gas Sci. Eng. 2019, 71, 102971. [Google Scholar] [CrossRef]
  5. El-Abbasy, M.S.; Zayed, T.; Mirahadi, F.; Parvizsedghy, L.; Senouci, A. Optimized maintenance plan for oil and gas pipelines. Can. J. Civ. Eng. 2022, 49, 1151–1162. [Google Scholar] [CrossRef]
  6. Dao, U.; Yarveisy, R.; Anwar, S.; Khan, F.; Zhang, Y.; Ngo, H.H. A Bayesian approach to assess under-deposit corrosion in oil and gas pipelines. Process Saf. Environ. Prot. 2023, 176, 489–505. [Google Scholar] [CrossRef]
  7. Farh, H.M.H.; Seghier, M.E.A.B.; Zayed, T. A comprehensive review of corrosion protection and control techniques for metallic pipelines. Eng. Fail. Anal. 2023, 143, 106885. [Google Scholar] [CrossRef]
  8. Zhang, J.; Lian, Z.; Zhou, Z.; Song, Z.; Liu, M.; Yang, K.; Liu, Z. Safety and reliability assessment of external corrosion defects assessment of buried pipelines—Soil interface: A mechanisms and FE study. J. Loss Prev. Process Ind. 2023, 82, 105006. [Google Scholar] [CrossRef]
  9. Chen, X.; Li, C.; Ming, N.; He, C. Effects of temperature on the corrosion behaviour of X70 steel in CO2-Containing formation water. J. Nat. Gas Sci. Eng. 2021, 88, 103815. [Google Scholar] [CrossRef]
  10. Singh, M.; Markeset, T.; Kumar, U. Some philosophical issues in modeling corrosion of oil and gas pipelines. Int. J. Syst. Assur. Eng. Manag. 2014, 5, 55–74. [Google Scholar] [CrossRef]
  11. Du, J.; Zheng, J.; Liang, Y.; Xu, N.; Liao, Q.; Wang, B.; Zhang, H. Deeppipe: Theory-guided prediction method based automatic machine learning for maximum pitting corrosion depth of oil and gas pipeline. Chem. Eng. Sci. 2023, 278, 118927. [Google Scholar] [CrossRef]
  12. Zeng, B.; Ma, X.; Shi, J. Modeling method of the grey GM (1, 1) model with interval grey action quantity and its application. Complexity 2020, 2020, 6514236. [Google Scholar] [CrossRef]
  13. Zhang, Y.-g.; Tang, J.; Liao, R.-p.; Zhang, M.-f.; Zhang, Y.; Wang, X.-m.; Su, Z.-y. Application of an enhanced BP neural network model with water cycle algorithm on landslide prediction. Stoch. Environ. Res. Risk Assess. 2021, 35, 1273–1291. [Google Scholar] [CrossRef]
  14. Wang, C.; Ma, G.; Li, J.; Dai, Z.; Liu, J. Prediction of corrosion rate of submarine oil and gas pipelines based on ia-svm model. IOP Conf. Ser. Earth Environ. Sci. 2019, 242, 022023. [Google Scholar] [CrossRef]
  15. Kumari, P.; Wang, Q.; Khan, F.; Kwon, J.S.-I. A unified causation prediction model for aboveground onshore oil and refined product pipeline incidents using artificial neural network. Chem. Eng. Res. Des. 2022, 187, 529–540. [Google Scholar] [CrossRef]
  16. Wang, Q.; Song, Y.; Zhang, X.; Dong, L.; Xi, Y.; Zeng, D.; Liu, Q.; Zhang, H.; Zhang, Z.; Yan, R. Evolution of corrosion prediction models for oil and gas pipelines: From empirical-driven to data-driven. Eng. Fail. Anal. 2023, 146, 107097. [Google Scholar] [CrossRef]
  17. Li, J.; Liu, Z.; Yi, H.; Liu, G.; Tian, Y. Stray current prediction model for buried gas pipelines based on multiple regression models and extreme learning machine. Int. J. Electrochem. Sci. 2021, 16, 210253. [Google Scholar] [CrossRef]
  18. Song, L.; Wang, Y.; Zhao, B.; Liu, Y.; Mei, L.; Luo, J.; Zuo, Z.; Yi, J.; Guo, X. Research on prediction of ammonia concentration in QPSO-RBF cattle house based on KPCA nuclear principal component analysis. Procedia Comput. Sci. 2021, 188, 103–113. [Google Scholar] [CrossRef]
  19. Zeng, D.; Wang, J.; Fan, M.; Yue, X.; Hou, Y. Parameter optimization of parallel mechanisms based on PCA. China Mech. Eng. 2017, 28, 2899. [Google Scholar]
  20. Anowar, F.; Sadaoui, S.; Selim, B. Conceptual and empirical comparison of dimensionality reduction algorithms (pca, kpca, lda, mds, svd, lle, isomap, le, ica, t-sne). Comput. Sci. Rev. 2021, 40, 100378. [Google Scholar] [CrossRef]
  21. Cheng, C.-Y.; Hsu, C.-C.; Chen, M.-C. Adaptive kernel principal component analysis (KPCA) for monitoring small disturbances of nonlinear processes. Ind. Eng. Chem. Res. 2010, 49, 2254–2262. [Google Scholar] [CrossRef]
  22. Qiu, S.; Chen, B.; Wang, R.; Zhu, Z.; Wang, Y.; Qiu, X. Atmospheric dispersion prediction and source estimation of hazardous gas using artificial neural network, particle swarm optimization and expectation maximization. Atmos. Environ. 2018, 178, 158–163. [Google Scholar] [CrossRef]
  23. Jitchaijaroen, W.; Keawsawasvong, S.; Wipulanusat, W.; Kumar, D.R.; Jamsawang, P.; Sunkpho, J. Machine learning approaches for stability prediction of rectangular tunnels in natural clays based on MLP and RBF neural networks. Intell. Syst. Appl. 2024, 21, 200329. [Google Scholar] [CrossRef]
  24. Boštík, J.; Kukal, J. ANN WHICH COVERS MLP AND RBF. Sign 4, e1. Available online: https://www2.humusoft.cz/www/papers/tcp11/023_bostik.pdf (accessed on 21 March 2024).
  25. Hussain, M.; Zhang, T.; Chaudhry, M.; Jamil, I.; Kausar, S.; Hussain, I. Review of prediction of stress corrosion cracking in gas pipelines using machine learning. Machines 2024, 12, 42. [Google Scholar] [CrossRef]
  26. Goel, L. An extensive review of computational intelligence-based optimization algorithms: Trends and applications. Soft Comput. 2020, 24, 16519–16549. [Google Scholar] [CrossRef]
  27. Mohammed, H.M.; Umar, S.U.; Rashid, T.A. A systematic and meta-analysis survey of whale optimization algorithm. Comput. Intell. Neurosci. 2019, 2019, 8718571. [Google Scholar] [CrossRef] [PubMed]
  28. Yarat, S.; Senan, S.; Orman, Z. A comparative study on PSO with other metaheuristic methods. In Applying Particle Swarm Optimization: New Solutions and Cases for Optimized Portfolios; Springer: Cham, Switzerland, 2021; pp. 49–72. [Google Scholar]
  29. Wang, D.; Tan, D.; Liu, L. Particle swarm optimization algorithm: An overview. Soft Comput. 2018, 22, 387–408. [Google Scholar] [CrossRef]
  30. Yang, X.-S. Firefly algorithms for multimodal optimization. In Proceedings of the International Symposium on Stochastic Algorithms, Sapporo, Japan, 26–28 October 2009; pp. 169–178. [Google Scholar]
  31. Mirjalili, S.; Mirjalili, S.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  32. Eberhart, R.; Kennedy, J. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  33. Xinsheng, Z.; Yingge, C. Prediction of external corrosion rate of offshore oil and gas pipelines based on FA-BAS-ELM. China Saf. Sci. J. 2022, 32, 99. [Google Scholar]
  34. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice Hall PTR: Hoboken, NJ, USA, 1998. [Google Scholar]
  35. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  36. Al-Daoud, E. A Comparison Between Three Neural Network Models for. J. Artif. Intell. 2009, 2, 56–64. [Google Scholar] [CrossRef]
  37. Wang, J.; Lu, S.; Wang, S.-H.; Zhang, Y.-D. A review on extreme learning machine. Multimed. Tools Appl. 2022, 81, 41611–41660. [Google Scholar] [CrossRef]
  38. Xia, Y.; Yi, W.; Zhang, D. Coupled extreme learning machine and particle swarm optimization variant for projectile aerodynamic identification. Eng. Appl. Artif. Intell. 2022, 114, 105100. [Google Scholar] [CrossRef]
  39. Li, S.; Du, H.; Cui, Q.; Liu, P.; Ma, X.; Wang, H. Pipeline Corrosion Prediction Using the Grey Model and Artificial Bee Colony Algorithm. Axioms 2022, 11, 289. [Google Scholar] [CrossRef]
  40. Cao, L.; Chua, K.S.; Chong, W.K.; Lee, H.P.; Gu, Q. A comparison of PCA, KPCA and ICA for dimensionality reduction in support vector machine. Neurocomputing 2003, 55, 321–336. [Google Scholar] [CrossRef]
  41. Fan, Z.; Wang, J.; Xu, B.; Tang, P. An efficient KPCA algorithm based on feature correlation evaluation. Neural Comput. Appl. 2014, 24, 1795–1806. [Google Scholar] [CrossRef]
  42. Wang, H.; Peng, M.-j.; Yu, Y.; Saeed, H.; Hao, C.-m.; Liu, Y.-k. Fault identification and diagnosis based on KPCA and similarity clustering for nuclear power plants. Ann. Nucl. Energy 2021, 150, 107786. [Google Scholar] [CrossRef]
  43. Han, B.; Li, B.; Qin, C. A novel hybrid particle swarm optimization with marine predators. Swarm Evol. Comput. 2023, 83, 101375. [Google Scholar] [CrossRef]
  44. Pace, F.; Santilano, A.; Godio, A. A review of geophysical modeling based on particle swarm optimization. Surv. Geophys. 2021, 42, 505–549. [Google Scholar] [CrossRef] [PubMed]
  45. Li, X.-L.; Serra, R.; Olivier, J. A multi-component PSO algorithm with leader learning mechanism for structural damage detection. Appl. Soft Comput. 2022, 116, 108315. [Google Scholar] [CrossRef]
  46. Garcia-Gonzalo, E.; Fernandez-Martinez, J.L. A brief historical review of particle swarm optimization (PSO). J. Bioinform. Intell. Control 2012, 1, 3–16. [Google Scholar] [CrossRef]
  47. Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
  48. Tang, S.; Zhu, C.; Cui, G.; Xing, X.; Mu, J.; Li, Z. Analysis of internal corrosion of supercritical CO2 pipeline. Corros. Rev. 2021, 39, 219–241. [Google Scholar] [CrossRef]
  49. Sevinç, E. The Effect of Hidden Neurons in Single-Hidden Layer Feedforward Neural Networks. Bilişim Teknol. Derg. 2019, 12, 277–286. [Google Scholar] [CrossRef]
  50. Huynh, H.T.; Won, Y.; Kim, J.-j. An improvement of extreme learning machine for compact single-hidden-layer feedforward neural networks. Int. J. Neural Syst. 2008, 18, 433–441. [Google Scholar] [CrossRef]
  51. Yang, L.; Tsang, E.C.; Wang, X.; Zhang, C. ELM parameter estimation in view of maximum likelihood. Neurocomputing 2023, 557, 126704. [Google Scholar] [CrossRef]
  52. Saccenti, E.; Hendriks, M.H.; Smilde, A.K. Corruption of the Pearson correlation coefficient by measurement error and its estimation, bias, and correction under different error models. Sci. Rep. 2020, 10, 438. [Google Scholar] [CrossRef]
  53. Zhe, N.; Yang, J.F.; Liu, W.B.; Chen, L.C. Prediction of corrosion rate of process pipeline based on KPCA and SVM. Corros. Prot. 2019, 40, 56–60. [Google Scholar]
Figure 1. Flowchart of the KPCA algorithm.
Figure 1. Flowchart of the KPCA algorithm.
Applsci 14 05120 g001
Figure 2. Flowchart of the PSO algorithm.
Figure 2. Flowchart of the PSO algorithm.
Applsci 14 05120 g002
Figure 3. Schematic of the ELM network structural model.
Figure 3. Schematic of the ELM network structural model.
Applsci 14 05120 g003
Figure 4. Flowchart of the KPCA-PSO-ELM model.
Figure 4. Flowchart of the KPCA-PSO-ELM model.
Applsci 14 05120 g004
Figure 5. External corrosion influencing factors of buried pipelines.
Figure 5. External corrosion influencing factors of buried pipelines.
Applsci 14 05120 g005
Figure 6. Histogram of corrosion rate.
Figure 6. Histogram of corrosion rate.
Applsci 14 05120 g006
Figure 7. Internal correlation matrix of corrosion data.
Figure 7. Internal correlation matrix of corrosion data.
Applsci 14 05120 g007
Figure 8. Cumulative contribution of the main elements.
Figure 8. Cumulative contribution of the main elements.
Applsci 14 05120 g008
Figure 9. Iterative error variation curves of PSO-optimized ELM parameters.
Figure 9. Iterative error variation curves of PSO-optimized ELM parameters.
Applsci 14 05120 g009
Figure 10. Model prediction results.
Figure 10. Model prediction results.
Applsci 14 05120 g010
Figure 11. Test set relative error.
Figure 11. Test set relative error.
Applsci 14 05120 g011
Figure 12. Comparison chart of model prediction results.
Figure 12. Comparison chart of model prediction results.
Applsci 14 05120 g012
Figure 13. Standard deviation for different models predicting the test set data.
Figure 13. Standard deviation for different models predicting the test set data.
Applsci 14 05120 g013
Figure 14. Linear fit of the prediction results of the four models. (a) ELM forecast value, (b) KPCA-ELM forecast value, (c) PSO-ELM forecast value, and (d) KPCA-PSO-ELM forecast value.
Figure 14. Linear fit of the prediction results of the four models. (a) ELM forecast value, (b) KPCA-ELM forecast value, (c) PSO-ELM forecast value, and (d) KPCA-PSO-ELM forecast value.
Applsci 14 05120 g014
Figure 15. Taylor diagram for different models predicting the test set data.
Figure 15. Taylor diagram for different models predicting the test set data.
Applsci 14 05120 g015
Table 1. Results of KPCA analysis.
Table 1. Results of KPCA analysis.
Principal ComponentEigenvalueContribution RateCumulative Contribution Rate
11.32039.72%39.72%
20.75122.60%62.32%
30.68920.73%83.06%
40.2758.28%91.33%
50.0932.80%94.13%
60.0742.23%96.36%
70.0461.38%97.74%
80.0371.11%98.86%
90.0250.75%99.61%
100.0130.39%100.00%
Table 2. PSO-ELM model simulation environment configuration.
Table 2. PSO-ELM model simulation environment configuration.
ParameterSpecification
Processor (CPU)Intel(R) Core (TM) i5-10300H @2.50 GHz
Device Manufacturer Name: Intel Corporation
City: Santa Clara
Country: United States
System Type64-bit operating system based on x64 processor
Operating System (OS)Windows 10 Home Chinese version
Programming SoftwareMATLAB R2020a
Table 3. PSO algorithm parameters.
Table 3. PSO algorithm parameters.
ParameterValue
Population size30
Maximum iterations50
Maximum   particle   velocity   ( V m a x )2
Minimum   particle   velocity   ( V m i n )−2
Acceleration   factor   1   ( c 1 )2
Acceleration   factor   2   ( c 2 )2
Maximum   inertia   weight   ( ω m a x )1.2
Minimum   inertia   weight   ( ω m i n )0.4
Maximum   particle   position ( x m a x )1
Minimum   particle   position   ( x m i n )−1
Table 4. Prediction results of KPCA-PSO-ELM model.
Table 4. Prediction results of KPCA-PSO-ELM model.
No.Measured Value/mm·a−1KPCA-PSO-ELM Forecast Value/mm·a−1Relative Error (%)
10.1030.101−1.94
20.0760.08410.53
30.2400.235−2.08
40.0710.0767.04
50.0580.0626.90
60.0190.0205.26
70.2040.200−1.96
80.0560.06516.07
90.1630.1714.91
100.0340.0352.94
110.0310.03512.90
120.1430.1494.20
Table 5. Model prediction error statistics.
Table 5. Model prediction error statistics.
No.Corrosion
Rate
/mm·a−1
ELMKPCA-ELMPSO-ELMKPCA-PSO-ELM
Forecast Value
/mm·a−1
Relative Error
/%
Forecast Value
/mm·a−1
Relative Error
/%
Forecast Value
/mm·a−1
Relative Error
/%
Forecast Value
/mm·a−1
Relative Error
/%
10.1030.11511.990.11310.040.1129.060.101−1.94
20.0760.070−8.140.0804.980.08410.230.08410.53
30.2400.2629.040.230−4.280.230−4.280.235−2.08
40.0710.0789.300.07910.700.08113.500.0767.04
50.0580.0627.310.05210.000.0615.580.0626.90
60.0190.01427.410.02319.260.0218.890.0205.26
70.2040.2155.470.2186.940.191−6.300.200−1.96
80.0560.06211.280.052−6.670.06822.050.06516.07
90.1630.1746.700.1693.630.1714.860.1714.91
100.0340.03913.540.04325.190.0364.810.0352.94
110.0310.04130.660.03614.730.03821.100.03512.90
120.1430.17019.000.1505.000.1494.200.1494.20
Table 6. Comparison of error comprehensive evaluation indicators.
Table 6. Comparison of error comprehensive evaluation indicators.
Predictive ModelMRE (%)RMSE (%)R2
ELM13.3700.015788.67%
KPCA-ELM10.5800.006194.35%
PSO-ELM9.8820.006994.62%
KPCA-PSO-ELM6.3940.002999.59%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ren, Z.; Chen, K.; Yang, D.; Wang, Z.; Qin, W. Predicting the External Corrosion Rate of Buried Pipelines Using a Novel Soft Modeling Technique. Appl. Sci. 2024, 14, 5120. https://doi.org/10.3390/app14125120

AMA Style

Ren Z, Chen K, Yang D, Wang Z, Qin W. Predicting the External Corrosion Rate of Buried Pipelines Using a Novel Soft Modeling Technique. Applied Sciences. 2024; 14(12):5120. https://doi.org/10.3390/app14125120

Chicago/Turabian Style

Ren, Zebei, Kun Chen, Dongdong Yang, Zhixing Wang, and Wei Qin. 2024. "Predicting the External Corrosion Rate of Buried Pipelines Using a Novel Soft Modeling Technique" Applied Sciences 14, no. 12: 5120. https://doi.org/10.3390/app14125120

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop