Next Article in Journal
Performance Analysis of Polymer Additive Manufactured Gear Bearings
Previous Article in Journal
Evaluation Model of Holographic Communication Experience in the 6G Era Based on Light Field Display
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cementing Quality Prediction in the Shunbei Block Based on Genetic Algorithm and Support Vector Regression

1
Hubei Provincial Key Laboratory of Oil and Gas Drilling and Production Engineering, Wuhan 430100, China
2
National Engineering Research Center for Oil & Gas Drilling and Completion Technology, School of Petroleum Engineering, Yangtze University, Wuhan 430100, China
3
Cooperative Innovation Center of Unconventional Oil and Gas, Yangtze University (Ministry of Education & Hubei Province), Wuhan 430100, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(22), 12382; https://doi.org/10.3390/app132212382
Submission received: 7 October 2023 / Revised: 7 November 2023 / Accepted: 8 November 2023 / Published: 16 November 2023

Abstract

:
There are a number of factors that can affect the quality of cementing, and they constrain each other. Current cementing quality prediction methods are still in the stage of development, and it is difficult to establish an analytical model for cementing quality prediction that meets the strict requirements of cementing design. In order to accurately predict the cementing quality in the Shunbei block of the Northwest Oilfield, in this study, we established a cementing quality prediction model based on support vector regression (SVR) and optimized the penalty parameter and kernel parameter by using grid search (GS), a Bayesian optimization algorithm (BOA), and a genetic algorithm (GA), which improve the prediction accuracy of SVR. The results show that the smallest root-mean-square error and average relative error (2.318% and 7.30%, respectively) and the highest accuracy are achieved when using GA–SVR as compared to SVR, GS–SVR, and BOA–SVR. Therefore, GA–SVR is suitable for cementing quality prediction in the Shunbei block.

1. Introduction

Cementing is an important and indispensable part of drilling and completion operations. The purpose of cementing is to protect and support the casing in oil and gas wells and to seal off oil, gas, and water formations. The quality of cementing has an important impact on the production of oil and gas wells. If the quality of cementing is low, it may lead to a decline in oil and gas well production, shorten the service life, and even cause accidents. Therefore, improving the quality of cementing is a very important task in drilling engineering. By predicting the quality of cementing, it is possible to optimize measures to put new wells into production, improve the quality of cementing, and prolong the service life of oil, gas, and water wells. Therefore, many domestic and foreign experts are attempting to improve the quality of cementing [1,2]. There are many factors affecting the quality of cementing, such as the geological environment, construction conditions, and material properties. The mutual constraints and uncertainties among these factors increase the difficulty of implementing traditional cementing quality analysis and even affect the analysis results. Therefore, it is very difficult to establish a cementing quality prediction model under multifactorial conditions. Scholars at home and abroad are using different methods to analyze the correlation between the main influencing factors and initially establish some cementing quality evaluation methods [3,4]. However, there are some shortcomings in the accuracy and stability of these methods; thus, further research is necessary. Nezhad et al. proposed a convolutional neural network (CNN) approach to evaluate the performance of neural networks in automatic interpretation and combined it with fuzzy systems [5]. Viggen et al. proposed an assisted cement log interpretation tool based on supervised ML, and the implemented tool, which can be used for cementing quality evaluation, allows the interpretation of logging results to be automated [6]. Fang Chunfei et al. proposed a multi-scale perceptual convolutional neural network with kernels of different sizes, which is suitable for recognizing logging variable density images and evaluating cementing quality [7]. Souza et al. combined an experimental setup simulating several well conditions with machine learning as a diagnostic tool to demonstrate that support vectors are more suitable than other vectors for cementing quality evaluation [8]. Kinoshita et al. investigated a method of evaluating the quality of cementing with an acoustic quantitative model [9,10]. Sun Zhicheng et al. used gray relational theory for multifactorial statistics on cementing quality [11].
An analysis shows that current cementing quality evaluation methods mainly rely on amplitude logging, whereas cementing quality prediction mainly relies on fuzzy evaluation theory and neural network theory to establish relevant models [12]. For the current situation, the existing cementing quality prediction methods do not satisfy the requirements of cementing design. In recent years, support vector regression (SVR) has proven to be a reliable machine learning algorithm that can overcome the many shortcomings of artificial neural networks; thus, it has been widely applied in many scientific and engineering fields [13,14]. In this paper, a method for cementing quality prediction based on a genetic support vector regression machine is proposed. Based on the cementing historical data of the Shunbei block in the Northwest Oilfield, a cementing quality prediction model is established and optimized. This method has high prediction accuracy and can be used to predict the cementing quality in the Shunbei block.

2. Background

2.1. Analysis of Cementing Quality Influencing Factors

Located in the hinterland of the Taklamakan Desert, the Northwest Oilfield is the basin with the largest remaining oil and gas resources in China. The Northwest Oilfield is characterized by a special geographic location, abundant oil and gas resources, and leading technology for drilling ultra-deep wells. The Shunbei block in the Northwest Oilfield is a carbonate formation that is structurally characterized by fractured formations, large well slopes, and leaky conditions. The Shunbei block of the Northwest Oilfield has deep wells, a long cement section, a high bottomhole temperature, and a small annular space gap. In the process of cementing, the annular space can be easily blocked, resulting in downhole leakage and well collapse. In addition, the horizontal casing center is difficult to set, and the displacement is limited, so it is difficult to ensure the quality of cementing. According to the existing literature, the main factors affecting the quality of cementing include the inclination angle, total angle change rate, borehole diameter enlargement rate, casing centers, grouting return rate, mud displacement return rate, and cement slurry contact time. When the inclination angle is large, the casing tends not to be centered, and when the total angle change rate is large, the casing tends to stick to the edge of the casing, which makes it difficult to drive away chips and mud effectively. It also tends to flow, which affects the cementing quality. In addition, the grout reflux rate, mud replacement reflux rate, and cement slurry contact time are key factors affecting replacement efficiency. In particular, they directly affect the efficiency of the replacement of cementing fluid with cement slurry, which in turn affects the quality of cementing. The borehole expansion rate directly affects the grout reflux rate and mud replacement rate, which indirectly affects the cementing quality.

2.2. Relevant Works

Many scholars have applied machine learning algorithms to the field of cementing quality prediction. For example, cementing quality prediction can be performed using a multi-scale perceptual convolutional neural network model. It is a deep learning model; specifically, it can utilize convolutional kernels of different scales to perform convolutional operations on the feature maps obtained at a certain moment to obtain new feature maps of different scales and then later up-sample the feature maps of different scales according to the scales of the input feature maps. In addition to this, there are other types. Deepak Kumar Voleti et al. examined the efficacy of random forest classification and neural network models in cement evaluation and proposed the use of nested models that have low bias, improving interpretation accuracy and cement evaluation in onshore wells; however, the work did not address the issues of model generalizability, data dependency, or potential overfitting [15]. Ni Hongmei et al. introduced a particle swarm optimization algorithm based on stochastic global optimization in neural network training. This method was used to improve defects, which include the slow convergence speed of the traditional BP network in cementing quality prediction. This work reduced the training time and improved the accuracy of cementing quality prediction, but it was not time-effective [16].

2.3. SVM Algorithm

The main idea of SVMs (support vector machines) is to minimize the structural risk and VC dimensionality theory in statistical learning [17,18]. The SVM algorithm can be used to solve small sample problems and nonlinear regression models, which can overcome the problems of “dimensionality catastrophe” and “overlearning” to a large extent [19,20].
For example, taking two types of data classification and letting the sample set be x 1 , y 1 , x 2 , y 2 , , x i , y i , the optimal interface linear equation for the sample space can be expressed as follows:
ω T x + b = 0
where ω denotes the normal vector and b denotes the displacement term [21,22].
ωTx: This is the inner product (dot product) operation, which multiplies the weight vector ω and the eigenvector x and adds their corresponding elements. The result of this part of the computation represents a linear combination between the hyperplane and the eigenvector x. The result of this part of the computation represents a linear combination between the hyperplane and the eigenvector x.
= 0: This is the equational part of the equation. The regression hyperplane is defined as the hyperplane that satisfies this equation. In SVR, the goal of this equation is to make the hyperplane fit the training data points as closely as possible, and data points within the ε-tube are considered to fall on the hyperplane, with ε being a user-defined parameter.
The goal of SVR is to minimize the error term and find a regression hyperplane such that the sum of its residuals with respect to the data points is as small as possible. This is why the SVR model uses this linear equation to describe the regression hyperplane, but by adjusting the weight vector ω and the bias term b it can fit nonlinear data relationships.
In summary, ωTx + b = 0 is the equation used in the SVR model to describe the regression hyperplane, where the weight vector ω and the bias term b are determined by training the model to fit the training data and perform the regression.
To ensure that the hyperplane (ω, b) can classify the training sample normally, the following should be satisfied:
ω T x i + b + 1 , y i = + 1 ω T x i + b 1 , y i = 1
The exact derivation of the algorithm is described in Appendix A.
Essentially, the SVR in the SVM is used to establish the margin of error, and predictions that fall within the margin of error are considered to be correct, which is different from the principle of the SVM model (maximum spacing between sample points closest to the hyperplane). The purpose of the SVR is to minimize the distance between the farthest sample point to the hyperplane.
Support vectors can be categorized into two types, depicted as the first kind and second kind in Figure 1. The first kind of support vectors are data points that lie within the ε-tube, which is a user-defined width parameter used to determine the threshold of prediction error allowed by the model. The second kind of support vectors are data points that lie outside the ε-tube, i.e., those that are more than ε away from the regression hyperplane. It is important to help regularize the model, avoid overfitting, and maintain the complexity of the model.
In summary, the first kind of support vectors and the second kind of support vectors represent different data points in SVR, and they are important for both model fitting and generalization performance. The first type of support vectors are used to construct regression hyperplanes, while the second type of support vectors help regularize the model and control the complexity of the model.

3. Data and Methods

3.1. Datasets

In this model training, 48 sections of historical cementing data of the Shunbei block were used, and each section included cementing quality (CBL acoustic amplitude) and its influencing factors: inclination angle, overall angle change rate, borehole diameter enlargement rate, casing centralizer, grouting return velocity, mud replacement return velocity, and cement slurry contact time. Various factors interacted, inhibited, or depended on each other, and an accurate cementing quality prediction model was obtained by substituting the model for post-training evaluation.
Here are histograms of the relevant parameters, as shown in Figure 2.

3.2. Data Preprocessing

The max–min method in the standardized method was used to process 48 datasets to prevent excessively large data dimensions, which can affect the prediction results. The max–min method can be expressed as Equation (3):
x i = x i x min x max x min
Overall, 40 groups were randomly selected from the processed data and set as training data; the remaining 8 groups were used as prediction data. To verify the accuracy of the model, the root-mean-square error (RMSE) and average relative error (MRE) for the prediction data were calculated. The calculation formulas for RMSE and MRE are shown in Equation (4) and Equation (5), respectively:
R M S E = i = 1 n p i o i 2 n
M R E = 1 n i = 1 n o i p i o i
where Oi denotes the true value of cementing quality, pi denotes the prediction value of the cementing quality, and n denotes the number of the predicted data group.

3.3. Model Training

A cementing quality prediction model was established based on SVR. The inclination angle, overall angle change rate, borehole diameter enlargement rate, casing centralizer, grouting return velocity, mud replacement return velocity, and cement slurry contact time are the input values; cementing quality is the output value. The processed data were used by the SVR model to build a model, in which the penalty coefficient takes 1, and the kernel function g takes 1/n. Table 1 shows the divided test set as well as the RMSE and MRE values for the prediction results.
From Table 1, it is evident that that the MRE between the SVR predicted cementing quality and actual cementing quality is roughly 9.80%. Accordingly, the SVR parameters must be optimized to increase the accuracy of the model.
As can be seen in Table 1, there is a large difference between entries 14 and 17. This is due to the fact that a lack of cementing data samples from a certain rare and specific environment can lead to a larger relative error.

3.4. Model Optimization

For the SVR algorithm, the penalty parameter and kernel parameter both directly affect the prediction error. Researchers have studied SVR regression and put forward some suggestions, but many contradictory opinions still exist. In this study, to reduce the SVR prediction error, the grid search (GS) method [24], Bayesian optimization algorithm (BOA) [25,26,27], and genetic algorithm (GA) [28,29] were used to optimize the penalty and kernel parameters, after which the optimized SVR algorithm was used to train and predict the relevant data. In this study, RBF was specified as the kernel function in Python, the scope of the penalty coefficient was [0.1, 100], and the range of the kernel parameter was [0.01, 1].

3.4.1. Grid Search Method to Optimize the Support Vector Regression Model

The GS method is one of the easiest methods in the global optimization algorithm, which is a learning method to optimize the model performance by traversing the given parameter combinations. Its principle is to determine the search range of the model, separate the possible values of the parameters to be searched in accordance with a certain step length, and generate a “grid”. Then, it searches for each node of the grid one by one, uses the objective function to make judgments, and adjusts the search range and the step length according to the value of the objective function until it reaches the optimal solution.
The optimization flow chart of the GS method is shown in Figure 3.
From Table 2, it is evident that the MRE between GS–SVR and the actual cementing quality is 8.46%, and the RMSE is 2.653, which suggests a certain optimization effect.

3.4.2. Bayesian Optimization Algorithm for the Support Vector Regression Model

The BOA [30,31] involves testing each sample point according to prior knowledge and updating prior distributions to identify the global optimal value.
Assume that a sample S is present, including t parameter groups and corresponding objective function value. Firstly, the posterior probability distribution Dt (namely, joint normal distribution) is calculated by the Gaussian process, and then a function a(x|Dt) is selected to collect a new parameter value xt+1 so as to satisfy Equation (6):
x = arg max E max 0 , f t + 1 x f b e s t D t
Thus, in the distribution, Dt is able to make the expected difference between the new estimate objective function value ft+1(x) and current maximum objective function value fbest the largest, the new parameter is used to calculate the new actual objective function value and update the corresponding Gaussian process, and a new posterior distribution is obtained. The above process is repeated until the stop condition is satisfied.
The optimization flow chart of the BOA is shown in Figure 4.
From Table 3, it is evident that the MRE between the predicted cementing quality of BOA–SVR and the actual cementing quality is 8.00%, and the RMSE is 2.533, which suggests good optimization.

3.4.3. Genetic Algorithm Optimization for the Support Vector Regression Model

The GA [32,33] is an efficient stochastic search and optimization method based on natural genetic mechanisms and biological evolution. The GA exhibits a global optimization performance and can find the most appropriate values of the penalty and kernel function parameters for the SVM, by which the SVM can achieve the optimal configuration to accurately predict cementing quality.
The optimization flow chart of the GA method is shown in Figure 5.
The GA–SVR code is run in Python, and the optimization results and model evaluation are shown in Table 4, from which it is evident that the prediction accuracy of the SVR model constructed after GA optimization is higher than that of SVR, GS–SVR, and BOA–SVR. The MSE is 7.30%, and the RMSE is 2.318, which suggests that the optimization effect is the best.

3.5. Analysis of Cementing Quality Prediction Results

In this study, a cementing quality prediction algorithm was established based on SVR, and the penalty parameter and kernel function parameter were optimized. The parameter optimization process was carried out via Python programming, and the GA, GS, and BOA optimization methods were adopted. The prediction results of the four algorithms are shown in Table 5.
From Table 5 and Figure 6 and Figure 7, it is evident that the MRE of GA–SVR is 7.30%, and the RMSE is 2.318, both of which are far smaller compared with SVR, GS–SVR, and BOA–SVR, indicating that this algorithm is the most suitable for cementing quality prediction. Table 5 indicates that GA–SVR is the least time-effective method. This suggests that, compared with SVR and BOA–SVR, it is more complex and takes more comprehensive factors into consideration. Meanwhile, there are nonlinearities and uncertainties in the influencing factors of cementing quality; while GA–SVR performs better in dealing with high-dimensional, multimodal, and nonlinear problems, GS–SVR is more suitable for dealing with large-scale datasets, and BOA–SVR is suitable for the problems that need to find the global optimal solution. Based on this property, it is more appropriate to use the GA–SVR algorithm.

4. Results and Discussion

Pearson correlation, which is a method that measures the degree of correlation between two variables X and Y, was adopted to verify the example. The Pearson correlation coefficient between X and Y is defined as the quotient of the covariance and standard deviation of the two variables, as shown by Equation (7). The Pearson correlation coefficient ranges between 1 and 1, where 1 suggests complete positive correlation of variables, 0 indicates irrelevance, and 1 means complete negative correlation [34].
ρ X , Y = cov ( X , Y ) σ X σ Y = E [ ( X μ X ) ( Y μ Y ) ] σ X σ Y
Equation (7) defines the population correlation coefficient and estimates the sample covariance and standard deviation. The calculation method is expressed by Equation (8).
r = i = 1 n ( X i X ¯ ) ( Y i Y ¯ ) i = 1 n ( X i X ¯ ) 2 i = 1 n ( Y i Y ¯ ) 2
The completion depth of a well in the Shunbei block of the Northwest Oilfield is 8725 m, the maximum deviation angle is 86.01°, the maximum angle change rate is 0.6127°/m, the maximum diameter expansion rate is 12%, the mud density is 1.68 g/cm3, and the plastic viscosity is 36 mPa · s. For the 139.7 mm liner cementing of the well, the centralizer cannot be placed in case the straight casing is upset. The casing is close to the lower shaft wall, and the casing is basically zero in the medium range. The displacement flushing fluid (1.05 g/cm3) is designed to be 7 m3, the weighted isolation fluid (1.75 g/cm3) is 4 m3, the cement slurry (1.88 g/cm3) is 11 m3, and the designed injection displacement is 0.6–0.8 m3/min. The sound amplitude logging of the 139.7 mm liner well cementing is shown in Figure 8, from which it is evident that the sound amplitude of the CBL logging in the whole sealing section is between 10% and 30%, and the cementing quality for 7850–8000 m is excellent.
Using GA–SVR to predict the cementing quality of the cementing data, the comparison between the predicted values and the actual values is shown in Figure 9, from which it can be seen that Pearson’s correlation coefficient is 0.9781; in addition, the predicted values of GA–SVR have a good correlation with the actual cementing quality values.
In order to further analyze the deviation between the predicted values of the GA–SVR model and the actual cementing quality values, the relative errors of the prediction results were analyzed, and the results are shown in Figure 10, from which it can be seen that the sixth data group has the largest relative error of 9.16%. The first data group has the smallest relative error of 1.27%. The prediction errors of most subgroups are concentrated around 5.5%, with MRE of 5.27% and RMSE of 1.166. The results show that the predicted values of GA–SVR match the actual cementing quality values to the greatest extent and have the smallest errors, which indicates that this method is suitable for predicting the cementing quality of the Shunbei block. GA–SVR is a complex model, which will always have some uncertainty in predicting future situations. For example, many factors, such as the environment, can affect the prediction results. In addition, there is randomness and noise in the data, which will have an impact on the prediction results. This method is more suitable for situations where the environment is similar, and it may not be able to adapt quickly to changes in the data when the conditions differ.

5. Conclusions

In this study, a cementing quality prediction model was established based on SVR. In the process of using SVR, the penalty coefficient and kernel coefficient have an important impact on data learning and prediction results, so the model was optimized using GS, BOA, and GA, and the cementing quality was predicted using SVR, GS–SVR, BOA–SVR, and GA–SVR in combination with the historical cementing data of a block in Shunbei. The prediction results show that the MRE and RMSE of GA–SVR are relatively small. This proves that the algorithm has high prediction accuracy and can be used to predict cementing quality in the Shunbei block. In addition, this method is suitable for the case of similar environments, and, when the environments differ, this method cannot adapt to the data changes quickly, so we plan to use incremental learning for the next optimization in the future.

Author Contributions

Data curation, J.W.; project administration, S.Z.; software, J.H.; validation, K.B.; writing the original preparation, J.W.; writing review J.W. and editing, S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Open Foundation of Cooperative Innovation Center of Unconventional Oil and Gas, Yangtze University (Ministry of Education and Hubei Province), grant number No. UOG2022-06, the scientific research project of the Hubei Provincial Department of Education, grant number D20201304, and the open fund of Hubei Key Laboratory of Drilling and Production Engineering for Oil and Gas (Yangtze University).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in this article. The data are not publicly available due to private reasons.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The decision function can be obtained as y x = sgn ω T x + b . To maximize the interval, it can be reduced to the following quadratic programming problem:
max 2 ω ω , b s . t . y i ω T x i + b 1 i = 1 , 2 , , m
Then, the language function can be introduced:
L ω , b , α = 1 2 ω + i = 1 m α i 1 y i ω T x i + b
Then, the dual problem can be identified:
max ω i = 1 m α i 1 2 i = 1 m j = 1 m α i α j y i y j x T x j
Then, α can be determined, and, by solving ω and b, the model can be obtained:
f x = ω T x + b = i = 1 m α i y i x i T x + b
where αi satisfies the Karush–Kuhn–Tucker condition:
α i 0 y i f x i 1 0 α i y i f x i 1 = 0
Because some of the points appear at the interval of the two lines and do not satisfy the constraint condition of the function interval, a slack variable is introduced to represent the degree to which the sample does not satisfy the constraint:
min ω , b , ε i 1 2 ω 2 + C i = 1 m ε i s . t . y i ω T x i + b 1 ε i ε i 0 , i = 1 , 2 , , m
where εi denotes the slack variable and C denotes the penalty factor (C > 0).
In the case of linear inseparability, the sample can be mapped to a higher dimensional eigenspace, so that the sample is linearly separable in said eigenspace. The model corresponding to the optimal hyperplane divided in the feature space can be expressed as follows:
f x = ω T ϕ x + b = i = 1 m α i y i ϕ x i T ϕ x + b
Finally, a kernel function is introduced to avoid the dimensional disaster in Equation (A7). Common kernel functions include linear, polynomial, and Gaussian functions (also known as the radial basis function, RBF). Accordingly, the SVM nonlinear regression model can be expressed as follows:
f x = i = 1 m α i y i k x , x i + b

References

  1. Gottschling, J.C. HZ Marcellus Well Cementing in Appalachia; Society of Petroleum Engineers: Richardson, TX, USA, 2009. [Google Scholar] [CrossRef]
  2. Ravi, K.; Weber, L. Drill-Cutting Removal in a Horizontal Wellbore for Cementing; Society of Petroleum Engineers: Richardson, TX, USA, 1996. [Google Scholar] [CrossRef]
  3. Zhao, M.; Ren, J.; Ji, L.; Fu, C.; Li, J.; Zhou, M. Parameter selection of support vector machines and genetic algorithm based on change area search. Neural Comput. Appl. 2012, 21, 1–8. [Google Scholar] [CrossRef]
  4. Chun, Z.; Ting, W.T.; Chi, A.; Dan, Y.J. The Study and Application of Gray-Orthogonal Wavelet Network Forecasting Mode on Cementing Quality. In Proceedings of the 2010 International Conference on E-Business and E-Government, Guangzhou, China, 7–9 May 2010; pp. 5327–5330. [Google Scholar] [CrossRef]
  5. Saleh Nezhad, M.; Akbarizadeh, G.; Seifi Kavian, Y.; Salehi-Moorkani, R. Automatic Interpretation of Oil and Gas Well Cement Evaluation Logs Using Fuzzy Convolutional Neural Networks. SPE Drill. Complet. 2023, 38, 535–545. [Google Scholar] [CrossRef]
  6. Viggen, E.M.; Singstad, B.J.; Time, E.; Mishra, S.; Berg, E. Assisted Cement Log Interpretation Using Machine Learning. SPE Drill. Complet. 2023, 38, 220–234. [Google Scholar] [CrossRef]
  7. Fang, C.; Wang, Z.; Song, X.; Zhu, Z.; Yang, D.; Liu, M. A Novel Cementing Quality Evaluation Method Based on Convolutional Neural Network. Appl. Sci. 2022, 12, 10997. [Google Scholar] [CrossRef]
  8. de Souza, L.P.B.; Ferreira, G.R.B.; Camerini, I.G.; de Magalhães Correia, T.; de Castro Ribeiro, M.G.; Hidalgo, J.A.S.; de São João, B.L.D.; Llerena, R.W.A.; Kubrusly, A.C.; Ayala, H.V.H.; et al. Machine learning-based cement integrity evaluation with a through-tubing logging experimental setup. Geoenergy Sci. Eng. 2023, 227, 211882. [Google Scholar] [CrossRef]
  9. Kinoshita, T.; Izuhara, W.; Valero, H.-P.; Blyth, M. Feasibility and Challenge of Quantitative Cement Evaluation with LWD Sonic. In Proceedings of the SPE Annual Technical Conference and Exhibition, New Orleans, LA, USA, 30 September–2 October 2013. [Google Scholar] [CrossRef]
  10. Pistre, V.; Kinoshita, T.; Blyth, M.; Saenz, E. Attenuation-Based Quantitative Cement Bond Index with LWD Sonic: A Novel Approach Applicable to all Casing and Cement Cases. In Proceedings of the SPE Annual Technical Conference and Exhibition, Amsterdam, The Netherlands, 27–29 October 2014. [Google Scholar] [CrossRef]
  11. Sun, Z. Research on Cementing Quality Evaluation and Statistical Analysis Technology of Daqing Adjustment Well Sub-Block; Daqing Petroleum Institute: Daqing, China, 2009. [Google Scholar]
  12. Xie, J.X.; Cheng, C.T.; Chau, K.W.; Pei, Y.Z. A hybrid adaptive timedelay neural network model for multi-step-ahead prediction of sunspot activity. Int. J. Environ. Pollut. 2006, 28, 364–381. [Google Scholar] [CrossRef]
  13. Hatami, S.; Ghaderi-Ardakani, A.; Niknejad-Khomami, M.; Karimi-Malekabadi, F.; Rasaei, M.R.; Mohammadi, A.H. On the prediction of CO2 corrosion in petroleum industry. J. Supercrit. Fluids 2016, 117, 108–112. [Google Scholar] [CrossRef]
  14. Asefa, T.; Kemblowski, M.; McKee, M.; Khalil, A. Multi-time scale stream flow predictions: The support vector machines approach. J. Hydrol. 2006, 318, 7–16. [Google Scholar] [CrossRef]
  15. Voleti, D.K.; Reddicharla, N.; Guntupalli, S. Smart Way for Consistent Cement Bond Evaluation and Reducing Human Bias Using Machine Learning. In Proceedings of the Abu Dhabi International Petroleum Exhibition & Conference, Abu Dhabi, United Arab Emirates, 11 September 2020. [Google Scholar]
  16. Ni, H.M.; Yi, Z.; Li, P.C.; Tong, X. Application of BP Network Based on PSO Algorithm in Cementing Quality Prediction. Adv. Mater. Res. 2014, 926–930, 4433–4436. [Google Scholar] [CrossRef]
  17. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  18. Vapnik, V.N. Statical Leaming Theory; John Wiley & Sons lnc.: New York, NY, USA, 1998. [Google Scholar]
  19. Vapnik, V.N. Statistical Learning Theory. Encycl. Ences Learn. 1998, 41, 3185. [Google Scholar]
  20. Vapnik, V.N.; Kotz, S. Estimation of Dependences Based on Empirical Data; Springer: New York, NY, USA, 2006. [Google Scholar]
  21. Thissen, U.; van Brakel, R.; de Weijer, A.; Melssen, W.; Buydens, L. Using support vector machines for time series prediction. Chemom. Intell. Lab. Syst. 2003, 69, 35–49. [Google Scholar] [CrossRef]
  22. Schölkopf, B.; Bartlett, P.; Smola, A.; Williamson, R. Support Vector Regression with Automatic Accuracy Control. In ICANN 98. ICANN 1998. Perspectives in Neural Computing; Niklasson, L., Bodén, M., Ziemke, T., Eds.; Springer: London, UK, 1998. [Google Scholar] [CrossRef]
  23. Jakkula, V. Tutorial on support vector machine (svm). Sch. EECS Wash. State Univ. 2006, 37, 3. [Google Scholar]
  24. Chauchard, F.; Svensson, J.; Axelsson, J.; Andersson-Engels, S.; Roussel, S. Localization of embedded inclusions using detection of fluorescence: Feasibility study based on simulation data, LS-SVM modeling and EPO pre-processing. Chemom. Intell. Lab. Syst. 2008, 91, 34–42. [Google Scholar] [CrossRef]
  25. Verwaeren, J.; Van der Weeën, P.; De Baets, B. A search grid for parameter optimization as a byproduct of model sensitivity analysis. Appl. Math. Comput. 2015, 261, 8–27. [Google Scholar] [CrossRef]
  26. Guo, G.; Zhang, J.; Yorke-Smith, N. A Novel Evidence-Based Bayesian Similarity Measure for Recommender Systems. ACM Trans. Web 2016, 10, 1–30. [Google Scholar] [CrossRef]
  27. Anishya, F.; Suresh Kumar, M. A novel approach for travel package recommendation using Bayesian approach. In Proceedings of the 2015 International Conference on Computing and Communications Technologies (ICCCT), Chennai, India, 26–27 February 2015; pp. 296–301. [Google Scholar] [CrossRef]
  28. Pourbasheer, E.; Aalizadeh, R.; Ganjali, M.R.; Norouzi, P. QSAR study of α1β4 integrin inhibitors by GA-MLR and GA-SVM methods. Struct. Chem. 2014, 25, 355–370. [Google Scholar] [CrossRef]
  29. Chen, P.-W.; Wang, J.-Y.; Lee, H.-M. Model selection of SVMs using GA approach. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 3, pp. 2035–2040. [Google Scholar] [CrossRef]
  30. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. Adv. Neural Inf. Process. Syst. 2012, 4, 2951–2959. [Google Scholar]
  31. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; De Freitas, N. Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proc. IEEE 2015, 104, 148–175. [Google Scholar] [CrossRef]
  32. Frenzel, J.F. Genetic algorithms. IEEE Potentials 1993, 12, 21–24. [Google Scholar] [CrossRef]
  33. Goldberg, D. Genetic Algorithms in Search, Optimization, and Machine; Addison-Wesley: Boston, MA, USA, 1989; pp. 232–237. [Google Scholar]
  34. Singh, K.; Miska, S.; Ozbayoglu, E.; Aydin, B.A. Using Supervised Machine Learning Algorithms to Predict Pressure Drop in Narrow Annulus. In Proceedings of the SPE/AAPG Eastern Regional Meeting, Pittsburgh, PA, USA, 7–11 October 2018. [Google Scholar] [CrossRef]
Figure 1. Optimal hyperplane diagram [23].
Figure 1. Optimal hyperplane diagram [23].
Applsci 13 12382 g001
Figure 2. Statistical analysis histograms.
Figure 2. Statistical analysis histograms.
Applsci 13 12382 g002aApplsci 13 12382 g002b
Figure 3. GS flow chart.
Figure 3. GS flow chart.
Applsci 13 12382 g003
Figure 4. BOA flow chart.
Figure 4. BOA flow chart.
Applsci 13 12382 g004
Figure 5. GA flow chart.
Figure 5. GA flow chart.
Applsci 13 12382 g005
Figure 6. Prediction results of the four algorithms.
Figure 6. Prediction results of the four algorithms.
Applsci 13 12382 g006
Figure 7. MRE changes of the four algorithms.
Figure 7. MRE changes of the four algorithms.
Applsci 13 12382 g007
Figure 8. Acoustic amplitude log of cementing quality of a well in the Shunbei block.
Figure 8. Acoustic amplitude log of cementing quality of a well in the Shunbei block.
Applsci 13 12382 g008
Figure 9. Correlation diagram between the actual value and the predicted value of cementing quality.
Figure 9. Correlation diagram between the actual value and the predicted value of cementing quality.
Applsci 13 12382 g009
Figure 10. GA–SVR model prediction cementing quality relative error analysis diagram.
Figure 10. GA–SVR model prediction cementing quality relative error analysis diagram.
Applsci 13 12382 g010
Table 1. Test set of random partition and model prediction evaluation.
Table 1. Test set of random partition and model prediction evaluation.
NumberActual Cementing Quality/%SVR Predicted Cementing Quality/%Relative ErrorRMSEMRETime/s
1413.3713.920.04112.8510.0980<1
1619.3420.490.0595
1717.2921.330.2337
2022.3122.540.0103
2620.1921.210.0505
4228.3423.690.1641
4329.7625.070.1576
4421.3922.760.0640
Table 2. Optimization results and model evaluation of GS method.
Table 2. Optimization results and model evaluation of GS method.
Penalty Coefficient CKernel Function Parameter gNumberActual/%GS–SVR Prediction Cementing Quality
/%
Relevant ErrorRMSEMRETime/s
25.10.131413.3713.650.02092.6530.084644.8
1619.3420.600.0652
1717.2921.320.2331
2022.3122.570.0117
2620.1920.190.0005
4228.3424.480.1362
4329.7625.040.1586
4421.3922.470.0505
Table 3. BOA results and model evaluation.
Table 3. BOA results and model evaluation.
Penalty Parameter
C
Kernel Function Parameter gNumberActual Cementing Quality
/%
BOA–SVR Predicted Cementing Quality
/%
Relevant ErrorRMSEMRETime/s
47.380.076351413.3713.520.01122.5330.080019.8
1619.3420.510.0605
1717.2921.210.2267
2022.3122.490.0081
2620.1920.240.0025
4228.3424.830.1239
4329.7625.180.1539
4421.3922.530.0533
Table 4. GA optimization results and model evaluation.
Table 4. GA optimization results and model evaluation.
Penalty Parameter CKernel Function Parameter
g
NumberActual Cementing Quality
/%
GA–SVR Predicted Cementing Quality
/%
Relevant ErrorRMSEMRETime
/s
85.080.02381413.3713.560.01422.3180.073036.3
1619.3420.090.0388
1717.2920.630.1932
2022.3122.060.0112
2620.1920.300.0054
4228.3425.090.1148
4329.7625.410.1462
4421.3922.670.0598
Table 5. Evaluation before and after model optimization.
Table 5. Evaluation before and after model optimization.
ModelMRERMSETime
SVR0.09802.851<1
GS–SVR0.08462.65344.8
BOA–SVR0.08002.53319.8
GA–SVR0.07302.31836.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, J.; Zheng, S.; Han, J.; Bai, K. Cementing Quality Prediction in the Shunbei Block Based on Genetic Algorithm and Support Vector Regression. Appl. Sci. 2023, 13, 12382. https://doi.org/10.3390/app132212382

AMA Style

Wei J, Zheng S, Han J, Bai K. Cementing Quality Prediction in the Shunbei Block Based on Genetic Algorithm and Support Vector Regression. Applied Sciences. 2023; 13(22):12382. https://doi.org/10.3390/app132212382

Chicago/Turabian Style

Wei, Juntao, Shuangjin Zheng, Jiafan Han, and Kai Bai. 2023. "Cementing Quality Prediction in the Shunbei Block Based on Genetic Algorithm and Support Vector Regression" Applied Sciences 13, no. 22: 12382. https://doi.org/10.3390/app132212382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop