Next Article in Journal
Experimental and Numerical Study of Nonlinear Vibration in Compressor Stator Blade with Holding Ring Structure Considering Contact Friction
Previous Article in Journal
Hybrid Particle Swarm Optimization Genetic LQR Controller for Active Suspension
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DPGWO Based Feature Selection Machine Learning Model for Prediction of Crack Dimensions in Steam Generator Tubes

by
Mathias Vijay Albert William
1,
Subramanian Ramesh
2,
Robert Cep
3,*,
Siva Kumar Mahalingam
4 and
Muniyandy Elangovan
5,6,*
1
Department of Electronics and Communication Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Avadi 600062, India
2
Department of Electrical and Electronics Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Avadi 600062, India
3
Department of Machining, Assembly and Engineering Metrology, Faculty of Mechanical Engineering, VSB-Technical University of Ostrava, 70800 Ostrava, Czech Republic
4
Department of Mechanical Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Avadi 600062, India
5
Department of Biosciences, Saveetha School of Engineering, Saveetha Nagar, Thandalam 602105, India
6
Department of R&D, Bond Marine Consultancy, London EC1V 2NX, UK
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(14), 8206; https://doi.org/10.3390/app13148206
Submission received: 21 June 2023 / Revised: 11 July 2023 / Accepted: 13 July 2023 / Published: 14 July 2023

Abstract

:
The selection of an appropriate number of features and their combinations will play a major role in improving the learning accuracy, computation cost, and understanding of machine learning models. In this present work, 22 gray-level co-occurrence matrix features extracted from magnetic flux leakage images captured in steam generator tubes’ cracks are considered for developing a machine learning model to predict and analyze crack dimensions in terms of their length, depth, and width. The performance of the models is examined by considering R2 and RMSE values calculated using both training and testing data sets. The F Score and Mutual Information Score methods have been applied to prioritize the features. To analyze the effect of different machine learning models, their number of features, and their selection methods, a Taguchi experimental design has been implemented and an analysis of variance test has been conducted. The dynamic population gray wolf algorithm (DPGWO) has been adopted to select the best features and their combinations. Due to the two contradictory natures of performance metrics, Pareto optimal solutions are considered, and the best one is obtained using Deng’s method. The effectiveness of DPGWO is proved by comparing its performance with Grey Wolf Optimization and Moth Flame Optimization algorithms using the Friedman test and performance indicators, namely inverted generational distance and spacing.

1. Introduction

In nuclear power plants, critical components such as steam generator tubes (SGT), feed water heaters, and pressure vessels have stringent design requirements due to the high temperature, pressure, and radiation-exposing environment, which induce stress corrosion cracking, pitting, fouling, and mechanical fretting [1]. Apart from that, steam generator tube rupture (SGTR) leads to the inevitable release of radiation into the environment [2]. For the safety and reliable operation of the plant, it is necessary to go for online monitoring and periodic inspection [3,4]. Non-destructive testing (NDT) is a mode of testing that has been adopted by industries for more than a decade to test mass-manufactured products for anomalies. This process is increasingly being used by sectors such as aerospace, oil and gas, petroleum, nuclear, and construction industries [5,6,7]. More part failures occur due to the complexity of the devices produced. To avoid such things, NDT findings forecast failure and enhance the safety and economy of enterprises. Sumptuous non-destructive evaluation techniques have been established for major systems such as power plants and airplanes to confirm the durability of the NDT test and their safety. Singh et al. [8] proposed the magnetic flux leakage (MFL) technique to identify the localized faults in the SGT. Zhang et al. [9] implemented the MFL technique to detect both shallow surface and deep sub-surface defects in ferromagnetic materials. Suresh et al. [10,11] suggested the MFL approach for the detection of defects and subsurface cracks in small-diameter SGT. The experimental setup for the measurement of MFL is detailed in Suresh et al. [10]. Daniel et al. [12] designed an ANN model to forecast the SGT’s defect in terms of the length, breadth, and depth of the crack by providing the gray-level co-occurrence matrix (GLCM) information extracted from the MFL images.
Wang et al. [13] studied a detailed review of the application of ML models in predicting the outcomes of stroke with structured data. The random forest (RF) ML algorithm has been implemented to predict the outcome of endovascular treatment [14] and acute stroke [15] and improve prediction accuracy in Ischemic stroke patients. Decision tree (DT) and extreme gradient boosting (XGB) ML algorithms were implemented to improve the accuracy of the prediction of Ischemic stroke patients [16].
In ML, feature selection, a method of selecting independent parameters, reduces the computation time and complexity of the problem by eliminating irrelevant, not-useful, and redundant features. Due to its combinatorial nature, feature selection is treated as a non-polynomial hard problem, which creates the opportunity for the researchers to implement meta-heuristic algorithms. Devi et al. [17] proposed a new variant of the Golden Jackel Optimization (GJO) algorithm called Improved GJO (IGJO) to solve the feature selection problem in ML. Qaraad et al. [18] introduced quadratic interpolation with a salp swarm-based local escape operator in the feature selection of 19 datasets and conducted Friedman and Wilcoxon tests to analyze the results. Houssein et al. [19] implemented a centroid mutation-based search and rescue optimization algorithm for feature selection in 15 disease data sets with different feature sizes extracted from the UCI machine learning repository and compared the performance with six existing meta-heuristic algorithms. Filter-based and wrapper-based techniques are the two basic categories under which feature selection may be classified [20,21]. While the wrapper-based strategy bases its evaluation of the solution throughout the searching and optimization procedures on the learning algorithm, the filter-based approach uses the correlation between the data and the relevant class label without consulting the learning algorithm. The wrapper-based technique is the most frequently employed feature selection method as compared to the less computationally expensive filter-based approach due to its higher performance accuracy [22,23].
In this work, the ML model is to be developed to predict the SGT’s defect in terms of length, width, and depth of crack from the given gray level co-occurrence matrix feature from the MFL image. The feature selection method and the number of selected features will affect the performance of the ML models. The R2 and root mean square error (RMSE) values are considered metrics to measure the performance of the ML models. Multiple contradictory performance measures require the conversion of multiple objectives into a single objective, which initiated the implementation of a multi-criterion decision-making method, namely Deng’s method. The dynamic population gray wolf optimization (DPGWO) algorithm is implemented to select the number of features and their combinations to minimize the prediction error. The effectiveness of the DPGWO algorithm is proved by comparing its performance with the gray wolf optimization (GWO) Algorithm [24] and moth flame optimization algorithm (MFO) [25] by implementing a non-parametric Friedman test [25] and performance indicators.
The paper is organized as follows. Section 2 describes the problem statement, and Section 3 describes the proposed methodology and the various stages, namely prioritizing the features, Taguchi orthogonal array experimental design [25], analysis of Variance (ANOVA) Test, and meta-heuristic algorithms for feature selection involved in solving the problem. Section 4 deals with the results and discussion, including a comparison of the performance of the DPGWO algorithm with the MFO and GWO algorithms. Finally, the conclusion part of the paper gives future scope.

2. Problem Statement

Any model must be appropriately selected to match the requirements of the application. Decision-makers analyze the behavior of data using prediction models. Regression models, artificial neural networks, and support vector machines are a few techniques used to build prediction models. Each technique has its own pros and cons based on the size of the dataset. Recently, machine learning techniques have assisted researchers in developing more accurate prediction models. The type of machine learning (ML) model will differ from problem to problem. The same ML model may not give similar performance across applications. Hence, the selection of a suitable ML model for the given problem is a challenging task.
Apart from that, the selection of appropriate features will play a major role in improving the learning accuracy, computation cost, and understanding of the model. Overfitting of models is one of the major problems that will reduce the application of ML in various fields. The problem considered in this work will be developing an accurate ML model with a smaller number of features to predict the crack dimensions from 22 GLCM features extracted from 105 MFL crack images presented by Daniel et al. [12].

3. Methodology

The main focus of the present work is to implement a suitable ML Model with an appropriate selection of features and their combinations to predict the crack dimensions more accurately as compared with the existing ANN model. Daniel et al. [12] from the given 22 GLCM features extracted from 115 MFL crack images. This has been carried out in four stages. In the first stage, the features are prioritized based on F Score (FS) and Mutual Information Score (MIS) values. L16 Taguchi orthogonal array design [25] has been constructed by considering feature selection methods (FSM), machine learning models (MLM), and number of features as parameters and R2 and RMSE values as response values. For each experiment, python codes are executed using the corresponding MLM, FSM, and target number of features as determined by the Taguchi array to determine the corresponding R2 and RMSE values for both training and test data sets. To analyze the effect of parameters and their interaction on the response values, an ANOVA test has been carried out in the third stage. The responses, namely R2 and RMSE values, for the training and testing data sets, are different for each experiment, which is similar to different alternatives among multiple design problems. To select the best design, these multiple responses are to be converted into a single value. Hence, in this work, a similarity-based multi-criterion ranking method called Deng’s method [26] is introduced to select the best ML model in the third stage. In the fourth stage, meta-heuristic algorithms, namely MFO, GWO, and DPGWO, a new variant of GWO proposed in this work, have been implemented to select the number of features and their combinations. Also, Performance indicators and Friedman’s test have been conducted in the fourth stage to prove the effectiveness of the DPGWO, along with statistical analysis to confirm the results. The proposed methodology is shown in Figure 1. The step-by-step algorithm of the proposed work is given below.
Step 1:
Read the GLCM features and their corresponding crack dimension datasets.
Step 2:
Prioritize the features’ order based on
(a)
Mutual Information Score (MIS)
(b)
F Score (FS)
Step 3:
Arrange the features based on MIS.
Step 4:
Select the first 15 features and fix one of the crack’s dimensions as the target value.
Step 5:
Set a machine learning model and select R2 and RMSE as performance metrics of the machine learning model.
Step 6:
Set the size of a training data set, and based on that, separate the training and testing data sets along their target values.
Step 7:
Fit the ML model for the data set and predict both training and testing target values.
Step 8:
Compute the response values (performance metrics, namely R2 and RMSE) of the ML model for both the training and testing data sets.
Step 9:
Repeat steps 5 to 8 by changing the different ML models.
Step 10:
Repeat steps 4 to 9 by changing the number of features to 17, 19, and 21.
Step 11:
Repeat steps 3 to 10 by arranging the features based on the F Score.
Step 12:
Conduct an ANOVA and statistical test to test the significance of the parameters on the response values.
Step 10:
Implement Deng’s method to select the best machine learning model based on R2 and RMSE values.
Step 11:
Tune the hyper-parameters of the best machine learning model.
Step 12:
Implement the MFO, GWO, and DPGWO algorithms by randomly selecting the given features along with their target values between 15 and 21.
Step 13:
Compare the performance of algorithms using performance indicators (Inverted Generational Distance and Spacing) and Friedman’s Test. Select the best number of features and their combinations.
Step 14:
Repeat the steps from 3 to 15 for other dimensions of the crack, like crack depth and crack width.

3.1. Stage 1: Prioritizing the Features

Consideration of all features in machine learning may lead to poor performance, excess computation time, and overfitting of the prediction model. To avoid this, feature selection techniques have been used in ML. Since the wrapper method of feature selection is an iterative process and slow in nature, in this work two filter methods, namely, the F score (FS) and the mutual information score (MIS), have been introduced for selection. In the F score method, features are selected statistically by identifying the relationship between parametric features and target features, whereas in the mutual information score method, features are selected based on their entropy. Table 1 represents the list of 22 GLCM features extracted from MFL images of crack dimensions. Figure 2 represents the prioritized features of crack dimensions. It is understood that in Figure 2a, the 3rd feature is given priority and the 2nd feature is given least the priority in the F-Score-based prioritized method, whereas in the MIS method, the 1st feature has first priority and the 19th feature has the least priority in crack length prediction. In crack width prediction, the 7th feature has first priority in both prioritized methods. Different features, namely the 20th and 6th features, have the least priority in crack depth prediction [27].

3.2. Stage 2: Taguchi Orthogonal Array Experimental Design for ML Model Selection

In this work, a total of six different ML models have been used, in which a different combination of four models is considered for each crack dimension. The ML models’ names are described in Table 2. The parameters and their levels considered for crack length are represented in Table 3. The ML models considered for crack depth and crack width are represented in Table A1 and Table A4. Table 4 represents the experimental parameter settings of the L16 Taguchi orthogonal array for the given combination of levels of parameters mentioned in Table 3, obtained using Minitab 19 Software, where 16 represents the minimum number of experimental runs required to conduct the experiments and to find the best parameter settings [28]. For each experimental run corresponding to its parameter settings, Python code has been executed for the training data set with a sample size of 95 images’ features and tested with a data size of 20 images. The performance of the ML model is recorded for each experiment based on the R2 and RMSE values of both the training and testing data sets. Deng’s method has been implemented to select the best ML model based on the overall performance index calculated using R2 and RMSE values, which are listed in Table 4. L16 Taguchi orthogonal array and Deng’s values for crack depth and crack width are presented in Table A2 and Table A5.

3.3. Stage 3: ANOVA Test

The ANOVA Test is used to analyze the differences among the means of different groups using variance and to prove which groups are statistically significant. To prove that the performance is significantly different for various ML models, an ANOVA test [24] is implemented in this work. Figure 3 represents the probability plot of performance measures of the ML model obtained by executing the Python codes. It is understood that the values of R2 and RMSE of both training and testing are within the 95% confidence interval and that their probability values are greater than 0.005. This shows that the experimentation results are acceptable for crack length prediction. Similarly, for crack depth and crack width, the p-value shown in Figure A1 and Table A6 is greater than 0.005, which reveals that the experimental results obtained are acceptable.
The main and interaction effect plots of parameters are depicted in Figure 4a–h for crack length prediction. It is inferred from Figure 4a,b that the mean value of R2Tr is higher when the parameters are MIS, 15, and XGB, whereas in R2Tt the parameters are FS and 18, respectively. It is understood from Figure 4c,d, that for getting the lower mean value of RMSETr and RMSETt, the parameters are to be set in a different way than R2Tr and R2Tt. From Figure 4e,f, it is noted that increasing the NoF value decreases the R2Tr value for all ML models, whereas in R2Tt except MLM and DT in all other models, it increases. Using XGB, a lower RMSE is reported in Figure 4g,h.
The main and interaction effects plots for crack depth and width are shown in Figure A2a–h and Figure A3a–h. It is observed that different parameter settings are required for getting an individual’s optimum response value. Hence, in this work, Deng’s method is introduced to select the best parameter settings based on R2Tr, R2Tt, RMSETr, and RMSETt values. It is inferred that FS, 19, and XGB are the optimum parameters for the highest Deng’s value, which will give optimum response values.
The p-values of factors and their interactions for the training dataset, except for the interaction between FSM and MLM, are all greater than 0.05 in Table 5, which confirms that none of the factors and their interactions influence the performance of ML Models. In the testing dataset, it is identified that NoF and MLM are the influencing parameters. A similar representation is observed in crack depth prediction, which is shown in Table A3. In Table A7, it is noted that only MLM influences the performance of the prediction of crack width.
Table 6 represents the best ML model, the feature selection method, and the number of features for each crack dimension. It is understood that out of the six ML models, XGB performed well for all the crack dimensions. For comparison purposes, the optimum parameters obtained by the desirability function method using Minitab R19 Software are given in Table 7.
Table 8 represents the list of hyper parameters [15] considered for tuning the performance of the XGB ML model. ‘RandomizedSearchCV’ function is used in Python to get the best hyper parameters for crack dimensions, and the same is presented in Table 9.

3.4. Stage 4: Meta-Heuristic Algorithms for Feature Selection

3.4.1. Gray Wolf Optimization

The GWO algorithm mimics the natural and social behavior of gray wolves in hunting [24]. It is a metaheuristic method used for solving several optimization issues, including signal processing, feature selection, and image processing. GWO was initially designed for one objective problem without constraint [24]. The following variants of GWO were found in a detailed literature review carried out by Mirjalili et al. [27].
  • Cross-over and mutation operators to solve economic dispatch problems.
  • Different penalty functions are used to solve constrained engineering problems.
  • Tournament selection and modified augmented Lagrangian multiplier methods to handle constraint problems.
  • Genetic operators to improve exploration and exploitation in multi-objective problems.
  • Non-dominated sorting operator to solve multi-objective problems.
  • Hybridization of other meta-heuristic algorithms with GWO.

3.4.2. Dynamic Population Gray Wolf Optimization (DPGWO)

The selection of the number of features and their combinations from the given 22 features is considered a non-polynomial hard problem. In this work, DPGWO, a new variant of GWO, is introduced to solve the above problem, and its effectiveness is proved by comparing its performance with MFO and GWO algorithms. The dynamic population size in DPGWO enables the optimization process to adjust the number of gray wolves based on their fitness values. As a result, gray wolves with higher fitness values, which are closer to the ideal solution, have a higher likelihood of surviving and reproducing, whereas those with lower fitness values have a larger probability of being replaced. This helps to maintain a varied population and prevent premature convergence to poor solutions. The distinction between the proposed algorithm and the existing enhancement of GWO will be
  • Introducing life for each wolf during population initialization.
  • Fixing the age and probability of reproducing capability of the wolf.
  • Choosing the disease probability to control the dynamic population size.
Apart from that, in the existing evolutionary population dynamics for the GWO algorithm, the repositioning of poor solutions has been carried out around the positions of alpha, beta, and delta wolves [29]. In this work, the following new concepts are incorporated into the existing GWO algorithm to enhance its performance with dynamic populations:
During population initialization, the life of each wolf has been assigned randomly within the maximum number of iterations. The life of ith wolf (Li) is defined as follows in Equation (1):
L i = j ; j [ 1 , n i t r ]
Removing the existing wolf and instead introducing a new wolf will improve exploration. If the life of the wolf is equal to the current iteration (itr), then that wolf will be removed from the population. Equation (2) expresses the life of ith wolf during the current iteration Li(itr).
L i ( i t r ) = 0   i t r = L i i t r L i i t r i t r < L i ( i t r )
The reproducing nature of the wolf is considered to introduce new offspring from the existing fittest wolves after a defined mature age (k) with reproduce probability (rp) expressed in Equation (3). This concept will improve the exploitation of the algorithm. In Equation (3), 1 indicates that the ith wolf can reproduce.
R i ( i t r ) = 1   L i i t r > k   a n d   P ( r 1 < r p ) 0   L i i t r > k   a n d   P ( r 1 r p )
To control the size of the wolves, after reaching the threshold size, the existence of each wolf (Ei) is confirmed with the disease probability (dp) as expressed in Equation (4). The values of r1 and r2 will be between 0 and 1.
E i ( i t r ) = 1   L i i t r > i t r   a n d   P ( r 2 < d p ) 0   L i i t r > i t r   a n d   P ( r 2 d )
The parameters used in the optimization algorithms are presented in Table 10 and Table 11. The pseudocodes of MFO, GWO, and DPGWO are given in Appendix B.
The implementation of DPGWO is presented as a flow chart in Figure 5. Python codes have been developed for all the algorithms and executed 25 times. The statistical analysis of the performance of the algorithm in terms of Deng’s value is presented in Figure 6.
The statistical analysis of output obtained using MFO, GWO, and DPGWO algorithms in terms of Deng’s value for both crack length training and testing data sets for 25 runs is shown in Figure 6. It is understood that the p-value shown in the summary report for all three algorithms will be above 0.005, which shows the output data obtained by the algorithms are normally distributed, hence the results are accepted. The p-value shown in Figure A4a–f confirmed that the results obtained by executing the Python codes for crack depth and width are acceptable.

4. Results and Discussion

Figure 7a,d represent sample pareto solutions obtained using different algorithms for crack length training and testing data sets. It is understood that the solution obtained by implementing DPGWO is better than the solution obtained by other algorithms, namely MFO and GWO. Figure 7b,c,e,f depict the quick convergence of R2 and RMSE in both training and testing data sets of crack length in DPGWO as compared with MFO and GWO. It is observed from Figure A5a,d that DPGWO has higher R2 and lower RMSE values as compared with MFO and GWO in both cases of crack depth training and testing data sets. Early convergence is recorded in iteration numbers 16, 18, 30, and 28 in the cases of R2Tr, RMSETr, R2Tt, and RMSETt, respectively, in Figure A5b,c,e,f. Similar kinds of Pareto solution and convergence plots are obtained in the case of crack width, as shown in Figure A6a–f. It is also observed that a major difference in RMSE value is achieved in DPGWO as compared to MFO and GWO in Figure A6c,f. The p-value shown in Figure A4 confirms that the results obtained for 25 runs are normally distributed and acceptable. Performance indicators are used to compare the pareto solutions generated by the algorithms. In this work, two different performance indicators, namely inverted generational distance (IGD) and spacing (SP), are implemented to check the effectiveness of the algorithms. For problems with more than three objectives, IGD is used to measure the quality of pareto optimal solutions in terms of distribution and convergence, whereas SP is used to measure the distribution and spread of pareto optimal solutions. In both cases, the lower the value, the better the performance of the algorithm. From Table 12, it is confirmed that the DPGWO algorithm has lower values of IGD and SP as compared to the other two algorithms, hence it is outperformed.
To further support the performance of the algorithms, Friedman’s test, a non-parametric test, has been implemented in this work. This test is used to determine whether statistically significant differences exist between the means of three or more groups. In this work, the Deng’s values obtained by 25 runs using three algorithms are considered for ranking and the calculation of mean values. Using the “Friedman ()” function in Matlab, this test is conducted. Table 13 represents Friedman’s Test values and the mean ranking of each algorithm. All the probability values of crack dimensions are less than 0.05, which shows that the algorithm’s performance differs from each other, i.e., a significant effect will be there on the selection of algorithms. Usually, a lower value will be allotted rank 1 in Friedman’s Test. But in this work, the higher the Deng’s value, the better the solution; the higher the mean rank value, the higher the performance of the algorithm; hence, it is proved that the mean rank value of 2.24 in the case of crack length prediction using DPGWO outperformed as compared to the mean rank values of 2.2 and 1.56 of the MFO and GWO algorithms, respectively. Similarly, in the case of crack depth and width prediction, the mean rank of DPGWO will be 2.23 and 2.12, which are higher than the mean rank of MFO and GWO.
The number of features and their combinations obtained by implementing the MFO, GWO, and DPGWO for each crack dimension are presented in Table 14, along with features considered in the Taguchi Orthogonal Array Method (TM) and Desirability Method (DM). In Table 14, ‘N’ represents that the particular feature is not selected by the methods/algorithms.
The feature that has a higher count of ‘N’ is decided to be an important feature. The list of not-important features compared by various methods is presented in Table 15. A total of 5 out of 22 features are not considered important features.
The performance metrics for the sample considered in the existing literature by Daniel et al. [12] are calculated based on the number of features and their combinations given in Table 16. Figure 8, Figure 9 and Figure 10 represent the actual crack dimensions of the sample along with the calculated crack dimensions using the existing methods, MFO, GWO, and DPGWO. It is clear that all the dimensions calculated by DPGWO are very close to the actual dimensions. Table 16 represents the comparison of the performance metrics with existing and proposed algorithms for the sample data set considered by Daniel et al. It is observed that the R2 value is higher in DPGWO as compared with the existing methods, MFO, and GWO. In the meantime, other metrics like RMSE, MAE, and MAPE are also lower in DPGWO as compared with others. It shows that the DPGWO performed well. The statistical comparison of computation time for the proposed DPGWO with MFO and GWO is shown in Table 17. It is understood that the minimum computation times of DPGWO, MFO, and GWO will be 32.1 s, 29.6 s, and 89.6 s, respectively. Due to the introduction of life, reproduction, and disease strategies in DPGWO, the computation time is 8.4% higher than the GWO algorithm and 64.2% lower than the MFO algorithm.

5. Conclusions

In this work, the best ML model suitable to predict the crack dimensions from the given 22 GLCM features extracted from MFL images reported by Daniel et al. [12], was determined using the Taguchi L16 orthogonal array experimental design. Out of six ML models, the XGB model is identified as the best suited for the proposed problem. Apart from that, two-filter-based feature selection methods are implemented for prioritizing the features. The performance of the ML models is examined based on the number of features and their combinations. Meta-heuristic algorithms, namely MFO, GWO, and a new variant of GWO called DPGWO, are implemented to identify the number of features and their combinations to achieve better performance of the selected XGB ML model. The proposed method proved that by eliminating 23% of the number of features as compared with the existing method, it was possible to get good performance using the XGB ML model. Also, in this work, it is identified that the features ETR, CST, ID, AC, and IDN are not important features, i.e., these features are eliminated during the prediction of crack dimensions. The effectiveness of the proposed DPGWO is proved by comparing Friedman’s Test and performance indicators’ values, viz., IGD and SP with MFO and GWO. Also, the proposed method is implemented on the 12-test data set carried out by Daniel et al. [12]. As compared to the existing method, the proposed DPGWO performed well in calculating performance metrics like R2, RMSE, MAE, and MAPE. Approximately 0.2 to 0.5% improvements in R2 value, 50% to 57% improvement in RMSE value, 52% to 62% improvement in MAE value, and 53% to 63% improvement in MAPE values are reported by using a smaller number of features, i.e., 17 and its combinations, and implementing a new variant of GWO called DPGWO. As a further scope, the wrapper method of features selection along with the multi-regressor concept of ML models may be considered to improve performance. Apart from that, the crack dimensions may be changed from numerical to categorical data type, and ML classifiers can be introduced instead of ML regressors. The proposed DPGWO may be implemented for constrained engineering problems.

Author Contributions

Conceptualization, M.V.A.W., S.R., R.C., S.K.M. and M.E.; data curation, M.V.A.W.; Formal analysis, M.V.A.W.; investigation, M.V.A.W.; methodology, S.R. and M.E.; resources, R.C. and S.K.M.; software, S.R., R.C., S.K.M. and M.E.; visualization, M.V.A.W. and S.R.; writing—original draft, M.V.A.W. and S.R.; writing—review and editing, R.C., S.K.M. and M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the project SP2023/088 supported by the Ministry of Education, Youth, and Sports, Czech Republic.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request through email to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

GitHub Link: https://github.com/lawansisa/ML_MFO_GWO_DPGWO (accessed on 1 June 2023).
Table A1. Parameters and Its Level—Crack Depth.
Table A1. Parameters and Its Level—Crack Depth.
ParameterNo. of LevelsLevels
1234
FSM2FSMIS
MLM4ABRLiRRFXGB
NoF415171921
Table A2. L16 OA—Performance of ML for Crack Depth.
Table A2. L16 OA—Performance of ML for Crack Depth.
E.No.FSMMLMNoFR2TrR2TtRmseTrRmseTtDeng’s Value
1FSABR150.77410.04411.39243.00910.3319
2FSABR170.85370.16801.28023.39720.4154
3MISABR190.86990.02211.18333.17800.3480
4MISABR210.91540.04460.91103.21480.3941
5FSLiR150.59560.36091.86653.55400.4292
6FSLiR170.60490.26981.83903.36890.3965
7MISLiR190.68830.40801.62043.66480.4693
8MISLiR210.72740.28821.53323.51750.4361
9MISRF150.92070.04440.81053.12080.4061
10MISRF170.92280.05840.76033.02850.4208
11FSRF190.93300.04690.69823.23780.4224
12FSRF210.94120.04550.62243.06680.4325
13MISXGB151.00000.44520.00063.64580.6737
14MISXGB171.00000.48290.00053.72460.6806
15FSXGB191.00000.51530.00063.90490.6836
16FSXGB211.00000.46630.00043.76020.6761
Figure A1. Statistical Inference of ML Models’ Performance for Crack Depth Prediction. (a) Probability Plot for Crack Length-R2Tr. (b) Probability Plot for Crack Length-R2Tt. (c) Probability Plot for Crack Length-RMSETr. (d) Probability Plot for Crack Length-RMSETt.
Figure A1. Statistical Inference of ML Models’ Performance for Crack Depth Prediction. (a) Probability Plot for Crack Length-R2Tr. (b) Probability Plot for Crack Length-R2Tt. (c) Probability Plot for Crack Length-RMSETr. (d) Probability Plot for Crack Length-RMSETt.
Applsci 13 08206 g0a1
Figure A2. Factorial and Interaction Plots for Performance of ML Models for Crack Depth Prediction. (a) R2 Factorial Plot for Training Dataset—D. (b) R2 Factorial Plot for Testing Dataset—D. (c) RMSEFactorial Plot for Training Dataset—D. (d) RMSEFactorial Plot for Testing Dataset—D. (e) R2 Interaction Plot for Training Dataset—D. (f) R2 Interaction Plot for Testing Dataset—D. (g) RMSE Interaction Plot for Training Dataset—D. (h) RMSE Interaction Plot for Testing Dataset—D.
Figure A2. Factorial and Interaction Plots for Performance of ML Models for Crack Depth Prediction. (a) R2 Factorial Plot for Training Dataset—D. (b) R2 Factorial Plot for Testing Dataset—D. (c) RMSEFactorial Plot for Training Dataset—D. (d) RMSEFactorial Plot for Testing Dataset—D. (e) R2 Interaction Plot for Training Dataset—D. (f) R2 Interaction Plot for Testing Dataset—D. (g) RMSE Interaction Plot for Training Dataset—D. (h) RMSE Interaction Plot for Testing Dataset—D.
Applsci 13 08206 g0a2aApplsci 13 08206 g0a2b
Table A3. ANOVA for Crack Depth Prediction.
Table A3. ANOVA for Crack Depth Prediction.
SourceDFR2TrR2TtRMSETrRMSETt
Adj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-Value
Regression130.290050.0223187.170000.011000.547410.0421165.070000.015006.582270.50633753.530000.001001.227280.094418.760000.10700
 NoF10.000180.000180.700000.491000.005300.005308.200000.103000.000540.000540.810000.463000.038260.038263.550000.20000
 FSM10.000010.000010.020000.893000.000280.000280.440000.577000.003290.003294.900000.157000.001920.001920.180000.71400
 MLM30.004160.001395.420000.160000.022470.0074911.570000.081000.058400.0194728.970000.034000.100520.033513.110000.25300
 NoF*NoF10.000000.000000.000000.983000.003360.003365.200000.150000.003770.003775.600000.142000.023720.023722.200000.27600
 NoF*FSM10.000010.000010.030000.884000.000050.000050.080000.809000.002360.002363.510000.202000.000010.000010.000000.98100
 NoF*MLM30.002410.000803.140000.251000.016300.005438.390000.108000.019770.006599.810000.094000.087680.029232.710000.28100
 FSM*MLM30.001130.000381.470000.429000.021820.0072711.240000.083000.006800.002273.370000.237000.098590.032873.050000.25700
Error20.000510.00026 0.001290.00065 0.001340.00067 0.021540.01077
Total150.29057 0.54870 6.58361 1.24882
Table A4. Parameters and Its Level—Crack Width.
Table A4. Parameters and Its Level—Crack Width.
ParameterNo. of LevelsLevels
1234
FSM2FSMIS
MLM4ABRLiRRFXGB
NoF415171921
Table A5. L16 OA—Performance of ML for Crack Width.
Table A5. L16 OA—Performance of ML for Crack Width.
E.No.FSMMLMNoFR2TrR2TtRMSETrRmseTtDeng’s Value
1FSABR150.84470.20310.13490.30080.4569
2FSABR170.86330.24350.12400.28960.4771
3MISABR190.85730.30260.12710.27980.4879
4MISABR210.87890.34110.11910.27240.5040
5FSLiR150.69830.32270.17950.26040.4434
6FSLiR170.70470.39610.17810.23910.4616
7MISLiR190.73820.36110.16630.25610.4643
8MISLiR210.76320.15820.15180.29090.4248
9MISRF170.93510.38890.08200.26480.5484
10MISRF190.93610.31630.08000.28220.5360
11FSRF190.93280.29490.08000.26650.5324
12FSRF210.93370.33320.07930.26350.5408
13MISXGB151.00000.19670.00060.28250.6073
14MISXGB171.00000.41960.00060.24400.6444
15FSXGB191.00000.30500.00050.26880.6263
16FSXGB211.00000.33860.00060.25700.6326
Table A6. Statistical Analysis of ML Models’ Performance for Crack Width Prediction.
Table A6. Statistical Analysis of ML Models’ Performance for Crack Width Prediction.
VariableMeanStDevVarianceMinimumQ1MedianQ3MaximumRangep-Value
R2Tr0.88040.10610.01130.69830.78360.90580.98401.00000.30170.0950
R2Tt0.30760.07430.00550.15820.25640.31950.35610.41960.26140.3030
RmseTr0.09400.06480.00420.00050.02030.10060.14760.17950.17900.0750
RmseTt0.26990.01700.00030.23910.25790.26760.28250.30080.06170.9400
Figure A3. Factorial and Interaction Plots for Performance of ML Models for Crack Width Prediction. (a) R2 Factorial Plot for Training Dataset—W. (b) R2 Factorial Plot for Testing Dataset—W. (c) RMSEFactorial Plot for Training Dataset—W. (d) RMSEFactorial Plot for Testing Dataset—W. (e) R2 Interaction Plot for Training Dataset—W. (f) R2 Interaction Plot for Testing Dataset—W. (g) RMSE Interaction Plot for Training Dataset—W. (h) RMSE Interaction Plot for Testing Dataset—W.
Figure A3. Factorial and Interaction Plots for Performance of ML Models for Crack Width Prediction. (a) R2 Factorial Plot for Training Dataset—W. (b) R2 Factorial Plot for Testing Dataset—W. (c) RMSEFactorial Plot for Training Dataset—W. (d) RMSEFactorial Plot for Testing Dataset—W. (e) R2 Interaction Plot for Training Dataset—W. (f) R2 Interaction Plot for Testing Dataset—W. (g) RMSE Interaction Plot for Training Dataset—W. (h) RMSE Interaction Plot for Testing Dataset—W.
Applsci 13 08206 g0a3aApplsci 13 08206 g0a3b
Table A7. ANOVA for Crack Width Prediction.
Table A7. ANOVA for Crack Width Prediction.
SourceDFR2TrR2TtRMSETrRMSETt
Adj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-Value
Regression130.168960.01300835.70.0010.065840.005070.60.7730.062900.00484298.850.0030.003830.000301.150.556
 NoF10.000000.000000.190.7030.014020.014021.660.3260.000000.000000.020.8960.000490.000491.930.299
 FSM10.000050.000053.430.2050.001260.001260.150.7360.000020.000020.910.4420.000110.000110.420.584
 MLM30.001490.0005031.840.0310.021750.007250.860.5770.000450.000159.180.10.000730.000240.950.55
 NoF*NoF10.000030.000031.790.3130.011750.011751.390.3590.000010.000010.30.6370.000380.000381.50.346
 NoF*FSM10.000020.000021.380.360.001180.001180.140.7440.000010.000010.410.5890.000110.000110.440.576
 NoF*MLM30.000330.000117.130.1260.018970.006330.750.6150.000070.000021.450.4340.000590.000200.780.606
 FSM*MLM30.000200.000074.280.1950.007230.002410.290.8360.000030.000010.620.6660.000560.000190.730.624
Error20.000030.00002 0.016860.00843 0.000030.00002 0.000510.00026
Total150.16899 0.08270 0.06294 0.00434
Figure A4. Statistical Analysis of Output using different algorithms for crack depth and crack width. (a) Statistical analysis of output using MFO for Crack Depth. (b) Statistical analysis of output using GWO for Crack Depth. (c) Statistical analysis of output using DPGWO for Crack Depth. (d) Statistical analysis of output using MFO for Crack Width. (e) Statistical analysis of output using GWO for Crack Width. (f) Statistical analysis of output using DPGWO for Crack Width.
Figure A4. Statistical Analysis of Output using different algorithms for crack depth and crack width. (a) Statistical analysis of output using MFO for Crack Depth. (b) Statistical analysis of output using GWO for Crack Depth. (c) Statistical analysis of output using DPGWO for Crack Depth. (d) Statistical analysis of output using MFO for Crack Width. (e) Statistical analysis of output using GWO for Crack Width. (f) Statistical analysis of output using DPGWO for Crack Width.
Applsci 13 08206 g0a4
Figure A5. Pareto and Convergence Plot for Crack Depth. (a) Pareto Plot for Training Data Set—Crack Depth. (b) R2 Convergence Plot for Training Data Set—Crack Depth. (c) RMSE Convergence Plot for Training Data Set—Crack Depth. (d) Pareto Plot for Testing Data Set—Crack Depth. (e) R2 Convergence Plot for Testing Data Set—Crack Length. (f) RMSE Convergence Plot for Testing Data Set—Crack Depth.
Figure A5. Pareto and Convergence Plot for Crack Depth. (a) Pareto Plot for Training Data Set—Crack Depth. (b) R2 Convergence Plot for Training Data Set—Crack Depth. (c) RMSE Convergence Plot for Training Data Set—Crack Depth. (d) Pareto Plot for Testing Data Set—Crack Depth. (e) R2 Convergence Plot for Testing Data Set—Crack Length. (f) RMSE Convergence Plot for Testing Data Set—Crack Depth.
Applsci 13 08206 g0a5
Figure A6. Pareto and Convergence Plot for Crack Width. (a) Pareto Plot for Training Data Set—Crack Width. (b) R2 Convergence Plot for Training Data Set—Crack Width. (c) RMSE Convergence Plot for Training Data Set—Crack Width. (d) Pareto Plot for Testing Data Set—Crack Width. (e) R2 Convergence Plot for Testing Data Set—Crack Width. (f) RMSE Convergence Plot for Testing Data Set—Crack Width.
Figure A6. Pareto and Convergence Plot for Crack Width. (a) Pareto Plot for Training Data Set—Crack Width. (b) R2 Convergence Plot for Training Data Set—Crack Width. (c) RMSE Convergence Plot for Training Data Set—Crack Width. (d) Pareto Plot for Testing Data Set—Crack Width. (e) R2 Convergence Plot for Testing Data Set—Crack Width. (f) RMSE Convergence Plot for Testing Data Set—Crack Width.
Applsci 13 08206 g0a6

Appendix B

Algorithm A1: Pseudocode of Moth Flame Optimization Algorithm [27]
Initialize the parameters for Moth-flame
Initialize no. of moths (Mj—i = 1,2,…nm) and their positions as ‘nof’ number of randomly selected features from the given 22 GLCM features list
For each i = 1:n do
Determine the fitness function fi (i = 1,2,…nm) of each moth—Performance of ML model
In terms of R2 and RMSE values for both training and testing data sets using Deng’s Method
End For
While (iteration ≤ max_iteration) do
    Update the position of Mi
    Calculate the no. of flames
    Evaluate the fitness function fi
If (iteration == 1) then
    F = sort (M)
       OF = sort (OM)
Else
       F = sort (Mt−1,Mt)
       OF = sort (Mt−1,Mt)
End if
For each i = 1:n do
For each j = 1:d do
         Update the values of r and t
         Calculate the value of D w.r.t. corresponding Moth
         Update M(i,j) w.r.t. corresponding Moth
End For
End For
End While
Display the best objective values with their no. of features and their combinations
Algorithm A2: Pseudocode for Grey Wolf Optimization Algorithm [28]
Initialize no. of grey wolf (Xj—i = 1,2,…ng) and its positions as ‘nof’ number of randomly selected features from the given 22 GLCM features list
While (itr < nitr)
Determine the fitness function fvi (i = 1,2,…ng) of each wolf—Performance of ML model
In terms of R2 and RMSE values for both training and testing data sets using Deng’s Method
Sort fvi in descending order and set as sfi and store the first wolf’s data as Xitr. and Fvitr.
Using the sorted data sfi, assign Xa. = X1., Xb. = X2. and Xd. = X3.
Compute  b = 2 i t r × ( 2 n i t r )
For each wolf
       Update the position using
       b1 = 2 × b × rand()-b and c1 = 2 × rand() and Da. = abs(c1 × Xa. − Xi.) and X1. = Xa. − b1 × Da.
       b2 = 2 × b × rand()-b and c2 = 2 × rand() and Db. = abs(c2 × Xb. − Xi.) and X2. = Xb. − b2 × Db.
       b3 = 2 × b × rand()-b and c3 = 2 × rand() and Dd. = abs(c3 × Xd. − Xi.) and X3. = Xd. − b3 × Dd.
       Xi. = (X1. + X2. + X3.)/3
       Check Xi. within bounds
    End
End
Display the best objective values with their no. of features and their combinations
Algorithm A3: Pseudocode for Dynamic Population Grey Wolf Optimization Algorithm
Initialize no. of grey wolf (Xj—i = 1,2,…ng) and its positions as ‘nof’ number of randomly selected features from the given 22 GLCM features list. Set maximum population size (Nm), reproduction probability (pr), and disease probability (pd)
Generate the life (Li) of each grey wolf within the maximum iteration number.
While (itr < nitr)
    For each wolf
       If Li = itr
        Remove the wolf from the population and update the size of the population
       End
    End
    For each wolf
       If Li ≥ Aa
        If rand() ≤ pr
Generate a child wolf and update the size of the population
        End
       End
    End
    If Ng > Nm
       For each wolf
        If rand() < pd
           Remove the wolf from the population and update the size of the population
        End
       End
    End
Determine the fitness function fvi (i = 1,2,…Ng) of each wolf—Performance of ML model
In terms of R2 and RMSE values for both training and testing data sets using Deng’s Method
Sort fvi in descending order and set as sfi and store the first wolf’s data as Xitr. and Fitr.
Using the sorted data sfi, assign Xa. = X1., Xb. = X2. and Xd. = X3.
Compute  b = 2 i t r × ( 2 n i t r )
For each wolf
       Update the position using
       b1 = 2 × b × rand()-b and c1 = 2 × rand() and Da. = abs(c1 × Xa. − Xi.) and X1. = Xa. − b1 × Da.
       b2 = 2 × b × rand()-b and c2 = 2 × rand() and Db. = abs(c2 × Xb. − Xi.) and X2. = Xb. − b2 × Db.
       b3 = 2 × b × rand()-b and c3 = 2 × rand() and Dd. = abs(c3 × Xd. − Xi.) and X3. = Xd. − b3 × Dd.
       Xi. = (X1. + X2. + X3.)/3
       Check Xi. within bounds
    End
End
Display the best objective values with their no. of features and their combinations

References

  1. Gupta, M.; Khan, M.A.; Butola, R.; Singari, R.M. Advances in applications of Non-Destructive Testing (NDT): A Review. Adv. Mater. Process. Technol. 2021, 8, 2286–2307. [Google Scholar] [CrossRef]
  2. Liu, C.; Peng, Z.; Cui, J.; Huang, X.; Li, Y.; Chen, W. Development of crack and damage in shield tunnel lining under seismic loading: Refined 3D finite element modeling and analyses. Thin-Walled Struct. 2023, 185, 110647. [Google Scholar] [CrossRef]
  3. Gao, Z.; Hong, S.; Dang, C. An experimental investigation of subcooled pool boiling on downward-facing surfaces with microchannels. Appl. Therm. Eng. 2023, 226, 120283. [Google Scholar] [CrossRef]
  4. Liu, H.; Yue, Y.; Liu, C.; Spencer, B.; Cui, J. Automatic recognition and localization of underground pipelines in GPR B-scans using a deep learning model. Tunn. Undergr. Space Technol. 2023, 134, 104861. [Google Scholar] [CrossRef]
  5. Xia, Y.; Shi, M.; Zhang, C.; Wang, C.; Sang, X.; Liu, R.; Zhao, P.; An, G.; Fang, H. Analysis of flexural failure mechanism of ultraviolet cured-in-place-pipe materials for buried pipelines rehabilitation based on curing temperature monitoring. Eng. Fail. Anal. 2022, 14, 106763. [Google Scholar] [CrossRef]
  6. Wang, Y.-Y.; Lou, M.; Wang, Y.; Wu, W.-G.; Yang, F. Stochastic Failure Analysis of Reinforced Thermoplastic Pipes Under Axial Loading and Internal Pressure. China Ocean Eng. 2022, 36, 614–628. [Google Scholar] [CrossRef]
  7. Zeng, L.; Lv, T.; Chen, H.; Ma, T.; Fang, Z.; Shi, J. Flow accelerated corrosion of X65 steel gradual contraction pipe in high CO2 partial pressure environments. Arab. J. Chem. 2023, 16, 104935. [Google Scholar] [CrossRef]
  8. Singh, W.S.; Rao, B.P.; Thirunavukkarasu, S.; Mahadevan, S.; Mukhopadhyay, C.; Jayakumar, T. Development of magnetic flux leakage technique for examination of steam generator tubes of prototype fast breeder reactor. Ann. Nucl. Energy 2015, 83, 57–64. [Google Scholar] [CrossRef]
  9. Zhang, J.; Liu, X.; Xiao, J.; Yang, Z.; Wu, B.; He, C. A comparative study between magnetic field distortion and magnetic flux leakage techniques for surface defect shape reconstruction in steel plates. Sens. Actuators A Phys. 2019, 288, 10–20. [Google Scholar] [CrossRef]
  10. Suresh, V.; Abudhahir, A.; Daniel, J. Development of magnetic flux leakage measuring system for detection of defect in small diameter steam generator tube. Measurement 2017, 95, 273–279. [Google Scholar] [CrossRef]
  11. Suresh, V.; Abudhahir, A.; Daniel, J. Characterization of defects on ferromagnetic tubes using magnetic flux leakage. IEEE Trans. Magn. 2019, 55, 6200510. [Google Scholar] [CrossRef]
  12. Daniel, J.; Abudhahir, A.; Paulin, J.J. Magnetic Flux Leakage (MFL) based defect characterization of steam generator tubes using artificial neural networks. J. Magn. 2017, 22, 34–42. [Google Scholar] [CrossRef] [Green Version]
  13. Wang, W.; Kiik, M.; Peek, N.; Curcin, V.; Marshall, I.J.; Rudd, A.G.; Wang, Y.; Douiri, A.; Wolfe, C.D.; Bray, B. A systematic review of machine learning models for predicting outcomes of stroke with structured data. PLoS ONE 2020, 15, e0234722. [Google Scholar] [CrossRef]
  14. Liu, M.; Gu, Q.; Yang, B.; Yin, Z.; Liu, S.; Yin, L.; Zheng, W. Kinematics Model Optimization Algorithm for Six Degrees of Freedom Parallel Platform. Appl. Sci. 2023, 13, 3082. [Google Scholar] [CrossRef]
  15. Xie, L.; Zhu, Y.; Yin, M.; Wang, Z.; Ou, D.; Zheng, H.; Liu, H.; Yin, G. Self-feature-based point cloud registration method with a novel convolutional Siamese point net for optical measurement of blade profile. Mech. Syst. Signal Process. 2022, 178, 109243. [Google Scholar] [CrossRef]
  16. Lu, H.; Zhu, Y.; Yin, M.; Yin, G.; Xie, L. Multimodal Fusion Convolutional Neural Network With Cross-Attention Mechanism for Internal Defect Detection of Magnetic Tile. IEEE Access 2022, 10, 60876–60886. [Google Scholar] [CrossRef]
  17. Devi, R.M.; Premkumar, M.; Kiruthiga, G.; Sowmya, R. IGJO: An improved golden jackel optimization algorithm using local escaping operator for feature selection problems. Neural Process. Lett. 2023, 1–89. [Google Scholar] [CrossRef]
  18. Qaraad, M.; Amjad, S.; Hussein, N.K.; Elhosseini, M.A. An innovative quadratic interpolation salp swarm-based local escape operator for large-scale global optimization problems and feature selection. Neural Comput. Appl. 2022, 34, 17663–17721. [Google Scholar] [CrossRef]
  19. Houssein, E.H.; Saber, E.; Ali, A.A.; Wazery, Y.M. Centroid mutation-based Search and Rescue optimization algorithm for feature selection and classification. Expert Syst. Appl. 2022, 191, 116235. [Google Scholar] [CrossRef]
  20. Ganesh, N.; Shankar, R.; Čep, R.; Chakraborty, S.; Kalita, K. Efficient feature selection using weighted superposition attraction optimization algorithm. Appl. Sci. 2023, 13, 3223. [Google Scholar] [CrossRef]
  21. Hu, F.; Qiu, L.; Xiang, Y.; Wei, S.; Sun, H.; Hu, H.; Weng, X.; Mao, L.; Zeng, M. Spatial network and driving factors of low-carbon patent applications in China from a public health perspective. Front. Public Health 2023, 11, 1121860. [Google Scholar] [CrossRef]
  22. Dai, X.; Xiao, Z.; Jiang, H.; Alazab, M.; Lui, J.C.S.; Min, G.; Dustdar, S.; Liu, J. Task Offloading for Cloud-Assisted Fog Computing With Dynamic Service Caching in Enterprise Management Systems. IEEE Trans. Ind. Inform. 2023, 19, 662–672. [Google Scholar] [CrossRef]
  23. Priyadarshini, J.; Premalatha, M.; Čep, R.; Jayasudha, M.; Kalita, K. Analyzing Physics-Inspired Metaheuristic Algorithms in Feature Selection with K-Nearest-Neighbor. Appl. Sci. 2023, 13, 906. [Google Scholar] [CrossRef]
  24. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  25. Mahalingam, S.K.; Nagarajan, L.; Velu, C.; Dharmaraj, V.K.; Salunkhe, S.; Hussein, H.M.A. An Evolutionary Algorithmic Approach for Improving the Success Rate of Selective Assembly through a Novel EAUB Method. Appl. Sci. 2022, 12, 8797. [Google Scholar] [CrossRef]
  26. Deng, H. A similarity-based approach to ranking multicriteria alternatives. In Advanced Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence: Third International Conference on Intelligent Computing, ICIC 2007, Qingdao, China, 21–24 August 2007; Huang, D.S., Heutte, L., Loog, M., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2007; Volume 4682, pp. 253–262. [Google Scholar]
  27. Mirjalili, S.; Aljarah, I.; Mafarja, M.; Heidari, A.A.; Faris, H. Grey Wolf Optimizer: Theory, Literature Review, and Application in Computational Fluid Dynamics Problems. In Nature-Inspired Optimizers; Springer: Cham, Switzerland, 2019; pp. 87–105. [Google Scholar] [CrossRef]
  28. Arivalagan, S.; Sappani, R.; Čep, R.; Kumar, M.S. Optimization and Experimental Investigation of 3D Printed Micro Wind Turbine Blade Made of PLA Material. Materials 2023, 16, 2508. [Google Scholar] [CrossRef]
  29. Saremi, S.; Mirjalili, S.Z.; Mirjalili, S.M. Evolutionary population dynamics and grey wolf optimizer. Neural Comput. Appl. 2015, 26, 1257–1263. [Google Scholar] [CrossRef]
  30. Khalilpourazari, S.; Naderi, B.; Khalilpourazary, S. Multi-Objective Stochastic Fractal Search: A Powerful Algorithm for Solving Complex Multi-Objective Optimization Problems. Soft Comput. 2019, 24, 3037–3066. [Google Scholar] [CrossRef]
Figure 1. Proposed Methodology.
Figure 1. Proposed Methodology.
Applsci 13 08206 g001
Figure 2. Prioritized features for crack dimensions. (a) F Score-based prioritized features for Crack Length. (b) MIS-based prioritized features for Crack Length. (c) F Score-based prioritized Features for Crack Depth. (d) MIS-based prioritized features for Crack Depth. (e) F Score-based prioritized Features for Crack Width. (f) MIS-based prioritized features for Crack Width.
Figure 2. Prioritized features for crack dimensions. (a) F Score-based prioritized features for Crack Length. (b) MIS-based prioritized features for Crack Length. (c) F Score-based prioritized Features for Crack Depth. (d) MIS-based prioritized features for Crack Depth. (e) F Score-based prioritized Features for Crack Width. (f) MIS-based prioritized features for Crack Width.
Applsci 13 08206 g002aApplsci 13 08206 g002b
Figure 3. Probability Plot of Performance Measures of ML Models in Crack Length. (a) Probability Plot for Crack Length-R2Tr. (b) Probability Plot for Crack Length-R2Tt. (c) Probability Plot for Crack Length-RMSETr. (d) Probability Plot for Crack Length-RMSETt.
Figure 3. Probability Plot of Performance Measures of ML Models in Crack Length. (a) Probability Plot for Crack Length-R2Tr. (b) Probability Plot for Crack Length-R2Tt. (c) Probability Plot for Crack Length-RMSETr. (d) Probability Plot for Crack Length-RMSETt.
Applsci 13 08206 g003aApplsci 13 08206 g003b
Figure 4. Factorial and Interaction Plots for Performance of ML Models for Crack Length Prediction. (a) R2 Factorial Plot for Training Dataset—L. (b) R2 Factorial Plot for Testing Dataset—L. (c) RMSEFactorial Plot for Training Dataset—L. (d) RMSEFactorial Plot for Testing Dataset—L. (e) R2 Interaction Plot for Training Dataset—L. (f) R2 Interaction Plot for Testing Dataset—L. (g) RMSE Interaction Plot for Training Dataset—L. (h) RMSE Interaction Plot for Testing Dataset—L.
Figure 4. Factorial and Interaction Plots for Performance of ML Models for Crack Length Prediction. (a) R2 Factorial Plot for Training Dataset—L. (b) R2 Factorial Plot for Testing Dataset—L. (c) RMSEFactorial Plot for Training Dataset—L. (d) RMSEFactorial Plot for Testing Dataset—L. (e) R2 Interaction Plot for Training Dataset—L. (f) R2 Interaction Plot for Testing Dataset—L. (g) RMSE Interaction Plot for Training Dataset—L. (h) RMSE Interaction Plot for Testing Dataset—L.
Applsci 13 08206 g004aApplsci 13 08206 g004b
Figure 5. Implementation of DPGWO.
Figure 5. Implementation of DPGWO.
Applsci 13 08206 g005
Figure 6. Statistical Analysis of Output using different algorithms for Crack Length. (a) Statistical analysis of output using MFO for Crack Length. (b) Statistical analysis of output using GWO for Crack Length. (c) Statistical analysis of output using DPGWO for Crack Length.
Figure 6. Statistical Analysis of Output using different algorithms for Crack Length. (a) Statistical analysis of output using MFO for Crack Length. (b) Statistical analysis of output using GWO for Crack Length. (c) Statistical analysis of output using DPGWO for Crack Length.
Applsci 13 08206 g006
Figure 7. Pareto and Convergence Plot for Crack Length. (a) Pareto Plot for Training Data Set—Crack Length. (b) R2 Convergence Plot for Training Data Set—Crack Length. (c) RMSE Convergence Plot for Training Data Set—Crack Length. (d) Pareto Plot for Testing Data Set—Crack Length. (e) R2 Convergence Plot for Testing Data Set—Crack Length. (f) RMSE Convergence Plot for Testing Data Set—Crack Length.
Figure 7. Pareto and Convergence Plot for Crack Length. (a) Pareto Plot for Training Data Set—Crack Length. (b) R2 Convergence Plot for Training Data Set—Crack Length. (c) RMSE Convergence Plot for Training Data Set—Crack Length. (d) Pareto Plot for Testing Data Set—Crack Length. (e) R2 Convergence Plot for Testing Data Set—Crack Length. (f) RMSE Convergence Plot for Testing Data Set—Crack Length.
Applsci 13 08206 g007
Figure 8. Experimental VS Predicted Value of Crack Length Using Different Algorithms (EV-Experimental Value; P_EM-Predicted value using the existing method; P_MFO, P_GWO and P_DPGWO-Predicted values based on the number of features and its combination obtained using MFO, GWO, and DPGWO algorithms.
Figure 8. Experimental VS Predicted Value of Crack Length Using Different Algorithms (EV-Experimental Value; P_EM-Predicted value using the existing method; P_MFO, P_GWO and P_DPGWO-Predicted values based on the number of features and its combination obtained using MFO, GWO, and DPGWO algorithms.
Applsci 13 08206 g008
Figure 9. Experimental VS Predicted Value of Crack Depth Using Different Algorithms.
Figure 9. Experimental VS Predicted Value of Crack Depth Using Different Algorithms.
Applsci 13 08206 g009
Figure 10. Experimental VS Predicted Value of Crack Width Using Different Algorithms.
Figure 10. Experimental VS Predicted Value of Crack Width Using Different Algorithms.
Applsci 13 08206 g010
Table 1. GLSM Feature Details (Daniel et al. [12]).
Table 1. GLSM Feature Details (Daniel et al. [12]).
FNo.FNameFeature Name
0UNFEnergy/Uniformity
1ETREntropy
2DSLDissimilarity
3CSTContrast
4IDInverse Difference
5CNCorrelation
6HHomogeneity
7ACAuto correlation
8CSCluster shade
9CPCluster prominence
10MPMaximum Probability
11SSSum of squares
12SASum average
13SVSum Variance
14SESum entropy
15DVDifference Variance
16DEDifference entropy
17IMC (1)Information measure of correlation1
18IMC (2)Information measure of correlation 2
19MCCMaximal Correlation Coefficient
20INNInverse Difference Normalized
21IDNInverse different moment normalized
FNo.—Feature Number and Fname—Feature Name.
Table 2. List of Machine Learning (ML) Models Used in Crack Length.
Table 2. List of Machine Learning (ML) Models Used in Crack Length.
ModelName
DTDecision Tree Regressor
LoRLogistic Regressor
LiRLinear Regressor
XGBExtreme Gradient Booster
ABRAdaptive Booster Regressor
RFRandom Forest Regressor
Table 3. Parameters and Its Level for Crack Length.
Table 3. Parameters and Its Level for Crack Length.
ParameterNo. of LevelsLevels
1234
FSM2FSMIS
MLM4DTLiRLoRXGB
NoF415171921
Table 4. L16 OA—Performance of ML for Crack Length.
Table 4. L16 OA—Performance of ML for Crack Length.
E.No.FSMMLMNoFR2TrR2TtRMSETrRMSETtDeng’s Value
1FSDT150.26190.37370.96841.40000.4569
2FSDT170.19980.45721.00001.30000.4771
3MISDT190.61000.14341.02781.64400.4879
4MISDT210.59400.14291.02821.64460.5040
5FSLiR150.95960.60910.23461.20900.4434
6FSLiR170.95920.62090.23891.18150.4616
7MISLiR190.92210.61090.39171.19770.4643
8MISLiR210.90780.54370.42251.28600.4248
9MISLoR150.60020.14490.82001.64250.5484
10MISLoR170.61200.14410.81501.64330.5360
11FSLoR190.58600.14340.80081.64410.5324
12FSLoR210.39600.14300.84001.64450.5408
13MISXGB150.86520.86460.44100.69420.6073
14MISXGB170.87570.87050.40550.68690.6444
15FSXGB191.00000.59880.00051.18010.6263
16FSXGB211.00000.58410.00061.20130.6326
Table 5. ANOVA for Crack Length Prediction.
Table 5. ANOVA for Crack Length Prediction.
SourceDFR2TrR2Tt
Adj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-Value
Regression131.04330.080331.94000.03101.04450.08032780.60000.0000
 NoF10.00080.00080.33000.62500.00330.0033113.42000.0090
 FSM10.00010.00010.05000.84800.00000.00000.47000.5640
 MLM30.00510.00170.67000.64500.00570.001965.80000.0150
 NoF*NoF10.00200.00200.81000.46400.00210.002172.58000.0130
 NoF*FSM10.00370.00371.49000.34700.00130.001344.06000.0220
 NoF*MLM30.00530.00180.70000.63200.00250.000828.74000.0340
 FSM*MLM30.04140.01385.49000.15800.03940.0131454.09000.0020
Error20.00500.0025 0.00010.0000
Total151.0484 1.0446
RMSETrRMSETt
Adj SSAdj MSF-Valuep-ValueAdj SSAdj MSF-Valuep-Value
Regression131.97310.1518711.90000.00101.50920.11611722.12000.0010
 NoF10.00020.00020.73000.48200.00570.005683.76000.0120
 FSM10.00050.00052.23000.27400.00000.00000.15000.7340
 MLM30.00570.00198.87000.10300.00630.002131.21000.0310
 NoF*NoF10.00040.00041.65000.32700.00370.003755.46000.0180
 NoF*FSM10.00050.00042.09000.28500.00220.002232.84000.0290
 NoF*MLM30.00090.00031.40000.44100.00340.001116.88000.0560
 FSM*MLM30.01790.006027.95000.03500.07780.0259384.54000.0030
Error20.00040.0002 0.00010.0001
Total151.9735 1.5093
Table 6. Best ML Model and Number of Features for Crack Dimension Prediction in Taguchi Method (TM).
Table 6. Best ML Model and Number of Features for Crack Dimension Prediction in Taguchi Method (TM).
CharacteristicFSMMLMNoFR2TrR2TtRMSETrRMSETtDeng’s Value
LengthMISXGB170.87570.87050.40550.68690.6444
DepthFSXGB191.00000.51530.00063.90490.6836
WidthMISXGB171.00000.41960.00060.24400.6444
Table 7. Optimum Parameters for Crack Dimensions using Composite Desirability Method (DM).
Table 7. Optimum Parameters for Crack Dimensions using Composite Desirability Method (DM).
Solution forFSMMLNoFR2Tr—FitR2Tt—FitRMSETr—FitRMSETt—FitComposite Desirability
Crack LengthMISXGB170.8905690.8758540.3878790.6552710.868386
Crack DepthFSXGB210.9999980.6218510.01992023.392330.840767
Crack WidthMISXGB190.9997060.6790070.00068790.2315550.887385
Table 8. List of Hyper Parameters and their Levels.
Table 8. List of Hyper Parameters and their Levels.
ParametersLevels
12345
n_estimators10050090011001500
base_score0.250.50.751
learning_rate0.050.10.15
BoostergbtreeGblinear
Table 9. Optimum Hyper Parameters of XGB for Crack Dimensions.
Table 9. Optimum Hyper Parameters of XGB for Crack Dimensions.
ParametersCrack Dimensions
LengthDepthWidth
n_estimators5001500100
base_score10.751
learning_rate0.050.150.05
BoosterGblineargblineargbtree
Table 10. MFO Parameters [28].
Table 10. MFO Parameters [28].
Name of the ParametersValue
Position of the moth close to the flame (t)−1 to −2
Update mechanismLogarithmic spiral
No. of moths (N)30
No. of iterations (nitr)100
Table 11. GWO [25] and DPGWO Parameters.
Table 11. GWO [25] and DPGWO Parameters.
Name of the ParametersGWO DPGWO
No. of grey wolf (Ng)2020
No. of Iterations (nitr)5050
Scale Factor (SF)2 to 02 to 0
Reproduction Probability (pr)Not applicable0.45 (0.35 to 0.55)
Disease Probability (pd)0.05 (0.03 to 0.07)
Table 12. Performance Indicators of Algorithms [30].
Table 12. Performance Indicators of Algorithms [30].
Crack DimensionAlgorithmIGDSP
Crack LengthMFO0.100660.05977
GWO0.104130.04707
DPGWO0.096520.04558
Crack LengthMFO0.238100.49592
GWO0.218440.24535
DPGWO0.193390.08908
Crack LengthMFO0.209320.15989
GWO0.343710.13514
DPGWO0.082450.06335
Table 13. Friedman’s Test Values and Mean Ranking of Algorithms [25].
Table 13. Friedman’s Test Values and Mean Ranking of Algorithms [25].
Crack DimensionsMean RankProbability
MFOGWODPGWO
Crack Length2.21.562.240.0263
Crack Depth1.981.782.230.0049
Crack Width1.842.042.120.0093
Table 14. List of Features Considered in Various Methods.
Table 14. List of Features Considered in Various Methods.
FNo.FnameCrack LengthCrack DepthCrack Width
TMDMOTTMDMOTTMDMOT
MFOGWODPGWOMFOGWODPGWOMFOGWODPGWO
0UNF N NN
1ETRNN NNN N N
2DSL N NNNN
3CST N N N N NNN
4ID NN NN NN N
5CN NN NNNN
6H N N
7AC N NNNN N N N
8CS
9CP
10MP
11SS
12SA
13SV
14SE
15DV
16DENNN N
17IMC(1) NN NN
18IMC(2) N
19MCC
20INN N
21IDNN NN N NNN NNN
No. of Features191717171719211717171719171717
Table 15. Not Importance Features.
Table 15. Not Importance Features.
FNo.FName
1ETR
3CST
4ID
7AC
21IDN
Table 16. Comparison of Performance with existing and proposed algorithms.
Table 16. Comparison of Performance with existing and proposed algorithms.
Performance MetricsEMMFOGWODPGWO
R20.99700.99870.99810.9992
0.99380.99750.99660.9989
0.99690.99930.99900.9994
RMSE0.1010.0660.0780.050
0.0320.0200.0230.014
0.2080.0990.1260.089
MAE0.08720.05630.05680.0417
0.02470.01470.01350.0099
0.19220.07920.10280.0727
MAPE2.931.881.601.38
3.091.721.601.16
2.250.911.090.91
Table 17. Comparison of computation time of algorithms.
Table 17. Comparison of computation time of algorithms.
VariableSamplesMeanStDevVarianceMinimumQ1MedianQ3MaximumSkewnessKurtosis
DPGWO12458.2367.6832.139.9543.752.7759.60.22−0.65
GWO1238.46.9848.7529.631.6239.2541.3851.80.57−0.09
MFO12108.4215.64244.7789.695.93103.75126133.70.42−1.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

William, M.V.A.; Ramesh, S.; Cep, R.; Mahalingam, S.K.; Elangovan, M. DPGWO Based Feature Selection Machine Learning Model for Prediction of Crack Dimensions in Steam Generator Tubes. Appl. Sci. 2023, 13, 8206. https://doi.org/10.3390/app13148206

AMA Style

William MVA, Ramesh S, Cep R, Mahalingam SK, Elangovan M. DPGWO Based Feature Selection Machine Learning Model for Prediction of Crack Dimensions in Steam Generator Tubes. Applied Sciences. 2023; 13(14):8206. https://doi.org/10.3390/app13148206

Chicago/Turabian Style

William, Mathias Vijay Albert, Subramanian Ramesh, Robert Cep, Siva Kumar Mahalingam, and Muniyandy Elangovan. 2023. "DPGWO Based Feature Selection Machine Learning Model for Prediction of Crack Dimensions in Steam Generator Tubes" Applied Sciences 13, no. 14: 8206. https://doi.org/10.3390/app13148206

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop