Next Article in Journal
Single-Machine Scheduling with Simultaneous Learning Effects and Delivery Times
Previous Article in Journal
Finite Element Method for a Fractional-Order Filtration Equation with a Transient Filtration Law
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Bio-Optimized Algorithms for Hyperparameter Tuning in Machine Learning Models: A Software Defect Prediction Case Study

Department of Information Technology, National Institute of Technology Karnataka, Surathkal, Mangalore 575025, India
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(16), 2521; https://doi.org/10.3390/math12162521
Submission received: 10 June 2024 / Revised: 27 July 2024 / Accepted: 2 August 2024 / Published: 15 August 2024
(This article belongs to the Section Mathematics and Computer Science)

Abstract

:
Addressing real-time optimization problems becomes increasingly challenging as their complexity continues to escalate over time. So bio-optimization algorithms (BoAs) come into the picture to solve such problems due to their global search capability, adaptability, versatility, parallelism, and robustness. This article aims to perform hyperparameter tuning of machine learning (ML) models by integrating them with BoAs. Aiming to maximize the accuracy of the hybrid bio-optimized defect prediction (HBoDP) model, this research paper develops four novel hybrid BoAs named the gravitational force Lévy flight grasshopper optimization algorithm (GFLFGOA), the gravitational force Lévy flight grasshopper optimization algorithm–sparrow search algorithm (GFLFGOA-SSA), the gravitational force grasshopper optimization algorithm–sparrow search algorithm (GFGOA-SSA), and the Lévy flight grasshopper optimization algorithm–sparrow search algorithm (LFGOA-SSA). These aforementioned algorithms are proposed by integrating the good exploration capacity of the SSA with the faster convergence of the LFGOA and GFGOA. The performances of the GFLFGOA, GFLFGOA-SSA, GFGOA-SSA, and LFGOA-SSA are verified by conducting two different experiments. Firstly, the experimentation was conducted on nine benchmark functions (BFs) to assess the mean, standard deviation (SD), and convergence rate. The second experiment focuses on boosting the accuracy of the HBoDP model through the fine-tuning of the hyperparameters in the artificial neural network (ANN) and XGBOOST (XGB) models. To justify the effectiveness and performance of these hybrid novel algorithms, we compared them with four base algorithms, namely the grasshopper optimization algorithm (GOA), the sparrow search algorithm (SSA), the gravitational force grasshopper optimization algorithm (GFGOA), and the Lévy flight grasshopper optimization algorithm (LFGOA). Our findings illuminate the effectiveness of this hybrid approach in enhancing the convergence rate and accuracy. The experimental results show a faster convergence rate for BFs and improvements in software defect prediction accuracy for the NASA defect datasets by comparing them with some baseline methods.

1. Introduction

In the real world, solving constrained and unconstrained optimization problems is practically very challenging due to their increased complexity. Optimization problems arise in various fields, including mathematics, engineering, economics, computer science, etc.
In recent years, metaheuristic approaches have gained popularity due to local and global convergence and have been applied to many optimization problems. Based on [1], metaheuristic algorithms have been categorized into single-based methods and population-based methods. Population-based algorithms consist of evolutionarily inspired, swarm-inspired, physics-based, human-based, bio-based, and math-based algorithms. The popularity of the genetic algorithm (GA) [2], the evolutionary algorithm (EA) [3], and the particle swarm optimization (PSO) [4] algorithm has created more interest among researchers to propose more bio-optimization algorithms (BoAs), such as grasshopper optimization algorithm (GOA) [5], the Wolf Pack Algorithm (WPA) [6], Artificial Bee Colony (ABC) [7], the Harris Hawks Optimizer (HHO) [8], etc. Bio-optimization algorithms (BoAs) follow a two-step process, consisting of the the exploration and exploitation stages. Due to their higher convergence speed, few parameters, and simple implementation, these BoAs are widely used in solving parameter optimization, NP-hard problems [9], fault prediction [10], bioinformatics [11], image processing [12], and many other real-time engineering problems.
In this paper, we considered the widely used NASA defect dataset [13] to predict software defects as a case study, where accuracy is considered as one of the performance metrics. To improve or enhance the accuracy, the optimization term comes into the picture. By combining BoAs with the machine learning (ML) and deep learning (DL) approaches for parameter tuning of the respective models, the accuracy level can be enhanced.
Despite the development of numerous BoAs, each algorithm possesses its own set of limitations. Based on the no free lunch (NFL) theorem [14], no single BoA can perform best in addressing all optimization problems. As a result, researchers are interested in exploring and proposing effective optimization methods to tackle a broader range of optimization problems. Referring to [15], innovating hybrid algorithms integrating two or more bio-optimization techniques can be a promising solution. These hybrids aim to amalgamate the strengths of all constituent algorithms.

1.1. Motivation

The need for more advanced optimization approaches increases with the complexity of the optimization problems. Our research in this field is motivated by the following factors:
  • Based on the aforementioned explanation, hybrid approaches are becoming more and more popular to enhance the convergence performance of individual algorithms.
  • The effectiveness and scalability of hybrid BoAs across diverse scientific domains inspired us to propose novel and competitive hybrid methodologies.
  • Hyperparameter tuning can be regarded as an optimization problem, as this process involves searching through a range of objective functions to maximize accuracy.
  • The use of hybrid BoAs for solving the hyperparameter tuning problem is motivated by their ability to handle efficient exploration–exploitation trade-offs and scalability.
Motivated by the preceding arguments, our study introduces an improved version of the GOA. This enhanced version incorporates the principles of gravitational force (GF) and Lévy flight (LF). Additionally, we developed a hybrid novel algorithm by merging the potent exploration feature of the SSA with the enhanced version of the GOA.

1.2. Contribution

The objective of this research work is to propose four novel hybrid BoAs for improving the accuracy of the SDP model. These novel hybrid algorithms are proposed based on the probabilistic selection mechanism, which carefully balances the strong exploration and exploitation tendencies. To evaluate their effectiveness and performance, two types of experiments were carried out. Initially, the experiments were conducted on benchmark functions (BFs) to assess the mean, standard deviation (SD), and convergence rates. Subsequently, a second experiment focused on boosting the accuracy of SDP through the fine-tuning of the hyperparameters in both the artificial neural network (ANN) and XGBOOST (XGB) models. The key contributions of the proposed work are as follows:
  • Design of an Enhanced GOA named gravitational force Lévy flight grasshopper optimization algorithm (GFLFGOA) by introducing LF and GF concepts to balance the exploration and exploitation nature.
  • Design of a novel hybrid gravitational force Lévy flight grasshopper optimization algorithm–sparrow search algorithm (GFLFGOA-SSA) that includes GFLFGOA and the sparrow search algorithm (SSA) concept to accelerate convergence rate.
  • Design of a hybrid gravitational force grasshopper optimization algorithm–sparrow search algorithm (GFGOA-SSA) by embedding the concepts of gravitational force grasshopper optimization algorithm (GFGOA) and SSA concepts.
  • Design of a hybrid Lévy flight grasshopper optimization algorithm–sparrow search algorithm (LFGOA-SSA).
  • Extensive experiments conducted on BFs and hyperparameter tuning of XGB and ANN models to prove the above-proposed algorithm’s superiority to the state-of-the-art techniques.
The subsequent sections of this paper are organized as follows. Section 2 delves into a literature survey, while Section 3 provides a concise overview of the background concepts utilized in the proposed method. Section 4 outlines the methodology employed for carrying out two extensive experiments. Following this, Section 5 presents the results obtained and subsequent analysis. Lastly, Section 6 concludes the paper by summarizing observations and suggesting potential directions for future research.

2. Literature Survey

This section presents some pertinent works that adopt various bio-optimized approaches with ML and DL models for solving various types of optimization problems. We categorized this section into two subsections, namely
  • bio-optimized approach to optimization problems;
  • bio-optimized approach to hyperparameter tuning problems.

2.1. Bio-Optimized Approach to Optimization Problems

It has recently been discovered that optimization approaches based on bio-optimized algorithms are effective in a variety of fields. Certain variations of PSO [4], GA [16], ABC [17], etc., algorithms have been successfully used in a variety of sectors to achieve restricted optimization.
Zhen et al. in [18] proposed a hybrid WPA-PSO to estimate and predict the software reliability parameters. The proposed algorithm was verified through five sets of data from the industry. Through the simulation, the authors justified higher accuracy and better optimization. Zulfiqar et al. in [19] proposed a hybrid model by integrating the multivariate empirical model decomposition (MEMD) with the adaptive differential evolution (ADE) algorithm to optimize the hyperparameter of a support vector machine (SVM) model for electricity load forecasting. The above-proposed MEMD-ADE-SVM model achieved good accuracy by tuning the parameters of SVM. Blume et al. in [20] implemented GA for hyperparameter optimization on ANN for designing software sensors. Akter et al. in [21] developed a new crossover technique to improve GA and implemented it on the traveling salesman problem for optimization.
Sajjad et al. [11] implemented a hyperparameter tuning approach using grey wolf optimization (GWO) on machine learning (ML) algorithms and deep neural networks. The proposed method was employed on eleven datasets of different kinds, including biological, biomedical, clinical diagnosis, etc. Binghui et al. [22] proposed a metamodel-based methodology for hyperparameter optimization of optimization algorithms in building energy optimization (BEO). The authors validated their methodology by considering 15 benchmark BEO problems with various properties. Meetesh et al. [10] investigated the impact of hyperparameter optimization on defect count prediction. In their paper, the validation of the approach was performed with 15 software defect datasets. Through their research work, the authors emphasized the importance of the exploration of parameter space.
These days, several fields have discovered bio-optimized strategies to be effective. Many studies have been carried out on swarm intelligence based on the behavior of sparrows, wolves, bees, ants, etc. On that behavior, some new algorithms and hybrid algorithms have been proposed. Xue et al. in [23] introduced a new optimization technique based on sparrow behavior, which signifies the improvement in terms of convergence speed and search precision. In the recent advancement of the SSA algorithm, researchers have introduced different variants of SSA with adaptive learning factor [24], chaotic mapping [25], robot path planning [26], and adaptive version with adaptive weight [27].
Zhao et al. [28] introduced an optimized technique based on the behavior of wolves. It is based on a greedy strategy, which leads to getting trapped in local optima because of excessive greed. It has been observed that it has faster convergence in the initial stages, but in later stages, the convergence speed slows down. There are many improvisations employed in this algorithm. Regarding the greedy problem, Li et al. [29] have shown that incorporating a chaotic approach into WPA’s search process can successfully prevent it from entering a local optimum. Additionally, the work by Xiu et al. [30] describes how to increase the slow convergence speed at later stages using Lévy behavior. Chen et al. [31] introduced the concept of opposition-based genetic learning to tackle the lack of influence of lead wolves, reducing the stability of the algorithm.
Jadon et al. [32] proposed a hybrid ABC algorithm with differential evolution (DE) that is tested on a welded beam design problem. He modified the position update equation for the employee bee phase and then applied DE to update the position for the onlooker bee phase. Hybrid PSOGA was proposed by Mirjalili et al. [33] for binary optimization by combining the social component of PSO with GA to accelerate the convergence speed. To solve high-dimensional complex problems, quantum particle swarm optimization (QPSO) was introduced by Li et al. [34]. A new hybrid SSA-PSO was developed by Yang et al. [35], which showed great improvements in convergence speed and stability.
As our research work focuses on the hybridization of the GOA with other BoAs, the survey emphasizes various variants and applications of GOA. Saremi et al. [36] proposed a novel BoA named GOA using swarm intelligence to solve various optimization problems. As this algorithm is proven to be efficient enough in solving optimization problems, it has gained a lot of interest among researchers. The authors in [37,38,39] performed a comprehensive analysis of GOA on various real-time problems.
Meraihi et al. in [5] conducted a comprehensive review of the hybridized version of GOA with other BoAs to address real-world problems in order to fully extend GOA’s performance. Arora et al. in [40] incorporated chaos theory into GOA to boost global convergence and justified it by testing it using thirteen benchmark functions (BFs). In [41], the authors proposed improved GOA (IGOA) by embedding trigonometric substitution into the original GOA to boost Cauchy mutation. The performance of the proposed IGOA was validated on the IEEE CEC2017 BFs compared with other BoAs.
The authors in [42] enhanced the exploration phase of the GOA by embedding the crossover operator and salp swarm algorithm into it. The correctness of the proposed method was validated through feature selection datasets and six real-time engineering problems. Yildiz et al. in [43] proposed an improved version of GOA by adding an elite opposite-based learning method into it, called the elite opposite-based learning grasshopper optimization method (EOBL-GOA). The EOBL-GOA was validated through various engineering design problems. Feng et al. [44] developed dynamic opposite learning-assisted GOA (DOLGOA) and validated the correctness of DOLGOA through CEC2014 BFs and the flexible job scheduling problem (FJSP). Peng et al. in [45] proposed an improved grasshopper optimization algorithm (IGOA) by taking the gravity force concept into GOA to optimize the parameters of the backpropagation neural network (BPNN).
The aforementioned algorithms offer the advantage of striking a balance between exploration and exploitation, but their performance is somewhat constrained. Given our focus on hyperparameter tuning of ML models using software defect datasets, we conducted a literature review on optimization problems.

2.2. Bio-Optimized Approach for Hyperparameter Tuning Problems

ML and DL models are widely used for solving regression and classification problems related to various types of real-time applications. The authors in [46] use various ML models, including naive Bayes (NB), support vector regression (SVR), decision tree (DT), and random forest (RF) algorithms, to evaluate the performance metrics for the software reliability assessment problem. The aforementioned models are cross-validated on DBS-1 and DBS-2 datasets. Besides ML, the authors also implemented a learning approach that incorporates a recurrent neural network (RNN) for comparison on the above-mentioned dataset. Consideration of the limited dataset in this work and the lack of an explanation of further optimization processes are the limitations of this work.
In [47], the authors implemented various DL techniques to predict software fault considering Chidamber and Kemerer (CK) metrics-based datasets. The above methods were verified and validated by comparing them with other existing methods. The authors in [48] implemented the feature reduction concept for software defect prediction (SDP). To carry out the proposed method, they used four NASA defect datasets.
An attention-based recurrent neural network (DPARNN) framework is proposed in [49] for SDP. Seven open-source java projects in apache were considered for validating the above-proposed method. The authors in [50] conducted a systematic literature review on fault prediction using various ML, DL, and data mining techniques. Besides the techniques, they also reviewed defect datasets and performance metrics.
The authors in [51] proposed an ANN model with one input layer (eight dimensions), two hidden layers, and one output layer with a sigmoid activation function to predict the software defect of dataset JM1. A comparative analysis was performed by Jindal et al. [52] with different DL models such as gated recurrent network (GRU), long short-term memory (LSTM), and RNN with ANN as a base model, and they found that LSTM performs better than others. Alghanim et al. [53] proposed an enhanced deep neural network (NN) model based on GRNN and tested it with repeated 10-fold cross-validation.
Clemente et al. [54] proved that within ANN, SVM, DT, and RF, RF performed better in the case of PC1, and ANN performed better in the case of KC2. Wongpheng et al. [55] proved that when a model was trained for 100 and 1000 epochs, it delivered high accuracy for a greater number of epochs, but extra experiments were still required for deciding the optimal learning rate and other model parameters.
In [56], the authors examined "five NASA defect public datasets" from the PROMISE data repository named CM1, KC1, KC2, JM1, and PC1 for defect prediction using ten ML classifiers but found no consistently accurate results. DL techniques were also explored, highlighting the complexity of the problem. By selecting the most informative features, the dimensionality of the data is reduced, improving the efficiency of the analysis and enhancing the quality of predictions. It also helps in understanding the underlying factors that contribute to software failures, thus allowing for targeted improvements and resource allocation.
Alsaeedi et al.’s work in [57] employed three classifiers and ensemble methods for defect anticipation on NASA datasets. They found that RF and Ada-Boost with RF outperformed other approaches in addressing software flaws. In [58], an ensemble approach was used with six algorithms on NASA datasets, demonstrating the effectiveness of RF as the best ensemble algorithm for defect prediction.
Iqbal et al.’s study [59] focuses on utilizing twelve NASA defect datasets and employs a range of classification algorithms, such as NB, radial basis function, multi-layer perceptron, and K-nearest neighbor (K-NN). The objective was to forecast software errors and enhance the reliability of software systems through ML-based classifiers and statistical methods.
R. Malhotra’s work in [60] conducted a comprehensive assessment of software bug prediction methods, evaluating ML techniques, comparing them with statistical approaches, and summarizing their strengths and weaknesses. Parashar et al. in [61] proposed a multicore parallel ML approach to classification problems for SDP. The proposed model was trained and tested on eleven software systems of NASA and other relevant repositories.
To the best of our knowledge, in the above-mentioned literary survey, there is a lack of explanation on further improvisation of the performance metrics, even though they had implemented various ML and DL techniques for SDP. As this research work focuses on optimization of the ML models to obtain accurate prediction, it motivates us to focus on parameter tuning of the models for SDP using hybrid BoAs.

3. Concepts

This section briefly describes the base BoAs that are used to propose four novel hybrid BoAs. The performance of the proposed algorithm is verified through BFs and hyperparameter tuning of ML models for the SDP classification problem. The ML models considered for this work are XGB and ANN. For the SDP classification problem, the NASA defect dataset (NASA Defect Dataset https://github.com/klainfo/NASADefectDataset, accessed on 5 January 2024) is considered.

3.1. Grasshopper Optimization Algorithm (GOA)

Saremi et al. [36] proposed GOA for solving optimization problems [62] by mimicking the behavior of grasshopper swarms in nature. The mathematical model of the original GOA is as follows:
Z i = S I i + G F i + A W i
where:
  • Z i   i t h grasshopper position;
  • S I i Social interaction;
  • G F i Gravitational force;
  • A W i Wind advection.
Social interaction in Equation (1) is an important factor for the evaluation of the position of the grasshopper and is evaluated as follows:
S I i = j = 1 , j i n s ( d i j ) d i j ^
where:
  • n Number of grasshoppers;
  • d i j Distance from i t h grasshopper to j t h grasshopper;
  • d i j ^ Unit Vector from i t h grasshopper to j t h grasshopper.
The evaluations of d i j and d i j ^ are presented in Equations (3) and (4), respectively.
d i j = Z j Z i
d i j ^ = ( Z j Z i ) d i j
where:
  • Z j   j t h grasshopper’s position;
  • Z i   i t h grasshopper’s position.
The social force strength (s) is defined in Equation (5).
s ( r ) = f e r / l e r
where:
  • f  Intensity of attraction;
  • l  Attractive length.
As mentioned in [36], the parameter values are l = 1.5 and f = 0.5. The social force (s) cannot be applied if the distance is large between grasshoppers, as s trends to 0. So, to avoid such situations, d should be mapped in the interval of [1, 4]. The gravitational force ( G F i ) is presented as follows:
G F i = g e g ^
where:
  • g  Gravitational constant;
  • e g ^ Unity vector towards the center of the earth.
The wind advection A W i is presented as follows:
A W i = u w i ^
where:
  • u  Constant draft;
  • w i ^ Unity vector in the wind direction.
Substituting the values of S I i , G F i , and A W i in Equation (1), Equation (1) becomes
Z i = j = 1 , j i n s ( Z j Z i ) ( Z j Z i ) d i j g e g ^ + u w i ^
Using Equation (8), the optimization model cannot be solved efficiently, as the grasshopper may be located in the comfort zone. Due to their comfort zone, the grasshoppers cannot be converged efficiently. So, for efficient convergence, some special parameters are added to Equation (8). So Equation (8) can be modified as follows:
Z i d = c j = 1 , j i n c u b d l b d 2 s ( Z j d Z i d ) Z j Z i d i j + T d ^
where:
  • u b d and l b d Upper bound and lower bound of the d t h dimension of the i t h grasshopper, respectively;
  • T d ^ Target or optimal position in the d t h dimension (best solution found so far);
  • c  A decreasing coefficient to shrink the comfort repulsion and attraction areas.
The parameter c is updated to reduce exploration and increase exploitation in accordance with the number of iterations, as indicated in Equation (10).
c = c m a x r c m a x c m i n G
where:
  • c m a x Maximum value;
  • c m i n Minimum value;
  • r  Current iteration;
  • G  Maximum iteration.
The values adopted in this work are c m i n = 0.00001 and c m a x = 1 .

3.2. Lévy Flight GOA (LFGOA)

Paul Lévy introduced the Lévy Flight (LF) concept, which is currently presented as Lévy statistics [63]. Generally, Lévy’s flight step can be represented by Equation (11).
L e v y _ s s 1 α
where:
  • s  Random step length of Lévy’s flight;
  • α Power-law index [0, 2].
Based on [64], the authors considered Mantegna’s algorithm to generate L e v y _ S for stable distribution of the Lévy flight. So for random walks, the step length L e v y _ S is determined using Mantegna’s algorithm and is defined by using Equations (12)–(14).
L e v y _ s = M P 1 α
M = N ( 0 , σ M 2 )
P = N ( 0 , σ P 2 )
where:
  • L e v y _ S Step length for random walk;
  • α 1.5;
  • M  Normal standard variable with standard deviation σ M that follows a normal distribution;
  • P  Normal standard variable with standard deviation σ P that follows a normal distribution.
σ M and σ P are derived from the normal distribution and are represented by using Equations (15) and (16), respectively.
σ M = Γ 1 + α × sin 0.5 π α Γ 0.5 1 + α × α × 2 0.5 α 1 1 / α
σ P = 1
So the new position of the grasshopper using Lévy’s flight is defined as follows:
Z i d = l e v y _ S c ( j = 1 , j i n c u b d l b d 2 s ( | Z j d Z i d | ) Z j Z i d i j ) + T ^ d

3.3. Gravitational Force GOA (GFGOA)

Peng et al. [45] proposed an improved GOA by considering gravitational force (GF) and the selected probability (p) by normalizing the distance between the grasshoppers. In our research work, only the GF concept is considered, as presented in Equation (6). Considering only the GF concept in Equation (9), Equation (9) can be updated as follows:
Z i d = c j = 1 , j i n c u b d l b d 2 s ( Z j d Z i d ) Z j Z i d i j g e g ^ + T d ^
where:
  • g  Gravitational constant (0.9);
  • e g ^ Z j Z i d i j .

3.4. Sparrow Search Algorithm (SSA)

SSA is inspired by the foraging behaviors of sparrows, developed by Xue et al. [23]. In SSA, the sparrows are divided into producer sparrows (PD) and scrounger sparrows (SD). The PD directs the whole population towards the food source. During each iteration, the PD position is updated by Equation (19).
Z i , j r + 1 = Z i , j r · exp i α · G R 2 < S T Z i , j r + Q · L R 2 S T
where:
  • r  Current iteration;
  • L  Matrix of (1 × d);
  • d  Dimension of the variable;
  • Z i , j r   Position of the j t h dimension of the i t h sparrow at iteration r;
  • G  Maximum iterations;
  • Q  Random number which obeys a normal distribution;
  • α   Random number [0, 1];
  • R 2   Alarm value [0, 1];
  • ST  Safety threshold [0.5, 1.0];
  • R 2 < ST Signals the predator to search on wide mode, as there are no predators;
  • R 2 ≥ ST Signals sparrows to fly to another place, as there are predators.
For the SD update, they immediately leave their place to compete for food once they know that the PD has discovered some food source. The position of the SD is updated by using Equation (20).
Z i , j r + 1 = Q · exp Z worst r Z i , j r i 2 if i > n 2 Z P r + 1 + Z i , j r Z P r + 1 · A + · L otherwise
where:
  • Z w o r s t r   Current global worst location at iteration r;
  • Z P r + 1    ith sparrow’s current best location obtained so far;
  • Q  Random number which obeys a normal distribution;
  • A + = A T ( A A T ) 1 ;
  • A  One-dimensional vector [−1, 1];
  • i    { 1 , 2 n } ;
  • n  Number of sparrows.
Based on the assumption that 20% of the sparrows are aware of the danger, the positions of those sparrows are updated using Equation (21).
Z i , j r + 1 = Z b e s t r + β · Z i , j r Z b e s t r f i > f g Z i , j r + K · Z i , j r Z w o r s t r f i f w + ε f i = f g
where:
  • Z b e s t r   Current global optimal location at iteration r;
  • β   Random number [mean = 0, variance = 1];
  • K  Random number [−1, 1];
  • f i   Current sparrow’s fitness value;
  • f g   Current global best value;
  • f w   Worst fitness value;
  • ϵ   Constant value (c);
  • f i > f g   Sparrows are at the edge of the group;
  • f i = f g   Sparrows are in the middle of the population.

4. Methodology

This section emphasizes four proposed algorithms and their implementations on the unimodal and multimodal BFs. Besides focusing on the verification of the proposed algorithms on the BFs, this paper also focuses on the validation of the proposed algorithm for the improvement of accuracy on the SDP problem.

4.1. Proposed Algorithm

This subsection explains in detail the proposed hybrid BoAs, as mentioned in Section 4.1.1, Section 4.1.2, Section 4.1.3 and Section 4.1.4. Table 1 lists the notations and their explanation used for the proposed algorithms’ flowcharts, as drawn in Figure 1, Figure 2, Figure 3 and Figure 4.

4.1.1. GFLFGOA Hybrid Algorithm

In the original GOA, to balance local exploitation and global exploration, the parameter “c” was introduced. However, in the original GOA, the optimization process is nonlinear and has limited exploration ability. Due to the limited exploration ability, GOA may get stuck in local optima, which results in slow convergence. So, to overcome the disadvantages of GOA, an enhanced version of GOA is proposed, called GFLFGOA. To make a good balance between exploration and exploitation, LF and GF are introduced into the original GOA. After embedding the GF and LF concepts into the original GOA, as explained in Section 3.1, Section 3.2 and Section 3.3, the position of the grasshopper can be updated as follows:
Z i d = l e v y _ S c j = 1 , j i N c u b d l b d 2 s ( Z j d Z i d ) Z j Z i d i j g e g ^ + T d ^
The pseudocode of the GFLFGOA algorithm is presented in Algorithm 1.
Algorithm 1: GFLFGOA
Mathematics 12 02521 i001
To visualize the pseudocode and the process of the proposed GFLFGOA algorithm, a flow diagram is drawn in Figure 1.

4.1.2. LFGOA-SSA Hybrid Algorithm

A new hybrid LFGOA-SSA algorithm is proposed in this paper. Its features allow the improvement of the convergence speed and accuracy by avoiding the local optima. It can be observed from the flowchart, as drawn in Figure 2, that the developed LFGOA-SSA is designed to combine the fast convergence of LFGOA and the high accuracy of SSA. The pseudocode of the proposed algorithm is written in Algorithm 2. Basically, this hybrid algorithm is developed based on the probabilistic selection mechanism after the initialization. If rand < 0.5, then the LFGOA is considered to generate the new fitness solution, while the SSA will be selected for rand ≥ 0.5. For SSA, the search agents are divided into PD and SD. If the search agent is PD and R 2 < ST, then the new position will be updated as follows:
Z i , j r + 1 = Z i , j r · exp i α · G
Otherwise, the position for the PD will be updated as mentioned below:
Z i , j r + 1 = Z i , j r + Q · L
In the case of the SD search agent, if i < S D 2 , then the position is updated by Equation (25), and otherwise followed by Equation (26).
Z i , j r + 1 = Q · exp Z worst r Z i , j r i 2
Z i , j r + 1 = Z P r + 1 + Z i , j r Z P r + 1 · A + · L
The pseudocode of the proposed LFGOA-SSA algorithm is presented in Algorithm 2.    
Algorithm 2: LFGOA-SSA
Mathematics 12 02521 i002
    The flow diagram of the LFGOA-SSA algorithm is presented in Figure 2.

4.1.3. GFGOA-SSA Hybrid Algorithm

An enhanced version of GOA named GFGOA, as explained in Section 3.3, is combined with the SSA to improve the balance between exploration and exploitation. The faster convergence of GFGOA and the good exploration capacity of SSA are combined based on the probabilistic selection mechanism. The GFGOA algorithm is selected if the randomly generated value is less than 0.5; otherwise, SSA will be considered. For better visualization of the steps of the GFGOA-SSA hybrid algorithm, the pseudocode and the flow diagram are presented in Algorithm 3 and Figure 3, respectively.
Algorithm 3: GFGOA-SSA
Mathematics 12 02521 i003

4.1.4. GFLFGOA-SSA Hybrid Algorithm

This novel hybrid algorithm is proposed based on a random selection mechanism. If the random value is less than 0.5, then our proposed algorithm GFLFGOA, as explained in Section 4.3.1, is selected for generating the best fitness value; otherwise, the SSA algorithm is used to generate the fitness value. The pseudocode of the GFLFGOA-SSA algorithm is presented in Algorithm 4.
Algorithm 4: GFLFGOA-SSA
Mathematics 12 02521 i004
The flow diagram of the GFLFGOA-SSA algorithm is drawn in Figure 4.

4.2. Optimization on Benchmark Functions (BFs)

BFs are artificial problems that can be used to assess the behavior and the performance of optimization algorithms in diverse and complex situations [65]. These functions are categorized into unimodal problem, multimodal problem, multidimensional problem, etc. For any optimization algorithm, it is mandatory to verify the performance of the algorithm by comparing it with other existing optimization algorithms. A total of nine BFs are considered to test the ability of our proposed algorithms. With respect to our experimentation, three unimodal BFs ( f n 1 ( x ) f n 3 ( x ) ) and six multimodal BFs ( f n 4 ( x ) f n 9 ( x ) ) are taken into consideration. Out of six multimodal BFs, three ( f n 4 ( x ) f n 6 ( x ) ) are of fixed dimension, and the remaining are multi-dimensional functions. The mathematical expressions of the BFs are listed in Table 2. As mentioned in Table 2, n represents population size, G is the maximum number of iterations, dim refers to the number of dimensions, Range represents the interval of search space, and f n m i n represents the optimal solution with respect to that mathematical function. The parameters considered for this experiment are n, G, dim, and Range.

4.3. Software Defect Prediction (SDP) Framework

To validate the performance of the proposed algorithms, as detailed in Section 4.1, we performed a second experiment on the proposed SDP framework to improve the accuracy of the HBoDP model. The improvisation of accuracy is based on the hyperparameter tuning approach. Figure 5 represents the SDP framework that is followed to carry out our experiment.

4.3.1. Data Source

As for the data source, this paper uses 13 NASA defect datasets (NASA Defect Dataset https://github.com/klainfo/NASADefectDataset, accessed on 5 January 2024) to validate the effectiveness of the hybrid BoAs. A detailed description of the NASA defect dataset is tabulated in Table 3.

4.3.2. Data Pre-Processing

Before inputting the datasets into the ML model, data pre-processing plays an important role in avoiding biases towards any particular features. In this paper, data pre-processing is carried out using four steps, as follows:
  • Label Encoding: Label encoding is a part of data transformation. It is defined as mapping the non-numeric value into the numeric value. In this research work, the dataset consists of yes (Y) and no (N) labels. Yes (Y) and no (N) are mapped to 0 and 1 in this step, respectively.
  • Data Cleaning: In this process, the data are cleaned by removing any outliers, and the inconsistent values are transformed with the mean of the attributes. The outliers are removed and replaced by the inter-quartile range (IQR) method.
  • Feature Selection: As the dataset consists of more features, to avoid bias towards only one kind of feature, feature selection is performed using the Pearson correlation coefficient, which is tabulated in Table 3.
  • Data Scaling: Normally known as normalization, we performed a min-max scaler for normalization.
After performing the data pre-processing step, the pre-processed data were split at a 75–25 ratio into training and testing sets, respectively.

4.3.3. Hybrid Bio-Optimized Defect Prediction (HBoDP) Model

The HBoDP model is established by integrating the novel hybrid BoAs (GFLFGOA, LFGOA-SSA, GFGOA-SSA, GFLFGOA-SSA) with the ML models (ANN, XGB). The model is trained and validated using the proposed hybrid BoA to predict software defects, with performance compared against baseline ML models such as ANN and XGB. This model aims to enhance accuracy while optimizing computational efficiency. To improve the accuracy of the HBoDP model, the hyperparameter tuning approach is applied. As the hyperparameter tuning approach is an optimization problem, BoA comes into the picture to tune the hyperparameters, thus enhancing the accuracy of the model.

5. Results and Discussion

This section focuses on the results and analyzes two different experiments, along with the parameter settings of the SDP framework for conducting the experiments.

5.1. Parameter Settings

This section details the parameter settings of the HBoDP model in two sections: first, a description of the NASA defect dataset, and second, the hyperparameters of the ML models.

5.1.1. NASA Defect Dataset Description

Based on our research work, this article considers 13 NASA defect datasets (NASA Defect Dataset https://github.com/klainfo/NASADefectDataset, accessed on 5 January 2024) to validate the effectiveness of the proposed algorithm of the HBoDP model. The NASA defect dataset is a collection of datasets that have been curated from software projects developed within the National Aeronautics and Space Administration (NASA). Given the nature of NASA’s operations, the software they develop is often safety-critical. Any defects or failures in such software can lead to catastrophic consequences, including mission failures, loss of expensive equipment, or even loss of life. Therefore, understanding and predicting defects in such systems is of great importance. This dataset provides a rich source of information on defects from real-world projects, making it highly relevant for studies aiming to predict failures in SCSs. A detailed description of the NASA defect dataset is tabulated in Table 3.
Based on [66], each instance represents one module from the original source code. A module is a self-contained unit of code that encapsulates one or more functions. Each of these datasets, as specified in Table 3, provides valuable insights into the nature of the defects in different types of software systems.

5.1.2. Hyperparameters of the ML Models

Hyperparameter tuning plays an important part in training the models. The ranges of hyperparameters considered to obtain high accuracy and faster convergence for the ANN and XGB models are tabulated in Table 4 and Table 5, respectively.
The proposed hybrid BoAs, as explained in the Section 4.1, are coded with Python language and run in the Jupyter Notebook environment. The experiments are conducted with the following system configuration as tabulated in Table 6.
All the experiments are conducted by considering the 13 NASA defect datasets, as explained and tabulated in Section 5.1.1 and Table 3, respectively.

5.2. Results

Based on the two different experiments, results are noted in the following Section 5.2.1 and Section 5.2.2.

5.2.1. BF Results

The experiments are conducted for 200 iterations; the values of the mean and SD, along with their ranking, are listed in Table 7 and Table 8, respectively. The convergence rates of the algorithms based on the fitness values are plotted in Figure 6.

5.2.2. SDP Framework Results

The experiments are conducted with the system configuration specified in Table 6 for 100 generations; the results are noted in Table 9, Table 10, Table 11 and Table 12. For evaluating the proposed hybrid algorithm, accuracy is considered a performance metric. To verify the computational effectiveness of the proposed algorithm, this paper compares the runtime of the proposed algorithms with some base algorithms. The runtime of the algorithm justifies the computational effectiveness of the algorithms proposed in this study.

5.3. Analysis

This paper focuses on two different experiments to assess the effectiveness of the proposed algorithms. The first one evaluates the statistical analysis of GFLFGOA, GFLFGOA-SSA, GFGOA-SSA, and LFGOA-SSA with the LFGOA, GFGOA, GOA, and SSA algorithms. The second one is performed to improve the prediction accuracy by optimizing the hyperparameters of the ANN and XGB models. For a comprehensive analysis, the discussion is divided into Section 5.3.1, Section 5.3.2 and Section 5.3.3.

5.3.1. BF Analysis

In order to evaluate the statistical performance of the GFLFGOA, GFLFGOA-SSA, GFGOA-SSA, and LFGOA-SSA algorithms on BFs, the mean and SD values, along with the rank of each method, are taken into account. A total of nine BFs are considered, consisting of unimodal, fixed-dimension multimodal, and multi-dimensional multimodal BFs, as detailed in Section 4.2. The statistical results of BFs are tabulated in Table 7 and Table 8. For fair analysis, each BF is experimented on for 200 iterations, and the convergence rates of the algorithms are plotted in Figure 6. The top rank, indicated in bold, is attained by having the lowest mean and SD values for each BF. To avoid ambiguity, we have categorized our discussion based on the types of statistical analysis.
  • Mean Analysis: For a better understanding of the mean analysis of BFs, we presented the discussion based on the types of BFs.
    • Unimodal BFs ( f n 1 f n 3 ): From Table 7, it can be clearly stated that for f n 2 , GFGOA-SSA ranks first, while for f n 1 and f n 3 , even though SSA ranks first, the convergence rates of the proposed algorithms are relatively faster, as plotted in Figure 6a,c, respectively.
    • Fixed-dimension Multimodal BFs ( f n 4 f n 6 ): GFLFGOA-SSA ranks first, achieving the lowest mean value for f n 6 , as tabulated in Table 7. But in the case of f n 4 and f n 5 , the ranking of the proposed algorithm is the same as the base algorithms. Even though the ranking is the same, the convergence rate is relatively equal to the base algorithms, as visualized in Figure 6d,e.
    • Multi-dimension Multimodal BFs ( f n 7 f n 9 ): Except f n 8 , GFLFGOA-SSA and LFGOA-SSA rank first for f n 7 and f n 9 , respectively. Even though SSA ranks first for f n 8 , the convergence rate is relatively close to GFLFGOA-SSA, as plotted in Figure 6h.
  • SD Analysis: To avoid ambiguity, we elaborated the discussion based on the types of BFs.
    • Unimodal BFs ( f n 1 f n 3 ): In f n 1 and f n 2 , the proposed algorithm (GFLFGOA, GFGOA-SSA) ranks first for SD values, as tabulated in Table 8. In the case of f n 3 , the GFLFGOA-SSA has a faster convergence rate despite the fact that SSA ranks first, as demonstrated in Figure 6c.
    • Fixed-dimension Multimodal BFs ( f n 4 f n 6 ): For f n 6 , GFLFGOA-SSA achieve the lowest SD value, as tabulated in Table 8. In the case of f n 4 , all of the other proposed algorithms have equal convergence rates, despite the fact that SSA ranks first, as plotted in Figure 6d, while in the case of f n 5 , the convergence rate of GFGOA is nearly identical to that of the proposed algorithm, despite the SD being zero.
    • Multi-dimension Multimodal BFs ( f n 7 f n 9 ): From Table 8, it can be clearly stated that any one of our proposed algorithms ranks first for f n 7 f n 9 . GFLFGOA-SSA, GFLFGOA, and LFGOA-SSA achieve the lowest SD value for f n 7 , f n 8 , and f n 9 , respectively.
From Table 7 and Table 8, there are seven out of nine statistical results ranking first, obtained by the proposed algorithms. However, these proposed algorithms give unsatisfactory results on f n 3 and f n 5 . Although the mean and SD results are not satisfactory for some BFs, the proposed algorithms’ convergence rate is relatively equal to that of the base algorithms. Through unimodal BF, we may claim the good exploitation capacity of the proposed algorithm, as it possesses one global optimum. Multimodal functions are more complicated than unimodal functions due to more than one local optimum. Through multimodal function, the proposed algorithm’s exploration capacity is evaluated. So based on the aforementioned discussion, we can justify the performance of the proposed algorithms.

5.3.2. SDP Framework Analysis

To verify the performance and the scalability of the proposed hybrid BoAs, we considered 13 NASA defect datasets, as they cover a wide range of instances, i.e., ranges from 125 to 10,878. Basically, BoA deals with the fitness function, which represents the evaluation of the quality of a candidate solution within the optimization problem. With respect to our research work, i.e., classification problem, the term fitness function used in the BoA is analogous to the evaluation metrics used to assess the performance of ML models, known as accuracy. To justify the effectiveness, the four proposed algorithms are compared with the four base algorithms. Except for accuracy, as noted in Table 9 and Table 11, the experimental runtime is also tabulated in Table 10 and Table 12. The algorithms that provide good accuracy and have the lowest runtime compared with the other base algorithms can be called good optimization algorithms. For a better and more precise analysis, we split our discussion into accuracy and runtime analysis with respect to the SDP problem.
  • Algorithm Accuracy Analysis: From Table 9 and Table 11, it can be clearly stated that for all datasets, there is an enhancement in accuracy when the XGB and ANN models are tuned with BoAs. As tabulated in Table 13, in the case of JM1, MW1, PC1, PC3, PC4, and PC5, our proposed algorithms have better optimization effects as compared to the four base algorithms, while in the cases of CM1, KC3, KC4, MC1, and PC2, the superiority of the proposed algorithm is relatively close to the base approaches. In hyperparameter tuning of the ANN model, GFGOA-SSA, GFLFGOA-SSA, and GFGOA-SSA show better accuracy for JM1, PC3, and PC5, respectively. For the remaining datasets, the superiority of our algorithms is relatively close or equal to the base approaches, as shown in Table 11 and Table 14.
  • Algorithm Runtime Analysis: From Table 10, we can justify that the computational runtime is lowest for all datasets except CM1 and MW1 by embedding either the LF or GF or both concepts into the SSA. For CM1 and MW1, the difference in runtime is not too great compared to LFGOA. For CM1, JM1, KC1, KC4, MC1, MW1, PC1, PC3, and PC5, the base algorithm’s computational runtime is low as compared to the proposed algorithm, as listed in Table 12, whereas for other datasets, our proposed algorithm’s runtime shows superiority.
For better visualization, the lowest runtime and highest accuracy for ML models are noted in Table 13 and Table 14. From the above-stated explanation, we may state that for the XGB model, our proposed algorithm’s runtime is low, whereas for the ANN model, the computational runtime is low only for a few datasets. From this, we may deduce that the runtime is affected by the complexity of the model. As the complexity of the model increases, the runtime may increase.
As for most of the datasets, at least one of our proposed algorithms works better, so based on experimental results and the NFL theorem [14], we may conclude that our proposed algorithms perform better in terms of global search and have higher stability. The improvements in the global search ability are due to the GF and LF concepts embedded with the SSA concept. Through the experimental results, the robustness of our hybrid approach is verified in terms of global search ability and accuracy.
In this paper, we also focus on comparing our proposed algorithms with the other state-of-the-art methods, except for the comparison with the other base algorithms. The observations are noted in Table 15.
From Table 15, it can be clearly stated that our proposed algorithm provides better accuracy for eight datasets than the other state-of-the-art methods. From the above-mentioned observation, we may conclude that our approaches show superiority in runtime and accuracy. Experimental results prove that our hybrid algorithm has better optimization and stable performance. So, the proposed algorithm has certain computational validity.

5.3.3. Computational Complexity Analysis

The computational complexity analysis of the hybrid BoAs, as detailed in Section 4.1, is analyzed in two ways:
  • Time Complexity: In general, the time complexity can be defined as follows:
    Time Complexity = O(Initialization) + [O(Fitness evaluation of each search agent) + O(Position updation of agents) + O(Sorting)] ∗ Maximum iteration.
    Mathematically, this can be coded as follows:
    T i m e C o m p l e x i t y = O ( n d ) + [ O ( n ) + O ( n d ) + O ( n l o g n ) ] G = O ( n d + G n ( 1 + d + l o g n ) )
  • Space Complexity: The maximum number of spaces occupied by the proposed algorithm at any time is decided by the random initialization of the population. So, the space complexity can be calculated as follows:
    S p a c e C o m p l e x i t y = O ( n d )
where d represents the dimension, G represents the maximum iteration, and n represents the size of the population. The above time and space complexity, as presented in Equations (27) and (28), respectively, is the same for GFLFGOA, GFGOA-SSA, LFGOA-SSA, and GFLFGOA-SSA. The above-mentioned time and space complexity, as presented in Equations (27) and (28), respectively, are the same not only for the proposed algorithm, but also for the base BoAs, as all BoAs go through the exploration and exploitation phases.

6. Concluding Remarks and Future Scope

This article presents four hybrid BoAs by employing GOA, SSA, LF, and GF concepts to enhance the exploration and exploitation nature of hybrid algorithms. To validate their performance and stability, experiments are conducted, first on BFs and second on the HBoDP model using the NASA defect dataset. This HBoDP model is established by integrating the proposed hybrid algorithms with the ML models (ANN, XGB). The first set of experiments achieved a good exploration and exploitation capacity of the proposed algorithms, as it performs better for seven BFs as compared to the baseline algorithms. But in the case of two BFs, even though the mean and SD values are lower for the proposed algorithm compared to the baseline algorithm, the convergence rate is relatively equal. The second set of experiments is conducted to tune the ML hyperparameters in the HBoDP model to improve the accuracy.
Achieving higher accuracy and the lowest runtime compared to the baseline algorithms validates the effectiveness of the algorithms proposed in the study. Based on the result and analysis, these proposed algorithms provide a compromise in terms of accuracy for some of the datasets but are effective in terms of runtime. For example, for the ANN model, there will be a change in weights in neurons while rerunning the experiments again. So, this may affect the global fitness values. Sometimes, different tuning parameter values in the optimization methods might lead to significant differences in their performance, which may result in a completely different conclusion. Also, it may lead to a different conclusion if there is a change in hyperparameter range value, such as learning rate, epochs, population size, number of iterations, etc.
So to justify this limitation and prove the correctness of the algorithm, future research should include experiments with these hybrid algorithms on variable dimensions of unimodal and multimodal BFs to prove scalability. The future scope of research also includes the implementation of other ML and DL models to tune the hyperparameters. The implementation of these novel algorithms can be extended to other engineering problems.

Author Contributions

Conceptualization, M.D. and B.R.M.; Methodology, M.D. and N.P.; Software, M.D. and N.P.; Formal analysis, M.D.; Investigation, M.D.; Data curation, M.D.; Writing—original draft, M.D.; Writing—review & editing, M.D., B.R.M. and R.M.R.G.; Visualization, B.R.M.; Supervision, B.R.M. and R.M.R.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Van Thieu, N.; Mirjalili, S. MEALPY: An open-source library for latest meta-heuristic algorithms in Python. J. Syst. Archit. 2023, 139, 102871. [Google Scholar] [CrossRef]
  2. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  3. Bäck, T.; Schwefel, H.P. An overview of evolutionary algorithms for parameter optimization. Evol. Comput. 1993, 1, 1–23. [Google Scholar] [CrossRef]
  4. Eberhart, R.; Kennedy, J. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  5. Meraihi, Y.; Gabis, A.B.; Mirjalili, S.; Ramdane-Cherif, A. Grasshopper optimization algorithm: Theory, variants, and applications. IEEE Access 2021, 9, 50001–50024. [Google Scholar] [CrossRef]
  6. Wu, H.S.; Zhang, F.M. Wolf pack algorithm for unconstrained global optimization. Math. Probl. Eng. 2014, 2014, 465082. [Google Scholar] [CrossRef]
  7. Karaboga, D.; Basturk, B. An artificial bee colony (ABC) algorithm for numeric function optimization. In Proceedings of the IEEE Swarm Intelligence Symposium, Indianapolis, IN, USA, 28–29 September 2006; IEEE: Piscataway, NJ, USA, 2006; Volume 2006. [Google Scholar]
  8. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  9. Hussain, K.; Mohd Salleh, M.N.; Cheng, S.; Shi, Y. Metaheuristic research: A comprehensive survey. Artif. Intell. Rev. 2019, 52, 2191–2233. [Google Scholar] [CrossRef]
  10. Nevendra, M.; Singh, P. Empirical investigation of hyperparameter optimization for software defect count prediction. Expert Syst. Appl. 2022, 191, 116217. [Google Scholar] [CrossRef]
  11. Nematzadeh, S.; Kiani, F.; Torkamanian-Afshar, M.; Aydin, N. Tuning hyperparameters of machine learning algorithms and deep neural networks using metaheuristics: A bioinformatics study on biomedical and biological cases. Comput. Biol. Chem. 2022, 97, 107619. [Google Scholar] [CrossRef]
  12. Lentzas, A.; Nalmpantis, C.; Vrakas, D. Hyperparameter tuning using quantum genetic algorithms. In Proceedings of the 2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI), Portland, OR, USA, 4–6 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1412–1416. [Google Scholar]
  13. Shepperd, M.; Song, Q.; Sun, Z.; Mair, C. Data quality: Some comments on the nasa software defect datasets. IEEE Trans. Softw. Eng. 2013, 39, 1208–1215. [Google Scholar] [CrossRef]
  14. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  15. Zang, H.; Zhang, S.; Hapeshi, K. A review of nature-inspired algorithms. J. Bionic Eng. 2010, 7, S232–S237. [Google Scholar] [CrossRef]
  16. McCall, J. Genetic algorithms for modelling and optimisation. J. Comput. Appl. Math. 2005, 184, 205–222. [Google Scholar] [CrossRef]
  17. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  18. Zhen, L.; Liu, Y.; Dongsheng, W.; Wei, Z. Parameter estimation of software reliability model and prediction based on hybrid wolf pack algorithm and particle swarm optimization. IEEE Access 2020, 8, 29354–29369. [Google Scholar] [CrossRef]
  19. Zulfiqar, M.; Kamran, M.; Rasheed, M.; Alquthami, T.; Milyani, A. Hyperparameter optimization of support vector machine using adaptive differential evolution for electricity load forecasting. Energy Rep. 2022, 8, 13333–13352. [Google Scholar] [CrossRef]
  20. Blume, S.; Benedens, T.; Schramm, D. Hyperparameter optimization techniques for designing software sensors based on artificial neural networks. Sensors 2021, 21, 8435. [Google Scholar] [CrossRef]
  21. Akter, S.; Nahar, N.; ShahadatHossain, M.; Andersson, K. A new crossover technique to improve genetic algorithm and its application to TSP. In Proceedings of the 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’s Bazar, Bangladesh, 7–9 February 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  22. Si, B.; Liu, F.; Li, Y. Metamodel-based hyperparameter optimization of optimization algorithms in building energy optimization. Buildings 2023, 13, 167. [Google Scholar] [CrossRef]
  23. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  24. Zhang, Z.; Han, Y. Discrete sparrow search algorithm for symmetric traveling salesman problem. Appl. Soft Comput. 2022, 118, 108469. [Google Scholar] [CrossRef]
  25. Yang, X.; Liu, J.; Liu, Y.; Xu, P.; Yu, L.; Zhu, L.; Chen, H.; Deng, W. A novel adaptive sparrow search algorithm based on chaotic mapping and t-distribution mutation. Appl. Sci. 2021, 11, 11192. [Google Scholar] [CrossRef]
  26. Ouyang, C.; Qiu, Y.; Zhu, D. Adaptive spiral flying sparrow search algorithm. Sci. Program. 2021, 2021, 1–16. [Google Scholar] [CrossRef]
  27. Liang, Q.; Chen, B.; Wu, H.; Han, M. A novel modified sparrow search algorithm based on adaptive weight and improved boundary constraints. In Proceedings of the 2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS), Chengdu, China, 23–26 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 104–109. [Google Scholar]
  28. Zhao, Q.; Tao, R.; Li, J.; Mu, Y. An improved wolf pack algorithm. In Proceedings of the 2020 Chinese Control And Decision Conference (CCDC), Hefei, China, 22–24 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 626–633. [Google Scholar]
  29. Li, H.; Wu, H. An oppositional wolf pack algorithm for Parameter identification of the chaotic systems. Optik 2016, 127, 9853–9864. [Google Scholar] [CrossRef]
  30. Xiu, Z.; Zhen-Hua, W. Improved Wolf Pack Algorithm Based on Tent Chaotic Mapping and Levy Flight. In Proceedings of the 2017 International Conference on Robots & Intelligent System (ICRIS), Huai’an, China, 15–16 October 2017; pp. 165–169. [Google Scholar] [CrossRef]
  31. Chen, X.; Cheng, F.; Liu, C.; Cheng, L.; Mao, Y. An improved Wolf pack algorithm for optimization problems: Design and evaluation. PLoS ONE 2021, 16, e0254239. [Google Scholar] [CrossRef] [PubMed]
  32. Jadon, S.S.; Tiwari, R.; Sharma, H.; Bansal, J.C. Hybrid artificial bee colony algorithm with differential evolution. Appl. Soft Comput. 2017, 58, 11–24. [Google Scholar] [CrossRef]
  33. Mirjalili, S.; Wang, G.G.; Coelho, L.d.S. Binary optimization using hybrid particle swarm optimization and gravitational search algorithm. Neural Comput. Appl. 2014, 25, 1423–1435. [Google Scholar] [CrossRef]
  34. Li, Z.; Yu, M.; Wang, D.; Wei, H. Using hybrid algorithm to estimate and predicate based on software reliability model. IEEE Access 2019, 7, 84268–84283. [Google Scholar] [CrossRef]
  35. Yang, L.; Li, Z.; Wang, D.; Miao, H.; Wang, Z. Software defects prediction based on hybrid particle swarm optimization and sparrow search algorithm. IEEE Access 2021, 9, 60865–60879. [Google Scholar] [CrossRef]
  36. Saremi, S.; Mirjalili, S.; Lewis, A. Grasshopper optimisation algorithm: Theory and application. Adv. Eng. Softw. 2017, 105, 30–47. [Google Scholar] [CrossRef]
  37. Abualigah, L.; Diabat, A. A comprehensive survey of the Grasshopper optimization algorithm: Results, variants, and applications. Neural Comput. Appl. 2020, 32, 15533–15556. [Google Scholar] [CrossRef]
  38. Razmjooy, N.; Estrela, V.V.; Loschi, H.J.; Fanfan, W. A comprehensive survey of new meta-heuristic algorithms. In Recent Advances in Hybrid Metaheuristics for Data Clustering; Wiley Publishing: Hoboken, NJ, USA, 2019. [Google Scholar]
  39. El-Henawy, I.; Abdelmegeed, N.A. Meta-heuristics algorithms: A survey. Int. J. Comput. Appl. 2018, 179, 45–54. [Google Scholar] [CrossRef]
  40. Arora, S.; Anand, P. Chaotic grasshopper optimization algorithm for global optimization. Neural Comput. Appl. 2019, 31, 4385–4405. [Google Scholar] [CrossRef]
  41. Zhao, S.; Wang, P.; Heidari, A.A.; Zhao, X.; Ma, C.; Chen, H. An enhanced Cauchy mutation grasshopper optimization with trigonometric substitution: Engineering design and feature selection. Eng. Comput. 2022, 38, 4583–4616. [Google Scholar] [CrossRef]
  42. Ewees, A.A.; Gaheen, M.A.; Yaseen, Z.M.; Ghoniem, R.M. Grasshopper optimization algorithm with crossover operators for feature selection and solving engineering problems. IEEE Access 2022, 10, 23304–23320. [Google Scholar] [CrossRef]
  43. Yildiz, B.S.; Pholdee, N.; Bureerat, S.; Yildiz, A.R.; Sait, S.M. Enhanced grasshopper optimization algorithm using elite opposition-based learning for solving real-world engineering problems. Eng. Comput. 2022, 38, 4207–4219. [Google Scholar] [CrossRef]
  44. Feng, Y.; Liu, M.; Zhang, Y.; Wang, J. A dynamic opposite learning assisted grasshopper optimization algorithm for the flexible jobscheduling problem. Complexity 2020, 2020, 8870783. [Google Scholar] [CrossRef]
  45. Qin, P.; Hu, H.; Yang, Z. The improved grasshopper optimization algorithm and its applications. Sci. Rep. 2021, 11, 23733. [Google Scholar] [CrossRef]
  46. Behera, R.K.; Shukla, S.; Rath, S.K.; Misra, S. Software reliability assessment using machine learning technique. In Computational Science and Its Applications–ICCSA 2018: Proceedings of the 18th International Conference, Melbourne, VIC, Australia, 2–5 July 2018, Proceedings, Part V 18; Springer: Berlin/Heidelberg, Germany, 2018; pp. 403–411. [Google Scholar]
  47. Batool, I.; Khan, T.A. Software fault prediction using deep learning techniques. Softw. Qual. J. 2023, 31, 1241–1280. [Google Scholar] [CrossRef]
  48. Jayanthi, R.; Florence, L. Software defect prediction techniques using metrics based on neural network classifier. Clust. Comput. 2019, 22, 77–88. [Google Scholar] [CrossRef]
  49. Fan, G.; Diao, X.; Yu, H.; Yang, K.; Chen, L. Software defect prediction via attention-based recurrent neural network. Sci. Program. 2019, 2019, 6230953. [Google Scholar] [CrossRef]
  50. Batool, I.; Khan, T.A. Software fault prediction using data mining, machine learning and deep learning techniques: A systematic literature review. Comput. Electr. Eng. 2022, 100, 107886. [Google Scholar] [CrossRef]
  51. Sobhana, M.; Preethi, G.S.S.; Sri, G.H.; Sujitha, K.B. Improved Reliability Prediction in Engineering Systems Based on Artificial Neural Network. In Proceedings of the 2022 International Mobile and Embedded Technology Conference (MECON), Noida, India, 10–11 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 455–460. [Google Scholar]
  52. Jindal, A.; Gupta, A. Comparative Analysis of Software Reliability Prediction Using Machine Learning and Deep Learning. In Proceedings of the 2022 Second International Conference on Artificial Intelligence and Smart Energy (ICAIS), Coimbatore, India, 23–25 February 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 389–394. [Google Scholar]
  53. Alghanim, F.; Azzeh, M.; El-Hassan, A.; Qattous, H. Software defect density prediction using deep learning. IEEE Access 2022, 10, 114629–114641. [Google Scholar] [CrossRef]
  54. Clemente, C.J.; Jaafar, F.; Malik, Y. Is predicting software security bugs using deep learning better than the traditional machine learning algorithms? In Proceedings of the 2018 IEEE International Conference on Software Quality, Reliability and Security (QRS), Lisbon, Portugal, 16–20 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 95–102. [Google Scholar]
  55. Wongpheng, K.; Visutsak, P. Software defect prediction using convolutional neural network. In Proceedings of the 2020 35th International Technical Conference on Circuits/Systems, Computers and Communications (ITC-CSCC), Nagoya, Japan, 3–6 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 240–243. [Google Scholar]
  56. Cetiner, M.; Sahingoz, O.K. A comparative analysis for machine learning based software defect prediction systems. In Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 1–3 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–7. [Google Scholar]
  57. Alsaeedi, A.; Khan, M.Z. Software defect prediction using supervised machine learning and ensemble techniques: A comparative study. J. Softw. Eng. Appl. 2019, 12, 85–100. [Google Scholar] [CrossRef]
  58. Li, R.; Zhou, L.; Zhang, S.; Liu, H.; Huang, X.; Sun, Z. Software defect prediction based on ensemble learning. In Proceedings of the 2019 2nd International Conference on Data Science and Information Technology, Seoul, Republic of Korea, 19–21 July 2019; pp. 1–6. [Google Scholar]
  59. Iqbal, A.; Aftab, S.; Ali, U.; Nawaz, Z.; Sana, L.; Ahmad, M.; Husen, A. Performance analysis of machine learning techniques on software defect prediction using NASA datasets. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 300–308. [Google Scholar] [CrossRef]
  60. Malhotra, R. Comparative analysis of statistical and machine learning methods for predicting faulty modules. Appl. Soft Comput. 2014, 21, 286–297. [Google Scholar] [CrossRef]
  61. Parashar, A.; Kumar Goyal, R.; Kaushal, S.; Kumar Sahana, S. Machine learning approach for software defect prediction using multi-core parallel computing. Autom. Softw. Eng. 2022, 29, 44. [Google Scholar] [CrossRef]
  62. Topaz, C.M.; Bernoff, A.J.; Logan, S.; Toolson, W. A model for rolling swarms of locusts. Eur. Phys. J. Spec. Top. 2008, 157, 93–109. [Google Scholar] [CrossRef]
  63. Kamaruzaman, A.F.; Zain, A.M.; Yusuf, S.M.; Udin, A. Levy flight algorithm for optimization problems-a literature review. Appl. Mech. Mater. 2013, 421, 496–501. [Google Scholar] [CrossRef]
  64. Luo, J.; Chen, H.; Xu, Y.; Huang, H.; Zhao, X. An improved grasshopper optimization algorithm with application to financial stress prediction. Appl. Math. Model. 2018, 64, 654–668. [Google Scholar] [CrossRef]
  65. Jamil, M.; Yang, X.S. A literature survey of benchmark functions for global optimisation problems. Int. J. Math. Model. Numer. Optim. 2013, 4, 150–194. [Google Scholar] [CrossRef]
  66. Lemon, B. The Effect of Locality Based Learning on Software Defect Prediction; West Virginia University: Morgantown, WV, USA, 2010. [Google Scholar]
  67. Ali, M.; Mazhar, T.; Arif, Y.; Al-Otaibi, S.; Ghadi, Y.Y.; Shahzad, T.; Khan, M.A.; Hamam, H. Software Defect Prediction Using an Intelligent Ensemble-Based Model. IEEE Access 2024, 12, 20376–20395. [Google Scholar] [CrossRef]
  68. Yadav, S. Software Reliability Prediction by using Deep Learning Technique. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 683–693. [Google Scholar] [CrossRef]
  69. Mumtaz, B.; Kanwal, S.; Alamri, S.; Khan, F. Feature Selection Using Artificial Immune Network: An Approach for Software Defect Prediction. Intell. Autom. Soft Comput. 2021, 29, 669–684. [Google Scholar] [CrossRef]
  70. Odejide, B.J.; Bajeh, A.O.; Balogun, A.O.; Alanamu, Z.O.; Adewole, K.S.; Akintola, A.G.; Salihu, S.A.; Usman-Hamza, F.E.; Mojeed, H.A. An empirical study on data sampling methods in addressing class imbalance problem in software defect prediction. In Computer Science On-Line Conference; Springer: Cham, Switzerland, 2022; pp. 594–610. [Google Scholar]
  71. Balogun, A.O.; Basri, S.; Capretz, L.F.; Mahamad, S.; Imam, A.A.; Almomani, M.A.; Adeyemo, V.E.; Alazzawi, A.K.; Bajeh, A.O.; Kumar, G. Software defect prediction using wrapper feature selection based on dynamic re-ranking strategy. Symmetry 2021, 13, 2166. [Google Scholar] [CrossRef]
  72. Balogun, A.O.; Lafenwa-Balogun, F.B.; Mojeed, H.A.; Adeyemo, V.E.; Akande, O.N.; Akintola, A.G.; Bajeh, A.O.; Usman-Hamza, F.E. SMOTE-based homogeneous ensemble methods for software defect prediction. In Computational Science and Its Applications–ICCSA 2020: Proceedings of the 20th International Conference, Cagliari, Italy, 1–4 July 2020, Proceedings, Part VI 20; Springer: Berlin/Heidelberg, Germany, 2020; pp. 615–631. [Google Scholar]
Figure 1. GFLFGOA flowchart.
Figure 1. GFLFGOA flowchart.
Mathematics 12 02521 g001
Figure 2. LFGOA-SSA flowchart.
Figure 2. LFGOA-SSA flowchart.
Mathematics 12 02521 g002
Figure 3. GFGOA-SSA flowchart.
Figure 3. GFGOA-SSA flowchart.
Mathematics 12 02521 g003
Figure 4. GFLFGOA-SSA flowchart.
Figure 4. GFLFGOA-SSA flowchart.
Mathematics 12 02521 g004
Figure 5. Software defect prediction (SDP) framework.
Figure 5. Software defect prediction (SDP) framework.
Mathematics 12 02521 g005
Figure 6. Convergence rate of benchmark functions (BFs). (a) f n 1 . (b) f n 2 . (c) f n 3 . (d) f n 4 . (e) f n 5 . (f) f n 6 . (g) f n 7 . (h) f n 8 . (i) f n 9 .
Figure 6. Convergence rate of benchmark functions (BFs). (a) f n 1 . (b) f n 2 . (c) f n 3 . (d) f n 4 . (e) f n 5 . (f) f n 6 . (g) f n 7 . (h) f n 8 . (i) f n 9 .
Mathematics 12 02521 g006aMathematics 12 02521 g006b
Table 1. Notations used in flowcharts.
Table 1. Notations used in flowcharts.
SymbolsExplanation
rCurrent iteration
GMaximum iteration
STSafety Threshold [0.5, 1.0]
SDScrounger Sparrow
PDProducer Sparrow
Z b e s t Return the best position and the corresponding fitness value. With respect to our flowchart, as drawn in Figure 1, Figure 2, Figure 3 and Figure 4, Z b e s t returns the current best individual after the initialization step, whereas, after the r < G condition becomes false, Z b e s t returns the global best position and the corresponding fitness value.
Table 2. Benchmark functions [65].
Table 2. Benchmark functions [65].
NameMathematical EquationType(n, G, dim)Range fn min
Quartic f n 1 ( x ) = i = 1 d i x i 4 + r a n d o m [ 0 , 1 ) Unimodal100, 200, 30[−1.28,1.28]0
Rosenbrock f n 2 ( x ) = i = 1 d [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] Unimodal100, 200, 30[−30, 30]0
Schwefel 2.21 f n 3 ( x ) = max 1 i d x i Unimodal100, 200, 30[−100, 100]0
Six Hump Camel f n 4 ( x ) = 4 2.1 x 1 2 + x 1 4 3 x 1 2 + x 1 x 2 + 4 x 2 2 4 x 2 2 Fixed-dimension Multimodal100, 200, 2[−5, 5]−1.0316
Branin f n 5 ( x ) = x 2 5.1 x 1 2 4 π 2 + 5 x 1 π 6 2 + 10 1 1 8 π cos ( x 1 ) + 10 Fixed-dimension Multimodal100, 200, 2[−5, 0, 10, 15]0.397887
Booth f n 6 ( x ) = x 1 + 2 x 2 7 2 + 2 x 1 + x 2 5 2 Fixed-dimension Multimodal100, 200, 2[−10, 10]0
Zakharov f n 7 ( x ) = i = 1 n x i 2 + 1 2 i = 1 n i x i 2 + 1 2 i = 1 n i x i 4 Mullti-dimension Multimodal100, 200, 30[−5, 10]0
Rastrigin f n 8 ( x ) = i = 1 d x 1 2 10 cos ( 2 π x i ) + 10 Multi-dimension Multimodal100, 200, 30[−5.12, 5.12]0
Schaffer 6 f n 9 ( x ) = i = 1 d 0.5 + s i n 2 x i 2 + x i + 1 2 0.5 1 + 0.001 x i 2 + x i + 1 2 2 Multi-dimension Multimodal100, 200, 30[−100, 100]0
Table 3. Dataset description.
Table 3. Dataset description.
DatasetDescriptionTarget FeatureOriginal FeaturesSelected FeaturesInstances
CM1Spacecraft instrument’s softwareDefective3826327
JM1Real-time predictive ground systemlabel221610,878
KC1spacecraft’s ground data system (storage management)Defective22152107
KC3Flight software systemDefective4025194
KC4Software system related to spacecraft operationsDefective4135125
MC1Spacecraft’s data processing systemDefective40309466
MC2spacecraft’s power distribution software systemDefective4027124
MW1Ground data system for a weather satelliteDefective3828250
PC1Air traffic control software systemDefective3826679
PC2Spacecraft’s altitude control software systemDefective3725722
PC3University’s administration software systemDefective38251053
PC4Spacecraft’s orbit determination software systemDefective38271270
PC5Satellite’s ground software systemDefective39261694
Table 4. Hyperparameters of ANN.
Table 4. Hyperparameters of ANN.
HyperparametersLower BoundUpper Bound
Learning rate0.0010.1
Neurons in layer1810
Neurons in layer 2610
Batch size432
Epochs1050
Table 5. Hyperparameters of XGB.
Table 5. Hyperparameters of XGB.
HyperparametersLower BoundUpper Bound
Learning rate0.0010.1
Max depth318
Subsample01
N estimator50200
Table 6. System configuration.
Table 6. System configuration.
SpecificationSize
RAM32 GB
Hard Disk512 GB
ProcessorIntel core I7
OS NameUbuntu
OS Type64-bit
Table 7. Mean value and ranking of eight BoAs (Bold indicates lowest mean value).
Table 7. Mean value and ranking of eight BoAs (Bold indicates lowest mean value).
BFIndexGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
f n 1 Mean 5.73 × 10 1 4.77 × 10 3 6.09 × 10 3 2.76 × 10 3 2.94 × 10 1 7.66 × 10 1 3.69 × 10 1 1 . 81 × 10 3
Rank73425861
f n 2 Mean 1.95 × 10 7 5.85 × 10 2 2.16 × 10 1 3 . 28 × 10 2 5.59 × 10 7 2.58 × 10 6 1.25 × 10 8 1.47 × 10 1
Rank62417583
f n 3 Mean 6.75 × 10 1 4.25 × 10 4 9.61 × 10 4 3.66 × 10 4 8.72 × 10 1 5.98 × 10 1 8.57 × 10 1 2 . 43 × 10 4
Rank63428571
f n 4 Mean 9.67 × 10 1 1.02 × 10 0 1 . 03 × 10 0 9.90 × 10 1 1.02 × 10 0 8.72 × 10 1 9.94 × 10 1 1 . 03 × 10 0
Rank52142631
f n 5 Mean 4.53 × 10 1 3.99 × 10 1 3.99 × 10 1 3.99 × 10 1 3 . 98 × 10 1 5.12 × 10 1 4.02 × 10 1 3 . 98 × 10 1
Rank42221531
f n 6 Mean 1.14 × 10 1 7 . 67 × 10 3 9.82 × 10 3 1.26 × 10 2 4.01 × 10 2 7.91 × 10 2 6.68 × 10 2 8.55 × 10 2
Rank81234657
f n 7 Mean 6.61 × 10 2 9 . 19 × 10 2 1.80 × 10 1 7.30 × 10 0 4.70 × 10 2 8.94 × 10 2 4.41 × 10 2 4.53 × 10 1
Rank71436852
f n 8 Mean 4.22 × 10 2 1.82 × 10 1 2.34 × 10 2 1.33 × 10 1 2.82 × 10 2 4.17 × 10 2 3.34 × 10 2 2 . 49 × 10 7
Rank84235761
f n 9 Mean 1.55 × 10 2 1.40 × 10 4 3 . 04 × 10 7 8.45 × 10 6 2.17 × 10 2 1.43 × 10 2 2.75 × 10 2 7.33 × 10 7
Rank64137582
Table 8. Standard deviation (SD) value and ranking of eight BoAs (Bold indicates lowest SD value).
Table 8. Standard deviation (SD) value and ranking of eight BoAs (Bold indicates lowest SD value).
BFIndexGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
f n 1 SD 7 . 11 × 10 15 3.01 × 10 2 5.05 × 10 2 1.76 × 10 2 2.79 × 10 1 2.84 × 10 14 1.70 × 10 1 9.78 × 10 3
Rank15648273
f n 2 SD 3.02 × 10 7 3.60 × 10 1 2.62 × 10 0 2 . 93 × 10 1 5.14 × 10 7 1.21 × 10 7 8.21 × 10 7 2.03 × 10 0
Rank62417583
f n 3 SD 1.31 × 10 1 3.12 × 10 3 1.01 × 10 2 3.15 × 10 3 6.09 × 10 1 1.87 × 10 1 3.19 × 10 0 5 . 43 × 10 4
Rank72435861
f n 4 SD 2.46 × 10 1 5.52 × 10 2 7.39 × 10 3 1.56 × 10 1 8.72 × 10 2 5.23 × 10 2 3.26 × 10 2 9 . 22 × 10 4
Rank85276431
f n 5 SD 4.29 × 10 2 3.87 × 10 3 9.16 × 10 3 3.63 × 10 4 1.48 × 10 3 0 . 00 × 10 0 1.91 × 10 2 3.31 × 10 4
Rank85634172
f n 6 SD 4.42 × 10 1 1 . 77 × 10 2 2.84 × 10 2 3.86 × 10 2 2.22 × 10 1 5.00 × 10 1 1.30 × 10 1 3.19 × 10 1
Rank71235846
f n 7 SD 8.36 × 10 1 7 . 22 × 10 1 2.20 × 10 2 6.27 × 10 1 1.51 × 10 2 1.21 × 10 2 1.49 × 10 2 6.40 × 10 0
Rank41837562
f n 8 SD 0 . 00 × 10 0 2.21 × 10 0 2.13 × 10 1 1.25 × 10 0 5.60 × 10 1 5.68 × 10 14 2.30 × 10 1 4.47 × 10 7
Rank16458273
f n 9 SD 3.25 × 10 2 1.14 × 10 3 2 . 08 × 10 6 6.28 × 10 5 3.15 × 10 2 8.71 × 10 3 3.76 × 10 2 1.03 × 10 5
Rank74136582
Table 9. Accuracy of XGB.
Table 9. Accuracy of XGB.
DatasetWithout OptimizationGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
CM10.890240.939020.939020.926830.939020.926830.926830.926830.93902
JM10.810660.814710.815070.814340.816910.814340.812130.813970.81654
KC10.863380.876660.876660.870970.876660.867170.878560.869070.87856
KC30.816330.857140.857140.857140.857140.836740.857140.836740.85714
KC40.68750.750.781250.781250.781250.718750.781250.750.75
MC10.994510.996200.996620.996200.996200.995350.996620.996200.99620
MC20.709680.838710.838710.838710.838710.806450.838710.8387100.90323
MW10.880.904760.904760.904760.920640.904760.904760.904760.90476
PC10.911760.929410.929410.929410.935290.917650.923530.929410.92941
PC20.988950.994480.994480.994480.994480.994480.994480.994480.99448
PC30.863640.882580.882580.8750.882580.878790.878790.8750.875
PC40.902520.940250.933060.930820.930820.937110.937110.924530.93082
PC50.778300.820750.808960.811320.816040.813680.816040.811320.81604
Table 10. Runtime (seconds) of XGB.
Table 10. Runtime (seconds) of XGB.
DatasetGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
CM18.638942.622275.978168.167116.800144.6706217.407784.14228
JM149.1133711.1651923.016136.4405520.1178818.97812100.4569332.35944
KC110.9880113.280124.717602.792419.856967.2117330.878826.74143
KC35.169924.896494.928473.615827.550462.6173311.181747.2783
KC43.370442.726151.030548.003153.018727.934597.8999910.81161
MC120.0637716.2893617.454093.1872120.9432710.6191836.697937.54776
MC26.510457.536656.083443.03464.420683.7969819.68635.38492
MW111.619928.840758.754498.915456.905717.6882717.2759614.46098
PC116.158045.143362.14377.263075.3330911.993687.7278213.36885
PC28.654407.131304.984628.606969.586455.799616.6619212.31117
PC315.0014916.574828.229843.4816911.093104.5478925.5978011.45645
PC411.3551912.562719.4651911.6012918.4175414.4198431.3173243.08894
PC514.160403.079288.786627.0295322.220738.2775320.3146821.10666
Table 11. Accuracy of ANN.
Table 11. Accuracy of ANN.
DatasetWithout OptimizationGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
CM10.878050.939020.951220.939020.939020.914630.939020.914630.95122
JM10.796690.814710.815440.813970.815810.815440.815440.813600.81544
KC10.848200.874760.869070.872870.872870.872870.870970.878560.88046
KC30.775510.857140.857140.877550.857140.836740.857140.877550.85714
KC40.56250.718750.718750.81250.781250.718750.843750.718750.75
MC10.994080.995780.995350.995780.995350.994710.995780.995350.99578
MC20.741940.838710.838710.870970.870970.838710.903230.870970.87097
MW10.888890.936510.936510.936510.936510.936510.936510.936510.93651
PC10.894120.941180.947060.947060.935290.929410.941180.947060.93529
PC20.9889511110.9944810.994480.99448
PC30.825760.867420.882580.878790.871210.8750.878790.856060.875
PC40.902520.930820.927670.924530.930820.933960.930820.937110.94339
PC50.740570.811320.804250.806600.813680.806600.806600.806600.80896
Table 12. Runtime (seconds) of ANN.
Table 12. Runtime (seconds) of ANN.
DatasetGFLFGOAGFLFGOA-SSALFGOA-SSAGFGOA-SSALFGOAGFGOAGOASSA
CM114.1514819.0398910.9355814.2583719.8845720.5397249.7011859.26953
JM1329.74139461.88978280.21241258.71992372.32476428.38037100.74072677.38826
KC167.2877540.9303834.2796843.7183833.2037627.49754127.15726370.54866
KC316.9692110.0337716.2472422.7908910.2770112.4828622.3223725.73986
KC413.6539317.9465511.1810110.9665928.478588.8513816.1485230.20686
MC1403.20674185.79484196.88806375.0098138.50669213.44733413.83031505.23935
MC220.731479.08239.8095916.0191812.0044212.0462813.4741826.23941
MW125.698910.8132611.250216.4026813.565279.4440317.9514827.54926
PC126.294124.4064419.0921319.3405324.4802433.9998517.2685158.35074
PC210.2454954.5234733.1708126.8333336.66479101.208622.2087341.70491
PC319.3297933.1006732.4525940.4949814.166631.1965623.1915271.15856
PC428.1179930.8226535.6409932.1514768.5749745.7024531.37875100.55374
PC5287.2792872.066562.8895370.0065827.6253530.3159346.8640691.27525
Table 13. Bio-optimized algorithm analysis (XGB).
Table 13. Bio-optimized algorithm analysis (XGB).
DatasetAccuracy (Highest)Runtime (Lowest)
CM1GFLFGOA-SSA, GFLFGOA, GFGOA-SSA, SSAGFLFGOA-SSA
JM1GFGOA-SSAGFGOA-SSA
KC1GFGOA, SSAGFGOA-SSA
KC3GFGOA, GFLFGOA-SSA, SSA, GFGOA-SSA, GFLFGOAGFGOA
KC4GFGOA-SSA, GFGOA, LFGOA-SSA, GFLFGOA-SSALFGOA-SSA
MC1GFLFGOA-SSA, GFGOAGFGOA-SSA
MC2SSAGFGOA-SSA
MW1GFGOA-SSALFGOA
PC1GFGOA-SSALFGOA-SSA
PC2GFLFGOA, GFLFGOA-SSA, LFGOA-SSA, GFGOA-SSA, LFGOA, GFGOA, GOA, SSALFGOA-SSA
PC3GFGOA-SSA, GFLFGOA, GFLFGOA-SSAGFGOA-SSA
PC4GFLFGOALFGOA-SSA
PC5GFLFGOAGFLFGOA-SSA
Table 14. Bio-optimized algorithm analysis (ANN).
Table 14. Bio-optimized algorithm analysis (ANN).
DatasetAccuracy (Highest)Runtime (Lowest)
CM1GFLFGOA-SSA, SSAGOA
JM1GFGOA-SSAGOA
KC1SSAGFGOA
KC3LFGOA-SSA, GOAGFLFGOA-SSA
KC4GFGOAGFGOA
MC1GFLFGOA, LFGOA-SSA, GFGOA, SSALFGOA
MC2GFGOAGFLFGOA-SSA
MW1GFLFGOA, GFLFGOA-SSA, LFGOA-SSA, GFGOA-SSA, LFGOA, GFGOA, GOA, SSAGFGOA
PC1GFLFGOA-SSA, LFGOA-SSA, GOAGOA
PC2GFGOA-SSA, GFGOA, GFLFGOA, GFLFGOA-SSA, LFGOA-SSAGFLFGOA
PC3GFLFGOA-SSALFGOA
PC4SSAGFLFGOA
PC5GFGOA-SSALFGOA
Table 15. Comparison of state-of-the-art methods (accuracy).
Table 15. Comparison of state-of-the-art methods (accuracy).
DatasetsCM1JM1KC1KC3KC4MC1MC2MW1PC1PC2PC3PC4PC5
[67]0.86870.7912----0.68420.89330.9216-0.87970.8714-
[57]0.830.78----0.68-0.91-0.840.84-
[56]0.8780.8030.850-----0.922----
[68]-0.890.84--0.95--0.850.860.830.890.91
[48]-0.810.79-------0.89--
[59]0.77550.7396----0.64860.8266--0.82590.8608-
[69]0.8179-------0.8979----
[70]----------0.8192--
[71]------0.68320.6005--0.7370.869-
[72]-----------0.8661-
[53] (GA-SVM)0.9035-0.85140.79890.90350.99500.67100.91830.93670.99590.90140.8821-
[53] (PSO-SVM)0.9037-0.85140.90400.68850.99500.67060.91790.93670.99590.90210.8793-
[53] (GAPSO-SVM)0.9049-0.84380.89950.67820.99510.67830.91570.93860.99590.90400.8779-
Proposed (GFLFGOA-XGB)0.93900.81470.87670.85710.750.99620.83870.90480.92940.99450.88260.94030.8208
Proposed (GFLFGOA-SSA-XGB)0.93900.81510.87670.85710.78130.99660.83870.90480.92940.99450.88260.93310.8090
Proposed (LFGOA-SSA-XGB)0.92680.81430.87100.85710.78130.99620.83870.90480.92940.99450.8750.93080.8113
Proposed (GFGOA-SSA-XGB)0.93900.81690.87670.85710.78130.99620.83870.92060.93530.99450.88260.93080.8160
Proposed (GFLFGOA-ANN)0.93900.81470.87480.85710.71880.99580.83870.93650.94121.00.86740.93080.8113
Proposed (GFLFGOA-SSA-ANN)0.95120.81540.86910.85710.71880.99540.83870.93650.94711.00.88260.92770.8043
Proposed (LFGOA-SSA-ANN)0.93900.81400.87290.87760.81250.99580.87100.93650.94711.00.87880.92450.8066
Proposed (GFGOA-SSA-ANN)0.93900.81580.87290.85710.78130.99540.87100.93650.93531.00.87120.93080.8137
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Das, M.; Mohan, B.R.; Guddeti, R.M.R.; Prasad, N. Hybrid Bio-Optimized Algorithms for Hyperparameter Tuning in Machine Learning Models: A Software Defect Prediction Case Study. Mathematics 2024, 12, 2521. https://doi.org/10.3390/math12162521

AMA Style

Das M, Mohan BR, Guddeti RMR, Prasad N. Hybrid Bio-Optimized Algorithms for Hyperparameter Tuning in Machine Learning Models: A Software Defect Prediction Case Study. Mathematics. 2024; 12(16):2521. https://doi.org/10.3390/math12162521

Chicago/Turabian Style

Das, Madhusmita, Biju R. Mohan, Ram Mohana Reddy Guddeti, and Nandini Prasad. 2024. "Hybrid Bio-Optimized Algorithms for Hyperparameter Tuning in Machine Learning Models: A Software Defect Prediction Case Study" Mathematics 12, no. 16: 2521. https://doi.org/10.3390/math12162521

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop