Next Article in Journal
Optimal Open-Loop Control of Discrete Deterministic Systems by Application of the Perch School Metaheuristic Optimization Algorithm
Next Article in Special Issue
Agglomerative Clustering with Threshold Optimization via Extreme Value Theory
Previous Article in Journal
A Review of Modern Audio Deepfake Detection Methods: Challenges and Future Directions
Previous Article in Special Issue
Partitioning of Transportation Networks by Efficient Evolutionary Clustering and Density Peaks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Binary Horse Optimization Algorithm for Feature Selection

Independent Researcher, 405200 Dej, Romania
Algorithms 2022, 15(5), 156; https://doi.org/10.3390/a15050156
Submission received: 11 April 2022 / Revised: 29 April 2022 / Accepted: 30 April 2022 / Published: 6 May 2022
(This article belongs to the Special Issue Nature-Inspired Algorithms in Machine Learning)

Abstract

:
The bio-inspired research field has evolved greatly in the last few years due to the large number of novel proposed algorithms and their applications. The sources of inspiration for these novel bio-inspired algorithms are various, ranging from the behavior of groups of animals to the properties of various plants. One problem is the lack of one bio-inspired algorithm which can produce the best global solution for all types of optimization problems. The presented solution considers the proposal of a novel approach for feature selection in classification problems, which is based on a binary version of a novel bio-inspired algorithm. The principal contributions of this article are: (1) the presentation of the main steps of the original Horse Optimization Algorithm (HOA), (2) the adaptation of the HOA to a binary version called the Binary Horse Optimization Algorithm (BHOA), (3) the application of the BHOA in feature selection using nine state-of-the-art datasets from the UCI machine learning repository and the classifiers Random Forest (RF), Support Vector Machines (SVM), Gradient Boosted Trees (GBT), Logistic Regression (LR), K-Nearest Neighbors (K-NN), and Naïve Bayes (NB), and (4) the comparison of the results with the ones obtained using the Binary Grey Wolf Optimizer (BGWO), Binary Particle Swarm Optimization (BPSO), and Binary Crow Search Algorithm (BCSA). The experiments show that the BHOA is effective and robust, as it returned the best mean accuracy value and the best accuracy value for four and seven datasets, respectively, compared to BGWO, BPSO, and BCSA, which returned the best mean accuracy value for four, two, and two datasets, respectively, and the best accuracy value for eight, seven, and five datasets, respectively.

1. Introduction

Algorithms inspired by nature have recently been considered in various categories of machine learning and deep learning problems, ranging from the selection of optimal parameters for deep neural networks [1] to the optimal hyper-parameter tuning of Support Vector Machine (SVM) classifiers [2] for medical diagnosis problems. The results obtained using bio-inspired heuristics are often comparable to the ones obtained using classical approaches such as random search or grid search.
The application of population-based bio-inspired algorithms [3] is competitive, due to the many optimization problems from various research areas, such as scheduling, routing, power, and energy or communication issues. These bio-inspired algorithms are often adapted using improved versions, mutation operators, binary modifications, chaos theory, quantum theory, or hybridization with other algorithms.
The sources of inspiration for these algorithms are various, ranging from the echolocation behavior of bats in the case of the Bat Algorithm (BA) [4] and the behavior of wolves in the case of the Grey Wolf Optimizer (GWO) [5] to the duration of the survival of trees in forests in the case of the Forest Optimization Algorithm (FOA) [6] and tree growth in the case of the Tree Growth Algorithm (TGA) [7]. Parts of the older, well-known bio-inspired algorithms are also sources of inspiration for novel algorithms inspired by nature, which are often much better for various categories of optimization problems and result in better results for specific benchmark functions.
The application of bio-inspired heuristics has led to improved performance in the case of various machine learning algorithms. The authors of [8] used a Particle Swarm Optimization (PSO) [9] Random Forest (RF) approach for the accurate diagnosis of spontaneous ruptures of ovarian endometriomas (OE). Compared to all other models, the PSO-RF approach led to the most accurate results. The authors of [10] improved the accuracy of the Naïve Bayes (NB) algorithm for hoax classification using PSO, while the authors of [11] used an integrated approach of NB and PSO for email spam detection. In [12], an improved weighted K-Nearest Neighbors (K-NN) algorithm which was based on PSO was applied in wind power system state recognition. PSO was applied in that approach in the optimization of the weight and the k parameter. The authors of [13] applied the Artificial Flora Algorithm (AFA) [14] for feature selection and Gradient Boosted Trees (GBT) for classification of diabetes data. Logistic Regression (LR) and PSO were applied in [15] in the optimization of the reliability of a bank.
Bio-inspired algorithms were applied in various types of machine learning optimization problems, ranging from the optimization of the sigma, C, and epsilon parameters of a Support Vector Regression (SVR) algorithm [16] and feature selection in classification problems [17], to tuning machine learning ensembles weights [18] and the selection of cluster centers in clustering problems [19].
Interest in the application of swarm intelligence algorithms to feature selection increased a great deal in the last decades, as can be seen in [20], where the review results showed that the number of articles searched in Google Scholar hitting the terms “swarm intelligence” and “feature selection” per year increased significantly in recent years.
The application of bio-inspired algorithms in feature selection for classification problems using machine learning therefore represents a good alternative to other methods considered in the research literature. A few state-of-the-art approaches for feature selection which do not involve bio-inspired algorithms are fuzzy entropy measures [21], feature distribution measures [22], and recursive feature elimination [23].
The principal contributions of this article are:
(1)
Presentation of the main steps of the original version of the Horse Optimization Algorithm (HOA) introduced in [24];
(2)
Adaptation of the HOA to the Binary Horse Optimization Algorithm (BHOA) for feature selection optimization problems;
(3)
Application of the BHOA to nine state-of-the-art datasets from the UCI machine learning repository [25], which are representative for classification problems, using the classification algorithms RF, SVM, GBT, LR, K-NN, and NB;
(4)
Comparison of the results with the ones obtained using three other bio-inspired approaches, based on binary versions of the GWO, PSO, and Crow Search Algorithm (CSA) [26].
The structure of this article is as follows: Section 2 presents related work and is organized in two parts; the first introduces the HOA [24] and presents its applications, while the second presents related work in the context of bio-inspired approaches for feature selection. Section 3 presents the methods and materials and is organized in three parts; the first describes the high-level view of the HOA, the second shows how the HOA is adapted to the BHOA for feature selection, and the third presents the BHOA machine learning methodology for feature selection. Section 4 shows the results obtained using the BHOA-based methodology. Section 5 discusses the results, comparing them to the ones obtained using three other bio-inspired approaches based on the GWO, PSO, and CSA. Finally, Section 6 presents the conclusions.

2. Related Work

2.1. The Horse Optimization Algorithm and Its Applications

The HOA is a novel bio-inspired algorithm introduced in [24], having its main source of inspiration the characteristics of horse herds. The characteristics considered in the original article were: (1) the organization of the horses in herds such that each herd is led by a dominant stallion, (2) the replacement of the dominant stallions of the herds by better rival stallions, (3) the variety of gaits, (4) the consideration of running as the best defense mechanism, (5) the excellent long term memory of horses, (6) the hierarchical organization of the herds, which regulates access to various resources such as shelter, water, and food, and (7) the redistribution of horses with low fitness values.
Compared to other bio-inspired algorithms which consider the properties of horses such as the Horse Herd Optimization Algorithm (HHOA) [27] and the Wild Horse Optimizer (WHO) [28], the proposed HOA is different because it shares similarities with PSO, Chicken Swarm Optimization (CSO) [29], and the Cat Swarm Optimization Algorithm (CSOA) [30]. Representative influences of PSO in the development of the HOA were swarm representation, the use of velocities and positions, and the formulas for the updating of velocities and positions. CSO influenced the HOA in terms of properties such as the reorganization of herds at various frequencies, the organization of the horses in different categories according to their fitness values, and the updating of their memory using the Gaussian operator. From the CSOA, the HOA considered the use of memory, movement of the horses towards the center of the herd rather than towards the best position of a cat, and the consideration of two types of equations for updating positions depending on whether the horses belong to a herd or not, which were inspired by the two behaviors of cats, namely, the tracing mode and the seeking mode.
The HOA was applied in [24] in feature selection. Another application of the HOA is the one from [31] where the HOA was considered in the tuning of the dropout and recurrent dropout parameters of a Long Short-Term Memory (LSTM) Recurrent Neural Network (RNN) studied in the classification of epileptic seizures. Compared to the case where the default values of the two parameters were used, the HOA-RNN method led to better precision, recall, F1 score, and accuracy values.

2.2. Bio-Inspired Approaches for Feature Selection

The approach presented in [32] applied an adapted binary version of PSO and two other bio-inspired approaches, Genetic Algorithm [33] and Differential Evolution [34], in feature selection for daily life activities. The data used in experiments were from the Daily Life Activities (DaLiAc) dataset [35] from the UCI machine learning repository. Other feature selection approaches which were compared in that article were RF, Backward Features Elimination (BFE), and Forward Feature Selection (FFS). The applied classifier was RF. The proposed binary PSO approach showed results similar to the ones obtained when classical feature selection approaches were applied.
Finally, the machine learning method presented in [36] used a multi-objective version of GWO for feature selection, using as experimental support the appliances energy prediction dataset from the UCI machine learning repository and other UCI datasets. The two objectives were the minimization of the number of selected features and the maximization of the prediction performance. For eight of the nine datasets, the GWO approach returned the best results compared to other multi-objective approaches based on PSO and GA.

3. Methods

3.1. HOA

Figure 1 presents the high-level view of the original version of the HOA [24], describing the main steps, configuration parameters, and instruction flow.
Some of the configurable parameters of the HOA are common to other bio-inspired algorithms, such as: the number of iterations ( I ), the number of dimensions of the search space ( D ), the size of the population ( N ), and the reorganization frequency ( M ). On the other hand, some of the configurable parameters are specific to the HOA, such as: the dominant stallions percent ( D S P ) , the horse distribution rate ( H D R ) , the horse memory pool ( H M P ), and the single stallions percent ( S S P ).
The main steps of the HOA are described in more detail next:
  • Step 1. N horses { h 1 , , h N } are initialized randomly in the D -dimensional search space.
  • Step 2. The algorithm computes the fitness value of each horse using the objective function and updates the value of the best horse h g b e s t .
  • Step 3. The value of i t e r a t i o n is initialized to 0 .
If i t e r a t i o n is less than I , then there are two cases: if i t e r a t i o n modulo M equals 0 , then the algorithm continues with Step 4, otherwise the algorithm continues with Step 8. If i t e r a t i o n is greater than or equal to I , then the algorithm continues with Step 18.
  • Step 4. The best N × D S P horses according to the fitness value are the leaders of the herds from the set T of newly initialized herds such that | T | = | N × D S P | .
  • Step 5. The next best N × S S P horses according to the fitness value are single stallions and form the set S .
  • Step 6. The remaining horses are assigned randomly to herds from T .
  • Step 7. The worst N × H D R horses in terms of fitness value are distributed randomly in the D -dimensional search space, their fitness values are recomputed, and the value of h g b e s t is updated accordingly.
  • Step 8. For each herd { h 1 , , h K } of size K from T , the ranks are from { 1 K , 2 K , , 1 } such that the highest rank values are assigned to the horses with the best fitness values, which are the minimum values in case of minimization problems or the maximum values in case of maximization problems. If two horses with indices i , j from { 1 , , K } have the same fitness value, such that i < j , then the horse with the index j has a higher rank than the horse with the index i . The center of the herd is computed as the weighted arithmetic mean of the positions of the horses such that the weights are the ranks.
    The instructions from Step 9 to Step 14 are performed for each horse.
  • Step 9. The gait is updated to a random value from [ 1 ,   2 ] .
If the horse is single then its velocity is updated in Step 10, otherwise its velocity is updated in Step 11.
  • Step 10. The formula for updating the position for a single stallion is:
    v i t e r a t i o n + 1 = v i t e r a t i o n + r × h g a i t × ( n h e r d c e n t e r x i t e r a t i o n ) ,
    such that v i t e r a t i o n + 1 , v i t e r a t i o n are the values of the velocity in i t e r a t i o n + 1 and i t e r a t i o n , r is a random value from [ 0 ,   1 ] , h g a i t is the gait of the horse h , n h e r d c e n t e r is the center of the nearest herd, namely, the one for which the Euclidean distance between the position of the stallion and the center of the herd is minimum, and x i t e r a t i o n is the position of the stallion in the current i t e r a t i o n .
  • Step 11. If the horse belongs to a herd, then its velocity is updated using the formula:
    v i t e r a t i o n + 1 = v i t e r a t i o n + h r a n k × h g a i t × ( h e r d c e n t e r x i t e r a t i o n ) ,
    such that h r a n k and h e r d c e n t e r are the rank of h and the center of the herd, respectively.
  • Step 12. The formula for the updating of the position is:
    x i t e r a t i o n + 1 = x i t e r a t i o n + v i t e r a t i o n + 1 ,
    where x i t e r a t i o n + 1 and x i t e r a t i o n represent the positions in i t e r a t i o n + 1 and i t e r a t i o n , respectively.
  • Step 13. The H M P -dimensional memory pool is updated as follows:
    M i t e r a t i o n + 1 = [ m 1 , 1 i t e r a t i o n + 1 m 1 , D i t e r a t i o n + 1 m H M P , 1 i t e r a t i o n + 1 m H M P , D i t e r a t i o n + 1 ] ,
    where for any k { 1 , , H M P } the formula for the updating of memory is:
    m k i t e r a t i o n + 1 = x i t e r a t i o n + 1 × N ( 0 , 1 ) ,
    such that N ( 0 , 1 ) is the normal distribution with mean 0 and standard deviation 1 .
  • Step 14. The new fitness value of the horse is computed as the best value between the fitness value of the position x i t e r a t i o n + 1 and the fitness values of the M i t e r a t i o n + 1 matrix elements. If an element of M i t e r a t i o n + 1 has a better fitness value than x i t e r a t i o n + 1 , then x i t e r a t i o n + 1 is updated to the value of that element. If x i t e r a t i o n + 1 has a better fitness value than h g b e s t , then h g b e s t is updated to x i t e r a t i o n + 1 .
  • Step 15. The instructions performed in this step are the ones performed in Step 8.
  • Step 16. For each single stallion from S , the nearest herd is determined. If the single stallion has a better fitness value than the stallion of the herd, then the two stallions are swapped as follows: (1) the single stallion becomes the leader of the herd, (2) the leader of the herd becomes a single stallion, and (3) the positions of the two stallions are switched.
  • Step 17. The value of i t e r a t i o n is incremented by 1 .
  • Step 18. The algorithm returns the value of h g b e s t .

3.2. BHOA for Feature Selection

The BHOA is adapted for feature selection, such that the positions are represented by arrays of 0 s and 1 s, where the number of 1 s is at least two to avoid exceptional cases when the classifiers cannot be applied to data samples with one feature or no features.
The function which is used for converting the continuous values in binary values is:
S ( x ) = { 1 ,     r a n d < | x 1 + x 2 | 0 ,     otherwise ,
where the function r a n d returns random values from [ 0 ,   1 ] .
The pseudo-code of the BHOA for feature selection is presented next (see Algorithm 1).
Algorithm 1. BHOA for feature selection.
1: Input I , D , N , M , D S P , H D R , H M P , S S P , O F
2:Output  h g b e s t
3: initialize a population of N horses in the D -dimensional space
4: adjust the positions of the horses
5: apply O F to compute the fitness values of the horses and update h g b e s t
6:for  i t e r a t i o n = 0   to I 1  do
7:  if  i t e r a t i o n   mod   M = = 0  then
8:   the best N × D S P horses are leaders of herds from T
9:   the next best N × S S P horses represent set S
10:   the remaining horses are distributed randomly to herds from T
11:   the worst N × H D R horses are positioned randomly
12:  end if
13:  compute the horse ranks and the left of each herd from T
14:  for each horse do
15:   update the gait
16:   if single then
17:    update the velocity using Formula (1)
18:   else
19:    update the velocity using Formula (2)
20:   end if
21:   update the position by applying the Formula (6) to the velocity and adjust it
22:   update the H M P -dimensional memory using the Formulas (4)–(6) and adjust it
23:   update the new fitness value, the position, and h g b e s t
24:  end for
25:  compute the horse ranks and the left of each herd from T
26:  swapping stallions operations
27:end for
28:return  h g b e s t
The inputs of the BHOA (line 1) are the same as in the case of the HOA, where O F is the objective function. The output is the position of the best horse, h g b e s t (line 2).
The population of N horses is initialized in the D -dimensional search space (line 3) with binary values. The position of each horse is adjusted in line 4 such that if it contains less than two 1′s and the rest are only 0′s, then the position is updated to a new array with two 1′s selected at random positions, the rest of the elements of the array being filled with 0′s. O F is applied in line 5 to compute the fitness of the horses and h g b e s t is updated to the position of the horse with the best fitness value.
The instructions which correspond to lines 7–26 are performed I times. If i t e r a t i o n is divisible by M (line 7), then the horses are reorganized as follows: the best N × D S P horses are leaders of herds from the set T (line 8), the next N × S S P horses compose the set S (line 9), and the remaining horses are assigned randomly to herds from T (line 10). In line 11, the worst N × H D R horses in terms of fitness value are positioned randomly in the D -dimensional search space such that their positions are represented by new arrays of 0 s and 1 s initialized randomly. The new positions are adjusted such that the number of 1 s is at least two for each position. The new fitness values of the worst horses are computed using O F and h g b e s t is updated accordingly.
The horse ranks and the center of each herd from the set T are evaluated in line 13.
The instructions from lines 15–23 are performed for each horse. The gait is updated to a random number from [ 1 ,   2 ] (line 15).
If the horse is single (line 16), then it updates its velocity using Formula (1), otherwise it updates its velocity using Formula (2). Formula (6) is applied in (line 21) to the velocity value to calculate the new value of the position. Then, the position is adjusted such that it contains at least two 1 s.
In line 22, the H M P -dimensional memory is updated using Formulas (4)–(6) such that Formula (6) is used to binarize the values computed using Formulas (4) and (5). Then, the values of the H M P -dimensional memory are adjusted such that each element of the memory has at least two 1 s.
In line 23, the new fitness value is updated to the best value between the fitness value of the position and the fitness values of the elements of the H M P -dimensional memory. Moreover, the position is also updated to the position of the element of the H M P -dimensional memory with the best fitness value. The value of h g b e s t is updated to the value of the position if the position has a better fitness value.
The instructions from line 25 for the computation of the horse ranks and the centers of the herds from T are the same as the ones from line 13. Swapping stallions operations are performed in line 26 for each single stallion from the set S . For each single stallion from S , if the leader of the nearest herd computed using the Euclidean distance has a better fitness value, then the two stallions are swapped such that the leader of the nearest herd becomes single, the single stallion becomes the new leader of the nearest herd, and the positions of the two stallions are swapped.
The value of h g b e s t is returned in line 28.

3.3. BHOA Machine Learning Methodology for Feature Selection

The proposed BHOA machine learning methodology for feature selection is developed for classification problems, such that the applicable datasets are characterized by the following properties:
(1)
the data samples have the same number of features;
(2)
the number of features is greater than or equal to 2 ;
(3)
each data sample has a label which is an integer value greater than or equal to 0 ;
(4)
the values corresponding to the features are real numerical values.
Figure 2 presents the high-level view of the BHOA machine learning methodology for feature selection.
The proposed methodology has the following steps:
  • Step 1. The input data is represented by a dataset characterized by data samples with an equal number of features, which is a number greater than 1 , a label which is an integer value greater than or equal to 0 , and features described by real numerical values.
  • Step 2. In the normalization phase, the values corresponding to the features are normalized to take values from [ 0 ,   1 ] , while the labels are not altered.
  • Step 3. In the holdout cross-validation phase, the input data is split randomly into 80% training data and 20% testing data.
  • Step 4. The BHOA algorithm is applied in feature selection. For each running of the algorithm, a different C l a s s i f i e r from the possible classifiers RF, SVM, GBT, LR, K-NN, and NB, is selected. The objective function O F is adapted to the following formula:
    O F ( x ) = 1 4 × x s e l e c t e d   f e a t u r e s x a l l   f e a t u r e s + 3 4 × R M S E ( C l a s s i f i e r ,   x ) ,
    where x s e l e c t e d   f e a t u r e s is the number of selected features which is equal to the number of 1 s of the position x , x a l l   f e a t u r e s is equal to D , which is the number of dimensions of the search space, and R M S E ( C l a s s i f i e r ,   x ) is the Root Mean Squared Error (RMSE), computed using as parameters the testing data and the values predicted by the C l a s s i f i e r when the features were selected according to x s e l e c t e d   f e a t u r e s .
  • Step 5. The best horse is the one for which the value of O F ( x ) is minimal, as the objective of the optimization problem is to obtain the best prediction results using as few features as possible.
  • Step 6. The C l a s s i f i e r performance is evaluated using the standard classification metrics:
    a c c u r a c y = T P + T N T P + T N + F P + F N ,
    p r e c i s i o n = T P T P + F P ,
    F 1   s c o r e = 2 × T P 2 × T P + F P + F N ,
    where T P , T N , F P , and F N are the true positives, true negatives, false positives, and false negatives for binary classification problems. The r e c a l l metric was not considered because for multi-label classification problems the value of the w e i g h t e d   r e c a l l is equal to the value of the a c c u r a c y . In particular, the values for p r e c i s i o n and F 1   s c o r e are computed as weighted values.

4. Results

The experiments were performed on a machine with the following characteristics: Intel(R) Core(TM) i7-7500 CPU @ 2.70 GHz processor, 8.00 GB installed RAM, 64-bit operating system, and x64-based processor.
For each dataset, a single experiment was run for each classifier from the possible classifiers RF, SVM, GBT, LR, K-NN, and NB, respectively. Each algorithm was trained using the training data and the classification metrics were computed using the testing data and the data predicted by the algorithm. No validation set was included in the experiments, as the methodology considered the simplest case of cross-validation, namely, holdout cross-validation, such that the set of all samples is represented by the union of the training data samples and the testing data samples.
Table 1 presents the sklearn configurations of the classifiers used in experiments.
The configurations of the classifiers were selected after a series of in-house experiments with the objective of improving the running time of the classifiers compared to the cases where the default values of those parameters were used. In the case of K-NN, the value 9 was selected, bearing in mind that some studies suggest that an optimal value of K is the nearest integer to n s a m p l e s , where n s a m p l e s is the number of samples of the training data. The dataset with the minimum number of samples, namely, Breast Cancer Coimbra Data [37], had n s a m p l e s = 93 for the training data; consequently a value of K equal to [ 93 ] = [ 9.643650 ] = 9 was used in the experiments. The value of this parameter led to good results for the other experimental datasets, therefore K = 9 was applied for all the experimental datasets. The value 100,000 for the parameter var smoothing for NB and the value 0.001 for the parameter learning rate for GBT were applied only to the Cervical Cancer (Risk Factors) Data Set [38]. For all other experimental datasets, the default value for the var smoothing parameter was considered for NB and a 1.0 learning rate value was considered for GBT.

4.1. Experimental Data Description

The datasets used in the experiments are presented next. All nine datasets were taken from the UCI machine learning repository, are representative for classification problems, and are from various research domains. The datasets were preprocessed for the proposed methodology as follows:
(1)
Smart Grids Data: The original data is the Electrical Grid Stability Simulated Data Dataset. The dataset was generated for a smart grid with one producer and three consumers, and the features describe the reaction time of the participants, the power consumed/produced, and the gamma coefficients of the price elasticity. The 13th feature, which represents the maximal real part corresponding to the characteristic equation root, is removed. The 14th feature is represented by the labels column. The values ‘stable’ and ‘unstable’ are converted into 1 and 0, respectively.
(2)
Raisin Data [39]: This dataset contains information from images of Besni and Kecimen raisin varieties that were grown in Turkey. The features describe the area, perimeter, major axis length, minor axis length, eccentricity, convex area, and extent. The labels column is represented by the ‘class’ column. The labels ‘Kecimen’ and ‘Besni’ were converted into 0 and 1, respectively.
(3)
Turkish Music Emotion Data [40]: The data are developed as a discrete model. The four classes of this data are ‘happy’, ‘sad’, ‘angry’, and ‘relaxed’. The dataset is prepared using non-verbal music and verbal music from various genres of Turkish music. The labels ‘happy’, ‘sad’, ‘angry’, and ‘relaxed’ are converted into 1, 2, 3, and 0, respectively.
(4)
Diabetes Risk Prediction Data [41]: The original data are from the Early Stage Diabetes Risk Prediction Dataset. The features contain information about the age of the monitored subjects and various conditions such as obesity, visual blurring, and delayed healing. The labels column is called ‘class’, and the two types of possible labels are ‘positive’ and ‘negative’. The categorical values are converted to numerical values as follows: ‘female’, ‘yes’, and ‘positive’ are converted to 1, while ‘male’, ‘no’, and ‘negative’ are converted to 0.
(5)
Rice Data [42]: The Rice (Cammeo and Osmancik) Data Set contains data about the Osmancik and Cammeo species of rice. For each grain of rice, seven morphological features were extracted, namely, the area, perimeter, major axis length, minor axis length, extent, convex area, and eccentricity. The ‘class’ column represents the labels. The values ‘Cammeo’ and ‘Osmancik’ were converted into 0 and 1, respectively.
(6)
Parkinson’s Disease Data [43]: The Parkinson’s Disease Classification Data Set was gathered from 108 patients with Parkinson’s Disease, as follows: 81 women and 107 men. The 64 healthy individuals are represented by 41 women and 23 men. The column ‘id’ was removed in the preprocessed data.
(7)
Cervical Cancer Data: The Cervical Cancer (Risk Factors) Data Set contains data which were collected at the Hospital Universitario de Caracas in Caracas, Venezuela. The data were collected from 858 patients, and describe demographic information, historical medical records, and habits. The missing values were replaced using the mean heuristic. The four target variables ‘Hinselmann’, ‘Schiller’, ‘cytology’, and ‘biopsy’ were converted into a single target variable equal to 2 3 × H i n s e l m a n n + 2 2 × S c h i l l e r + 2 1 × C y t o l o g y + 2 0 × B i o p s y . This conversion led to the following set of labels: {0, 1, 2, 3, 4, 5, 6, 7, 8, 12, 13, 14, 15}. Then, the labels {12, 13, 14, 15} were converted into {9, 10, 11, 12}, such that the final labels were {0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12}.
(8)
Chronic Kidney Disease Data: These data can be used to predict chronic kidney disease, and the attributes contain information about various characteristics such as age, blood pressure, sugar, specific gravity, and red blood cell count. The values ‘abnormal’, ‘notpresent’, ‘no’, ‘poor’, and ‘notckd’ were converted to 0, while the categorical values ‘normal’, ‘present’, ‘yes’, ‘good’, and ‘ckd’ were converted to 1. The missing values were replaced using the mean heuristic.
(9)
Breast Cancer Coimbra Data: The data are characterized by nine features which describe quantitative data, such as age, Body Mass Index (BMI), and glucose level, and one ‘labels’ column. In the initial version of the data, the healthy controls were labeled with 1, while the patients were labeled with 2. In the processed data, the healthy controls and the patients were labeled with 0 and 1, respectively.
Table 2 summarizes the properties of the datasets after the preprocessing. For each dataset, the table presents the number of labels, the number of features, the number of samples of the training data, and the number of samples of the testing data.

4.2. BHOA Experimental Results

Table 3 presents the configuration parameters of the BHOA used in experiments.
The value of D is different for each dataset, and is equal to the number of features from Table 2. The objective function O F is represented by Formula (7).
Table 4 presents the BHOA experimental results. The columns are: (1) Dataset—the preprocessed data used in experiments, (2) Classifier—one of the classifiers from the possible set of classifiers {RF, SVM, GBT, LR, K-NN, NB}, (3) Accuracy—the accuracy for testing data, (4) Precision—the precision for testing data, (5) F1 score—the F1 score for testing data, (6) EvalGBest—the evaluated value of the global best individual after the running of all iterations, (7) (nf, RMSE)—the number of selected features and the RMSE computed using the testing data and the applied classifier, and (8) Time (ms)—the running time of the algorithm for I iterations.
As can be seen in the table, the classifier that returned the best accuracy results was SVM, as it returned the best accuracy results for six out of the nine datasets used in the experiments. Another observation is that all classifiers returned the best accuracy result for the Chronic Kidney Disease Data.
The classifier which returned the best E v a l G B e s t value in most cases was SVM as it returned the best value for five datasets.
Table 5 presents the summary of the BHOA G B e s t values. The selected features are marked with 1, while the features which are not selected are marked with 0.
As can be seen in the table, the features selected by the BHOA differ significantly for each classifier, even in cases when the same values for accuracy, precision, and F1 score were obtained, as in the case of Chronic Kidney Disease Data, which can be seen in Table 4.

5. Discussion

This section compares the BHOA to binary versions of the GWO, PSO, and CSA as follows:
(1)
Binary Grey Wolf Optimizer (BGWO);
(2)
Binary Particle Swarm Optimization (BPSO);
(3)
Binary Crow Search Algorithm (BCSA).
Table 6, Table 7 and Table 8 present the configuration parameters for the BGWO, BPSO, and BCSA which were used in experiments.
Table 9, Table 10 and Table 11 present the BGWO, BPSO, and BCSA experimental results, respectively.
In the case of the BGWO, the classifier which led to the best accuracy results for most of the datasets was SVM, as it returned the best accuracy for five out of the nine datasets. The classifiers which returned the best E v a l G B e s t were SVM and RF, because each one returned the best value for three out of the nine datasets.
The experimental results for the BPSO show that the algorithms which returned the best accuracy for most of the datasets (five out of nine), were RF and SVM. On the other hand, RF returned the best E v a l G B e s t value for four out of the nine datasets.
Finally, in the case of the BCSA, the SVM classifier returned the best accuracy for six out of the nine datasets, and the best E v a l G B e s t value for five out of the nine datasets as well.
Next, Table 12, Table 13 and Table 14 summarize the G B e s t values returned by the BGWO, BPSO, and BCSA, respectively.
The G B e s t values are mostly distinct for all datasets for all bio-inspired approaches, except for the Cervical Cancer Data set. One justification for this exception might be the nature of the Cervical Cancer Data set, which has the highest number of labels, a value equal to 13. As can be seen in the table, in the case of the Cervical Cancer Data set, multiple algorithms, namely SVM, GBT, LR, and NB, returned the same features for all bio-inspired algorithms.
Table 15 compares the accuracy obtained by the BHOA for each classifier with the values obtained by the BGWO, BPSO, and BCSA, respectively.
As can be seen in the table, the BHOA results are comparable to the results obtained by the other bio-inspired algorithms. In particular, the BHOA, BGWO, BPSO, and BCSA returned the best accuracy results in 33, 32, 33, and 33 experiments, respectively. These values were calculated as the number of bolded values of each column.
Several shortcomings of the BHOA compared to the BGWO, BPSO, and BCSA are: (1) the running time of the BHOA is higher than the running time of the BGWO, BPSO, and BCSA, (2) the number of configurable parameters of the BHOA is the highest of the group, a value equal to seven, compared to the numbers of configurable parameters of the BGWO, BPSO, and BCSA which are equal to two, five, and four, respectively, (3) determination of the optimal values for the configurable parameters of the BHOA is complex because of the large number of parameters that must be set, and (4) the BHOA is less performant than the BGWO, BPSO, and BCSA in the case of datasets that have many dimensions, as in the case of the Parkinson’s Disease Data set, where the BHOA returned the worst accuracy results.
Finally, Table 16 presents a statistical analysis of the results. For each dataset and for each algorithm, the table displays the mean, the standard deviation, the maximum value, and the minimum value with respect to the accuracy values obtained after the application of RF, SVM, GBT, LR, K-NN, and NB, respectively.
The results show that the BHOA, BGWO, BPSO, and BCSA returned the best mean accuracy values for four, four, two, and two datasets, respectively. Regarding the best maximum accuracy values, the BHOA, BGWO, BPSO, and BCSA returned the best values for seven, eight, seven, and five datasets, respectively.

6. Conclusions

This article presented the novel bio-inspired algorithm BHOA for feature selection for classification problems. The BHOA algorithm was applied with the classifiers RF, SVM, GBT, LR, K-NN, and NB on nine datasets from the UCI machine learning repository that are representative for classification problems. The obtained results were compared to the results obtained by the BGWO, BPSO, and BCSA. The experimental results show that the BHOA results are comparable to the results obtained by the BGWO, BPSO, and BCSA, as it returned the best accuracy results in 33 cases, compared to the 32, 33, and 33 cases for the other approaches. The statistical accuracy results analysis shows that the BHOA is a performant approach, as the BHOA and BGWO returned the best mean accuracy values for four datasets, and the BPSO and BCSA returned the best mean accuracy values for two datasets, respectively. On the other hand, the running time of the BHOA is overall longer than the running time of the other algorithms, and the BHOA requires the largest number of configurable parameters, a numerical value which is equal to seven, compared to the other approaches of the BGWO, BPSO, and BCSA, which require two, five, and four configurable parameters, respectively. The following directions are considered as future research work: (1) development of alternative versions of the BHOA using other functions for the conversion of the continuous values into binary values, (2) improvement of the performance of the BHOA using hybridizations with other bio-inspired algorithms, (3) comparison of the BHOA to more bio-inspired algorithms, and (4) adaption of the HOA to a multi-objective form that can be applied for feature selection in classification problems.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets which support this article are from the UCI machine learning repository, https://archive.ics.uci.edu/ml/index.php (accessed on 19 March 2022).

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Brodzicki, A.; Piekarski, M.; Jaworek-Korjakowska, J. The Whale Optimization Algorithm Approach for Deep Neural Networks. Sensors 2021, 21, 8003. [Google Scholar] [CrossRef]
  2. Rojas-Dominguez, A.; Padierna, L.C.; Valadez, J.M.C.; Puga-Soberanes, H.J.; Fraire, H.J. Optimal Hyper-Parameter Tuning of SVM Classifiers With Application to Medical Diagnosis. IEEE Access 2017, 6, 7164–7176. [Google Scholar] [CrossRef]
  3. Deb, S.; Gao, X.-Z.; Tammi, K.; Kalita, K.; Mahanta, P. Recent Studies on Chicken Swarm Optimization algorithm: A review (2014–2018). Artif. Intell. Rev. 2020, 53, 1737–1765. [Google Scholar] [CrossRef]
  4. Yang, X.-S. A New Metaheuristic Bat-Inspired Algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010). Studies in Computational Intelligence; González, J.R., Pelta, D.A., Cruz, C., Terrazas, G., Krasnogor, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; Volume 284, pp. 65–74. [Google Scholar]
  5. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  6. Ghaemi, M.; Feizi-Derakhshi, M.R. Forest Optimization Algorithm. Expert Syst. Appl. 2014, 41, 6676–6687. [Google Scholar] [CrossRef]
  7. Cheraghalipour, A.; Hajiaghaei-Keshteli, M.; Paydar, M.M. Tree Growth Algorithm (TGA): A novel approach for solving optimization problems. Eng. Appl. Artif. Intell. 2018, 72, 393–414. [Google Scholar] [CrossRef]
  8. Zhou, M.; Lin, F.; Hu, Q.; Tang, Z.; Jin, C. AI-Enabled Diagnosis of Spontaneous Rupture of Ovarian Endometriomas: A PSO Enhanced Random Forest Approach. IEEE Access 2020, 8, 132253–132264. [Google Scholar] [CrossRef]
  9. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  10. Wijaya, A.P.; Santoso, H.A. Improving the Accuracy of Naïve Bayes Algorithm for Hoax Classification Using Particle Swarm Optimization. In Proceedings of the 2018 International Seminar on Application for Technology of Information and Communication, Semarang, Indonesia, 21–22 September 2018; pp. 482–487. [Google Scholar]
  11. Agarwal, K.; Kumar, T. Email Spam Detection Using Integrated Approach of Naïve Bayes and Particle Swarm Optimization. In Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 14–15 June 2018; pp. 685–690. [Google Scholar]
  12. Lee, C.-Y.; Huang, K.-Y.; Shen, Y.-X.; Lee, Y.-C. Improved Weighted k-Nearest Neighbor Based on PSO for Wind Power System State Recognition. Energies 2020, 13, 5520. [Google Scholar] [CrossRef]
  13. Nagaraj, P.; Deepalakshmi, P.; Mansour, R.F.; Almazroa, A. Artificial Flora Algorithm-Based Feature Selection with Gradient Boosted Tree Model for Diabetes Classification. Diabetes Metab. Syndr. Obes. Targets Ther. 2021, 14, 2789–2806. [Google Scholar]
  14. Cheng, L.; Wu, X.-h.; Wang, Y. Artificial Flora (AF) Optimization Algorithm. Appl. Sci. 2018, 8, 329. [Google Scholar] [CrossRef] [Green Version]
  15. Ravi, V.; Madhav, V. Optimizing the Reliability of a Bank with Logistic Regression and Particle Swarm Optimization. In Data Management, Analytics and Innovation. Lecture Notes on Data Engineering and Communications Technologies; Sharma, N., Chakrabarti, A., Balas, V.E., Bruckstein, A.M., Eds.; Springer: Singapore, 2021; Volume 70, pp. 91–107. [Google Scholar]
  16. Li, M.-W.; Geng, J.; Wang, S.; Hong, W.-C. Hybrid Chaotic Quantum Bat Algorithm with SVR in Electric Load Forecasting. Energies 2017, 10, 2180. [Google Scholar] [CrossRef] [Green Version]
  17. Fang, M.; Lei, X.; Cheng, S.; Shi, Y.; Wu, F.-X. Feature Selection via Swarm Intelligence for Determining Protein Essentiality. Molecules 2018, 23, 1569. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Koohestani, A.; Abdar, M.; Khosravi, A.; Nahavandi, S.; Koohestani, M. Integration of Ensemble and Evolutionary Machine Learning Algorithms for Monitoring Diver Behavior Using Physiological Signals. IEEE Access 2019, 7, 98971–98992. [Google Scholar] [CrossRef]
  19. Cai, J.; Wei, H.; Yang, H.; Zhao, X. A Novel Clustering Algorithm Based on DPC and PSO. IEEE Access 2020, 8, 88200–88214. [Google Scholar] [CrossRef]
  20. Brezocnik, L.; Fister, I., Jr.; Podgorelec, V. Swarm Intelligence Algorithms for Feature Selection: A Review. Appl. Sci. 2018, 8, 1521. [Google Scholar] [CrossRef] [Green Version]
  21. Tran, M.Q.; Elsisi, M.; Liu, M.-K. Effective feature selection with fuzzy entropy and similarity classifier for chatter vibration diagnosis. Measurement 2021, 184, 109962. [Google Scholar] [CrossRef]
  22. Tran, M.-Q.; Li, Y.-C.; Lan, C.-Y.; Liu, M.-K. Wind Farm Fault Detection by Monitoring Wind Speed in the Wake Region. Energies 2020, 13, 6559. [Google Scholar] [CrossRef]
  23. Vo, T.-T.; Liu, M.-K.; Tran, M.-Q. Identification of Milling Stability by using Signal Analysis and Machine Learning Techniques. Int. J. iRobotics 2021, 4, 30–39. [Google Scholar]
  24. Moldovan, D. Horse Optimization Algorithm: A Novel Bio-Inspired Algorithm for Solving Global Optimization Problems. In CSOC 2020: Artificial Intelligence and Bioinspired Computational Methods; Silhavy, R., Ed.; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2020; Volume 1225, pp. 195–209. [Google Scholar]
  25. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 19 March 2022).
  26. Askarzadeh, A. A novel metaheuristic method for solving constrained engineering optimization problems: Crow search algorithm. Comput. Struct. 2016, 169, 1–12. [Google Scholar] [CrossRef]
  27. MiarNaeimi, F.; Azizyan, G.; Rashki, M. Horse herd optimization algorithm: A nature-inspired algorithm for high-dimensional optimization problems. Knowl.-Based Syst. 2021, 213, 106711. [Google Scholar] [CrossRef]
  28. Naruei, I.; Keynia, F. Wild horse optimizer: A new meta-heuristic algorithm for solving engineering optimization problems. Eng. Comput. 2021. [Google Scholar] [CrossRef]
  29. Meng, X.; Liu, Y.; Gao, X.; Zhang, H. A New Bio-inspired Algorithm: Chicken Swarm Optimization. In Advances in Swarm Intelligence. ICSI 2014. Lecture Notes in Computer Science; Tan, Y., Shi, Y., Coello, C.A.C., Eds.; Springer: Cham, Switzerland, 2014; Volume 8794, pp. 86–94. [Google Scholar]
  30. Chu, S.-C.; Tsai, P.-W.; Pan, J.-S. Cat Swarm Optimization. In PRICAI 2006: Trends in Artificial Intelligence; Yang, Q., Webb, G., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4099, pp. 854–858. [Google Scholar]
  31. Moldovan, D. Horse Optimization Algorithm Based Recurrent Neural Network Method for Epileptic Seizures Classification. In Proceedings of the 7th International Conference on Advancements of Medicine and Health Care through Technology. MEDITECH 2020. IFMBE Proceedings; Vlad, S., Roman, N.M., Eds.; Springer: Cham, Switzerland, 2022; Volume 88, pp. 183–190. [Google Scholar]
  32. Moldovan, D.; Anghel, I.; Cioara, T.; Salomie, I. Adapted Binary Particle Swarm Optimization for Efficient Features Selection in the Case of Imbalanced Sensor Data. Appl. Sci. 2020, 10, 1496. [Google Scholar] [CrossRef] [Green Version]
  33. Alirezazadeh, P.; Fathi, A.; Abdali-Mohammadi, F. A genetic algorithm-based feature selection for kinship verification. IEEE Signal Process. Lett. 2015, 22, 2459–2463. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Gong, D.-W.; Gao, X.-Z.; Tian, T.; Sun, X.-Y. Binary differential evolution with self-learning for multi-objective feature selection. Inform. Sci. 2020, 507, 67–85. [Google Scholar] [CrossRef]
  35. Leutheuser, M.; Schludhaus, D.; Eskofier, B.M. Hierarchical, multi-sensor based classification of daily life activities: Comparison with state-of-the-art algorithms using a benchmark dataset. PloS ONE 2013, 8, e75196. [Google Scholar] [CrossRef] [Green Version]
  36. Moldovan, D.; Slowik, A. Energy consumption prediction of appliances using machine learning and multi-objective binary grey wolf optimization for feature selection. Appl. Soft Comput. 2021, 111, 107745. [Google Scholar] [CrossRef]
  37. Patricio, M.; Pereira, J.; Crisostomo, J.; Matafome, P.; Gomes, M.; Seica, R.; Caramelo, F. Using Resistin, glucose, age and BMI to predict the presence of breast cancer. BMC Cancer 2018, 18, 1–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Fernandes, K.; Cardoso, J.S.; Fernandes, J. Transfer Learning with Partial Observability Applied to Cervical Cancer Screening. In Pattern Recognition and Image Analysis, IbPRIA 2017; Alexandre, L., Salvador Sánchez, J., Rodrigues, J., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2017; Volume 10255, pp. 243–250. [Google Scholar]
  39. Cinar, I.; Koklu, M.; Tasdemir, S. Classification of Raisin Grains Using Machine Vision and Artificial Intelligence Methods. GJES 2020, 6, 200–209. [Google Scholar]
  40. Er, M.B.; Aydilek, I.B. Music Emotion Recognition by Using Chroma Spectrogram and Deep Visual Features. Int. J. Comput. Intell. Syst. 2019, 12, 1622–1634. [Google Scholar]
  41. Faniqul Islam, M.M.; Ferdousi, R.; Rahman, S.; Bushra, H.Y. Likelihood Prediction of Diabetes at Early Stage Using Data Mining Techniques. In Computer Vision and Machine Intelligence in Medical Image Analysis; Gupta, M., Konar, D., Bhattacharyya, S., Biswas, S., Eds.; Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; Volume 992, pp. 113–125. [Google Scholar]
  42. Cinar, I.; Koklu, M. Classification of Rice Varieties Using Artificial Intelligence Methods. Int. J. Intell. Syst. Appl. Eng. 2019, 7, 188–194. [Google Scholar] [CrossRef] [Green Version]
  43. Sakar, C.O.; Serbes, G.; Gunduz, A.; Tunc, H.C.; Nizam, H.; Sakar, B.E.; Tutuncu, M.; Aydin, T.; Isenkul, M.E.; Apaydin, H. A comparative analysis of speech signal processing algorithms for Parkinson’s disease classification and the use of the tunable Q-factor wavelet transform. Appl. Soft Comput. 2019, 74, 255–263. [Google Scholar] [CrossRef]
Figure 1. HOA high-level view.
Figure 1. HOA high-level view.
Algorithms 15 00156 g001
Figure 2. High-level view of the BHOA for features selection.
Figure 2. High-level view of the BHOA for features selection.
Algorithms 15 00156 g002
Table 1. Classifier Configurations.
Table 1. Classifier Configurations.
ClassifierConfiguration
RF10 estimators, 42 random state
SVM0.00001 tolerance, 42 random state
GBT10 estimators, 1.0 or 0.001 learning rate
LRdefault parameters
K-NN9 neighbors
NBdefault or 100,000 var smoothing
Table 2. Experimental Datasets Summary.
Table 2. Experimental Datasets Summary.
Experimental DatasetLabels
Number
Features NumberTraining Data Samples NumberTesting Data Samples Number
Smart Grids Data21280002000
Raisin Data27720180
Turkish Music Emotion Data45032080
Diabetes Risk Prediction Data216416104
Rice Data263048762
Parkinson’s Disease Data2753605151
Cervical Cancer Data1332686172
Chronic Kidney Disease Data22432080
Breast Cancer Coimbra Data299323
Table 3. BHOA Configuration Parameters.
Table 3. BHOA Configuration Parameters.
Configuration ParameterValue
I (number of iterations)50
N (number of horses)30
M (horses reorganization frequency)5
D S P (dominant stallions percent)10%
H D R (horses distribution rate)10%
H M P (horse memory pool)3
S S P (single stallions percent)10%
Table 4. BHOA Experimental Results Summary.
Table 4. BHOA Experimental Results Summary.
DatasetClassifierAccuracyPrecisionF1 ScoreEvalGBest(nf, RMSE)Time (ms)
Smart Grids DataRF0.9120.9125480.9109720.409985(9, 0.296647)998,676
SVM0.6940.6869130.6663020.456546(2, 0.553172)361,459
GBT0.8990.8986270.8987360.425853(9, 0.317804)1,411,226
LR0.69350.6854250.6673920.456884(2, 0.553624)127,434
K-NN0.8970.8977190.8955220.407368(8, 0.320936)1,509,808
NB0.70050.6972740.6709530.452116(2, 0.547265)19,477
Raisin DataRF0.850.8583480.8496980.361902(2, 0.387298)109,878
SVM0.8722220.8808950.8719650.339523(2, 0.357460)17,181
GBT0.8666660.8700970.8666660.345289(2, 0.365148)89,852
LR0.8555550.8652770.8551800.356472(2, 0.380058)26,454
K-NN0.850.8609970.8495130.361902(2, 0.387298)38,053
NB0.850.8640390.8492890.361902(2, 0.387298)5984
Turkish Music
Emotion Data
RF0.90.9050830.9007210.545(34, 0.5)125,597
SVM0.9250.9238720.9228070.485473(39, 0.387298)169,451
GBT0.7750.8111600.7867730.589279(26, 0.612372)797,135
LR0.9250.9296730.9231350.530756(35, 0.474341)308,861
K-NN0.8250.8324990.8238490.685330(31, 0.707106)40,878
NB0.850.8579030.8513680.514260(26, 0.512347)11,822
Diabetes Risk
Prediction Data
RF0.9903840.9906190.9904060.198543(8, 0.098058)82,800
SVM0.9423070.9423070.9423070.305144(8, 0.240192)15,158
GBT1.01.01.00.1875(12, 0.0)45,600
LR0.9326920.9331720.9328420.319577(8, 0.259437)26,535
K-NN0.9326920.9351000.9331010.288327(6, 0.259437)29,988
NB0.9519230.9541330.9522150.305073(9, 0.219264)6495
Rice DataRF0.9133850.9133330.9133020.304060(2, 0.294302)240,551
SVM0.9238840.9242800.9236660.290251(2, 0.275890)36,531
GBT0.9146980.9147670.9145380.302381(2, 0.292064)286,505
LR0.9225720.9227700.9223970.292027(2, 0.278258)72,552
K-NN0.9225720.9228900.9223660.292027(2, 0.278258)145,352
NB0.9225720.9225530.9224840.292027(2, 0.278258)8526
Parkinson’s Disease DataRF0.8476820.8527200.8498140.333878(124, 0.390279)759,970
SVM0.8609270.8566190.8487570.300941(64, 0.372924)2,663,289
GBT0.8609270.8623740.8616050.335037(187, 0.363937)4,859,351
LR0.8609270.8596750.8461850.317210(113, 0.372924)845,271
K-NN0.8609270.8596750.8461850.318538(117, 0.372924)1,279,114
NB0.8211920.8230180.8220640.355323(115, 0.422856)519,532
Cervical Cancer DataRF0.8895340.9089490.8479671.574075(14, 1.952934)140,590
SVM0.8895340.9017370.8375311.732502(4, 2.268336)197,975
GBT0.8895340.9017370.8375311.732502(4, 2.268336)969,271
LR0.8895340.9017370.8375311.732502(4, 2.268336)214,834
K-NN0.8895340.9017370.8375311.732502(4, 2.268336)70,139
NB0.8895340.9017370.8375311.732502(4, 2.268336)15,678
Chronic Kidney
Disease Data
RF1.01.01.00.09375(9, 0.0)91,848
SVM1.01.01.00.083333(8, 0.0)13,385
GBT1.01.01.00.083333(8, 0.0)60,803
LR1.01.01.00.104166(10, 0.0)30,407
K-NN1.01.01.00.09375(9, 0.0)26,898
NB1.01.01.00.145833(14, 0.0)6673
Breast Cancer
Coimbra Data
RF0.9130430.9264210.9127140.304496(3, 0.294883)84,961
SVM0.8260860.8260860.8260860.423882(4, 0.417028)7934
GBT0.9130430.9130430.9130430.360051(5, 0.294883)38,080
LR0.7391300.8260860.7163040.438620(2, 0.510753)17,734
K-NN0.8260860.8695650.8191850.368327(2, 0.417028)11,024
NB0.7826080.8505430.7741330.460800(4, 0.466252)5248
Table 5. BHOA GBest Values Results Summary.
Table 5. BHOA GBest Values Results Summary.
DatasetClassifierGBest Value
Smart Grids DataRF[111110001111]
SVM[100000000010]
GBT[111100101111]
LR[100000000010]
K-NN[111100001111]
NB[100000000010]
Raisin DataRF[1001000]
SVM[10001]
GBT[1000001]
LR[110]
K-NN[10001]
NB[100001]
Turkish Music
Emotion Data
RF[01011111111101111111110110101110110110101000101010]
SVM[11111111101110101111011111111011101110111001100111]
GBT[01111110111010011010101101010101010101100010010000]
LR[11110010110111010111111011111101000110111011110110]
K-NN[11111101011111000001111011011101111101011000010010]
NB[01111100111101110001110000010011111110000000010011]
Diabetes Risk
Prediction Data
RF[1101100000111100]
SVM[0101101110110000]
GBT[1101011101110111]
LR[1101101110100000]
K-NN[1101101000010000]
NB[1101101010110010]
Rice DataRF[100001]
SVM[001010]
GBT[110000]
LR[110000]
K-NN[001010]
NB[110000]
Parkinson’s Disease DataRF[000011000000000000001100000000001000000000000001000001000001000000101010000000000110000000100010000010000100001000011011000000000000000000010000000001000000000000000000100000001000000010010010011000000011010000100000000000000100100000000010000100001000000000100000000100000001011100000001000000001000001100000000000000000101000001000000000110000000111000100011100000000000010000100000000000001101000001110000000000000000000100001010000000000000000000000000100100100000100010000000000000000000000000100000110010100010000000010000000010000100000000000101000000000010000000000001000000100001000100000000010000010001000000000000001000101000000000011000000011001000010010000000010110000110001000000000001000000000000100100000000010010000000100000010110011000]
SVM[000010000000000000000000100000010000000000000100010000000000000000000100000000000000000000000000010000010000000000001000000000010000000000010000000000000000000000000000010001001000000000010000000000001000100000000000000000000000000000000011000000000000000000000001001000000011010000000000000000000000000000000000000000000000000100000000000000000000000000000001000000100000000001000000001000000000000000000100001000100000000000000010000001001000010010000001001000001000000000000000000011000000000000000001000000000001000000000000000001000000000000000000000000000000001000000000000000000010000000000000001000000010011000100000000000000000000000000000010000000000000000000000000010001000001000000000000000000011000000100100000000100000000001010000000000100]
GBT[100001010010110000110000010001010000000010101000100001101001010000000000011010100100000010000000010100000000100110001100101001100010100001010010000001010000010110011000101000000010000001000000000000110010010000000000000000000000001010001000000110100101001001100001000001010001010001001110001000000000000100000100001000000000010000100000000010100011000100010000100000000000101000010000000011001100000001110001101000010000000001000010010000100100100100110100000000011000000000001111001000000110000010101011000100110001000000110110000010100001000000000100000010111000000001001000010100000000000100001001111100000000011000100011000100001101001000000000000001000101000001000000000010000110001010000110000000000001100000000000000000000010000001000111010110001]
LR[000010000001000000000000100000110000000001000101010001000100000000101100000010000100000000000000010000011000000000001000000000010001000001010010000000100000000000000001010001001000000000010001000001001000100000000000000000000100000000000011000000100000000000000001101000000011110000000010100000000000001000010001000000000100000100000000010000000000001000001001000001100000000001000000001000100000000000000100001000100000100000000010000001001000010010010001001000001000000000000000000011001000000001000001100000010001000000000000000001000110000000001100000000100000001000000000001000000011000100000000001000000000011000100001000000000000000000001000010000001000000000000000001010001000001000000000000000000010000000100100000000100000000001010000000000100]
K-NN[000010000100100100000001000100001001000000000000000000011000000000000000011000000010000100000000000010001000001000000001000001010000101001101000000001001000001001000000100000110000100000000100110001000100000000000010000100100000001000000000000000100000000100001000001010100000010001000000000000010000110010000000010001000001000000000010000000000000000000100001000001000000000010000101000000010000000000100000100000001100000000000000001000100000001011000000000000101101000000010010000000000000000001000001000000100000101010000000000000000001000000000000001100000000010000000000000001000010010010000000000000000001100000001000001000001000110001000100000001000000010001000100000000000100001101000000000000110000001000000000000000100001000000000000000000001]
NB[000000001000100001010000000101000000100000010001000000000000001100100000000100000000010100000100000001000000010000100000010000000000000000001000000010000000000000000000000000000010001011000000110000000010100000110110000000000100000000000000000000000000000000000100100000010000000101000000000010000010100010000110100000000000010000000000011000010010010010100000010000101000001010000000001000010000000000101000100000100000100001100010001000001000000000000000100000000011000000010000010000000000000100000000000101000000100000100000000000000000000000110000000010000101010000000000000110000000000000000000000000010100100101000000000000100110000001110001000000001000000000000001000000000001000000101000000000000000000000101010100000000010100000001000001000000]
Cervical
Cancer Data
RF[11110101010001010101100100100000]
SVM[00010000000100010000000000000001]
GBT[00010000000100010000000000000001]
LR[00010000000100010000000000000001]
K-NN[00100000000000000001001000000100]
NB[00010000000100010000000000000001]
Chronic
Kidney
Disease Data
RF[000101110000101100101000]
SVM[111011010000001000001000]
GBT[101010000010110100000010]
LR[111011010100001001001000]
K-NN[000110001000011011010100]
NB[111101001011001101111000]
Breast Cancer
Coimbra Data
RF[101001000]
SVM[101010010]
GBT[101001011]
LR[001100000]
K-NN[100100000]
NB[111000010]
Table 6. BGWO Configuration Parameters.
Table 6. BGWO Configuration Parameters.
Configuration ParameterValue
I (number of iterations)50
N (number of wolves)30
Table 7. BPSO Configuration Parameters.
Table 7. BPSO Configuration Parameters.
Configuration ParameterValue
I (number of iterations)50
N (number of particles)30
w (initial inertia value)0.901
c 1 (cognitive component constant)2
c 2 (social component constant)2
Table 8. BCSA Configuration Parameters.
Table 8. BCSA Configuration Parameters.
Configuration ParameterValue
I (number of iterations)50
N (number of crows)30
f l (flight length)0.9
A P (awareness probability)0.5
Table 9. BGWO Experimental Results Summary.
Table 9. BGWO Experimental Results Summary.
DatasetClassifierAccuracyPrecisionF1 scoreEvalGBest(nf, RMSE)Time (ms)
Smart Grids DataRF0.9120.9125480.9109720.409985(9, 0.296647)25,1259
SVM0.7480.7432910.7390870.459830(4, 0.501996)60,674
GBT0.8990.8986270.8987360.405020(8, 0.317804)332,609
LR0.74950.7447810.7410970.458708(4, 0.500499)21,082
K-NN0.8970.8977190.8955220.407368(8, 0.320936)326,848
NB0.73150.7285420.7148100.451127(3, 0.518169)4087
Raisin DataRF0.850.8583480.8496980.361902(2, 0.387298)25,023
SVM0.8611110.8696220.8608320.350937(2, 0.372677)3343
GBT0.8666660.8700970.8666660.345289(2, 0.365148)16,658
LR0.8555550.8589420.8555550.356472(2, 0.380058)6225
K-NN0.8722220.8785140.8720920.339523(2, 0.357460)8892
NB0.8444440.8599800.8435770.367232(2, 0.394405)1479
Turkish Music
Emotion Data
RF0.88750.8894230.8865500.515756(32, 0.474341)35,934
SVM0.93750.9403680.9371430.418107(28, 0.370809)31,204
GBT0.8250.8347030.8281570.579262(32, 0.559016)190,036
LR0.91250.9130350.9102880.437334(27, 0.403112)68,773
K-NN0.850.8617570.8484890.544260(32, 0.512347)11,434
NB0.86250.8618070.8592480.52(29, 0.5)3132
Diabetes Risk
Prediction Data
RF1.01.01.00.1875(12, 0.0)21,041
SVM0.9326920.9351000.9331010.288327(6, 0.259437)3359
GBT0.9903840.9906190.9904060.229793(10, 0.098058)11,020
LR0.9326920.9427160.9334530.272702(5, 0.259437)5473
K-NN0.9423070.9460370.9427500.273894(6, 0.240192)6570
NB0.8942300.8948740.8944670.273823(5, 0.260931)1477
Rice DataRF0.9107610.9107610.9107610.307379(2, 0.298728)47,690
SVM0.9238840.9242800.9236660.290251(2, 0.275890)7404
GBT0.9160100.9161270.9158370.300690(2, 0.289809)51,724
LR0.9199470.9202490.9197340.295534(2, 0.282935)10,566
K-NN0.9160100.9161270.9158370.300690(2, 0.289809)33,522
NB0.9212590.9214000.9210970.293788(2, 0.280606)2140
Parkinson’s Disease DataRF0.9205290.9188770.9187950.355813(462, 0.269903)133,329
SVM0.9470190.9504370.9444840.301116(387, 0.230174)637,241
GBT0.9006620.8978800.8978890.358514(392, 0.304491)965,475
LR0.9072840.9060970.9028480.347226(358, 0.304491)183,896
K-NN0.9205290.9208770.9167270.347218(409, 0.281904)235,658
NB0.8344370.8299440.8317870.430336(378, 0.406451)22,674
Cervical Cancer DataRF0.8895340.9099690.8453341.563974(11, 1.970715)32,820
SVM0.8895340.9017370.8375311.724689(3, 2.268336)39,130
GBT0.8895340.9017370.8375311.724689(3, 2.268336)180,010
LR0.8895340.9017370.8375311.724689(3, 2.268336)32,845
K-NN0.8895340.9017370.8375311.724689(3, 2.268336)10,959
NB0.8895340.9017370.8375311.724689(3, 2.268336)3971
Chronic Kidney
Disease Data
RF1.01.01.00.0625(6, 0.0)20,445
SVM1.01.01.00.0625(6, 0.0)2410
GBT1.01.01.00.083333(8, 0.0)13,438
LR1.01.01.00.052083(5, 0.0)10,193
K-NN1.01.01.00.052083(5, 0.0)5905
NB1.01.01.00.09375(9, 0.0)1539
Breast Cancer
Coimbra Data
RF0.8260860.8350580.8240800.396104(3, 0.417028)20,000
SVM0.7826080.8010860.7775230.423882(3, 0.454065)1665
GBT0.8695650.8719060.8690680.409757(5, 0.361157)9323
LR0.8260860.8695650.8191850.396104(3, 0.417028)4061
K-NN0.8695650.8719060.8690680.409757(5, 0.361157)2635
NB0.7826080.7839460.7817800.433022(3, 0.466252)1276
Table 10. BPSO Experimental Results Summary.
Table 10. BPSO Experimental Results Summary.
DatasetClassifierAccuracyPrecisionF1 scoreEvalGBest(nf, RMSE)Time (ms)
Smart Grids DataRF0.9120.9125480.9109720.409985(9, 0.296647)248,483
SVM0.75750.7546170.7474010.452665(4, 0.492442)89,348
GBT0.8990.8986270.8987360.405020(8, 0.317804)329,124
LR0.7580.7549430.7482610.452284(4, 0.491934)31,514
K-NN0.8970.8977190.8955220.407368(8, 0.320936)316,462
NB0.77550.7785010.7631770.438694(4, 0.473814)4771
Raisin DataRF0.8222220.8370500.8212310.361902(3, 0.339679)27,490
SVM0.850.8560790.8498470.339523(3, 0.309841)4111
GBT0.8666660.8700970.8666660.345289(2, 0.365148)23,406
LR0.8555550.8627890.8553410.356472(2, 0.380058)6817
K-NN0.8722220.8785140.8720920.339523(2, 0.357460)9630
NB0.8555550.8606810.8554660.356472(2, 0.380058)1610
Turkish Music
Emotion Data
RF0.81250.8287810.8170750.559279(20, 0.612372)37,558
SVM0.8750.8810370.8768750.518303(25, 0.524404)35,201
GBT0.81250.8192720.8132360.609341(27, 0.632455)156,980
LR0.83750.8410590.8356010.504262(17, 0.559016)53,802
K-NN0.76250.7843970.7637810.668072(7, 0.844097)8584
NB0.86250.8626340.8589070.525732(36, 0.460977)2265
Diabetes Risk
Prediction Data
RF1.01.01.00.125(8, 0.0)20,847
SVM0.8846150.8890440.8855010.286012(2, 0.339683)3908
GBT0.9903840.9905320.9903610.214168(9, 0.098058)11,035
LR0.8846150.8890440.8855010.286012(2, 0.339683)6369
K-NN0.9615380.9626010.9617020.256462(7, 0.196116)7427
NB0.8846150.8890440.8855010.286012(2, 0.339683)1488
Rice DataRF0.9133850.9133850.9133850.304060(2, 0.294302)60,536
SVM0.9238840.9242800.9236660.290251(2, 0.275890)9229
GBT0.9160100.9161270.9158370.300690(2, 0.289809)68,857
LR0.9199470.9202490.9197340.295534(2, 0.282935)18,202
K-NN0.9225720.9228900.9223660.292027(2, 0.278258)36,409
NB0.9212590.9214000.9210970.293788(2, 0.280606)2196
Parkinson’s Disease DataRF0.9006620.8984390.8990720.295813(179, 0.315178)107,492
SVM0.9205290.9208770.9167270.270857(179, 0.281904)420,036
GBT0.9072840.9049980.9052600.263561(106, 0.304491)743,155
LR0.9006620.9020480.8936500.286184(150, 0.315178)130,233
K-NN0.8940390.8907710.8904080.298585(164, 0.325515)151,805
NB0.8476820.8462180.8469050.318605(78, 0.390279)17,699
Cervical Cancer DataRF0.8895340.8947360.8401161.662037(7, 2.143133)49,912
SVM0.8895340.9017370.8375311.716877(2, 2.268336)45,406
GBT0.8895340.9017370.8375311.716877(2, 2.268336)220,140
LR0.8895340.9017370.8375311.716877(2, 2.268336)40,927
K-NN0.8895340.9017370.8375311.716877(2, 2.268336)16,559
NB0.8895340.9017370.8375311.716877(2, 2.268336)3805
Chronic Kidney
Disease Data
RF1.01.01.00.052083(5, 0.0)20,926
SVM1.01.01.00.09375(9, 0.0)2718
GBT1.01.01.00.052083(5, 0.0)14,586
LR1.01.01.00.083333(8, 0.0)7743
K-NN1.01.01.00.083333(8, 0.0)6437
NB0.98750.9877350.9874450.167185(8, 0.111803)1539
Breast Cancer
Coimbra Data
RF0.9130430.9264210.9127140.332274(4, 0.294883)20,374
SVM0.6521730.6521730.6521730.423882(6, 0.342954)1862
GBT0.6086950.6119560.5995420.360051(6, 0.257846)9458
LR0.8260860.8695650.8191850.396104(3, 0.417028)7177
K-NN0.9130430.9130430.9130430.332274(4, 0.294883)2644
NB0.7826080.7839460.7817800.433022(3, 0.466252)1433
Table 11. BCSA Experimental Results Summary.
Table 11. BCSA Experimental Results Summary.
DatasetClassifierAccuracyPrecisionF1 scoreEvalGBest(nf, RMSE)Time (ms)
Smart Grids DataRF0.8910.8914290.8894780.414280(8, 0.330151)226,327
SVM0.7130.7136930.6853120.443459(2, 0.535723)72,052
GBT0.8990.8986270.8987360.405020(8, 0.317804)258,876
LR0.7130.7122690.6867810.443459(2, 0.535723)24,238
K-NN0.8970.8977190.8955220.407368(8, 0.320936)188,415
NB0.70850.7111540.6768420.446597(2, 0.539907)4146
Raisin DataRF0.8555550.8681580.8549830.356472(2, 0.380058)26,458
SVM0.8722220.8808950.8719650.339523(2, 0.357460)3743
GBT0.8666660.8700970.8666660.345289(2, 0.365148)19,151
LR0.8555550.8627890.8553410.356472(2, 0.380058)6513
K-NN0.8722220.8785140.8720920.339523(2, 0.357460)9179
NB0.8555550.8606810.8554660.356472(2, 0.380058)1441
Turkish Music
Emotion Data
RF0.86250.8721730.8624130.532142(26, 0.536190)30,483
SVM0.9250.9261990.9224360.454759(26, 0.433012)35,100
GBT0.80.8167920.8040530.568705(25, 0.591607)140,459
LR0.88750.8912180.8863300.530791(24, 0.547722)47,209
K-NN0.76250.7906900.7679700.630330(20, 0.707106)8163
NB0.81250.8217410.8149350.535791(25, 0.547722)1939
Diabetes Risk
Prediction Data
RF0.9903840.9905320.9903610.198543(8, 0.098058)20,801
SVM0.9038460.9168950.9050110.295065(4, 0.310086)3723
GBT0.9903840.9906190.9904060.198543(8, 0.098058)10,362
LR0.9038460.9168950.9050110.295065(4, 0.310086)5936
K-NN0.9807690.9816840.9808510.244631(9, 0.138675)6770
NB0.9423070.9435180.9425530.289519(7, 0.240192)1393
Rice DataRF0.9133850.9133330.9133020.304060(2, 0.294302)54,663
SVM0.9238840.9242800.9236660.290251(2, 0.275890)8067
GBT0.9120730.9124600.9118030.300690(2, 0.289809)57,991
LR0.9225720.9227700.9223970.292027(2, 0.278258)14,810
K-NN0.9225720.9228900.9223660.292027(2, 0.278258)34,673
NB0.9225720.9225530.9224840.292027(2, 0.278258)2132
Parkinson’s Disease DataRF0.9139070.9131340.9134680.337923(355, 0.293415)118,080
SVM0.9337740.9327560.9323290.319168(380, 0.257342)499,655
GBT0.9072840.9050610.9041070.339922(336, 0.304491)864,187
LR0.9072840.9082930.9014740.357186(388, 0.304491)134,454
K-NN0.8940390.8929250.8873980.369634(378, 0.325515)149,951
NB0.8211920.8269240.8236940.435004(355, 0.422856)13,405
Cervical Cancer DataRF0.8430230.8410240.8241691.662473(11, 2.102047)29,424
SVM0.8895340.9017370.8375311.755939(7, 2.268336)47,669
GBT0.8895340.9017370.8375311.755939(7, 2.268336)207,039
LR0.8895340.9017370.8375311.755939(7, 2.268336)38,510
K-NN0.8895340.9017370.8375311.755939(7, 2.268336)14,752
NB0.8895340.9017370.8375311.755939(7, 2.268336)3392
Chronic Kidney
Disease Data
RF1.01.01.00.072916(7, 0.0)20,432
SVM1.01.01.00.0625(6, 0.0)2326
GBT1.01.01.00.072916(7, 0.0)12,809
LR1.01.01.00.0625(6, 0.0)7381
K-NN1.01.01.00.0625(6, 0.0)5904
NB1.01.01.00.09375(9, 0.0)1415
Breast Cancer
Coimbra Data
RF0.9130430.9264210.9127140.304496(3, 0.294883)21,385
SVM0.8260860.8260860.8260860.423882(4, 0.417028)1723
GBT0.9130430.9130430.9130430.304496(3, 0.294883)8984
LR0.8260860.8695650.8191850.396104(3, 0.417028)4358
K-NN0.9130430.9130430.9130430.332274(4, 0.294883)2646
NB0.7826080.7839460.7817800.433022(3, 0.466252)1240
Table 12. BGWO GBest Values Results Summary.
Table 12. BGWO GBest Values Results Summary.
DatasetClassifierGBest Value
Smart Grids DataRF[111110001111]
SVM[101100000100]
GBT[111100001111]
LR[101100000100]
K-NN[111100001111]
NB[100100000100]
Raisin DataRF[1001000]
SVM[0000110]
GBT[1000001]
LR[0011000]
K-NN[1000100]
NB[0100010]
Turkish Music
Emotion Data
RF[01111111111100110111101100000111010001111010011011]
SVM[01110001101110001011000101011001110101110010011111]
GBT[11100110101110110111011001010001100011111101111011]
LR[11011111001111010111001101010001000100011100010110]
K-NN[10010111110111101111011101100110101111010000111010]
NB[11011110010010111010011101010111100000011001110111]
Diabetes Risk
Prediction Data
RF[1001111111011011]
SVM[0101101000110000]
GBT[0001101011111110]
LR[0101101000001000]
K-NN[0101001100110000]
NB[0011100000100100]
Rice DataRF[011000]
SVM[001010]
GBT[011000]
LR[011000]
K-NN[010001]
NB[011000]
Parkinson’s Disease DataRF[111011110111000110110001010100101100100111101010010110111111011101100111011011110100111001101101001101101011111000101001001111110100100011010011000111010010111001111010100010100100100110010010010101100101000111111000110110110111111111110010111101001011111101011110100011010100001001100011111100011010110011101111011101111101001111110101110111111111111111011110001101101101011101001110110000010111010001000111110111001111011001010100011100101010011111111100010101010100010010101010100110111001010110111110100110010111111111010011111110111101111111111100011110100001011001111110110010111011011111100111111110100011001010011011111011111011101101101111101011111001101100100101010011011111111111100001010111110011101101100001101100101111011111001001101101011]
SVM[010011110011111001100010110100101001100100011101010000111101010101111101111111111100000100101100001011100011011010101110111010001101111111101110100010111001011110101111000101110000110010000000000101010011100111110110110110010110010101100111100011011101101100011101111000101010110101101011100110011100000111000000100110100101011001110010001011011101101101001011010100111101101100001001110001001111100111101110100011111100101011110100001001011100000111100000110011110111110010101010000000001111110101001011010011010000011100110100000000010001010110001001111101000000010011010001001010010010111001110010011001001001101100110000001000001011000100001111010101111011110010011011110111011100101110101011110011110000010000001111010010011100101111001010111100100]
GBT[010000110110110101001000000010100101101000001111101101010001100001011101111011111101001000110111000100110010111100111110011011100000110100100100101110000010010100010101110101010101101001101010101011111101101000011100001000111110101000101001011101110011001001110100011111110001110100111011010110010101110011000011111011011011111100010110001000111001101101010011011001110000111110110011100101000001001111100000101100001011000100101100101011111001011011111111001100001011111110011010111110010100001000011110011110111011111111100011101000011000000010101010001110110100011110011101100010101000111011000110011011001001111010101110100110100111100001001100101001101011101010111100010010111110001110000100010000111000110010000111010101111111100110001010000010110]
LR[000101010111110010000111111011110010111110001011000000010110011001011111110110100111010100001100001101001111100100101000011100101100110011001111000001011011010000010001000000010100101011100000010001110111101001001100010000011110111011100011010111100101100100110100111011010011011111111110000110111010111111110010111011000011111001110010100000001101100101001110100010001110101010111000100011011001101000001010100011000111001010101000011010101000011100100100100101001110110001101001001111000011110011011000010100111010111000110000101011010001011010000001000101000100001010001011111011101100111111010000110100010110000010110011110101000010100011011100100000000001101011001010110100100101110011100100101000000101010001001000000111101000111000100100010100111]
K-NN[101000100101010010111001011111110101111011001101011010111111000001011010111100100001011110101110101100011110101110001011011101010110110111011001110011010110011010010100110001010110010111010000111100111110101001111001010010011010011011111011010101110010001101010100010101011011000011001010101011001001011110110010101001000001011001011100101000011111010010000010111001011001011001000000001010000100011011111010010000011010010011111000101011010111111100001101111010010111111111010101001001010100111011011111110110011111100011010111111011110011001101111111111100011010011100100011110111011110101011000011011001110010110000001000101110111011011100011100001010011111101010011110110011000101110011101100000101110000000001111101001101100000110010001110010111111]
NB[101110100011001010000000010001110000100100000111010111010011001011101100001000101001001110000111001111110110110100110010111000100010101110111100010100011001011000110100101010010101101101111110101110100110001001100100111010101011011001110101011111110010010001100100110100000000011110001000111101011001001111010001101000111001011101110001100100000111111011010100101011001111010100010001010101111000011001101000001011010010100000101110011111110111010101100101101111100111110011100111100000101001011000100110110111001001111010110100111100011111011111001111011100011100101101010001010111011101001001100101010010110100100001111000101000111110001000101100101010110010001001000110110100100101111100010000001001101101010010011011010101010000010100000110101101101]
Cervical
Cancer Data
RF[00110010010000100010001110010010]
SVM[00100000000010000001000000000000]
GBT[00100000000010000001000000000000]
LR[00100000000010000001000000000000]
K-NN[00100000000010000001000000000000]
NB[00100000000010000001000000000000]
Chronic
Kidney
Disease Data
RF[001100000010000001010100]
SVM[001000000100101100010000]
GBT[110100000110100001000001]
LR[001000000000001110010000]
K-NN[001000000100001100010000]
NB[000101011001000101110000]
Breast Cancer
Coimbra Data
RF[001001010]
SVM[101100000]
GBT[111010010]
LR[001010010]
K-NN[101110010]
NB[110000010]
Table 13. BPSO GBest Values Results Summary.
Table 13. BPSO GBest Values Results Summary.
DatasetClassifierGBest Value
Smart Grids DataRF[111110001111]
SVM[111000000001]
GBT[111100001111]
LR[111000000001]
K-NN[111100001111]
NB[111000000001]
Raisin DataRF[1101000]
SVM[1101000]
GBT[1000001]
LR[0100100]
K-NN[1000100]
NB[0011000]
Turkish Music
Emotion Data
RF[10100011101101001001000100000100010001010100110101]
SVM[11111011010100011010001100010001101100010011011010]
GBT[01001100111011011011000101111110001111000100100110]
LR[01110010000000001001001000110001100110011000110000]
K-NN[00001010000000000000100001010001000000000000010000]
NB[01100111011011010110011111111111111100111010110111]
Diabetes Risk
Prediction Data
RF[1101000000111110]
SVM[0101000000000000]
GBT[1101001111100010]
LR[0101000000000000]
K-NN[0001001100101101]
NB[0101000000000000]
Rice DataRF[100100]
SVM[001010]
GBT[011000]
LR[011000]
K-NN[001010]
NB[011000]
Parkinson’s Disease DataRF[000011100000001011000001001000000000000000010010001000101010000000000010001001001001000000000100000000110100000001100001000110001000001100100101001010000010010001110001000000000000100101010010111000101001000000000000010100000000010100000110100100000100100000100000001010101000000001000110000000100000001000100000001000100000001010001000001110010010111101101000001101010101010010000010000000000010010000000000000010000000010010110000100001001011000000001000000100000000101000110001001101010000000000000000000000001011111010000001000000000000010010000010100000100000001000000100001100000010010001000011000010000000110000000000100011100000100000100010011000010000100001000010000000100000010010100000000110000010000000001001010000010000000001100110100001010]
SVM[011010011000000011100000110100010000000010000100100000101111001000010100000111001001000010100100110100000001001000101100000011001000101000001000000010000001000101010000100011010001000010110110000000010001000000000001010000000011000000000101110000000000000100010000001010110000010000100000000000100101000000000000100000001100011010000101000000000001000000000000000000000100000000000010100000010010010000001100100000000001011011000000000000001011010000010000000010000000100000000010000000100000001010011000100100100000011101100010000101000111000100001001100000110001001100000100010000100010100010000010000110000000110000001000001001011000000000000000000000001010101000000000001000001000011000101010100011010000010100001000000000001000100001001000010000000]
GBT[000101101000000000000000000001001000000001000100110000001000000000000000000010001000000010001100100000001000000010000000000001100010001000000001000100000010000000101000010010000000000010000010000010010010001000000001000001000010001000000101010000000000000000001001000000011000000100000010100000000000001000000000010000010000000100000000001100000001000001100000000000000010000001000000000000000100000001000001010000000000010000100001000000000000100000000001110000110000000010000000000000001000000000000000000000000011000000010001100001000000000000000000000000001000100000000000010000110000110110000101000000001000000000000000100101000000000000000001000000000000000010000000000000001000000100010000000010000000000000000001000000000010000000000000010000000]
LR[100000000000000010000000010000110000000001100100100000000000000001000000001100000010000000000000000000010000001001100000011100100111000000000000000000000010100000000100000001001000100001000000110000001000100010010111001000001010000011000010000000000000100001000000000000110000000000100000110100010000000010000100100000001001010000000000100100000101000000101001000000000000010101010001010000000000000000000001000000000001000011001100000100000001011000010011000000000001001111000000000000010010000000111010000110000000000000000000000001001011000000000100100100100010001000000111010000000011000001000100010000000010000010100101000000000000000000010000010111000101000010000100010001000101000001001010001010000000000000000000010100101010010000010000001000000]
K-NN[100100000010000000000000000100000000100100000000000000001000000001100100001100000000000000000000001000100000100000001000010101100010000010000000000100000000100000000001000000001100110000110000010000000010101010011110101010000000000010000000000000000001110000001000000010011100000100100000100000110001000010001101000000101011001000110000100010000101010010000000000010100010010101000000010000110000100000000001000000001101000000000100000100001001011010110011000010000110100110000001011000010011000001111000000000000000000000000010000000000011001101000100000100010000100010010111000000000100100010100010000000000100000000100100000000010000001000100001000101000101000010000100110101000000000001010000000000000000000010000000000110100000000001010100101000000]
NB[000000000000000000000000000000000000000000000000110000000000110100000000000000001000000000010100000100001000000000000000000000000010000100000100011000000001000000000100010000000000100000000000000000000000100000000000000000000100000000010000100000001000000000000010000000100000000000000000100010010000000000000000000000100000001000100000000010000000000010000010000000000000000000001000000000000010000000001001000100000100000000100000000000010001010000000000000010100100000000000000000000000001000000000011000000000000000001000101100000000000110000100000000000000001000000000001010010000000000110000000000010000100010000000000010000000000000010000000000000010000000000000000010000000010000010000010001001000000000000000000000000000000000000000100100000000]
Cervical
Cancer Data
RF[00000000100010101010000011000000]
SVM[00000000000000000000000000000101]
GBT[00000000000000000000000000000101]
LR[00000000000000000000000000000101]
K-NN[00000000000000000000000000000101]
NB[00000000000000000000000000000101]
Chronic
Kidney
Disease Data
RF[000100010000000100010010]
SVM[100010000101101010100100]
GBT[100000000001100000100010]
LR[001111000100001110000000]
K-NN[000110010010011100000001]
NB[001111000100001110000000]
Breast Cancer
Coimbra Data
RF[111000010]
SVM[001111110]
GBT[001111110]
LR[001010010]
K-NN[101100001]
NB[110000010]
Table 14. BCSA GBest Values Results Summary.
Table 14. BCSA GBest Values Results Summary.
DatasetClassifierGBest Value
Smart Grids DataRF[111100001111]
SVM[001000000001]
GBT[111100001111]
LR[001000000001]
K-NN[111100001111]
NB[001000000001]
Raisin DataRF[0011000]
SVM[0010001]
GBT[1000001]
LR[0100100]
K-NN[1000100]
NB[0011000]
Turkish Music
Emotion Data
RF[00111011101100011011010110010000101101101010010101]
SVM[00111101110000011001001000100111011001111001100111]
GBT[01100110001001010001001100011100111110110100011101]
LR[00010101101110011011100001010010010110100001111100]
K-NN[10101011001001001011001101000001000000100001001111]
NB[11010101001110111000001110010010111010100010110001]
Diabetes Risk
Prediction Data
RF[1001111110010000]
SVM[0001101000100000]
GBT[0001001111100011]
LR[0001101000100000]
K-NN[0001001110111011]
NB[1101100010100100]
Rice DataRF[100001]
SVM[001010]
GBT[011000]
LR[110000]
K-NN[001010]
NB[110000]
Parkinson’s Disease DataRF[001100010110001010101011001001011100011010110110011000110110111110010010001111000010111001010000100100000000000001010001000000010111001111001010111001101000011010101000111010001111000110011001110011111100111110100001000101110111001010101110101110011111111000011001100100101001100100000101101100000101101110010010100100111011100001010011110101000100010011010001101011010001000000011101011011001110101111010011010010010010000100011110011010100111010100001001110101100010011110010101011010001100001101001001110011110111111100111110011000010010100010111001011110000000100010110011101000110000111000100111111110000001000111011011000010011110010101101010100110111110000000011101000110001000001010110010101001101000001110011100111100010010110100100100001010000]
SVM[110011001000010111111001101100110001001111010000101111010111110000101010001011111110000001110000101110100110111011010010100001010011110010110011010101100101101100011100011110110100111011010110011101100110000100101111000001111101011101000011011010000001100010001001111110001011000101110111001000010000100011100100000100010101001100000111101000100110011101111011011101001101010101010101000101010100011111111110110110010111100001000000000001110000110111011111111100100111110111110000001000101001100010110001000001111000110011010110101101010000001001001100110001010010011101000111010100111100001101010111001110101111000011100011111101011110101001100010011111101010000001000111100110101001111001010010000111100110010001100011000100010000111011111111111110010]
GBT[001101111011011000101000000001010000101001000001101001001001010110011101000111001011010010000110000110000100100101011101001011110001000101100011101010011010000011000011001010000100110101010110110000000101010110101101010010100100110000100001000101001001010011000000101000000001001001101000011010110001100111001101111100000010010100011101010111000010101010010000000011011000110000010110010001010111110000101111001000111010010101010000101000110110101011100011011100110001000011000110011100001010000001010010101010101110100000110010000110110101110110011111011001111001011010011111110000111011100010110100010000000010011110011101001101110100110001011011010011110110010110110100101111100011101100100100101000111001001100010101011000110000110001100011100001101]
LR[000010100101101001111111101010101110111110101111110110001011001000101000111101000100000111101001010110101100010010101011010000101111110001111000011100110110010011101111001101100100101011001010110101100111010000111011101101110001001110110101100001000010000111010110110010011110001100001101001110110000110000011001011111001011000010110111000000110000010011100111101011101110100000010011000001001100010100010110100111100000001001000101111111000011001011111001010000010101111010101011000010101001011101010011011101101101001110101101101101000111001001010111101100010011101101110110101100100110110111110110100011011110111101010111101011101111011000111100101101101110111100000011011000001000011111000000011011001111000011010011010010101010001111000010001111000]
K-NN[001010001000010011001111011010010001001111111111111010011001001110001101101110001010010101011110101010101101101001001110010110000100010011000011101111101001010111101000110010100010101111001010011111000001011110100110101001100011111000101001100000111101000010001111001000011100101100000000110110000101011111011101110001011111101001000110111111010111100010110001100010110011110100101010110001100000111111000010110100011001001001100011010111010000011101010101001011001000110111110001100100010101001000101110111011000010110111110010011010000000111001010101010011100100011100011001100010001011011110000101000110000010011111101000000110011010001111000100001111101100010111010101110100110010101100101011010000100011110101100111110001110010010011011111110110110]
NB[100000100000111010101100101110001111100000010101110011010011010100110010010000000101010001011101101000000101100011011110011011000011001110000101110000111111100000010100101001110100000000100010001111100000101111100111111101010011101100100110100001010100110011111010100011001011110011001111101000001001000000110010011010111110011110111011110000100001001010111000010100011011100011110101110001010001001100101010110110110100000001000110101101010010011000100110101011111101000000111011100010010011101110101000111010001111011001110001001001111101001010000010111111000001101001101011010110000011000000000110010100000001101001010011000100011001001101110100001101111000001101000111111101110000001000011110001010000100000111001010111011001101000110011011011111100]
Cervical
Cancer Data
RF[10100001001110101000100000001010]
SVM[00000100110100010000000010000100]
GBT[00000100110100010000000010000100]
LR[00000100110100010000000010000100]
K-NN[00000100110100010000000010000100]
NB[00000100110100010000000010000100]
Chronic
Kidney
Disease Data
RF[001100010100001100010000]
SVM[000010000100001010101000]
GBT[110000000111000100001000]
LR[001100000100001100010000]
K-NN[000000000000001110110001]
NB[001100001011000111010000]
Breast Cancer
Coimbra Data
RF[101001000]
SVM[101010010]
GBT[101000001]
LR[001010010]
K-NN[101100001]
NB[110000010]
Table 15. Comparison of BHOA Accuracy Results to BGWO, BPSO, and BCSA Accuracy Results.
Table 15. Comparison of BHOA Accuracy Results to BGWO, BPSO, and BCSA Accuracy Results.
DatasetClassifierBHOABGWOBPSOBCSA
Smart Grids DataRF0.9120.9120.9120.891
SVM0.6940.7480.75750.713
GBT0.8990.8990.8990.899
LR0.69350.74950.7580.713
K-NN0.8970.8970.8970.897
NB0.70050.73150.77550.7085
Raisin DataRF0.850.850.8222220.855555
SVM0.8722220.8611110.850.872222
GBT0.8666660.8666660.8666660.866666
LR0.8555550.8555550.8555550.855555
K-NN0.850.8722220.8722220.872222
NB0.850.8444440.8555550.855555
Turkish Music
Emotion Data
RF0.90.88750.81250.8625
SVM0.9250.93750.8750.925
GBT0.7750.8250.81250.8
LR0.9250.91250.83750.8875
K-NN0.8250.850.76250.7625
NB0.850.86250.86250.8125
Diabetes Risk
Prediction Data
RF0.9903841.01.00.990384
SVM0.9423070.9326920.8846150.903846
GBT1.00.9903840.9903840.990384
LR0.9326920.9326920.8846150.903846
K-NN0.9326920.9423070.9615380.980769
NB0.9519230.8942300.8846150.942307
Rice DataRF0.9133850.9107610.9133850.913385
SVM0.9238840.9238840.9238840.923884
GBT0.9146980.9160100.9160100.912073
LR0.9225720.9199470.9199470.922572
K-NN0.9225720.9160100.9225720.922572
NB0.9225720.9212590.9212590.922572
Parkinson’s Disease DataRF0.8476820.9205290.9006620.913907
SVM0.8609270.9470190.9205290.933774
GBT0.8609270.9006620.9072840.907284
LR0.8609270.9072840.9006620.907284
K-NN0.8609270.9205290.8940390.894039
NB0.8211920.8344370.8476820.821192
Cervical Cancer DataRF0.8895340.8895340.8895340.843023
SVM0.8895340.8895340.8895340.889534
GBT0.8895340.8895340.8895340.889534
LR0.8895340.8895340.8895340.889534
K-NN0.8895340.8895340.8895340.889534
NB0.8895340.8895340.8895340.889534
Chronic Kidney
Disease Data
RF1.01.01.01.0
SVM1.01.01.01.0
GBT1.01.01.01.0
LR1.01.01.01.0
K-NN1.01.01.01.0
NB1.01.00.98751.0
Breast Cancer
Coimbra Data
RF0.9130430.8260860.9130430.913043
SVM0.8260860.7826080.6521730.826086
GBT0.9130430.8695650.6086950.913043
LR0.7391300.8260860.8260860.826086
K-NN0.8260860.8695650.9130430.913043
NB0.7826080.7826080.7826080.782608
Table 16. Statistical Accuracy Results Analysis.
Table 16. Statistical Accuracy Results Analysis.
DatasetAccuracy ResultBHOABGWOBPSOBCSA
Smart Grids Datamean0.7993330.8228330.8331660.803583
standard deviation0.1133400.0878320.0765820.100919
maximum0.9120.9120.9120.899
minimum0.69350.73150.75750.7085
Raisin Datamean0.8574070.8583330.8537030.862962
standard deviation0.0097280.0103930.0174500.008364
maximum0.8722220.8722220.8722220.872222
minimum0.850.8444440.8222220.855555
Turkish Music
Emotion Data
mean0.8666660.8791660.8270830.841666
standard deviation0.0605530.0415830.0406330.060553
maximum0.9250.93750.8750.925
minimum0.7750.8250.76250.7625
Diabetes Risk
Prediction Data
mean0.9583330.9487170.9342940.951922
standard deviation0.0295840.0397230.0558740.041245
maximum1.01.01.00.990384
minimum0.9326920.894230.8846150.903846
Rice Datamean0.9199470.9179780.9195090.919509
standard deviation0.0046210.0046760.0040370.005293
maximum0.9238840.9238840.9238840.923884
minimum0.9133850.9107610.9133850.912073
Parkinson’s Disease Datamean0.8520970.9050760.8951430.896246
standard deviation0.0160400.0380810.0249260.038992
maximum0.8609270.9470190.9205290.933774
minimum0.8211920.8344370.8476820.821192
Cervical Cancer Datamean0.8895340.8895340.8895340.881782
standard deviation1.22 × 10−161.22 × 10−161.22 × 10−160.018988
maximum0.8895340.8895340.8895340.889534
minimum0.8895340.8895340.8895340.843023
Chronic Kidney
Disease Data
mean1.01.00.9979161.0
standard deviation0.00.00.0051030.0
maximum1.01.01.01.0
minimum1.01.00.98751.0
Breast Cancer
Coimbra Data
mean0.8333320.8260860.7826080.862318
standard deviation0.0696550.0388880.1289770.057789
maximum0.9130430.8695650.9130430.913043
minimum0.739130.7826080.6086950.782608
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Moldovan, D. Binary Horse Optimization Algorithm for Feature Selection. Algorithms 2022, 15, 156. https://doi.org/10.3390/a15050156

AMA Style

Moldovan D. Binary Horse Optimization Algorithm for Feature Selection. Algorithms. 2022; 15(5):156. https://doi.org/10.3390/a15050156

Chicago/Turabian Style

Moldovan, Dorin. 2022. "Binary Horse Optimization Algorithm for Feature Selection" Algorithms 15, no. 5: 156. https://doi.org/10.3390/a15050156

APA Style

Moldovan, D. (2022). Binary Horse Optimization Algorithm for Feature Selection. Algorithms, 15(5), 156. https://doi.org/10.3390/a15050156

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop