Next Article in Journal
Supporting Students’ Academic Performance Using Explainable Machine Learning with Automated Prescriptive Analytics
Previous Article in Journal
Supporting Meteorologists in Data Analysis through Knowledge-Based Recommendations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved African Vulture Optimization Algorithm for Feature Selection Problems and Its Application of Sentiment Analysis on Movie Reviews

by
Aitak Shaddeli
1,
Farhad Soleimanian Gharehchopogh
1,*,
Mohammad Masdari
1 and
Vahid Solouk
1,2
1
Department of Computer Engineering, Urmia Branch, Islamic Azad University, Urmia 5756151818, Iran
2
Faculty of Information Technology and Computer Engineering, Urmia University of Technology, Urmia 5756151818, Iran
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2022, 6(4), 104; https://doi.org/10.3390/bdcc6040104
Submission received: 22 July 2022 / Revised: 13 September 2022 / Accepted: 21 September 2022 / Published: 28 September 2022

Abstract

:
The African Vulture Optimization Algorithm (AVOA) is inspired by African vultures’ feeding and orienting behaviors. It comprises powerful operators while maintaining the balance of exploration and efficiency in solving optimization problems. To be used in discrete applications, this algorithm needs to be discretized. This paper introduces two versions based on the S-shaped and V-shaped transfer functions of AVOA and BAOVAH. Moreover, the increase in computational complexity is avoided. Disruption operator and Bitwise strategy have also been used to maximize this model’s performance. A multi-strategy version of the AVOA called BAVOA-v1 is presented. In the proposed approach, i.e., BAVOA-v1, different strategies such as IPRS, mutation neighborhood search strategy (MNSS) (balance between exploration and exploitation), multi-parent crossover (increasing exploitation), and Bitwise (increasing diversity and exploration) are used to provide solutions with greater variety and to assure the quality of solutions. The proposed methods are evaluated on 30 UCI datasets with different dimensions. The simulation results showed that the proposed BAOVAH algorithm performed better than other binary meta-heuristic algorithms. So that the proposed BAOVAH algorithm set is the most accurate in 67% of the data set, and 93% of the data set is the best value of the fitness functions. In terms of feature selection, it has shown high performance. Finally, the proposed method in a case study to determine the number of neurons and the activator function to improve deep learning results was used in the sentiment analysis of movie viewers. In this paper, the CNNEM model is designed. The results of experiments on three datasets of sentiment analysis—IMDB, Amazon, and Yelp—show that the BAOVAH algorithm increases the accuracy of the CNNEM network in the IMDB dataset by 6%, the Amazon dataset by 33%, and the Yelp dataset by 30%.

1. Introduction

Feature selection is usually used as a pre-processing step to find an optimal subset of features considered in a set of all features in the data analysis process. It can be said that the problem is to find the optimal subset of features for better prediction of class tags in a data set with machine learning algorithms and deep learning by removing irrelevant or redundant features. The primary purpose of feature selection methods is to find an optimal subset of features from the original data set [1,2]. It improves the performance of machine learning algorithms and saves memory and CPU time. A data set with n samples is provided that i = 1, 2, … n. Each sample consists of d different features that = 1, 2, … D, and A¡j refer to samples i and j. Moreover, each sample consists of a class or tag, defined by i = 1, 2, …, m, as shown in Figure 1.
Figure 1 shows that class samples may have similar values and features. Moreover, the samples of different classes do not have various features [3] (in terms of number, type of features). Assuming that the original set consists of n samples and d features, then D ⊊RN contains all the features of d. Moreover, feature selection methods aim to find the optimal feature subset from the large D set. Thus, the binary mode for feature selection can be shown as X= (yi1, yi2, yi3, … yd). Here, yij can contain 0 and 1; zero means the feature is not selected, and one means the feature is selected. Solution x is considered a binary solution for feature selection. Feature selection methods can be divided into two general categories [4]: filter-based and wrapper-based. Filter-based methods use mathematical and statistical methods to find essential and dependent output features [5]. However, wrapper-based methods use machine learning and deep learning classification algorithm to evaluate optimal features. Filter-based methods are independent of classification algorithms and are relatively fast. Wrapper-based methods achieve better results than filter-based methods due to classification algorithms in their evaluation model. The challenging discussion in wrapper-based methods is the selection of a subset of features by the optimization algorithm and evaluating this set on a classifier in each step.
A heuristic is an automatic method of selecting or generating a set of heuristics. Hyper-heuristic was divided into two general and main categories [6]. These two categories are selection hyper-heuristic and generation hyper-heuristic. These categories are named ‘heuristics to select heuristics’ and ‘heuristics to generate’ heuristics’ defined and expressed. Hyper-heuristics and selection and generation can be selected or generated based on the nature of heuristics and are divided into two other categories: constructive or perturbative hyper-heuristics [7]. A constructive hyper-heuristic incrementally creates a perfect solution from scratch, while a disruptive hyper-heuristic iteratively improves an existing solution by using its chaotic mechanisms. The heuristics selected or generated by hyper-heuristic are called low-level heuristics (LLHs). A hyper-heuristic model based on selection includes two general levels. The low level has problem representation, evaluation function(s), and specific LLH problems. In the case of the high level, the high level manages which of the LLHs is selected to generate the solution for the new solution. The LLH Generates a new solution to be used, then decides to accept this new solution or not.
A cross-domain assessment was proposed for multi-objective optimization issues in the real world [8]. This paper uses five multi-objective evolutionary algorithms as Low-Level meta-Heuristics (LLHs) to govern a set of election-based, reinforcement learning, and mathematical function hyper-heuristics. According to their findings, hyper-heuristics perform better across domains than a single meta-heuristic, making them an excellent choice for novel multi-objective optimization problems. The authors of [9] proposed a new algorithm for feature selection concerning metaheuristics and hyper-heuristic algorithms. They use hyper-heuristics to give an efficient technique for dealing with complicated optimization challenges for feature selection in industrial and scientific-based domains for text classification. It can be done by using the power of the internet. This paper introduced a brand-new Hyper-Heuristic Feature Selection technique for locating an effective suitable feature subset [10]. Two categories of low-level heuristics are distinguished: the exploiters, which effectively exploit the search space by enhancing the quality of the candidate solution at hand, and the explorers, which thoroughly explore the solution space by focusing on random disturbances.
Recently many meta-heuristic algorithms have been used for the feature selection problem based on the wrapper-based model. Moreover, show high performance compared to traditional feature selection methods. Optimal meta-heuristic algorithms can select the optimal subset or close to an optimal subset of features in good time [11]. Searching for all subsets requires creating possible subcategories to find the solution, which is almost impractical and time-consuming work on a large data set. Most meta-heuristic optimization algorithms are designed to solve continuous optimization problems [12]. In addition to that, to solve the feature selection problem, these algorithms need to be designed with binary operators. Besides this, due to the lack of attention to the features of the feature selection problem, the operators of updating the optimization and optimal meta-heuristic algorithms in high dimensions are of collision constraint. They cannot solve the problem of feature selection well. Therefore, designing an optimization algorithm with powerful operators and its performance in a large data set is a big challenge.
The AVOA is a new meta-heuristic algorithm inspired by African vultures’ feeding and orienting behaviors. It consists of powerful operators while maintaining the balance of exploration and efficiency in solving continuous optimization problems. Feature selection can be defined as identifying relevant features and removing irrelevant and repetitive features to observe a subset of features that describe the problem well with minimal loss of efficiency. The importance of feature selection includes understanding the data, gaining knowledge about the process and helping to visualize it and reduce overall data, limiting storage requirements, possibly helping reduce costs, and gaining speed. Therefore, it encouraged us to offer a binary version of the AVOA. In addition, determining the values of the hyper-parametric algorithms of deep learning is a big challenge for analyzing the emotions of the film viewer, which has not yet provided a suitable and accurate solution. Because the parameters of a deep neural network have always been essential, the function of the deep neural network depends mainly on them. Finding the optimal values of deep parameters requires much work, experience, and time, which can be considered an optimization problem. One of the effective methods is to use binary meta-heuristic algorithms in bits to determine the appropriate amount of deep learning Artificial Neural Network hyper-parameters in analyzing movie viewers’ emotions. In this paper, to solve these challenges and problems and further evaluate and determine the appropriate value of hyper parameters, we used the BAOVAH approach, presented in the paper’s final section with its results. The contribution of this article is as follows:
  • It introduces three new binary versions of the AVOA algorithm.
  • Introduced a new hybrid version of AVOA with the sine cosine algorithm.
  • Presenting a new model of combining meta-heuristic algorithms based on hyper-heuristics.
  • Using the Disruption operator mechanism to maximize the performance of the proposed model.
  • Using IPRS strategy to generate quality initial population based on ranking strategy.
  • Using mutation neighborhood search strategy (MNSS) to maintain the balance between exploration and exploitation in the AVOA algorithm.
  • Using multi-parent crossover strategy to increase exploitation and properly implement exploitation step in AVOA algorithm.
  • Using the Bitwise method to binaries the operators of the AVOA algorithm.
  • Evaluating proposed approaches on 30 UCI datasets, including small, medium, and large datasets.
  • Comparing the proposed approaches with filter-based methods.
  • Designing CNNEM deep network for emotion analysis.
  • Optimizing the parameters of the deep learning method in emotion analysis using the BAOVAH approach.
Section 2 examines the previous works of different types of binary versions of meta-heuristic algorithms. Section 3 describes the AVOA. Section 4 presents three new binary versions of the AVOA algorithm. Section 5 evaluates the proposed methods and compares them with other algorithms. Section 6 considers a case study of in-depth learning and sentiment analysis. The paper’s final section presents conclusions and future work in general.

2. Related Works

Researchers have developed many wrapper-based methods based on meta-heuristic algorithms to solve the feature selection problem [13]. We will discuss some of the essential algorithms in the following. The operators of meta-heuristic methods usually work continuously; the authors have attempted to use different strategies to binaries these algorithms to solve the feature selection problem. Transfer functions are often considered the most straightforward methods and usually do not impose any change on the structure of meta-heuristic algorithms.
Moreover, there is sometimes a lack of proper exploration and exploitation in this transfer function. Genetic operators such as mutation and crossover can be used to overcome this problem. However, we need different strategies to balance exploration and exploitation to use meta-heuristic methods while providing higher quality solutions and greater diversity. For this reason, some authors have used new operators to maintain a balance between exploration and exploitation. Some disadvantages are as follows:
  • The inefficiency of algorithms in high dimensional datasets;
  • Poor convergence of some algorithms due to weak operators;
  • Use a single strategy to maintain a balance between exploration and exploitation;
  • Getting trapped in local optima due to an imbalance between exploration and exploitation;
  • Evaluation of algorithms on a few datasets.
The above disadvantages indicate that using only a few strategies can lead to poor results. However, using several different techniques can ensure the maintenance of a balance between exploration and exploitation and higher quality solutions and diversity. Therefore, the proposed method applies different strategies such as IPRS, mutation neighborhood search strategy (MNSS) (balancing exploration and exploitation), multi-parent crossover (increasing exploitation), and Bitwise strategy (increasing diversity and inquiry) to provide solutions with greater variety and to ensure the quality of the solution. We also proposed the AVOA BAOVAS and BAOVAV for the two versions based on the S-shaped and V-shaped transfer functions, which do not change the structure and operators.
In [14], only the S-shape and V-shape transfer functions are simply used to binary the PSO algorithm. No data sets and feature selection were used to test and evaluate the proposed method, but benchmark mathematical functions were used. Besides this, in this paper, new binary operators are not designed due to the feature selection, and the basic PSO algorithm is used with the transfer function. In addition, no particular operator is designed to escape the local optimization. In [15], a binary version of the GWO algorithm is presented with a change in the structure of the GWO algorithm to solve the feature selection problem. However, the proposed BGWO method has been evaluated on a relatively small and medium-sized data set. Moreover, no special operator is designed to escape the local optimization. In [16], a binary version of the BALO algorithm is presented by changing the structure of the BALO algorithm to solve the feature selection problem. The proposed BALO method is evaluated on a relatively small and medium-sized data set.
In [17], only simple V-shape transfer functions are used to binary the CSA algorithm. Moreover, no changes have been made to the algorithm’s structure to increase exploration and productivity. The proposed BCSA method was evaluated on a small dimension and numbers data set. In addition, fewer criteria have been assessed for the proposed method and other algorithms. In [18], only the new transfer functions TV-S shape and TV-V shape are used to binary the BDA algorithm. The innovation of this paper is more related to new transfer functions, but no changes have been made to the operators of the BDA algorithm. The proposed BDA method is evaluated on a relatively small and medium-sized data set. In addition, no special operator is designed to escape local optimization.
In [19], the author tries to use a version based on transfer functions and genetic operators to binary the SSA algorithm. Although no new operator is provided for the feature selection problem, this algorithm has compared algorithms with more and better statistical criteria. In [20], only an S-shape transfer and chaotic functions are used to binary the CSA algorithm. The innovation of this article is more about using ten chaotic maps, and no changes have been made to the operators of the CSA algorithm. The proposed algorithm has not been evaluated in the data set with dimensions of more than 82 items. In [21], only the S-shape and V-shape transfer functions are used to binary the BBO algorithm. The innovation of this paper is more related to transfer functions, but no changes have been made to the operators of the BBO algorithm. The proposed BBO method is evaluated on a relatively small and medium-sized data set. Moreover, no special operator is designed to escape local optimization. In [22], they tried to use opposition and local search operators to improve the SA algorithm in feature selection. However, the proposed algorithm was evaluated on a smaller data set. It also achieved worse results in several datasets in terms of several features.
In [23], only one V-shape transfer function is used to binary the COA algorithm, and there is no innovation in the structure of the COA algorithm. In [24], transfer functions and two mutation steps are used to binary the GWO algorithm to solve the feature selection problem. Indeed, the evaluation of the proposed method is done only on the data set with low dimensions (less than 62). In [25], a new version of the HHO algorithm is presented based on the mutation and opposition operator to solve the feature selection problem. The proposed algorithm may have a more complex time than the base version of the HHO algorithm.
In [26], only the S-shape and V-shape transfer functions are used to binary the WOA algorithm. The innovation of this paper is more related to transfer functions, but no changes have been made to the operators of the WOA algorithm. Moreover, no changes have been made to the operators of the WOA algorithm. A new binary version of the GSKO algorithm for feature selection is proposed [22]. The evaluation of the proposed algorithm is not based on a real problem. In [27], two binary versions of the SOS algorithm with an S-shape transfer function and new binary operators are presented. This paper evaluates both proposed methods with a smaller data set, and a special operator is designed to escape the local optimization.
In [28], four S-shape and four V-shape transfer functions are used to binary the EPO algorithm. The innovation of this paper is more related to transfer functions, but no changes have been made to the operators of the BBO algorithm. In addition, no special operator is designed to escape local optimization. In [29], two binary versions of the FFA algorithm with an S-shape transfer function and new binary operators are presented. This paper evaluates both proposed methods with a smaller data set. Moreover, no special operator is designed to escape local optimization.
In this paper, due to the deficiency and disadvantages of other methods, we have tried to provide a powerful algorithm for solving feature selection. Therefore, this paper presents a practical study of the AVOA algorithm using transfer functions, four different strategies, and new binary operators.

3. Enhanced AVOA with Hyper-Heuristic (Approach-1)

In this section, the mechanisms used in the first proposed model, as well as how to integrate the above mechanisms, are discussed.

3.1. AVOA

The AVOA is a new meta-heuristic algorithm inspired by African vultures’ feeding and orientation behaviors [30]. According to the conduct of vultures in nature, vultures are divided into two categories based on physical strength. Moreover, the desire to eat of vultures and hours of searching for food by vultures causes escaping from the hungry trap. Moreover, two of the best solutions are considered the strongest and best vultures. In this section, the general steps of this algorithm are summarized, which can be referred to [30] for more details.

3.1.1. Determining the Best Vulture in Each Group

In the first step of the AVOA algorithm, the first and second vultures are selected for global optimization based on the fitness function. Then other solutions according to the following formula show the possibility of choosing the vultures.
R ( i ) = { B e s t   V u l t u r e 1     i f   p i = α B e s t   V u l t u r e 2     i f   p i = β
In Equation (1), α and β are quantified parameters before the search operation. The value of these parameters is between zero and one, and the sum of the importance of both parameters is equal to 1. Furthermore, choose any of the best solutions by using the roulette wheel. If α is close to one, it increases exploitation. Vulture vigor rate due to vulture behavior, lack of energy, and aggressive behavior while searching for food is formulated with a mathematical form in Equation (2).
F = ( 2 × r a n d 1 + 1 ) × y × ( 1 I t i   m a x I t   )
In Equation (2), f indicates the degree of satiety of the vulture, It i Indicates the current iteration number, maxIt represents the total number of iterations, and y is a random number between −1 and 1 that changes each time it is repeated. rand 1 is a random number between 0 and 1; when the value of y is less than 0, the vulture is hungry; otherwise, vultures are total. In the AVOA algorithm, the value of the variable F strikes a balance between exploration and exploitation. When the value of |F| is more than 1, it enters the exploration phase; otherwise, it enters the exploitation phase [30].

3.1.2. Exploration Phase

Population exploration and diversity are formulated at this phase of the AVOA algorithm. For this phase, two strategies can be adjusted according to a parameter called p1 with a value between 0 and 1. To choose any of the strategies in the exploration phase p1, which is a random number between 0 and 1, if this number is more significant than or equal to parameter 1, Equation (3) is used, but if randp1 is smaller than parameter p1, Equation (4) is used.
V ( i + 1 ) = R ( i ) D ( i ) × F D ( i ) = | X × R ( i ) V ( i ) |
According to Equation (3), V ( i   + 1 ) indicates the vector position of the vulture in the next iteration, f indicates the degree of the vulture satiety, R ( i ) indicates one of the best-chosen vultures and x, which means the random movement of the lead vulture to protect food from other vultures. X is used as a coefficient vector that increases a random motion that changes with each iteration and is obtained using the X = 2 × Rand formula that Rand is a random number between zero and one. Vi is the current vector position of the vulture [30].
V ( i + 1 ) = R ( i ) F + r a n d 2 × ( ( u b l b ) × r a n d 3 + l b )
In Equation (4), rand 3 is a random number between zero and one. Moreover, ub and lb represent the upper and lower bounds. rand 3 has been used to increase the coefficient of random quiddity.

3.1.3. Exploitation Phase

At this phase of the AVOA algorithm, efficiency and convergence are formulated. For this phase, two strategies with values between 0 and 1, considering two parameters, p2 and p3, are adjusted [30]. In the following, each strategy for the exploitation phase is listed.
Phase 1
In the first phase, two different strategies of rotation flight and siege and quiet quarrel are performed. To determine the selection of each strategy, the p2 parameter is formulated as Equation (5) [30]. That randp2 is a random number between zero and one.
V ( i + 1 ) = {   E q u a t i o n   ( 6 )   i f         P 3     r a n d P 2   E q u a t i o n   ( 8 )   i f       P 3 < r a n d P 2
Siege and a quiet quarrel: At this phase of the AVOA algorithm, the severe quarrels of weak vultures on food acquisition and vital vultures are formulated, as shown in Equation (6).
V ( i + 1 ) = D ( i ) × ( F + r a n d 4 ) d ( t )
d ( t ) = R ( i ) V ( i )
In Equation (6), r a n d 4 is a random number between 0 and 1 increases the random coefficient. Ri is one of the best-selected vultures of the two groups. Vi is the current vector position of the vulture. Using this equation, the distance between the vulture and one of the best vultures of the two groups is obtained [30].
Rotational movement of vultures: In this step of the AVOA algorithm, rotational motion is modeled using spiral motion. The distance between the vulture and one of the two best vultures is first calculated in this method. Moreover, a spiral equation between the vulture and one of the best vultures is created. The motion is proportional to f.
V ( i + 1 ) = D ( i ) × e S F × c o s ( 2 π F ) + R ( i )
In Equation (8), S is a parameter to determine the logarithmic model of spiral shape.
Phase 2
In this phase of the AVOA algorithm, the movement of all vultures towards the food source is examined. To determine the selection rate of each of the strategies, the p2 parameter is formulated as Equation (9), which r a n d P 3 is a random number between zero and one.
V ( i + 1 ) = {   E q 11   i f       P 3     r a n d P 3 E q 12   i f       P 3 < r a n d P 3
The gartering of several types of vultures in the food source: In this phase, the movement of all vultures towards the food source is examined. There are times when vultures are hungry, and there is much competition for food, and several types of vultures may accumulate on the same food source, which is modeled according to Equations (11) and (12).
A 1 = B e s t   V u l t u r e 1 ( i ) D ( i ) × F A 2 = B e s t   V u l t u r e 2 ( i ) D ( i ) × F
In Equation (10), best vulture 1(i) is the best vulture of the first group in the current iteration, and best vulture 2(i) is the best vulture of the second group in the current iteration.
V ( i + 1 ) = A 1 + A 2 2
Finally, all vultures are aggregated using Equation (11).
Siege and fierce quarrels: At this phase of the AVOA algorithm, the lead vultures are hungry and weak; on the other hand, vultures have also become aggressive. Moreover, move in several directions towards the leading vulture. Equation (12) has been used to model this movement [30].
V ( i + 1 ) = R ( i ) | d ( t )   | × F × L e v y ( d )
In Equation (12), d ( t ) indicates the distance of the vulture to one of the best vultures of the two groups. The flight patterns of the levy flight function have been used to increase the effectiveness of the AVOA algorithm in Equation (12).

3.2. Sine Cosine Algorithm (SCA)

SCA is one of the powerful meta-heuristic algorithms presented by Mirjalili [31], which has been used to solve many problems. This algorithm uses two mathematical functions, Sine and Cosine, for optimization operations. In the SCA algorithm, Equation (13) is used to change the position of the search agent vector.
x i ( t + 1 ) = { x i ( t ) + r 1 × s i n ( r 2 ) × | r 3 P i ( t ) x i ( t ) | ,   r 4 < 0.5 x i ( t ) + r 1 × c o s ( r 2 ) × | r 3 P i ( t ) x i ( t ) | ,   r 4 0.5  
In Equation (13), x i ( t ) shows the current location of the ith particle and the tth iteration.  r 1 ,   r 2 and r 3 are random numbers and also r 3 are random numbers between 0 and 1.  P i is the destination of the particle. SCA meta-heuristic Algorithm 1 has four main parameters r 1 ,     r 2 , r 3 , and r 4 . The parameter r 1 represents the next position of the particle. This place may be inside the answer space or outside it. The parameter r 2 shows how much the particle’s movement moves towards the destination or against the direction of the destination. The parameter r 3 gives a certain weight to the destination position, and finally, the parameter r 4 indicates whether to use the sine component or the cosine component. Equation (14) is used to balance the extraction and exploration phases.
r 1 = a t a T
In Equation (14), t is the current iteration number, T represents the total number of iterations, and a is a constant value.
Algorithm 1: Pseudo-code of SCA algorithm.
Initialize a set of search agents (solutions)(X)
Do
Evaluate each of the search agents by the objective function
Update the best solution obtained so far (P = X * )
Update r 1 , r 2 , r 3 , and r 4
Update the position of search agents using Equation (14)
While (t< maximum number of iterations)
Return the best solution obtained so far as the global optimum

3.3. Modified Choice Function

In [32], the authors presented a hyper-heuristic based on the Choice function and scoring approach. The values of each LLH are calculated based on the previous performance of the LLH. The value of each LLH is calculated based on three separate criteria f 1 , f 2 and f 3 . The f 1 criterion, which is the recent performance of each LLH, is calculated using Equation (15):
f 1   ( h j ) = n α n 1 I n ( h j ) T n ( h j )
In Equation (15), h j is the same as LLH j . I n ( h j ) Shows the difference value between the current solution and the new solution with the n th application of h j , T n ( h j ) expresses the amount of time spent by the n th application of h j to propose a new solution, α is a parameter between zero and one that prioritizes recent performance. The second criterion, f 2 , indicates the dependence between a consecutive pair of LLHs, calculated using Equation (16).
f 2   ( h k . h j ) = n β n 1 I n ( h k . h j ) T n ( h k . h j )
In Equation (16), I n ( h k . h j ) , shows the amount of difference between the current solution and the new solution using the n t h consecutive application of h k and h j (that is, h j is executed right after h k ), T n ( h k . h j ) specifies the amount of time spent by the n t h consecutive application of h k and h j to propose a new solution, β is a parameter between zero and one that prioritizes the recent performance. These two criteria f 1 and f 2 are the components of the Choice function’s escalation, which increase the selection of LLHs with better performance. The third criterion, f 3 , the time elapsed since the last execution of a particular LLH is calculated using Equation (17).
f 3   ( h j ) = τ ( h j )
In Equation (17), τ ( h j ) represents the elapsed time since the last execution of h j (in seconds). Note that f 3 plays a role as a diversity component in the selection function and prioritizes those LLHs that have not been used for a long time. The score of each LLH is calculated using the sum of the weights of all three criteria f 1 , f 2 . and f 3 as shownin in Equation (18).
F ( h j ) = α f 1   ( h j ) + β f 2   ( h k . h j ) + δ f 3   ( h j )
In Equation (18), α, β and δ are the parameters that show the weights of f 1 , f 2 and f 3 criteria, which are fixed values in the initial model. In [33], the authors presented an improved version of the hyper-heuristic to increase efficiency and performance. In this version, the parameters can be dynamically controlled during execution. In this version, if an LLH improves the solution, the values of the α and β parameters increase compared to the improvement of the new solution with the previous solution. Still, if the selected LLH does not improve the solution, the values of the α and β parameters decrease due to the difference in the costs of the new solution and the previous solution. In this version of the selection function, parameters α and β are combined in a single parameter called μ, and finally, the score of each LLH is calculated using Equation (19).
F t ( h j ) = μ t   [ f 1   ( h j ) + f 2   ( h k . h j ) ] + δ f 3   ( h j )
If an LLH improves the solution, the resonance component is prioritized, and the μ parameter becomes the maximum static value close to one. At the same moment, the δ parameter decreases to the minimum static value close to zero. If LLH does not improve the solution, the μ parameter is penalized linearly, and the lower limit is 0.01. This mechanism causes the δ parameter to grow at a uniform and low rate so that the resonance components do not lose their effectiveness quickly. The parameters μ and δ are calculated using Equations (20) and (21), which show the difference between the cost of the previous solution and the cost of the new solution in Equation (20).
μ t ( h j ) = {                       0.99 .     d > 0 max [ 0.01 . μ t 1 ( h j ) 0.01 ] , d 0
δ t ( h j ) = 1 μ t ( h j )
In the proposed model, the Modified Choice Function has been used to change the optimization algorithm during optimization so that the performance and efficiency of the proposed model can be greatly increased by using this mechanism.

3.4. Disruption Operator (DO)

DO is used to increase population diversity, inspired by astrophysical phenomena. This operator has been used to improve the ability in the search process of the proposed model in the problem space and to create a balance between exploration and exploitation mechanisms. The perturbation operator is formulated using Equation (22).
D o p = { D i s , i , j × δ ( 2.2 ) .   i f D i s . i . b e s t 1   1 + D i s . i . b e s t × δ ( 10 4 2 . 10 4 2 ) .     o t h e r w i s e
In Equation (22), D i s , i , j represents the Euclidean distance between the ith and jth solutions, where the jth solution is located near and adjacent to the ith solution. The Euclidean distance is between the ith and the best solution. Moreover, δ ( x . y ) is a random number generated in [x,y] interval [34].

3.5. Bitwise Strategy (BS)

To escape from the local optimal points and solve the problem of low population diversity, two operators, bitwise OR operation and bitwise AND operation, have been used. In the first step, a random solution is created in these two operators. Then a bitwise AND operation is performed between the generated new, unexpected solution and the best solution. The main goal in the bitwise AND operator is to obtain good features in the best and the random new solutions.
Meanwhile, the solution obtained from AND operation and the new solution generated from the proposed model are used as input for the OR operation. Moreover, the primary purpose of the OR operation is to transfer features that are better and more useful than other features that come from the AND operation, which leads to the production of new and high-quality solutions. In addition, Bitwise operations can lead to escape from local optimal points because random shuffles can increase population diversity and improve solution quality. Figure 2 shows the Bitwise operation.

4. Hyper-Heuristic Binary African Vultures Optimization Algorithm (Approach-1)

The AVOA and its steps are described in Section 3. The standard AVOA is designed for continuous problems, and the solutions produced in this algorithm include decimal values. On the other hand, according to the nature of the problem, it is necessary to integrate certain mechanisms into the AVOA to achieve a powerful model with the ability to reach quality solutions. For this reason, we have used and benefited from several mechanisms in the proposed model, which are fully explained in Section 3. According to the feature of the AVOA, initially presented to solve continuous optimization problems, it focuses more on exploration mechanisms at the beginning of the optimization operation. Still, in the later stages, it tends more towards exploitation mechanisms. This approach is suitable for solving continuous problems, but the AVOA can get stuck in local optimal points in solving discrete problems.
We have integrated the AVOA with the sine and cosine algorithms. The reason for our choice of the sine–cosine algorithm is the unique features of this algorithm in increasing the diversity component. By using this algorithm, diversity can be optimized in all stages increased. However, to integrate the sine-cosine algorithm with the African vulture algorithm, the Modified Choice Function has been used. The main reason for choosing MCF is the intelligent choice of AVOA and SCA algorithms. You can switch between algorithms because each can perform differently in different optimization stages. In the next step, after updating the solutions using optimization algorithms, two other mechanisms are used, which are DO and BS, because these two mechanisms can significantly increase the solutions’ quality. To increase the computational complexity of the proposed model, after updating the solutions using optimization algorithms, only one of the DO and BS mechanisms is used to update the solutions. A random number between 0 and 1 is generated to select these mechanisms. If the generated random number is greater than 0.5, the DO mechanism is selected; otherwise, the BS mechanism is selected.
Due to their simplicity and lack of complexity, the transfer functions are used with the help of random thresholding to convert continuous space to binary space. In general, transfer functions are S-shaped or V-shaped, and each of these functions has different types. Figure 3 and Figure 4 illustrate the output of S-shaped and V-shaped transfer functions, respectively. This section presents two versions of the S-shaped and V-shaped proposed methods, BAOVAS and BAOVAV. In the BAOVAS version, we used four S-shaped transfer functions to binarize the AVOA, as described in the following.
S G 1 ( V i t + 1 ) = 1 1 + e 2 V i t + 1
S G 2 ( V i t + 1 ) = 1 1 + e V i t + 1
S G 3 ( V i t + 1 ) = 1 1 + e ( V i t + 1 2 )
S G 3 ( V i t + 1 ) = 1 1 + e ( V i t + 1 3 )
In the BAOVAS version, the solution V i t + 1 is transferred from the continuous state to the new space by each vs. transfer function between 0 and 1. In the BAOVAS version, we used four V-shaped transition functions to binary the AVOA, described in the following four functions.
V S 1 ( X i t + 1 ) = | erf ( π 2 V i t + 1 ) |
V S 2 ( X i t + 1 ) = | tan h ( V i t + 1 ) |
V S 3 ( X i t + 1 ) = | V i t + 1 1 + V i t + 1 2 |
V S 4 ( X i t + 1 ) = |   2 π a r c   tan ( π 2 V i t + 1 ) |
In the BAOVAS version, the solution V i t + 1 is transferred from the continuous state to the new space by each vs. transfer function between 0 and 1. After each solution in the AVOA becomes a constant space between zero and one, the following interface can convert binary vectors.
X i t + 1 = { 0         i f   r   < V S | S G ( V i t + 1 ) 1         i f   r   V S | S G ( V i t + 1 )
In Equation (31), r is a number between 0 and 1, which determines the final value of each of the dimensions of the solution V i t + 1 , which eventually produces a new binary solution. BAOVAH based S-shape and V-shape are shown in Algorithm 2. And flowchart of the this approach are shown in Figure 5.
Algorithm 2: BAOVAH based S-shape and V-shape.
01: setting parameter
02: Initialize the random binary population V i ( i = 1 .   2 .     N )
03: For it = 1: MaxIt, do
04:      Calculate the fitness according to Equation (23)
05:      get first and second Vulture Best_vulture1 and Best_vulture2
06:      for i = 1: N, do
07:     C a l c u l a t e   L L H   B a s e d   O n   M C F .
08:            if  ( L L H 1     L L H 2 ) ,  then
09:                Select the AVOA algorithm
10:          Else
11:          Select the SCA algorithm
12:      End if
13:      Update the location V S | S G ( V i t + 1   ) According to Equation (31)
14:        V i t + 1 = { 0         i f   r   V S | S G ( V i t + 1 ) 1         i f   r   V S | S G ( V i t + 1 )
15:       u p d a t e C h o i c e F u n c t i o n ( s e l e c t e d L L H )   / /   Equation   ( 19 )  
16:      if  ( r a n d   0.5 ) ,  then
17:        Apply D o p the formula for all individuals and their opposite makes more suitable positions for them
18:      Else
19:       Generate a random solution
20:       Apply AND operation (Selected Leader, random solution)
21:       Apply OR operation (AND Solution, V b i n a r y ( t + 1 ) )
22:       If fitness (OR solution) < fitness (Leader) then Update Leader = OR solution
23:      End if

5. Multi-Strategy Binary African Vultures Optimization Algorithm (Approach-2)

This paper presents a new approach to the AVOA, i.e., BAVOA-v1, using several robust strategies for solving feature selection and binarization problems. In the proposed BAVOA-v1approach, techniques such as IPRS strategy mutation neighborhood search strategy (MNSS) (balancing exploration and exploitation), multi-parent crossover strategy (increasing exploitation), and Bitwise strategy (increasing diversity and exploration) are applied to provide solutions with a greater variety and to ensure the quality of the solution. Considering these four different strategies and the exploration and exploitation steps of the AVOA, each of the operators is correctly implemented in the desired step; see Figure 6 to better understand the flowchart of the proposed BAVOA-v1 approach.

5.1. Initial Population Based on Ranking Strategy (IPRS)

The first step in any meta-heuristic algorithm is randomly generating the initial population. Nonetheless, considering the nature of the problem, there are different ways to improve the quality of the solution produced by AVOA from the beginning. Feature selection seeks two goals: selecting fewer features and acquiring higher accuracy. Suppose these two essential goals are intelligently incorporated into the AVOA from the start in the initial population generation step. In that case, they can play a critical role in the exploitation and convergence rate of the AVOA. In the proposed algorithm, we use the Information Gain Ranking (IGR) strategy to rank the features of each dataset. The IGR strategy helps the AVOA in the selection of more efficient features and thus the production of a higher quality initial solution. IGR is an entropy-based metric that measures the importance of a feature (the amount of information each feature carries). It is a good criterion for determining feature relevance for classification. As a statistical method, IG assigns the weight of features according to the correlation between features and classes. Assuming a dataset called D, the gain of feature F can be expressed by Equation (32).
G a i n f = E n t r o p y ( D a t a ) V | D a t a v | | D a t a |   E n t r o p y ( D a t a v )
In Equation (32), the number of elements D a t a and D a t a v are respectively shown as |   D a t a |, and |   D a t a v |; v is all unique values of the feature. After obtaining each feature’s IG, it can be represented for the k feature as GF= [Gainf1, Gainf2, …, Gainfk]. The production of an initial solution can be expressed as Equation (34).
S G = j = 1 k G F j
P r o p j = G F j S G
X ( i , j ) = { 1   i f   r a n d < P r o p j 0   i f   r a n d > P r o p j
In Equation (34), the GF represents the feature rankings by IG, and SG is the sum of the rankings of all the features. In addition, Propj calculates the probability of selection of a feature based on SG and GF. Finally, in Equation (35), we used a clever and relatively random technique to generate the initial solution. The higher the ranking of a feature by IG, the more likely it is to be selected by the variable Propj. Moreover, we used a rand to generate relatively random solutions based on higher-ranking features.

5.2. Mutation Neighborhood Search Strategy (MNSS)

The Mutation Neighborhood Search Strategy (MNSS) is an essential technique for balancing global and local search, first proposed by Das et al. [35] in 2009. The critical point in MNSS is that the mutation operator is used to search the neighborhood of the best solution (small area). In this paper, we use this strategy to improve the results of the BAVOAH algorithm. Note that, to control MNSS, we use the greedy rule: first, the nutation is carried out on the best current solution, then if the new position is better than the position of the current best solution, the current best solution is replaced by the MNS mutated solution; thereby, MNS-based local search is implemented. In the MNS strategy, to create neighbors for the first and last features, all the solutions are generated in the form of rings so that it is possible to make mutations to the right and left quickly. Figure 7 makes it easier to understand the MNS strategy in the form of a ring and a right neighborhood.
According to Figure 7, first, a solution is usually selected at two mutation points (Figure 7a); then, the solution is considered as a ring (Figure 7b); finally, the MNS strategy produces a new solution based on the right neighbor’s neighborhood from two points (Figure 7c). Note that the MNS strategy can also use the left neighborhood. So, the MNS strategy first generates a right neighborhood, then compares them to each other, then generates a left neighborhood on the final solution and compares the two solutions.

5.3. Multi-Parent Crossover Strategy (MPCS)

Genetic operators (selection, mutation, crossover) are inseparable in most discrete problems, including feature selection. Crossing over two parents, the crossover operator plays a vital role in exploitation. However, the point to be noted while using this operator is that if the parents are diverse, it will generate more diverse new solutions, and this will cause late convergence of the algorithm. On the other hand, if the parents are very similar in different iterations, it generates a repetitive and similar solution, thus causing early convergence. A multi-parent crossover operator was proposed by Al-Sharhan et al. [36] in 2019 to overcome this problem, where three parents can be defined as follows:
  σ 1 = X 1 + β × ( X 2 X 3 )
σ 2 = X 2 + β × ( X 3 X 1 )
σ 3 = X 3 + β × ( X 1 X 2 )
In Equation (36), β is a randomly selected position of the first parent used in crossover operations. In the AVOA algorithm, we have three main solutions in the exploitation phase (first solution: first-best solution; second solution: second-best solution; third solution: current solution); we can use the MPCS operator at this stage, which can be expressed as:
  σ 1 = V u l t u r e B e s t 1 + β × ( V u l t u r e B e s t 2 V c u r r e n t   )
σ 2 = V u l t u r e B e s t 2 + β × ( V c u r r e n t   V u l t u r e B e s t 1 )
σ 3 = V c u r r e n t   + β × ( V u l t u r e B e s t 1 V u l t u r e B e s t 2 )
In the exploitation step in the basic AVOA algorithm (important sub-steps of the accumulation of several vultures around the food source, the siege, and aggressive quarrel), the current solution is updated by the best solution in the current iteration VultureBest1 (i) and the second-best current solution VultureBest2 (ii). Therefore, in the binary version of BAVOA-v1, we designed a binary operator based on the same behavior, so that these three solutions can be combined to generate a binary value (Figure 8).
Since we use the two best solutions at this stage, premature convergence may occur in the algorithm. The current solution is updated by one of the best solutions and levying light on the siege and aggressive quarrel sub-steps to avoid it. Therefore, in the binary version of BAVOA-v1, we used the same behavior due to a direct levy flight and converted all of its values to binary with thresholding. The pseudo-code of this binary behavior is shown in Algorithm 3 (8:6).
Algorithm 3: Binary exploitation phase 2.
01: dim = length( V c u r r e n t   );
02: if abs(F)<0.5
03:   if rand<p2
04:      Multi parent Crossover strategy: V u l t u r e B e s t 1 ,   V u l t u r e B e s t 2 , V c u r r e n t  
05:   else
06:    update V c u r r e n t   by levy Flight and V random  
07:    convert binary V c u r r e n t   by threshold
08:   end if
The first exploitation stage in the basic AVOA algorithm consists of two sub-stages: siege, gentle quarrel, and rotational movement of vultures. We can implement the exploitation step using the crossover operator. Therefore, we used a variety of crossover operators to implement this step in binary. As a result, the first exploitation step in the binary version of BAVOA-v1 is defined as a single-point and a double-point crossover operator, as shown in Equation (42).
V ( i + 1 ) = {   S i n g l e P o i n t C r o s s o v e r   i f         P 3     r a n d P 2   D o u b l e P o i n t C r o s s o v e r   i f         P 3 < r a n d P 2  
In Equation (42), the Single Point Crossover and Double Point Crossover operators are applied to increase exploitation in the binary version of BAVOA-v1.

5.4. Bitwise Strategy (BS)

The meta-heuristic algorithm uses the Bitwise operator to increase population diversity and avoid getting trapped in local optima. On the other hand, in [37], it has been proven that the Bitwise operator performs better regarding the final solution accuracy, convergence speed, and robustness. The operators AND, OR, and NOT are three critical Bitwise operators that can play a crucial role in generating a new binary solution (example in Figure 9). Therefore, we also used the three operators for exploration in this step. The binary pseudo-code of the exploration step of the BAVOA-v1 algorithm is shown in Algorithm 3.
In Algorithm 4, the basic AVOA algorithm is applied so that Vrandom is one of the best solutions in the population, selected using Equation (1) in each iteration; Vcurrent represents the position of the current solution. Based on the exploration step of the basic AVOA algorithm, the new p value-based solution may be updated by the solution of Vrandom or take a new value according to the upper (ub) and the lower (lb) bounds. In the binary version of BAVOA-v1, considering these principles of the basic AVOA algorithm, this step is performed in two ways: (1) If p1 is larger than rand, then the two solutions Vrandom and Vcurrent are updated by two operators “AND” and “OR”, and thereby a binary vector is generated based on these two solutions; (2) If p1 is less than rand, then the solution Vcurrent is updated by the operator “NOT” with a probability of 50%, and thereby, a binary vector is generated based on these two solutions.
Algorithm 4: Binary Exploration base Bitwise Operators.
01: dim = length( V c u r r e n t   );
02: if rand < p1%--use “AND” and “OR” Bitwise—
03:     V e c or   =BitwiseOR( V c u r r e n t   ,   V random   )
04:     V e c and   = BitwiseAND ( V c u r r e n t   ,   V random   )
05:    for i = 1 to dim
06:       r = rand;
07:       if (r < 0.25)
08:           V c u r r e n t   (i) =   V e c and   (i);
09:       elseif(r < 0.5)
10:           V c u r r e n t   (i) =   V e c or   (i);
11:       else
12:           V c u r r e n t   (i) = V c u r r e n t   (i)
13:       end if
14:     end for
15: else if %-- use “NOT” Bitwise—
16:    for i=1 to dim
17:       if (rand > 0.5)
18:           V c u r r e n t   (i) = BitwiseNOT ( V c u r r e n t   (i))
19:       end if
20:     end for
21:   end if

5.5. Fitness Function

According to Equation (43), the fitness function in the feature selection problem can be a combination of reducing classifier error and reducing the number of selected features. The fitness function states that the number of selected features should be minimized so that the accuracy of the classification algorithm increases. The fitness function is described in detail in Equation (43).
F i t n e s s = α × ( 1 C l a s s i f i e r a c c ) + ( 1 α ) × f s S
In Equation (43), Classifieracc refers to the accuracy of a classification algorithm such as KNN. The α parameter represents the importance of classification accuracy, and (1 − α) shows the importance of the number of selected features. Fs refers to the number of selected features, and S is the total number of features in a dataset.

6. Results and Evaluation

In this section, the proposed algorithm and other meta-heuristic algorithms are evaluated. This section’s tests have been performed in a MATLAB software environment and on a system with five processors and 6 GB of RAM. The value of an in Equation (43) is equal to 99.0, and the classifier is also set in the KNN classification. Comparative algorithms include BBA [38], V-shaped BPSO [14], BGWO [39], BCCSA [20], and BFFA [29]. The proposed algorithm and other algorithms on 30 data sets are in Table 1.

6.1. Data Set

This phase will introduce the data set taken from the UCI source. This paper considered different data sets regarding the number of features, classes, and samples. In this paper, we divide the data set into three categories: small dimensions (less than 15) and medium (between 15 and 50 features), and large (with more than 50 features). A list of 24 different datasets with the total number of samples, number of features, and number of classes is given in Table 1.

6.2. Setting Parameters

This section sets the basic parameters of the proposed algorithms and five other well-known binary meta-heuristic algorithms. In all experiments, the population number of the proposed algorithm and different algorithms is set to 10 and the number of iterations to 100, unless, for a specific experiment, the number is stated separately in the same section. Table 2 shows the initial parameter value in comparative algorithms.
Table 3 shows that the three proposed approaches—BAOVAH-S, BAOVAH-V, and BAVOA-v1—have the same parameters. First, BAOVAS and BAOVAV versions are compared in eight different versions, and then one will be selected as the version based on transfer functions.

6.3. Evaluating the Three Proposed Methods of BAOVAH-S and BAOVAH-V

This section compares BAOVAH-S1, BAOVAH-S2, BAOVAH-S3, and BAOVAH-S4 with BAOVAH-V1, BAOVAH-V2, BAOVAH-V3, and BAOVAH-V4. Then, an approach is selected as the final approach based on the transfer functions. In these experiments, the iteration number was set to 100, and the population was assigned to 20. Moreover, all methods based on S-shaped and V-shaped are evaluated based on the criteria of an average number of selected features, classification accuracy, and fitness function. The first experiment evaluates the BAOVAH-S1, BAOVAH-S2, BAOVAH-S3, and BAOVAH-S4 approaches with BAOVAH-V1, BAOVAH-V2, BAOVAH-V3, and BAOVAH-V4 in terms of the number of features selected. The results of this experiment are shown in Table 3.
According to Table 3, V-shaped methods’ performance is much more successful and efficient in identifying valuable and essential features. With the BAOVAS4-based method, the results are more effective and efficient than other methods based on S-shaped. Moreover, BAOVAS2 and BAOVAS3 methods provided tolerable performance among the 30 datasets case study. The weakest result is in the S-shaped method related to BAOVAS1.
In the V-shaped-based method, the BAOVAV4 algorithm obtains the most successful results. BAOVA-V1, BAOVA-V2, and BAOVA-V3 algorithms also incur average results in the average number of selected features. The following results of S-shaped- and V-shaped-based algorithms based on the average accuracy evaluation criterion are reviewed. The results of this experiment are shown in Table 4. In the following, we will analyze and compare the methods based on S-shaped- and V-shaped-based methods in terms of average accuracy.
According to Table 4, it can be seen that the performance of V-shaped-based methods is much more efficient and better in obtaining the values of the fitness function. BAOVAS4-based methods have achieved more successful results than other S-Shaped-based methods. BAOVAS2 and S3-BAOVA methods also provided acceptable performance among the 30 datasets studied. The weakest effect in the average fitness function is also related to BAOVAS1 in the S-shaped method. The most successful results in the BAOVAV4 algorithm are in the V-shaped-based method. BAOVAV1 and BAOVAV2 algorithms have also achieved relatively weak results. As a result, the experiments prove the superiority of BAOVAV4 in terms of average selected features, accuracy, and fitness function. Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12 show the degree of convergence of each method based on S-shaped and V-shaped.
It can be seen from Table 5 that the performance of V-shaped-based methods is much more efficient and successful in obtaining the average accuracy. BAOVAS1 and BAOVAS4-based methods have achieved more successful results than other S-shaped-based methods. The most successful results in the BAOVAV4 algorithm are in the V-shaped-based method. BAOVAV1 and BAOVAV2 algorithms also achieved relatively weak effects. The test results prove the superiority of the BAOVA-4 in terms of the mean of the selected features, the mean accuracy, and the mean of the fitness function. Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 show the degree of convergence of the S-shaped- and V-shaped-based methods.

6.4. Evaluating the BAVOA-v1 Approach

In this subsection, we aim to evaluate the proposed approaches, BAOVAH (hyper-heuristic) and BAVOA-Version1 (multi-strategy). The BAOVAH method has been selected based on Section 5.3 as a transfer function-based method among eight transfer functions according to its results. This section will compare these two proposed methods with the basic algorithms BBA, BPSO, BGWO, and new algorithms such as BDA, BCCSA, and BCCSA regarding different statistical criteria. The fitness function specified in Equation (10) states that the proposed approach and other algorithms should consider feature selection and accuracy objectives. However, 99% of the scale is set on the accuracy of the classification. Therefore, in the first experiment, the average accuracy of the algorithms and the proposed approach on 30 data sets are presented in Table 6.
Comparing the mean accuracy of all methods in Table 6 shows that the BAOVAH method is more accurate than comparative algorithms out of 30 data sets in 20 sets, which is 67% of the data set has obtained the highest accuracy. Moreover, in the other 33%, it has shown perfect accuracy. Besides this, the BAVOA-V1 approach has shown relatively weak performance compared to the BAOVAH approach and different algorithms due to its easy use of the transfer function. After examining the algorithms’ average accuracy, the features’ average is considered in the following: in the fitness function, only 01% of the number of features are considered. Evaluation of the proposed approach and other algorithms in terms of the average number of selected features is given in Table 7.
A comparison of the average choice of method features in Table 7 shows that the BAOVAH method obtained a better average of feature selection than comparative algorithms from 30 data sets in 12 data sets. At 40%, the data set has the lowest accuracy. In the other 60%, it has shown an excellent selection. Besides this, obtaining the optimal value or the best fitness function can determine the superiority of algorithms. To further prove the proposed method, Table 8 compares the proposed BAOVAH approach based on the best value of the fitness function
Comparing the best value of the fitness function of all methods in Table 8 shows that the BAOVAH method has obtained a better fitness function value out of 30 data sets in 28 data sets than comparative algorithms. 93% of the data set, which has the best fitness function value, and the other 7%, showed perfect accuracy. Moreover, the BAVOA-V1 approach has a relatively weak performance compared to the BAOVAH approach and different algorithms due to the easy use of the transfer function. So, statistical results based on Table 8 show the approach. The proposed method of BAOVAH can be presented as a robust algorithm in feature selection. The degree of convergence of the algorithm shows how the algorithm has maintained a balance between exploration and exploitation. The convergence results of the proposed approach and other comparative algorithms to prove the convergence of the proposed method are shown in Figure 15, Figure 16, Figure 17, Figure 18 and Figure 19 on 30 data sets, respectively.
The convergence function of the proposed method and other algorithms in the above figures shows that the BAOVAH approach has a relatively good convergence. It has been able to maintain the balance between exploration and exploitation. Finally, in 93% of the data set, it has shown better convergence.

7. Case Study

This section aims to apply a practical study to evaluate further the BAOVAH approach in setting the parameters of deep learning algorithms in sentiment analysis. This section will combine the proposed BAOVAH approach with in-depth learning methods for sentiment analysis. In this section, BAOVAH’s proposed approach for setting hyper-parameters of the deep learning algorithm, such as the number of neurons in each layer, activation function, and deep learning, has been used in sentiment analysis. This section will use the UCI Sentiment Analysis Textual Database, consisting of three different and primary sentiment analysis datasets. This dataset includes three datasets IMDB, Amazon, and Yelp. Each data set contains 1000 samples with 0 labels and 1. the results of this section have been done in the Google collab environment and with Python language. The Tensor Flow and Keras libraries have also implemented in-depth methods. We used the proposed BAOVAH approach, a binary algorithm in setting the parameters of deep learning algorithms in sentiment analysis that has advantages such as exploring and creating a variety of solutions and creating quality solutions over continuous solutions. In [42], a partial swarm quantum binary algorithm called BQPSO is used to determine the appropriate parameters of the deep convolution network. The experimental results of this paper prove that the proposed binary-based method can have better performance and stability than the traditional method. This method is more commonly known as fixed-length string binary. In [43], a binary string with a genetic algorithm is used as a three-step evolutionary process consisting of selection, synthesis, and mutation to optimize the structure of the DenseNet topology. The results of various experiments confirm the superiority of the proposed method. Moreover, the number of parameters has been significantly reduced. In [44], the BPSO is being used to optimize the parameter values of a deep CNN network. The bit string is used to encode this problem that the results of the CNN hybrid method or BPSO have a significant advantage over other methods.

7.1. Pre-Processing

Since this dataset is textual, a series of initial pre-processing is required; in this phase of the proposed method, a series of pre-processing operations, including data clearing, tokenization removing the stop words, and stemming, is being applied to any comments. In natural language processing, algorithms do not understand the text. Therefore, the first and most crucial step is to identify or separate the words (signs and words) and the unit maker’s responsibility to separate their words. The next step is to remove the stop words. Stop words are repetitive words in the text that do not contain information and only connect words in a sentence [45]. Rooting is the last step that is done in the pre-processing phase. The root refers to the central meaning and concept of the word. Thus, a limited number of roots are formed in the natural language, and the rest of the words are derived from these roots [46,47]. The purpose of the root finder is to extract the root and delete the fixes attached to the word [48,49]. Hence, rooting is one of the main steps in natural language processing that must be done.

7.2. CNN Deep Neural Network Proposed Based on the Embedding Layer (CNNEM)

CNN is a particular type of multilayer perceptron that includes an input layer, an output layer, and a cannulation layer with several filters with different dimensions, followed by a pooling layer [50]. However, when this network is used for text and sentiment analysis, it is defined with the first layer of embedding. This embedding technique maps words and phrases from a dictionary to numeric vectors. We have designed a deep CNN model based on the embedding layer called CNNEM. The detailed specifications are given in Table 9. The CNNEM model shows some variable parameters in the Changeable column determined by the BAOVAH proposed approach.
To better understand the determination of CNNEM model parameters by the BAOVAH algorithm, we have shown it in Figure 18. In Figure 20, we mentioned the range of numbers, the number of bits required for each layer, and finally, the dimensions of the problem for the BAOVAH proposed approach. Finally, the CNNEM model requires 61 bits witch these bits will be generated with zero and one by the BAOVAH algorithm and will optimize the CNNEM model. The improved model by the BAOVAH approach is named CNNEMBH.

7.3. The Results of Improving the CNNEM Model with the BAOVAH Proposed Approach

In this subsection, the parameters specified in Figure 20, including the size of the embedding layer, the numbers of the convolutional layer filters 1 and 2, and the dense layer of neurons one to three. The name of the activating function for each layer is determined by the BAOVAH approach to increase the accuracy of the CNNEM model and its efficiency in sentiment analysis. For the initial evaluation in this section, first, the CNNEM model with the basic parameters, shown in Table 10, is being performed, and we display its different results. Then we offer the CNNEM model with the help of the parameters implemented by the BAOVAH approach and its various effects. In this part, the base model with the initial CNNEM parameters is named, and the improved model is called CNNEMBH by the BAOVAH approach. A comparison of the accuracy of the two in-depth models, CNNEM and CNNEMBH, for the three sentiment analysis datasets is shown in Table 10.
Comparing the accuracy of two in-depth models, CNNEM and CNNEMBH, for the three sets of sentiment analysis in Table 10 shows that the improved CNNEMBH method has been able to increase the accuracy to 79 in the IMDB dataset to 0.78 in the Amazon and Yelp datasets and has achieved better results compared to the basic model.

8. Conclusions and Future Work

This paper introduced and implemented two approaches to AVOAs. In the first proposed model, a hyper-heuristic method is used to increase the performance of the proposed model to combine and integrate the mechanisms of the sine and cosine algorithm with the African vulture algorithm. This combination’s purpose is to choose both algorithms’ mechanisms intelligently and avoid increasing the computational complexity. On the other hand, in the first proposed model, two mechanisms, the Disruption operator and the Bitwise strategy, have been used to maximize the capability and efficiency of the first proposed model.
Each used four S-shaped transfer functions and four V-shaped transfers to binarize the basic AVOA. In addition, this paper presented an improved version of the AVOA called BAVOA-v1, where four different strategies were applied to improve the performance of the AVOA in the feature selection problem. These four strategies included: IPRS, mutation neighborhood search strategy (MNSS) (balancing exploration and exploitation), multi-parent crossover strategy (increasing exploitation), and Bitwise strategy (increasing diversity and exploration). These strategies were used to provide solutions with more variety and to assure the quality of solutions. Each of these four strategies was designed and implemented in the exploration and exploitation step of the AVOA algorithm. Finally, the proposed BAOVAH approach was evaluated on 30 UCI datasets. The results of different simulations showed that the proposed BAOVAH algorithm performed better than the basic binary meta-heuristic algorithms, and it has shown better performance. Hence, the proposed BAVOA algorithm in 67% of the data set is the most accurate, and in 93% of the data set, it is the best fitness function value. In terms of feature selection, it has shown high performance. A practical study in which the BAOVAH approach was used to determine the appropriate values of hyper-parameters of the deep learning algorithm was performed in sentiment analysis.
The string fixed-length binary coding method was used, fitting deep learning algorithms. Moreover, a new deep convolutional network called CNNEM based on Embedded layers was designed. Determining the appropriate values of CNNEM hyper-parameters is done with the proposed BAOVAH approach. The results of various experiments on three basic sets of sentiment analysis called IMDB, Amazon, and Yelp show that the BAOVAH algorithm increases the accuracy of the CNNEM network in the IMDB dataset by 6%, in the Amazon dataset by 33%, and in the Yelp dataset by 30%. It has optimized the appropriate values of CNNEM hyper parameters well. Different primary population production strategies are considered new local search and binary operators for future work. The BAOVAH approach will increase the precision of profound algorithms in image processing and data with large dimensions. In addition, the BAOVAH approach can be used for various engineering and medical research.

Author Contributions

Conceptualization, A.S. and F.S.G.; writing—original draft preparation, A.S. and F.S.G.; Review and editing, M.M. and V.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are grateful to the reviewers for providing advice and guidance as to how to improve the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ABCArtificial Bee Colony
ALOAnt Lion Optimization
AVOAAfrican Vulture Optimization Algorithm
BALOBinary Ant Lion Optimization
BBOBinary Butterfly Optimization
BAOVAHBinary African Vulture Optimization Algorithm with Hyper-heuristic
BCCSABinary Chaotic Crow Search Algorithm
BCSABinary Crow Search Algorithm
BDABinary Dragonfly Algorithm
BGOBinary Grasshopper Optimization
BGWOBinary Gray Wolf Algorithm
BOAButterfly Optimization Algorithm
BSSABinary Salp Swarm Algorithm
CCSAChaotic Crow Search Algorithm
COACoyote Optimization Algorithm
CSACrow Search Algorithm
DADragonfly Algorithm
EPOEmperor Penguin Optimizer
FFAFruit Fly Algorithm
FFAFarmland Fertility Algorithm
FSFeature Selection
GAGenetic Algorithm
GOAGrasshopper Optimization Algorithm
GSGravitational Search
GSKOGaining–Sharing Knowledge-Based Optimization
GWOGrey Wolf Optimization
HHOHarris Hawks Optimization
MBAMine Blast Algorithm
PSOParticle Swarm Optimization
SASimulated Annealing
SOSSymbiotic Organisms Search
SPSASalp Swarm Algorithm
SSASalp Swarm Algorithm
WOAWhale Optimization Algorithm
IPRSInitial Population generation based on Ranking Strategy

References

  1. Hancer, E.; Xue, B.; Zhang, M. A survey on feature selection approaches for clustering. Artif. Intell. Rev. 2020, 53, 4519–4545. [Google Scholar] [CrossRef]
  2. Nadimi-Shahraki, M.H.; Banaie-Dezfouli, M.; Zamani, H.; Taghian, S.; Mirjalili, S. B-MFO: A binary moth-flame optimization for feature selection from medical datasets. Computers 2021, 10, 136. [Google Scholar] [CrossRef]
  3. Abdollahzadeh, B.; Gharehchopogh, F.S. A multi-objective optimization algorithm for feature selection problems. Eng. Comput. 2021, 38, 1845–1863. [Google Scholar] [CrossRef]
  4. Nadimi-Shahraki, M.H.; Zamani, H.; Mirjalili, S. Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study. Comput. Biol. Med. 2022, 148, 105858. [Google Scholar] [CrossRef] [PubMed]
  5. Shaddeli, A.; Gharehchopogh, F.S.; Masdari, M.; Solouk, V. BFRA: A New Binary Hyper-Heuristics Feature Ranks Algorithm for Feature Selection in High-Dimensional Classification Data. Int. J. Inf. Technol. Decis. Mak. 2022, 1–66. [Google Scholar] [CrossRef]
  6. Hosseini, F.; Gharehchopogh, F.S.; Masdari, M. A Botnet Detection in IoT Using a Hybrid Multi-objective Optimization Algorithm. New Gener. Comput. 2022, 1–35. [Google Scholar] [CrossRef]
  7. Gharehchopogh, F.S. Advances in tree seed algorithm: A comprehensive survey. Arch. Comput. Methods Eng. 2022, 1–24. [Google Scholar] [CrossRef]
  8. de Carvalho, V.R.; Özcan, E.; Sichman, J.S. Comparative Analysis of Selection Hyper-Heuristics for Real-World Multi-Objective Optimization Problems. Appl. Sci. 2021, 11, 9153. [Google Scholar] [CrossRef]
  9. Abiodun, E.O.; Alabdulatif, A.; Abiodun, O.I.; Alawida, M.; Alabdulatif, A.; Alkhawaldeh, R.S. A systematic review of emerging feature selection optimization methods for optimal text classification: The present state and prospective opportunities. Neural Comput. Appl. 2021, 33, 15091–15118. [Google Scholar] [CrossRef]
  10. Montazeri, M. HHFS: Hyper-heuristic feature selection. Intell. Data Anal. 2016, 20, 953–974. [Google Scholar] [CrossRef]
  11. Gharehchopogh, F.S.; Nadimi-Shahraki, M.H.; Barshandeh, S.; Abdollahzadeh, B.; Zamani, H. CQFFA: A Chaotic Quasi-oppositional Farmland Fertility Algorithm for Solving Engineering Optimization Problems. J. Bionic Eng. 2022, 1–26. [Google Scholar] [CrossRef]
  12. Nadimi-Shahraki, M.H.; Fatahi, A.; Zamani, H.; Mirjalili, S. Binary Approaches of Quantum-Based Avian Navigation Optimizer to Select Effective Features from High-Dimensional Medical Data. Mathematics 2022, 10, 2770. [Google Scholar] [CrossRef]
  13. Gharehchopogh, F.S. An Improved Tunicate Swarm Algorithm with Best-random Mutation Strategy for Global Optimization Problems. J. Bionic Eng. 2022, 1–26. [Google Scholar] [CrossRef]
  14. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  15. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  16. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary ant lion approaches for feature selection. Neurocomputing 2016, 213, 54–65. [Google Scholar] [CrossRef]
  17. De Souza, R.C.T.; dos Santos Coelho, L.; De Macedo, C.A.; Pierezan, J. A V-shaped binary crow search algorithm for feature selection. In Proceedings of the 2018 IEEE congress on evolutionary computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  18. Mafarja, M.; Aljarah, I.; Heidari, A.A.; Faris, H.; Fournier-Viger, P.; Li, X.; Mirjalili, S. Binary dragonfly optimization for feature selection using time-varying transfer functions. Knowl. Based Syst. 2018, 161, 185–204. [Google Scholar] [CrossRef]
  19. Faris, H.; Mafarja, M.M.; Heidari, A.A.; Aljarah, I.; Ala’M, A.-Z.; Mirjalili, S.; Fujita, H. An efficient binary salp swarm algorithm with crossover scheme for feature selection problems. Knowl. Based Syst. 2018, 154, 43–67. [Google Scholar] [CrossRef]
  20. Sayed, G.I.; Hassanien, A.E.; Azar, A.T. Feature selection via a novel chaotic crow search algorithm. Neural Comput. Appl. 2019, 31, 171–188. [Google Scholar] [CrossRef]
  21. Arora, S.; Anand, P. Binary butterfly optimization approaches for feature selection. Expert Syst. Appl. 2019, 116, 147–160. [Google Scholar] [CrossRef]
  22. Tubishat, M.; Idris, N.; Shuib, L.; Abushariah, M.A.; Mirjalili, S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020, 145, 113122. [Google Scholar] [CrossRef]
  23. de Souza, R.C.T.; de Macedo, C.A.; dos Santos Coelho, L.; Pierezan, J.; Mariani, V.C. Binary coyote optimization algorithm for feature selection. Pattern Recognit. 2020, 107, 107470. [Google Scholar] [CrossRef]
  24. Abdel-Basset, M.; El-Shahat, D.; El-henawy, I.; de Albuquerque, V.H.C.; Mirjalili, S. A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection. Expert Syst. Appl. 2020, 139, 112824. [Google Scholar] [CrossRef]
  25. Sihwail, R.; Omar, K.; Ariffin, K.A.Z.; Tubishat, M. Improved harris hawks optimization using elite opposition-based learning and novel search mechanism for feature selection. IEEE Access 2020, 8, 121127–121145. [Google Scholar] [CrossRef]
  26. Hussien, A.G.; Hassanien, A.E.; Houssein, E.H.; Amin, M.; Azar, A.T. New binary whale optimization algorithm for discrete optimization problems. Eng. Optim. 2020, 52, 945–959. [Google Scholar] [CrossRef]
  27. Mohmmadzadeh, H.; Gharehchopogh, F.S. An efficient binary chaotic symbiotic organisms search algorithm approaches for feature selection problems. J. Supercomput. 2021, 77, 9102–9144. [Google Scholar] [CrossRef]
  28. Dhiman, G.; Oliva, D.; Kaur, A.; Singh, K.K.; Vimal, S.; Sharma, A.; Cengiz, K. BEPO: A novel binary emperor penguin optimizer for automatic feature selection. Knowl. Based Syst. 2021, 211, 106560. [Google Scholar] [CrossRef]
  29. Hosseinalipour, A.; Gharehchopogh, F.S.; Masdari, M.; Khademi, A. A novel binary farmland fertility algorithm for feature selection in analysis of the text psychology. Appl. Intell. 2021, 51, 4824–4859. [Google Scholar] [CrossRef]
  30. Abdollahzadeh, B.; Gharehchopogh, F.S.; Mirjalili, S. African vultures optimization algorithm: A new nature-inspired metaheuristic algorithm for global optimization problems. Comput. Ind. Eng. 2021, 158, 107408. [Google Scholar] [CrossRef]
  31. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl. Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  32. Cowling, P.; Kendall, G.; Soubeiga, E. A hyperheuristic approach to scheduling a sales summit. In Proceedings of the International Conference on the Practice and Theory of Automated Timetabling, Konstanz, Germany, 16–18 August 2000; Springer: Berlin/Heidelberg, Germany, 2000; pp. 176–190. [Google Scholar]
  33. Cowling, P.; Kendall, G.; Soubeiga, E. A parameter-free hyperheuristic for scheduling a sales summit. In Proceedings of the 4th Metaheuristic International Conference, MIC, Fairfax, VA, USA, 26–28 July 2001; pp. 127–131. [Google Scholar]
  34. Neggaz, N.; Ewees, A.A.; Abd Elaziz, M.; Mafarja, M. Boosting salp swarm algorithm by sine cosine algorithm and disrupt operator for feature selection. Expert Syst. Appl. 2020, 145, 113103. [Google Scholar] [CrossRef]
  35. Das, B.; Prasad, K.E.; Ramamurty, U.; Rao, C. Nano-indentation studies on polymer matrix composites reinforced by few-layer graphene. Nanotechnology 2009, 20, 125705. [Google Scholar] [CrossRef] [PubMed]
  36. Al-Sharhan, S.; Bimba, A. Adaptive multi-parent crossover GA for feature optimization in epileptic seizure identification. Appl. Soft Comput. 2019, 75, 575–587. [Google Scholar] [CrossRef]
  37. Droste, S. Analysis of the (1 + 1) EA for a dynamically bitwise changing OneMax. In Proceedings of the Genetic and Evolutionary Computation Conference, Chicago, IL, USA, 9–11 July 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 909–921. [Google Scholar]
  38. Nakamura, R.Y.M.; Pereira, L.A.M.; Rodrigues, D.; Costa, K.A.P.; Papa, J.P.; Yang, X.-S. Binary bat algorithm for feature selection. In Swarm Intelligence and Bio-Inspired Computation; Elsevier: Amsterdam, The Netherlands, 2013; pp. 225–237. [Google Scholar]
  39. Abdel-Basset, M.; Sallam, K.M.; Mohamed, R.; Elgendi, I.; Munasinghe, K.; Elkomy, O.M. An improved binary grey-wolf optimizer with simulated annealing for feature selection. IEEE Access 2021, 9, 139792–139822. [Google Scholar] [CrossRef]
  40. Mirjalili, S.; Mirjalili, S.M.; Yang, X.-S. Binary bat algorithm. Neural Comput. Appl. 2014, 25, 663–681. [Google Scholar] [CrossRef]
  41. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  42. Li, Y.; Xiao, J.; Chen, Y.; Jiao, L. Evolving deep convolutional neural networks by quantum behaved particle swarm optimization with binary encoding for image classification. Neurocomputing 2019, 362, 156–165. [Google Scholar] [CrossRef]
  43. Fang, Z.; Ren, J.; Marshall, S.; Zhao, H.; Wang, S.; Li, X. Topological optimization of the densenet with pretrained-weights inheritance and genetic channel selection. Pattern Recognit. 2021, 109, 107608. [Google Scholar] [CrossRef]
  44. Wang, B.; Sun, Y.; Xue, B.; Zhang, M. A hybrid differential evolution approach to designing deep convolutional neural networks for image classification. In Proceedings of the Australasian Joint Conference on Artificial Intelligence, Wellington, New Zealand, 11–14 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 237–250. [Google Scholar]
  45. Rani, R.; Lobiyal, D. Automatic Construction of Generic Stop Words List for Hindi Text. Procedia Comput. Sci. 2018, 132, 362–370. [Google Scholar] [CrossRef]
  46. Porter, M.F. An algorithm for suffix stripping. Program 1980, 14, 130–137. [Google Scholar] [CrossRef]
  47. Xu, J.; Croft, W.B. Corpus-based stemming using cooccurrence of word variants. ACM Trans. Inf. Syst. (TOIS) 1998, 16, 61–81. [Google Scholar] [CrossRef]
  48. Weikum, G. Foundations of statistical natural language processing. ACM SIGMOD Rec. 2002, 31, 37–38. [Google Scholar] [CrossRef]
  49. Porter, M.F. Snowball: A Language for Stemming Algorithms. 2001. Available online: http://snowball.tartarus.org/texts/introduction.html (accessed on 10 September 2022).
  50. Xu, S.; Shijia, E.; Xiang, Y. Enhanced attentive convolutional neural networks for sentence pair modeling. Expert Syst. Appl. 2020, 151, 113384. [Google Scholar] [CrossRef]
Figure 1. An example of a dataset with features.
Figure 1. An example of a dataset with features.
Bdcc 06 00104 g001
Figure 2. Bitwise operations.
Figure 2. Bitwise operations.
Bdcc 06 00104 g002
Figure 3. A graphic view of various types of S-shaped transfer function.
Figure 3. A graphic view of various types of S-shaped transfer function.
Bdcc 06 00104 g003
Figure 4. A graphic view of various types of V-shaped transfer function.
Figure 4. A graphic view of various types of V-shaped transfer function.
Bdcc 06 00104 g004
Figure 5. Flowchart of the proposed BAOVAH approach.
Figure 5. Flowchart of the proposed BAOVAH approach.
Bdcc 06 00104 g005
Figure 6. Flowchart of the proposed BAVOA-v1 approach.
Figure 6. Flowchart of the proposed BAVOA-v1 approach.
Bdcc 06 00104 g006
Figure 7. Mutation neighborhood search strategy (a) feature solution, (b) ring feature solution, (c) right neighbors MNSS.
Figure 7. Mutation neighborhood search strategy (a) feature solution, (b) ring feature solution, (c) right neighbors MNSS.
Bdcc 06 00104 g007
Figure 8. Multi-parent Crossover strategy with three solutions.
Figure 8. Multi-parent Crossover strategy with three solutions.
Bdcc 06 00104 g008
Figure 9. Bitwise Operators (AND, OR).
Figure 9. Bitwise Operators (AND, OR).
Bdcc 06 00104 g009
Figure 10. Comparison of the convergence of S-shaped- and V-shaped-based methods on the first six sets.
Figure 10. Comparison of the convergence of S-shaped- and V-shaped-based methods on the first six sets.
Bdcc 06 00104 g010
Figure 11. Comparison of the convergence of S-shaped- and V-shaped-based methods on the second 6 data sets.
Figure 11. Comparison of the convergence of S-shaped- and V-shaped-based methods on the second 6 data sets.
Bdcc 06 00104 g011
Figure 12. Comparison of the convergence of S-shaped- and V-shaped-based methods on the third six data sets.
Figure 12. Comparison of the convergence of S-shaped- and V-shaped-based methods on the third six data sets.
Bdcc 06 00104 g012
Figure 13. Comparison of the convergence of S-shaped- and V-shaped-based methods on the fourth six data sets.
Figure 13. Comparison of the convergence of S-shaped- and V-shaped-based methods on the fourth six data sets.
Bdcc 06 00104 g013
Figure 14. Comparison of the convergence of S-shaped- and V-shaped-based methods on the fifth six data sets.
Figure 14. Comparison of the convergence of S-shaped- and V-shaped-based methods on the fifth six data sets.
Bdcc 06 00104 g014
Figure 15. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the first six data sets.
Figure 15. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the first six data sets.
Bdcc 06 00104 g015
Figure 16. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the second six data sets.
Figure 16. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the second six data sets.
Bdcc 06 00104 g016
Figure 17. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the third six sets.
Figure 17. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the third six sets.
Bdcc 06 00104 g017
Figure 18. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the fourth six sets.
Figure 18. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the fourth six sets.
Bdcc 06 00104 g018
Figure 19. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the fifth six data sets.
Figure 19. Comparison of the degree of convergence of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) method on the fifth six data sets.
Bdcc 06 00104 g019
Figure 20. Determining CNNEM in-depth model parameters and problem dimensions for the proposed BAOVAH approach.
Figure 20. Determining CNNEM in-depth model parameters and problem dimensions for the proposed BAOVAH approach.
Bdcc 06 00104 g020
Table 1. Different types of data sets have been used in this paper.
Table 1. Different types of data sets have been used in this paper.
IDType (Size)NameNo. of InstancesNo. of FeaturesNo. of ClassesArea
Data1lowabalone4177928Life
Data2breastcancerw699102Biology
Data3tictactoe958102Game
Data4glass214116Physical
Data5heart270142Life
Data6wine178143Chemistry
Data7mediumletterrecognition20,0001726computer
Data8seismicbumps2584192-
Data9hepatitis155202Life
Data10waveform5000223Physical
Data11spect267232Life
Data12german1000252Financial
Data13breastEW569312Biology
Data14Steel1941342Physical
Data15Dermatology366356Biology
Data16ionosphere351352Physical
Data17soybean3073619Life
Data18krvskp3195372Game
Data19highlungcancer32572Life
Data20spambase4601582computer
Data21sonar208612Physical
Data22audiology1997124Life
Data23libras3609115-
Data24LSVT1253112Life
Data25PersonGaitDataSet4832216Computer
Data26pd_speech7567552Computer
Data27ORL400102540image
Data28warppie210242110image
Data29lung20333135voice
Data30SMK-CAN-18718719,9942Biology
Table 2. Initial value of parameters in comparative algorithms.
Table 2. Initial value of parameters in comparative algorithms.
AlgorithmParameterValue
BBA [40]A0.5
r0.5
Qmin0
Qmax2
BPSO [14]C12.05
C22.05
W2
BGWO [15]--
BDA [41]--
BCCSA [20]AP0.1
f12
BFFAG [29]W1
Q0.7
R0.9
BAVOAVp10.6
(hyper-heuristic)p20.4
BAVOASp30.6
(hyper-heuristic)alpha0.8
BAVOAbetha0.2
(multi-strategy)gamma0.25
Table 3. Comparison of proposed BAVOA methods based on S-shaped and V-shaped based on the criterion of the average number of features.
Table 3. Comparison of proposed BAVOA methods based on S-shaped and V-shaped based on the criterion of the average number of features.
DatasetsBAOVAH-S1BAOVAH-S2BAOVAH-S3BAOVAH-S4BAOVAH-V1BAOVAH-V2BAOVAH-V3BAOVAH-V4
Data153.53.03.23.73.55.55.3
Data24.65.24.64.244.14.55.8
Data33.14.44.92.44.24.24.54.6
Data48.68.38.38.64.24.74.13.6
Data56.68.56.310.45.84.85.66.1
Data64.47.25.77.27.56.36.87.5
Data76.3615.57.79.67.210.411.3
Data814.916.216.116.58.96.77.36.7
Data99.311.116.511.69.410.17.56.1
Data1010.38.16.25.69.310.812.915.2
Data1113.412.518.718.27.78.18.410.2
Data1216.49.213.111.411.37.113.115.3
Data1316.215.112.313.411.212.310.214.1
Data1419.117.516.217.516.414.116.316.2
Data1514.317.212.111.514.115.121.320.4
Data1623.125.327.525.39.213.114.115.3
Data1712.410.217.313.415.318.418.318.5
Data1815.515.417.618.819.318.616.713.2
Data1945.528.239.736.126.219.719.122.9
Data2022.130.919.519.120.828.928.441.5
Data2134.931.828.433.228.114.535.438.1
Data2231.528.435.233.435.130.836.329.8
Data2361.543.254.358.841.140.147.866.5
Data24202.8163.1153.7151.4133134.8163.5133
Data25224.9189.5223.7197.193120.595.693
Data26212.7349.5404.5335.6337.8297.5312.8444.6
Data27561.8660.8669.9719.5552.8546.4438.2422.1
Data282047.21648.81620.31732.5996.7871.51081.21498.6
Data292791.62077.62291.72831.71319.71337.41491.41251.6
Data3016324.414680.214373.915590.27649.38405.68426.27665.9
Rank-low30|0130|0130|0230|0630|0430|0530|0330|12
Rank-mid30|0130|0130|0230|0630|0430|0530|0330|12
Rank-high30|0130|0130|0230|0630|0430|0530|0330|12
Ranking all30|0130|0130|0230|0630|0430|0530|0330|12
Table 4. Comparison of proposed BAOVA methods based on S-shaped and V-shaped based on average criteria.
Table 4. Comparison of proposed BAOVA methods based on S-shaped and V-shaped based on average criteria.
DatasetsBAOVAH-S1BAOVAH-S2BAOVAH-S3BAOVAH-S4BAOVAH-V1BAOVAH-V2BAOVAH-V3BAOVAH-V4
Data10.2130.2160.2090.2160.2160.2160.2160.216
Data20.970.9720.9720.970.9720.9720.970.97
Data30.7930.7930.7930.7930.7930.7930.7930.793
Data40.9470.9470.9470.9470.9470.9470.9470.947
Data50.8390.8160.8150.8140.8240.8160.8320.831
Data60.9890.9890.9780.9890.9890.9890.9780.989
Data70.9440.9360.9450.9430.9450.9470.9470.947
Data80.9380.9380.9380.9380.9380.9380.9380.938
Data90.7160.7450.7080.7290.7290.7290.7550.729
Data100.7750.780.7780.7810.7850.7780.7940.797
Data110.7660.7650.7820.7820.7820.7740.7680.773
Data120.7150.720.7340.7210.7240.7190.730.725
Data130.9440.9460.9440.9470.9520.9520.9460.967
Data140.9980.9980.9970.9980.99910.9991
Data150.9650.9650.9770.9650.9760.9790.9790.985
Data160.910.9220.9170.9270.9290.9350.9110.935
Data170.9050.9120.9120.9250.9250.9310.9440.924
Data180.9530.9540.9650.9530.9780.9750.9750.976
Data190.940.940.940.940.940.9411
Data200.9130.910.9060.9150.910.9090.9230.917
Data210.9060.9060.8870.9060.9060.9450.9250.902
Data220.740.750.710.750.80.780.790.79
Data230.8140.8190.8410.8240.8240.8240.8350.83
Data240.8270.8580.8580.8430.8430.8750.8590.875
Data250.6270.6270.6690.6270.7520.6690.5850.705
Data260.8790.8870.890.8860.8880.8880.890.896
Data270.9070.9130.9130.9170.9170.9260.9170.926
Data280.9150.9260.9260.9260.9320.9260.9440.944
Data290.9620.9720.9720.9620.9720.980.9720.972
Data300.6290.6190.6260.6190.6370.6190.6370.639
Rank-low30|0530|0730|0730|0730|0830|1330|1230|18
Rank-mid30|0530|0730|0730|0730|0830|1330|1230|18
Rank-high30|0530|0730|0730|0730|0830|1330|1230|18
Ranking all30|0530|0730|0730|0730|0830|1330|1230|18
Table 5. Comparison of proposed BAOVA methods based on Z-shaped and V-shaped based on the average objective function criterion.
Table 5. Comparison of proposed BAOVA methods based on Z-shaped and V-shaped based on the average objective function criterion.
DatasetsBAOVAH-S1BAOVAH-S2BAOVAH-S3BAOVAH-S4BAOVAH-V1BAOVAH-V2BAOVAH-V3BAOVAH-V4
Data10.8050.810.810.8110.80910.7240.8090.801
Data20.0690.0670.0730.060.0670.0960.0720.044
Data320.3210.3890.370.9250.3770.310.98510.247
Data40.2790.2730.3190.2880.25420.2270.3260.22
Data50.9180.3070.3160.2940.9630.3020.9480.952
Data60.1530.1080.2640.1170.0710.0860.1020.058
Data70.4040.4280.5760.4020.1310.3270.1480.131
Data80.1220.1140.1160.1220.1130.9840.910.106
Data90.4420.4680.4440.4670.41720.3470.3890.954
Data100.3390.3540.3920.4150.3220.3020.2650.262
Data110.4590.4090.4060.4170.4050.3640.3830.932
Data120.3530.3640.3730.3620.3350.3070.3490.323
Data130.110.1080.1150.1190.1070.1140.2040.101
Data140.2160.2050.1860.2750.1630.1750.1020.095
Data150.1950.1660.220.3040.2230.1380.0980.107
Data160.1290.1270.1290.1250.1330.1330.9010.12
Data170.46590.3850.3010.3520.2150.2240.2780.258
Data180.3190.330.3320.3160.2080.2430.2030.502
Data190.3210.4220.3630.3890.31210.2920.2950.182
Data200.1480.1350.1460.1650.1390.1320.1440.108
Data210.2090.2160.2520.2240.1790.2460.1790.164
Data220.6390.6120.5550.5980.5530.4820.4530.568
Data230.2190.2370.2320.2310.2130.2280.2110.205
Data240.3210.3480.3370.3920.3080.2620.2570.388
Data250.690.7220.7230.7040.6180.6160.6720.584
Data260.1670.1690.1580.1690.1570.1650.1560.148
Data270.1450.1310.1390.1360.1350.1220.1210.121
Data280.1020.1020.1020.1020.0940.1160.0970.101
Data290.0620.0530.0530.0560.0520.0520.0550.051
Data300.4530.4550.4520.4450.4210.4220.4330.416
Rank-low30|0530|0730|0730|0730|0830|1330|1230|18
Rank-mid30|0530|0730|0730|0730|0830|1330|1230|18
Rank-high30|0530|0730|0730|0730|0830|1330|1230|18
Ranking all30|0530|0730|0730|0730|0830|1330|1230|18
Table 6. Comparison of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) approach based on the average accuracy criterion.
Table 6. Comparison of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) approach based on the average accuracy criterion.
BBABPSOBGWOBDABCCSABFFAGBAVOA-V1
(Multi-Strategy)
BAVOAH
Data10.1740.1740.2060.2130.2060.2070.1810.207
Data20.9340.9340.9650.970.9650.5740.9270.972
Data30.5240.5240.7830.7850.7770.5880.6070.787
Data40.7360.7360.8710.9240.9440.5440.7160.878
Data50.6750.6750.8040.8070.8070.520.6730.824
Data60.8740.8740.9690.9580.9660.5260.8600.954
Data70.5590.5590.9420.9440.9310.5470.9450.945
Data80.9040.9040.9270.9370.9370.5470.9340.934
Data90.5760.5760.7290.7540.6850.5240.7320.739
Data100.6720.6720.7930.7850.7560.5160.7970.797
Data110.570.570.7310.750.7690.5460.7710.694
Data120.650.650.7150.7150.6960.5160.6240.742
Data130.9010.9010.9580.9510.9330.5320.8760.945
Data140.7370.7370.9970.9910.9820.5320.7900.992
Data150.8560.8560.9770.9760.9450.5290.8450.955
Data160.8780.8780.9040.9030.9090.5270.8810.935
Data170.6470.6470.9420.9420.870.5230.6860.946
Data180.6960.6960.9570.9210.9140.510.6380.922
Data190.7030.7030.9280.9840.8750.5760.5830.929
Data200.8480.8480.910.90.8990.5660.8350.920
Data210.7970.7970.9170.9270.8650.5530.7780.876
Data220.4940.4940.7840.8310.680.5280.7920.833
Data230.7810.7810.8250.8250.8250.5360.7860.786
Data240.7070.7070.8170.8540.8410.5540.6640.900
Data250.3380.3380.540.6790.5830.570.6720.550
Data260.8410.8410.9140.8750.8760.570.8350.903
Data270.8740.8740.9220.9010.910.5840.8640.924
Data280.9010.9010.9230.9260.9260.9260.9260.926
Data290.9510.9510.970.960.9710.5650.9600.971
Data300.560.560.6240.6240.6170.5750.5990.636
Rank-low06|0006|0006|0106|0206|0106|0006|0006|04
Rank-mid12|0012|0012|0212|0412|0012|0012|0412|07
Rank-high12|0012|0012|0112|0412|0212|0012|0112|09
Ranking all30|0030|0730|0430|0730|0330|1330|0530|20
Table 7. Comparison of proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (Multi-strategy), an approach based on the criterion of the average number of features.
Table 7. Comparison of proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (Multi-strategy), an approach based on the criterion of the average number of features.
BBABPSOBGWOBDABCCSABFFAGBAVOA-V1
(Multi-Strategy)
BAVOAH
Data12.54.565.654.92.46.83.5362.3
Data24.36.305.4543.27.64.7543.9
Data32.65.285.75.0598.34.56.14
Data457.508.4118.158.7293.64
Data56.158.5210.46.95811.655.5345.534
Data64.357.807.94.05510.957.47.2
Data75.256.5613.1111613.911.574.5
Data84.910.256.051114.615.7996.63
Data96.410.5812.28.05316.5512.3846.38
Data107.210.3617.6515.12118.5515.976.912
Data116.45216.96.3419.1519.59912.24
Data1210.1511.7417.613.7421.059.6219.59
Data1312.4520.3015.811.0511.520.314.6711.48
Data149.157.521.37.53326.7522.6914.12
Data1514.218.52410.651829.116.5718.15
Data1613.517.7517.7514.2628.6529.4618.10
Data1710.353227.0516.953530.514.7824.09
Data1815.220.325.813.352730.118.92120.21
Data1918.429.1231.212.71325.3529.0845.79
Data2020.526.9540.9539.25750.420.4720.46
Data2119.7537.6541.0523.951953.0530.39418.60
Data2224.326.352.3526.32537.632.25459.45
Data2330.5531.2854.7539.73854.826.69577.12
Data24118.45191.5223.35147.25151272161.622111.49
Data25119.2159.45203.6122.3123274.1593.01124.43
Data26257.4673.4571.85472.85221673.4436.011345.23
Data27366.45715.8715.8689.95273905.45574.222270.047
Data28791.652042.831297.651053.35987.52062.451563.67652.325
Data29964.611936.751375.18632834.11960.242876.363
Data306042114129.259682.8989417652.251163919742
Rank-low06|0006|0006|0106|0206|0106|0006|0006|04
Rank-mid12|0012|0012|0212|0412|0012|0012|0412|07
Rank-high12|0012|0012|0112|0412|0212|0012|0112|09
Ranking all30|0030|0730|0430|0730|0330|1330|0530|20
Table 8. Comparison of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) an approach based on the criterion of the best value of the fitness function.
Table 8. Comparison of the proposed BAOVAH (hyper-heuristic) and BAVOA-V1 (multi-strategy) an approach based on the criterion of the best value of the fitness function.
BBABPSOBGWOBDABCCSABFFAGBAVOA-V1
(Multi-Strategy)
BAVOAH
Data10.7920.7840.7830.7830.7880.7830.7830.783
Data20.0380.0360.0360.0330.0370.0330.0330.033
Data30.2510.2120.2120.2120.2310.2310.2120.212
Data40.0580.0570.1290.0570.0570.0570.0570.057
Data50.2270.2050.1840.1740.1940.1760.1740.173
Data60.0150.0270.0280.0140.0370.0170.0220.013
Data70.130.0640.0640.060.0780.060.0600.059
Data80.0640.0720.0730.0630.0630.0630.0630.063
Data90.3190.2970.2720.2450.2810.2580.2690.244
Data100.2350.2170.210.2140.2520.210.2270.205
Data110.2990.1970.2670.2320.230.2350.2130.195
Data120.3060.2530.2760.2530.3020.2660.2760.250
Data130.0560.0460.050.0490.0690.0470.0520.039
Data140.0170.0040.0090.0020.0270.0020.0040.002
Data150.0640.0260.0280.0140.0590.020.0250.019
Data160.1040.0830.090.0660.0920.0590.0530.063
Data170.1720.070.0650.0560.1390.0830.0810.049
Data180.0710.0360.0480.0230.0920.0220.0260.022
Data190.0020.1270.0670.0020.1260.0020.0020.001
Data200.1030.0940.0940.0850.110.0880.0920.081
Data210.1090.0620.0830.0510.1360.0830.0590.032
Data220.2820.2110.2150.1620.320.1430.2410.141
Data230.190.1630.1760.1640.1730.1590.1750.157
Data240.160.0670.1010.0990.1620.1170.1430.051
Data250.3750.2930.460.2920.4150.2930.3280.250
Data260.1250.0750.0910.0890.1260.1050.1060.080
Data270.0980.0740.0810.0750.0920.0860.0770.073
Data280.0870.080.0810.080.0780.070.0690.066
Data290.0410.0430.0350.0430.0320.0340.0310.021
Data300.4030.3630.3760.3630.3840.3760.3610.325
Rank-low06|0006|0006|0106|0206|0106|0006|0006|04
Rank-mid12|0012|0012|0212|0412|0012|0012|0412|07
Rank-high12|0012|0012|0112|0412|0212|0012|0112|09
Ranking all30|0030|0730|0430|0730|0330|1330|0530|20
Table 9. Specifications of CNNEM model and its modifiable neurons using proposed BAOVAH approach.
Table 9. Specifications of CNNEM model and its modifiable neurons using proposed BAOVAH approach.
Model: “CNNEM”
Layer (Type)Output ShapeParam #Changeable
Embedding_layer (Embedding)(None, 100, 100)261,2000–1024
Conv_1 (Conv1D)(None, 97, 64)25,6640–512
drop_1 (Dropout)(None, 97, 64)0-
MaxPool_1 (MaxPooling1D)(None, 49, 64)0-
drop_2 (Dropout)(None, 49, 64)0-
Conv_2 (Conv1D)(None, 48, 32)41280–512
drop_3 (Dropout)(None, 48, 32)0-
MaxPool_2 (MaxPooling1D)(None, 24, 32)0-
flatten_2 (Flatten)(None, 768)0-
dense_1 (Dense)(None, 16)12,3040–1024
drop_4 (Dropout)(None, 16)0-
dense_2 (Dense)(None, 8)1360–1024
drop_5 (Dropout)(None, 8)0-
dense_3 (Dense)(None, 4)360–1024
output (Dense)(None, 2)10-
Table 10. A comparison of the accuracy of the two in-depth models.
Table 10. A comparison of the accuracy of the two in-depth models.
ModelCNNEMCNNEMBH
DatasetTrainTestTrainTest
IMDB0.980.730.990.79
Amazon0.540.450.980.78
Yelp0.520.480.980.78
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shaddeli, A.; Soleimanian Gharehchopogh, F.; Masdari, M.; Solouk, V. An Improved African Vulture Optimization Algorithm for Feature Selection Problems and Its Application of Sentiment Analysis on Movie Reviews. Big Data Cogn. Comput. 2022, 6, 104. https://doi.org/10.3390/bdcc6040104

AMA Style

Shaddeli A, Soleimanian Gharehchopogh F, Masdari M, Solouk V. An Improved African Vulture Optimization Algorithm for Feature Selection Problems and Its Application of Sentiment Analysis on Movie Reviews. Big Data and Cognitive Computing. 2022; 6(4):104. https://doi.org/10.3390/bdcc6040104

Chicago/Turabian Style

Shaddeli, Aitak, Farhad Soleimanian Gharehchopogh, Mohammad Masdari, and Vahid Solouk. 2022. "An Improved African Vulture Optimization Algorithm for Feature Selection Problems and Its Application of Sentiment Analysis on Movie Reviews" Big Data and Cognitive Computing 6, no. 4: 104. https://doi.org/10.3390/bdcc6040104

APA Style

Shaddeli, A., Soleimanian Gharehchopogh, F., Masdari, M., & Solouk, V. (2022). An Improved African Vulture Optimization Algorithm for Feature Selection Problems and Its Application of Sentiment Analysis on Movie Reviews. Big Data and Cognitive Computing, 6(4), 104. https://doi.org/10.3390/bdcc6040104

Article Metrics

Back to TopTop