Next Article in Journal
A Review on Reinforcement Learning in Production Scheduling: An Inferential Perspective
Next Article in Special Issue
Integrating Eye Movement, Finger Pressure, and Foot Pressure Information to Build an Intelligent Driving Fatigue Detection System
Previous Article in Journal
Fast Minimum Error Entropy for Linear Regression
Previous Article in Special Issue
New Multi-View Feature Learning Method for Accurate Antifungal Peptide Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of Gene Selection for Cancer Classification in High-Dimensional Data Using an Improved African Vultures Algorithm

by
Mona G. Gafar
1,2,*,
Amr A. Abohany
3,†,
Ahmed E. Elkhouli
4,† and
Amr A. Abd El-Mageed
5,†
1
Department of Computer Engineering and Information, College of Engineering in Wadi Alddawasir, Prince Sattam bin Abdulaziz University, Kharj 16278, Saudi Arabia
2
Machine Learning and Information Retrieval Department, Artificial Intelligence, Kafrelsheikh University, Kafrelsheikh 33511, Egypt
3
Faculty of Computers and Information, Kafrelsheikh University, Kafrelsheikh 33511, Egypt
4
Department of Biomedical Engineering, Faculty of Electrical Engineering, Menofia University, Menofia 32951, Egypt
5
Department of Information Systems, Sohag University, Sohag 82511, Egypt
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Algorithms 2024, 17(8), 342; https://doi.org/10.3390/a17080342
Submission received: 27 April 2024 / Revised: 16 July 2024 / Accepted: 1 August 2024 / Published: 6 August 2024
(This article belongs to the Special Issue Algorithms for Feature Selection (2nd Edition))

Abstract

:
This study presents a novel method, termed RBAVO-DE (Relief Binary African Vultures Optimization based on Differential Evolution), aimed at addressing the Gene Selection (GS) challenge in high-dimensional RNA-Seq data, specifically the rnaseqv2 lluminaHiSeq rnaseqv2 un edu Level 3 RSEM genes normalized dataset, which contains over 20,000 genes. RNA Sequencing (RNA-Seq) is a transformative approach that enables the comprehensive quantification and characterization of gene expressions, surpassing the capabilities of micro-array technologies by offering a more detailed view of RNA-Seq gene expression data. Quantitative gene expression analysis can be pivotal in identifying genes that differentiate normal from malignant tissues. However, managing these high-dimensional dense matrix data presents significant challenges. The RBAVO-DE algorithm is designed to meticulously select the most informative genes from a dataset comprising more than 20,000 genes and assess their relevance across twenty-two cancer datasets. To determine the effectiveness of the selected genes, this study employs the Support Vector Machine (SVM) and k-Nearest Neighbor (k-NN) classifiers. Compared to binary versions of widely recognized meta-heuristic algorithms, RBAVO-DE demonstrates superior performance. According to Wilcoxon’s rank-sum test, with a 5% significance level, RBAVO-DE achieves up to 100% classification accuracy and reduces the feature size by up to 98% in most of the twenty-two cancer datasets examined. This advancement underscores the potential of RBAVO-DE to enhance the precision of gene selection for cancer research, thereby facilitating more accurate and efficient identification of key genetic markers.

1. Introduction

Deoxyribonucleic Acid (DNA) makes up our genetic code, containing the recipe for our existence. Although every cell has the same DNA, each tissue structure is unique and serves a distinct purpose. The RNA transcription mechanism determines which genes in a cell are active, enabling RNA to be transformed into proteins responsible for the cell’s structure and functionality [1]. We analyze RNA-Seq gene expression data (GED) to determine genetic changes and evaluate disease biomarkers. Differential expression analysis identifies quantitative distinctions in gene expression, allowing us to categorize genes whose expression changes under various conditions. This method helps us understand illnesses and find ways to manage them. Gene expression profiling technologies have significantly developed over the years [2].
Two popular techniques are micro-array, which uses a hybridization-based approach, and RNA-Seq, which is based on next-generation sequencing. These technologies have demonstrated their value in advancing our understanding of gene expression [3]. Notable studies, such as those by Chen et al. [4] and Nunez et al. [5], have highlighted the potential of these techniques, while Wang et al. [6] have provided valuable insights into the RNA-Seq approach.
The two methods under consideration serve the purpose of quantifying gene expression for classification and statistical investigation. This paper has chosen the quantification data obtained through the next-generation sequencing (NGS)-based RNA-Seq approach because it offers more accuracy in detecting RNA quantification levels compared to micro-array data [7]. RNA-Seq has overcome several limitations of micro-array analysis, such as dependence on prior sequencing knowledge, which restricted the detection range [8,9]. Unlike micro-arrays, RNA-Seq does not require any previous knowledge and significantly expands the dynamic detection range. This enhancement improves the accuracy of results and enables the identification of a comprehensive set of genes, providing a precise understanding of disease biomarkers. As a result, RNA-Seq has emerged as a potent tool for researchers, offering valuable insights into the underlying molecular mechanisms of various diseases [10].
The “dimensionality curse” [11] has become a common challenge in contemporary times owing to the abundance of available data. This has caused a surge in the development of feature selection (FS) techniques and algorithms. FS algorithms can be classified into four distinct methods: filter, wrapper, embedded, and hybrid [12,13]. These approaches aim to identify the most informative features that can distinguish between different classes. In our case, we are concerned with identifying the genes linked to tumors.
The filter approach is a widely used technique in FS that involves evaluating the relevance of individual genes based on statistical scores. This method is known for its high accuracy in selecting the most satisfactory group of genes. Nevertheless, the filter approach has some limitations, as functioning on each gene individually ignores the interrelationships between genes, which can result in a local optima issue [14]. It is worth noting that the filter approach can be categorized into two sub-types—univariate and multivariate—with the latter considering the correlations between genes. Some examples of the filter approach include Relief [15], Fisher score [16], t-test [17], and information gain [18].
The wrapper method investigates all feasible gene subsets to test and create a subset of genes to determine their implications. A specific classifier is employed to determine the outcome of each subset, and the categorization technique is used multiple times for each assessment. Compared to the filter approach, the wrapper approach delivers superior performance by using a categorization technique that directs the learning procedure. However, this method requires substantial time and computational resources, especially when dealing with large-scale data [19].
Meta-heuristic methods (MHMs) are high-level heuristics used in mathematical optimization and computer science to find satisfactory solutions to optimization problems when information is imperfect or resources are limited [20]. MHMs sample a subset of solutions, making them useful for various optimization issues due to their few assumptions about the problem [21]. However, they do not guarantee globally optimal solutions. Many MHMs employ stochastic optimization, meaning the solution is based on random variables generated during the process [22].
Meta-heuristic methods (MHMs) are more practical than traditional iterative methods and optimization algorithms because they can explore a broader range of possible solutions. Consequently, MHMs have become a preferred strategy for solving optimization problems [23]. Numerous research papers have demonstrated that among the different wrapper methods, MHMs are well suited to address the feature selection (FS) issue. Stochastic methods, including MHMs, can produce optimal or near-optimal results quickly. They offer benefits such as flexibility, self-management without detailed mathematical properties, and the ability to assess numerous outcomes simultaneously. Various MHMs have recently been developed to solve the FS problem, providing reliable near-optimal solutions at significantly reduced computational costs [24].
Embedded FS methods utilize a learning technique to select appropriate genes that interact with the classification process. This method combines the FS technique as part of the learning model. The learning algorithm is trained with an initial attribute subset to estimate a measure for evaluating the rank values of attributes [25]. The ultimate goal is to decrease the computation time for reclassifying different subsets by incorporating the FS stage into the training process. Some techniques in this approach perform feature weighting based on regularization models with objective functions that minimize fitting errors while enforcing feature coefficients to be small or precisely zero. Examples of embedded methods are the First-Order Inductive Learner (FOIL) rule-based feature subset selection algorithm and SVM based on Recursive Feature Elimination (SVM-RFE) [26].
The hybrid method is a well-crafted technique that amalgamates the filter and wrapper methods to leverage the strengths of each. This approach begins with a filter that reduces the feature space dimensionality, generating multiple subsets with intermediate complexity. Subsequently, a wrapper is employed as a learning technique to choose the most suitable candidate subset [27]. The hybrid approach integrates the accuracy of wrappers and the efficiency of filters, resulting in an optimal methodology [28,29].

1.1. Motivation and Contributions

An effective global meta-heuristic optimization algorithm is presented in this paper, which is called the African Vultures Optimization (AVO) algorithm [30], which imitates the living and eating habits of African vultures. The proposed algorithm consists of four basic phases: population division into three groups—the best solution, the second-best solution, and the remaining solutions; famine-level measurement to formulate a mathematical model for the vultures’ exploitation, exploration, and transfer; exploration of vultures employing two strategies to allow vultures to cover large distances over extended periods to find food at random sites; and exploitation of vultures involving two sub-phases. Two distinct strategies are used in the first sub-phase: siege fight for strong vultures and rotational flight, which forms a spiral activity between the outstanding vulture and the rest. Two strategies are utilized in the second sub-phase: assembling vultures around the food source and a hostile siege fight, where vultures become more aggressive and attempt to steal food left behind by healthy vultures.
This study proposes an enhanced binary variant of the AVO technique, namely the RBAVO-DE technique, which is a productive approach that demonstrates accurate performance in handling the GS problem. At first, there is a high possibility that the recommended technique will steer clear of local optima and accomplish sufficient examination precision, quick convergence, and improved stability. Compared to recent MHMs, the proposed RBAVO-DE technique achieves enhanced efficacy by producing optimal or substantially optimal solutions for many of the analyzed situations. The Relief algorithm is utilized to identify only the related features to produce the final classification dataset. RBAVO-DE combines the Relief algorithm with the DE approach to increase exploration capability and obtain the best results within the solution space via repetitions. It also utilizes a transfer function to transform real position values into binary values. The RBAVO-DE approach makes sense in GS because it is simple to comprehend and implement, can deal with various optimization problems, produces valuable results in an acceptable amount of time, requires less computing power, and uses a small number of control parameters. This paper’s primary contributions are as follows:
  • Level 3 data based on next-generation sequencing (RNA-Seq) have been pre-processed.
  • The ability of the proposed AVO meta-heuristic algorithm to address the GS problem has not been investigated, as RNA-Seq GED has never been used with the AVO algorithm.
  • To construct a binary version known as the RBAVO-DE algorithm, the AVO algorithm is altered and recreated.
  • The proposed RBAVO-DE algorithm combines the binary variant of the AVO with a Relief approach and DE to improve the exploration ability of the search space and enhance the achieved optimal results.
  • The proposed RBAVO-DE algorithm is applied to RNA-Seq GED for the first time.
  • Several performance indicators, including average fitness, classification accuracy, number of selected genes, precision, recall, and F1-score, are used to assess the outcomes.
  • A comparison is made between the impact of the presented RBAVO-DE algorithm, employing the two recommended ML classifiers (SVM and k-NN), and other algorithms in the literature.
  • Twenty-two distinct cancer datasets are used to assess the proposed RBAVO-DE method, and the results are shown.
  • The chosen genes are investigated using biomarkers associated with cancer.

1.2. Structure

This paper is structured into five main sections. Section 2 reviews previous research on FS using RNA-Seq GED. This is followed by Section 3, which offers an in-depth discussion of the RBAVO-DE algorithm, an enhanced version of AVO, including its parameters for addressing GS. Section 4 showcases the experimental findings compared to several recent MHMs. Lastly, Section 6 concludes this study and proposes avenues for future investigation.

2. Related Works

In this section, we discuss the literature that focuses on the methods researchers use to classify RNA-Seq GED, which typically has high dimensionality. To achieve optimal performance of classification methods, it is crucial to disregard irrelevant and unrelated genes; hence, selecting appropriate genes is a vital stage before utilizing ML and deep learning (DL) techniques [31] or any other classification techniques. In this regard, we have explored some relevant papers in this domain to accomplish the objective of RNA-Seq categorization for cancer identification.
Yaqoob et al. [32] introduced a cutting-edge technique for GS known as the Sine-Cosine–Cuckoo Search Algorithm (SCCSA), a hybrid method tailored to function alongside established ML models such as SVM. This innovative GS algorithm was evaluated using a breast cancer benchmark dataset, where its performance was meticulously analyzed and compared with other GS methodologies. To refine the selection of features, the minimum Redundancy Maximum Relevance method was initially applied as a preliminary filtering step. Subsequently, the hybrid SCCSA approach was deployed to further improve and fine-tune the GS process. The final stage involved using the SVM classifier to classify the dataset based on the selected genes. Considering the critical importance of GS in decoding complex biological datasets, SCCSA emerges as a crucial asset for the classification of cancer-related datasets.
Joshi et al. [33] introduced an innovative optimization strategy named PSO-CS integrated with DL for brain tumor classification. This method enhances the efficiency of the Particle Swarm Optimization (PSO) technique by incorporating the Cuckoo Search (CS) algorithm, optimizing the classification process. Following this optimization, PSO-CS utilizes DL to classify GED related to brain tumors, identifying various classes associated with specific tumors alongside the PSO-CS optimization method. By integrating the PSO-CS technique with DL, it significantly outperformed other DL and ML models regarding classification accuracy, as evidenced by various performance measures.
Mahto et al. [34] unveiled a groundbreaking approach for cancer classification through an integrated method based on CS-SMO (CS and Spider Monkey Optimization) for GS. Initially, the fitness function of the Spider Monkey Optimization (SMO) algorithm is modified using the CS algorithm. This modification leverages the strengths of both MHMs to identify a subset of genes capable of predicting cancer at an early stage. To further refine the accuracy of the CS-SMO algorithm, a pre-processing step, MRMR, is employed to reduce the complexity of cancer gene expression datasets. Subsequently, these gene subsets are processed using DL to classify different cancer types. The efficacy of the CS-SMO approach coupled with DL was evaluated using eight benchmark micro-array gene expression datasets for cancer, examining its performance across various measures. The CS-SMO method integrated with DL demonstrated superior classification accuracy across all examined large-scale gene expression datasets for cancer, outperforming existing DL and ML models.
Neggaz et al. [35] introduced an improved version of the manta ray foraging optimization, called MRFO-SC, which used trigonometric operators inspired by the sine-cosine (SC) algorithm to handle the GS issue. The k-NN model was used for gene-set selection. In addition, the statistical significance of the MRFO-SC was evaluated using Wilcoxon’s rank-sum test at a 5% significance level. The results were evaluated and compared with some recent MHMs. The comparison and experimental results confirmed the effective performance of the proposed MRFO-SC on high- and low-dimensional benchmark datasets by obtaining the greatest classification accuracy on 85% of the GS benchmark datasets.
Lyu et al. [36] explored cancer biomarkers by focusing on genes’ significance in their impact on classification. They followed a two-phase approach—data pre-processing and utilizing a convolutional neural network—to classify the type of tumor. In the second phase, they created heat maps for each category to identify genes related to pixels with the highest intensities in the heat maps. They then evaluated the selected genes’ pathways. During pre-processing, they removed the levels of gene expression that had not been modified during the GS phase, using a variance threshold of 1.19, which decreased the number of genes from 19,531 to 10,381. The final classification accuracy obtained was 95.59, which is good but could still be improved using a more effective FS methodology to further decrease data dimensionality.
Khalifa et al. [37] built on the aforementioned paper [36]. The focus of their study was on five types of cancer data: Uterine Corpus Endometrial Carcinoma (UCEC), Lung Squamous Cell Carcinoma (LUSC), Lung Adenocarcinoma (LUAD), Kidney Renal Clear Cell Carcinoma (KIRC), and Breast Invasive Carcinoma (BRCA). The benchmark dataset used for the study comprised 2086 records and 972 attributes. Each record provided detailed sample information, while each attribute included the RNA-Seq values for a specific gene, represented as RPKM (Reads Per Kilobase per Million) [38]. The researchers employed a mixed approach using binary PSO with the decision trees (BPSO-DT) method to pre-process the data. Out of 971 attributes, 615 were selected as the best attributes of RNA-SEquation The proposed method achieved an overall testing classification accuracy of 96.90%, as demonstrated by the suggested outcomes and evaluation measures used.
Xiao et al. [39] assessed their methodology using three RNA-Seq datasets: Stomach Adenocarcinoma (STAD), BRCA, and LUAD. Their approach relied on DL techniques, wherein they employed five classifiers and subsequently utilized the DL technique to ensemble each output of the five classifiers. This led to an improvement in the classification accuracy of all the predictions, with the BRCA dataset achieving 98.4% accuracy, STAD achieving 98.78% accuracy, and LUAD achieving 99.20% accuracy.
Liu et al. [40] used micro-array data and a hybrid approach to address the problem. Unlike the previously mentioned papers, they studied each cancer class separately. To assess performance, they employed four gene benchmarks related to small round blue cell tumors, colon cancer, lung cancer, and leukemia. Their proposed method used Relief as the pre-processing technique to eliminate genes with a lower correlation with the specific cancer class, followed by PSO as the search technique. Finally, they employed the SVM model to evaluate the classification accuracy of the selected subset of genes and obtain the conclusive optimum gene subset for each cancer type.
Based on the existing research, it seems that most studies using RNA-Seq GED are still in their early stages, with researchers attempting to implement and test various ideas in this promising area. While the literature contains a plethora of experiments employing multiple techniques, such as FS and DL recent methods, no single technique is perfect due to the high dimensionality of RNA-Seq GED. FS of RNA-Seq GED plays a crucial role in determining the relationship between a gene and its category. It is a critical pre-processing task for validating gene biomarkers of cancer and overcoming the dimensionality curse. Consequently, this study aims to introduce a new wrapper approach, the RBAVO-DE algorithm, and apply it to RNA-Seq data for the first time. It also compares the proposed algorithm’s effectiveness with that of other FS techniques.

3. Proposed RBAVO-DE for GS

An improved variant of AVO known as RBAVO-DE is proposed in this paper to discover the smallest relevant gene subsets from the classification process and to disregard irrelevant genes. RBAVO-DE utilizes a Relief-based binary AVO algorithm combined with the DE technique. RBAVO-DE’s primary feature is its ability to maximize accuracy while utilizing the fewest features possible. The proposed RBAVO-DE consists of two primary steps. First, there is a pre-processing step in which the Relief algorithm is used to determine which features are significant by assigning each feature a weight and then removing the features that are irrelevant and have the lowest weights. In the second step, the binary AVO algorithm and the DE technique are applied to identify the more pertinent and unique features. The AVO algorithm is prone to the local optimum trap when handling large-scale problems. The AVO algorithm incorporates the DE technique to prevent this.
The proposed RBAVO-DE algorithm for tackling the GS strategy requires applying the Relief algorithm, initializing, position boosting using the AVO algorithm, binary conversion, fitness appraisal, and integration with DE. The following subsections explain these steps.

3.1. Applying the Relief Algorithm for Feature Filtration

This step aims to pre-process the population using the Relief algorithm [41], which is considered a fast, easy, and efficient filtering technique for finding features related to one another. This algorithm’s primary goal is to find characteristics that differentiate sample values and group similar samples close together. As a result, the method depends on the weighted ranking of features, where a feature with a higher weight indicates better classification performance.
After choosing a sample randomly, the Relief algorithm examines two different kinds of closest samples: near-hit samples, which are associated with samples from the same class, and near-miss samples, which are associated with samples from other classes. The near-hit and near-miss values can be used to determine the features’ weights. To assess the importance in the classification process, the features’ weights are arranged from most significant to least significant. Finally, the features with the most significant weights are selected. The following formula can be used to determine the weight W for feature A:
W A = j = 1 N x A j N M ( x j ) A 2 x A j N H ( x j ) A 2 .
where W A denotes the weight of feature A, x A j denotes the value of feature A for data x j , and N denotes the number of samples. The nearest data points to x j in the same or different classes are N H ( x j ) and N M ( x j ) .
The Relief algorithm narrows its focus to only the necessary features and minimizes the search area to help the AVO algorithm find better features more quickly.

3.2. Initializing the Population

The proposed BAVO algorithm starts by randomly generating a population of N positions. A vector of dimensions D equivalent to the number of features in the original dataset describes each position as a possible solution at its constrained lower and upper bounds. This randomly initialized step, bounded within the [ 1 , 1 ] range for each position vector’s variable, uses the randomly produced position.

3.3. Boosting Positions via the AVO Algorithm

This paper presents the AVO algorithm [30], a meta-heuristic optimization algorithm inspired by vultures in Africa and their living and feeding behaviors. Based on fundamental ideas about vultures, the AVO algorithm is configured as follows: The African vulture population size is initially presumed by the AVO algorithm to consist of N vultures. The population of African vultures is then divided into three groups to reflect the primary natural function of vultures, based on the computed fitness function. The first group consists of the strongest vulture, which is the best solution; the second group contains a vulture weaker than the first, which is the second-best solution; and the final group encompasses the remaining weaker vultures, which are the worst solutions.
Based on the above, the proposed AVO algorithm is composed of four fundamental phases to simulate the behavior of different types of vultures. The following subsections clarify these phases.

3.3.1. Phase of Dividing the Population

The initial population is split into groups by assessing the solution’s fitness function. The best solution is chosen as the best vulture in the first group, while the second group contains the second-best solution. The third group contains the remaining solutions. As the solutions constantly strive to approach the best and second-best solutions, the population needs to be re-evaluated for each iteration, as follows:
R g = B e s t V u l t u r e 1 g , if p r g = L 1 , S e c o n d B e s t V u l t u r e 2 g , if p r g = L 2 ,
p r g = υ g g = 1 N υ g .
The best vulture in the first group and the second-best vulture in the second group at the g t h iteration are denoted by the expressions B e s t V u l t u r e 1 g and S e c o n d B e s t V u l t u r e 2 g , respectively. The probability of selecting the best solution for each group at the g t h iteration is defined using the roulette wheel approach, as shown in Equation (3), p r g . Two random parameters within the range [ 0 , 1 ] are L 1 and L 2 .

3.3.2. Phase of Measuring the Famine Level

This phase is utilized to formulate a mathematical model for the exploitation, exploration, and transfer among the vultures. When they are not famished, the vultures have a greater ability to search for food and can fly further. Furthermore, when famished, vultures cannot fly long distances in search of food and may become violent. The i t h vulture’s famine level ( F i g ) during the g t h iteration can be written as follows, which is used to develop a mathematical model for the vultures’ exploitation, exploration, and transfer:
F i g = ( 2 × r a n d + 1 ) × z × 1 g i G m a x + t ,
t = h × sin w π 2 × g i G m a x + cos π 2 × g i G m a x 1 .
The vultures are presumably full since the variable F i g shows the vultures’ shift from exploration to exploitation. A random number between 0 and 1 is called a r a n d . z denotes a random value within the interval [ 1 , 1 ] . The current iteration number is denoted by g i , while the maximum iteration number is denoted by G m a x . Equation (5) calculates the t value to help solve complicated optimization problems more effectively and prevent reaching a local optimum. An arbitrary value within the interval [ 2 , 2 ] is represented by h. The probability of carrying out the exploration process is regulated by the predefined constant parameter w; the exploration likelihood grows as its value increases. Exploration is less probable as its value declines.
Equation (4) states that as the number of iterations increases, F i g gradually lessens. As a result, the next step in the proposed AVO algorithm can be defined as follows:
Exploration phase ( Search food in diverse areas ) , if | F i g | 1 , Exploitation phase ( Search food in surrounding area ) , if | F i g | < 1 .

3.3.3. Phase of Exploration

The vultures are characterized by their great ocular ability to locate appropriate food during this phase. There are two different strategies used in this phase to enable vultures to travel great distances for lengthy periods to search for food in random locations. These locations are selected using a random number r a n d P 1 and a preset parameter P 1 , both of which have values within the range [ 0 , 1 ] . In the exploration phase, the famine level | F i g | is greater than or equal to 1. The exploration techniques are described as follows:
X i g + 1 = R g D i g × F i g , if r a n d P 1 P 1 , R g F i g + r a n d × ( ( U B L B ) × r a n d + L B ) , if r a n d P 1 > P 1 , if | F i g | 1 ,
D i g = | ( 2 × r a n d ) × R g X i g | .
where the upcoming updated position at the next ( g + 1 ) t h iteration is denoted by X i g + 1 . The best vulture chosen for the current iteration g is denoted by R g . This is determined using Equation (2). The vultures move randomly to protect the food from other vultures and to provide a high degree of randomness in their search behavior. r a n d is a random value between zero and one. The variables’ upper bound is denoted by U B ; their lower bound by L B ; and the current position at the g t h iteration is X i g .

3.3.4. Phase of Exploitation

In this phase, | F i g | is less than 1. The proposed AVO algorithm’s efficacy is evaluated in two sub-phases that comprise the exploitation phase. Each of these sub-phases uses two different strategies. For each internal sub-phase, the appropriate strategy is chosen using two predefined parameters: P 2 for the first sub-phase and P 3 for the second, with values ranging from 0 to 1. These two internal sub-phases are explained as follows:
  • First sub-phase of exploitation: This sub-phase applies two different strategies when | F i g | is less than 1 and greater than or equal to 0.5 . The selection of one of these two strategies is made using a random value r a n d P 2 within the range [ 0 , 1 ] and a specified parameter P 2 .
    The initial strategy of this sub-phase is called siege fight, which involves sufficiently powerful and somewhat satiated vultures. More robust and healthier vultures try not to share food with others because they convene around a single food source. By swarming near the healthy vultures and engaging in small fights, the weaker vultures try to take food from them. Conversely, the second strategy is called rotational flight; it creates a spiral movement between a superior vulture and the others. The following illustrates the strategies for the first exploitation sub-phase:
    X i g + 1 = D i g × ( F i g + r a n d ) d t g , if r a n d P 2 P 2 , R g ( S 1 g + S 2 g ) , if r a n d P 2 > P 2 , if 1 > | F i g | 0.5 ,
    d t g = R g X i g ,
    S 1 g = R g × r a n d × X i g 2 π × cos ( X i g ) ,
    S 2 g = R g × r a n d × X i g 2 π × sin ( X i g ) .
    where the vulture’s next updated position at the next ( g + 1 ) t h iteration is denoted by X i g + 1 , D i g is defined using Equation (8), and r a n d is an arbitrary value between 0 and 1. The distance d t g between the vulture and one of the best two vultures is estimated using Equation (10); R g denotes the appropriate best vulture in the current g t h iteration, which is computed using Equation (2); S 1 g and S 2 g are computed employing Equations (11) and (12), respectively; and X i g denotes the current position at the g t h iteration.
  • Second sub-phase of exploitation: This sub-phase is carried out when | F i g | is less than 0.5 . Various vulture species gather around the food supply and engage in many sieges and brawls during this sub-phase. This sub-phase employs two different strategies. A predefined parameter P 3 and an arbitrary value r a n d P 3 , with values ranging from 0 to 1, are used to decide which of these two strategies to use.
    This sub-phase’s initial strategy is called assemble vultures around the food source. In this strategy, different kinds of vultures search for food and may compete near a single source. The second strategy is known as hostile siege fight. In this strategy, the vultures become more aggressive and try to plunder the leftover food from the healthy vultures by flocking around them in different ways. The healthy vultures, on the other hand, deteriorate and are unable to fend off the other vultures. The following illustrates the strategies for the second exploitation sub-phase:
    X i g + 1 = A 1 g + A 2 g 2 , if r a n d P 3 P 3 , R g | d t g | × F i g × L e v y d , if r a n d P 3 > P 3 , if | F i g | < 0.5 , 0.5 ,
    A 1 g = B e s t V u l t u r e 1 g B e s t V u l t u r e 1 g × X i g B e s t V u l t u r e 1 g ( X i g ) 2 × F i g ,
    A 2 g = S e c o n d B e s t V u l t u r e 2 g S e c o n d B e s t V u l t u r e 2 g × X i g S e c o n d B e s t V u l t u r e 2 g ( X i g ) 2 × F i g ,
    L e v y d = 0.01 × μ × σ | ν | 1 β ,
    s i g m a = Γ ( 1 + β ) × sin ( Π β 2 ) Γ ( 1 + β 2 ) × β × 2 ( β 1 2 ) 1 β .
    where the next updated position of the vulture at the next ( g + 1 ) t h iteration, reflecting the assembly of vultures, is denoted by X i g + 1 . A 1 g and A 2 g are evaluated using, respectively, Equations (14) and (15). To increase the effectiveness of the AVO algorithm, L e v y d is the levy flight distribution function obtained using Equation (16). d is the dimensional space; σ is defined by Equation (17), where β = 1.5 is a constant value; and μ and ν are random numbers distributed equally within the range [ 0 , 1 ] .

3.4. Converting to Binary Nature

In the presented AVO algorithm, the positions are shown as real values. As such, they are not immediately applicable to the GS binary issue. Therefore, there is a need to convert these real position values into binary values to conform to GS’s binary nature and maintain the original algorithm’s structure. In this conversion, 1s represent the real values of the pertinent selected genes in the binarization vector, whereas 0s represent the real values of the unselected genes, which are not pertinent. At each iteration g, the following mathematical expression can be used to convert the real position X i g to a binary position ( X i g ) b i n :
( X i g ) b i n = 1 if X i g > δ , 0 otherwise .
where an arbitrary threshold point within [ 0 , 1 ] is represented by δ . According to this fundamental binary conversion approach, the binary “1” (selected feature) replaces its real value if ( X i g ) b i n is greater than δ . On the other hand, if ( X i g ) b i n is smaller than d e l t a , its real value is set to the binary “0” (a feature that was not chosen).

3.5. Appraising the Fitness Function Value

Finding the fewest number of selected features and optimizing the classification accuracy of the available classifiers (k-NN and SVM models) are two conflicting objectives that should be balanced to achieve the best solution and determine its quality. Since the k-NN and SVM classifiers’ accuracies might be hampered if the number of selected features is lower than the optimal, the fitness function balances the selected features’ size and accuracy. The fitness function concentrates on lowering the classification error rate rather than accuracy, as follows:
f i t = w 1 × E r r r a t e + w 2 × | f e a t p i c k e d | | D | , w 1 [ 0 , 1 ] , w 2 = 1 w 1 .
where f e a t p i c k e d denotes the number of selected features, D denotes the total number of features in the dataset, and E r r r a t e represents the classification error rate from the k-NN and SVM classifiers. The classification accuracy importance and the number of selected features are denoted by the weight parameters w 1 and w 2 , respectively. The values of w 1 ( 0.99 ) and w 2 ( 0.01 ) are determined through extensive trials conducted in previous studies [21,42,43].

3.6. Incorporating the DE Technique

DE [44] is characterized by its high efficiency and simplicity in finding a suitable solution for complex optimization problems. It can quickly produce value-added results. Three main processes—mutation, crossover, and selection—are necessary for DE. The differential mutation process seeks to produce a modified vector υ i for each iteration’s solution vector. This can be computed mathematically as follows:
υ i = X r 1 + W M ( X r 2 X r 3 )
Within the range [1, population size], three randomly selected asymmetric vectors are denoted by the symbols X r 1 , X r 2 , and X r 3 . W M denotes the mutation weighting factor within the interval [ 0 , 1 ] .
After the mutation operation, DE performs a crossover operation to increase the population’s variety. An offspring vector u i is produced by combining values from the target vector X i and the altered vector υ i . The most commonly used and basic factor of crossover search is binary crossover, which has the following mathematical expression:
u i , d = υ i , d , if r a n d C R o r d = j r a n d , X i , d , otherwise .
where a uniformly distributed random number j r a n d [ 1 , 2 , . . . , D X ] is used to ensure that the mutated vector has at least one dimension. The possibility of crossing each element is calculated using the crossover rate C R , frequently determined to be a big amount ( C R = 0.9).
The selection operation is then carried out, as shown in Equation (22). Here, a comparison between the fitness function f ( X i ) of the target vector and the corresponding offspring vector f ( u i ) is performed, and the lowest-valued fitness function is kept for the upcoming iteration.
X i = u i , if f ( u i ) < f ( X i ) X i , otherwise .

3.7. The Complete RBAVO-DE Algorithm

To handle the GS strategy, the steps of the recommended RBAVO-DE algorithm are described in the following subsections. The pseudo-code for the proposed RBAVO-DE algorithm is provided in Algorithm 1. A flowchart of the proposed RBAVO-DE algorithm is shown in Figure 1, illustrating its main steps.
Algorithm 1 The proposed RBAVO-DE algorithm.
Input:
  •    N—total number of positions (size of population)
  •     G m a x —maximum number of permitted iterations
  •    D—problem’s dimensional space
  •     L B —lower bounds of variables
  •     U B —upper bounds of variables
  •     C R —crossover rate
  •     W M —mutation weighting factor
Output:
  •     X B e s t —the global best vulture’s position found while searching
  •     f i t ( X B e s t ) —the global best fitness function value found, which should be lessened
 1:
Start
 2:
       Apply the Relief approach for selecting the related features and filtering them, as demonstrated in Section 3.1;
 3:
       Initialize a population of N positions, and provide the values of the necessary parameters ( L 1 , L 2 , w, P 1 , P 2 , and P 3 );
 4:
       Set a random position X in the initial population;
 5:
       Evaluate the fitness values f i t ( X ) for each position in the initial population;
 6:
       Arrange the positions in ascending order depending on their fitness function f i t ( X ) ;
 7:
        g 1 ;                                        ▹ Current number of iterations
 8:
       while  g < G m a x  do
 9:
             Assign the positions of both the first-best vulture X B e s t 1 g and the second-best vulture X S e c o n d B e s t 2 g , as well as their fitness values f i t ( X B e s t 1 g ) and f i t ( X S e c o n d B e s t 2 g ) , among all positions in the population;
10:
           for vulture’s position i = 1 : N  do
11:
                  Find the best vulture R g using Equation (2);
12:
                  Adjust the vulture’s famine level F i g using Equation (4);
13:
                  Modify the levy flight distribution function L e v y d using Equation (16);
14:
                  if  | F i g | 1  then
15:
                        if  r a n d P 1 P 1  then
16:
                              Amend the vulture’s position X i g + 1 based on the first stage of the exploration phase using Equation (7);
17:
                        else if  r a n d P 1 > P 1  then
18:
                              Upgrade the vulture’s position X i g + 1 based on the second stage of the exploration phase using Equation (7);
19:
                        end if
20:
                  else if  | F i g | < 1  then
21:
                        if  | F i g | 0.5  then
22:
                              if  r a n d P 2 P 2  then
23:
                                    Adjust the vulture’s position X i g + 1 based on the first condition of the first exploitation sub-phase using Equation (9);
24:
                              else if  r a n d P 2 > P 2  then
25:
                                    Amend the vulture’s position X i g + 1 based on the second condition of the first exploitation sub-phase using Equation (9);
26:
                              end if
27:
                     else if  | F i g | < 0.5  then
28:
                              if  r a n d P 3 P 3  then
29:
                                    Upgrade the vulture’s position X i g + 1 based on the first status of the second exploitation sub-phase using Equation (13);
30:
                              else if  r a n d P 3 > P 3  then
31:
                                   Adjust the vulture’s position X i g + 1 based on the second status of the second exploitation sub-phase using Equation (13);
32:
                              end if
33:
                          end if
34:
                      end if
35:
                       f i t ( X i g + 1 ) Compute the fitness function value for X i g + 1 ;
36:
                      if  f i t ( X i g + 1 ) < f i t ( X i g ) then
37:
                           X i g X i g + 1 ;
38:
                           f i t ( X i g ) f i t ( X i g + 1 );
39:
                      end if
40:
                  end for
41:
                  Re-arrange the positions in ascending order depending on their fitness function f i t ( X ) ;
42:
                  Detect the global best position X B e s t g + 1 and its global best fitness value f i t ( X B e s t g + 1 ) when the current iteration g + 1 is over;
43:
                  Perform the DE technique for every position to ameliorate X B e s t g + 1 , as shown in Section 3.6;
44:
                   X B e s t X B e s t g + 1 ;
45:
                   f i t ( X B e s t ) f i t ( X B e s t g + 1 ) ;
46:
                   g g + 1 ;
47:
      end while
48:
End

4. Experimental Results and Discussion

This section presents the empirical findings of the proposed RBAVO-DE and its counterparts: Binary Artificial Bee Colony (BABC) [45], Binary Salp Swarm Algorithm (BSSA) [46], Binary PSO (BPSO) [47], Binary Bat Algorithm (BBA) [48], Binary Grey-Wolf Optimization (BGWO) [49], Binary Grasshopper Optimization Algorithm (BGOA) [50], Binary Whale Optimization Algorithm (BWOA) [51], Binary ASO (BASO) [52], Binary Bird Swarm Algorithm (BBSA) [53], Binary HGSO (BHGSO) [54], and Binary Harris Hawks Optimization (BHHO) [55]. The optimizers undergo evaluation through training and testing benchmarks, with conclusive results derived from the average values of the evaluation metrics. The benchmarks employed for assessing the performance of the proposed model are detailed in Section 4.1. The parameters used in the operational environments are outlined in Section 4.2. The metrics used for evaluation are described in Section 4.3. The analysis of the experimental outcomes is discussed in Section 4.4.

4.1. Dataset Description

In-depth experimental approaches and various wrapper algorithms were applied to twenty-two datasets of gene descriptions; the data comprise normalized Level 3 RNA-Seq gene expression data for twenty-two kinds of tumors from the Broad Institute. These data are publicly accessible and can be found in [56]. We adhered to the methodology described in [36] and observed discrepancies between the data referenced from GitHub in the paper and the figures reported within the document, which were sourced from the website. The website’s data included a mix of tumor and standard samples, whereas the paper treated the data uniformly as tumor samples. Consequently, we conducted a detailed examination of the data. Initially, the website offered various formats of the identical dataset we intended to analyze. Upon delving into the data, we encountered the following challenges:
  • Certain genes are identified by ID but lack an associated symbol.
  • Some genes are absent from the annotation file.
  • There is a mix-up of samples, including both normal and tumor types.
Consequently, pre-processing was necessary to segregate and identify samples, distinguish between normal and tumor samples for use in binary classification, and streamline the GS process. We addressed the challenges above in the following manner:
  • We looked up the corresponding gene symbol for each ID found in the annotation file.
  • After cross-referencing with the annotating file, over one hundred genes were eliminated.
  • Based on the sample report, every Excel sheet’s row was organized according to the kind of sample for binary classification.
Moreover, the Relief approach, as detailed in Section 3.1, was utilized in the pre-processing stage to calculate the weight of each gene within the benchmark. These weights were subsequently ordered from highest to lowest. Genes with lower weights were then removed. The Relief approach was useful for discarding genes that do not contribute to classification.
Following the pre-processing phase, the data were refined and ready for the GS process. Contrary to the approach used in [36], which tackled the multi-classification of all types of cancer together, we opted to analyze every cancer type individually for greater specificity. Table 1 [57] presents a comprehensive overview of all twenty-two tumor types and their respective sample counts. The number of features utilized in the benchmark datasets is 32,000 features (genes).

4.2. Parameter Setting

The proposed RBAVO-DE algorithm was compared against binary variants of different meta-heuristic optimizers, such as BABC, BSSA, BPSO, BBA, BGWO, BWOA, BBSA, BGOA, BHHO, BASO, and BHGSO. The critical parameters for the ML models employed in this study are detailed in Table 2.
Within the proposed framework, the validity of the results was verified using a 10-fold cross-validation approach to ensure the reliability of the outcomes. This involved randomly dividing each dataset into two distinct subsets: 80% for training and the remaining 20% for testing. The training portion was employed to train the presented classifiers using optimization techniques, while the testing portion was used to evaluate the selected genes’ effectiveness. The parameters most commonly applied across all compared algorithms are summarized in Table 3. To execute all experiments in this paper, Python was utilized in a computational environment with a Dual Intel® Xeon® Gold 5115 2.4 GHz CPU and 128 GB of RAM on the Microsoft Windows Server 2019 operating system.

4.3. Evaluation Criteria

To evaluate the proposed RBAVO-DE’s effectiveness relative to other methods, each strategy was independently tested thirty times on each benchmark to ensure statistical validation of the results. For this purpose, the following established performance metrics for the GS issues were employed:
  • Mean classification accuracy ( M e a n A C ): This measure represents the accuracy of correctly classifying data, calculated by running the algorithm independently thirty times. It is determined in the following way:
    M e a n A C = k = 1 30 r = 1 m m a t c h ( P L r , A L r )
    In this formula, m denotes the number of samples in the test dataset, while P L r and A L r denote the predicted labels from the classifier and the actual class labels for sample r, respectively. The function m a t c h ( P L r , A L r ) serves as a comparison mechanism, where if P L r is equal to A L r , then m a t c h ( P L r , A L r ) is assigned a value of 1; if they do not match, it is assigned a value of 0.
  • Mean fitness  ( M e a n F i ) : This measure calculates the average m e a n F i , achieved by running the proposed approach thirty times independently. It gauges the balance between lowering the classification error rate and minimizing the number of genes selected. A lower value indicates a superior solution, assessed based on fitness:
    M e a n F i = 1 30 k = 1 30 f * k
    where f * k shows the optimum fitness value achieved in the k t h execution.
  • Mean number of chosen genes  ( M e a n G e ) : This metric calculates the average number of genes chosen (or GS ratio) by running the proposed approach thirty times independently, and it is defined as follows:
    M e a n F e = 1 30 k = 1 30 | d * k | | D |
    Here, | d * k | denotes the total count of genes chosen in the optimal solution for the k t h execution, and | D | represents the total number of genes present in the initial benchmark.
  • Standard Deviation  ( S T D E ) : Based on the outcomes above, the overall average results derived from thirty separate executions of each optimizer on every benchmark were assessed for stability in the following manner:
    S T D E = 1 29 k = 1 30 ( Y k M e a n Y ) 2
    where Y shows the measure to be utilized, Y k denotes the measure value Y in the k t h run, and M e a n Y is the mean of the measure from thirty independent executions.
The data in the subsequent tables represent the mean values obtained from thirty independent executions, focusing on classification accuracy, the number of chosen genes, average fitness, precision, recall, and F1-score. The ensuing subsections thoroughly examine and discuss these experimental findings, with bold figures highlighting the optimal results.

4.4. Results of Comparing the Proposed RBAVO-DE with Various Meta-Heuristic Algorithms

The proposed RBAVO-DE approach with the SVM and k-NN models was compared with other recent meta-heuristic methods executed under the same conditions to show its superiority over its counterparts. The proposed RBAVO-DE algorithm was compared with binary variants of several optimizers, including BABC, BSSA, BPSO, BBA, BGWO, BWOA, BBSA, BGOA, BHHO, BASO, and BHGSO.

4.4.1. Results Employing the k-NN Model

Table 4 shows the results of the proposed RBAVO-DE algorithm and other optimization techniques employing the k-NN model, based on accuracy metrics compared under similar conditions. The experimental outcomes reveal that the proposed RBAVO-DE achieved the most promising outcomes in four benchmarks. It is noteworthy that all competitive methods, including the proposed RBAVO-DE employing the k-NN model, produced comparable outcomes across eighteen benchmarks.
The rest of the metrics for the k-NN model are shown in Appendix A.1. In terms of the average fitness values, the proposed RBAVO-DE algorithm demonstrated higher efficiency compared to its counterparts using the k-NN model under similar conditions. The RBAVO-DE algorithm yielded the lowest fitness results and the most competitive STDE across all benchmarks. Also, all utilized benchmarks are high-dimensional, demonstrating that the proposed RBAVO-DE can run effectively on all benchmarks regardless of size. The RBAVO-DE algorithm shows promise by effectively balancing exploitation and exploration of the search space, avoiding becoming trapped in local optima during iterations. Unlike many other algorithms that tend to become trapped, it demonstrated the ability to escape such traps.
Regarding the mean results for the genes selected by the proposed RBAVO-DE algorithm and its counterparts employing k-NN, the proposed RBAVO-DE algorithm surpassed the other methods across all benchmarks concerning the number of chosen genes. Furthermore, the RBAVO-DE’s capacity to identify significant genes is due to its effective exploration of possible areas while enhancing accuracy.
In terms of the mean precision of the proposed RBAVO-DE algorithm and its counterparts with k-NN, the proposed RBAVO-DE outperformed alternative approaches for three of the twenty-two datasets. For nineteen datasets, BWOA achieved similar results, while BABC, BPSO, BGWO, BGOA, and BHHO obtained results similar to those of the proposed RBAVO-DE. BASO and BHGSO, which produced the same outcomes as the proposed RBAVO-DE on seventeen datasets, ranked fourth. Ultimately, BBA ranked lowest among all approaches, producing similar outcomes to the proposed RBAVO-DE for fifteen datasets.
Regarding the mean recall of the proposed RBAVO-DE and its counterparts employing k-NN, RBAVO-DE outperformed the alternative approaches for three of the twenty-two datasets. On the other hand, BSSA, BABC, BPSO, BGWO, BWOA, BGOA, and BBSA produced the same outcomes as RBAVO-DE for 19 datasets, while BHHO achieved similar results for 18 datasets. BHGSO, which obtained the same outcomes as the proposed RBAVO-DE for seventeen datasets, ranked fourth. Ultimately, BASO and BBA ranked lowest among all approaches, producing similar results to RBAVO-DE for sixteen datasets.
With regard to the mean F1-score of the proposed RBAVO-DE and its counterparts employing k-NN, the proposed RBAVO-DE outperformed the alternative approaches for three of the twenty-two datasets. In contrast, for eighteen datasets, BSSA, BABC, BPSO, BGWO, BGOA, and BBSA produced the same outcomes as RBAVO-DE, while BWOA and BHHO produced comparable outcomes for seventeen datasets. BASO and BHGSO, which obtained the same outcomes for 16 datasets as the proposed RBAVO-DE, ranked fourth. Ultimately, BBA ranked lowest among all approaches, producing similar outcomes to the proposed RBAVO-DE for fourteen datasets.

4.4.2. Results Employing the SVM Model

Table 5 displays the outcomes of the proposed RBAVO-DE and various optimization techniques employing the SVM model concerning classification accuracy results assessed under identical running conditions. The experimental outcomes demonstrate that the proposed RBAVO-DE algorithm outperformed the other approaches by obtaining the most promising values for four datasets. All competitive methods, including the proposed RBAVO-DE employing SVM, obtained equivalent values across 18 benchmarks.
The rest of the metrics utilizing SVM are presented in Appendix A.2. Regarding the mean fitness values of the proposed RBAVO-DE and its counterparts employing the SVM model under equivalent running conditions, the proposed RBAVO-DE proved to be more efficient than the other methods. The proposed RBAVO-DE employing the SVM model yielded the lowest fitness values and the most competitive STDE across all benchmarks. Also, all utilized benchmarks are high-dimensional, demonstrating that the proposed RBAVO-DE can run effectively on all benchmarks regardless of size. The RBAVO-DE algorithm shows promise by effectively balancing exploitation and exploration of the search space, avoiding becoming trapped in local optima during iterations. Unlike many other algorithms that tend to become trapped, it demonstrated the ability to escape such traps.
Regarding the mean results for the genes selected by the proposed RBAVO-DE and its counterparts employing SVM, the proposed RBAVO-DE achieved more promising results than other techniques across all benchmarks utilized in this study. Also, the superiority of the proposed RBAVO-DE employing SVM in this context demonstrates its capability to effectively explore valuable regions of the search space while avoiding regions with non-feasible solutions.
With regard to the average precision values of the proposed RBAVO-DE and its counterparts employing SVM, the proposed RBAVO-DE outperformed the alternative approaches for three of the twenty-two datasets. On the other hand, for eighteen datasets, BABC, BGWO, BWOA, BBSA, and BASO yielded comparable results. For seventeen datasets, BSSA, BPSO, BGOA, and BHHO yielded results similar to those of RBAVO-DE, thereby ranking fourth. Ultimately, BBA ranked lowest among all approaches, producing similar outcomes to those of the proposed RBAVO-DE for twelve datasets.
In terms of the mean recall of the proposed RBAVO-DE method and its counterparts using SVM, the proposed RBAVO-DE outperformed the alternative approaches for three of the twenty-two datasets. In contrast, BPSO and BHHO performed similarly for eighteen datasets, while BSSA, BABC, BGWO, BWOA, BGOA, and BBSA produced the same results as the proposed RBAVO-DE for nineteen datasets. BHGSO, which obtained the same outcomes as RBAVO-DE for seventeen datasets, ranked fourth. In the end, BASO and BBA ranked lowest among all approaches, producing outcomes similar to those of RBAVO-DE for sixteen datasets.
Regarding the mean F1-score of the proposed RBAVO-DE algorithm and its counterparts employing SVM, the proposed RBAVO-DE outperformed the alternative approaches for five of the twenty-two datasets. On the other hand, BWOA produced outcomes similar to those of the proposed RBAVO-DE for seventeen datasets, while BSSA, BABC, BPSO, BGWO, BGOA, and BBSA produced comparable results for sixteen datasets. BHHO, which obtained the same outcomes as the proposed RBAVO-DE for fifteen datasets, ranked fourth. Ultimately, BBA ranked lowest among all approaches, producing similar outcomes to those of RBAVO-DE for 12 datasets.

4.5. Convergence Analysis

It is evident in Appendix B [57] that the proposed RBAVO-DE employing the SVM and k-NN models attained optimum convergence behavior across all benchmarks. Therefore, the convergence performance of the proposed RBAVO-DE employing the SVM and k-NN models demonstrates its capability to reach optimal outcomes promptly while maintaining an efficient equilibrium between search and exploitation.

4.6. Wilcoxon’s Rank-Sum Test

This paper compares the fitness values achieved by the proposed RBAVO-DE and its counterparts pairwise using the Wilcoxon rank-sum test [60], which aims to determine whether there is a statistically significant difference between the different approaches. This test is a crucial tool for evaluating the success of the proposed algorithm. The Wilcoxon test is employed in hypothesis testing to compare matched data. It involves sorting the absolute differences in the outcomes of paired procedures on the j t h of N issues. Next, the lowest value is found by summing the positive ( R + ) and negative ( R ) rankings. The null hypothesis is rejected if the resultant significance level is less than 5%, and not rejected otherwise.
From examining the results of the comparison made between the RBAVO-DE and its counterparts pairwise using the Wilcoxon test, we found that all R + values equal 253, all R values equal 0, and all p-values equal 4.768 × 10 7 which are less than 0.05 (significance level of 5%). Hence, it can be concluded that the proposed RBAVO-DE employing SVM and k-NN performs better than any other algorithm in all cases. Therefore, all p-values less than 0.05 (significance level of 5%) offer compelling evidence that the outcomes of the proposed strategy are statistically significant and not just coincidental.

4.7. Computational Complexity of the RBAVO-DE Algorithm and Various Meta-Heuristic Algorithms

4.7.1. Computational Complexity of the Execution Time of the RBAVO-DE Algorithm

Each of RBAVO-DE’s five core steps can be analyzed separately to determine its computational complexity. These steps include filtering features, initializing the population, boosting and amending the position, appraising the fitness function, and using DE. Then, O e x e c u t e _ t i m e ( R B A V O D E ) can be utilized to represent the overall computational complexity of the proposed RBAVO-DE algorithm. This can be computed using the following big-O notation formulas:
O e x e c u t e _ t i m e ( R B A V O D E ) = O e x e c u t e _ t i m e ( Filtering   features ) + O e x e c u t e _ t i m e ( Initializing   population ) + O e x e c u t e _ t i m e ( Boosting   and   amending   position ) + O e x e c u t e _ t i m e ( Appraising   fitness   function ) + O e x e c u t e _ t i m e ( Incorporating   DE   technique ) .
Since N determines the population size, G m a x is the maximum number of iterations allowed, and D is the problem’s dimensional space. Hence,
  • O e x e c u t e _ t i m e ( Filtering   features ) = O e x e c u t e _ t i m e ( D ) .
  • O e x e c u t e _ t i m e ( Initializing   population ) = O e x e c u t e _ t i m e ( N ) .
  • O e x e c u t e _ t i m e ( Boosting   and   amending   position ) = O e x e c u t e _ t i m e ( G m a x × N × D ) .
  • O e x e c u t e _ t i m e ( Appraising   fitness   function ) = O e x e c u t e _ t i m e ( G m a x × N ) .
  • O e x e c u t e _ t i m e ( Incorporating   DE   technique ) = O e x e c u t e _ t i m e ( N × D ) .
For that,
O e x e c u t e _ t i m e ( R B A V O D E ) = O e x e c u t e _ t i m e ( D ) +                                                               O e x e c u t e _ t i m e ( N ) + O e x e c u t e _ t i m e ( G m a x × N × D ) +                                                               O e x e c u t e _ t i m e ( G m a x × N ) + O e x e c u t e _ t i m e ( N × D ) =                                                                                                                                                                                                                                                         O e x e c u t e _ t i m e ( G m a x × N × D ) .

4.7.2. Computational Complexity of the Memory Usage of the RBAVO-DE Algorithm

This involves measuring the amount of memory required by an algorithm to tackle a problem as the quantity of the input increases. It is frequently expressed as the additional memory needed by the algorithm beyond the input. It entails merging the following two primary elements:
  • Memory usage of input variables: This is the amount of memory required for the algorithm to store the input data. There are 13 input variables related to the proposed RBAVO-DE algorithm, as follows: N, G m a x , D, L B , U B , C R , W M , L 1 , L 2 , w, P 1 , P 2 , and P 3 . Since each variable stores numerical values, 4 bytes of memory are utilized by each one. Consequently, the total memory usage complexity of these 13 input variables is 52 bytes ( 13 × 4 bytes = 52 bytes). The memory usage complexity of the input values is constant.
  • Additional memory usage: This shows how much more memory the algorithm needs in addition to the input. It includes the memory required for data structures, internal variables, and other components of the algorithm. Regardless of the size of the input, the RBAVO-DE algorithm requires a specific amount of extra memory. The following variables are involved:
    • The memory usage complexity consumed by X i n i t i a l is ( 4 × N × D ) bytes. This is because each position in the positions vector X i n i t i a l requires 4 bytes of memory, and its size is ( N × D ) , proportionate to the initial population of N positions with dimension size D. Because the amount of memory required rises linearly with the value ( N × D ) , its memory use complexity is linear.
    • The variables p r , g, F i , t, R, D i , d t , S 1 , S 2 , A 1 , A 2 , s i g m a , f i t ( X i ) , f i t ( X B e s t 1 ) , f i t ( X S e c o n d B e s t 2 ) , and f i t ( X B e s t ) require 4 bytes of memory space each since they only represent numerical values. As a result, the total memory usage complexity of these 16 variables is 64 bytes ( 16 × 4 bytes = 64 bytes). This memory usage complexity is constant.
    • The position vectors B e s t V u l t u r e 1 , S e c o n d B e s t V u l t u r e 2 , X i , L e v y d , μ , σ , X i b i n , υ i , X r 1 , X r 2 , X r 3 , u i , and X B e s t each require 4 bytes of memory space, and the size of each vector is D, proportionate to the dimension size of the acquired positions. As a result, the vectors at these 13 points have a total memory usage complexity of ( 52 × D ) bytes ( 13 × 4 × D bytes). Since the memory needed grows linearly with the value D, its memory consumption complexity is linear.
    As a result, the overall memory usage complexity for all of the additional variables listed above is ( 4 × N × D ) + 64 + ( 52 × D ) bytes.
Lastly, the following formula can be used to determine the computational complexity of the overall memory usage of the proposed RBAVO-DE algorithm:
Memory   usage   complexity ( RBAVO DE ) = Input   values   memory   usage +                                           Additional   memory   usage = 52 + ( 4 × N × D ) + 64 + ( 52 × D )   bytes .
Keep in mind that there are constant bytes that are not taken into account. In big-O notation, the computational complexity of the total memory consumption of RBAVO-DE can be represented as O m e m o r y _ u s a g e ( R B A V O D E ) . This can then be calculated in big-O notation after eliminating all constants in the following way:
O m e m o r y _ u s a g e ( R B A V O D E ) = O m e m o r y _ u s a g e ( Input   values   memory   usage ) +         O m e m o r y _ u s a g e ( Additional   memory   usage ) =     O m e m o r y _ u s a g e ( 1 ) + O m e m o r y _ u s a g e ( N × D ) + O m e m o r y _ u s a g e ( 1 ) + O m e m o r y _ u s a g e ( D ) =     O m e m o r y _ u s a g e ( N × D ) .
It can be difficult to compile a thorough comparison of the memory usage and execution time complexity of various meta-heuristic algorithms since these complexities might differ based on the particular implementation, size of the problem, and other operators. Furthermore, not all of the algorithms listed have comprehensive evaluations of the execution time and memory usage complexity available, and their properties may vary depending on the problem at hand.

5. Benefits and Drawbacks of the RBAVO-DE Algorithm

This part presents a balanced discussion of the benefits and drawbacks of the RBAVO-DE algorithm. The benefits of the RBAVO-DE algorithm can be listed as follows:
  • High Accuracy in Classification: The RBAVO-DE algorithm has demonstrated high classification accuracy, achieving up to 100% in some cases. This is a noteworthy achievement, particularly regarding cancer datasets, as accurate gene selection can directly impact diagnostic and therapeutic outcomes.
  • Effective Feature Size Reduction: The algorithm has shown remarkable capability in reducing the feature size by up to 98% while maintaining or improving classification accuracy. This dimensionality reduction is crucial for processing high-dimensional datasets efficiently.
  • Robustness Across Diverse Datasets: The effectiveness of the RBAVO-DE algorithm across twenty-two cancer datasets indicates its robustness and adaptability to various genetic data characteristics. This versatility is beneficial for broader applications in genomics research.
  • Superior Performance Over Competitors: When compared with binary variants of widely recognized meta-heuristic algorithms, RBAVO-DE stands out for its outstanding performance in accuracy and feature reduction, highlighting its innovative approach to gene selection.
On the other hand, the drawbacks of the proposed methodology can be listed as follows:
  • Computational Complexity: The high dimensionality of the datasets and the iterative nature of meta-heuristic algorithms suggest that RBAVO-DE may have a significant computational cost. This aspect could limit its applicability in environments with constrained computational resources.
  • Generalizability Concerns: Despite its success across various cancer datasets, the algorithm’s generalizability to other types of biological data or diseases remains to be thoroughly investigated. It is crucial to test RBAVO-DE in broader contexts to confirm its applicability beyond the datasets examined.
  • Parameter Sensitivity: Like many meta-heuristic algorithms, RBAVO-DE might be sensitive to its parameter settings, affecting performance and efficiency. Detailed studies on the impact of different parameter configurations and strategies for their optimization could enhance the algorithm’s utility.
In summary, the RBAVO-DE algorithm represents a significant advancement in gene selection for cancer classification, with notable strengths in accuracy and efficiency. However, addressing its potential drawbacks through further research could broaden its applicability and improve its performance, making it an even more valuable tool in genomics and personalized medicine.

6. Conclusions and Future Directions

The RBAVO-DE approach proposed in this paper was the first to be implemented for handling GS issues in RNA-Seq gene expression data and determining potential biomarkers for different tumor classes to enhance the best solution discovered. The outcomes were promising, revealing that the proposed RBAVO-DE algorithm’s effectiveness and capabilities were significantly improved. SVM and k-NN, two widely used classification models, were employed to evaluate the efficacy of each set of selected genes. The proposed RBAVO-DE algorithm’s capability was compared to binary variants of eleven widely used meta-heuristic techniques to assess it on different tumor classes with various instances. The assessment was executed employing a combination of evaluation measures, such as average fitness, classification accuracy, the number of selected genes, precision, recall, and F1-score. The proposed RBAVO-DE algorithm using the SVM and k-NN classifiers achieved more promising results than other optimizers in handling GS issues. Despite the promising outcomes, this research opens several avenues for future exploration to further advance the field:
  • Algorithmic Enhancements: Improving the algorithm’s efficiency and reducing computational complexity.
  • Cross-disease Applicability: Testing the proposed algorithm on genetic data from various diseases beyond cancer.
  • Comparative Analyses with Deep Learning Models: Evaluating RBAVO-DE against advanced deep learning models for genetic data analysis.
  • Real-world Clinical Validation: Collaborating with clinical experts to validate the practical utility of selected genes in cancer treatment.
  • Scalability and Parallelization: Enhancing the algorithm’s scalability through parallel computing to handle larger genetic datasets efficiently.
  • Interdisciplinary Applications: Exploring the algorithm’s potential in other fields dealing with high-dimensional data, such as finance and environmental modeling.
In conclusion, while the RBAVO-DE algorithm represents a significant step forward in the field of gene selection for cancer classification, the paths outlined for future research highlight the potential for further advancements and broader applications of this work. Continued interdisciplinary collaboration and innovation will be crucial for unlocking the full capacity of gene selection methodologies to enhance healthcare outcomes and advance our understanding of complex diseases.

Author Contributions

M.G.G.: Resources, Validation, and Writing—review and editing. A.A.A.: Resources, Formal analysis, Data Curation, Validation, Writing—original draft, and Writing—review and editing. A.E.E.: Investigation, Visualization, Data Curation, Validation, Writing—original draft, and Writing—review and editing. A.A.A.E.-M.: Conceptualization, Methodology, Software, Formal analysis, Investigation, Data Curation, Validation, Writing—original draft, and Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

Prince Sattam bin Abdulaziz University funded this research work under project number PSAU/2023/01/25612.

Data Availability Statement

For enhancing the applicability and reproducibility of this research, the developed software and relevant Python code of this research are publicly available and obtainable in [61].

Acknowledgments

The authors express their gratitude to Prince Sattam bin Abdulaziz University for providing financial support for this research project.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Abbreviations

The following abbreviations are used in this research:
AVOAfrican Vultures Optimization
HNSCHead and Neck Squamous Cell Carcinoma
DEDifferential Evolution
KICHKidney Chromophobe
RBAVO-DERelief Binary African Vultures Optimization based on Differential Evolution
KIRPKidney Renal Papillary
GSGene Selection
LIHCLiver Hepatocellular Carcinoma
RNA-SeqRNA Sequencing
PAADPancreatic Adenocarcinoma
DNADeoxyribonucleic Acid
PCPGPheochromocytoma and Paraganglioma
SVMSupport Vector Machine
READRectum adenocarcinoma
k-NNk-Nearest Neighbor
SARCSarcoma
FSFeature Selection
SKCMSkin Cutaneous Melanoma
MHMsMeta-heuristic methods
M e a n F i Mean fitness
MLMachine Learning
M e a n G e Mean size of chosen genes
DLDeep learning
S T D E Standard Deviation
SASimulated Annealing
RFRandom Forest
FOILFirst-Order Inductive Learner
GSOGravitational Optimizer
SCCSASine-Cosine–Cuckoo Search Algorithm
SMOSpider Monkey Optimization
BHGSOBinary Henry Gas Solubility Optimization
BPSOBinary Particle Swarm Optimization
BABCBinary Artificial Bee Colony
BSSABinary Salp Swarm Algorithm
BBABinary Bat Algorithm
BGWOBinary Grey-Wolf Optimization
BGOABinary Grasshopper Optimization Algorithm
BWOABinary Whale Optimization Algorithm
BASOBinary Atom Search Optimization
BBSABinary Bird Swarm Algorithm
BHHOBinary Harris Hawks Optimization
BBBOBinary Brown-Bear Optimization
BAOBinary Aquila Optimization
BMOABinary Meerkat Optimization Algorithm
LUADLung Adenocarcinoma
LUSCLung Squamous Cell Carcinoma
BRCABreast Invasive Carcinoma
KIRCKidney Renal Clear Cell Carcinoma
UCECUterine Corpus Endometrial Carcinoma
RPKMReads Per Kilobase per Million mapped reads
STADStomach Adenocarcinoma
SDAEStacked Denoising Autoencoder
CGACancer Genome Atlas
BLCABladder Urothelial Carcinoma
THCAThyroid Cancer
CESCCervical and Endocervical Cancers
CHOLCholangiocarcinoma
COADColon Adenocarcinoma
ESCAEsophageal Cancer
GBMGlioblastoma Multiforme
THYMThymoma

Appendix A. Comparison Results of the Proposed RBAVO-DE and Its Counterparts

Appendix A.1. Comparison Results Based on the k-NN Model

Table A1. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average fitness values.
Table A1. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average fitness values.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage0.001900.011100.007100.011200.014700.009400.010700.011400.013100.010300.010800.01590
STDE0.004300.005300.004400.005400.002400.005500.005400.005400.004400.005500.005400.00180
CESCAverage0.016100.019900.020200.020500.019800.020000.020200.020200.019700.020200.020300.02060
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
CHOLAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
COADAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
ESCAAverage0.025600.029300.029600.029900.029300.029400.029700.029600.029100.029600.029800.03000
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00020
GBMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
HNSCAverage0.007100.012500.012900.013300.012800.012400.013100.013200.012700.012900.013200.01360
STDE0.003500.001600.001400.000100.000300.002100.000100.000100.000200.001500.000100.00010
KICHAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
KIRCAverage0.002500.011100.010400.012200.012000.010200.012000.010700.010400.011200.011700.01290
STDE0.003800.002700.003400.001900.000200.003400.001900.003200.003000.002800.002200.00010
KIRPAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
LIHCAverage0.000000.003900.004300.004500.003900.004000.004200.004300.003800.004300.004200.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
LUADAverage0.008700.012600.012900.013100.012400.012700.012900.012900.012400.012900.013000.01340
STDE0.000000.000200.000100.000100.000200.000200.000100.000100.000200.000100.000100.00010
LUSCAverage0.000000.004000.004300.004500.003800.004100.004300.004300.003900.004300.004300.00480
STDE0.000000.000200.000100.000100.000200.000100.000200.000100.000100.000100.000100.00010
PAADAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
PCPGAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
READAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
SARCAverage0.018900.022600.022900.023200.022600.022700.023000.022900.022400.022900.023000.02330
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
SKCMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
STADAverage0.000700.008300.006200.009300.014700.008100.009500.008000.011700.009400.008400.01570
STDE0.002800.005000.003600.005100.000200.005000.005100.004800.004800.005200.004900.00010
THCAAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
THYMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
UCECAverage0.000000.019200.017100.027800.035400.020100.027700.026900.025600.025200.028600.03670
STDE0.002500.012000.012200.006100.010000.011900.006200.007200.008300.009100.004500.01050
Ranking W | T | L 22 | 0 | 0 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22
Boldface values denote the best results.
Table A2. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average values of selected genes.
Table A2. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average values of selected genes.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage142.03210.17237.70235.00196.03218.17228.50226.30195.23228.27232.03237.13
STDE017.18020.50019.07017.12023.83018.27019.43014.81013.17014.60018.90014.14
CESCAverage122.50195.13213.03225.30194.10199.50213.63212.67185.60213.33215.90233.43
STDE005.45011.17005.31003.11009.66009.30008.35005.66006.49005.88006.02005.64
CHOLAverage122.13195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.18011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
COADAverage120.30195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE005.82011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
ESCAAverage120.17195.13213.03225.30194.10199.50213.63212.67185.60211.57218.87230.07
STDE006.81011.17005.31003.11009.66009.30008.35005.66006.49006.94004.68007.64
GBMAverage120.73195.13212.77225.30194.10199.73213.17212.80185.43211.00211.83232.63
STDE006.54011.17005.40003.11009.66009.05008.51005.72006.40006.44004.99004.39
HNSCAverage133.43202.87222.67227.43203.00211.37217.97220.87195.93219.83223.63241.27
STDE017.63009.26011.96003.91013.24007.65007.41005.32008.26007.71005.68006.21
KICHAverage121.43195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.58011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
KIRCAverage137.57200.83220.17227.70192.17209.87216.13222.30193.27218.73218.60234.10
STDE017.10009.63011.74008.27008.73014.08010.11015.18017.36013.82012.78004.21
KIRPAverage122.33195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE009.08011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
LIHCAverage123.40197.00214.03225.67193.33200.63212.33213.63189.30213.10212.03232.43
STDE006.59007.78005.22003.36009.17010.02009.35004.89006.89005.69004.81004.97
LUADAverage127.03197.47216.13222.93190.83205.33212.57215.67191.80215.27217.83237.90
STDE008.62008.98004.74006.16008.61010.57006.15004.43008.50006.13004.71004.94
LUSCAverage125.73198.87215.63223.00189.13203.60213.23216.37194.67215.53215.23241.23
STDE006.35008.77003.54004.89010.69007.20009.70003.44006.40005.54006.16004.66
PAADAverage121.10195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.71011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
PCPGAverage122.17195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.42011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
READAverage120.13195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.72011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
SARCAverage122.47195.13213.03225.30194.10199.50213.63212.67185.60212.90216.43231.80
STDE005.79011.17005.31003.11009.66009.30008.35005.66006.49006.30004.98003.40
SKCMAverage123.03195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE008.83011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
STADAverage142.60215.00236.17243.33187.50221.17234.87232.77201.73231.13237.10236.97
STDE013.31018.43013.94016.29010.27019.24022.63014.25014.00014.04015.35005.86
THCAAverage117.97195.43213.40225.13194.10199.63214.13212.80186.53211.00211.57231.90
STDE006.67011.20005.29003.10009.66009.81008.59005.69005.88006.44005.08004.78
THYMAverage120.37195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE004.89011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
UCECAverage143.93216.33237.87236.00242.37220.07230.80229.17206.77228.40235.00266.67
STDE013.18015.41015.82010.11034.13011.26011.27011.66010.84010.32006.14030.55
Ranking W | T | L 22 | 0 | 0 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22
Boldface values denote the best results.
Table A3. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average precision values.
Table A3. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average precision values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000000.999601.000001.000001.000001.000001.000001.000001.00000
CESC0.983900.983900.983900.983900.983900.983900.983900.983900.983900.983900.983900.98390
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000000.998901.000001.000001.000001.000001.000001.000001.00000
ESCA0.973700.973700.973700.973700.973700.973700.973700.973700.973700.973700.973700.97370
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC0.991800.964700.965400.965000.963500.966900.966600.965700.965400.965400.961000.96290
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD1.000000.997700.997400.998400.998700.997700.998000.995400.997100.997400.996700.99800
LUSC1.000001.000001.000001.000000.997901.000000.999701.000001.000001.000000.979000.99930
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC0.981100.981100.981100.981100.981100.981100.981100.981100.981100.981100.981100.98110
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD0.999200.951800.951800.951800.948700.951800.951800.951800.951800.951800.990000.95090
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000001.000001.000001.000000.999101.000001.000001.000001.000001.000000.933300.99910
Ranking ( W | T | L ) 3 | 19 | 0 0 | 19 | 3 0 | 19 | 3 0 | 19 | 3 0 | 15 | 7 0 | 19 | 3 0 | 18 | 4 0 | 19 | 3 0 | 19 | 3 0 | 19 | 3 0 | 17 | 5 0 | 17 | 5
Boldface values denote the best results.
Table A4. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average recall values.
Table A4. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average recall values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CESC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
ESCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC1.000000.995200.994500.994800.995900.992800.993100.994200.994500.994500.997900.99730
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC0.997100.990400.990400.990400.989700.990400.990400.990400.990400.990400.990100.99040
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD1.000000.992400.992700.991700.991400.992400.992100.994700.993100.992700.993400.99210
LUSC1.000001.000001.000001.000000.995801.000001.000001.000000.999301.000000.993100.99270
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD1.000001.000001.000001.000000.997501.000001.000001.000001.000001.000000.996200.99750
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000001.000001.000001.000000.993701.000001.000001.000001.000001.000000.996401.00000
Ranking ( W | T | L ) 3 | 19 | 0 0 | 19 | 3 0 | 19 | 3 0 | 19 | 3 0 | 16 | 6 0 | 19 | 3 0 | 19 | 3 0 | 19 | 3 0 | 18 | 4 0 | 19 | 3 0 | 16 | 6 0 | 17 | 5
Boldface values denote the best results.
Table A5. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average F1-score values.
Table A5. The results of the proposed RBAVO-DE algorithm and its counterparts employing k-NN regarding average F1-score values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000000.999801.000001.000001.000001.000001.000001.000001.00000
CESC0.991900.991900.991900.991900.991900.991900.991900.991900.991900.991900.991900.99190
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000000.999401.000001.000001.000001.000001.000001.000001.00000
ESCA0.986700.986700.986700.986700.986700.986700.986700.986700.986700.986700.986700.98670
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC0.995900.979700.979700.979700.979400.979700.979700.979700.979700.979700.979800.97970
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC0.998600.995200.995200.995200.994800.995200.995200.995200.995200.995200.995000.99520
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD0.995000.995000.995000.995000.995000.995000.995000.995000.995000.995000.995000.99500
LUSC1.000001.000001.000001.000000.996901.000000.999801.000000.999701.000000.996300.99600
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC0.990500.990500.990500.990500.990500.990500.990500.990500.990500.990500.990500.99050
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD0.999600.975300.975300.975300.972400.975300.975300.975300.975300.975300.971800.97360
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000001.000001.000001.000000.996401.000001.000001.000001.000001.000000.996400.99960
Ranking ( W | T | L ) 3 | 19 | 0 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 14 | 8 0 | 18 | 4 0 | 17 | 5 0 | 18 | 4 0 | 17 | 5 0 | 18 | 4 0 | 16 | 6 0 | 16 | 6
Boldface values denote the best results.

Appendix A.2. Comparison Results Based on the SVM Model

Table A6. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average fitness values.
Table A6. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average fitness values.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
CESCAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
CHOLAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
COADAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
ESCAAverage0.013700.028400.027200.029900.029300.026900.029700.028800.028300.028800.029800.03000
STDE0.012800.004600.007500.000100.000200.007500.000200.004500.004600.004400.000100.00020
GBMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
HNSCAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
KICHAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
KIRCAverage0.000000.003900.004300.004500.003800.004000.004300.004300.003800.004300.004300.00460
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00020
KIRPAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
LIHCAverage0.000000.004100.004400.004500.005000.004300.004400.004500.004000.004500.004500.00500
STDE0.000000.000100.000100.000100.002800.000100.000200.000100.000200.000100.000100.00020
LUADAverage0.008400.012500.012900.013100.012500.012600.012900.012900.012300.012900.012600.01330
STDE0.001600.000200.000100.000100.000200.000200.000200.000100.000100.000100.001500.00010
LUSCAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
PAADAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
PCPGAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
READAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
SARCAverage0.005700.017800.015700.018400.022600.014800.017600.018700.020000.018700.020000.02330
STDE0.008600.008000.008800.008000.000200.009000.008200.007700.006100.007700.006700.00010
SKCMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
STADAverage0.010400.014900.015300.015500.014500.015000.015300.015300.014700.015200.015300.01560
STDE0.002800.000200.000100.000100.002000.000200.000200.000100.000100.000100.000100.00010
THCAAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
THYMAverage0.000000.003900.004300.004500.003900.004000.004300.004300.003700.004200.004200.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000100.000100.000100.00010
UCECAverage0.000000.003900.004300.004500.003800.004000.004300.004300.003800.004200.004300.00470
STDE0.000000.000200.000100.000100.000200.000200.000200.000100.000200.000100.000100.00010
Ranking W | T | L 22 | 0 | 0 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22
Boldface values denote the best results.
Table A7. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average chosen gene values.
Table A7. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average chosen gene values.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage121.37194.70213.03225.30194.10199.50213.40212.67186.27211.00211.83232.37
STDE007.49011.44005.18003.11009.66009.30008.44005.66005.77006.44005.25004.42
CESCAverage119.83195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.02011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
CHOLAverage125.57195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.14011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
COADAverage122.63195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.73011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
ESCAAverage125.50195.13215.40225.30194.10201.77213.63213.13185.70212.40218.87230.07
STDE011.18011.17009.58003.11009.66012.33008.35006.44006.47008.98004.68007.64
GBMAverage119.73194.93213.03225.30193.90198.77213.70212.83185.50211.00211.00232.33
STDE008.07011.27005.31003.11009.47010.28008.36005.69006.51006.44004.92004.15
HNSCAverage122.17196.40213.90225.13194.03202.10213.83213.50186.93211.47210.87232.17
STDE008.29008.49005.22003.19009.56010.76009.34005.22005.74005.66004.79005.02
KICHAverage121.73195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE008.49011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
KIRCAverage123.87197.10214.70225.00191.93200.30212.97213.13187.57213.83214.63231.27
STDE007.57009.42005.09004.23008.70011.59007.64004.94005.31005.81004.20008.02
KIRPAverage120.07195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.87011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
LIHCAverage122.63204.27220.37226.40212.83214.07219.07223.70200.60223.03225.17249.90
STDE008.01006.99006.29006.80017.60007.06007.88004.13007.66006.48004.81007.97
LUADAverage121.17195.13213.03225.30194.10199.43213.63212.67185.87212.40214.50233.00
STDE007.56011.17005.31003.11009.66009.37008.35005.66006.40006.10005.98004.34
LUSCAverage122.67195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.28011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
PAADAverage119.23195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.90011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
PCPGAverage122.43195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE006.68011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
READAverage123.93195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.45011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
SARCAverage135.27202.90226.90233.77194.10210.53223.87217.80189.93220.57220.87231.53
STDE018.57020.86020.25014.44009.66017.74017.39013.90013.41016.40011.76003.70
SKCMAverage122.47195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.66011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
STADAverage123.53194.60213.40225.03194.30198.80214.10212.70185.70211.97215.07231.77
STDE010.65010.22005.38003.29009.79009.09008.68005.54006.92005.88005.12004.58
THCAAverage119.47195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE005.58011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
THYMAverage121.80195.13213.03225.30194.10199.50213.63212.67185.60211.00211.57232.63
STDE007.21011.17005.31003.11009.66009.30008.35005.66006.49006.44005.08004.36
UCECAverage122.23196.00213.27224.83192.03200.97212.60213.03189.23212.23212.93233.93
STDE005.93009.06006.31004.27007.85008.24007.74005.67008.78004.98004.97003.97
Ranking W | T | L 22 | 0 | 0 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22 0 | 0 | 22
Boldface values denote the best results.
Table A8. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average precision values.
Table A8. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average precision values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CESC1.000001.000001.000001.000000.998901.000001.000001.000001.000001.000001.000001.00000
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000000.998901.000001.000001.000001.000001.000001.000001.00000
ESCA1.000001.000001.000001.000000.998201.000001.000001.000001.000001.000001.000000.99210
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC1.000000.941400.941700.937500.928400.941400.939900.941700.941100.940500.845500.93150
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC1.000001.000001.000001.000000.999401.000001.000001.000001.000001.000001.000001.00000
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD1.000001.000001.000001.000000.998701.000001.000001.000001.000001.000001.000000.99930
LUSC1.000000.998600.998600.998600.998300.996600.998600.999000.999300.998600.814000.99970
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC1.000000.999401.000000.999400.995000.999401.000000.999400.996200.999401.000000.99120
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD0.988300.948400.951100.941700.936100.947300.946900.946900.939100.946500.952000.93500
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000000.999100.998200.998200.993001.000000.998200.998200.999101.000001.000000.99390
Ranking ( W | T | L ) 3 | 19 | 0 0 | 17 | 5 0 | 18 | 4 0 | 17 | 5 0 | 12 | 10 0 | 18 | 4 0 | 18 | 4 0 | 17 | 5 0 | 17 | 5 0 | 18 | 4 0 | 19 | 3 0 | 15 | 7
Boldface values denote the best results.
Table A9. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average recall values.
Table A9. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average recall values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CESC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000000.999401.000001.000001.000001.000001.000000.999401.00000
ESCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC1.000001.000001.000000.999700.999301.000001.000001.000001.000001.000000.999300.99930
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC1.000000.990400.990400.990400.990100.990400.990400.990400.990400.990400.987800.99040
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD1.000001.000001.000001.000000.995001.000001.000001.000000.999301.000000.998300.99770
LUSC1.000000.991000.991000.991000.991300.989600.991000.990600.990300.991000.991000.98990
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000000.995500.996400.983800.982000.994600.991000.990100.984700.993700.978400.97930
Ranking ( W | T | L ) 3 | 19 | 0 0 | 19 | 3 0 | 19 | 3 0 | 18 | 4 0 | 16 | 6 0 | 19 | 3 0 | 19 | 3 0 | 19 | 3 0 | 18 | 4 0 | 19 | 3 0 | 16 | 6 0 | 17 | 5
Boldface values denote the best results.
Table A10. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average F1-score values.
Table A10. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average F1-score values.
DatasetRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
CESC1.000001.000001.000001.000000.999501.000001.000001.000001.000001.000001.000001.00000
CHOL1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
COAD1.000001.000001.000001.000000.999101.000001.000001.000001.000001.000000.999701.00000
ESCA1.000001.000001.000001.000000.999101.000001.000001.000001.000001.000000.998200.99600
GBM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
HNSC1.000000.969800.970000.967600.962500.969800.969000.969700.969700.969400.957300.96420
KICH1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
KIRC1.000000.995200.995200.995200.994700.995200.995200.995200.995200.995200.993200.99520
KIRP1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LIHC1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
LUAD1.000001.000001.000001.000000.996901.000001.000001.000000.999701.000000.998800.99850
LUSC1.000000.994800.994800.994800.994800.994800.994800.994800.994800.994800.994800.99480
PAAD1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
PCPG1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
READ1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
SARC0.999000.999700.998100.999700.997500.999701.000000.999700.998100.999700.997100.99560
SKCM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
STAD0.994100.973500.974900.969900.967000.972900.972700.972700.968600.972500.965600.96640
THCA1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
THYM1.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.000001.00000
UCEC1.000000.997300.997300.990900.987300.997300.994500.994100.991800.996800.986800.98640
Ranking ( W | T | L ) 5 | 17 | 0 0 | 16 | 6 0 | 16 | 6 0 | 16 | 6 0 | 12 | 10 0 | 16 | 6 0 | 17 | 5 0 | 16 | 6 0 | 15 | 7 0 | 16 | 6 0 | 13 | 9 0 | 14 | 8
Boldface values denote the best results.

Appendix B. Convergence Graphs of the Proposed RBAVO-DE and Competitive Methods

Appendix B.1. Convergence Graphs Employing the SVM Model

Figure A1. Convergence graphs of the proposed RBAVO-DE algorithm and competitive methods employing the SVM model on the overall benchmarks [57].
Figure A1. Convergence graphs of the proposed RBAVO-DE algorithm and competitive methods employing the SVM model on the overall benchmarks [57].
Algorithms 17 00342 g0a1aAlgorithms 17 00342 g0a1bAlgorithms 17 00342 g0a1cAlgorithms 17 00342 g0a1d

Appendix B.2. Convergence Graphs Employing the k-NN Model

Figure A2. Convergence graphs of the proposed RBAVO-DE algorithm and competitive methods employing the k-NN model on the overall benchmarks (Cont.) [57].
Figure A2. Convergence graphs of the proposed RBAVO-DE algorithm and competitive methods employing the k-NN model on the overall benchmarks (Cont.) [57].
Algorithms 17 00342 g0a2aAlgorithms 17 00342 g0a2bAlgorithms 17 00342 g0a2cAlgorithms 17 00342 g0a2d

References

  1. Estrada-Meza, C.; Torres-Copado, A.; Loreti González-Melgoza, L.; Ruiz-Manriquez, L.M.; De Donato, M.; Sharma, A.; Pathak, S.; Banerjee, A.; Paul, S. Recent insights into the microRNA and long non-coding RNA-mediated regulation of stem cell populations. 3 Biotech 2022, 12, 270. [Google Scholar] [CrossRef] [PubMed]
  2. Kakati, T.; Bhattacharyya, D.K.; Kalita, J.K.; Norden-Krichmar, T.M. DEGnext: Classification of differentially expressed genes from RNA-seq data using a convolutional neural network with transfer learning. BMC Bioinform. 2022, 23, 17. [Google Scholar] [CrossRef]
  3. Zhao, S.; Fung-Leung, W.P.; Bittner, A.; Ngo, K.; Liu, X. Comparison of RNA-Seq and microarray in transcriptome profiling of activated T cells. PLoS ONE 2014, 9, e78644. [Google Scholar] [CrossRef]
  4. Chen, Z.; Luo, Z.; Zhang, D.; Li, H.; Liu, X.; Zhu, K.; Zhang, H.; Wang, Z.; Zhou, P.; Ren, J.; et al. TIGER: A web portal of tumor immunotherapy gene expression resource. Genom. Proteom. Bioinform. 2023, 21, 337–348. [Google Scholar] [CrossRef] [PubMed]
  5. Nunez-Garcia, J.; AbuOun, M.; Storey, N.; Brouwer, M.; Delgado-Blas, J.; Mo, S.S.; Ellaby, N.; Veldman, K.; Haenni, M.; Châtre, P.; et al. Harmonisation of in-silico next-generation sequencing based methods for diagnostics and surveillance. Sci. Rep. 2022, 12, 14372. [Google Scholar] [CrossRef]
  6. Wang, Z.; Gerstein, M.; Snyder, M. RNA-Seq: A revolutionary tool for transcriptomics. Nat. Rev. Genet. 2009, 10, 57–63. [Google Scholar] [CrossRef]
  7. Kim, W.J.; Choi, B.R.; Noh, J.J.; Lee, Y.Y.; Kim, T.J.; Lee, J.W.; Kim, B.G.; Choi, C.H. Comparison of RNA-Seq and microarray in the prediction of protein expression and survival prediction. Front. Genet. 2024, 15, 1342021. [Google Scholar] [CrossRef]
  8. Wang, M.; Chen, X.; Dai, Y.; Wu, D.; Liu, F.; Yang, Z.; Song, B.; Xie, L.; Yang, L.; Zhao, W.; et al. Concordance study of a 520-gene next-generation sequencing-based genomic profiling assay of tissue and plasma samples. Mol. Diagn. Ther. 2022, 26, 309–322. [Google Scholar] [CrossRef]
  9. Metzker, M.L. Sequencing technologies—The next generation. Nat. Rev. Genet. 2010, 11, 31–46. [Google Scholar] [CrossRef] [PubMed]
  10. Pandey, D.; Onkara Perumal, P. A scoping review on deep learning for next-generation RNA-Seq. data analysis. Funct. Integr. Genom. 2023, 23, 134. [Google Scholar] [CrossRef]
  11. Liu, S.; Yao, W. Prediction of lung cancer using gene expression and deep learning with KL divergence gene selection. BMC Bioinform. 2022, 23, 175. [Google Scholar] [CrossRef] [PubMed]
  12. Houssein, E.H.; Oliva, D.; Celik, E.; Emam, M.M.; Ghoniem, R.M. Boosted sooty tern optimization algorithm for global optimization and feature selection. Expert Syst. Appl. 2023, 213, 119015. [Google Scholar] [CrossRef]
  13. Joshi, A.A.; Aziz, R.M. A two-phase cuckoo search based approach for gene selection and deep learning classification of cancer disease using gene expression data with a novel fitness function. Multimed. Tools Appl. 2024, 83, 71721–71752. [Google Scholar] [CrossRef]
  14. Ramaswamy, R.; Kandhasamy, P.; Palaniswamy, S. Feature selection for Alzheimer’s gene expression data using modified binary particle swarm optimization. IETE J. Res. 2023, 69, 9–20. [Google Scholar] [CrossRef]
  15. Cui, X.; Li, Y.; Fan, J.; Wang, T. A novel filter feature selection algorithm based on relief. Appl. Intell. 2022, 52, 5063–5081. [Google Scholar] [CrossRef]
  16. Alhenawi, E.; Al-Sayyed, R.; Hudaib, A.; Mirjalili, S. Feature selection methods on gene expression microarray data for cancer classification: A systematic review. Comput. Biol. Med. 2022, 140, 105051. [Google Scholar] [CrossRef] [PubMed]
  17. Parlak, B.; Uysal, A.K. A novel filter feature selection method for text classification: Extensive Feature Selector. J. Inf. Sci. 2023, 49, 59–78. [Google Scholar] [CrossRef]
  18. Albulayhi, K.; Abu Al-Haija, Q.; Alsuhibany, S.A.; Jillepalli, A.A.; Ashrafuzzaman, M.; Sheldon, F.T. IoT intrusion detection using machine learning with a novel high performing feature selection method. Appl. Sci. 2022, 12, 5015. [Google Scholar] [CrossRef]
  19. Fatima, A.; Nazir, T.; Nazir, A.K.; Din, A.M.U. An efficient Incremental Wrapper-based Information Gain Gene Subset Selection (IG based on IWSSr) method for Tumor Discernment. Multimed. Tools Appl. 2024, 83, 64741–64766. [Google Scholar] [CrossRef]
  20. Kaur, S.; Kumar, Y.; Koul, A.; Kumar Kamboj, S. A systematic review on metaheuristic optimization techniques for feature selections in disease diagnosis: Open issues and challenges. Arch. Comput. Methods Eng. 2023, 30, 1863–1895. [Google Scholar] [CrossRef]
  21. Abd El-Mageed, A.A.; Gad, A.G.; Sallam, K.M.; Munasinghe, K.; Abohany, A.A. Improved Binary Adaptive Wind Driven Optimization Algorithm-Based Dimensionality Reduction for Supervised Classification. Comput. Ind. Eng. 2022, 167, 107904. [Google Scholar] [CrossRef]
  22. Gad, A.G.; Sallam, K.M.; Chakrabortty, R.K.; Ryan, M.J.; Abohany, A.A. An improved binary sparrow search algorithm for feature selection in data classification. Neural Comput. Appl. 2022, 34, 15705–15752. [Google Scholar] [CrossRef]
  23. Hussien, R.M.; Abohany, A.A.; Abd El-Mageed, A.A.; Hosny, K.M. Improved Binary Meerkat Optimization Algorithm for efficient feature selection of supervised learning classification. Knowl.-Based Syst. 2024, 292, 111616. [Google Scholar] [CrossRef]
  24. Abd El-Mageed, A.A.; Abohany, A.A.; Elashry, A. Effective Feature Selection Strategy for Supervised Classification based on an Improved Binary Aquila Optimization Algorithm. Comput. Ind. Eng. 2023, 181, 109300. [Google Scholar] [CrossRef]
  25. Yin, Y.; Jang-Jaccard, J.; Xu, W.; Singh, A.; Zhu, J.; Sabrina, F.; Kwak, J. IGRF-RFE: A hybrid feature selection method for MLP-based network intrusion detection on UNSW-NB15 dataset. J. Big Data 2023, 10, 15. [Google Scholar] [CrossRef]
  26. Nakao, H.; Imaoka, M.; Hida, M.; Imai, R.; Nakamura, M.; Matsumoto, K.; Kita, K. Determination of individual factors associated with hallux valgus using SVM-RFE. BMC Musculoskelet. Disord. 2023, 24, 534. [Google Scholar] [CrossRef] [PubMed]
  27. Sarafrazi, S.; Nezamabadi-Pour, H. Facing the classification of binary problems with a GSA-SVM hybrid system. Math. Comput. Model. 2013, 57, 270–278. [Google Scholar] [CrossRef]
  28. Cadenas, J.M.; Garrido, M.C.; MartíNez, R. Feature subset selection filter–wrapper based on low quality data. Expert Syst. Appl. 2013, 40, 6241–6252. [Google Scholar] [CrossRef]
  29. Oh, I.S.; Lee, J.S.; Moon, B.R. Hybrid genetic algorithms for feature selection. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1424–1437. [Google Scholar]
  30. Abdollahzadeh, B.; Gharehchopogh, F.S.; Mirjalili, S. African vultures optimization algorithm: A new nature-inspired metaheuristic algorithm for global optimization problems. Comput. Ind. Eng. 2021, 158, 107408. [Google Scholar] [CrossRef]
  31. El-Shafeiy, E.; Hassanien, A.E.; Sallam, K.M.; Abohany, A. Approach for training quantum neural network to predict severity of COVID-19 in patients. Comput. Mater. Contin. 2020, 66, 1745–1755. [Google Scholar] [CrossRef]
  32. Yaqoob, A.; Verma, N.K.; Aziz, R.M. Optimizing gene selection and cancer classification with hybrid sine cosine and cuckoo search algorithm. J. Med. Syst. 2024, 48, 10. [Google Scholar] [CrossRef]
  33. Joshi, A.A.; Aziz, R.M. Deep learning approach for brain tumor classification using metaheuristic optimization with gene expression data. Int. J. Imaging Syst. Technol. 2023, 34, e23007. [Google Scholar] [CrossRef]
  34. Mahto, R.; Ahmed, S.U.; Rahman, R.u.; Aziz, R.M.; Roy, P.; Mallik, S.; Li, A.; Shah, M.A. A novel and innovative cancer classification framework through a consecutive utilization of hybrid feature selection. BMC Bioinform. 2023, 24, 479. [Google Scholar] [CrossRef] [PubMed]
  35. Neggaz, N.; Neggaz, I.; Abd Elaziz, M.; Hussien, A.G.; Abulaigh, L.; Damaševičius, R.; Hu, G. Boosting manta rays foraging optimizer by trigonometry operators: A case study on medical dataset. Neural Comput. Appl. 2024, 36, 9405–9436. [Google Scholar] [CrossRef]
  36. Lyu, B.; Haque, A. Deep learning based tumor type classification using gene expression data. In Proceedings of the 2018 ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Washington, DC, USA, 29 August–1 September 2018; pp. 89–96. [Google Scholar]
  37. Khalifa, N.E.M.; Taha, M.H.N.; Ali, D.E.; Slowik, A.; Hassanien, A.E. Artificial intelligence technique for gene expression by tumor RNA-Seq data: A novel optimized deep learning approach. IEEE Access 2020, 8, 22874–22883. [Google Scholar] [CrossRef]
  38. Dillies, M.A.; Rau, A.; Aubert, J.; Hennequet-Antier, C.; Jeanmougin, M.; Servant, N.; Keime, C.; Marot, G.; Castel, D.; Estelle, J.; et al. A comprehensive evaluation of normalization methods for Illumina high-throughput RNA sequencing data analysis. Briefings Bioinform. 2013, 14, 671–683. [Google Scholar] [CrossRef]
  39. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A deep learning-based multi-model ensemble method for cancer prediction. Comput. Methods Programs Biomed. 2018, 153, 1–9. [Google Scholar] [CrossRef]
  40. Liu, M.; Xu, L.; Yi, J.; Huang, J. A feature gene selection method based on ReliefF and PSO. In Proceedings of the 2018 10th International Conference on Measuring Technology and Mechatronics Automation (ICMTMA), Changsha, China, 7–8 January 2018; IEEE: New York, NY, USA, 2018; pp. 298–301. [Google Scholar]
  41. Kononenko, I. Estimating attributes: Analysis and extensions of RELIEF. In Proceedings of the European Conference on Machine Learning, Catania, Italy, 6–8 April 1994; Springer: Berlin/Heidelberg, Germany, 1994; pp. 171–182. [Google Scholar]
  42. Faris, H.; Mafarja, M.M.; Heidari, A.A.; Aljarah, I.; Ala’M, A.Z.; Mirjalili, S.; Fujita, H. An efficient binary salp swarm algorithm with crossover scheme for feature selection problems. Knowl.-Based Syst. 2018, 154, 43–67. [Google Scholar] [CrossRef]
  43. Abdel-Basset, M.; Ding, W.; El-Shahat, D. A hybrid Harris Hawks optimization algorithm with simulated annealing for feature selection. Artif. Intell. Rev. 2020, 54, 593–637. [Google Scholar] [CrossRef]
  44. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  45. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  46. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  47. Khanesar, M.A.; Teshnehlab, M.; Shoorehdeli, M.A. A novel binary particle swarm optimization. In Proceedings of the 2007 Mediterranean Conference on Control & Automation, Athens, Greece, 27–29 June 2007; IEEE: New York, NY, USA, 2007; pp. 1–6. [Google Scholar]
  48. Mirjalili, S.; Mirjalili, S.M.; Yang, X.S. Binary bat algorithm. Neural Comput. Appl. 2014, 25, 663–681. [Google Scholar] [CrossRef]
  49. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  50. Hichem, H.; Elkamel, M.; Rafik, M.; Mesaaoud, M.T.; Ouahiba, C. A new binary grasshopper optimization algorithm for feature selection problem. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 316–328. [Google Scholar] [CrossRef]
  51. Hussien, A.G.; Hassanien, A.E.; Houssein, E.H.; Bhattacharyya, S.; Amin, M. S-shaped binary whale optimization algorithm for feature selection. In Recent Trends in Signal and Image Processing: ISSIP 2017; Springer: Berlin/Heidelberg, Germany, 2019; pp. 79–87. [Google Scholar]
  52. Zhao, W.; Wang, L.; Zhang, Z. Atom search optimization and its application to solve a hydrogeologic parameter estimation problem. Knowl. Based Syst. 2019, 163, 283–304. [Google Scholar] [CrossRef]
  53. Meng, X.B.; Gao, X.Z.; Lu, L.; Liu, Y.; Zhang, H. A new bio-inspired optimisation algorithm: Bird Swarm Algorithm. J. Exp. Theor. Artif. Intell. 2016, 28, 673–687. [Google Scholar] [CrossRef]
  54. Hashim, F.A.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W.; Mirjalili, S. Henry gas solubility optimization: A novel physics-based algorithm. Future Gener. Comput. Syst. 2019, 101, 646–667. [Google Scholar] [CrossRef]
  55. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  56. Normalized-level3 RNA-Seq Gene Expression Dataset. Available online: https://gdac.broadinstitute.org/ (accessed on 20 December 2023).
  57. El-Mageed, A.A.A.; Elkhouli, A.E.; Abohany, A.A.; Gafar, M. Gene selection via improved nuclear reaction optimization algorithm for cancer classification in high-dimensional data. J. Big Data 2024, 11, 46. [Google Scholar] [CrossRef]
  58. Mafarja, M.; Mirjalili, S. Whale optimization approaches for wrapper feature selection. Appl. Soft Comput. 2018, 62, 441–453. [Google Scholar] [CrossRef]
  59. Thaher, T.; Heidari, A.A.; Mafarja, M.; Dong, J.S.; Mirjalili, S. Binary Harris Hawks optimizer for high-dimensional, low sample size feature selection. In Evolutionary Machine Learning Techniques; Springer: Berlin/Heidelberg, Germany, 2020; pp. 251–272. [Google Scholar]
  60. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  61. Python Code for Gene Selection via Relief Binary African Vultures Optimization Integrated with Differential Evolution. Available online: https://github.com/D-Amr-Atef/Gene_Selection_RBAVO-DE.git (accessed on 18 May 2024).
Figure 1. Flowchart of the RBAVO-DE algorithm.
Figure 1. Flowchart of the RBAVO-DE algorithm.
Algorithms 17 00342 g001
Table 1. Datasets employed in this paper [57].
Table 1. Datasets employed in this paper [57].
#CancerNormalTumor
RecordsRecords
1Bladder Urothelial Carcinoma (BLCA)19408
2Thyroid Cancer (THCA)59501
3Cervical and Endocervical Cancers (CESCs)3304
4Cholangiocarcinoma (CHOL)936
5Colon Adenocarcinoma (COAD)41458
6Esophageal Cancer (ESCA)13184
7Glioblastoma Multiforme (GBM)5153
8Thymoma (THYM)2120
9Head and Neck Squamous Cell Carcinoma
(HNSC)44520
10Kidney Chromophobe (KICH)2566
11KIRC72533
12Kidney Renal Papillary32290
(KIRP)
13Liver Hepatocellular Carcinoma (LIHC)50371
14LUAD59515
15LUSC51501
16Pancreatic Adenocarcinoma (PAAD)4178
17UCEC35176
18Pheochromocytoma and3179
Paraganglioma (PCPG)
19Rectum adenocarcinoma (READ)1094
20Sarcoma (SARC)2259
21Skin Cutaneous Melanoma (SKCM)1103
22STAD37415
Table 2. The key parameters of the employed ML models.
Table 2. The key parameters of the employed ML models.
ModelParameters
SVMPolynomial kernel = 2
k-NNEuclidean distance metric k = 5 [21,58,59]
Table 3. Parameter configurations of all utilized optimizers.
Table 3. Parameter configurations of all utilized optimizers.
OptimizerParameters
All optimizersExecution count = 30
Maximum count of iterations G m a x = 100
Size of population N = 10
Dimensionality D = The count of genes
in the employed datasets
Upper boundaries U B
Lower boundaries L B
Proposed RBAVO-DEParameter L 1 = 0.7
Parameter L 2 = 0.2
Parameter w = 2
Parameter P 1 = 0.6
Parameter P 2 = 0.6
Parameter P 3 = 0.5
BSSACount of scroungers S D = 0.1 × N
Count of producers P D = 0.2 × N
Safety threshold S T = 0.8
BABCCount of employed bees = 16
Count of scout bees = 3
Count of onlooker bees = 4
BPSOInertia weight ω m a x = 0.9 ω m i n = 0.4
Acceleration coefficients c 2 = c 1 = 1.2
BBALoudness A = 0.8
Pulse emission rate r = 0.95
Lower and upper pulse frequencies = 0 , 10
BWOAa is linearly decreased from 2 to 0
p = 0.5
b = 1.0
BHHORabbit energy E [ 1 , 1 ]
BGWOa is linearly decreased from 2 to 0
BGOA C min = 0.00004 and C max = 1
BBSAFrequency of flight f f = 10
Effect on birds’ vigilance
Followed coefficient f l = 0.5
Probability of foraging for food p = 0.8
Acceleration coefficients c 1 = c 2 = 1.5
behaviors a 1 = a 2 = 1.0
BASODepth weight α = 50
Multiplier weight β = 0.2
BHGSOCount of clusters = 2
α = β = 0.1 K = 1.0 , and l 3 = 1 E 02
l 1 = 5 E 03 , l 2 = 1 E + 02
Table 4. The results of the proposed RBAVO-DE algorithm and its counterparts using k-NN regarding average classification accuracy results.
Table 4. The results of the proposed RBAVO-DE algorithm and its counterparts using k-NN regarding average classification accuracy results.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage0.9980.9930.9980.9930.9890.9950.9940.9930.9910.9940.9940.989
STDE0.0040.0060.0050.0060.0030.0060.0060.0060.0050.0060.0060.002
CESCAverage0.9840.9840.9840.9840.9840.9840.9840.9840.9840.9840.9840.984
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
CHOLAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
COADAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
ESCAAverage0.9740.9740.9740.9740.9740.9740.9740.9740.9740.9740.9740.974
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
GBMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
HNSCAverage0.9930.9910.9910.9910.9910.9920.9910.9910.9910.9910.9910.991
STDE0.0040.0020.0020.0000.0000.0020.0000.0000.0000.0020.0000.000
KICHAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
KIRCAverage0.9980.9930.9940.9920.9920.9940.9920.9940.9930.9930.9930.992
STDE0.0030.0030.0040.0020.0000.0040.0020.0040.0030.0030.0030.000
KIRPAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
LIHCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
LUADAverage0.9910.9910.9910.9910.9910.9910.9910.9910.9910.9910.9910.991
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
LUSCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
PAADAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
PCPGAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
READAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
SARCAverage0.9810.9810.9810.9810.9810.9810.9810.9810.9810.9810.9810.981
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
SKCMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
STADAverage0.9990.9960.9990.9960.9890.9960.9950.9970.9920.9950.9960.989
STDE0.0030.0050.0040.0050.0000.0050.0060.0050.0050.0060.0050.000
THCAAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
THYMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
UCECAverage0.9980.9850.9880.9770.9690.9840.9770.9780.9780.9790.9760.968
STDE0.0080.0120.0130.0060.0110.0120.0060.0080.0090.0090.0050.011
Ranking W | T | L 4 | 18 | 0 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4
Boldface values denote the best results.
Table 5. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average classification accuracy values.
Table 5. The results of the proposed RBAVO-DE algorithm and its counterparts employing SVM regarding average classification accuracy values.
DatasetMetricRBAVO-DEBSSABABCBPSOBBABGWOBWOABGOABHHOBBSABASOBHGSO
BLCAAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
CESCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
CHOLAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
COADAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
ESCAAverage0.9860.9750.9770.9740.9740.9770.9740.9750.9750.9750.9740.974
STDE0.0130.0040.0080.0000.0000.0080.0000.0050.0050.0050.0000.000
GBMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
HNSCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
KICHAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
KIRCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
KIRPAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
LIHCAverage1.0001.0001.0001.0000.99921.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.00290.0000.0000.0000.0000.0000.0000.000
LUADAverage0.9930.9910.9910.9910.9910.9910.9910.9910.9910.9910.9920.991
STDE0.0030.0000.0000.0000.0000.0000.0000.0000.0000.0000.0020.000
LUSCAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
PAADAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
PCPGAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
READAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
SARCAverage0.9940.9860.9890.9860.9810.9890.9870.9860.9840.9860.9840.981
STDE0.0080.0080.0090.0080.0000.0090.0090.0080.0060.0080.0070.000
SKCMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
STADAverage0.9900.9890.9890.9890.9890.9890.9890.9890.9890.9890.9890.989
STDE0.0030.0000.0000.0000.0020.0000.0000.0000.0020.0000.0000.000
THCAAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
THYMAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
UCECAverage1.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.0001.000
STDE0.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
Ranking W | T | L 4 | 18 | 0 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4 0 | 18 | 4
Boldface numbers indicate the best results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gafar, M.G.; Abohany, A.A.; Elkhouli, A.E.; El-Mageed, A.A.A. Optimization of Gene Selection for Cancer Classification in High-Dimensional Data Using an Improved African Vultures Algorithm. Algorithms 2024, 17, 342. https://doi.org/10.3390/a17080342

AMA Style

Gafar MG, Abohany AA, Elkhouli AE, El-Mageed AAA. Optimization of Gene Selection for Cancer Classification in High-Dimensional Data Using an Improved African Vultures Algorithm. Algorithms. 2024; 17(8):342. https://doi.org/10.3390/a17080342

Chicago/Turabian Style

Gafar, Mona G., Amr A. Abohany, Ahmed E. Elkhouli, and Amr A. Abd El-Mageed. 2024. "Optimization of Gene Selection for Cancer Classification in High-Dimensional Data Using an Improved African Vultures Algorithm" Algorithms 17, no. 8: 342. https://doi.org/10.3390/a17080342

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop