Next Article in Journal
Phytochemical Targeting of Mitochondria for Breast Cancer Chemoprevention, Therapy, and Sensitization
Next Article in Special Issue
Fluoride in the Central Nervous System and Its Potential Influence on the Development and Invasiveness of Brain Tumours—A Research Hypothesis
Previous Article in Journal
Glycogen Synthase Kinase 3β: A True Foe in Pancreatic Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Voting-Based Feature Selection and Ensemble Learning Model Scheme for Glioma Grading with Clinical and Molecular Characteristics

Radiation Oncology Branch, Center for Cancer Research, National Cancer Institute, National Institutes of Health, Building 10, Bethesda, MD 20892, USA
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(22), 14155; https://doi.org/10.3390/ijms232214155
Submission received: 16 September 2022 / Revised: 31 October 2022 / Accepted: 12 November 2022 / Published: 16 November 2022
(This article belongs to the Special Issue Biomechanics and Molecular Research on Glioblastoma)

Abstract

:
Determining the aggressiveness of gliomas, termed grading, is a critical step toward treatment optimization to increase the survival rate and decrease treatment toxicity for patients. Streamlined grading using molecular information has the potential to facilitate decision making in the clinic and aid in treatment planning. In recent years, molecular markers have increasingly gained importance in the classification of tumors. In this study, we propose a novel hierarchical voting-based methodology for improving the performance results of the feature selection stage and machine learning models for glioma grading with clinical and molecular predictors. To identify the best scheme for the given soft-voting-based ensemble learning model selections, we utilized publicly available TCGA and CGGA datasets and employed four dimensionality reduction methods to carry out a voting-based ensemble feature selection and five supervised models, with a total of sixteen combination sets. We also compared our proposed feature selection method with the LASSO feature selection method in isolation. The computational results indicate that the proposed method achieves 87.606% and 79.668% accuracy rates on TCGA and CGGA datasets, respectively, outperforming the LASSO feature selection method.

1. Introduction

Tumor grading, i.e., the ability to determine the biological aggressiveness of glioma, is very important for adequate management, including treatment planning and monitoring and the rate of survival in patients [1,2]. Gliomas are rapidly progressive, neurologically devastating, and the most common primary brain tumor originating from glial cells [3,4,5]. Standard-of-care treatment of gliomas is predicated on tumor grade, although it generally involves maximal surgical resection followed by radiation therapy (RT) with systemic treatment in the form of temozolomide (TMZ) chemotherapy administered concurrently and/or sequentially with the alternative of sequential PCV or PC (Procarbazine, CCNU with or without vincristine) [2,5]. According to the guidelines of the World Health Organization (WHO) classification of Central Nervous System (CNS) tumors, gliomas can be currently broadly categorized into two groups, low-grade gliomas (LGG) and high-grade gliomas (HGG), with glioblastoma multiforme (GBM), a high-grade glioma, based on histological and molecular parameters [6]. GBM is the most common, aggressive, invasive, and primary type of tumor. Glioma accounts for almost 80% of all primary malignant tumors of the brain, and GBM also accounts for more than 60% of all brain tumors in adults, emphasizing the importance of grading in this neuro-oncology scenario [7].
In recent years, as molecular alterations have gained importance for the classification of CNS tumors [6,8,9], a push for valued-added care has also added complexity to the discussion. Two facets emerged: the vast and heterogenous number of available molecular parameters emphasizing the importance of the identification and selection of the necessary molecular alterations, and the goal of reducing the cost of molecular testing to allow for its more widespread use and mitigation of health disparities. One example is the isocitrate dehydrogenase (IDH) mutation, a significant molecular marker to distinguish low-grade glioma from high-grade gliomas [10,11]. IDH testing has prompted discussion of cost and turnaround time, with USD 135 per p.R132H-specific IDH1 immunohistochemistry, USD 420 for single gene sequencing, and USD 1800 for next-generation sequencing [10]. Turnaround time utilized can take about two days for immunohistochemistry and 14 days for next-generation sequencing [10]. Clinical features such as age and gender [12] contribute to the tumor grading process, but there is a paucity of higher-level robust clinical annotation in publicly available datasets limiting links between pertinent molecular features and clinical data, which could advance value-added care as more widespread molecular testing may eventually benefit from the increase in reimbursement. Thus, selecting the best discriminative molecular and clinical markers not only reduces the cost to healthcare systems and patients while helping to curb growing health disparities in access to testing but also improves tumor grading performance, which can enable the selection of pertinent molecular features for future analyses and bench to bedside work as well as testing of novel targeted agents. The pattern recognition necessary to optimally leverage fragmented available molecular information is arguably not possible in the absence of computational analysis, hence our hypothesis that feature selection carries a significant role in this space.
Feature selection is concerned with the selection of the best feature subset from all features or patterns according to strategies that remove unrelated, insignificant, and redundant features [13]. Specifically, this process guarantees the best class prediction performance and reduces the computational demand/cost, increasing efficiency and providing more cost-effective features, increasing the classification accuracy rate, and improving the clarity of the results [14,15]. Today, feature selection is widely used in many data analysis applications, pattern recognition, and mining tasks [16,17]. As example studies of brain tumor grading, Ref. [18] performed subtyping and grading of lower-grade gliomas using integrated SVM recursive feature elimination and a correlation method using transcriptome data. In another study [19], molecular three-subtype classification of low-grade gliomas using magnetic-resonance-imaging-based radiomic features and employing genetic-algorithm-based feature selection with an eXtreme Gradient Boosting classifier was provided. In [20], a joint similarity network fusion (Joint-SNF) method was proposed to integrate different omic data types for subtype identification of Chinese lower-grade glioma. In [21], a deep-learning-based framework was developed for the precise and accurate classification of GBM subtypes by employing transcriptome and methylome data types. Although there is no single algorithm that could outperform every other machine learning model for all problems [22,23] (i.e., no free lunch theorem), many studies in the literature have shown that an ensemble of models generally outperforms individual models [22,24,25]. In recent years, ensemble learning methods have been considered state-of-the-art approaches for solving machine learning challenges [23,24,25,26,27]. In [28], the performances of advanced ensemble learning methods have been found to be more robust than the well-known Random Forest and AdaBoost ensemble classifiers. Currently, bagging and boosting models are one of the most popular ensemble learning methods, with Random Forest and AdaBoost being the most prominent and common implementations, respectively [29]. These methods cover a large scale of applications including face recognition, anomaly detection, and medicine [29]. The performance (e.g., the accuracy rate) of a machine learning model can be improved by training multiple learning models and combining their predictions (namely, classifier ensembles) [23,26,27,30,31]. This operation can be performed by employing various schemes such as bagging, boosting, or voting. Voting approaches can also be hard or soft voting based on the usage of the prediction results. Hard voting relies on only summing and majority voting the class outputs. Soft voting approaches yield more flexible and fine-grained results than hard majority voting due to handling probability scores for various machine learning tasks [32].
To this end, we aim to select the best subset of molecular and clinical features by reducing the number of features with various feature selection methods and ensemble-learning-based models in this study. We introduce a novel hierarchical voting-based methodology for the selection of the relevant features and improving the performance results of the machine learning models for the classification tasks. To reach the best performance result, we carried out experiments employing the possible combinations of supervised models in the soft-voting-based process employing two glioma datasets, TCGA and CGGA, that are currently in widespread use.
The main contributions of this study are summarized as follows:
  • To the best of our knowledge, our study illustrates the first method that employs the hierarchical voting-based approach in the processes of both feature selection methods and ensemble learning models together to improve glioma grading results, which is an innovative approach to large-scale molecular glioma data.
  • To the best of our knowledge, our study also constructs the first organized and structural dataset from TCGA LGG and GBM data that aggregates molecular and clinical information to allow for validation and data sharing and to speed up research progress for researchers working in this field.
  • We focus on fusing the advantages of various feature selection methods and machine learning models via a voting-based procedure for glioma grading on two commonly used glioma datasets (TCGA and CGGA).
  • We carry out comprehensive computational results for the comparison of our novel voting-based feature selection method with the LASSO feature selection method in isolation, given that these are two commonly employed glioma datasets that share similarities but also exhibit differences.
  • We aim to identify the best-performing combination of a voting-based ensemble learning model for the feature selection stage to obtain the most accurate results given dataset variability in large-scale datasets with transferable use.
  • The performances of voting-based ensemble models are evaluated with six measures for glioma grading.
  • The extensive effects of the combinations of sixteen ensemble learning models are presented.
The remaining sections of this study are structured as follows: First, we briefly give an overview of the proposed methodology and explain the related feature selection methods and supervised learning models for glioma grading in Section 2. We describe the experimental procedures, datasets employed, and evaluation metrics and give the comprehensive experimental results of our method with discussions in Section 3. Finally, Section 4 contains the conclusion, discussion of the results, and possible future directions of this study.

2. Results

This section defines the experimental processes, the dataset used and its characteristics, and evaluation metrics. Then, we present the comprehensive experimental results of the performance metrics of tumor grading in the following subsections.

2.1. Experimental Process

In this study, the proposed methods were implemented in Python programming language with the scikit-learn [33] package for machine learning and the xverse [34] package for feature selection operations, and all the experiments were conducted on a system running a MacBook Pro laptop PC with macOS Monterey, 2.3 GHz 8-core Intel Core i9 CPU, and 16 GB 2667 MHz DDR4 RAM. The predictive models utilized are LR, SVM, KNN, RF, and AdaBoost. The combinations of 3, 4, and 5 of these models are used for the soft-voting-based ensemble learning scheme. The total number of ensemble learning model combinations is 16 (i.e., C(5,3) + C(5,4) + C(5,5) = 16). The corresponding combination set numbers and ensemble model combinations for the voting process are illustrated in Table 1.
As a preliminary study, we focused on decreasing the number of combinations of preprocessing stage by selecting appropriate normalization techniques from [0, 1] min-max normalization, [−1, 1] min-max normalization, or z-score normalization/standardization for age feature values. We experimented with these techniques on the various classifiers employed and decided to use the z-score normalization technique, given that it resulted in the best performance metric (i.e., accuracy rate) value obtained from the models.
For the minimum number of votes based feature selection, the weight of evidence, recursive feature elimination, Random Forest, and LASSO methods were applied. The minimum number of votes was set to 1 to eliminate features not chosen unanimously. For LASSO-based linear model operation, 10-fold cross-validation was carried out to find the best alpha parameter value through the iterations. A 10-fold cross-validation technique was also applied on the TCGA and CGGA datasets to construct and test the mean performance results of the learning models employed. In this study, the GBM class was considered positive and the LGG class was used as negative for the evaluation of learning models. The default values were assigned as the corresponding parameter values for the utilized classifiers (i.e., num of neighbors = 5, metric = ‘minkowski’ for KNN; C = 1, and kernel = ‘rbf’, gamma = ’scale’ for SVM; penalty = ‘l2’, *, dual = False, tol = 0.0001, C = 1.0, fit_intercept = True, intercept_scaling = 1, class_weight = None for LR; n_estimators = 100, *, criterion = ‘gini’, max_depth = None, min_samples_split = 2, min_samples_leaf = 1 for RF; n_estimators = 50, learning_rate = 1.0, algorithm = ‘SAMME.R’ for AdaBoost). We set the random state number to 0 for all employed learning models in order to obtain the same computational results with the same random state number on the datasets used.

2.2. Dataset

To evaluate our proposed methodology for the hierarchical voting-based processes, we utilized the two most widely employed genome atlas databases, namely The Cancer Genome Atlas (TCGA) and the Chinese Glioma Genome Atlas (CGGA) [35], for analyzing brain tumor (i.e., glioma) grading.
TCGA’s original dataset consists of 3 clinical features, the most frequently mutated 20 molecular/mutation features, and class labels for glioma grading, resulting in 23 total features (Table 2). The molecular features are captured as mutated or not_mutated (wildtype) depending on the TCGA Case_ID. The grades can be dichotomized into LGG (lower-grade glioma) or GBM (glioblastoma multiforme). The total numbers of instances for the TCGA dataset without and with preprocessing are 862 and 839, respectively. In the preprocessing stage, missing data from this dataset for gender, Age_at_diagnosis, or race were removed when feature values were captured as ‘--’, or ‘not reported’. Age_at_diagnosis feature values were also converted from string to continuous value by adding day information to the corresponding year information in the dataset as a floating-point number for the preprocessing stage. The CGGA dataset has 286 instances and 22 features (one less than TCGA) with the same characteristics illustrated in Table 2 except for the race feature, which is not directly specified in CGGA; however, CGGA documentation describes data origin as based on a Chinese cohort. There are no missing data for the CGGA dataset. The TCGA preprocessed dataset consists of 352 GBM and 487 LGG patients and there is also 102 GBM and 184 LGG patients for the CGGA dataset. No balancing/sampling strategy was applied for this study. The dataset query and storing operations were provided by NIDAP [36].
According to TCGA, the distribution of the most frequently mutated 20 genes according to percentage of cases affected for glioma data are given in Figure 1. As shown in Figure 1, the percentage of cases affected decreases considerably after 20 molecular features, and computational load increases due to the number of features. Thus, we found it appropriate to choose 20 molecular features in this study.

2.3. Evaluation Metrics

To measure the tumor grading performance of the proposed hierarchical voting-based feature selection and ensemble learning methodology, we employed 6 evaluation metrics, classification accuracy rate (ACC), Area Under the ROC Curve (AUC), F-Measure (F1), precision (PRE), recall (REC), and specificity (SPEC) [37].
The classification accuracy rate was computed by dividing the total number of true positives and true negatives by the total number of true positives, false negatives, false positives, and true negatives. The equation is defined in Equation (1).
ACC = TP + TN TP + TN + FP + FN
where TP, TN, FP, and FN represent the number of true positives, true negatives, false positives, and false negatives, respectively.
AUC, the Area Under the Receiver Operating Characteristic (ROC) curve, was constructed by plotting the true-positive rate against the false-positive rate for the performance of the binary learning model. The area of the maximum AUC value (i.e., 1) indicates a perfect test, and the AUC value of 0 specifies that the predictor classifies all instances incorrectly. A value of 0.5 for AUC indicates that the ROC curve will be diagonal (i.e., 45-degree line), and hence suggests that the diagnostic test has no discriminatory ability and prediction problem [38].
F-Measure is calculated by the harmonic mean of precision and recall. It is represented in Equation (2).
F 1 = 2   ×   PRE   ×   REC PRE + REC
Precision means the positive predictive value. It is computed by dividing the number of true positives by the total number of true positives and false positives. The equation is illustrated in Equation (3).
PRE = TP TP + FP
Recall is the true-positive rate/hit rate or sensitivity. It is calculated by dividing the number of true positives by the total number of true positives and false negatives. The equation is shown in Equation (4).
REC = TP TP + FN
Specificity is expressed as the true negative rate. It is calculated by dividing the number of true negatives by the total number of false positives and true negatives. The equation of specificity is defined in Equation (5).
SPEC = TN TN + FP

2.4. Experimental Results

In this study, we present the computational results to evaluate the effects of the features selection and ensemble learning model scheme stages for the TCGA and CGGA datasets (Figure 2 and Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11). Bold values indicate the best results.
We initially performed voting-based feature selection results using four methods (the weight of evidence, recursive feature elimination, Random Forest, and LASSO) for the ensemble combination sets in terms of ACC, AUC, F1, PRE, REC, and SPEC. Computational results revealed the best performance to originate from the ensemble of SVM, RF, and AdaBoost learning models with the values of 0.876, 0.858, and 0.815 for accuracy rate, F1, and precision values, respectively, for the TCGA dataset (Table 3) reflecting set combination #9 (Table 1) and the best accuracy rate value of 0.797 originating from the ensemble of SVM, KNN, RF, and AdaBoost for CGGA (Table 4) reflecting set combination #15 (Table 1).
The mean numbers of selected features after applying voting-based feature selection through 10-fold cross-validation on the TCGA and CGGA datasets are 14.9 and 17.6 for TCGA and CGGA, respectively (Table 5). The feature selection stage results in an approximate 35% and 20% cost gain in terms of feature reduction for the TCGA and CGGA datasets, respectively. A comprehensive comparison of the six performance metrics (ACC, AUC, F1, PRE, REC, and SPEC) on TCGA and CGGA datasets reveals CGGA’s superiority over TCGA in only one domain, that of specificity (Figure 2). A comparison of performance metrics with analysis carried out in the absence of the feature selection process reveals that the feature selection operation with the best ensemble combination yields approximately 1.38% accuracy rate improvement for the TCGA dataset as compared to CGGA (2.68%) (Table 4 with feature selection process vs. Table 6 without feature selection process).
We also investigated the effects of only using the LASSO feature selection method through individual supervised models on TCGA and CGGA datasets in comparison to the novel method, given the widespread use of LASSO in isolation (Table 7 and Table 8). The Logistic Regression classifier resulted in the best accuracy rate, AUC, precision, and specificity values with the values of 0.871, 0.920, 0.808, and 0.847, respectively, on the TCGA dataset, with the SVM model providing the best accuracy rate, AUC, precision, and specificity values (0.786, 0.817, 0.781, and 0.916), respectively, for CGGA.
Furthermore, we carried out experiments to observe the effects of only using the LASSO feature selection method through ensemble learning models on TCGA and CGGA datasets with the related combination sets (Table 9 and Table 10). The highest predictive performance was achieved with combination sets of models #2 and #13 (LR + SVM + RF and LR + SVM + RF + AdaBoost) with ACC 0.874 for TCGA (Table 9). These combination sets also resulted in the highest F-Measure result, with a value of 0.855. Since both combination sets resulted in the same results, we selected combination set #2, given its superior computational efficiency. The CGGA dataset combination set #9 (SVM + RF + AdaBoost) resulted in the highest predictive performance result, with an ACC of 0.786. Voting of SVM + RF + AdaBoost models also resulted in the highest F-Measure and recall results, with values of 0.648 and 0.594, respectively (Table 10).
When comparing the LASSO-only-based results with the voting-based feature selection results, we noted that the hierarchical voting-based feature selection and ensemble learning process provided more accurate results as compared to employing the LASSO feature selection method. The number of selected features when employing only the LASSO operation on TCGA and CGGA datasets (Table 11) resulted in a lower number of features than the total in both datasets but in particular in the CGGA set (11.9). The main advantage of the voting-based feature selection process is that it provides superior performance with cost improvement and gives the authorization to remove redundant features with unanimous selection.

3. Discussion

Based on the empirical results of our novel hierarchical voting-based feature selection and ensemble learning methodology including sixteen ensemble model combinations on two datasets with molecular and clinical characteristics, several insights follow.
Considering the predictive performance of the ensemble learning models utilized for the voting-based feature selection process in this study, our proposed feature selection method outperforms the results of the using only LASSO feature selection method. The following results were obtained:
  • The empirical results indicate that our novel hierarchical voting-based feature selection and ensemble learning methodology can provide promising results for the glioma grading tasks. We employed four different feature selection methods and the combinations of five different individual learning models using the voting-based strategies to reach optimal results for this study. In this regard, to validate our ensemble feature selection method, we compared our methodology on two datasets with different ensemble model combinations and we outperformed the results of using the LASSO-only feature selection method.
  • Our primary goal in this study was to reduce the cost and number of features of the proposed feature selection method and increase the prediction performance. In addition, although the selection of the best ensemble learning models and features varied in both datasets, we can conclude that our proposed novel ensemble-based feature selection method results in more accurate results as compared to only using the LASSO method. In this context, when datasets with a high sample number are obtained, factors such as the number of selected features, performance, and cost increase will become more stable.
  • Our novel hierarchical voting-based methodology can be extended with a various number of feature selection methods and machine learning models as a framework to improve the many classifications, or pattern recognition tasks if desired.
  • Voting-based ensemble schemes can give more flexibility as well as trustworthy and efficient results than those using individual schemes for related problems.
  • Our feature selection methodology can also be combined with deep-learning-based feature reduction approaches.
The shortcomings of the dataset constructed are that we have a small size of data or features in this study since we accessed a limited number of important features; however, we need to explore higher-dimensional biomedical datasets to try these methods.

4. Materials and Methods

In this section, we present an overview of the proposed methodology for glioma grading and explain the feature selection and classification methods applied for this study. We also describe the ensemble learning process in the following subsections.

4.1. The Proposed Methodology for Glioma Grading

The overview of the proposed methodology for glioma grading is illustrated in Figure 3. The hierarchical voting-based methodology includes two phases: (a) the minimum number of votes based feature selection and (b) soft-voting-based ensemble model selection.
As an initial step, the glioma datasets (TCGA and CGGA) are split into training–test datasets via a ten-fold cross-validation technique. Then, all clinical and molecular features and grade class label information are constructed for the subsequent machine-learning-related tasks. Next, data type conversion and removal of missing or insignificant values and columns are performed in the data preprocessing stage, followed by the hierarchical voting-stages-based ensemble learning scheme that is applied as a novel methodology for the glioma grading process. The substages of this methodology are briefly explained as follows.
The minimum number of votes based feature selection stage uses four different feature selection methods, weight of evidence, recursive feature elimination, Random Forest, and least absolute shrinkage and selection operator (LASSO), to rank features individually by employing the five supervised learning models according to their importance levels (e.g., if the importance level of the feature is equal to 0, it is disregarded). If the minimum number of votes is equal to 1, this means the feature to be selected must have at least a total of 1 vote from the four feature selection methods. Then, the selected features that meet the minimum number of votes following the administration of the four feature selection methods are transferred to the next level to experiment with all the possible combination sets for the voting-based ensemble learning process.
The soft-voting-based ensemble model selection stage consists of five supervised learning models (Logistic Regression, Support Vector Machine, K Nearest Neighbors, Random Forest, and AdaBoost). To merge the advantages of the individual learning models, we explored all three, four, and five combinations of all employed learning models to improve the classification performance results of the soft-voting-based ensemble learning models. Each method is explained in the following subsections in detail.

4.2. Feature Selection Methods

Feature selection methods are one of the types of dimensionality reduction approaches. They aim to find optimal feature subset n from all feature set m where n ≤ m [39]. These methods allow to improve the predictor performance, reduce the computational or memory cost, and facilitate data visualization by removing redundant and insignificant attributes in the pattern [40,41]. Feature selection methods often are categorized into three categories depending on if the learning model is utilized for the selection process as filter, wrapper, or embedded methods.
We explain the four feature selection methods (weight of evidence, recursive feature elimination, Random Forest, and LASSO) employed in this study in the following subsections.

4.2.1. Weight of Evidence (WOE)

The weight of evidence is a quantitative method that estimates the predictive power of an independent variable in relation to the dependent variable in an easy way. The WOE method is a probabilistic approach based on a log-linear form of Bayesian rule [42]. WOE is also a helpful tool for identifying the relative risk based on the available information from the related pattern [43]. It is a type of data transformation technique that encodes categorical variables for classification. WOE handles outliers and helps to construct a linear relationship with log odds. The WOE method is defined in Equation (6) [44]. It is computed by taking the natural logarithm (ln) of the division of % of non-events and % of events.
WOE = ln ( %   of   non - events %   of   events )
Given a hypothesis and some evidence, the WOE seeks to respond to the following question [45]: “How much does the evidence speak in favor of or against the hypothesis?” The WOE is usually defined for some evidence e, a hypothesis h, and its logical complement h. For example, in a simple binary classification setting, e = (X1,…, Xn), h: Y = 1, and h: Y = 0.
As a concrete example [45], suppose that a doctor wants to know whether a patient’s symptoms indicate the presence of a certain disease, say, the flu. Denote e = “the patient has a fever,” h = “the patient has the flu,” and h = “the patient doesn’t have the flu.” The doctor might know that for a patient, the odds of having the flu roughly double once the patient’s fever is taken into account (i.e., the hypothesis-odds interpretation), which corresponds to WOE(h: e) ≈ log 2 [45]. Additional information about WOE can be obtained from [45].

4.2.2. Recursive Feature Elimination (RFE)

The recursive feature elimination method is an embedded-based backward feature selection strategy [46,47]. RFE searches for a subset of features, starting with all features in the dataset and successfully removing features by fitting the specific machine learning algorithm until the specified number of features remains. This process is achieved by ranking the features in order of importance, discarding the least important features, and refitting the model through iterations. In this study, the Logistic Regression classifier is considered a specific machine learning model for the recursive feature elimination method in the feature selection stage.

4.2.3. Random Forest (RF)

The Random Forest method selects features with respect to the importance values of features by ranking them. The importance of a feature is identified as how much this attribute is used in each tree of the forest. This is defined as the normalized total reduction in the metrics provided by that feature. In other words, important features are computed as the mean and standard deviation of accumulation of the impurity decrease within each tree [31].

4.2.4. Least Absolute Shrinkage and Selection Operator (LASSO)

In statistics and machine learning, the least absolute shrinkage and selection operator is a linear model that estimates sparse coefficients of parameters. It is used for attribute selection, regularization, and regression analysis [48]. Mathematically, LASSO includes a linear model with an added regularization term. Suppose y = (y1,…, yn) T is the response vector and xj = (x1j,…, xnj) T, j = 1,…, p, are the linearly independent predictors. Let X = [x1,…, xp] be the predictor matrix. Assume the data are standardized. The LASSO estimates for the coefficients of a linear model are obtained by Equation (7) [49].
β ^ = a r g   m i n β   y j = 1 p x j β j + λ j = 1 p | β j |
where λ is called the lasso regularization parameter and β ^ is an exact unbiased estimate of the degrees of freedom of the lasso, and this result can be used to create an adaptive model selection metric for efficiently selecting the optimal LASSO fit [49]. The LASSO is a special case of the penalized least squares regression with an L1-penalty function, and if there is a high correlation in the feature groups, the LASSO method chooses only one among them and shrinks the coefficients of others to zero [50]. Thanks to this method, LASSO can contribute to both model accuracy and interpretability by removing irrelevant or insignificant features from the pattern.

4.3. Classification

In machine learning, the classification stage is a supervised learning approach that predicts corresponding class labels from the given set of examples/instances. Supervised learning algorithms use labeled datasets and outputs to train machine learning algorithms on how to predict a class label or outcome [51].
In this study, five supervised learning algorithms, namely Logistic Regression, Support Vector Machine, K Nearest Neighbors, Random Forest, and AdaBoost are employed for the machine learning tasks. We also applied a voting-based ensemble learning process to improve the classification results of these models. In the following subsections, we briefly describe these models.

4.3.1. Logistic Regression

Logistic Regression is one of the simplest and most utilized statistical models for binary classification. It is a type of regression and predictive analysis technique that is used to explain the relationship between one dependent binary variable and one or more independent variables. The Logistic Regression model gives each variable a coefficient that measures its independent contribution to variation in the dependent variable [52].

4.3.2. Support Vector Machine

Support Vector Machine is a supervised learning algorithm proposed by Vapnik et al. [53,54]. It originated from the structural risk minimization idea and is based on statistical learning theory [55]. For binary classification problems, input data are mapped to higher-dimensional space using various kernel structures. The SVM classifier explores and tries to find a hyperplane that has the maximum margin. This algorithm is very attractive, systematic, and effective for two-class and linear or non-linear classification problems [55].

4.3.3. K Nearest Neighbors

K nearest neighbors is one of the most popular, simple, effective, robust, and fundamental classification models and lazy learning algorithms [56]. It categorizes test instances into the class of the closest instances based on distance measures and the number of neighbors. In other words, the test instance is assigned to the most common class among its k nearest neighbors by using the distance function that measures the difference or similarity between instances [4,56].

4.3.4. Random Forest

Random Forest, proposed by Leo Breiman [57], is one of the most accurate, simple, easily parallelized, general-purpose machine learning predictor models [58,59]. It constructs an ensemble model with a set of decision trees that grow in randomly selected subspaces of data [58]. This bagging ensemble learning model is robust to outliers and noise and faster than bagging and boosting [59].

4.3.5. AdaBoost

AdaBoost, short for Adaptive Boosting, is a type of statistical ensemble learning method. The output of the base learners is combined into a weighted sum that represents the final output of the boosted learning model. AdaBoost focuses on instances that were previously misclassified and, in each iteration, the weights of misclassified samples are increased while the weights of correctly classified samples are decreased [26].

4.4. Ensemble Learning

Ensemble learning methods are meta-approaches that combine the results of the multiple learning algorithms to achieve better prediction final performance results by using various schemes such as bagging, boosting, voting [27,32,60], or stacking. Voting approaches can be hard or soft based on their majority rule or predicted class probabilities. Soft voting approaches provide more well-calibrated, flexible, and fine-grained results than the hard-voting-based ensemble learning scheme [32,61]. In this study, soft-voting-based ensemble learning model combination sets are applied to obtain the best classification results with feature selection methods.

5. Conclusions

This study introduces a novel hierarchical voting-based feature selection and ensemble learning methodology for glioma grading. Our study uniquely explores ensemble feature selection methods and leverages ensemble learning models. Sixteen ensemble model combinations were employed to identify the best tumor grading performance on novel constructed datasets originating in the most widely employed molecular glioma databases, TCGA and CGGA, with captured molecular and clinical characteristics. Voting-based feature selection surpassed the results of using either the LASSO method or its combination with ensemble learning models on the TCGA and CGGA datasets. Considering the widespread use of LASSO and TCGA/CGGA, an optimized limit-free approach that confidently and consistently achieves superior performance is key. Our study illustrates that optimal feature selection and learning model selection are subject to both the number of instances and number of features in large-scale datasets, and that, given existing heterogeneity, ensemble approaches are beneficial for transferability and for obtaining superior results in healthcare datasets such as TCGA and CGGA, in spite of their trade-off strategy between computational workload and performance metrics (e.g., accuracy rate). We hope that our findings will serve as a basis for the development of efficient methods for value-added care pertaining to molecular markers while improving the predictive performance of models. The proposed methodology can also be utilized as a template framework to advance optimization considering the growing and already wide array of feature selection methods and machine learning models currently available for classification, pattern recognition, and data mining tasks.
As a future direction of this study, we plan to use additional biomedical datasets (clinical, imaging, and omic) with higher-dimensional features to both leverage and compare the performance of this method with other approaches. Additional or novel machine learning predictors, instead of the models in our study, can also be employed to analyze our methodology and potentially further improve the performance results for specific large-scale medical data scenarios.

Author Contributions

E.T.: Conceptualization, Data Curation, Methodology, Software, Investigation, Writing—Original Draft Preparation, Visualization, and Writing—Review and Editing. Y.Z.: Writing—Review and Editing. H.K.: Data Curation. K.C.: Supervision and Funding Acquisition. A.V.K.: Conceptualization, Investigation, Data Curation, Supervision, Project administration, Funding Acquisition, Writing—Original Draft Preparation, Visualization, and Writing—Review and Editing. All authors have read and agreed to the published version of the manuscript.

Funding

Funding was provided in part by the NCI NIH intramural program (ZID BC 010990).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data in this manuscript has been obtained from The Cancer Genome Atlas (TCGA) Research Network (https://www.cancer.gov/tcga) and the Chinese Glioma Genome Atlas (CGGA) (http://www.cgga.org.cn/).

Acknowledgments

The results shown here are in whole or part based upon data generated by the TCGA Research Network and the Chinese Glioma Genome Atlas (CGGA). Palantir Foundry was used in the integration, harmonization, and analysis of TCGA data inside the secure NIH Integrated Data Analysis Platform (NIDAP).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ACCAccuracy
AdaBoostAdaptive Boosting
AUCArea Under the ROC Curve
CGGAChinese Glioma Genome Atlas
CNSCentral Nervous System
F1F-Measure
GBMGlioblastoma Multiforme
HGGHigh-Grade Glioma
IDHIsocitrate Dehydrogenase
KNNK Nearest Neighbors
LGGLow-Grade Glioma
LASSOLeast Absolute Shrinkage and Selection Operator
LRLogistic Regression
NCINational Cancer Institute
NIHNational Institutes of Health
PREPrecision
RECRecall
RFRandom Forest
RTRadiation Therapy
RFERecursive Feature Elimination
ROCReceiver Operating Characteristics
SPECSpecificity
SVMSupport Vector Machine
TCGAThe Cancer Genome Atlas
TMZTemozolomide
WHOWorld Health Organization
WOEWeight of Evidence

References

  1. Naser, M.A.; Deen, M.J. Brain tumor segmentation and grading of lower-grade glioma using deep learning in MRI images. Comput. Biol. Med. 2020, 121, 103758. [Google Scholar] [CrossRef] [PubMed]
  2. Pereira, S.; Meier, R.; Alves, V.; Reyes, M.; Silva, C.A. Automatic brain tumor grading from MRI data using convolutional neural networks and quality assessment. In Understanding and Interpreting Machine Learning in Medical Image Computing Applications; Springer: Berlin/Heidelberg, Germany, 2018; pp. 106–114. [Google Scholar]
  3. Zhuge, Y.; Ning, H.; Mathen, P.; Cheng, J.Y.; Krauze, A.V.; Camphausen, K.; Miller, R.W. Automated glioma grading on conventional MRI images using deep convolutional neural networks. Med. Phys. 2020, 47, 3044–3053. [Google Scholar] [CrossRef] [PubMed]
  4. Tasci, E.; Ugur, A.; Camphausen, K.; Zhuge, Y.; Zhao, R.; Krauze, A.V. 3D Multimodal Brain Tumor Segmentation and Grading Scheme based on Machine, Deep, and Transfer Learning Approaches. Int. J. Bioinform. Intell. Comput. 2022, 1, 77–95. [Google Scholar]
  5. Krauze, A. Using Artificial Intelligence and Magnetic Resonance Imaging to Address Limitations in Response Assessment in Glioma. Oncol. Insights 2022, 2022. Available online: https://mediterraneanjournals.com/index.php/oi/article/view/616 (accessed on 11 November 2022). [CrossRef]
  6. Gaillard, F. WHO Classification of CNS Tumors. Reference Article, Radiopaedia.org. Available online: https://radiopaedia.org/articles/who-classification-of-cns-tumours-1?lang=us (accessed on 2 September 2022).
  7. Hanif, F.; Muzaffar, K.; Perveen, K.; Malhi, S.M.; Simjee, S.U. Glioblastoma multiforme: A review of its epidemiology and pathogenesis through clinical presentation and treatment. Asian Pac. J. Cancer Prev. APJCP 2017, 18, 3. [Google Scholar]
  8. Mirchia, K.; Richardson, T.E. Beyond IDH-mutation: Emerging molecular diagnostic and prognostic features in adult diffuse gliomas. Cancers 2020, 12, 1817. [Google Scholar] [CrossRef]
  9. Vigneswaran, K.; Neill, S.; Hadjipanayis, C.G. Beyond the World Health Organization grading of infiltrating gliomas: Advances in the molecular genetics of glioma classification. Ann. Transl. Med. 2015, 3, 95. [Google Scholar]
  10. DeWitt, J.C.; Jordan, J.T.; Frosch, M.P.; Samore, W.R.; Iafrate, A.J.; Louis, D.N.; Lennerz, J.K. Cost-effectiveness of IDH testing in diffuse gliomas according to the 2016 WHO classification of tumors of the central nervous system recommendations. Neuro-Oncol. 2017, 19, 1640–1650. [Google Scholar] [CrossRef]
  11. Krauze, A.; Zhuge, Y.; Zhao, R.; Tasci, E.; Camphausen, K. AI-Driven Image Analysis in Central Nervous System Tumors-Traditional Machine Learning, Deep Learning and Hybrid Models. J. Biotechnol. Biomed. 2022, 5, 1–19. [Google Scholar]
  12. Diaz Rosario, M.; Kaur, H.; Tasci, E.; Shankavaram, U.; Sproull, M.; Zhuge, Y.; Camphausen, K.; Krauze, A. The Next Frontier in Health Disparities—A Closer Look at Exploring Sex Differences in Glioma Data and Omics Analysis, from Bench to Bedside and Back. Biomolecules 2022, 12, 1203. [Google Scholar] [CrossRef]
  13. Guan, D.; Yuan, W.; Lee, Y.-K.; Najeebullah, K.; Rasel, M.K. A review of ensemble learning based feature selection. IETE Tech. Rev. 2014, 31, 190–198. [Google Scholar] [CrossRef]
  14. Mitchell, T.M. Generalization as search. Artif. Intell. 1982, 18, 203–226. [Google Scholar] [CrossRef]
  15. Zanella, L.; Facco, P.; Bezzo, F.; Cimetta, E. Feature Selection and Molecular Classification of Cancer Phenotypes: A Comparative Study. Int. J. Mol. Sci. 2022, 23, 9087. [Google Scholar] [CrossRef] [PubMed]
  16. Tasci, E.; Ugur, A. A novel pattern recognition framework based on ensemble of handcrafted features on images. Multimed. Tools Appl. 2022, 81, 30195–30218. [Google Scholar] [CrossRef]
  17. Taşcı, E.; Uğur, A. Shape and texture based novel features for automated juxtapleural nodule detection in lung CTs. J. Med. Syst. 2015, 39, 46. [Google Scholar] [CrossRef]
  18. Munquad, S.; Si, T.; Mallik, S.; Li, A.; Das, A.B. Subtyping and grading of lower-grade gliomas using integrated feature selection and support vector machine. Brief. Funct. Genom. 2022, 21, 408–421. [Google Scholar] [CrossRef]
  19. Lam, L.H.T.; Do, D.T.; Diep, D.T.N.; Nguyet, D.L.N.; Truong, Q.D.; Tri, T.T.; Thanh, H.N.; Le, N.Q.K. Molecular subtype classification of low-grade gliomas using magnetic resonance imaging-based radiomics and machine learning. NMR Biomed. 2022, 35, e4792. [Google Scholar] [CrossRef]
  20. Li, L.; Wei, Y.; Shi, G.; Yang, H.; Li, Z.; Fang, R.; Cao, H.; Cui, Y. Multi-omics data integration for subtype identification of Chinese lower-grade gliomas: A joint similarity network fusion approach. Comput. Struct. Biotechnol. J. 2022, 20, 3482–3492. [Google Scholar] [CrossRef]
  21. Munquad, S.; Si, T.; Mallik, S.; Das, A.B.; Zhao, Z. A deep learning–based framework for supporting clinical diagnosis of glioblastoma subtypes. Front. Genet. 2022, 13, 855420. [Google Scholar] [CrossRef]
  22. Cawood, P.; van Zyl, T. Evaluating State of the Art, Forecasting Ensembles-and Meta-learning Strategies for Model Fusion. arXiv 2022, arXiv:2203.03279. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Liu, J.; Shen, W. A Review of Ensemble Learning Algorithms Used in Remote Sensing Applications. Appl. Sci. 2022, 12, 8654. [Google Scholar] [CrossRef]
  24. McGowan, C.J.; Biggerstaff, M.; Johansson, M.; Apfeldorf, K.M.; Ben-Nun, M.; Brooks, L.; Convertino, M.; Erraguntla, M.; Farrow, D.C.; Freeze, J. Collaborative efforts to forecast seasonal influenza in the United States, 2015–2016. Sci. Rep. 2019, 9, 683. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Reich, N.G.; Brooks, L.C.; Fox, S.J.; Kandula, S.; McGowan, C.J.; Moore, E.; Osthus, D.; Ray, E.L.; Tushar, A.; Yamana, T.K. A collaborative multiyear, multimodel assessment of seasonal influenza forecasting in the United States. Proc. Natl. Acad. Sci. USA 2019, 116, 3146–3154. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  27. Tasci, E. Voting combinations-based ensemble of fine-tuned convolutional neural networks for food image recognition. Multimed. Tools Appl. 2020, 79, 30397–30418. [Google Scholar] [CrossRef]
  28. Colkesen, I.; Ozturk, M.Y. A comparative evaluation of state-of-the-art ensemble learning algorithms for land cover classification using WorldView-2, Sentinel-2 and ROSIS imagery. Arab. J. Geosci. 2022, 15, 942. [Google Scholar] [CrossRef]
  29. Ngo, G.; Beard, R.; Chandra, R. Evolutionary bagging for ensemble learning. Neurocomputing 2022, 510, 1–14. [Google Scholar] [CrossRef]
  30. Tasci, E.; Ugur, A. Image classification using ensemble algorithms with deep learning and hand-crafted features. In Proceedings of the 2018 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018; pp. 1–4. [Google Scholar]
  31. Feature Importances with a Forest of Trees. Available online: https://scikit-learn.org/stable/auto_examples/ensemble/plot_forest_importances.html (accessed on 19 August 2022).
  32. Tasci, E.; Uluturk, C.; Ugur, A. A voting-based ensemble deep learning method focusing on image augmentation and preprocessing variations for tuberculosis detection. Neural Comput. Appl. 2021, 33, 15541–15555. [Google Scholar] [CrossRef]
  33. Scikit-Learn. Available online: https://scikit-learn.org/stable/ (accessed on 25 August 2022).
  34. Xverse. Available online: https://pypi.org/project/xverse/ (accessed on 25 August 2022).
  35. Zhao, Z.; Zhang, K.-N.; Wang, Q.; Li, G.; Zeng, F.; Zhang, Y.; Wu, F.; Chai, R.; Wang, Z.; Zhang, C. Chinese Glioma Genome Atlas (CGGA): A comprehensive resource with functional genomic data from Chinese glioma patients. Genom. Proteom. Bioinform. 2021, 19, 1–12. [Google Scholar] [CrossRef]
  36. Palantir Foundry—The NIH Integrated Data Analysis Platform (NIDAP); NCI Center for Biomedical Informatics & Information Technology (CBIIT); Software Provided by Palantir Technologies Inc. Available online: https://www.palantir.com (accessed on 2 September 2022).
  37. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  38. Mandrekar, J.N. Receiver operating characteristic curve in diagnostic test assessment. J. Thorac. Oncol. 2010, 5, 1315–1316. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Gokalp, O.; Tasci, E.; Ugur, A. A novel wrapper feature selection algorithm based on iterated greedy metaheuristic for sentiment classification. Expert Syst. Appl. 2020, 146, 113176. [Google Scholar] [CrossRef]
  40. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  41. Tasci, E.; Zhuge, Y.; Camphausen, K.; Krauze, A.V. Bias and Class Imbalance in Oncologic Data—Towards Inclusive and Transferrable AI in Large Scale Oncology Data Sets. Cancers 2022, 14, 2897. [Google Scholar] [CrossRef] [PubMed]
  42. Chen, W.; Li, H.; Hou, E.; Wang, S.; Wang, G.; Panahi, M.; Li, T.; Peng, T.; Guo, C.; Niu, C. GIS-based groundwater potential analysis using novel ensemble weights-of-evidence with logistic regression and functional tree models. Sci. Total Environ. 2018, 634, 853–867. [Google Scholar] [CrossRef]
  43. Zdravevski, E.; Lameski, P.; Kulakov, A. Weight of evidence as a tool for attribute transformation in the preprocessing stage of supervised learning algorithms. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 181–188. [Google Scholar]
  44. Zeng, G. A necessary condition for a good binning algorithm in credit scoring. Appl. Math. Sci. 2014, 8, 3229–3242. [Google Scholar] [CrossRef]
  45. Alvarez-Melis, D.; Kaur, H.; Daumé III, H.; Wallach, H.; Vaughan, J.W. From human explanation to model interpretability: A framework based on weight of evidence. In Proceedings of the AAAI Conference on Human Computation and Crowdsourcing (HCOMP), Virtual, 14–18 November 2021. [Google Scholar]
  46. Stambaugh, C.; Yang, H.; Breuer, F. Analytic feature selection for support vector machines. In Proceedings of the International Workshop on Machine Learning and Data Mining in Pattern Recognition, New York, NY, USA, 19–25 July 2013; pp. 219–233. [Google Scholar]
  47. Escanilla, N.S.; Hellerstein, L.; Kleiman, R.; Kuang, Z.; Shull, J.; Page, D. Recursive feature elimination by sensitivity testing. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 40–47. [Google Scholar]
  48. Muthukrishnan, R.; Rohini, R. LASSO: A feature selection technique in predictive modeling for machine learning. In Proceedings of the 2016 IEEE international conference on advances in computer applications (ICACA), Coimbatore, India, 24 October 2016; pp. 18–20. [Google Scholar]
  49. Zou, H.; Hastie, T.; Tibshirani, R. On the “degrees of freedom” of the lasso. Ann. Stat. 2007, 35, 2173–2192. [Google Scholar] [CrossRef]
  50. Lasso. Available online: https://scikit-learn.org/stable/modules/linear_model.html#lasso (accessed on 19 August 2022).
  51. Jiang, T.; Gradus, J.L.; Rosellini, A.J. Supervised machine learning: A brief primer. Behav. Ther. 2020, 51, 675–687. [Google Scholar] [CrossRef]
  52. Boateng, E.Y.; Abaye, D.A. A review of the logistic regression model with emphasis on medical research. J. Data Anal. Inf. Process. 2019, 7, 190–207. [Google Scholar] [CrossRef] [Green Version]
  53. Cristianini, N.; Ricci, E. Support vector machines. Encycl. Algorithm 2008, 13, 18–28. [Google Scholar] [CrossRef]
  54. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; pp. 144–152. [Google Scholar]
  55. Othman, M.F.B.; Abdullah, N.B.; Kamal, N.F.B. MRI brain classification using support vector machine. In Proceedings of the 2011 Fourth International Conference on Modeling, Simulation and Applied Optimization, Kuala Lumpur, Malaysia, 19–21 April 2011; pp. 1–4. [Google Scholar]
  56. Jiang, L.; Cai, Z.; Wang, D.; Jiang, S. Survey of improving k-nearest-neighbor for classification. In Proceedings of the Fourth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD 2007), Haikou, China, 24–27 August 2007; pp. 679–683. [Google Scholar]
  57. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  58. Biau, G. Analysis of a random forests model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar]
  59. Tasci, E. A meta-ensemble classifier approach: Random rotation forest. Balk. J. Electr. Comput. Eng. 2019, 7, 182–187. [Google Scholar] [CrossRef] [Green Version]
  60. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  61. Voting Classifier. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.VotingClassifier.html (accessed on 31 August 2022).
Figure 1. The distribution of the most frequently mutated genes for glioma data on TCGA.
Figure 1. The distribution of the most frequently mutated genes for glioma data on TCGA.
Ijms 23 14155 g001
Figure 2. Graphical representations of the computational results for the TCGA and CGGA datasets.
Figure 2. Graphical representations of the computational results for the TCGA and CGGA datasets.
Ijms 23 14155 g002
Figure 3. Overview of the proposed methodology.
Figure 3. Overview of the proposed methodology.
Ijms 23 14155 g003
Table 1. The set numbers and the related learning model combinations for the voting process.
Table 1. The set numbers and the related learning model combinations for the voting process.
Set #Model Combination
1LR + SVM + KNN
2LR + SVM + RF
3LR + SVM + AdaBoost
4LR + KNN + RF
5LR + KNN + AdaBoost
6LR + RF + AdaBoost
7SVM + KNN + RF
8SVM + KNN + AdaBoost
9SVM + RF + AdaBoost
10KNN + RF + AdaBoost
11LR + SVM + KNN + RF
12LR + SVM + KNN + AdaBoost
13LR + SVM + RF + AdaBoost
14LR + KNN + RF + AdaBoost
15SVM + KNN + RF + AdaBoost
16LR + SVM + KNN + RF + AdaBoost
Table 2. Features and class information for the datasets used. TCGA has 23 class features (3 clinical, 20 molecular), whereas CGGA has 22 given it is comprised of a Chinese population with race not explicitly specified in the database.
Table 2. Features and class information for the datasets used. TCGA has 23 class features (3 clinical, 20 molecular), whereas CGGA has 22 given it is comprised of a Chinese population with race not explicitly specified in the database.
#TypeName#TypeName#TypeName
1ClinicalGender9MolecularCIC17MolecularBCOR
2ClinicalAge_at_diagnosis10MolecularMUC1618MolecularCSMD3
3ClinicalRace11MolecularPIK3CA19MolecularSMARCA4
4MolecularIDH112MolecularNF120MolecularGRIN2A
5MolecularTP5313MolecularPIK3R121MolecularIDH2
6MolecularATRX14MolecularFUBP122MolecularFAT4
7MolecularPTEN15MolecularRB123MolecularPDGFRA
8MolecularEGFR16MolecularNOTCH124ClassGrade
Table 3. Computational results for the TCGA dataset with the voting-based feature selection process.
Table 3. Computational results for the TCGA dataset with the voting-based feature selection process.
Set #ACCAUCF1PRERECSPEC
10.8690.9140.8520.7980.9220.835
20.8700.9170.8530.8010.9180.838
30.8700.9200.8540.7980.9250.835
40.8610.9120.8380.8070.8780.850
50.8680.9130.8500.8000.9130.839
60.8580.9170.8340.8070.8700.870
70.8680.9090.8480.8090.8980.847
80.8710.9110.8540.8030.9180.840
90.8760.9140.8580.8150.9120.852
100.8450.9080.8150.8040.8350.854
110.8740.9130.8560.8080.9180.844
120.8690.9150.8520.7980.9220.835
130.8700.9180.8530.8010.9180.838
140.8620.9140.8390.8090.8780.852
150.8680.9110.8480.8090.8980.847
160.8750.9140.8580.8080.9210.844
Bold values indicate the best results.
Table 4. Computational results for the CGGA dataset with the voting-based feature selection process.
Table 4. Computational results for the CGGA dataset with the voting-based feature selection process.
Set #ACCAUCF1PRERECSPEC
10.7760.7870.5970.7120.5430.910
20.7900.8010.6460.7460.6000.897
30.7860.8070.6420.7320.6020.891
40.7930.7800.6460.7740.5830.915
50.7760.7760.5960.7340.5350.917
60.7690.7880.6110.7180.5610.885
70.7930.7850.6430.7660.5850.913
80.7760.7900.6000.7020.5510.905
90.7930.7980.6500.7390.6060.898
100.7650.7640.6050.7070.5630.882
110.7830.7880.6280.7480.5760.903
120.7760.7890.5970.7120.5430.910
130.7900.8020.6460.7460.6000.897
140.7900.7830.6380.7740.5750.915
150.7970.7880.6480.7640.5930.914
160.7830.7910.6280.7480.5760.903
Bold values indicate the best results.
Table 5. The mean number of selected features after applying voting-based feature selection.
Table 5. The mean number of selected features after applying voting-based feature selection.
DatasetTCGACGGA
Total # of Features2322
Selected # of Features14.917.6
Table 6. The computational results without the feature selection process for the best ensemble set.
Table 6. The computational results without the feature selection process for the best ensemble set.
DatasetModelACCAUCF1PRERECSPEC
TCGASVM + RF + AdaBoost0.8640.9140.8420.8070.8850.850
CGGASVM + KNN + RF + AdaBoost0.7760.8040.6160.7750.5500.911
Bold values indicate the best results.
Table 7. Computational results for TCGA dataset with LASSO feature selection method and individual models.
Table 7. Computational results for TCGA dataset with LASSO feature selection method and individual models.
ModelACCAUCF1PRERECSPEC
LR0.8710.9200.8530.8080.9110.847
SVM0.8710.9090.8540.8010.9210.838
KNN0.8390.9010.8090.7870.8380.840
RF0.8250.9020.7920.7800.8130.836
AdaBoost0.8630.9140.8420.8030.8950.845
Bold values indicate the best results.
Table 8. Computational results for CGGA dataset with LASSO feature selection method and individual models.
Table 8. Computational results for CGGA dataset with LASSO feature selection method and individual models.
ModelACCAUCF1PRERECSPEC
LR0.7800.7930.6490.7300.7300.887
SVM0.7860.8170.6390.7810.5660.916
KNN0.7440.7520.5550.6820.4970.888
RF0.7550.7880.6130.6870.5830.858
AdaBoost0.7480.7720.6000.6530.5720.849
Bold values indicate the best results.
Table 9. Computational results for TGGA dataset with LASSO feature selection method and ensemble learning models.
Table 9. Computational results for TGGA dataset with LASSO feature selection method and ensemble learning models.
Set #Voting of Model CombinationACCAUCF1PRERECSPEC
1LR + SVM + KNN0.8670.9210.8470.8030.9010.844
2LR + SVM + RF0.8740.9210.8550.8080.9140.847
3LR + SVM + AdaBoost0.8700.9230.8520.8020.9150.841
4LR + KNN + RF0.8620.9180.8380.8080.8760.852
5LR + KNN + AdaBoost0.8650.9200.8450.8060.8920.848
6LR + RF + AdaBoost0.8580.9210.8340.8070.8700.853
7SVM + KNN + RF0.8580.9140.8340.8020.8730.848
8SVM + KNN + AdaBoost0.8670.9160.8460.8080.8920.851
9SVM + RF + AdaBoost0.8670.9140.8440.8090.8870.853
10KNN + RF + AdaBoost0.8460.9090.8160.8020.8370.855
11LR + SVM + KNN + RF0.8630.9190.8400.8040.8850.848
12LR + SVM + KNN + AdaBoost0.8680.9210.8490.8060.9010.846
13LR + SVM + RF + AdaBoost0.8740.9210.8550.8080.9140.847
14LR + KNN + RF + AdaBoost0.8630.9180.8390.8100.8760.854
15SVM + KNN + RF + AdaBoost0.8610.9140.8370.8050.8760.850
16LR + SVM + KNN + RF + AdaBoost0.8650.9190.8430.8070.8880.850
Bold values indicate the best results.
Table 10. Computational results for CGGA dataset with LASSO feature selection method and ensemble learning models.
Table 10. Computational results for CGGA dataset with LASSO feature selection method and ensemble learning models.
Set #Voting of Model CombinationACCAUCF1PRERECSPEC
1LR + SVM + KNN0.7830.8060.6360.7610.5670.908
2LR + SVM + RF0.7790.8180.6370.7520.5770.897
3LR + SVM + AdaBoost0.7760.8100.6370.7390.5870.884
4LR + KNN + RF0.7790.8050.6290.7490.5680.903
5LR + KNN + AdaBoost0.7720.7980.6190.7360.5600.897
6LR + RF + AdaBoost0.7720.8100.6290.7290.5760.885
7SVM + KNN + RF0.7680.8030.6190.7220.5700.887
8SVM + KNN + AdaBoost0.7720.7970.6130.7300.5580.899
9SVM + RF + AdaBoost0.7860.8110.6480.7520.5940.898
10KNN + RF + AdaBoost0.7720.7940.6200.7150.5730.889
11LR + SVM + KNN + RF0.7760.8110.6280.7390.5690.897
12LR + SVM + KNN + AdaBoost0.7790.8080.6330.7530.5670.902
13LR + SVM + RF + AdaBoost0.7790.8160.6370.7520.5780.897
14LR + KNN + RF + AdaBoost0.7790.8030.6290.7490.5680.903
15SVM + KNN + RF + AdaBoost0.7680.8020.6190.7220.5700.887
16LR + SVM + KNN + RF + AdaBoost0.7760.8110.6280.7390.5690.897
Bold values indicate the best results.
Table 11. The mean number of selected features after applying LASSO feature selection.
Table 11. The mean number of selected features after applying LASSO feature selection.
DatasetTCGACGGA
Total # of Features2322
Selected # of Features20.511.9
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tasci, E.; Zhuge, Y.; Kaur, H.; Camphausen, K.; Krauze, A.V. Hierarchical Voting-Based Feature Selection and Ensemble Learning Model Scheme for Glioma Grading with Clinical and Molecular Characteristics. Int. J. Mol. Sci. 2022, 23, 14155. https://doi.org/10.3390/ijms232214155

AMA Style

Tasci E, Zhuge Y, Kaur H, Camphausen K, Krauze AV. Hierarchical Voting-Based Feature Selection and Ensemble Learning Model Scheme for Glioma Grading with Clinical and Molecular Characteristics. International Journal of Molecular Sciences. 2022; 23(22):14155. https://doi.org/10.3390/ijms232214155

Chicago/Turabian Style

Tasci, Erdal, Ying Zhuge, Harpreet Kaur, Kevin Camphausen, and Andra Valentina Krauze. 2022. "Hierarchical Voting-Based Feature Selection and Ensemble Learning Model Scheme for Glioma Grading with Clinical and Molecular Characteristics" International Journal of Molecular Sciences 23, no. 22: 14155. https://doi.org/10.3390/ijms232214155

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop