Next Article in Journal
A Novel Hybrid Path Planning Method Based on Q-Learning and Neural Network for Robot Arm
Next Article in Special Issue
A Two-Phase Evolutionary Method to Train RBF Networks
Previous Article in Journal
Credit Card Fraud Detection in Card-Not-Present Transactions: Where to Invest?
Previous Article in Special Issue
Optimization of Truck-Drone Parcel Delivery Using Metaheuristics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Outlier Detection Based Feature Selection Exploiting Bio-Inspired Optimization Algorithms

by
Souad Larabi-Marie-Sainte
Computer Science Department, College of Computer and Information Sciences, Prince Sultan University, P.O. Box 53073, Riyadh 11586, Saudi Arabia
Appl. Sci. 2021, 11(15), 6769; https://doi.org/10.3390/app11156769
Submission received: 25 June 2021 / Revised: 15 July 2021 / Accepted: 20 July 2021 / Published: 23 July 2021

Abstract

:
The curse of dimensionality problem occurs when the data are high-dimensional. It affects the learning process and reduces the accuracy. Feature selection is one of the dimensionality reduction approaches that mainly contribute to solving the curse of the dimensionality problem by selecting the relevant features. Irrelevant features are the dependent and redundant features that cause noise in the data and then reduce its quality. The main well-known feature-selection methods are wrapper and filter techniques. However, wrapper feature selection techniques are computationally expensive, whereas filter feature selection methods suffer from multicollinearity. In this research study, four new feature selection methods based on outlier detection using the Projection Pursuit method are proposed. Outlier detection involves identifying abnormal data (irrelevant features of the transpose matrix obtained from the original dataset matrix). The concept of outlier detection using projection pursuit has proved its efficiency in many applications but has not yet been used as a feature selection approach. To the author’s knowledge, this study is the first of its kind. Experimental results on nineteen real datasets using three classifiers (k-NN, SVM, and Random Forest) indicated that the suggested methods enhanced the classification accuracy rate by an average of 6.64% when compared to the classification accuracy without applying feature selection. It also outperformed the state-of-the-art methods on most of the used datasets with an improvement rate ranging between 0.76% and 30.64%. Statistical analysis showed that the results of the proposed methods are statistically significant.

1. Introduction

The data is continuously evolving, leading to a big challenge in several domains. It is generally high-dimensional, which causes the curse of the dimensionality problem. In this study, we investigated this problem with a high number of attributes (features). Reducing the space of features is performed through feature extraction or feature selection (FS). Feature extraction transforms the initial feature space into a novel low-dimensional feature space, for example, Kernel Dimensionality Reduction [1], Principal Component Analysis (PCA) [2], and Kernel PCA [3]. While FS does not produce new features but employs the original features to select the best ones, it consists of choosing a subgroup of significant features that mainly contribute to enhancing the output. FS is an essential preprocessing step in clustering and classification techniques to improve data quality. The effectiveness of FS method is one of the crucial aspects that affect the accuracy outcomes. The goal of the algorithm is to minimize the cardinality of the features set. Finding the most relevant features is not an easy task. For a dataset with a dimension d, there are 2 d 1 potential subsets of features (excluding the empty subset) [4]. For even a reasonable value of d, it is of interest to choose a less computationally expensive method. FS is required in numerous domains and applications, for example, in bio-informatics [5,6], where the data can hold a large number of variables, most of wihich might be extremely well correlated with other variables. These variables are called dependent variables and do not deliver further information in the classification process, but they are considered as a noise for the classifier, which requires removing them to enhance the data quality. Another example in the biomedical domain wherein protein interaction discovery is vitally needed [7]. In biometric identification, FS is requested in facial recognition [8] to diminish the data dimensionality and then increase the classification accuracy. In addition, text mining suffers from the curse of dimensionality, and some FS methods were addressed in text categorization [9] and text clustering [10].
FS methods are classified into four categories, as displayed in Figure 1, called wrapper, filter, embedded, and outlier detection methods. Wrapper methods consist of evaluating features based on their predictive power using a classifier and usually a cross-validation technique. Several wrapper techniques have been proposed to solve the curse of dimensionality problem. They are based on a specific search method [11], such as the sequential selection [12], meta-heuristic (for example [13,14]), and local search [15]. The filter method is based on evaluating features using a criterion function that is relatively independent of classification such as distance [16], information [17], and dependency [18]. It could be based on information gain, Fisher score, variance threshold, chi-square test, correlation coefficient, and others [11]. Most (if not all) of the filter FS methods can be found in several statistical and ML software such that Weka, R, and SPSS. The embedded methods consist of applying FS in the learning process. The data are not divided into training and testing sets because they do not optimize an objective function or performance of a machine learning algorithm or a model like wrapper methods do. However, they focus on the use of a basic model constructing metrics during the learning. These methods are mainly based on the Least Absolute Shrinkage and Selection Operator (Lasso) technique combined with the Linear Regression method (for example, [19,20]). The other approaches can comprise outlier detection, which consist of detecting and removing abnormal features in order to reduce the dimensionality and enhance the quality.
The aforementioned FS methods have some limitations. The wrapper FS are efficient but time-consuming [21,22]. The filter FS techniques are efficient and not computationally expensive. However, they are affected by the multicollinearity [23,24]. The embedded FS methods are not well investigated compared to wrapper and filter FS methods due to their complex structure. The limitations of FS methods are deeply illustrated in Section 3.
To overcome these limitations, this article proposes novel FS techniques stimulated by outlier detection where the outliers cover a small part of the dataset.
Outlier detection involves searching for data objects that are different from standard data [25]. It is mainly required in numerous applications, including fraud detection, image processing, intrusion detection, public safety and security, and industry damage detection.
In this article, the irrelevant features represent the outliers, and the dataset is represented as a matrix. The matrix is then transposed so that the rows represent the features (the attributes) and the columns stand for the samples. Thus, the outliers will be found in the samples of the transposed dataset, which are the features set of the original dataset.
Outlier detection is performed using the Projection Pursuit (PP). The PP approach comprises looking for low dimensional linear projections that reveal potential outlying observations. Finding interesting projections is performed through the optimization of a specific function called a projection index (or PP index). The concept of this proposal is defined in Section 4. PP has successfully been applied to the search for outliers [26,27], which is performed by assuming data normality and that the data not following the model are outliers.
The present study focuses on developing FS based on outlier detection using PP called Projection Pursuit based Feature Selection (PPFS). To the author’s knowledge, only one study used outlier detection for FS [28] but without PP. It involved feature ranking and isolation forest, which is different from this proposal. Therefore, this proposal is a new of its kind. The contribution of PPFS is described in the following section along with the research objectives.
To validate this model, nineteen real-word datasets obtained from repositories are used along with three classifiers, including the k-Nearest Neighbors, Support Vector Machine, and Random Forest. Four classification measures are utilized, namely the accuracy, precision, F1 measure, and recall to confirm the results and compare them with those obtained with the state-of-the-art techniques.
The article is structured as follows. Section 2 illustrates the contributions and the research objectives of this study. Section 3 surveys the existing works related to the four FS categories and addresses their limitations. Section 4 describes PP and the PP indices used in this study. Section 5 discusses the optimization methods applied to the PP indices. Section 6 presents the methodology followed in this study. Section 7 investigates the experimentation results, including the comparison study along with a deep discussion illustrating the results, the advantages and the limitations of this study, in addition to the research impacts. Section 8 concludes this work and presents some perspectives and future work.

2. Contributions and Research Objectives

The contribution is fourfold:
  • Suggesting a new FS method based on PP, a well-known outlier detection technique.
  • Providing a one-stage approach based on detecting and removing outliers that correspond to the irrelevant features.
  • Discovering relevant features using optimization methods based on common bio-inspired algorithms, algorithms (GA) and a variant of PSO (Tribes).
  • Providing a software-based outlier-detection method to determine and remove the non-relevant features.
This study aims at achieving the following objectives.
  • Investigate the limitations of the existing FS techniques.
  • Develop a new model for FS based on PP (called PPFS) to enhance the classification accuracy.
  • Propose four different PPFS methods based on two optimization methods (GA and Tribes) and two PP indices (Kurtosis and Friedman).
  • Suggest the use of EPP-Lab to select the non-relevant features.
  • Investigate a statistical analysis to show that the four PPFS methods are equivalent and their results are statistically significant compared to the state-of-the-art methods.
  • Perform a comparison study with the state-of-the-art methods to confirm the efficiency of the proposed model.

3. Review of Related Literature

This section describes the recent related work regarding the four FS categories. Four subsections were presented, Wrapper FS methods, Filter FS, Embedded FS, and outlier FS methods.

3.1. Wrapper Feature Selection Methods

In the following, the most recent wrapper FS are discussed.
For the wrapper FS based meta-heuristics, the authors in [13] developed a new FS based on a Quantum Whale Optimization Algorithm. Four classifiers were used: Decision Tree (C4.5), Support Vector Machine with a linear kernel, k-Nearest Neighbors (k-NN), and Linear discriminant classifier with 13 datasets. The results revealed that the proposed method performed better than the state-of-the art methods, with an average accuracy rate achieving 88.57%. While in [14], the authors proposed a new FS using Whale optimization algorithm, the k-NN classifier was utilized to assess the effectiveness of the suggested method using different UCI datasets. A comparison study was addressed using GA, PSO, and Ant Lion optimizer wrapper FS methods along with five standard filter FS techniques. The results indicated that the mutation enhanced the results for most datasets.
In sequential search, the authors of [12] proposed sequential forward search (SRS) and sequential backward search (SBS) approaches combined with two optimization methods called AdaBoost and Bagging algorithms. The Decision Tree and the Naïve Bayes classifiers along with thirteen UCI datasets were used to assess the model using ten-fold cross validation. The results showed that the SFS combined with the Bagging technique achieved a higher accuracy of 89.60%.
For the Wrapper FS based local search, the authors of [15] proposed an FS method based on Greedy Search (IG) for Sentiment analysis. The Multi-nomial Naïve Bayes classifier was applied to four Amazon reviews and nine public datasets. The classification accuracy rate reached an average of 90.74% and 96.45% for each type of datasets, respectively.

3.2. Filter Feature Selection Methods

In the following, the most recent filter FS studies from the three categories (distance, information, and dependency) are presented.
The authors of [16] proposed a new filter FS technique based on the dependency concept using Discriminative Correlation Filter (DCF) for visual object tracking. Many benchmark datasets were used to measure the effectiveness of the proposed method, along with the Deep Neural Network. The comparison results confirmed the dominance of the proposed FS over the existing techniques. The study [18] used the concept of distance. The authors suggested a new filter FS using an artificial immune system. They minimized the intra-class distance and maximized the inter-class distance in the defined regions. Then, a comparison study was successfully performed with a global and local FS algorithms using UCI and synthetic datasets. The obtained results were encouraging. Another study mainly based on the information concept [17]. It consists of analyzing several filter FS methods. Twenty-two filter FS were presented such as oneR, Variance, permutation, Info gain, Gain ratio, and others. The authors compared their performance in terms of predictive accuracy and run time. They applied the k-NN, the Logistic Ridge Regression, and SVM classifiers to determine the accuracy of each filter FS method. They concluded that there is no best filter FS method for all the datasets.

3.3. Embedded Feature Selection Methods

In the following, the most recent embedded FS studies are presented.
The authors in [19] suggested a new embedded FS method related to within-class and the least-squares loss function. This study aimed at recommending a treatment to avoid or decelerate the progress of dementia disease. An SVM classifier based on the one-versus-one class was used. In addition, ten-fold cross-validation was reiterated for each classification. The Medical Reasoning Image (MRI) dataset was utilized. The proposed embedded method reduced the number of features to 24 over 90 reaching a classification accuracy of 81.9%. In another study [20], the authors developed a new embedded FS method to transform the multi-label FS problem into several single-label ones considering the correlation among labels as an evaluation measure. The authors used five bioinformatics datasets. Three classifiers based on multi-label were utilized, including Multi-Label k-NN, Binary Relevance, and Classifier Chain. The proposed methods were compared to X 2 -based maximal, Multi-label embedded FS, and filter multi-label FS. The proposed method surpassed the compared methods.

3.4. Outlier Feature Selection Methods

To the author’s knowledge, only one research study has been found that conducted on reducing the dimensionality by eliminating the outliers. In other words, they performed the outliers’ detection and elimination in the feature set [28]. The concept was based on feature ranking and isolation forest. The one-class SVM, the local outlier factor and IFOR were employed. In addition, several synthetic and real-world datasets were utilized to validate the efficiency of the proposed FS. A comparison study was introduced using several FS and outlier detection methods such as Laplacian, variance, and kurtosis. The results showed that the proposed FS was similar to kurtosis but better than Laplacian and the variance. Many other studies combined FS to outlier detection to strengthen the FS approach. For example, the authors in [29] suggested a novel FS method combined with an outlier detection technique and classification to identify the attacks. Other studies used FS methods to detect outliers. For example, the authors of [30] aimed to develop an outlier detection approach based on FS and association rules. In [31], the authors proposed a new FS method for unsupervised outlier detection. The authors used the inter and intra distance along with recursive backward elimination to extract the outliers.

3.5. Discussion

Many wrapper, filter, and embedded FS were proposed. However, several studies addressed their limitations:
  • The wrapper FS methods are time consuming [21,22] because the learning steps and the cross-validation are repeated many times.
  • The filter FS methods suffer from the multicollinearity; only one feature is randomly selected from a set of highly correlated features. This selection might negatively affect the result, for example, in gene expression [23,24].
  • The embedded FS methods are not well investigated due to their complex structure.
  • The hybrid wrapper-filter FS techniques [12,32] are two-stage approaches and thus require more knowledge in addition to their time-consuming nature.
  • The only work presenting the outlier-based FS method [28] used feature ranking and isolation forest. It is a random-based approach that makes the attributes selection and the location of the best split for building the isolation forest unclear [33].
  • The combination of wrapper FS with outlier detection [21,29] is a two-stage approach that works by reducing the features and then removing the outliers. This approach is time consuming due to the limitation of wrapper FS. In addition, the knowledge of wrapper FS and outlier detection are both required.
  • Some studies [30,31,34] performed the outlier detection using FS methods, which is the opposite of this proposed study.
To show the significance of this study, the difference between the PPFS methods and the related works is described in Table 1.

4. Projection Pursuit

PP belongs to the same family of the Principal Component Analysis (PCA) method. The central target of PP is to look for linear projections (combinations of variables) that reveal interesting structures in multidimensional and large data [35,36]. These projections are in one, two, or three dimensions. The interesting projections are found through the optimization of the projection index using a robust optimization algorithm capable of determining the optimum as well as the local optima that can also reveal hidden interesting projections [27]. The PP indices are categorized according to the search purpose, for example, detecting outliers or finding clusters. In this research study, the interesting projections are those that detect outliers. Huber [35] and Jones and Sibson [36] showed that the Gaussian distribution is the least interesting one, while the most interesting structures are those that bear a minimum resemblance to the Gaussian distribution. Sometimes the outliers can form a small cluster, PP can also identify unbalanced clusters, and it considers the smallest one as outliers.
Before going through the details of PP, it is necessary to introduce the principal idea of this study. The concept of this proposal is to detect the outliers in the feature set instead of the data objects. To do this, the dataset matrix must be transposed. Given the mathematical definition, let us denote by X an ( P × N ) dataset matrix, and X an ( N × P ) transposed dataset matrix, where the rows represent the features (N features) and the columns stand for the observations or the samples (P observations). It is worth knowing that the transposed matrix is used in PP (instead of the original dataset matrix) to find out the optimal projection that reveal the outliers (i.e., the irrelevant features). This research study considers a one-dimensional projection that can be defined as P-dimensional vector called a. The coordinates of the projected data are presented as an N-dimensional vector w such that
w = X a
where X is the transposed dataset matrix and a is the projection vector.
Therefore, finding the projection vector a is just optimizing the PP index I ( a ) . For more details about PP, one can refer to [27,37].
In this research work, two well-known PP indices dedicated to the search of outliers [27,38] are proposed, including the kurtosis index [39] and Friedman index [40]. The detection of the outliers is univariate.

4.1. Friedman Index

This index pertains to the polynomial based family indices. It measures the difference between the Gaussian distribution and the distribution of the projected data using expansions based on orthogonal polynomials introduced by Friedman [40]. The maximization of this index leads to finding the optimal projection. It is expressed as follows:
I ( a ) = j = 1 m [ 2 j + 1 2 ] 1 / N i = 1 N [ L j [ 2 ϕ ( a T X i ) 1 ] 2
where the recursive definition of Legendre polynomials L j is: L 0 ( r ) = 1 , L 1 ( r ) = r , L 2 ( r ) = 1 2 ( 3 r 2 1 ) , L j ( r ) = 1 j ( 2 j 1 ) r L j 1 ( r ) ( j 1 ) L j 2 ( r ) for j = 3 ,
The number of terms m is set to 3 according to the recommendations in [40,41]. ϕ be the cumulative distribution function of an N ( 0 , 1 ) random variable.
X is the transposed dataset matrix and N is the number of samples. X i is a P-dimensional vector representing the sample i of the transposed dataset matrix X , i = 1 , , N and P is the number of features.

4.2. Kurtosis Index

This index is the fourth moment of the projected data proposed by Peña and Prieto [39]. The authors stressed that maximizing the kurtosis index of the projected data leads to identifying outliers. This index is based on the bimodality that split the data into two clusters. Maximizing this index means getting unbalanced clusters, the smallest one is considered as outliers. It is defined as:
I ( a ) = i = 1 N ( a T X i ) 4
where X i is a P-dimensional vector representing the sample i of the transposed dataset matrix X , i = 1 , , N , N is the number of samples and P is the number of features. a is the projection vector.

5. Optimization Methods

To optimize the PP indices defined above, bio-inspired algorithms are used because of their success in finding different optimal values of the indices. GA and Tribes are investigated. These techniques are very suitable in the resolution of PP [27], and they are implemented in EPP-Lab [38]. The investigated methods are described below. For more details, one can refer to [38].

5.1. Genetic Algorithms

Genetic Algorithm (GA) is an evolutionary algorithm inspired by the natural evolution developed by Holland [42]. GA relies on three bio-inspired operators, including selection, crossover, and mutation. It involves a population of candidate solutions called individuals. An individual (or a solution) is a set of genes. In this study, it is characterized by an array of real numbers. More precisely, it represents the projection vector. A population is a set of individuals. The individuals are randomly initialized, searching for better solutions (close to the optimum) in the search space. At each iteration, every individual is assessed based on an objective function (PP index in this study). Then, the most suitable individuals are designated from the existing population using the tournament selection with three individuals. After that, the individuals are modified by applying the recombination and mutation operators to produce a novel population. The recombination consists of merging the information of two existing individuals (parents) to produce new individuals (offspring). In this study, two-point crossover is operated in the whole population with a probability equal to 0.65. It involves randomly selecting one part (of equal-size) from two parents to generate two offsprings. The two offsprings will have the same genes (real numbers) as the two parents, except for those of the selected part. Later, all the individuals are mutated with a probability equal to 0.05 by randomly selecting one gene and substituting it with a random real number. This population is involved in the following iteration of the algorithm. In this study, the algorithm terminates after reaching a maximum number of iterations. The algorithm that summarizes the GA mentioned above can be found in [27,38].

5.2. Particle Swarm Optimization

PSO is a population-based stochastic optimization technique, developed by Kennedy and Eberhart [43]. It is inspired by the behavior of a bird. In PSO, the potential solutions are called particles and form the swarm. The initial particles are randomly generated. They are assessed using the fitness function (or the objective function). The swarm direction is dominated by the movement of the best particle, i.e., the particle with the best performance. The position of the particle is calculated according to its best position achieved so far (called pbest) and the velocity. This calculation also includes the best position of the best particle of the swarm (called global best or gbest) or the best position of the neighbor particle (called the local best or lbest) if the neighborhood notion is applied. The notion of the neighborhood is related to the search purpose. Moreover, the movement of the particles depends on some parameters such as the maximum velocity (Vmax) or the inertia parameter. Both parameters are set to prevent a quick displacement of the particles and maintain the exploration of the search space. More details about PSO can be found in [38,44].

5.3. Tribes

Tribes is a PSO version suggested by Clerc [45]. It is a free parameter method and differs from the original PSO. A particle (or a solution) consists of an array of real numbers which represents the projection vector in this study. The swarm is a set of particles. Unlike PSO, the swarm is divided into subsets of particles called Tribes. The particles and Tribes are linked to ensure the recognition of both the best particle in each tribe and the best particle of the swarm. Contrary to PSO, this method does not require the computation of the velocity: the displacement of the particles in the problem domain uses specific movement strategies related to independent Gaussian or hyperspherical probability distributions with or without noise. It can been noted that this method is highly recommended for the search of local optima [45]. This is due to its structure, allowing simultaneous exploration of numerous potential regions around local optima before meeting the global optimum. The Tribes method is described in [44,45]. This method does not involve parameters. The fitness function is the PP index to be optimized. The algorithm of this technique can be found in [38,44].

6. Methodology

The present study is based on four main stages described as follows.

6.1. Outlier Detection

The detection of the outliers is univariate. First, let us represent the dataset as a matrix. The dataset matrix is transposed so that the rows represent the features (the attributes) and the columns are the samples. This is because the outliers can be found in the feature set of the original dataset (samples of the transposed dataset). Figure 2 shows an example of PPFS processing.
After transposing the dataset matrix, the PP index is chosen (either Friedman or Kurtosis) as well as the optimization method (either GA or Tribes). Then, the number of iterations and the number of runs are set (defined in the next section). Later, the optimization is applied to find the best projection a (see Section 4) associated with the best value of the PP index I ( a ) . The next step is to determine the outliers. For that, purpose the coordinate of the projected data are calculated according to Equation (1). In order to point out an observation (in the transposed dataset matrix) as an outlier, the rule based on k-sigma principle was applied. In fact, an observation is considered as an outlier when its distance to the mean is larger than 3 times the standard deviation of the projected data. Thus, the parameter k is set to 3 based on the recommendation of [37,38].
The optimization methods are performed several times (number of runs) to ensure that the obtained result is the best and cannot be changed. When selecting the best projection and determining the outliers, a threshold is set to confirm the outlying status of the recognized outliers. This threshold is set to half the number of runs. In fact, the proposed framework checks how many times these observations were detected as outliers throughout the runs. Thus, each outlier is associated with the frequency number showing how many times it was detected as outlier. In case of the previous example displayed in Figure 2, the last observation of the transposed matrix can be considered as an outlier because it is far from data points as displayed in Figure 3.
By applying PPFS, suppose that the number of runs is set to 100, and the two samples number 8 and 9, from the previous transposed dataset matrix example (see Figure 2) are detected as outliers. Therefore, the proposed technique will detect these two outliers with the number of times they were recognized as outliers after performing 100 runs. The sample number 8 will not be detected as an outlier, or it will be detected but with a frequency number less than the threshold (=50), while sample 9 will be detected with a frequency greater than 50 as it is considered as outlier. This frequency number ensures the status (outlier or not) of each observation. In this study, an outlier is confirmed if it is detected more than the fixed threshold (half the number of runs). The list of confirmed outliers corresponds to the non-relevant features in the original dataset.

Outlier Detection Using EPP-Lab

The extraction of the outliers is performed using EPP-Lab software available at https://www.researchgate.net/project/Projection-Pursuit (accessed on 22 July 2021) or https://github.com/fischuu/EPP-lab (accessed on 22 July 2021).
After the transposed dataset matrix is uploaded in EPP-Lab, the PP index is chosen (either Friedman or Kurtosis) as well as the optimization method (either GA or Tribes) from the EPP-Lab software. Then, the number of iterations and the number of runs are set (defined in the next section). Figure 4 shows the EPP-Lab interface used for this purpose.
After running, the best projection, the best values of the PP index obtained throughout the runs, and the projected data are provided. EPP-Lab allows the user to set the k parameter of the sigma rule. The software provides a list of detected outliers and the number of times (the frequency) they are detected as outliers. The functioning of EPP-Lab is described in [38].

6.2. Feature Selection

This stage involves removing the outliers obtained from the original dataset and keeping only the relevant features. Considering the previous example, feature 6 will be removed from the feature set of the original dataset, as shown in Figure 2. Therefore, the new dataset (after removing the irrelevant features) will be used in the classification phase.

6.3. Classification

This stage involves performing the classification of the dataset before and after removing the outliers (reducing the data dimentionality) using three classifiers, k-Nearest Neighbors (k-NN), Support Vector Machine (SVM), and Random Forest (RF) with ten-fold cross-validation. The reason behind the choice of these classifiers is that k-NN is the simplest classifier that is recommended for small datasets [46], while SVM and RN are well known for their efficiency in tackling large datasets [46,47].

6.4. Validation

The last step consists of validating the classification model using four classification measures including the accuracy, precision, recall, and F1 measure.
Figure 5 summarizes the four steps that define the proposed approach including transposing the dataset, detecting the outliers using EPP-Lab, FS, classification and evaluation.

7. Experimentation Results

The proposed outlier detection-based FS techniques (PPFS) were tested using a Dell XPS 9343 with an Intel core TM i7-5500U CPU 2.40 GHz and 8 GB RAM. In the following experimentation, 19 datasets (described in Table 2) were used to validate the effectiveness of this study. The first nine datasets are from the UCI repository (https://archive.ics.uci.edu/ml/datasets.php (accessed on 14 January 2021)). The next eight datasets are from (https://www.openml.org/search?q=australian&type=data (accessed on 14 January 2021)). Finally, August_senate_polls (August S.P) dataset is taken from https://data.world/fivethirtyeight/august-senate-polls (accessed on 14 January 2021), and Fifa2019 is from https://www.kaggle.com/karangadiya/fifa19 (accessed on 14 January 2021). August S.P contains polls from 1990 to 2016, the results of the election related to each poll, and the average error of each poll. This dataset is composed of three classes. The Fifa2019 dataset is composed of more than 13,000 instances and 80 attributes. An extensive preprocessing operation was applied to this dataset. Then, two classes were created in the body_type attribute, namely Lean and Normal, using R programming language.
As EPP-Lab software is used for outlier detection, only the number of iterations and the number of runs are set. All the parameters related to GA (set in EPP-Lab [38]) were not changed. This is why no validation phase for setting the parameters of GA is needed. In fact, for GA, the population is randomly generated with 50 individuals. The tournament selection was used with three individuals. The crossover was employed with two points, and the probability equals 0.65. The mutation is applied to all individuals with a probability 0.05. As stated before, the Tribes method does not require parameters. For the initialization phase, the EPP Lab randomly generates an initial solution (population) for each method. The numbers of iterations and runs are set to 100 according to the recommendations of [38]. The number of runs guarantees that the obtained results (the removed outliers) cannot be improved. It also allows determining the number of times a feature is detected as an outlier. The threshold that confirms the outlying status is then set to 50 (see Section 6.1). For more details, one can refer to [38,44].
Table 3 shows the parameters’ values.
In the classification stage, three classifiers are applied: Support Vector Machine (SVM) with a Linear kernel, the k Nearest Neighbors (k-NN), and the Random Forest (RF). R (version 3.5.1) is used for the classification step. The datasets are split into training (70%) and testing (30%). The split was done randomly by R using the “createDataPartition” function from the caret library. To validate the parameter setting of the classifiers, ten-fold Cross-Validation, repeated ten times, is applied to the training set. The accuracy with the largest value is used to select the optimal training model. The optimal model with its best parameters’ values is then used in the testing stage. The training and validation accuracy values are omitted to save space. The training set is centered and scaled.
For the k-NN classifier, several values of k were tested (between 5 and 200) for each dataset. For the RF classifier, the number of attributes used for splitting at each node in the tree was studied. The optimal training model was selected using the values (0.1, 0.8, 2, 6, 10). For the SVM classifier, the cost parameter was investigated, which is considered as the essential parameter for building the best useful model. After several experiments, the values (0.5, 1, 1.25, and 2) were used to select the optimal training model. Table 4 shows the obtained accuracy values for SVM using the four aforementioned values of the cost for the Breast Cancer dataset after applying PPFS-GAF. The final value used for the model was cost = 0.5. Note that this parameter setting approach was applied to the three classifiers before and after applying the four PPFS methods for all the datasets. Therefore, to save space, only Table 4 is shown.

7.1. Projection-Pursuit-Based Feature-Selection Techniques

The four PPFS methods include PP based on Friedman index and the GA optimization method (called GAF), PP based on Friedman index and the Tribes optimization method (called TF), PP based on Kurtosis index and the GA optimization method (GAK), and PP based on Kurtosis index and the Tribes optimization method (called TK). After detecting the outliers using each PPFS method, they were removed, and then the three classifiers were applied to determine the accuracy, precision, recall, and F1 measures for all the datasets, as explained in the Methodology section. Prior to starting the experiments, the features were numbered, starting from 1, according to their appearance in the dataset. For example, if a dataset had four features, then the features were numbered from 1 to 4 starting with the first feature in the dataset until the last one. Table 5 displays the number of outliers detected by each approach (column 4) for each dataset. In addition, the features detected as outliers and the number of times (over 100 runs) they were recognized as such (frequency) are shown in column 3. The last column yields the running time of PPFS to find out the outliers.
Most of the outliers are commonly revealed by the four approaches as shown in Column 3. Moreover, their frequency values are greater than the threshold (half of the number of runs = 50). For example, for the Diabetes dataset, the features 1, 2, 5, and 7 were detected as outliers by both PPFS-GAF and PPFS-TF. However, the features 2 and 7 were detected by the four PPFS methods. In addition, these features were recognized as outliers more than 68 times over 100 runs. Hence, the four features can be considered non-relevant and were removed from the dataset to perform the classification. The positive point in this result is that the four PPFS methods provide almost similar outputs. This fact confirms the consistency of the proposed study. The small difference can be explained by the behavior of the projection indices as well as the effect of the optimization methods. The Kurtosis index split the data into two unbalanced clusters, and the smallest one is considered to comprise outliers. The Friedman index detects the outliers by weighting distances in the center of the Normal distribution rather than in the tails [40]. For the optimization methods, GA is well-known in finding the best solution (the optimum or close to the optimum), whereas the Tribes is dedicated to finding local optima as well as the optimum. Therefore, combining these two optimization methods along with both PP indices does not yield exactly the same result; however, the accuracy and the different measures (precision, recall, and F1) provided by each classifier for the four approaches are slightly different for the used datasets. For the experiment time, EPP-Lab is fast when combining Tribes and the Kurtosis index [39]. The shortest average time noticed with the smallest dataset (Wine, 178 samples) was 79 ms using TK and 162 ms using TF. With the largest dataset, Fifa2019, there were 13,000 samples and 80 features, while Colon had 62 samples and 2000 features. TK for Fifa2019 spent an average of 12 s 580 ms, whereas TF took an average of 44 s 143 ms. TK for Colon was 1 s 654 ms, and TF was 6 s 443 ms. Again, the running time does not show a major difference (see Table 5).
Table 6 presents the accuracy, precision, recall, and the F1 values of each classifier before and after applying PPFS methods (columns 3–6 for k-NN, columns 7–10 for SVM, and columns 11–14 for RF). The PPFS methods have globally improved the classification accuracy compared to the accuracy obtained without applying FS. The accuracy obtained using k-NN (resp. SVM and RF) indicates an improvement rate ranging between 0.08 and 6.79 for 9 datasets, while the accuracy for SVM and RF were 0.36–12.9 and 0.01–12.7, respectively, for 16 datasets each. This result is explained by the fact that k-NN is not always efficient with large and noisy datasets. The classification accuracy of the datasets that is not enhanced with k-NN achieved a significant improvement rate with the other classifiers. This is the case for Wine, Spam, August S.P, Diabetes, Australian, Glass, and Ionosphere datasets.
The PPFS methods also managed well for the Fifa2019 dataset even though the accuracy remains below 70%. This dataset is large and required massive preprocessing efforts. However, PPFS methods did not enhance the accuracy of the Waveform and Segment datasets using the three classifiers because Waveform contains 40 features, all of which include noise, especially the last 19, which are all noise, and Segment contains many instances of noise. PPFS did not remove all these noise features.
Moreover, considering the results of the three classifiers, the classification accuracy of some datasets was enhanced, but they remained below or around 70%. This is the case of Liver D., Vehicle (enhanced only once reaching 83.33%), Diabetes (enhanced only once reaching 81.30%), Glass, and Fifa2019. To improve the results, it recommended to preprocess the datasets and/or apply centering and sphering procedure [37] before peforming PPFS.
The four proposed PPFS methods provided approaching accuracy, precision, recall, and F1 values for each dataset. This confirms the consistency of PPFS again. Figure 6 shows the number of times each PPFS method enhances the classification accuracy (compared to that obtained without applying any FS method) using the three classifiers for all datasets. As displayed in this figure, GAF and GAK performed well at 36 and 37 times, respectively, using the three classifiers. In fact, GAF enhanced the accuracy 10 times with k-NN, 12 times with SVM, and 14 times with RF. TF and TK yielded similar results (34 times). On the other hand, the SVM classifier performed well with all the PPFS methods for all the datasets. It enhanced the results 58 times. The RF is a comparative classifier as it enhanced the accuracy results 48 times using the PPFS for all the datasets. The k-NN does not perform well with large datasets. It improved the accuracy 35 times using all the PPFS methods for all the datasets. Consequently, SVM can be considered as the classifier providing the best results.

7.2. What Is the Best PPFS Technique?

It is necessary to show that the results of the PPFS approaches are equivalent. To do this, the non-parametric statistical test “Friedman” was employed with a significance level of 0.01 . This test is conducted with the best classification accuracy obtained by SVM for all the datasets after performing PPFS. The test does not adopt a normal distribution assumption [48].
It is worth knowing that the Friedman test [48] is different from the Friedman index used in PP [40]. The Friedman index is defined in Section 4, and Friedman test shows the existence of a significant statistical difference between the methods.
The Friedman test provides two main results, the p-value and the effect size. If the obtained p-value is less than the significance level ( 0.01 ), then the statistical hypothesis (that the PPFS methods are equivalent) is rejected. The effect size has a value ranging between 0 (corresponding to no relationship) and 1 (stating an ideal relationship). The Friedman test resulted in χ 2 ( 2 ) = 6.68 and p _ v a l u e = 0.0204 > 0.01 , with an effect size W = 0.172 . According to the p _ v a l u e , the hypothesis is not rejected. Moreover, the effect size is small. Consequently, there is no statistical difference between the PPFS methods. The obtained PPFS results are equivalent.
Figure 7 displays the boxplot of the accuracy obtained by SVM after using the PPFS methods for all the datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values. As indicated, the four PPFS have a similar median (the line inside the box) equal to 83.3 for GAF, TF, and TK, 83.5 for GAK. As the p _ v a l u e = 0.0204 > 0.01 , the dissimilarity between the median values are not statistically significant. This confirms that the PPFS techniques are equivalent and that their results are statistically significant. Furthermore, the mean values of the four PPFS do not show a significant difference. The lowest value was provided by GAF (82.0), whereas the highest mean was obtained by GAK (84.2). TF yielded 82.6 and TK 83.3.

7.3. Comparison Study

The comparison study was performed based on three wrapper and five filter-based FS methods. Unfortunately, the existing embedded FS methods did not use the available datasets or apply clustering techniques (instead of classification techniques). Furthermore, the related work applying outlier detection to select the relevant features did not use the available data, and their evaluation measures are different from those used in this article.
The first comparison study consisted of applying the Correlation (CFS), Principal Component Analysis (PCA), Gain ratio, the Mutual Information, and Gini index FS techniques. The reason behind this choice is that PCA belongs to the same family of the PP and the other FS are well-known filter FS methods. The implementation of the filter FS techniques was performed using both FSelector and FSinR R packages [11]. The Gain ratio, Mutual Information, and Gini index were applied using the Sequential Forward Selection (FSF) search method as it provided better results than Sequential Backward Selection (SBS) and Breadth First Search (BFS). In addition, the FSF search method was the fastest.
Table 7 displays the classification accuracy obtained by SVM (as it is the best classifier in this study) after applying the four PPFS methods along with the five filter FS for all the datasets. As shown in this table, the four PPFS outperformed all the Filter FS for 15 datasets (over 19 datasets). The accuracy improvement rate ranges between 0.76% and 30.64%. However, PCA managed well in eliminating the noisy features of Waveform dataset. In addition, it enhanced the accuracy of Sonar dataset despite its correlated features. This is due to its process in projecting the feature space into a new low-dimensional feature space without being affected by the noise or high correlation. Furthermore, CFS also improved the accuracy of Sonar dataset. This is done thanks to the concept of removing the correlated features. Besides, Mutual Information and Gini Index outperformed the PPFS methods when tackling the Spam dataset. Moreover, CFS, Gain ratio, Mutual Information, and Gini index (resp. CFS) yielded similar results to the four PPFS methods when handling the Climate (resp. Heart) dataset. For the Colon dataset, the last three FS methods did not manage to find the relevant features. For the running time, the five filter FS are similar to PPFS in terms of speed, except when handling Spam, Colon, and Fifa2019. For these datasets, the filter FS took an average of 12 min and 10 s compared to PPFS, which took an average of 7 s.
To show the statistical significance of the obtained results, the same experiment discussed in Section 7.2 was performed. It aimed to show the existence of a significant statistical difference between the methods. As the PPFS methods are equivalent, only GAK was used. Table 8 displays the mean and median of the accuracy values obtained by SVM classifier after applying GAK and the five filter FS methods. These results are visually shown in Figure 8. As indicated, the median and the mean of GAK are greater than those of the four filter methods, but slightly different from those of PCA. Moreover, the Friedman test was applied with a confidence level of 0.01. The p _ v a l u e achieved 0.00994 < 0.01 , χ 2 ( 2 ) = 15.1 , with an effect size W = 0.168 (small effect), which means that there is no relationship. Thus, the statistical hypothesis (the methods are not statistically different) is rejected since the p _ v a l u e is less than the significant level and the effect size is small.
This result indicates that the methods are statistically different.
Now, to demonstrate the differences between each pair of methods, the Friedman test was followed by the pairwise Wilcoxon signed-rank test. Table 8 (see the first line) pointed out the p _ v a l u e calculated between GAK and the filter methods. As displayed, GAK, CFS, and PCA are competitive methods. This is because the hypothesis is not rejected (GAK is statistically different from CFS and PCA) due to the p _ v a l u e not being less than the significance level 0.01 . However, GAK is better than Gain Ratio, Mutual Information, and Gini index FS methods, since their p-value is less than 0.01 . Consequently, GAK outperforms Gain Ratio, Mutual Information, and Gini index, but it is competitive with PCA and CFS. This result is again confirmed with the medians and the means computed in Table 8.
The second comparison study involved comparing PPFS approaches with three wrapper-based FS methods published in [14], named Enhanced Whale Optimization Algorithm (WOA-CM), GA, and PSO using k-NN. Note that GA is also employed in GAF and GAK, and Tribes is one variant of PSO used in TF and TK. Table 9 and Figure 9 show the classification accuracy obtained by k-NN after applying PPFS techniques along with the three wrapper methods using eight UCI datasets. The PPFS approaches outperformed the wrapper FS methods on five datasets. However, only WOA outperformed the four PPFS methods with an improvement of 5.06% for Ionosphere, 13.12% for Colon, and 16.09% for the Sonar datasets. In addition, the three wrapper FS outperformed the four PPFS methods when handling the Colon dataset. The reason behind this success is that WOA is mainly based on the exploration and exploitation concepts leading to the global search. This aspect allowed WOA to find out the optimal solutions for the Waveform, Colon, and Sonar datasets in revealing the important features despite the noise and high correlation containing in these datasets.
Since the number of datasets used in this comparison study is small (eight datasets), the result of the statistical tests (Friedman or Wilcoxon) might be inaccurate [48]. Therefore, their results were omitted. However, the median and the mean of the accuracy values obtained by k-NN after applying GAF, WOA, PSO, and GA were calculated and displayed in Table 10. In addition, the Boxplot of these four methods was presented in Figure 10. As displayed, the mean and the median of the accuracy values obtained by GAF are higher than those obtained by PSO and GA, but somewhat similar to WOA. Therefore, it is worth mentioning that GAF yielded results better than or comparable to the aforementioned FS. Therefore, PPFS methods are competitive methods.

7.4. Discussion

Many experiments were performed in this paper to fulfill the four reseach objectives. Firstly, the related works were deeply investigated to bring out their limitations. Consequently, the first objective was achieved.
To prove the efficacy of these four PPFS, nineteen datasets and three classifiers were utilized. The first experiment consists of showing the detected outliers for the nineteen datasets using the four FS methods. It was shown that the PPFS methods provide equivalent results in terms of the detecting outliers. In addition, the detection frequency of most of the outliers reached 80% using the four PPFS methods. This confirms that they are outliers. To show that the detected outliers correspond to non-relevant features, classification was performed on the datasets before and after applying the four PPFS methods. The results showed that the classification performance along with the precision, the recall, and the F1 measure were superior when applying PPFS. The k-NN achieved an improvement, reaching 6.79 for 11 datasets. The SVM and RN enhanced the results of 16 datasets, with an improvement reaching 12.91 and 12.4, respectively. Hence, SVM was selected as the best classifier for this study. Although the proposed PPFS did not improve the accuracy of the Waveform and Segment datasets, they enhanced the classification accuracy for most of the datasets. This allowed the second and third objectives to be realized.
On the other hand, the success of this study could not be realized without the utilization of EPP-Lab software. Therefore, EPP-Lab is efficient in detecting the non-important features. This software is open-source and can be used as a desktop application in any environment. This is the fourth objective of the present study.
One of the main experiments performed in this article was to show that the four presented PPFS methods provide equivalent results. This was carried out through the Friedman test. The test’s result indicated that the p-value is greater than the significace level and that the effect size is small. Consequently, the hypothesis is not rejected, and hence the difference between the PPFS results is not statistically significant. This experiment realized the fifth objective.
To demonstrate that the four PPFS are competitive FS methods, two comparison studies were performed. The first comparison included applying five filter FS methods (CFS, PCA, Gain Ratio, Mutual Information, Gini Index) using all the datasets. The results showed that CFS was superior to the four PPFS methods when testing only the Sonar dataset. PCA was also superior to the four PPFS methods when testing both the Waveform and Sonar datasets. Again, Mutual Information and Gini Index outperformed the PPFS for the Spam dataset. As stated previously, these three datasets have noise and high correlation, which negatively affect PPFS in enhancing the classification accuracy. These limitations deserve further investigation. The classification accuracy results of PPFS-GAK and the five Filter FS methods were used to compute the Friedman and pairwise Wilcoxon tests. The statistical results show that GAK is superior to Mutual Information, Gain Ratio, and Gini Index but equivalent to PCA and CFS.
The second comparison involved comparing PPFS with three wrapper FS methods, called WOA, PSO, and GA, using eight datasets. The four PPFS methods performed better than the three wrapper FS using five datasets. However, WOA surpassed all the FS methods using Sonar, Waveform, and Colon datasets. Again, the PPFS methods failed in enhancing the accuracy of these three datasets. Like the first comparison study, the statistical analysis validated the significant results of the PPFS methods. Finally, the last objective was achieved by demonstrating the efficiency of the PPFS methods. To sum up, the aforementioned experiments successfully carried out the research objectives of this study.
In the following, the research impacts, the advantages, and the limitations of PPFS methods are discussed.

7.4.1. Research Impacts

The research impacts of PPFS methods are important and deserve to be addressed:
  • They contribute to the scientific society by providing a new PPFS model that overcomes the limitations of the existing FS methods.
  • They provide four FS techniques that enhance the classification accuracy by an average of 6.64%.
  • They contribute to technological development by providing EPP-Lab software handling FS.
  • They contribute to decision making by providing four PPFS methods yielding different but equivalent results.

7.4.2. Advantages and Liminations of PPFS

The present research has six main advantages.
  • PPFS methods are competitive and efficient FS methods.
  • The four PPFS methods are different but provide similar results. The user can either apply all of them and choose the suitable result or use only one of them.
  • PPFS methods can be used directly through EPP-Lab software.
  • The EPP-Lab can be used to avoid the parameter setting of the optimization methods.
  • PPFS methods provide a list of outliers, and their frequency allows validating the non-relevant features.
  • EPP-Lab is an open-source desktop application, downloadable in any environment.
In spite of the previous advantages, PPFS have two limitations.
  • PPFS failed to find the non-significant features when most of the features are highly correlated or include noise.
  • PPFS methods do not work well when the dataset is binary (the values of features’ vectors are either 0 or 1).

8. Conclusions

The present study proposed four FS methods based on outliers detection using the Project Pursuit. Two PP indices along with two AI optimization methods were used to find the optimal projections that reveal the outliers. The proposed PPFS methods were validated using three classifiers and 19 datasets. They yielded promising results.Moreover, these methods were successfully compared with three Wrapper FS methods on eight datasets. They also yielded the best or equivalent accuracy against five Filter FS techniques on 15 datasets (among 19 datasets). Statistical analysis was performed and showed that PPFS are equivalent, and their results are statistically significant. It is also demonstrated that PPFS are comparative with PCA, CFS, and WOA. Therefore, our assumption to consider the outlier detection methods for revealing the irrelevant features from a dataset to increase the classification accuracy was validated. For the computation time, PPFS was faster than the five Filter FS methods when handling large datasets (Colon, Spam, and Fifa2019).
However, no computation time was reported for the three wrapper FS methods. Furthermore, PPFS yielded promising results in handling a large dataset reaching 13,000 instances (Fifa2019 dataset) and 2000 features (Colon dataset). The only limitation of this study is not getting the best result when dealing with the Waveform and Segment dataset. Hence, some future research directions are suggested:
  • Make the dataset spherical before seeking the outliers [27] to reduce the noise.
  • Use another optimization method such as PSO, which is already available in EPP-Lab.
  • Use large datasets reaching 30,000 features to test the extent of PPFS in enhancing accuracy.
  • Explore other PP indices such as Friedman Turkey and Discriminant indices.
  • Investigate other Machine Learning classifiers.

Funding

The authors would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges (APC) of this publication.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The links to datasets used in this study are provided in Section 7.

Acknowledgments

The author would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges (APC) of this publication. This work was also supported by Artificial Intelligence and Data Analytics Lab (AIDA), Prince Sultan University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

All the abbreviations used in this paper are defined below.
FSFeature Selection
PPProjection Pursuit
PPFSProject Pursuit based Feature Selection
AIArtificial Intelligence
GAGenetic Algorithms
PSOParticle Swarm Optimization
EPP-LabExploratory Projection Pursuit Laboratory
PPFS-GAFProjection-Pursuit-based Feature Selection
using Genetic Algorithm and Friedman index
PPFS-GAKProjection-Pursuit-based Feature Selection
using Genetic Algorithm and Kurtosis index
PPFS-TFProjection-Pursuit-based Feature Selection
using Tribes and Friedman index
PPFS-TKProjection-Pursuit-based Feature Selection
using Tribes and Kurtosis index
LASSOLeast Absolute Shrinkage and Selection Operator
LRLinear Regression
CVCross Validation
k-NNK-Nearest Neighbor
SVMSupport Vector Machine
RFRandom Forest
PCAPrincipal Component Analysis
CFSCorrelation Feature Selection
WOAWhale Optimization Algorithm
AccuAccuracy
PrecPrecision
RecRecall
mutualInfoMutual Information

References

  1. Fukumizu, K.; Bach, F.R.; Jordan, M.I. Dimensionality reduction for supervised learning with reproducing kernel hilbert spaces. J. Mach. Learn. Res. 2004, 5, 73–99. [Google Scholar]
  2. Sophian, A.; Tian, G.Y.; Taylor, D.; Rudlin, J. A feature extraction technique based on principal component analysis for pulsed Eddy current NDT. NDT E Int. J. 2003, 36, 37–41. [Google Scholar] [CrossRef]
  3. Scholkopf, B.; Smola, A.; Muller, K.-R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. J. 1998, 10, 1299–1319. [Google Scholar] [CrossRef] [Green Version]
  4. Chandrashekar, G.; Sahin, F. Survey on feature selection methods. J. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  5. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 57–82. [Google Scholar]
  6. Taminau, J.L.; Meganck, S.; Steenhoff, D.; Coletta, A.; Molter, C. A survey on filter techniques for feature selection in gene expression microarray analysis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2012, 9, 6–19. [Google Scholar]
  7. Eom, J.; Zhang, B. PubMiner: Machine learning-based text mining for biomedical information analysis. Lect. Notes Artif. Intell. 2000, 3192, 216–225. [Google Scholar]
  8. Larabi-Marie-Sainte, S.; Ghouzali, S. Multi-Objective Particle Swarm Optimization-based Feature Selection for Face Recognition. Stud. Inform. Control. J. 2020, 29, 99–109. [Google Scholar] [CrossRef]
  9. Larabi-Marie-Sainte, S.; Alayani, N. Firefly Algorithm based Feature Selection for Arabic Text Classification. J. King Saud Univ. Comput. Inf. Sci. (JKSUCIS) 2020, 32, 320–328. [Google Scholar] [CrossRef]
  10. Wen, J.; Canbing, L.; Rui, L. A Heuristic Feature Selection Approach for Text Categorization by Using Chaos Optimization and Genetic Algorithm. Math. Probl. Eng. 2013, 2013. [Google Scholar] [CrossRef] [Green Version]
  11. Aragon-Royon, F.; Jimenez-Vilchez, A.; Arauzo-Azofra, A.; Benitez, J.M. FSINR: An exhaustive package for feature selection. arXiv 2020, arXiv:2002.10330v1. [Google Scholar]
  12. Rattanawadee, P.; Anongnart, S. Wrapper Feature Subset Selection for Dimension Reduction Based on Ensemble Learning Algorithm. Procedia Comput. Sci. 2015, 72, 162–169. [Google Scholar]
  13. Agrawala, R.K.; Kaur, B.; Sharma, S. Quantum based Whale Optimization Algorithm for wrapper feature selection. Appl. Soft Comput. J. 2020, 89, 106092. [Google Scholar] [CrossRef]
  14. Mafarja, M.; Mirjalili, S. Whale optimization approaches for wrapper feature selection. Appl. Soft Comput. J. 2018, 62, 441–453. [Google Scholar] [CrossRef]
  15. Gokalp, O.; Tasci, E.; Ugur, A. A novel wrapper feature selection algorithm based on iterated greedy metaheuristic for sentiment classification. Expert Syst. Appl. 2020, 146, 113–176. [Google Scholar] [CrossRef]
  16. Xu, T.; Feng, Z.H.; Wu, X.J.; Kittler, J. Learning Adaptive Discriminative Correlation Filters via Temporal Consistency Preserving Spatial Feature Selection for Robust Visual Object Tracking. IEEE Trans. Image Process. J. 2019, 28, 5596–5609. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Bommerta, A.; Sun, X.; Bischl, B.; Rahnenführer, J.; Lang, M. Benchmark for filter methods for feature selection in high-dimensional classification data. Comput. Stat. Data Anal. J. 2020, 143. [Google Scholar] [CrossRef]
  18. Wang, Y.; Tao, L. Local feature selection based on artificial immune system for classification. Appl. Soft Comput. J. 2020, 87, 105989. [Google Scholar] [CrossRef]
  19. Cai, J.; Hu, L.; Liu, Z.; Zhou, K.; Zhang, H. An Embedded Feature Selection and Multi-Class Classification Method for Detection of the Progression from Mild Cognitive Impairment to Alzheimer’s Disease. J. Med. Imaging Health Inform. 2020, 10, 370–379. [Google Scholar] [CrossRef]
  20. Guo, Y.; Chung, F.-L.; Li, G.; Zhang, L. Multi-Label Bioinformatics Data Classification With Ensemble Embedded Feature Selection. IEEE Access J. 2019, 7, 103863–103875. [Google Scholar] [CrossRef]
  21. Azmandian, F.; Dy, J.G.; Aslam, J.A.; Kaeli, D.R. Local Kernel Density Ratio-Based Feature Selection for Outlier Detection. In Proceedings of the Asian Conference on Machine Learning, Singapore, 4–6 November 2012; Volume 25, pp. 49–64. [Google Scholar]
  22. Suresh, M.S.S.; Narayanan, A. Improving Classification Accuracy Using Combined Filter+Wrapper Feature Selection Technique. In Proceedings of the 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), Coimbatore, India, 20–22 February 2019; pp. 1–6. [Google Scholar]
  23. Hallac, D.; Leskovec, J.; Boyd, S. Network lasso: Clustering and optimization in large graphs. In Proceedings of the KDD, Sydney, Australia, 10–13 August 2015. [Google Scholar]
  24. Osanaiy, O.; Cai, H.; Raymond-Choo, K.K.; Dehghantanha, A.; Xu, Z.; Dlodlo, M. Ensemble-based multi-filter feature selection method for DDoS detection in cloud computing. EURASIP J. Wirel. Commun. Netw. 2016, 130, 1–10. [Google Scholar] [CrossRef] [Green Version]
  25. Hadi, S.; Imon, A.H.M.R.; Werner, M. Detection of outliers. Wiley Interdiscip. Rev. Comput. Stat. 2009, 1, 57–70. [Google Scholar] [CrossRef]
  26. Ruiz-Gazen, A.; Larabi-Marie-Sainte, S.; Berro, A. Detecting Multivariate Outliers Using Projection Pursuit with Particle Swarm Optimization. In Proceedings of the 19th International Conference on Computational Statistics (Book Chapter), Paris, France, 22–27 August 2010; pp. 89–98. [Google Scholar]
  27. Berro, A.; Larabi-Marie-Sainte, S.; Ruiz-Gazen, A. Genetic Algorithms and Particle rm Optimization for Exploratory Projection Pursuit. Ann. Math. Artiffitial Intell. J. 2010, 60, 153–178. [Google Scholar] [CrossRef]
  28. Yang, Q.; Singh, J.; Lee, J. Isolation-based feature Selection for Unsupervised Outlier Detection. In Proceedings of the Annual Conference of the PHM Society, Scottsdale, AZ, USA, 21–22 September 2019; Volume 11. [Google Scholar]
  29. Kovarasan, R.K.; Rajkumar, M. An Effective Intrusion Detection System Using Flawless Feature Selection, Outlier Detection and Classification. In Progress in Advanced Computing and Intelligent Engineering. Advances in Intelligent Systems and Computing; Pati, B., Panigrahi, C., Misra, S., Pujari, A., Bakshi, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2019; p. 713. [Google Scholar]
  30. Karthikeyan, G.; Balasubramanie, P. Robust Feature Selection Model for Outlier Detection Using Fuzzy Clustering and Rule Mining. Int. J. Appl. Eng. Res. 2017, 12, 7019–7028. [Google Scholar]
  31. Pang, G.; Cao, L.; Chen, L. Learning Homophily Couplings from Non-IID Data for Joint Feature Selection and Noise-Resilient Outlier Detection. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-17), New York, NY, USA, 9–15 July 2016. [Google Scholar]
  32. Solorio-Fernández, S.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J.F. A new hybrid filter–wrapper feature selection method for clustering based on ranking. Neurocomput. J. 2016, 214, 866–880. [Google Scholar] [CrossRef]
  33. Zhen, L.; Xin, L.; Jin, M.; Hui, G. An Optimized Computational Framework for Isolation Forest. Math. Probl. Eng. 2018, 2018. [Google Scholar] [CrossRef] [Green Version]
  34. Koupaie, H.M.; Suhaimi, I.; Hosseinkhani, J. Outlier Detection in Stream Data by Machine Learning and Feature Selection Methods. Int. J. Adv. Comput. Sci. Inf. Technol. (IJACSIT) 2013, 2, 17–24. [Google Scholar]
  35. Huber, P. Projection pursuit. Ann. Stat. 1985, 13, 435–475. [Google Scholar] [CrossRef]
  36. Jones, M.; Sibson, R. What is projection pursuit? (With discussion). J. R. Stat. Soc. A 1987, 150, 1–36. [Google Scholar] [CrossRef]
  37. Larabi-Marie-Sainte, S. Biologically Inspired Algorithms for Exploratory Projection Pursuit. Ph.D. Thesis, Toulouse 1 University (Tolbiac), Toulouse, France, 2011. Available online: https://scanr.enseignementsup-recherche.gouv.fr/publication/these2011TOU10021 (accessed on 20 July 2021).
  38. Larabi-Marie-Sainte, S. Detection and visualization of non-linear structures in large datasets using Exploratory Projection Pursuit Laboratory (EPP-Lab) software. J. King Saud Univ. Comput. Inf. Sci. 2017, 29, 2–18. [Google Scholar]
  39. Penã, P.; Prieto, F. Cluster Identification using projections. J. Am. Stat. Assoc. 2001, 456, 1433–1445. [Google Scholar] [CrossRef] [Green Version]
  40. Friedman, J. Exploratory Projection Pursuit. Am. Stat. Assoc. 1987, 82, 249–266. [Google Scholar] [CrossRef]
  41. Sun, J. Some practical aspects of exploratory projection pursuit. SIAM J. Sci. Comput. 1993, 14, 68–80. [Google Scholar] [CrossRef]
  42. Holland, J.H. Adaptation in Natural and Artificial Systems; University of Michigan Press: Ann Harbor, MI, USA, 1975. [Google Scholar]
  43. Kennedy, J.; Eberhart, R. Swarm Intelligence; Shi, Y., Ed.; Springer: Boston, MA, USA, 1995. [Google Scholar]
  44. Larabi-Marie-Sainte, S.; Berro, A.; Ruiz-Gazen, A. An efficient Optimization Method for Revealing Local optima of Projection Pursuit Indices. In Proceedings of the Swarm Intelligent, ANTS2010, Brussels, Belgium, 8–10 September 2010; pp. 60–71. [Google Scholar]
  45. Clerc, M. Particle swarm optimization. In International Scientific and Technical Encyclopaedia; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  46. Tomar, A.; Nagpal, A. Comparing Accuracy of K-Nearest-Neighbor and Support-Vector-Machines for Age Estimation. Int. J. Eng. Trends Technol. (IJETT) 2016, 38, 326–329. [Google Scholar] [CrossRef]
  47. Huang, M.W.; Chen, C.W.; Lin, W.C.; Ke, S.W.; Tsai, C.F. SVM and SVM Ensembles in Breast Cancer Prediction. PLoS ONE 2017, 12. [Google Scholar] [CrossRef]
  48. Derrac, J.; Garcia, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. J. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
Figure 1. Feature selection categories.
Figure 1. Feature selection categories.
Applsci 11 06769 g001
Figure 2. An example of PPFS Processing.
Figure 2. An example of PPFS Processing.
Applsci 11 06769 g002
Figure 3. Representation of the transposed matrix of the above example.
Figure 3. Representation of the transposed matrix of the above example.
Applsci 11 06769 g003
Figure 4. Outlier detection interface using EPP-Lab [37].
Figure 4. Outlier detection interface using EPP-Lab [37].
Applsci 11 06769 g004
Figure 5. Projection Pursuit-based feature-selection approach.
Figure 5. Projection Pursuit-based feature-selection approach.
Applsci 11 06769 g005
Figure 6. Performance rate (represented in the y axis) of each PPFS method using each classifier for all the datasets.
Figure 6. Performance rate (represented in the y axis) of each PPFS method using each classifier for all the datasets.
Applsci 11 06769 g006
Figure 7. Boxplot of the accuracy obtained by SVM after applying PPFS methods for all the datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Figure 7. Boxplot of the accuracy obtained by SVM after applying PPFS methods for all the datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Applsci 11 06769 g007
Figure 8. Boxplot of the accuracy obtained by SVM after applying GAK and the 5-filter FS for all the datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Figure 8. Boxplot of the accuracy obtained by SVM after applying GAK and the 5-filter FS for all the datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Applsci 11 06769 g008
Figure 9. The classification accuracy (presented in the y axis) provided by k-NN classifier after applying the four PPFS, WOA, PSO, and GA feature selection.
Figure 9. The classification accuracy (presented in the y axis) provided by k-NN classifier after applying the four PPFS, WOA, PSO, and GA feature selection.
Applsci 11 06769 g009
Figure 10. Boxplot of the accuracy obtained by k-NN after applying PPFS-GAF, WOA, PSO, and GA for 8 datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Figure 10. Boxplot of the accuracy obtained by k-NN after applying PPFS-GAF, WOA, PSO, and GA for 8 datasets. The y-axis represents the minimum, first quartile, median, third quartile, and maximum value of the accuracy values.
Applsci 11 06769 g010
Table 1. Difference between the existing studies and the proposed approaches.
Table 1. Difference between the existing studies and the proposed approaches.
Related WorkPPFS
[28] Outlier-based FS usingOutlier-based FS using PP
feature ranking and isolation forest
[28] Outlier-based FS is a random-based on PSO and GA
based approachoptimization methods
Filter FS based on specific measuresbased on PP that finds interesting
(Distance, Correlation, ⋯)projections by optimizing the PP index
Wrapper FS based on AI methodsBased on PP
or a combination of methods
Embedded FS based on the LassoBased on PP and optimization methods
combined with the Linear Regression
Table 2. Datasets used in this study.
Table 2. Datasets used in this study.
DatasetFeaturesSampleClasses
1Breast cancer106992
2Sonar602082
3Waveform4050003
4Wine131783
5Liver Disorder63452
6Spam5746012
7Diabetes87682
8Vehicle188474
9Ionosphere343512
10Climate205402
11WDBC305692
12Australian146902
13Stock99502
14Glass92142
15Segment1923107
16Heart132702
17Colon2000622
18August S.P105953
19Fifa20198013,0002
Table 3. Parameter Setting for GA.
Table 3. Parameter Setting for GA.
ParameterValue
Population size50
Number of individuals in the tournament selection3
Crossover type2-points
Tournament Probability0.65
Mutation Probability0.05
Number of iterations100
Number of runs100
Table 4. Parameter Setting for SVM after applying PPFS-GAF for the Breast Cancer dataset.
Table 4. Parameter Setting for SVM after applying PPFS-GAF for the Breast Cancer dataset.
CostAccuracyKappa
0.500.96588720.9248802
1.000.96520690.9233597
1.250.96452660.9218380
2.000.96452660.9218380
Table 5. The features (outliers) detected by the four proposed PPFS methods for the 19 datasets.
Table 5. The features (outliers) detected by the four proposed PPFS methods for the 19 datasets.
DatasetProposed PPFSFeature Number (Frequency)Outliers NumberTime
Breast CancerPPFS-GAF1(97), 3(80), 6(81), 10(81)4245 ms
PPFS-GAK1(97)1130 ms
PPFS-TF1(97)1497 ms
PPFS-TK1(97)1401 ms
Liver disorderPPFS-GAF1(100), 2(78), 6(99)382 ms
PPFS-GAK1(100), 6(100)242 ms
PPFS-TF1(98), 2(63), 6(87)3164 ms
PPFS-TK1(98), 6(98)2115 ms
SonarPPFS-GAF1(96), 2(94), 3(91), 19(75), 20(98), 21–29(100), 30(97), 49(84), 50–52(97), 53–55(98), 56–60(99)27359 ms
PPFS-GAK1(100), 2(97), 3(72), 20(96), 21–29(100), 30(98), 50–60(100)24115 ms
PPFS-TF1(94), 2(90), 3(79), 4(51), 19(57), 20(79), 21(89), 22(91), 23(95), 24(96), 25(97), 26(99), 27(100), 28(100), 29(95), 30(89), 49(80), 50–60(98)23510 ms
PPFS-TK1(88), 2(73), 20(67), 21(90), 22(97), 23–25(100), 26(99), 27(100), 28(99), 29(98), 30(90), 50(91), 51(94), 52(95), 53–60(95)23207 ms
WinePPFS-GAF4(64), 5(64), 13(98)382 ms
PPFS-GAK13(100)132 ms
PPFS-TF13(98)1162 ms
PPFS-TK13(100)179 ms
SpamPPFS-GAF5(60), 16(54), 21(52), 26(56), 46(52), 55(50)67 s 13 ms
PPFS-GAK57(100)13 s 259 ms
PPFS-TF2(55), 19(53), 21(56), 27(67), 42(52), 45(61), 46(53), 55(53), 56(57), 57(60)108 s 14 ms
PPFS-TK56(93), 57(100)23 s 241 ms
WaveformPPFS-GAF1(84), 2(75), 10(88), 11(88), 12(82), 20(80), 21(91)73 s 758 ms
PPFS-GAK1(76), 2(64), 10(63), 11(75), 12(63), 13(53), 20(59), 21(78)81 s 627 ms
PPFS-TF1(79), 2(63), 10(83), 11(90), 12(71), 20(64), 21(79)75 s 1 ms
PPFS-TK1(71), 10(52), 11(71), 12(55), 15(50), 21(74)62 s 759 ms
VehiclePPFS-GAF6(100), 12(100), 15–16(100)495 ms
PPFS-GAK6(100), 12(100), 15–16(100)434 ms
PPFS-TF6(100), 9(62), 12(100), 15–16(100)5182 ms
PPFS-TK6(100), 12(100), 15–16(100)4114 ms
DiabetesPPFS-GAF1(90), 2(95), 5(84), 7(94)4232 ms
PPFS-GAK2(100), 7(100)2127 ms
PPFS-TF1(93), 2(99), 5(68), 7(97)4451 ms
PPFS-TK2(98), 7(86)2351 ms
IonospherePPFS-GAF1(98), 8(57), 13(50), 14(56), 15(60), 17(51), 18(71), 20(85), 24(65), 26(67), 27(52), 28(58)12310 ms
PPFS-GAK1(98), 3(81), 5(84), 7(73), 9(58), 20(65), 24(66), 26(50), 28(67)9133 ms
PPFS-TF1(81), 8(51), 10(54), 13(59), 14(51), 15(65), 17(57), 18(62), 19(56), 20(63), 21(52), 24(65), 26(53), 28(55)14420 ms
PPFS-TK1(98), 3(86), 5(77), 7(57), 24(55), 28(62)6206 ms
ClimatePPFS-GAF1(81)1257 ms
PPFS-GAK2(93), 3(100)2112 ms
PPFS-TF2(88), 3(92)2398 ms
PPFS-TK2(82), 3(100)2398 ms
WDBCPPFS-GAF2(53), 3(73), 4(68), 14(70), 22(64), 23(74), 24(66)7386 ms
PPFS-GAK4(100), 24(100)2163 ms
PPFS-TF4(85), 24(86)2574 ms
PPFS-TK4(100), 24(86)2267 ms
AustralianPPFS-GAF2(86), 3(81)2247 ms
PPFS-GAK2(100), 3(60)2136 ms
PPFS-TF2(99)1386 ms
PPFS-TK2(98)1229 ms
StockPPFS-GAF3(91), 5(80), 7(88), 8(82)4213 ms
PPFS-GAK3(100), 5(99), 6(94), 7(100), 8(100)5114 ms
PPFS-TF3(98), 5(97), 7(99), 8(95)4339 ms
PPFS-TK3(99), 5(83), 6(78), 7(100), 8(89)5193 ms
GlassPPFS-GAF3(51), 7(51)2121 ms
PPFS-GAK5(100)154 ms
PPFS-TF5(100)1203 ms
PPFS-TK5(100)1119 ms
SegmentPPFS-GAF1(56), 2(56), 14(76), 16(62)4937 ms
PPFS-GAK1(100), 2(100)2470 ms
PPFS-TF1(100), 2(99), 12(57), 14(98), 16(85), 17(60)61 s 622 ms
PPFS-TK1(99), 2(100)2901 ms
HeartPPFS-GAF4(98), 5(95), 8(94)396 ms
PPFS-GAK4(100), 5(100), 8(100)341 ms
PPFS-TF4(100), 5(100), 8(99)3161 ms
PPFS-TK4(98), 5(100), 8(99)382 ms
ColonPPFS-GAF2(62), 6(52), 10(54), 16(55), 24(58), 30(55), 42(64), 46(64), 50(55), 58(52), 64(50), 236(67), 332(61), 610(79), 712(59), 1612(59), 1754(76)177 s 181 ms
PPFS-GAK1(100), 2(100), 4(100), 6(100), 8(100), 10(100), 12(100), 14(100), 16(100), 18(100), 20(100), 22(100), 24(100), 26(100), 28(100), 30(100), 32(100), 34(99), 36(100), 38(100), 40(100), 42(100), 44(100), 46(100), 48(99), 50(100), 54(100), 56(100), 58(100), 60(100), 64(100), 70(100), 72(92), 76(99), 78(99), 80(99), 82(99), 84(84), 86(100), 92(100), 98(57), 100(57), 102(57), 104(57), 206(99), 236(100), 332(100), 610(98), 712(100), 1612(85), 1754(99)51839 ms
PPFS-TF1(98), 2(96), 4(96), 6(95), 8(95), 10(96), 12(100), 14(94), 16(97), 18(93), 20(96), 22(92), 24(95), 26(95), 28(93), 30(96), 32(93), 34(88), 36(93), 38(94), 40(96), 42(95), 44(96), 46(94), 48(73), 50(97), 54(95), 56(87), 58(93), 60(92), 64(96), 70(75), 72(76), 76(86), 78(86), 80(86), 82(86), 84(73), 86(93), 92(93), 98(52), 100(52), 102(52), 104(52), 206(86), 236(97), 332(97), 610(87), 712(95), 1612(66), 1754(88)516 s 443 ms
PPFS-TK1(100), 2(100), 4(100), 6(99), 8(99), 10(100), 12(100), 14(100), 16(99), 18(99), 20(100), 22(98), 24(100), 26(98), 28(99), 30(100), 32(98), 34(95), 36(98), 38(99), 40(98), 42(99), 44(98), 46(99), 48(85), 50(100), 54(100), 56(99), 58(97), 60(99), 64(99), 70(94), 72(81), 76(87), 78(87), 80(87), 82(87), 84(83), 86(99), 92(99), 98(51), 100(51), 102(51), 104(51), 206(88), 236(97), 332(98), 610(87), 712(97), 1612(72), 1754(91)511 s 654 ms
August S.PPPFS-GAF3–4(100)2221 ms
PPFS-GAK1(100), 3–4(100)3123 ms
PPFS-TF1(100), 3–4(100)3388 ms
PPFS-TK1(100), 3–4(100)3274 ms
Fifa2019PPFS-GAF5(61), 6(76), 8(83), 9(86), 17(82), 18(84), 80(67)743 s 952 ms
PPFS-GAK1(100)121 s 483 ms
PPFS-TF1(75), 5(65), 6(79), 8(72), 9(83), 17(75), 18(74), 80(70)844 s 143 ms
PPFS-TK1(100)112 s 580 ms
Table 6. The classification results obtained after (and before) applying the PPFS approaches using KNN, SVM, and RF classifiers for the 19 datasets.
Table 6. The classification results obtained after (and before) applying the PPFS approaches using KNN, SVM, and RF classifiers for the 19 datasets.
DatasetsPPFSk-NNSVMRF
AccuPrecRecF1AccuPrecRecF1AccuPrecRecF1
Breast C.Without95.2295.6897.0896.3898.0999.2697.8198.5394.6497.7394.1695.91
PPFS-GAF97.6197.8398.5498.1896.1797.0897.0897.0895.2296.3596.3596.35
PPFS-GAK/PPFS-TF/PPFS-TK95.2295.6897.0896.3898.5610097.8198.8997.6198.5397.8198.17
SonarWithout70.9770.2759.0954.2069.3566.6784.8574.6782.2684.3881.8283.08
PPFS-GAF75.8171.4390.9180.0082.2680.5689.6847.3288.7188.2490.9189.55
PPFS-GAK66.1363.6484.8572.7380.6581.8288.0048.1872.5876.6769.7073.02
PPFS-TF75.8171.4390.9180.0080.6583.8785.6449.4880.6580.0084.8582.35
PPFS-TK67.7464.4487.8874.3669.3568.4294.6141.9780.6586.2175.7680.65
Liver D.Without63.1164.8680.0071.6476.7075.0090.0081.8263.1167.7470.0068.85
PPFS-GAF69.9068.3590.0077.7065.0567.1478.3372.3171.8475.4176.6776.03
PPFS-GAK66.0265.4388.3375.1873.7973.9185.0079.0771.8475.4176.6776.03
PPFS-TF69.9068.3590.0077.7065.0567.1478.3372.3171.8475.4176.6776.03
PPFS-TK66.0265.4388.3375.1873.7973.9185.0079.0771.8475.4176.6776.03
WineWithout98.0898.4197.7898.0492.3191.6993.0992.2698.0898.4198.1598.23
PPFS-GAF98.0898.4198.1598.2396.1595.2497.1095.92100100100100
PPFS-GAK/PPFS-TF/PPFS-TK96.1596.8396.4996.4896.1596.0396.4996.10100100100100
SpamWithout90.7292.4592.2292.3492.6893.4994.5093.9996.0195.8997.6196.74
PPFS-GAF88.6990.0991.3990.7493.0493.0295.6994.3493.7693.5096.4194.94
PPFS-GAK89.4190.5992.1191.3493.6293.9095.6994.7995.0094.7597.1395.92
PPFS-TF90.7290.7894.2692.4993.0494.7194.2694.4895.4395.5297.0196.26
PPFS-TK89.9990.4993.3091.8793.0493.5395.1094.3195.3694.9997.4996.22
WaveformWithout86.0683.9584.5283.7786.3986.3986.3886.3785.4685.4185.5485.35
PPFS-GAF/PPFS-TF82.4582.3883.5181.8983.7983.7683.7783.6780.6580.5980.8680.24
PPFS-GAK82.1282.0782.8581.4881.6581.6381.5581.5781.1278.8979.0478.50
PPFS-TK81.9981.9083.8081.3281.4581.4281.3681.3581.7881.7281.9881.49
VehicleWithout67.4667.8465.9266.7379.7680.1079.7479.7573.0273.4171.5772.35
PPFS-GAF/PPFS-GAK/PPFS-TK70.2470.6169.7370.1383.3383.5283.1083.2774.6074.9873.6774.10
PPFS-TF73.0273.3072.0472.4079.3679.5778.8179.1076.9877.2876.0076.51
DiabetesWithout72.1777.2281.3379.2280.0081.7189.3385.3575.6578.6686.0082.17
PPFS-GAF/PPFS-TF67.3970.7285.3377.3481.3082.4290.6786.3576.9682.5582.0082.27
PPFS-GAK/PPFS-TK65.6569.8383.3375.9981.3082.4290.6786.3576.5281.1783.3382.24
IonosphereWithout87.5083.6489.1985.4589.4286.3590.5987.8991.3590.2690.7790.50
PPFS-GAF87.5084.2588.2485.6980.7777.8179.4378.4794.2393.7193.7193.71
PPFS-GAK81.7376.1483.7577.9690.3887.7091.2989.0893.2792.9692.4692.70
PPFS-TF84.6279.5987.1481.6280.7778.4279.1778.7690.3887.7091.2989.08
PPFS-TK84.6279.5987.1481.6281.7374.9386.7977.0491.3590.8690.3990.62
ClimateWithout91.9250.00NaNNaN87.5858.1658.1658.1691.9250.00NaNNaN
PPFS-GAF91.9350.00NaNNaN91.9350.00NaNNaN91.9350.00NaNNaN
PPFS-GAK/PPFS-TF/PPFS-TK91.9350.00NaNNaN91.9350.00NaNNaN90.6849.3245.9145.91
WDBCWithout92.9491.1393.8392.2194.7193.1895.6094.1894.7193.8494.7794.27
PPFS-GAF95.2993.9896.0394.8593.5392.2593.8492.9494.7194.4994.2194.34
PPFS-GAK/PPFS-TF/PPFS-TK98.2497.9598.2798.1098.2497.9698.2898.1095.2994.6395.2594.92
AustralianWithout85.9284.9786.9385.4584.9585.2584.8684.8985.9286.5586.1885.91
PPFS-GAF/PPFS-GAK83.5082.8883.6683.1485.9286.5586.1885.9186.8986.2787.2686.59
PPFS-TF/PPFS-TK83.5082.8883.6683.1485.9286.5586.1885.9186.8986.2787.2686.59
StockWithout91.9092.0292.1691.9081.6981.8582.1481.6695.0795.1795.2195.07
PPFS-GAF/PPFS-TF95.4295.4395.4295.4283.1082.8384.3382.8596.4896.4896.4896.48
PPFS-GAK/PPFS-TK95.4295.3995.4695.4286.6286.4387.3086.5195.4295.3795.5295.41
GlassWithout69.8471.5669.6169,0153.9645.6844.5844.5274.6065.7476.7966.49
PPFS-GAF57.1454.4354.2154.1960.3252.6653.3352.1769.8465.2466.4565.65
PPFS-GAK/PPFS-TF/PPFS-TK60.3261.0960.0859.2969.8463.1466.4663.6587.3088.1485.7686.54
SegmentWithout94.8194.8194.7894.7594.6694.6694.7794.6897.8397.8497.8697.84
PPFS-GAF89.1889.1889.1889.0986.5886.5886.6986.5894.3794.3794.3794.37
PPFS-GAK/PPFS-TK90.3390.3390.4690.3789.4789.4789.8089.4592.5092.5092.7892.50
PPFS-TF87.8887.8887.8987.7785.1485.1485.6085.1591.3491.3491.5491.36
HeartWithout80.2579.7280.1379.8781.4881.3981.2381.3083.9583.6183.8283.70
PPFS-GAF/PPFS-GAK/PPFS-TF/PPFS-TK81.4881.3981.2381.3082.7283.3382.9782.6987.6588.0687.5987.61
ColonWithout77.7866.6787.5067.8672.2270.8369.4869.9077.7875.0075.0075.00
PPFS-GAF72.2266.6768.4667.2783.3387.5083.3382.8683.3383.3381.1781.93
PPFS-GAK/PPFS-TF/PPFS-TK77.7875.0075.0075.0083.3379.1782.3180.3677.7870.8376.7872.31
August S.PWithout68.3866.9766.2066.5855.3754.6550.7852.6594.9294.6994.9594.82
PPFS-GAF65.5464.6562.4863.5456.552.2050.9651.5797.1897.0796.8796.97
PPFS-GAK/PPFS-TF/PPFS-TK62.7160.4058.8459.6161.0257.6856.7057.1895.4895.8294.6895.25
Fifa2019Without60.0450.0351.9337.9360.0950.00NANNAN59.4350.0750.4840.49
PPFS-GAF60.1250.0957.3437.9060.0950.00NANNAN59.2249.9349.5840.50
PPFS-GAK/PPFS-TK60.7049.9447.0838.1460.8550.00NANNAN59.9850.2651.2842.01
PPFS-TF60.1050.0956.9837.9160.0850.00NANNAN60.0850.0031.38NAN
Table 7. The Classification accuracy provided by SVM Classifier after applying the 4 PPFS, CFS, and PCA feature selection techniques.
Table 7. The Classification accuracy provided by SVM Classifier after applying the 4 PPFS, CFS, and PCA feature selection techniques.
SVM AccuracyPPFS-GAF (%)PPFS-GAK (%)PPFS-TF (%)PPFS-TK (%)CFS (%)PCA (%)GainRatio (%)MutualInfo (%)GiniIndex (%)
Breast Cancer96.1798.5698.5698.5695.8094.9690.9085.6585.65
Liver Disorder65.0573.7965.0573.7958.1259.8359.2258.2553.40
Sonar82.2685.5886.0687.9890.7693.2853.2377.4277.42
Wine96.1596.1596.1596.1563.3878.8780.7788.4688.46
Spam93.0493.693.0493.6285.7686.4183.4796.6196.61
Waveform83.7981.6583.7981.4534.4398.3669.8566.1866.18
Vehicle83.3383.3379.3683.3365.8773.0253.9748.4155.56
Diabetes81.3081.3081.3081.3078.7077.7373.9173.9173.91
Climate91.9391.9391.9391.9391.9391.4891.9391.9391.93
WDBC95.2998.2498.2498.2495.2993.6790.5991.1891.18
Autralian83.5083.5085.9285.9279.1385.5161.6563.5963.59
Stock95.4295.4283.1086.6275.0082.7484.1584.5084.50
Glass60.3269.8469.8469.8465.0764.4965.0865.0865.08
Segment86.5889.4789.4785.1458.8382.2951.8075.9074.31
Heart82.7282.7282.7282.7282.7281.8577.7855.5655.56
Ionosphere80.7790.3880.7781.7382.6988.0375.0064.4264.42
Colon83.3383.3383.3383.3381.7380.65---
August S.P56.5061.0261.0261.0244.0752.1944.0648.5948.59
Fifa201960.0960.8560.0860.8560.0960.0860.0960.0960.09
Table 8. The statistical test results of GAK and the 5 filter FS using 18 datasets.
Table 8. The statistical test results of GAK and the 5 filter FS using 18 datasets.
PPFS-GAKCFSPCAGainRatioMutualInfoGiniIndex
p _ V a l u e -0.0430.3250.0070.0080.009
Median84.576.882.571.970.070.0
Mean84.372.680.370.472.072.0
Table 9. The classification accuracy provided by the k-NN classifier after applying the four PPFS, WOA, PSO, and GA feature selection.
Table 9. The classification accuracy provided by the k-NN classifier after applying the four PPFS, WOA, PSO, and GA feature selection.
kNN/PPFSPPFS-GAF (%)PPFS-GAK (%)PPFS-TF (%)PPFS-TK (%)WOA (%)PSO (%)GA (%)
Breast Cancer97.6196.8596.8596.8596.8395.495.5
Sonar75.8166.1375.8167.7491.974.072.6
Wave F.82.4582.1282.4581.9975.3376.176.7
Wine98.0896.1596.1596.1595.9095.093.30
WDBC95.2998.2498.2498.2497.1094.1093.80
Heart81.4881.4881.4881.4880.7078.4082.20
Ionosphere87.5081.7384.6284.6292.5684.3083.40
Colon72.2277.7877.7877.7890.9084.3083.40
Table 10. The statistical results of GAF, WOA, PSO, and GA using 4 datasets.
Table 10. The statistical results of GAF, WOA, PSO, and GA using 4 datasets.
PPFS-GAFWOAPSOGA
Median86.5888483
Mean86.586.884.885
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Larabi-Marie-Sainte, S. Outlier Detection Based Feature Selection Exploiting Bio-Inspired Optimization Algorithms. Appl. Sci. 2021, 11, 6769. https://doi.org/10.3390/app11156769

AMA Style

Larabi-Marie-Sainte S. Outlier Detection Based Feature Selection Exploiting Bio-Inspired Optimization Algorithms. Applied Sciences. 2021; 11(15):6769. https://doi.org/10.3390/app11156769

Chicago/Turabian Style

Larabi-Marie-Sainte, Souad. 2021. "Outlier Detection Based Feature Selection Exploiting Bio-Inspired Optimization Algorithms" Applied Sciences 11, no. 15: 6769. https://doi.org/10.3390/app11156769

APA Style

Larabi-Marie-Sainte, S. (2021). Outlier Detection Based Feature Selection Exploiting Bio-Inspired Optimization Algorithms. Applied Sciences, 11(15), 6769. https://doi.org/10.3390/app11156769

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop