Next Article in Journal
Algorithm to Correct Measurement Offsets Introduced by Inactive Elements of Transducer Arrays in Ultrasonic Flow Metering
Previous Article in Journal
MWA Performed at 5.8 GHz through ‘Side Firing’ Approach: An Exploratory Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RHSOFS: Feature Selection Using the Rock Hyrax Swarm Optimization Algorithm for Credit Card Fraud Detection System

by
Bharat Kumar Padhi
1,
Sujata Chakravarty
1,
Bighnaraj Naik
2,
Radha Mohan Pattanayak
3,* and
Himansu Das
4
1
Department of Computer Science & Engineering, Centurion University of Technology & Management, Bhubaneswar 761211, Odisha, India
2
Department of Computer Application, VSSUT, Burla 768018, Odisha, India
3
School of Computer Science & Engineering, VIT-AP University, Amravati 522237, Andhra Pradesh, India
4
School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar 751024, Odisha, India
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(23), 9321; https://doi.org/10.3390/s22239321
Submission received: 29 September 2022 / Revised: 17 November 2022 / Accepted: 17 November 2022 / Published: 30 November 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In recent years, detecting credit card fraud transactions has been a difficult task due to the high dimensions and imbalanced datasets. Selecting a subset of important features from a high-dimensional dataset has proven to be the most prominent approach for solving high-dimensional dataset issues, and the selection of features is critical for improving classification performance, such as the fraud transaction identification process. To contribute to the field, this paper proposes a novel feature selection (FS) approach based on a metaheuristic algorithm called Rock Hyrax Swarm Optimization Feature Selection (RHSOFS), inspired by the actions of rock hyrax swarms in nature, and implements supervised machine learning techniques to improve credit card fraud transaction identification approaches. This approach is used to select a subset of optimal relevant features from a high-dimensional dataset. In a comparative efficiency analysis, RHSOFS is compared with Differential Evolutionary Feature Selection (DEFS), Genetic Algorithm Feature Selection (GAFS), Particle Swarm Optimization Feature Selection (PSOFS), and Ant Colony Optimization Feature Selection (ACOFS) in a comparative efficiency analysis. The proposed RHSOFS outperforms existing approaches, such as DEFS, GAFS, PSOFS, and ACOFS, according to the experimental results. Various statistical tests have been used to validate the statistical significance of the proposed model.

1. Introduction

Feature selection (also known as variable selection) is an important topic in the field of data mining. It is motivated by the requirement to choose the “optimal” selection of variables for prediction. The purpose of FS is to find the “best” subsets of features (or variables) for statistical analysis or to build a machine learning model [1,2]. Preprocessing is frequently required in FS to help the classification, prediction, or clustering stages better distinguish or represent the data and different approaches to be followed for feature selection [3,4]. In data mining and machine learning applications, FS is a critical activity that eliminates unnecessary and redundant characteristics and improves learning performance [5,6]. FS decreases dimensionality, removes irrelevant input, improves learning accuracy, and improves result comprehension as a preprocessing step for machine learning [7]. The difficulty for a learning algorithm is focusing its attention on a subset of features while ignoring the rest of the problem. As a result, processing and analyzing such large amounts of data is quite difficult. Without the use of an automated system, extracting valuable information from enormous amounts of data is a difficult task. FS is essential for detecting credit card fraud in large, multi-dimensional [8], and imbalanced datasets [9]. Many optimization algorithms have been used in the past decades to solve the FS algorithm by creating a subset of important features from dimensional datasets [10,11,12,13,14,15,16,17,18,19,20,21,22].
Filtering and wrapping are two types of FS approaches. If the FS technique is unaffected by the learning algorithm, then it is referred to as a filter approach; otherwise, it is referred to as a wrapper approach. The filter method is faster than the wrapper method in terms of processing time. On the other hand, the filter technique has the major disadvantage of being susceptible to inductive biases in the learning algorithms used to build the classifier. The wrapper approach has a higher processing overhead because it uses learning algorithms to evaluate a subset of features. However, in terms of accuracy, the wrapper strategy may outperform the filter method [1,2,10,11]. A preprocessing step is used in the filter technique [1,3,11] to select the best features. The filter approach’s fundamental flaw is that it completely disregards the impact of the selected feature subset on the induction algorithm’s performance. The wrapper methodology [1,3,11] introduced by Kohavi and John in 1997 is a simple and effective method for dealing with the problem of variable selection. The feature subset selection algorithm is used to wrap around the induction process in the wrapper approach. As part of the function evaluating feature subsets, the feature subset selection algorithm searches for a good subset using the induction process. The wrapper approach works on the simple principle of treating the induction process as a black box. The feature subset selection uses the induction technique as a black box in the wrapper approach, as shown in Figure 1. (i.e., no knowledge of the algorithm is needed, just the interface). The feature subset selection algorithm employs the induction algorithm as part of the evaluation function to conduct a search for a good subset. The wrapper-based FS technique is used in this article for optimal FS. For each individual combination of features, FS employs a search strategy to identify the best-suited features. The number of features increases exponentially with the number of viable solutions in the classic search approach [1,3,10,11].
Grasshopper Optimization [12], Differential Evolution algorithm (DE) [17], Genetic Algorithm (GA) [18], Particle Swarm Optimization (PSO) [19], and Ant Colony Optimization (ACO) [20,21] have all been used to solve the FS problem using wrapper-based FS methods. The disadvantage of this FS method is that it necessitates the tuning of various parameters for better performance. In this context, this paper proposes a new wrapper-based FS approach based on the Rock Hyrax Optimization (RHO) algorithm [23], which can detect credit card fraud [24,25,26,27,28,29,30] in massive and high-dimensional datasets and is considered very important for improving classification performance and fraud detection processes. This method also identifies FS models with small abnormalities in large datasets [31,32] with high precision and focuses on low computation that does not require extensive model-specific parameter settings. It finds the most necessary and pertinent features by upgrading the worst features using the RHO algorithm’s strength. These important characteristics are appropriate for developing a supervised machine-learning model for categorization.
Table 1 demonstrates that the Internet Crime Complaint Center (IC3) was formed in May 2000 to process complaints concerning internet crime as defined by the Certified Fraud Examiner Association (CFEA) [33]. Figure 1 and Figure 2 below display aggregate complaints and loss data annually from 2016 to 2020. During that time, IC3 received a total of 2,211,396 complaints and USD 13.3 billion in revenue losses.
The main contributions of this study are that it presents a novel FS method based on the Rock Hyrax Swarm Optimization algorithm, as well as a detailed experimental comparison of numerous FS approaches such as DEFS, GAFS, PSOFS, ACOFS, and RHSOFS. Thus, it also presents optimum features for creating an effective credit card fraud detection system.
Several performance measures have been applied to FS approaches, and performance evaluation was carried out using extensive experiments on credit card datasets using classification algorithms such as NB, SVM [21,25,32], KNN, and DT, and the results were compared to show that the experimental data were significant. The key consequences of this presented technique are that it reduces the overfitting concerns that arise when datasets are imbalanced [27] and increases the model’s generalizability. Intrusion detection, spam mail detection systems, important medical disease classification, sophisticated picture classification, and industrial automation systems are examples of applications that require large and complex data processing.
The remaining sections of the paper are organized as follows. Section 2 discusses the related work on feature selection algorithms and their impact on application research areas. Section 3 briefly discusses feature subset selection modeling and key problem formulation. Section 4 thoroughly describes the proposed RHOFS methodology. Section 5 presents experimental and statistical result analysis of various FS methods, as well as key issue discussion, and Section 6 concludes with futile suggestions and future work.

2. Literature Review

In the paper [1], the relationship between optimal feature subset selection and relevance was investigated, as well as the wrapper method to feature subset selection utilizing naïve Bayes and decision trees. This paper [2] discusses the fundamental difficulties in FS, such as feature relevance, redundancy, the characteristics and performance of different FS methods, and how to choose the best method for a given application. In the paper [3], the proper definitions of the objective function, as well as feature creation, were explored. It also looked at feature ranking and multivariate FS, as well as efficient search algorithms and ways for determining feature validity. The paper [4] focused on various typical methods of FS and extraction, as well as comparisons of each method. The paper [5] provided a detailed assessment of semi-supervised FS strategies, outlining the benefits and drawbacks of each method. The paper [7] focused on the filer model and created a new FS method that can successfully remove both unnecessary and redundant features while being less computationally expensive than existing algorithms. The paper [9] provided a wrapper-based FS strategy for selecting the most relevant features based on the artificial electric field optimization algorithm. A new method for selecting features in biomedical data is proposed in the paper [11]. A novel FS technique based on the real-valued grasshopper optimization algorithm was proposed in this paper [12]. The jaya optimization algorithm [15] has been used to construct a unique and prominent wrapper-based FS model with an emphasis on a low-computing FS model that does not require sophisticated algorithm-specific parameter tuning. The goal of the paper [16] was to find a way to reduce the time complexity of wrapper-based FSS with an embedded K-Nearest-Neighbor (KNN) classifier by building a classifier distance matrix and incrementally updating it to speed up the calculation of relevance criteria in evaluating the quality of candidate features. The paper [17] proposed a Differential Evolution (DE) optimization technique for FS, and the result of DE was compared with GA and PSO. The use of genetic algorithms to solve the feature subset selection problem using neural network classifiers was presented in the paper [18]. The implementation of FS in intrusion detection in wireless sensor networks is provided in the paper [19], which is based on PSO and PCA space. The outcome of the approach is compared to that of GA. The Ant Colony Optimization technique for FS is provided in the paper [20,21]. The paper [23] suggested a new swarm intelligence technique based on the behavior of rock hyrax swarms. The proposed algorithm can also balance the exploration and exploitation phases, making it suitable for a wide range of optimization problems. In the paper [24], an improved Credit Card Risk Identification (CCRI) technique for detecting fraud risk is described, which is based on feature selection algorithms such as Random Forest and Support Vector Machine (SVM) classifier. The paper [25] describes an SVM-type FS strategy that uses artificial variables and mutual information to filter out noisy variables from high-dimensional metabolome data. In the paper [26], they presented a credit card fraud detection model and the necessity of using a feature selection approach. Some of the most prominent supervised and unsupervised machine learning algorithms were used to detect credit card thefts in a severely skewed dataset [30]. The analysis and comparison studies of different machine learning algorithms and boosting machine learning algorithms for fraud detection are discussed in papers [31,32].

3. Problem Definition

This section describes the problem formulation applied in this paper. To increase the performance of classification models, FS refers to a method for selecting an optimum subset of input features from the entire dataset. It simply picks out the elements that matter in the decision-making process. To lower the computing cost of the problem, it generates and selects the most effective subset of features by removing redundant and irrelevant features. This is an NP-hard problem [13], meaning it cannot be solved in polynomial time. The goal is to obtain the best subset of features to increase the classification process’ performance. The following are the four steps that make up the best FS: (i) a subset of features is generated; (ii) using these subsets of characteristics, evaluate and compare fitness levels; (iii) verify that the termination conditions have been met, if not, repeat steps (i) and (ii); validate the results using the best subset of characteristics.
The problem formulation for FS is performed by selecting d important features from a set of D features, which is represented in Equation (1), shown below:
f ( x ) = min _ e r r ( d )   and   d D
Minimize   f x ,   Subject   to   Condition , x = | D |   and   x 0 .
Using the optimized subset features, Equation (1) reduces error in each iteration stage, thus increasing classification accuracy in the proposed model.

4. Proposed Model

RHSO (Rock Hyrax Swarm Optimization) is a meta-heuristic based on the natural behavior of rock hyrax swarms. The RHSO algorithm simulates the collective behavior of rock hyraxes to find food and their unique way of looking at it. Rock hyraxes live in colonies or groups, with a dominant male keeping a close eye on the colony to ensure its protection. The algorithm seeks out the best solutions by incorporating both local heuristics and prior knowledge into the construction of the best subset of features to improve the classification process’ performance [23].
The RHSOFS detailed functioning model is depicted in Figure 3, which separates the whole dataset into training and testing sets. The training data are entered into the optimization technique. (i.e., f(x)) to find the best suitable optimum features. The classification algorithm is fed the optimum subset of features (i.e., f(x)), train, and test data to evaluate the model’s performance. Equation (1) can be used to represent the selection of the most optimal features. Equation (2) reduces the error in each iteration using the specified characteristics, increasing the classification accuracy in the process.
Population size, generation number, initial weighting factors, cognitive and social scaling factors, and probabilities of mutation and crossover are some of the regulatory parameters for population-based algorithms. To achieve the best results, these parameters must be fine-tuned, and the performance of optimization methods is determined by parameter fine-tuning; otherwise, these parameter values may end up in the optimization algorithm in a local optimum stage, increasing the computational cost of the optimization algorithm problem. To address the aforementioned concerns, the RHOSFS technique is applied. This approach is applied to create the best subset of input features to increase the efficiency of the classification process. The RHSOFS comprehensive functional model is depicted in Figure 3. Below is a detailed description of the proposed RHOSFS approach.
First, generate, select, and examine a random sample of a binary (0, 1) population for the total number of input features for FS. Create a feature subset that is equal to 1 for each representation of the input population. For the computation of fitness, the extracted optimal input features are fed into classification models such as NB, SVM, KNN, and DT to compute fitness. The goal of this research work is to find the best subset of input features that reduce the model’s fitness while also improving its accuracy.
e r r x i   = a c t u a l _ o u t p u t   x i   m o d e l _ e s t i m a t e d _ o u t p u t ( x i )
f i t n e s s ( x ) = x = 0 n e r r ( x ) n
l e a d e r = r 1 × x l e a d e r p o s , j
where r 1 denotes a random number between [0, 1], x is the previous position of the leader, l e a d e r p o s denotes the old position of the leader, and j refers to “each diminution”. After the leader’s position is updated, all members (or search agents) update their positions using Equation (5).
x i , j = x i , j c i r c × x i , j + l e a d e r
where circ denotes circular motion, it is calculated as follows to try to replicate the circle system in Equation (6):
c i r c = s q r t ( n 1 2 + n 2 2 )
n 1 = r 2 × cos a n g
n 2 = r 2 × sin a n g
where r 2 is the radius and is a random number between [0, 1], and a n g denotes the angle of a move and is a random value between [0, 360] in Equation (7) and (8). Every generation, the a n g is updated as well, and this update is based on the lower and upper bounds of the variables, where lb and ub are the lower and upper bands of the random number generator, respectively.
d a l t a = r a n d o m l b , u b
a n g = a n g + d a l t a
If the value of the output grows larger than 360, or less than 0, the angle (ang) can be set to 360 or 0 to keep it within the desired range.
Algorithm 1 explains the RHSOFS pseudo-code. The RHSOFS begins by producing a binary population of P agents at random and examining all of the features. For each instance of the population to be studied, create a feature subset equal to 1. These chosen attributes are fed into classification models in order to calculate fitness value. Equation (2) calculates the err(x) by the differences between an actual and a predicted value of the model, where x = 1, 2, …, n and n is the number of testing observations.
The model’s fitness is calculated by dividing the sum of errors by the number of observations, as shown in Equation (3). The algorithm then attempts to update the position of the Leader according to Equation (4) and the position of each search agent according to Equation (5). Then, using Equation (3), determine each search agent’s new fitness. According to Equations (9) and (10), this algorithm progresses toward angle updating. The bestX persons are those who have the lowest fitness value. This algorithm then tries to update each search agent’s position in accordance with Equation (5).
The new individuals are chosen only if their new fitness value is greater than or equal to their prior fitness value, and their new fitness value is flipped. For the next generation, only those with the lowest fitness value are chosen. Finally, the algorithm selects the most suitable candidates.
Algorithm 1 Proposed RHSOFS Algorithm
  • Create an initial population of 0 & 1 of P agents randomly.
  • Set the dimension of the problem, D = P, where P is the number of agents.
  • Set Low to 1 and High to D, where Low and High refer to the low and high dimensions, respectively.
  • Generate the value of r1 and r2, where r1 is a random number (0, 1) and r2 is a random radius (0, 360).
  • Generate training and testing data.
  • Set max_iter = maximum number of iterations.
  • Calculate each agent’s fitness using Equation (3).
  • Set Leader = the best agent.
  • Set t = 1.
  • while (t < max_iter)
  •  for (i = 1 to n) do
  •   Update Leader position, according to Equation (4).
  •   Update the position of each search agent according to Equation (5).
  •   Calculate Newfitness of each search agent using Equation (3).
  •   Select the best member of the population → bestX = X (min (fitness))
  •   Update the angle according to Equations (9) and (10).
  •   If (Newfitness (i) <= fitness (i); then
  •    Update the position of each search agent according to Equation (5).
  •    fitness (i) = Newfitness (i).
  •   end if
  •  end for
  •  t = t + 1.
  • end while
  •   Return the best agent

5. Experimental and Statistical Result Analysis and Discussion

To evaluate and examine the performance and effectiveness of the proposed FS approach, called RHSOFS, we have compared it with other useful approaches such as DEFS, GAFS, PSOFS, and ACOFS. Numerical experiments have been conducted on a real-world credit card fraud dataset using a range of data mining approaches to test the efficiency of the presented approach. The stratified cross-validation approaches have been used to create ten identical datasets due to a shortage of real credit card fraud datasets. The stratified cross-validation (SCV) method, which is related to the k-fold cross-validation method, is used to provide training as well as test indices for dividing the entire dataset into train and test sets by keeping the percentage of samples for each class the same for each fold. The stratified cross-validation approach is used for classification problems and when the dataset is imbalanced. Imbalanced datasets can create overfit results. The SCV technique creates new datasets by preserving the target class ratio in each fold the same as it is in the full dataset rather than randomly splitting the entire dataset.
Steps followed to create ten identical datasets using a stratified cross-validation approach:
  • Initially, the entire original dataset has been randomly shuffled;
  • The randomly shuffled dataset is split into k folds (we have set k = 10);
  • For each fold, the training dataset samples are selected by using stratified sampling in order to maintain the class distribution as per the original datasets. The test set is formed from the remaining datasets. This process is repeated for each fold.
In September 2013, European cardholders using credit cards collected the dataset in two days, and it is now available via the ULB machine learning group. This dataset is transformed using the PCA approach, which has 28 principal components or features spanning from V1 to V28. However, only 30 are included in the evaluation. There are a total of 284,807 transactions in this dataset, with 492 of them being fraud transactions, making it highly lopsided and skewed toward fraud [32].
The tests are performed on a PC with the following specifications: a 1.60 GHz Intel Core i5-8250U processor and 8 GB of RAM. Matlab 2014b is used to implement these approaches. The size of the population (Pop) and the maximum number of possible generations (MaxGen) are the variables that are used to train and test the model in the experiment.
The values of Pop and MaxGen have been set to 10 and 100, respectively, for the optimum overall performance. This section examines the performance of DEFS, GAFS, PSOFS, ACOFS, and RHSOFS techniques with respect to the number of selected features and the accuracy achieved by each technique. Because the procedures are stochastic, ten trials were carried out with a random sample population. The average classification accuracy for each dataset and FS approach is shown in Table 1. The results suggest that the proposed RHSOFS approach can achieve greater values than other approaches, such as DEFS, GAFS, PSOFS, and ACOFS, with the best optimal features. There are significantly fewer selected features compared to the original input features, as shown in Table 2. In comparison to other existing approaches (DEFS, GAFS, PSOFS, and ACOFS), which produce similar accuracy values with few variations, the new RHSOFS methodology produces significant accuracy results for almost all datasets using Equation (11).
Accuracy = T P + T N T P + T N + F P + F N
The average accuracy (%) over the 10 datasets using NB, KNN, SVM, and DT classifiers has been studied and analyzed, as depicted in Figure 4, applied over the FS approaches such as DEFA, GAFS, PSOFS, ACOFS, and RHSOFS, respectively. The recall comparison, as shown in Figure 5, identifies the effectiveness of the proposed method by comparing the recall values of all the models over each dataset. The results show that the proposed RHSOFS approaches have outperformed other FS approaches.
The goal of FS approaches such as DEFA, GAFS, PSOFS, ACOFS, and RHSOFS is to locate the best features and cut down on execution time to build a reliable credit card fraud detection system. For most FS algorithms, controlling elements such as the size of the population and the number of iterations are considered 10 and 100, respectively. Only approach-specific regulating parameters, such as probabilities of mutation, crossover, selection operators, initial weighting factors, and cognitive and social scaling factors, differ across all other FS techniques.
Table 3 and Table 4 show performance indicators for all datasets, including precision, recall, f1-score, Matthews correlation coefficient (MCC), and specificity with and without the feature (WTFS) subset selection. Finally, the proposed approach has a significantly lower number of selected features than previous approaches, resulting in a significant improvement in classification accuracy. Different performance measures from the confusion matrix include classification accuracy (Equation (11)), precision (Equation (12)), recall (Equation (13)), f-measure (Equation (14)), MCC (Equation (15)), and specificity (Equation (16)). True positive, true negative, false positive, and false negative are represented by the letters TP, TN, FP, and FN, respectively.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + T N
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
M C C = T P × T N F P × F N T P + F P T P + F N T N + F P T N + F N
S p e c i f i c i t y = T N T N + F P
The performance of the proposed RHSOFS model has been compared to that of other models using statistical analysis [14]. The Friedman test is a non-parametric statistical method for analyzing the results of models. Friedman proposes two hypotheses (H0 and H1), with H0 implying that there is no significant variance among all approaches and that all approaches are considered equivalent, whereas H1 implies the opposite. The Friedman test is one of the best techniques to determine the importance of statistics across all approaches, and an individual approach is rated based on its accuracy.
This test has been conducted by considering the smallest number is assigned the highest rank, while the highest number is assigned the lowest rank. Table 5 shows the average rank of many models in relation to four different classifiers (NB, KNN, DT, and SVM) using Equation (17). It divides the number of classifiers by the total of all of their ranks. Similarly, Equation (18) is used to determine the average rank of various models in relation to the datasets shown in Table 6. The average rankings of all models are calculated by dividing the total number of datasets by the summation of the average rank (AR) of all models. Finally, the AR of all models represents RHSOFS (AR1 = 1.525), GAFS (AR2 = 1.725), PSOFS (AR3 = 3.2), ACOFS (AR4 = 3.85), DEFS (AR5 = 3.975), and WTFS (AR6 = 5.825).
A R M o d e l s = S u m m a t i o n   o f   r a n k   o f   a l l   c l a s s i f i e r s N o . o f   c l a s s i f i e r s
A R D a t a s e t s = S u m m a t i o n   o f   A R M o d e l s N o . o f   d a t a s e t s
X F 2 = 12 × N P P + 1 × j A R 2 P × P + 1 2 4
F F = N 1 × X F 2 N × P 1 X F 2
Using Equation (19), the Friedman test statistic has a chi-square ( X F 2 ) distribution with (P − 1) degrees of freedom. Whereas Equation (20) yields a value of 5.442917 for the Friedman test statistic (FF), where N is the number of datasets and P denotes the number of models that were employed in this experiment, The number of datasets used in this study is ten, the number of models employed is six, and the threshold of significance of a is 0.05 with degrees of freedom (5, 45). The null hypothesis is rejected because the crucial value of FF is 2.42, which is smaller than the Friedman statistics of FF = 5.442917.
z = ( A R i - A R j ) P ( P + 1 ) / 6 N
The number of algorithms employed in this experiment is P, z is used to calculate the z-score value using Equation (21), and the number of datasets is N. ARi and ARj denote the average rank of the ith and jth models, respectively. All the models are compared to the suggested model using the z-value, p-value, and α/(P − i), and the results are shown in Table 7.
There may be some insignificant, duplicated, or noisy data in the feature set, which increases processing time while also affecting the model’s performance. Only optimal features are processed by the model, and all unwanted, redundant, and noisy characteristics are eliminated. These relevant features boost the model’s performance while cutting down on computation time. FS methods, on the other hand, pick a subset of significant and relevant original features. This paper compares the model’s performance, including NB, KNN, SVM, and DT, using wrapper-based FS methods such as DEFS, GAFS, PSOFS, ACOFS, and RHSOFS. To compute to select the best features, modeling FS processes can be used in two ways: first, to choose fixed optimum features, and second, to select variable optimum features. It is a difficult task to determine the set number of optimum features for all models in the fixed optimal FS technique. The dataset used in the experiment was already processed using the PCA method, which contains 28 principal components or features ranging from V1 to V28. There may be some features that are duplicated. Another difficult task is to reduce the redundant features and replace redundant features with the next most important features. On the other hand, distinct FS techniques based on several optimization algorithms select the fewest number of variable optimum features. Individual features are assigned a value of 1 or 0 in this method. The goal of the presented RHSOFS approach is to remove irrelevant features by selecting the most relevant and appropriate features to improve the model’s performance. The number of relevant features used is determined by the performance of the optimization methods. Furthermore, applying different optimization techniques does not guarantee that you will select the same number of optimum features.
Finally, the model has learned to use these selected relevant features gained by optimization procedures. The goal of this paper is to boost the model’s performance by optimizing the number of relevant features. A scale limit can often be provided based on the rank of the relevant features or the number of features to be chosen in a fixed optimal FS technique, ensuring that the predefined number of features is selected. The comparison is based on the best FS using different optimization algorithms.

6. Conclusions

To determine an optimal subset of features, a new unique FS technique depending on the RHO algorithm named RHSOFS, has been presented. The RHSOFS approach is used to explore the most relevant features by updating the irrelevant, redundant, and noisy features. Four classifiers, such as NB, KNN, SVM, and DT, have been employed on ten datasets for evaluating the efficacy of the proposed RHSOFS approach.
The proposed RHSOFS strategy effectively reduces duplicate features and outperforms existing approaches, according to the experimental data. In terms of classification accuracy and the number of characteristics chosen, all the models are compared. The data points that are insignificant, duplicated, or noisy should be removed. This could result in data loss, which is a drawback of this method. Certainly, the use of AI techniques for the prediction of future behaviors can have good results, although, in the field of cybercrime, it is more complex since it is trained with some acquired knowledge, not counting on the evolution in the learning of the offender. Although this article presents a very interesting proposal, we would recommend taking into account works such as “Evolution Oriented Monitoring oriented to Security Properties for Cloud Applications” in the sense of providing applications with the ability to evolve securely by integrating acquired knowledge. On the other hand, it would be interesting to study how to endow Trusted Computing-type trusted hardware with this type of intelligence so that they can provide hybrid hardware–software certification mechanisms in different scenarios, as proposed in “Software and Hardware Certification Techniques in a Combined Certification Model”. This proposed strategy could even be applied to a wide range of complex and different applications with many features, such as mail fraud detection, intrusion detection, and fake insurance analysis.

Author Contributions

Conceptualization, B.K.P.; Methodology, B.K.P.; Software, R.M.P.; Formal analysis, S.C. and H.D.; Investigation, B.K.P.; Writing—review & editing, B.N. and R.M.P.; Supervision, S.C., B.N. and H.D.; Funding acquisition, R.M.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded VIT-AP University, Amravati-522237, Andhra Pradesh, India.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  2. Huang, S.H. Supervised feature selection: A tutorial. Artif. Intell. Res. 2015, 4, 22–37. [Google Scholar] [CrossRef] [Green Version]
  3. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar] [CrossRef]
  4. Ghojogh, B.; Samad, M.N.; Mashhadi, S.A.; Kapoor, T.; Ali, W.; Karray, F.; Crowley, M. Feature selection and feature extraction in pattern analysis: A literature review. arXiv 2019, arXiv:1905.02845. [Google Scholar]
  5. Sheikhpour, R.; Sarram, M.A.; Gharaghani, S.; Chahooki, M.A.Z. A Survey on semi-supervised feature selection methods. Pattern Recognit. 2017, 64, 141–158. [Google Scholar] [CrossRef]
  6. Molina, L.C.; Belanche, L.; Nebot, À. Feature selection algorithms: A survey and experimental evaluation. In Proceedings of the 2002 IEEE International Conference on Data Mining, Maebashi City, Japan, 9–12 December 2002; pp. 306–313. [Google Scholar] [CrossRef] [Green Version]
  7. Yu, L.; Liu, H. Feature Selection for High-Dimensional Data: A Fast Correlation-Based Filter Solution. In Proceedings of the 20th International Conference on Machine Learning (ICML-03), Washington, DC, USA, 21–24 August 2003; Volume 2, pp. 856–863. [Google Scholar]
  8. Gheyas, I.A.; Smith, L.S. Feature subset selection in large dimensionality domains. Pattern Recognit. 2010, 43, 5–13. [Google Scholar] [CrossRef] [Green Version]
  9. Van Hulse, J.; Khoshgoftaar, T.M.; Napolitano, A. Experimental perspectives on learning from imbalanced data. In Proceedings of the 24th International Conference on Machine Learning, New York, NY, USA, 20–24 June 2007; Volume 227, pp. 935–942. [Google Scholar] [CrossRef]
  10. Peng, Y.; Wu, Z.; Jiang, J. A novel feature selection approach for biomedical data classification. J. Biomed. Inform. 2010, 43, 15–23. [Google Scholar] [CrossRef] [Green Version]
  11. Das, H.; Naik, B.; Behera, H.S. Optimal Selection of Features Using Artificial Electric Field Algorithm for Classification. Arab. J. Sci. Eng. 2021, 46, 8355–8369. [Google Scholar] [CrossRef]
  12. Zakeri, A.; Hokmabadi, A. Efficient feature selection method using real-valued grasshopper optimization algorithm. Expert Syst. Appl. 2019, 119, 61–72. [Google Scholar] [CrossRef]
  13. Drotár, P.; Gazda, J.; Smékal, Z. An experimental comparison of feature selection methods on two-class biomedical datasets. Comput. Biol. Med. 2015, 66, 1–10. [Google Scholar] [CrossRef]
  14. Caruana, R.; Niculescu-Mizil, A. An empirical comparison of supervised learning algorithms. In Proceedings of the 23rd International Conference on Machine Learning, New York, NY, USA, 25–29 June 2006; Volume 2006, pp. 161–168. [Google Scholar]
  15. Das, H.; Naik, B.; Behera, H.S. A Jaya algorithm based wrapper method for optimal feature selection in supervised classification. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 3851–3863. [Google Scholar] [CrossRef]
  16. Wang, A.; An, N.; Chen, G.; Li, L.; Alterovitz, G. Accelerating wrapper-based feature selection with K-nearest-neighbor. Knowl.-Based Syst. 2015, 83, 81–91. [Google Scholar] [CrossRef]
  17. Khushaba, R.N.; Al-Ani, A.; Al-Jumaily, A. Differential Evolution based feature subset selection. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, J.; Honavar, V. Feature Subset Selection Using A Genetic Algorithm Feature Subset Selection Using 1 Introduction. Intell. Syst. Appl. 1997, 13, 44–49. [Google Scholar] [CrossRef] [Green Version]
  19. Ahmad, I. Feature selection using particle swarm optimization in intrusion detection. Int. J. Distrib. Sens. Netw. 2015, 2015. [Google Scholar] [CrossRef]
  20. Ahmed, A.-A. Feature subset selection using ant colony optimization. Int. J. Comput. 2005, 2, 53–58. Available online: http://epress.lib.uts.edu.au/research/handle/10453/6002 (accessed on 1 January 2005).
  21. Deriche, M. Feature selection using ant colony optimization. In Proceedings of the 2009 6th International Multi-Conference on Systems, Signals and Devices, Djerba, Tunisia, 23–26 March 2009. [Google Scholar] [CrossRef]
  22. Urbanowicz, R.J.; Meeker, M.; la Cava, W.; Olson, R.S.; Moore, J.H. Relief-based feature selection: Introduction and review. J. Biomed. Inform. 2018, 85, 189–203. [Google Scholar] [CrossRef]
  23. Al-Khateeb, B.; Ahmed, K.; Mahmood, M.; Le, D.N. Rock Hyraxes Swarm Optimization: A New Nature-Inspired Metaheuristic Optimization Algorithm. Comput. Mater. Contin. 2021, 68, 643–654. [Google Scholar] [CrossRef]
  24. Rtayli, N.; Enneya, N. Selection features and support vector machine for credit card risk identification. Procedia Manuf. 2020, 46, 941–948. [Google Scholar] [CrossRef]
  25. Lin, X.; Yang, F.; Zhou, L.; Yin, P.; Kong, H.; Xing, W.; Lu, X.; Jia, L.; Wang, Q.; Xu, G. A support vector machine-recursive feature elimination feature selection method based on artificial contrast variables and mutual information. J. Chromatogr. B Anal. Technol. Biomed. Life Sci. 2012, 910, 149–155. [Google Scholar] [CrossRef]
  26. Bhattacharyya, S.; Jha, S.; Tharakunnel, K.; Westland, J.C. Data mining for credit card fraud: A comparative study Decis. Support Syst. 2011, 50, 602–613. [Google Scholar] [CrossRef]
  27. Bahnsen, A.C.; Aouada, D.; Stojanovic, A.; Ottersten, B. Feature engineering strategies for credit card fraud detection. Expert Syst. Appl. 2016, 51, 134–142. [Google Scholar] [CrossRef]
  28. Bolton, R.J.; Hand, D.J.; Hand, D.J. Unsupervised Profiling Methods for Fraud Detection. In Credit Scoring Credit Control VII; University of Edinburgh: Edinburgh, UK, 2001; pp. 5–7. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.24.5743 (accessed on 22 September 2001).
  29. Carcillo, F.; le Borgne, Y.A.; Caelen, O.; Kessaci, Y.; Oblé, F.; Bontempi, G. Combining unsupervised and supervised learning in credit card fraud detection. Inf. Sci. 2021, 557, 317–331. [Google Scholar] [CrossRef]
  30. Mishra, A. Credit Card Fraud Detection on the Skewed Data Using Various Classification and Ensemble Techniques. In Proceedings of the 2018 IEEE International Students’ Conference on Electrical, Electronics and Computer Science (SCEECS), Bhopal, India, 24–25 February 2018; pp. 1–5. [Google Scholar] [CrossRef]
  31. Mittal, S.; Tyagi, S. Performance evaluation of machine learning algorithms for credit card fraud detection. In Proceedings of the 2019 9th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 10–11 January 2019; pp. 320–324. [Google Scholar] [CrossRef]
  32. Padhi, B.; Chakravarty, S.; Biswal, B. Anonymized credit card transaction using machine learning techniques. In Advances in Intelligent Computing and Communication Lecture Notes in Networks and Systems; Springer: Berlin/Heidelberg, Germany, 2020; Volume 109. [Google Scholar]
  33. IC3 Report. Available online: https://www.ic3.gov/Media/PDF/AnnualReport/2020_IC (accessed on 22 September 2001).
Figure 1. IC3 Report on complaints received.
Figure 1. IC3 Report on complaints received.
Sensors 22 09321 g001
Figure 2. IC3 Report on loss.
Figure 2. IC3 Report on loss.
Sensors 22 09321 g002
Figure 3. Feature Selection Models Block Diagram.
Figure 3. Feature Selection Models Block Diagram.
Sensors 22 09321 g003
Figure 4. Average accuracy (%) of all models over 10 datasets.
Figure 4. Average accuracy (%) of all models over 10 datasets.
Sensors 22 09321 g004
Figure 5. Recall comparison of all models over 10 datasets.
Figure 5. Recall comparison of all models over 10 datasets.
Sensors 22 09321 g005
Table 1. Classification accuracy (%) of all feature selection approaches.
Table 1. Classification accuracy (%) of all feature selection approaches.
Sl. NoDatasetsFeature Selection AlgorithmsNBKNNDTSVMNo. of Selected Features
1D1WT FS97.0224599.4503099.4044996.1978928
DEFS97.9798499.5006999.4777898.5661921
GAFS99.0380299.6335399.4503099.0838314
PSOFS98.6257499.6060599.4503098.8089815
ACOFS98.8089899.4961199.4961197.8011914
RHSOFS98.8547999.6793499.6335399.2670613
2D2WTFS97.7095799.3586899.0380296.5643628
DEFS98.5707799.4884799.4895698.6165817
GAFS99.4044999.5877299.5419199.1754513
PSOFS98.9006099.5190199.5266598.8089814
ACOFS99.3037199.3815899.3174598.1676614
RHSOFS98.9922199.6335399.5877299.2212612
3D3WT FS97.7553899.3586899.4503096.1978928
DEFS98.4562599.4961199.5327598.4654118
GAFS99.3128799.5419199.5877298.9006012
PSOFS98.9006099.5190199.4961198.6853013
ACOFS99.1754599.4228199.5419198.0760413
RHSOFS98.9464099.6335399.6335399.2670611
4D4WT FS97.1140699.2212699.2212696.5185528
DEFS98.1081199.4732099.3999198.3234117
GAFS99.3128799.4961199.4044998.8089812
PSOFS98.8547999.3770099.3220398.5020613
ACOFS98.8089899.2991399.2670698.0760412
RHSOFS98.8547999.4961199.4503098.7631712
5D5WT FS97.4805399.4961199.1296496.1978928
DEFS97.8378499.5061199.3128798.3875424
GAFS99.0838399.5419199.3128799.0838312
PSOFS98.7631799.5261199.2441698.6715514
ACOFS98.9784799.5161199.2670698.3967012
RHSOFS98.8089899.5877299.4503099.2670612
6D6WT FS97.1140699.4503099.3586895.0984928
DEFS97.5950599.5877299.5739898.4654121
GAFS99.4503099.6793499.6793499.1754514
PSOFS98.4425199.6152199.5680998.7631714
ACOFS99.0884199.5164799.4044997.2056815
RHSOFS98.7173699.6335399.6793499.2670613
7D7WT FS97.8011999.4503099.3586896.3582228
DEFS98.2730299.5877299.3999198.5982620
GAFS99.0838399.6335399.6335398.9464014
PSOFS98.9006099.5968999.6008198.4883214
ACOFS99.3037199.4777899.4503097.9844312
RHSOFS98.9006099.7251599.6793499.2670612
8D8WT FS97.1598799.3586899.4044996.5689428
DEFS97.7370699.4961199.4961198.3921221
GAFS99.2212699.5877299.5877298.9464013
PSOFS98.6257499.5098599.5571898.7585913
ACOFS99.1388099.4044999.5419198.0760414
RHSOFS99.1754599.6335399.6793499.0380212
9D9WT FS97.8928199.4503099.3586896.7934028
DEFS98.1355999.5520999.5648298.5341321
GAFS99.0380299.6793499.5419198.9922113
PSOFS98.8089899.6014799.5953698.6257413
ACOFS99.2395899.5006999.4961197.1782012
RHSOFS98.9006099.6793499.6793498.9922112
10D10WT FS97.1127499.2667398.9000996.6590328
DEFS97.3877299.3583999.0834197.9560022
GAFS98.9642599.4042299.2667398.5334613
PSOFS98.5609599.2850699.1475798.6709414
ACOFS98.7259499.3125699.1292498.1668211
RHSOFS98.3501499.4042299.4042298.8084310
Table 2. List of the subset of selected features for each dataset.
Table 2. List of the subset of selected features for each dataset.
DatasetsNumber of Selected FeaturesSubset of Selected Features
D1133, 4, 6, 9, 11, 12, 14, 15, 16, 17, 18, 22, 25
D2124, 6, 7, 9, 10, 11, 12, 13, 16, 17, 18, 22
D3114, 5, 11, 12, 13, 15, 16, 17, 19, 25, 26
D4123, 6, 11, 12, 13, 14, 15, 16, 17, 19, 24, 26
D5123, 4, 5, 6, 9, 11, 12, 14, 15, 16, 18, 22
D6134, 7, 9, 11, 12, 13, 14, 16, 17, 23, 24, 25, 25
D7126, 11, 12, 13, 14, 15, 16, 18, 19, 22, 24, 26
D8123, 4, 9, 10, 11, 14, 16, 18, 19, 24, 25, 26
D9123, 4, 6, 9, 11, 12, 14, 16, 17, 18, 24, 26
D10101, 6, 9, 13, 14, 15, 16, 18, 24, 28
Table 3. Recall, Precision, F1-score MCC, and Specificity for all approaches with NB and KNN classifier.
Table 3. Recall, Precision, F1-score MCC, and Specificity for all approaches with NB and KNN classifier.
Sl. NoDatasetFS AlgorithmNBKNN
RecallPrecisionF1 ScoreMCCSpecificityRecallPrecisionF1 ScoreMCCSpecificity
1D 1WTFS0.44000.83020.57520.59180.99570.90200.86790.88460.88200.9967
DEFS0.55490.84910.67110.67720.99620.92870.86040.89320.89140.9965
GAFS0.75000.90570.82050.81940.99761.00000.84910.91840.91970.9963
PSOFS0.65750.90570.76190.76530.99760.96440.86980.91470.91390.9968
ACOFS0.90910.56600.69770.71220.98930.93750.84910.89110.88970.9963
RHSOFS0.71210.88680.78990.78910.99721.00000.86790.92930.93010.9967
2D 2WTFS0.46840.82220.59680.61060.99620.86050.82220.84090.83790.9963
DEFS0.61310.83110.70570.70700.99640.91770.82590.86940.86800.9963
GAFS0.86360.84440.85390.85090.99670.95000.84440.89410.89360.9967
PSOFS0.69090.84440.76000.75840.99670.91770.84220.87830.87670.9967
ACOFS0.83410.82670.83040.82680.99640.87410.81780.84500.84230.9962
RHSOFS0.71700.84440.77550.77310.99670.95120.86670.90700.90610.9972
3D 3WTFS0.46050.81400.58820.60250.99620.83720.83720.83720.83390.9967
DEFS0.57460.83260.68000.68450.99660.90820.82790.86620.86460.9965
GAFS0.83330.81400.82350.82010.99630.92310.83720.87800.87680.9967
PSOFS0.68630.81400.74470.74190.99620.91770.83020.87180.87050.9966
ACOFS0.79900.77670.78770.78360.99550.87070.83020.85000.84730.9966
RHSOFS0.70000.81400.75270.74960.99620.94870.86050.90240.90170.9972
4D 4WTFS0.39560.81820.53330.55720.99620.81400.79550.80460.80070.9958
DEFS0.51990.80000.63030.63620.99580.93920.78980.85800.85870.9957
GAFS0.82220.84090.83150.82800.99670.94590.79550.86420.86500.9958
PSOFS0.70210.75000.72530.71980.99490.89180.78640.83570.83430.9956
ACOFS0.87500.47730.61760.64130.98930.84580.79770.82110.81780.9958
RHSOFS0.67920.81820.74230.73980.99620.92310.81820.86750.86650.9963
5D 5WTFS0.50000.80000.61540.62090.99470.95830.83640.89320.89280.9958
DEFS0.55160.75820.63860.63620.99370.95120.83570.88970.88900.9957
GAFS0.84310.78180.81130.80720.99440.97070.84360.90270.90270.9960
PSOFS0.73330.80000.76520.75960.99480.93140.84820.88790.88610.9960
ACOFS0.83990.73450.78370.78030.99320.94000.85450.89520.89370.9962
RHSOFS0.74580.80000.77190.76630.99480.97920.85450.91260.91270.9963
6D 6WTFS0.42000.89360.57140.60150.99760.85710.89360.87500.87240.9977
DEFS0.46920.89150.61480.63700.99760.91300.89360.90320.90120.9977
GAFS0.85710.89360.87500.87240.99770.95450.89360.92310.92200.9977
PSOFS0.59150.89360.71190.72010.99760.91960.90000.90970.90780.9978
ACOFS0.75610.85110.80080.79760.99670.88140.89600.88860.88620.9977
RHSOFS0.65570.85110.74070.74080.99670.93330.89360.91300.91140.9977
7D 7WTFS0.48860.93480.64180.66720.99860.86960.86960.86960.86680.9972
DEFS0.55530.90650.68870.70200.99800.90440.89950.90190.89980.9978
GAFS0.72410.91300.80770.80870.99810.93180.89130.91110.90950.9977
PSOFS0.67740.91300.77780.78130.99810.90260.90650.90460.90250.9980
ACOFS0.83480.83480.83480.83120.99640.86810.88700.87740.87480.9976
RHSOFS0.68970.86960.76920.76910.99720.93480.93480.93480.93340.9986
8D 8WTFS0.40220.84090.54410.57020.99670.84090.84090.84090.83760.9967
DEFS0.46620.84550.60100.61820.99680.89660.84770.87150.86930.9969
GAFS0.77550.86360.81720.81450.99720.90700.88640.89660.89450.9977
PSOFS0.62070.81820.70590.70600.99620.90910.84090.87370.87190.9967
ACOFS0.77630.80450.79020.78590.99600.86050.84090.85060.84760.9967
RHSOFS0.76000.86360.80850.80600.99720.95000.86360.90480.90400.9972
9D 9WTFS0.57140.88140.69330.70020.99670.89830.89830.89830.89550.9972
DEFS0.60600.88640.71990.72440.99680.92840.90400.91600.91380.9973
GAFS0.77140.91530.83720.83550.99761.00000.88140.93690.93730.9967
PSOFS0.72600.89830.80300.80180.99720.95640.89320.92380.92230.9970
ACOFS0.85450.86610.86030.85640.99630.92270.88980.90600.90350.9969
RHSOFS0.74650.89830.81540.81350.99720.96430.91530.93910.93780.9976
10D 10WTFS0.46320.78570.58280.59030.99430.91670.78570.84620.84500.9944
DEFS0.49440.78750.60740.61200.99430.93750.80360.86540.86480.9948
GAFS0.79610.80180.79890.79360.99480.93880.82140.87620.87520.9953
PSOFS0.68870.80180.74090.73580.99480.91220.79820.85140.84980.9947
ACOFS0.77220.71430.74210.73620.99250.90200.82140.85980.85730.9953
RHSOFS0.64290.80360.71430.71060.99480.93880.82140.87620.87520.9953
Table 4. Recall, Precision, F1-score MCC, and Specificity for all approaches with DT and SVM classifier.
Table 4. Recall, Precision, F1-score MCC, and Specificity for all approaches with DT and SVM classifier.
Sl. NoDatasetFS AlgorithmDTSVM
RecallPrecisionF1 ScoreMCCSpecificityRecallPrecisionF1 ScoreMCCSpecificity
1D 1WTFS0.93480.81130.86870.86790.99530.38520.85450.53110.55900.9961
DEFS0.94070.83770.88620.88510.99600.65750.85470.74320.74270.9964
GAFS0.93620.83020.88000.87890.99580.80700.83640.82140.81690.9958
PSOFS0.88370.84440.86360.86110.99670.70770.86790.77970.77790.9967
ACOFS0.92000.86790.89320.89100.99670.52750.90570.66670.68190.9976
RHSOFS0.97870.86790.92000.91990.99670.87760.81130.84310.84010.9953
2D 2WTFS0.74000.82220.77890.77520.99620.36840.93330.52830.57460.9986
DEFS0.91000.83490.87090.86910.99650.62050.84670.71620.71830.9967
GAFS0.92680.84440.88370.88240.99670.76470.86670.81250.81000.9972
PSOFS0.91940.84440.88030.87870.99670.66670.84440.74510.74450.9967
ACOFS0.83970.82670.83310.82970.99640.53330.88890.66670.68060.9976
RHSOFS0.95000.84440.89410.89360.99670.79170.84440.81720.81370.9967
3D 3WTFS0.91890.79070.85000.84970.99580.32760.88370.47800.52490.9976
DEFS0.94320.81160.87250.87270.99620.58740.80230.67820.67910.9959
GAFS0.94740.83720.88890.88860.99670.67920.83720.75000.74870.9967
PSOFS0.92860.80620.86310.86270.99610.62840.81400.70920.70880.9962
ACOFS0.94590.81400.87500.87520.99630.50680.86050.63790.65210.9972
RHSOFS0.97300.83720.90000.90080.99670.86490.74420.80000.79860.9949
4D 4WTFS0.81400.79550.80460.80070.99580.35190.86360.50000.53850.9971
DEFS0.89510.79550.84240.84080.99580.55910.79550.65670.65890.9958
GAFS0.89740.79550.84340.84190.99580.67310.79550.72920.72580.9958
PSOFS0.85780.79550.82550.82260.99580.59560.80000.68280.68300.9959
ACOFS0.83330.79550.81400.81040.99580.51430.81820.63160.64000.9962
RHSOFS0.92110.79550.85370.85320.99580.66040.79550.72160.71860.9958
5D 5WTFS0.87500.76360.81550.81310.99390.38520.85450.53110.55900.9961
DEFS0.92170.79480.85360.85250.99470.64020.82180.71970.71750.9954
GAFS0.93480.78180.85150.85150.99440.80700.83640.82140.81690.9958
PSOFS0.88890.80000.84210.83950.99480.70830.80360.75300.74780.9949
ACOFS0.89800.80000.84620.84390.99480.63160.87270.73280.73490.9967
RHSOFS0.95740.81820.88240.88240.99530.89800.80000.84620.84390.9948
6D 6WTFS0.81130.91490.86000.85830.99810.29730.93620.45130.51240.9985
DEFS0.90890.89150.90010.89800.99760.59390.90850.71830.72770.9980
GAFS0.95450.89360.92310.92200.99770.75440.91490.82690.82680.9981
PSOFS0.91220.88450.89810.89610.99750.65150.91490.76110.76640.9981
ACOFS0.85420.87230.86320.86020.99720.43140.93620.59060.62520.9986
RHSOFS0.97620.87230.92130.92120.99720.79250.89360.84000.83780.9977
7D 7WTFS0.82000.89130.85420.85170.99770.36080.94350.52190.57120.9987
DEFS0.82970.90000.86340.86110.99780.61220.91300.73300.74140.9981
GAFS0.91300.91300.91300.91120.99810.68250.93480.78900.79390.9986
PSOFS0.89670.91610.90630.90430.99820.59420.89130.71300.72090.9976
ACOFS0.84000.91300.87500.87300.99810.51190.93480.66150.68370.9986
RHSOFS0.93330.91300.92310.92150.99810.78850.89130.83670.83460.9977
8D 8WTFS0.84440.86360.85390.85090.99720.36020.90450.51520.55870.9980
DEFS0.88370.86360.87360.87110.99720.56690.85680.68240.68960.9970
GAFS0.92680.86360.89410.89260.99720.69090.86360.76770.76730.9972
PSOFS0.91870.85610.88630.88460.99700.64350.86140.73660.73860.9971
ACOFS0.90480.86360.88370.88160.99720.51350.86360.64410.65760.9972
RHSOFS0.95120.88640.91760.91660.99770.71700.86360.78350.78220.9972
9D 9WTFS0.88140.88140.88140.87810.99670.45600.96610.61960.65190.9990
DEFS0.95080.88470.91660.91500.99680.66920.90510.76950.77130.9973
GAFS0.94550.88140.91230.91050.99670.76810.89830.82810.82560.9972
PSOFS0.95470.89270.92260.92110.99700.68830.89830.77940.77970.9972
ACOFS0.92860.88140.90430.90210.99670.48850.93900.64270.66620.9983
RHSOFS0.96430.91530.93910.93780.99760.77610.88140.82540.82200.9967
10D 10WTFS0.80770.75000.77780.77270.99340.42820.90000.58030.60780.9973
DEFS0.86140.76610.81100.80770.99390.57140.81430.67160.67240.9951
GAFS0.90000.80360.84910.84670.99480.66670.85710.75000.74880.9962
PSOFS0.88160.77140.82290.82040.99400.71430.80360.75630.75090.9948
ACOFS0.86270.78570.82240.81890.99440.59760.87500.71010.71470.9967
RHSOFS0.92160.83930.87850.87650.99580.75860.78570.77190.76590.9944
Table 5. Rank of all FS models.
Table 5. Rank of all FS models.
Sl. NoDatasetFS AlgorithmAccuracyAvg. Rank
NBKNNDTSVM
1D 1WTFS97.02245 (6)99.45030 (6)99.40449 (5)96.19789 (6)5.75
DEFS97.97984 (5)99.50069 (4)99.47778 (3)98.56619 (4)4
GAFS99.03802 (1)99.63353 (2)99.45030 (4)99.08383 (2)2.25
PSOFS98.62574 (4)99.60605 (3)99.45030 (4)98.80898 (3)3.5
ACOFS98.80898 (3)99.49611 (5)99.49611 (2)97.80119 (5)3.75
RHSOFS98.85479 (2)99.67934 (1)99.63353 (1)99.26706 (1)1.25
2D 2WTFS97.70957 (6)99.35868 (6)99.03802 (6)96.56436 (6)6
DEFS98.57077 (5)99.48847 (4)99.48956 (4)98.61658 (4)4.25
GAFS99.40449 (1)99.58772 (2)99.54191 (2)99.17545 (2)1.75
PSOFS98.90060 (4)99.51901 (3)99.52665 (3)98.80898 (3)3.25
ACOFS99.30371 (2)99.38158 (5)99.31745 (5)98.16766 (5)4.25
RHSOFS98.99222 (3)99.63353 (1)99.58772 (1)99.22126 (1)1.5
3D 3WTFS97.75538 (6)99.35868 (6)99.45030 (6)96.19789 (6)6
DEFS98.45625 (5)99.49611 (4)99.53275 (4)98.46541 (4)4.25
GAFS99.31287 (1)99.54191 (2)99.58772 (2)98.90060 (2)1.75
PSOFS98.90060 (4)99.51901 (3)99.49611 (5)98.68530 (3)3.75
ACOFS99.17545 (2)99.42281 (5)99.54191 (3)98.07604 (5)3.75
RHSOFS98.94640 (3)99.63353 (1)99.63353 (1)99.26706 (1)1.5
4D 4WTFS97.11406 (5)99.22126 (5)99.22126 (6)96.51855 (6)6
DEFS98.10811 (4)99.47320 (2)99.39991 (3)98.32341 (4)3.25
GAFS99.31287 (1)99.49611 (1)99.40449 (2)98.80898 (1)1.25
PSOFS98.85479 (2)99.37700 (3)99.32203 (4)98.50206 (3)3
ACOFS98.80898 (3)99.29913 (4)99.26706 (5)98.07604 (5)4.25
RHSOFS98.85479 (2)99.49611 (1)99.45030 (1)98.76317 (2)1.5
5D 5WTFS97.48053 (6)99.49611 (6)99.12964 (5)96.19789 (6)5.75
DEFS97.83784 (5)99.50611 (5)99.31287 (2)98.38754 (5)4.25
GAFS99.08383 (1)99.54191 (2)99.31287 (2)99.08383 (2)1.75
PSOFS98.76317 (4)99.52611 (3)99.24416 (4)98.67155 (3)3.5
ACOFS98.97847 (2)99.51611 (4)99.26706 (3)98.39670 (4)3.25
RHSOFS98.80898 (3)99.58772 (1)99.45030 (1)99.26706 (1)1.5
6D 6WTFS97.11406 (6)99.45030 (6)99.35868 (5)95.09849 (6)5.75
DEFS97.59505 (5)99.58772 (4)99.57398 (2)98.46541 (4)3.75
GAFS99.45030 (1)99.67934 (1)99.67934 (1)99.17545 (2)1.25
PSOFS98.44251 (2)99.61521 (3)99.56809 (3)98.76317 (3)2.75
ACOFS99.08841 (3)99.51647 (5)99.40449 (4)97.20568 (5)4.25
RHSOFS98.71736 (4)99.63353 (2)99.67934 (1)99.26706 (1)2
7D 7WTFS97.80119 (5)99.45030 (6)99.35868 (6)96.35822 (6)5.75
DEFS98.27302 (4)99.58772 (3)99.39991 (5)98.59826 (3)3.75
GAFS99.08383 (2)99.63353 (2)99.63353 (2)98.94640 (2)2
PSOFS98.90060 (3)99.59689 (4)99.60081 (3)98.48832 (4)3.5
ACOFS99.30371 (1)99.47778 (5)99.45030 (4)97.98443 (5)3.75
RHSOFS98.90060 (3)99.72515 (1)99.67934 (1)99.26706 (1)1.5
8D 8WTFS97.15987 (6)99.35868 (6)99.40449 (6)96.56894 (6)6
DEFS97.73706 (5)99.49611 (4)99.49611 (5)98.39212 (4)4.5
GAFS99.22126 (1)99.58772 (2)99.58772 (2)98.94640 (2)1.75
PSOFS98.62574 (4)99.50985 (3)99.55718 (3)98.75859 (3)3.25
ACOFS99.13880 (3)99.40449 (5)99.54191 (4)98.07604 (5)4.25
RHSOFS99.17545 (2)99.63353 (1)99.67934 (1)99.03802 (1)1.25
9D 9WTFS97.89281 (6)99.45030 (5)99.35868 (6)96.79340 (5)5.5
DEFS98.13559 (5)99.55209 (3)99.56482 (3)98.53413 (3)3.5
GAFS99.03802 (1)99.67934 (1)99.54191 (4)98.99221 (1)1.75
PSOFS98.80898 (4)99.60147 (2)99.59536 (2)98.62574 (2)2.5
ACOFS99.23958 (2)99.50069 (4)99.49611 (5)97.17820 (4)3.75
RHSOFS98.90060 (3)99.67934 (1)99.67934 (1)98.99221 (1)1.5
10D 10WTFS97.11274 (6)99.26672 (5)98.90009 (6)96.65903 (6)5.75
DEFS97.38772 (5)99.35839 (2)99.08341 (5)97.95600 (5)4.25
GAFS98.96425 (1)99.40422 (1)99.26672 (2)98.53346 (3)1.75
PSOFS98.56095 (3)99.28506 (4)99.14757 (3)98.67094 (2)3
ACOFS98.72594 (2)99.31256 (3)99.12924 (4)98.16682 (4)3.25
RHSOFS98.35014 (4)99.40422 (1)99.40422 (1)98.80843 (1)1.75
Table 6. Average rank of all models.
Table 6. Average rank of all models.
DatasetWTFSDEFSGAFSPSOFSACOFSRHSOFS
D 15.7542.253.53.751.25
D 264.251.753.254.251.5
D 364.251.753.753.751.5
D 463.251.2534.251.5
D 55.754.251.753.53.251.5
D 65.753.751.252.754.252
D 75.753.7523.53.751.5
D 864.51.753.254.251.25
D 95.53.51.752.53.751.5
D 105.754.251.7533.251.75
Average5.825 (6)3.975 (5)1.725 (2)3.2 (3)3.85 (4)1.525 (1)
Table 7. Holm Test.
Table 7. Holm Test.
Sl. NoFS Algorithmsz-Valuesp-Valuesα/(P − i)
1RHSOFS: WTFS5.1390.000010.01
2RHSOFS: DEFS2.9280.001710.0125
3RHSOFS: GAFS0.2390.405530.016667
4RHSOFS: PSOFS2.0020.022640.025
5RHSOFS: ACOFS2.7790.002730.05
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Padhi, B.K.; Chakravarty, S.; Naik, B.; Pattanayak, R.M.; Das, H. RHSOFS: Feature Selection Using the Rock Hyrax Swarm Optimization Algorithm for Credit Card Fraud Detection System. Sensors 2022, 22, 9321. https://doi.org/10.3390/s22239321

AMA Style

Padhi BK, Chakravarty S, Naik B, Pattanayak RM, Das H. RHSOFS: Feature Selection Using the Rock Hyrax Swarm Optimization Algorithm for Credit Card Fraud Detection System. Sensors. 2022; 22(23):9321. https://doi.org/10.3390/s22239321

Chicago/Turabian Style

Padhi, Bharat Kumar, Sujata Chakravarty, Bighnaraj Naik, Radha Mohan Pattanayak, and Himansu Das. 2022. "RHSOFS: Feature Selection Using the Rock Hyrax Swarm Optimization Algorithm for Credit Card Fraud Detection System" Sensors 22, no. 23: 9321. https://doi.org/10.3390/s22239321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop