Next Article in Journal
Comparative Electromyographic Analysis in Leg Press of Traditional Fitness Equipment, Traditional Outdoor Fitness Equipment, and a New Model of Outdoor Fitness Equipment in Trained Young Men
Previous Article in Journal
The Noise Exposure of Urban Rail Transit Drivers: Hazard Classification, Assessment, and Mitigation Strategies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Credit Card Fraud: Analysis of Feature Extraction Techniques for Ensemble Hidden Markov Model Prediction Approach

1
Department of Computer Science, Tai Solarin University of Education, Ijagun, Ijebu-Ode 2118, Nigeria
2
French-South African Institute of Technology, Department of Electrical, Electronic, and Computer Engineering, Cape Peninsula University of Technology, Bellville Campus, Bellville 7535, South Africa
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(16), 7389; https://doi.org/10.3390/app14167389 (registering DOI)
Submission received: 22 July 2024 / Revised: 17 August 2024 / Accepted: 20 August 2024 / Published: 21 August 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
In the face of escalating credit card fraud due to the surge in e-commerce activities, effectively distinguishing between legitimate and fraudulent transactions has become increasingly challenging. To address this, various machine learning (ML) techniques have been employed to safeguard cardholders and financial institutions. This article explores the use of the Ensemble Hidden Markov Model (EHMM) combined with two distinct feature extraction methods: principal component analysis (PCA) and a proposed statistical feature set termed MRE, comprising Mean, Relative Amplitude, and Entropy. Both the PCA-EHMM and MRE-EHMM approaches were evaluated using a dataset of European cardholders and demonstrated comparable performance in terms of recall (sensitivity), specificity, precision, and F1-score. Notably, the MRE-EHMM method exhibited significantly reduced computational complexity, making it more suitable for real-time credit card fraud detection. Results also demonstrated that the PCA and MRE approaches perform significantly better when integrated with the EHMM in contrast to the conventional HMM approach. In addition, the proposed MRE-EHMM and PCA-EHMM techniques outperform other classic ML models, including random forest (RF), linear regression (LR), decision trees (DT) and K-nearest neighbour (KNN).

1. Introduction

The progress in online banking systems has motivated the use of credit cards to ease the payment of products and services. Credit cards are payment cards issued to cardholders to purchase products and services based on their accumulated debt. In most cases, before a credit card payment is approved, some vital information is requested from the cardholders such as the personal identification number (PIN), card verification value (CVV) number, expiration date, and so on. This information helps to validate the authenticity of the cardholders, to prevent fraud; yet cases of credit card fraud are recorded daily. Nonetheless, credit cards are becoming favoured and widely employed for online banking payments, which implies that the number of credit card transactions has increased exponentially despite the theft risk it involves.
This growth in the acceptability of credit cards has made it difficult to differentiate between a fraudulent and a non-fraudulent credit card transaction. Accordingly, credit card theft is on the rise irrespective of the security information (such as PIN and CVV) required before a transaction is approved. Although security checks such as tokenization and data encryption are used to prevent credit card theft [1], they cannot completely prevent fraudulent credit card transactions. In particular, credit card fraud occurs remotely, whereby the simple card information is all that is required. The time and place for the fraudulent transactions do not call for a PIN, card imprint, or handwritten signature. Most of the time, the victims of fraud are oblivious that the perpetrators have access to their credit card information, especially when these credit cards are used for payment on phishing websites [2,3]. The quickest way to spot these credit card frauds is to examine each card’s spending patterns and look for any differences from regular spending patterns. However, it is complex to achieve because the daily number of credit card transactions is huge, resulting in large transactional information. As a result, there has been a lot of recent study into the exact, rapid, and efficient prediction of credit card fraud.
In recent times, machine learning (ML) tools have been deployed in the literature to efficiently predict credit card frauds. ML tools enable computers to improve their forecasting abilities by learning from previous datasets. Different ML tools, such as hidden Markov models (HMMs), decision trees, K-nearest neighbour, logistic regression, and so on, have been deployed for the prediction of credit card frauds [4,5,6,7,8]. However, work is being carried out to improve the predictive ability of these ML tools.
HMM is a popular and flexible ML tool that can easily model randomly changing datasets. HMMs can easily predict fraudulent transactions from a sequence of defined observations. Nonetheless, the ability of HMM to effectively predict these credit card frauds depend on the adopted feature extraction technique. This implies that the more reliable the output of the feature extraction technique, the better the prediction performance of HMM [9,10,11]. Likewise, the length of the outputted feature vector, which doubles as the dimension of HMM, determines the computational time complexity imposed on HMM. The longer the length of the feature vector, the more the computational time complexity imposed on HMM [12]. Therefore, it is paramount to carefully select the feature attraction technique that will be combined with HMM to balance the trade-off between its performance gain and the computational time complexity.
Accordingly, this article analyses two feature extraction techniques that can be combined with HMM for the prediction of credit card fraud. First, the principal component analysis (PCA) [13,14] technique is combined with HMM for the prediction of credit card fraud. In addition to employing PCA, this article computationally determines the “optimal” feature vector length that will balance the trade-off between the PCA performance gain and the computational time complexity it imposed on HMM. It was discovered that this ‘optimal’ feature vector length is not computational time-efficient when PCA is combined with HMM. Therefore, the features derived using PCA are converted to statistical features to reduce the computational time complexity imposed on HMM. The proposed robust but simple statistical features denoted as MRE (Mean, Relative Amplitude, and Entropy) are merged to form a feature vector that can be combined with HMM to effectively predict credit card frauds.
Furthermore, as highlighted in [12], the Gaussian emission distribution parameters of HMM are sensitive to a flat start or random values, which impedes prediction performance. Hence, the K-mean clustering (K-MC) technique [15,16] and the Gaussian mixture model (GMM) [17,18] are sequentially used to initialise the HMM process. Since the K-MC and GMM techniques are embedded in the Gaussian emission process of HMM, the latter is referred to as an ensemble hidden Markov model (EHMM) in this article. Therefore, this article evaluates the performance of PCA and MRE when combined with EHMM (PCA-EHMM and MRE-EHMM). Likewise, the performance of PCA and MRE is verified when combined with the conventional HMM (PCA-HMM and MRE-HMM), demonstrating the necessity of the initialisation. Moreover, the performance of the proposed MRE-EHMM and PCA-EHMM techniques is validated with other classic ML models such as random forest (RF), linear regression (LR), decision trees (DT), and K-nearest neighbour (KNN). The performances are documented for different performance metrics such as recall/sensitivity ( S e ), specificity ( S p ), precision ( P ), and F1-score ( F 1 ) using the credit card transactions dataset of European cardholders gathered within two days in September 2013 (https://www.kaggle.com/mlg-ulb/creditcardfraud, accessed on 21 July 2024).
The key contributions and significance of this article are outlined as follows. While various studies have proposed the use of HMM for predicting credit card fraud, these works often lack clarity regarding the feature extraction techniques employed and their integration with the HMM. Feature extraction is crucial for enhancing the model’s predictive accuracy [11]. In this article, two feature extraction techniques are proposed, which are tailored for use with the EHMM to effectively predict fraudulent credit card transactions. The EHMM approach presented here distinguishes itself from the conventional HMM models used in credit card fraud detection in two significant ways: (1) HMM is systematically integrated with K-MC and GMM techniques to form an EHMM, enhancing the model’s predictive capability, and (2) two feature extraction techniques (PCA and MRE) are introduced and can be seamlessly integrated with the EHMM to further improve its performance. Moreover, the results obtained from applying these techniques demonstrate high reliability and are reproducible for real-time credit card transaction monitoring. This article is expected to contribute to reducing financial losses for cardholders and institutions while bolstering confidence in online banking involving credit card transactions.
The remaining part of this article is organised as follows. Section 2 briefly reviews some of the recent works on the prediction of credit card frauds using HMMs. In Section 3, the dataset used for result verification is discussed. The PCA and MRE feature extraction techniques are explained in Section 4. Section 5 discusses the EHMM in detail while also explaining its training process. In Section 6, the metrics used in documenting the performances of the proposed feature extraction techniques are briefly explained. Section 7 discusses the obtained results from the PCA-EHMM and MRE-EHMM prediction techniques. In addition, this section highlights the performance difference between the EHMM and conventional HMM approaches, and also compares the PCA-EHMM and MRE-EHMM techniques to other existing ML techniques in the literature. The article is concluded with some final remarks in Section 8.

2. Related Work

The widespread use of credit cards for online purchases has also increased the possibility of fraud. According to the Nigerian Deposit Insurance Corporation (NDIC) annual report of 2018, between 2016 and 2018, the number of credit card fraud incidents in Nigeria increased by 33%, while the actual amount lost to this credit card fraud climbed by 84% [19]. Likewise, the federal trade commission (FTC) affirmed that there were around 1579 data breaches, totalling 179 million data points, with credit card fraud being the most widespread [6,20]. Therefore, a wide range of ML and data mining techniques have been deployed in the literature to proffer solutions to this menace. This section reviews some recent and related credit card prediction techniques.
A weighted average ensemble algorithm for credit card fraud detection was proposed in [21]. This algorithm integrated multiple models, including logistic regression (LR), K-nearest neighbour (KNN), random forest (RF), AdaBoost, and Bagging, to enhance detection accuracy. The study utilized a dataset from a European credit card company and achieved an accuracy of 99%, surpassing the baseline models (LR, AdaBoost, RF, KNN, and Bagging) in performance. The authors demonstrated the potential of utilizing an ensemble model for real-time credit card fraud detection. However, the lack of details regarding the feature extraction process limits the reproducibility of their work.
In [22], a comparative study on four ML techniques, namely, DT, RF, LR, and Naive Bayes (NB), for the detection of credit card frauds was carried out, with attention on imbalanced datasets. Their analysis revealed that RF was effective at detecting credit card fraud, while LR and NB performed comparably, and DT had the worst performance. However, similar to [21], the study did not provide insight into the feature extraction process, raising concerns about the reproducibility of their results.
The authors in [23] conducted an extensive review of various machine learning techniques used for detecting credit card fraud. The review encompassed methods such as NB, support vector machines (SVMs), multilayer feed forward (MLFF) neural networks, Bayesian belief networks (BBNs), genetic algorithms (GAs), and classification and regression trees (CARTs). As this article was a critical analysis of prior studies, no specific dataset was employed. The findings highlighted the superior performance of NB, which achieved the highest accuracy, followed by SVM and GA.
In [24], the sequence of operation of credit card transactions was modelled using HMMs.Also, the paper provided information on how HMMs can be used to detect fraudulent credit card transactions. Yet, the paper did not document any clear results to buttress the performance of their developed HMM. More importantly, the paper did not provide information on the feature extraction technique used or combined with their developed HMM. As mentioned earlier, the feature extraction technique used with any ML technique including HMMs determines the performance of the ML technique, hence, the focus of this article.
The authors in [25] modelled a fraud detection system that would attempt to detect credit card fraud as accurately as possible by producing clusters and analysing the clusters formed by the dataset for anomalies. Therefore, their work examined the detection accuracy of two hybrid techniques: the K-MC technique with multilayer perceptron (MLP) and the K-MC technique with HMM. The authors show that the detection accuracy of the two models examined was fairly the same. Nonetheless, the paper was silent on the adopted feature extraction technique which is a major factor in analysing the performances of their proposed models.
The process of fraud detection using HMMs was described in [26]. In their work, the researchers combined the HMM process with the K-MC technique to form what this article described as an EHMM. The K-MC technique was used to initialise the HMM process to improve its performance. Their model’s performance was presented in terms of recall and precision. Although their model training flowchart displayed the importance of the feature extraction step, there was no discussion on their adopted feature extraction technique. This article combines the K-MC technique and GMM sequentially with the HMM to form an EHMM and focuses on the adopted feature extraction technique, which determines the performance of the EHMM and any other ML tool.
A credit card fraud detection model was developed using a multiple-perspective HMM-based approach in [4]. The study developed eight HMMs to model sequences of credit card transactions. They employed history-based features with HMM based on three perspectives that were aided with different assumptions. Their result was documented in terms of recall and precision. On the other hand, this article derives the feature vector using PCA and a simple statistical method termed MRE. More so, an EHMM is used in this article, which demonstrates significant performance improvement in comparison to the conventional HMM method.

3. Dataset

A secondary dataset for credit card fraud was obtained for this article from an internet repository made available by Kaggle.com (https://www.kaggle.com/mlg-ulb/creditcardfraud, accessed on 21 July 2024).This dataset dating from September 2013 contains different credit card transactions made by cardholders in Europe.This dataset includes 492 fraud cases out of 284,807 transactions within two days, indicating that 0.172% of card transactions were fraudulent. Each transaction contains 31 numerical features (31 columns), 28 ( V 1 V 28 ) of which are transformed using PCA to support seclusion [6,20]. The three remaining columns are time, amount, and the class of transaction. The time column depicts the time difference between each transaction in the dataset from the first one to the last, whereas the class of transaction column, which is key to this study, indicates if the transaction is fraudulent ‘1’ or non-fraudulent ‘0’. Table 1 provides a summary of some statistical properties ( S P ) of the dataset.

4. Feature Extraction Techniques

4.1. Principal Component Analysis (PCA)

PCA is a technique used for reducing the dimension of a dataset, by converting the variables in the dataset into a smaller set while preserving the information in the original dataset [12,13]. Consequently, this reduced dataset, called features, can be seamlessly visualised and evaluated using different ML tools. PCA uses basic equations to derive variables, called principal components, from the original dataset. PCA is a very robust feature extraction technique, especially when the sampling points in the dataset to be analysed carry both negative and positive values [12,13]. The operation of PCA can be sequentially broken down into three steps: (I) standardisation, (II) computation of covariance matrix, and (III) eigenvector and eigenvalue computation [12].

4.1.1. Standardisation

The standardisation step is a scaling process that allows all the variables or sampling points in a continuous signal S i to evenly participate in the feature analysis. This step limits any bias in the feature analysis process, in case there are large differences between the dataset variables. Given a continuous signal, S i = ( s 1 , s 2 , , s i ) , the standardisation step is achieved as:
S ¯ i = s 1 M ( S i ) D ( S i ) , s 2 M ( S i ) D ( S i ) , , s i M ( S i ) D ( S i ) = ( s ¯ 1 , s ¯ 2 , , s ¯ i ) ,
where S ¯ i is the standardised continuous signal, M is the mean of S i , and D is the standard deviation of S i .

4.1.2. Covariance Matrix Computation, ∑

The covariance matrix ∑ is computed to identify the variation between each variable in S ¯ i and the mean of S ¯ i . This step is used to eliminate redundant points or information since the variables are presumed to be highly correlated. ∑ is a symmetric d × d matrix, where its entries are covariances linked to all the possible pairs of the initial variables. For example, given a d-dimensional matrix of the form:
G = G 1 G 2 G d ,
where G 1 , G 2 , and G d are defined as:
G 1 = g 1 1 g 1 2 g 1 a , G 2 = g 2 1 g 2 2 g 2 a , G d = g d 1 g d 2 g d a .
Then, the covariance of matrix G can be computed as:
G ¯ = c o v ( G ) = G 1 2 a G 1 G 2 a G 1 G d a G 2 G 1 a G 2 2 a G 2 G d a G d G 1 a G d G 2 a G d 2 a ,
where the variance of G 1 , G 2 , and G d is the diagonal of Equation (3). Also, it should be noted from Equation (3) that G 1 G 2 a = G 2 G 1 a is the covariance of two vectors, which is defined as:
c o v ( G 1 , G 2 ) = c o v ( G 2 , G 1 ) = ( G 1 M ( G 1 ) ) ( G 2 M ( G 2 ) ) a 1 .

4.1.3. Eigenvector and Eigenvalue Computation

The covariance matrix ∑ is used to compute the eigenvectors U and eigenvalues V to derive the principal components of the continuous signal S i . The principal components are computed by finding the solution to Equation (5).
U = V U .
where V are scalar values and U are non-zero vectors called principal components. The direction of the PCA space is depicted by U while the corresponding V shows the length and scaling factor of U [12,27]. The principal components are arranged according to the variance of V from the highest to the lowest. Thus, the U with the highest V is the first principal component because it has the highest variance and carries the maximum possible information. This arrangement enables a reduction in the dimension of the principal components with no or minimal information lost. It should be noted that these principal components can be reconstructed linearly as a combination of the original signal. Also, the number of derived principal components is a function of the dimension of the analysed signal. That is, the d-dimension matrix of Equation (2) will produce d principal components after complete decomposition.
To solve for U and V , Equation (5) is decomposed using the singular vector decomposition (SVD) technique into three matrices given as:
G ¯ = X Y Z ,
where Y is the a × a diagonal matrix, X is the a × d matrix called the left singular vector, and Z is the d × d matrix referred to as the right singular vector. The elements on the diagonal of Y arranged from highest to lowest are the V while the principal components ( U ) are the columns of the right singular vector, Z. The principal components after complete SVD are usually of the form:
Z = z 1 , 1 z 1 , 2 z 1 , d z 2 , 1 z 2 , 2 z 2 , d z d , 1 z d , 2 z d , d .
Therefore, to derive a suitable feature vector f of dimensions 1 × d that can be combined with HMM, Equation (7) is transformed as:
f d = 1 d 1 d | z r | 1 , 1 d 1 d | z r | 2 , , 1 d 1 d | z r | d = [ f 1 , f 2 , , f d ] .
Hence, given α credit card transactions used for training HMM, the feature vector can be represented in the form of a matrix as:
f α , d = f 1 , 1 f 1 , 2 f 1 , d f 2 , 1 f 2 , 2 f 2 , d f α , 1 f α , 2 f α , d .
It should be noted that because of privacy concerns, each transaction in the dataset has been converted to principal components with dimensions 1 × d , where d = 28 . Nonetheless, if given the original dataset, the principal components which serve as the feature vector can be derived using these three sequential steps. This proposed PCA feature extraction technique is summarised in Algorithm 1.
Algorithm 1: PCA Feature Extraction
Input S i
Output: f α , d
1: Compute S ¯ i as in (1)
2: Compute G ¯ as in (3)
3: Compute V and U as in (6)
4: Derive f d and f α , d as in (8) and (9)

4.2. Statistical Features: MRE

This statistical feature extraction method, termed MRE (Mean, Relative Amplitude, and Entropy), is based on the observation of the output of PCA; that is, the principal components. First, it is observed that the principal components are of dimensions 1 × 28 , which automatically increases the complexity of the EHMM. However, as mentioned earlier, the principal components are arranged with respect to the variance of the eigenvector V from the highest to the lowest. This implies that the dimension of the principal components can be reduced to 1 × k (k is the ‘optimal’ length of the principal components) without necessarily losing information, as shown in Section 7.1. Yet, the value of k is still large enough to increase the computational burden of the EHMM. Therefore, the proposed MRE is used to statistically reduce the dimension of k so as to limit the computational time complexity of the EHMM. Accordingly, given that the principal component is of 1 × k dimensions ( f = [ f 1 , f 2 , , f k ] ) , the MRE feature vector can be constructed by merging the following statistical parameters in no specified order.

4.2.1. Mean, M

The mean M is an important statistical tool used to average a collection of numbers. The mean is selected as a component of the MRE because it is noticed that the 1 × k principal components fall within a defined range. Therefore, given a single card transaction with 1 × k principal components, M is computed as:
M = 1 k 1 k f k .
Consequently, for α credit card transactions, the mean can be represented as:
M = M 1 M 2 M α = 1 k 1 k f k 1 1 k 1 k f k 2 1 k 1 k f k α

4.2.2. Relative Amplitude

The relative amplitude R is also a statistical tool used in this article to reduce the dimension of the principal components. It is selected because it determines the bands of the principal components. Thus, given a single card transaction with 1 × k principal components, R is computed as:
R = [ f m a x , f m i n , f d i f f ] ,
where f m a x is the maximum relative amplitude, f m i n is the minimum relative amplitude, and f d i f f is the difference between the maximum and minimum relative amplitude, computed as:
f m a x = m a x ( f ) , f m i n = m i n ( f ) and f d i f f = | f m a x f m i n | .
For α credit card transactions, the relative amplitude can be represented as:
R = f 1 m a x f 1 m i n f 1 d i f f f 2 m a x f 2 m i n f 2 d i f f f α m a x f α m i n f α d i f f .

4.2.3. Entropy, E

The theory of entropy plays a key part in the description of many intricate processes such as communications, statistics, thermodynamics, and so on. Entropy is scientifically associated with the terms disorder, uncertainty, or randomness [28]. Different types of entropy have been used for feature extraction such as approximate entropy (ApEn), sample entropy (SampEn), permutation entropy (PeEn), and so on [29,30]. However, the Shannon entropy E is used in this article because it can easily measure the mean or average amount of information conveyed in each data point. Therefore, given a single card transaction with 1 × k principal components, E is estimated as:
E = 1 k P k l o g 2 1 P k ,
where P k is the principal component at point k. Accordingly, for α credit card transactions, E will be formulated as:
E = E 1 E 2 E α = 1 k P k l o g 2 1 P k 1 1 k P k l o g 2 1 P k 2 1 k P k l o g 2 1 P k α
Thus, the mean, relative amplitude, and entropy are combined in no particular order to form a feature vector that can be adapted with HMM as shown:
f M R E = [ M , R , E ] = M 1 R 1 E 1 M 2 R 2 E 2 M α R α E α
It should be noted that the f M R E is h-dimensional (where h = 5), which is a significant reduction from the d-dimension ( d = 28 ) or k-dimension (k is determined in Section 7.1). However, with HMM, the computational complexity increases with an increase in the dimension of the feature vector. This implies that MRE-HMM simplifies the HMM process in comparison to PCA-HMM. This proposed MRE feature extraction technique is summarised in Algorithm 2.
Algorithm 2: MRE Feature Extraction
Input: f
Output: f M R E
1: Compute M as in (10) and (11)
2: Compute R as in (12)–(14)
3: Compute E as in (15) and (16)
4: Derive f M R E as in (17)

5. Ensemble Hidden Markov Model (EHMM)

HMM is a probabilistic ranking classifier that assigns a tag to each observational segment in a sequence. As a result, it determines the probability distribution over the set of observations and produces the sequence of observations that is most probable [10,12]. Because of this, it is able to model and categorise the set of observations derived from this credit card transaction with ease. There are various forms of HMMs; however, this article employs the ergodic HMM, which permits transitions between states without any transition probabilities being zero [12]. In general, an HMM consist of the following components.
  • The number of the transition states, s.
  • The sequence of observation, Q = { q 1 , q 2 , , q z } , where z is the number of observation which determines the output of the modelled system.
  • The transition probability matrix, ρ : The transition probability between two states, s 1 and s 2 , is denoted as ρ = ρ s 1 , s 2 . Thus, ρ is the probability of moving from state s 1 to s 2 . It is important to note that the likelihood of transitioning from state s 1 to s 2 is only determined by the present state s 1 , not by any previous state. Mathematically, ρ is of the form of an s × s matrix given as:
    ρ = ρ 1 , 1 ρ 1 , 2 ρ 1 , s ρ 2 , 1 ρ 2 , 2 ρ 2 , s ρ s , 1 ρ s , 2 ρ s , s .
  • Gaussian emission distribution parameters, ϕ : The probability of emitting an observation from a particular state s is determined by ϕ . Mathematically, ϕ is of the form of an s × z matrix given as
    ϕ = ϕ 1 , 1 ϕ 1 , 2 ϕ 1 , z ϕ 2 , 1 ϕ 2 , 2 ϕ 2 , z ϕ s , 1 ϕ s , 2 ϕ s , z .
    ϕ consist of three important parameters; namely, the mean M, covariance matrix ∑, and mixture weight β , represented as ϕ = { M , , β } .
  • The start probability, ν = { ν 1 , ν 2 , , ν z } : ν indicates the beginning of the Gaussian distribution process. ν always adds up to 1 at every point of the distribution process. That is,
    1 z ν z = 1
As a result, a given HMM can be represented using these three parameters as:
H = { ν , ρ , ψ }
In applying HMMs, three primary problems are typically addressed to find the optimal sequence of observations:
  • Computing the observation probability: The HMM uses an observation sequence, Q = { q 1 , q 2 , , q z } , and a known model, H = { ν , ρ , ψ } , to estimate the likelihood of a specific sequence of observations.
  • Computing the model parameters (Training): During this phase, HMM selects the model that best fits the input data. The HMM parameters, H = { ν , ρ , ψ } , are continually re-estimated to maximize the likelihood of each observation sequence. Maximum likelihood estimation (MLE) techniques like the Baum–Welch algorithm [31] can be used for this process. The Baum–Welch algorithm iteratively optimizes the HMM parameters, H = { ν , ρ , ψ } , based on the Gaussian emission distribution parameters, ϕ = { M , , β } until the condition is satisfied. The Gaussian emission distribution parameters, ϕ = { M , , β } , are used in the MLE process. These Gaussian parameters, ϕ , assume random values or flat values at the start of the MLE process. However, HMMs are sensitive to flat start or random values of the Gaussian parameters M, ∑, and β because it limits the prediction performance of the model in general. Accordingly, the K-MC [15,16] and GMM [17,18] techniques are sequentially used to initialise the Gaussian emission parameters. As a result, this article refers to HMM as an ensemble hidden Markov model (EHMM) because the K-MC and GMM techniques are embedded in the Gaussian emission process, as shown in Figure 1.
  • Decoding: HMM identifies the hidden states or path based on the observation, Q. The Viterbi algorithm ( V a l g ) [32] is used to complete this phase. The V a l g is an error-correcting algorithm that searches for a new path using the knowledge from a previous path. It subsequently outputs the path with optimal probability by calculating all probable hidden paths.
This conventional HMM approach, utilizing MLE and the V a l g , has been applied in prior studies to predict credit card fraud. However, the reliability of the extracted feature vector is crucial to the performance of EHMM. In fact, the more reliable the feature vector, the better the performance of EHMM. Therefore, this article places significant emphasis on the feature extraction technique.

EHMM Training

The credit card dataset, which includes 284,807 transactions, is divided into two portions. The model is tested on a small subset of the dataset, while the majority is used for training ( χ = 70–80%). The training portion is divided into two groups: fraudulent transactions and non-fraudulent transactions. As indicated in the dataset, fraudulent transactions are represented as ’1’ while non-fraudulent transactions are designated with ’0’. As a result, two HMMs are created to represent fraudulent transactions ( ν 1 , ρ 1 , ϕ 1 ) and non-fraudulent transactions ( ν 2 , ρ 2 , ϕ 2 ). In each scenario, a four-state ergodic HMM with two mixture weights is used. The dimension of EHMM is dictated by the resulting feature extraction vector dimension. MRE-EHMM assumes five dimensions, whereas PCA-EHMM assumes k dimensions.
During testing, the two HMMs are merged and processed through the V a l g . This combined HMM (defined by ν = ν 1 ,   ν 2 , ρ = ρ 1 ,   ρ 2 , and ϕ = ϕ 1 ,   ϕ 2 ) forms an eight-state model with four mixture weights. States 1–4 represent the fraudulent transactions, while states 5–8 correspond to the non-fraudulent transactions. The ϕ = ϕ 1 ,   ϕ 2 parameter is used to fine-tune the resulting feature vector ( f M R E or f α , d ) for the unknown card transaction to be predicted. The fine-tuned feature vector ( f M R E ζ or f α , d ζ ) is subsequently fed into the V a l g . The V a l g uses the ν = ν 1 ,   ν 2 and ρ = ρ 1 ,   ρ 2 to predict the sequence of states form the fine-tuned feature vector, thereby predicting whether the unknown card transaction is fraudulent or not. Also, the V a l g transits between the two states (states 1–4 and 5–8) with equal probability as defined in the transmission matrix.

6. Performance Parameters

The performance of the two feature vectors is verified using the following metrics, which are explained in the context of this study.
  • Recall/Sensitivity, S e : S e measures the ability of the models to correctly predict the non-fraudulent transactions; that is, class ‘0’. It is expressed as:
    S e = T P T P + F N ,
    where true positives (TP) refer to the number of accurately predicted non-fraudulent transactions and false negatives (FN) refer to the number of times the models miss the manually identified non-fraudulent transactions.
  • Specificity, ( S p ): S p measures the ability of the models to correctly predict the fraudulent transactions; that is, class ‘1’. It is expressed as:
    S p = T N T N + F P ,
    where true negatives (TN) refer to the number of accurately predicted fraudulent transactions and false positives (FP) refer to the number of times the models miss the manually identified fraudulent transactions.
  • Precision, P : P is defined as the capacity of the model to accurately predict the class of the card transaction; that is, ‘0’ of ‘1’. It is expressed as:
    P = T P T P + F P ,
    where true positives (TP) refer to the number of accurately predicted transactions and false negatives (FP) refer to the number of times the models miss a manually identified transaction class. A high value of P indicates a good model performance.
  • F1-score, F 1 : F 1 is a measure that combines precision and recall. It is commonly known as the harmonic mean of the two. It is expressed as:
    F 1 = 2 P S e P + S e .

7. Results and Discussion

7.1. PCA-EHMM

Table 2 shows the PCA-EHMM prediction performance for specific values (k = 7–12). As previously stated, the principal components after complete decomposition are organized from left to right in order of importance. As a result, reducing the number of principal components from the right has no effect on their strength. In this vein, the performance of PCA-EHMM was verified for k = 7–12 (rather than d = 28 ), which reduced the computational cost imposed by PCA on EHMM. This is because the higher the dimension, the more complex the EHMM process, where k represents the dimension of EHMM.
From Table 2, one can notice that the performance of PCA-EHMM improves as the value of k increases from 7 to 9. For example, at χ = 80 , there is a performance improvement of 0.20%, 0.21%, 0.20%, and 0.21% in S e , S p , P , and F 1 , respectively, as k increases from 7 to 8. Likewise, at χ = 80 , there is a performance improvement of 0.21%, 0.23%, 0.22%, and 0.22% in S e , S p , P , and F 1 , respectively, as k increases from 8 to 9. Therefore, increasing the value of k does not result in a significant improvement in the prediction performance of PCA-EHMM; instead, it increases the computing cost of the EHMM process. Thus, k = 9 is chosen as the ideal number for k because it strikes a suitable compromise between performance gain and computational time complexity. Notwithstanding, this value of k is large enough to impose a high computational burden on EHMM. As such, MRE-EHMM is introduced in this article, as discussed earlier.
Figure 2, Figure 3, Figure 4 and Figure 5 depict the PCA-EHMM performance as k increases in size. One can notice how the curves of each figure flatten as k increases. This clearly shows that there is no significant increase in the performance of PCA-EHMM, even if all derived principle components are used (that is, d = 28 ). Also, it is important to mention that the performance of PCA-EHMM improves as the training size χ increases from 70% to 80%. This performance improvement is evident in the four performance metrics considered.

7.2. MRE-EHMM

Table 3 shows the performance of the introduced MRE-EHMM in comparison to PCA-EHMM at k = 9 . From the table, MRE-EHMM and PCA-EHMM exhibited approximately the same S e , S p , P , and F 1 performances, even as χ increases. However, one should keep in mind that MRE-EHMM is five-dimensional while PCA-EHMM is nine-dimensional. However, the computational time complexity of EHMM is proportional to the dimension size and increases as the dimension grows. Thus, when compared to the PCA approach, the MRE approach requires significantly less processing time from EHMM. As a result, the MRE approach can be used as a more efficient performance alternative to the PCA approach for real-time credit fraud prediction systems.

7.3. PCA and MRE Performance Comparison with HMM and Proposed EHMM

In this section, the performance of PCA and MRE is verified for the proposed EHMM and conventional HMM (that is, when the K-MC and GMM components are excluded from the HMM process). The prediction models’ performance is shown in Table 4. The table shows that PCA-EHMM and MRE-EHMM outperform PCA-HMM and MRE-HMM, respectively. For example, at χ = 80 % , PCA-EHMM outperforms PCA-HMM by 28.01%, 28.16%, 28.22%, and 28.12% in terms of S e , S p , P , and F 1 , respectively. Likewise, at χ = 80 % , MRE-EHMM outperforms MRE-HMM by 28.02%, 28.14%, 28.22%, and 28.13% in terms of S e , S p , P , and F 1 , respectively. This indicates that sequentially initialising the HMM process with the K-MC and GMM techniques is vital for improving its performance and the prediction system as a whole.

PCA-EHMM and MRE-EHMM Comparison with other Models

This section compares the performance of PCA-EHMM and MRE-EHMM with other relevant ML models in the literature. As highlighted in Section 2, RF, LR, DT, and KNN have been extensively used for detecting credit card frauds. RF combines different decision trees to limit overfitting [33,34], LR predicts the class probabilities based on some certain dependant variable [22], DT utilises a tree-like model for feature classification [35], and KNN classifies transactions based on the K-nearest data point [36]. Table 5 shows the performance of PCA-EHMM and MRE-EHMM in comparison to the RF, LR, DT, and KNN models.
From Table 5, the proposed MRE-EHMM and PCA-EHMM techniques achieved better S e , S p , P , and F 1 performance in comparison to the RF, LR, DT, and KNN models. Furthermore, this improvement in performance is noticeable regardless of changes in χ . Thus, the proposed MRE-EHMM and PCA-EHMM techniques are more suited to detecting online banking fraudulent transactions, followed by RF, LR, and DT, with KNN having the worst performance.

8. Conclusions

This article analyses the performance of two feature extraction techniques (PCA and MRE) that can be combined with EHMM for the prediction of credit card fraud. The PCA method offered good S e , S p , P , and F 1 performance. However, it exhibited a high computation burden on the entire prediction system. The MRE was introduced in the article as an efficient performance alternative to the PCA technique. Although the MRE technique offered the same S e , S p , P , and F 1 performances as the PCA technique, it exhibited significantly less processing time, which makes it more suitable in real time. Nonetheless, both approaches offer solutions to the loss experienced by cardholders and financial institutions as a result of online banking which requires credit card details. Moreover, the results demonstrated that the two feature extraction techniques perform significantly better when integrated with EHMM compared to the conventional HMM approach (that is, when the K-MC and GMM components are excluded). In addition, the article demonstrated that the proposed MRE-EHMM and PCA-EHMM techniques are suitable for credit card fraud detection, outperforming the RF, LR, DT, and KNN models. Future research could further compare these models with other ML and deep learning techniques to highlight their relative advantages and potential limitations.

Author Contributions

O.O. (Olayinka Ogundile) and O.B. conceptualized the study. O.O. (Olayinka Ogundile), A.O. and O.O. (Olabisi Ogundile) completed the theoretical derivation. O.O. (Olayinka Ogundile) and O.B. established the model, carried out the simulation, and finished the writing of this manuscript. V.B. proofread and supervised the writing of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study analyzed publicly available datasets. The data are available at https://www.kaggle.com/mlg-ulb/creditcardfraud (accessed on 1 June 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Iwasokun, G.B.; Omomule, T.G.; Akinyede, R.O. Encryption Tokenization-Based Syst. Credit Card Inf. Security. Int. J. Cyber Secur. Digit. Forensics 2018, 7, 283–293. [Google Scholar]
  2. Bhasin, M.L. The Role of Technology in Combating Bank Frauds: Perspectives and Prospects. Ecoforum J. 2016, 5, 200–212. [Google Scholar]
  3. Rushin, G.; Stancil, C.; Sun, S.; Adam, S.; Beling, P. Horse Race Analysis in Credit Card Fraud—Deep Learning, Logistic Regression, and Gradient Boosted Tree. In Proceedings of the 2017 Systems and Information Engineering Design Symposium (SIEDS), Charlottesville, VA, USA, 28 April 2017; IEEE: New York, NY, USA, 2017; pp. 117–121. [Google Scholar]
  4. Lucas, Y.; Portier, P.-E.; Laporte, L.; He-Guelton, L.; Caelen, O.; Granitzer, M.; Calabr, S. Towards Automated Feature Engineering for Credit Card Fraud Detection Using Multi-Perspective HMMs. Future Gener. Comput. Syst. 2020, 102, 393–402. [Google Scholar]
  5. Robinson, W.N.; Aria, A. Sequential Fraud Detection for Prepaid Cards Using Hidden Markov Model Divergence. Expert Syst. Appl. 2018, 91, 235–251. [Google Scholar]
  6. Ileberi, E.; Sun, Y.; Wang, Z. A Machine Learning Based Credit Card Fraud Detection using the GA Algorithm for Feature Selection. J. Big Data 2022, 9, 24. [Google Scholar]
  7. Khare, N.; Sait, S.Y. Credit Card Fraud Detection using Machine Learning Models and Collating Machine Learning Models. Int. J. Pure Appl. Math. 2018, 118, 825–837. [Google Scholar]
  8. Xuan, S.; Liu, G.; Li, Z.; Zheng, L.; Wang, S.; Jiang, C. Random Forest for Credit Card Fraud Detection. In Proceedings of the 2018 IEEE 15th International Conference on Networking, Sensing and Control (ICNSC), Zhuhai, China, 27–29 March 2018; IEEE: Zhuhai, China, 2018; pp. 1–6. [Google Scholar]
  9. Ogundile, O.M.; Owoade, A.A.; Ogundile, O.O.; Babalola, O.P. Linear Discriminant Analysis Based Hidden Markov Model for Detection of Mysticetes’ Vocalisations. Sci. Afr. 2024, 24, e02128. [Google Scholar]
  10. Ogundile, O.O.; Usman, A.M.; Versfeld, D.J.J. An Empirical Mode Decomposition Based Hidden Markov Model Approach for Detection of Bryde’s Whale Pulse Calls. J. Acoust. Soc. Am. 2020, 147, EL125–EL131. [Google Scholar] [PubMed]
  11. Babalola, O.P.; Usman, A.M.; Ogundile, O.O.; Versfeld, D.J.J. Detection of Bryde’s Whale Short Pulse Calls using Time Domain Features with Hidden Markov Models. S. Afr. Inst. Electr. Eng. 2021, 112, 15–23. [Google Scholar]
  12. Ogundile, O.O.; Babalola, O.P.; Odeyemi, S.G.; Rufai, K.I. Hidden Markov Models for Detection of Mysticetes Vocalisations Based on Principal Component Analysis. Bioacoustics 2022, 31, 710–738. [Google Scholar]
  13. Alkarkhi, A.F.M.; Alqaraghuli, W.A.A. Chapter 8—Principal Components Analysis. In Easy Statistics for Food Science with R; Alkarkhi, A.F.M., Alqaraghuli, W.A.A., Eds.; Academic Press: London, UK, 2019; pp. 125–141. [Google Scholar]
  14. Syms, C. Principal Components Analysis. In Encyclopedia of Ecology; Jørgensen, S.E., Fath, B.D., Eds.; Elsevier: London, UK, 2008; pp. 2940–2949. [Google Scholar]
  15. MacQueen, J. Some Methods for Classification and Analysis of Multivariate Observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Statistics; University of California Press: Berkeley, CA, USA, 1967; pp. 281–297. [Google Scholar]
  16. Forgy, E.W. Cluster Analysis of Multivariate Data: Efficiency versus Interpretability of Classification. Biometrics 1965, 21, 768–780. [Google Scholar]
  17. Duda, R.; Hart, P.; Stork, D.G. Pattern Classification, 2nd ed.; Wiley: New York, NY, USA, 2001. [Google Scholar]
  18. Reynolds, D. Gaussian Mixture Models. In Encyclopedia of Biometrics; Li, S.Z., Jain, A., Eds.; Springer: Boston, MA, USA, 2009; pp. 659–663. [Google Scholar]
  19. Ololade, B.M.; Salawu, M.K.; Adekanmi, A.D. E-Fraud in Nigerian Banks: Why and How? J. Financ. Risk Manag. 2020, 9, 211–228. [Google Scholar]
  20. Dornadula, V.N.; Geetha, S. Credit Card Fraud Detection using Machine Learning Algorithms. Proc. Comput. Sci. 2019, 165, 631–641. [Google Scholar]
  21. Sahithi, G.L.; Roshmi, V.; Sameera, Y.V.; Pradeepini, G. Credit Card Fraud Detection using Ensemble Methods in Machine Learning. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 28–30 April 2022; pp. 1237–1241. [Google Scholar]
  22. Tanouz, D.; Subramanian, R.R.; Eswar, D.; Reddy, G.V.P.; Kumar, A.R.; Praneeth, C.H.V.N.M. Credit Card Fraud Detection Using Machine Learning. In Proceedings of the 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 967–972. [Google Scholar]
  23. Sadgali, I.; Sael, N.; Benabbou, F. Performance of machine learning techniques in the detection of financial frauds. Procedia Comput. Sci. 2019, 148, 45–54. [Google Scholar]
  24. Khan, A.; Singh, T.; Sinhal, A. Observation Probability in Hidden Markov Model for Credit Card Fraudulent Detection System. In Proceedings of the Second International Conference on Soft Computing for Problem Solving (SocProS 2012), Jaipur, India, 28–30 December 2012; Springer: New Delhi, India, 2014; Volume 236, pp. 751–760. [Google Scholar]
  25. Fashoto, S.G.; Owolabi, O.; Adeleye, O.; Wandera, J. Hybrid Methods for Credit Card Fraud Detection Using K-means Clustering with Hidden Markov Model and Multilayer Perceptron Algorithm. Br. J. Appl. Sci. Technol. 2016, 13, 1–11. [Google Scholar]
  26. Wang, X.; Wu, H.; Yi, Z. Research on Bank Anti-fraud Model Based on K-means and Hidden Markov Model. In Proceedings of the 2018 3rd IEEE International Conference on Image, Vision and Computing, Chongqing, China, 27–29 June 2018; IEEE: New York, NY, USA, 2018; pp. 780–784. [Google Scholar]
  27. Tharwat, A. Principal Component Analysis—A Tutorial. Int. J. Appl. Sci. Eng. 2009, 7, 41–61. [Google Scholar]
  28. Aristov, V.V.; Buchelnikov, A.S.; Nechipurenko, Y.D. The Use of the Statistical Entropy in Some New Approaches for the Description of Biosystems. Entropy 2022, 24, 172. [Google Scholar] [CrossRef] [PubMed]
  29. Nalband, S.; Prince, A.; Agrawal, A. Entropy-Based Feature Extraction and Classification of Vibroarthographic Signal Using Complete Ensemble Empirical Mode Decomposition with Adaptive Noise. IET Sci. Meas. Technol. 2018, 12, 350–359. [Google Scholar]
  30. Richman, J.S.; Lake, D.E.; Moorman, J.R. Sample Entropy. In Numerical Computer Methods, Part E; Walker, J.M., Ed.; Academic Press: London, UK, 2004; Volume 384, pp. 172–184. [Google Scholar]
  31. Baum, L.E.; Petrie, T.; Soules, G.; Weiss, N. A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains. Ann. Math. Stat. 1970, 41, 164–171. [Google Scholar]
  32. Viterbi, A. Error Bounds for Convolutional Codes and an Asymptotically Optimum Decoding Algorithm. IEEE Trans. Inf. Theory 1967, 13, 260–269. [Google Scholar]
  33. Randhawa, K.; Loo, C.K.; Seera, M.; Lim, C.P.; Nandi, A.K. Credit card fraud detection using AdaBoost and majority voting. IEEE Access 2018, 6, 14277–14284. [Google Scholar]
  34. Khalid, A.R.; Owoh, N.; Uthmani, O.; Ashawa, M.; Osamor, J.; Adejoh, J. Enhancing Credit Card Fraud Detection: An Ensemble Machine Learning Approach. Big Data Cogn. Comput. 2024, 8, 6. [Google Scholar] [CrossRef]
  35. Jadhav, S.D.; Channe, H.P. Comparative study of K-NN, naive Bayes and decision tree classification techniques. Int. J. Sci. Res. 2016, 5, 1842–1845. [Google Scholar]
  36. Ogundile, O.O.; Owoade, A.A.; Emeka, P.B.; Olaniyan, O.S. Animals’ Classification: A Review of Different Machine Learning Classifiers. J. Sci. Logics ICT Res. 2023, 9, 106–114. [Google Scholar]
Figure 1. EHMM credit card prediction model.
Figure 1. EHMM credit card prediction model.
Applsci 14 07389 g001
Figure 2. PCA-EHMM sensitivity versus dimension performance.
Figure 2. PCA-EHMM sensitivity versus dimension performance.
Applsci 14 07389 g002
Figure 3. PCA-EHMM specificity versus dimension performance.
Figure 3. PCA-EHMM specificity versus dimension performance.
Applsci 14 07389 g003
Figure 4. PCA-EHMM precision versus dimension performance.
Figure 4. PCA-EHMM precision versus dimension performance.
Applsci 14 07389 g004
Figure 5. PCA-EHMM F1-score versus dimension performance.
Figure 5. PCA-EHMM F1-score versus dimension performance.
Applsci 14 07389 g005
Table 1. Summary of some statistical properties of the dataset.
Table 1. Summary of some statistical properties of the dataset.
S P Time V 1 V 2 V 3 V 4 V 5
Minimum0−56.4075−72.7157−48.3256−5.6832−113.7433
Maximum1727922.454922.05779.382616.875334.8017
Mean948140.00000.00000.00000.00000.0000
Median846920.01810.06550.1799−0.0199−0.5434
1st quartile54202−0.9204−5.5986−0.8904−0.8486−0.6916
3rd quartile1393211.31560.80371.02720.74330.6119
S P V 6 V 7 V 8 V 9 V 10 V 11
Minimum−26.1605−43.5572−73.2167−13.4341−24.5883−4.7975
Maximum73.3016120.589520.007215.595023.745112.0189
Mean0.00000.00000.00000.00000.00000.0000
Median−0.27420.04010.0224−0.0514−0.0929−0.0328
1st quartile−0.7683−0.5541−0.2086−0.6431−0.5354−0.7625
3rd quartile0.39860.57040.327350.59710.453920.7396
S P V 12 V 13 V 14 V 15 V 16 V 17
Minimum−12.6837−5.7919−19.2143−4.4989−14.1295−25.1628
Maximum7.84847.126810.52688.877717.31519.2535
Mean0.00000.00000.00000.00000.00000.0000
Median0.1400−0.01360.05060.04810.0664−0.6568
1st quartile−0.4056−0.6485−0.4256−0.5829−0.4680−0.4838
3rd quartile0.61820.66250.49310.64880.52330.3997
S P V 18 V 19 V 20 V 21 V 22 V 23
Minimum−9.4987−7.2135−54.4977−34.8304−10.9331−44.8077
Maximum5.04115.592039.420927.202810.503022.5284
Mean0.00000.00000.00000.00000.00000.0000
Median−0.00360.0037−0.0625−0.02950.0068−0.0112
1st quartile−0.4989−0.4563−0.2117−0.2284−0.5424−0.1619
3rd quartile0.50080.45890.13300.18640.52860.1476
S P V 24 V 25 V 26 V 27 V 28 Amount
Minimum−2.8366−10.2954−2.6046−22.5657−15.43010.0000
Maximum4.58467.51963.517331.612233.847825,691.1600
Mean0.00000.00000.00000.00000.000088.3500
Median0.04100.0166−0.05210.00130.011222.0000
1st quartile−0.3546−0.3172−0.3270−0.0708−0.05305.6000
3rd quartile0.43950.35070.24100.09100.078377.1700
S P Class
Minimum0.0000
Maximum1.0000
Mean0.0017
Median0.0000
1st quartile0.0000
3rd quartile0.0000
Table 2. PCA-EHMM performance as a function of k.
Table 2. PCA-EHMM performance as a function of k.
k = 7
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.3399.2799.1999.26
7599.3999.3999.3199.35
8099.4099.3899.3099.35
d = 8
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.5899.4999.4199.49
7599.6099.6199.5299.56
8099.6099.5999.5399.57
k = 9
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.6999.7099.6399.66
7599.7999.8299.7199.75
8099.8199.8299.7299.77
k = 10
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.6999.7199.6399.66
7599.8099.8199.7099.75
8099.8099.8299.7299.76
k = 11
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.6899.7099.6199.65
7599.7999.7999.7199.75
8099.8099.8199.7299.76
k = 12
χ (%) S e (%) S p (%) P (%) F 1 (%)
7099.6999.7199.6199.65
7599.7999.8099.7299.75
8099.7999.8199.7199.75
Table 3. MRE-EHMM and PCA-EHMM performance comparison; h = 5 and k = 9 .
Table 3. MRE-EHMM and PCA-EHMM performance comparison; h = 5 and k = 9 .
χ (%)MRE-EHMMPCA-EHMMMRE-EHMMPCA-EHMMMRE-EHMMPCA-EHMMMRE-EHMMPCA-EHMM
S e (%) S p (%) P (%) F 1 (%)
7099.6899.6999.7099.7099.6299.6399.6599.66
7599.8099.7999.8099.8299.7299.7199.7699.75
8099.8199.8199.8199.8299.7399.7299.7799.77
Table 4. MRE and PCA performance comparison with EHMM and HMM; h = 5 and k = 9 .
Table 4. MRE and PCA performance comparison with EHMM and HMM; h = 5 and k = 9 .
χ (%)Predictor S e (%) S p (%) P (%) F 1 (%)
70MRE-EHMM99.6899.7099.6299.65
MRE-HMM71.2270.9171.0571.13
PCA-EHMM99.6999.7099.6399.66
PCA-HMM71.2070.9671.0771.13
75MRE-EHMM99.8099.8099.7299.76
MRE-HMM71.6571 6171.4471.54
PCA-EHMM99.7999.8299.7199.75
PCA-HMM71.6671.6071.4571.56
80MRE-EHMM99.8199.8199.7399.77
MRE-HMM71.7971.6771.5171.64
PCA-EHMM99.8199.8299.7299.77
PCA-HMM71.8071.6671.5071.65
Table 5. MRE-EHMM and PCA-EHMM performance comparison with other ML models.
Table 5. MRE-EHMM and PCA-EHMM performance comparison with other ML models.
χ (%)Predictor S e (%) S p (%) P (%) F 1 (%)
70MRE-EHMM99.6899.7099.6290.65
PCA-EHMM99.6999.7099.6399.66
RF97.0197.0996.9897.00
LR96.7796.7996.6596.71
DT96.2396.2596.1096.17
KNN95.8795.8695.6295.75
75MRE-EHMM99.8099.8099.7299.76
PCA-EHMM99.7999.8299.7199.75
RF97.2297.2497.1997.16
LR96.9996.9796.8896.94
DT96.4496.4396.2196.32
KNN96.0496.0395.8995.96
80MRE-EHMM99.8199.8199.7399.77
PCA-EHMM99.8199.8299.7299.77
RF97.5697.5497.2997.42
LR97.1197.0997.0097.06
DT96.8896.8996.7896.83
KNN96.5496.5796.4896.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ogundile, O.; Babalola, O.; Ogunbanwo, A.; Ogundile, O.; Balyan, V. Credit Card Fraud: Analysis of Feature Extraction Techniques for Ensemble Hidden Markov Model Prediction Approach. Appl. Sci. 2024, 14, 7389. https://doi.org/10.3390/app14167389

AMA Style

Ogundile O, Babalola O, Ogunbanwo A, Ogundile O, Balyan V. Credit Card Fraud: Analysis of Feature Extraction Techniques for Ensemble Hidden Markov Model Prediction Approach. Applied Sciences. 2024; 14(16):7389. https://doi.org/10.3390/app14167389

Chicago/Turabian Style

Ogundile, Olayinka, Oluwaseyi Babalola, Afolakemi Ogunbanwo, Olabisi Ogundile, and Vipin Balyan. 2024. "Credit Card Fraud: Analysis of Feature Extraction Techniques for Ensemble Hidden Markov Model Prediction Approach" Applied Sciences 14, no. 16: 7389. https://doi.org/10.3390/app14167389

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop