Next Article in Journal
Fibre Refractometry for Minimally Invasive Sugar Content Measurements within Produce
Previous Article in Journal
Covert Information Mapped Generalized Spatial and Direction Modulation toward Secure Wireless Transmission
Previous Article in Special Issue
Green Intrusion Detection Systems: A Comprehensive Review and Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Focal Causal Temporal Convolutional Neural Networks: Advancing IIoT Security with Efficient Detection of Rare Cyber-Attacks

by
Meysam Miryahyaei
,
Mehdi Fartash
* and
Javad Akbari Torkestani
Department of Computer Engineering, Arak Branch, Islamic Azad University, Arak 38361-1-9131, Iran
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6335; https://doi.org/10.3390/s24196335
Submission received: 5 July 2024 / Revised: 29 July 2024 / Accepted: 9 August 2024 / Published: 30 September 2024
(This article belongs to the Special Issue Intrusion Detection Systems for IoT)

Abstract

:
The Industrial Internet of Things (IIoT) deals with vast amounts of data that must be safeguarded against tampering or theft. Identifying rare attacks and addressing data imbalances pose significant challenges in the detection of IIoT cyberattacks. Innovative detection methods are important for effective cybersecurity threat mitigation. While many studies employ resampling methods to tackle these issues, they often face drawbacks such as the use of artificially generated data and increased data volume, which limit their effectiveness. In this paper, we introduce a cutting-edge deep binary neural network known as the focal causal temporal convolutional neural network to address imbalanced data when detecting rare attacks in IIoT. The model addresses imbalanced data challenges by transforming the attack detection into a binary classification task, giving priority to minority attacks through a descending order strategy in the tree-like structure. This approach substantially reduces computational complexity, surpassing existing methods in managing imbalanced data challenges in rare attack detection for IoT security. Evaluation of various datasets, including UNSW-NB15, CICIDS-2017, BoT-IoT, NBaIoT-2018, and TON-IIOT, reveals an accuracy of over 99%, demonstrating the effectiveness of FCTCNNs in detecting attacks and handling imbalanced IoT data with efficiency.

1. Introduction

The widespread adoption of the Internet of Things (IoT) has significantly impacted various domains, including smart cities, cameras, smart industrial equipment, and other aspects of daily life. On the other hand, the IoT environment deals with massive amounts of data that need to be protected from tampering or theft. Detecting security attacks against the IoT domain requires intelligent techniques rather than relying on signature matching. Machine learning (ML) and deep learning (DL) approaches are effective in identifying these attacks and predicting intrusion behavior based on unknown patterns [1,2,3]. Machine learning plays a crucial role in building network intrusion detection systems. However, machine learning models trained on imbalanced cybersecurity data cannot effectively detect minority data and, consequently, attacks [4,5].
Using imbalanced data in machine learning leads models to successfully predict normal data (majority class) but fail to detect minority and rare attacks. In such cases, the models assign rare data to the normal or majority data class [6,7]. This challenge becomes more problematic in multi-class classification, as accurately detecting minority attack types is more critical than identifying normal traffic [8,9]. One of the existing solutions to address and manage this challenge is the generation of synthetic data, achieved through methods like oversampling [10,11,12] and undersampling [13], to balance the data [14,15,16,17,18].
Studies focusing on this challenge often utilize methods such as the SMOTE family of algorithms [19,20,21], Generative Adversarial Networks (GANs) [22], and similar techniques. These methods primarily aim to produce synthetic data to augment the minority class [17,23]. In [24], the Bee Colony and Borderline SMOTE were used to address data imbalances for attack detection. In [5,25], ADASYN was employed for oversampling rare attacks. GANs were utilized in [4,22,23,26,27]; these methods have been somewhat more successful compared to SMOTE-based models. However, the quality of generated synthetic data and increased computational complexity due to the generation of new data are major limitations of such methods, leading to their partial success. Network traffic data are inherently classified as big data, and generating data in highly rare attack classes, which significantly differ from the majority class, results in a substantial increase in data volume and consequently in computational and temporal complexity.
In this paper, to overcome the data imbalance problem in attack detection within IoT, unlike existing methods, we propose a deep focal architecture based on causal temporal convolutional networks (TCNNs) that effectively manages data imbalances without data manipulation. This model transforms the multi-class classification problem into a binary classification and considers the population of each class for classification. In this model, a novel approach to addressing the data imbalance issue is proposed, which, unlike existing data-centric methods, is model-centric. This method considers and classifies both classes with nearly equal sample sizes at each level of its hierarchical structure. As a result, the model does not encounter data imbalance problems at any level of the proposed hierarchical structure. Consequently, the issue of data imbalance is managed without data manipulation or the generation of new synthetic data. The main components of our method are as follows:
  • Proposing a new binary deep model based on causal temporal convolutional networks (TCNN) for detecting rare attacks in traffic data streams.
  • Managing the data imbalance challenge structurally, without altering the data.
  • Transforming the multi-class attack detection problem into a binary classification problem and detecting two balanced types of attacks at each level of the deep hierarchical structure to address the challenge of data imbalance.
  • Utilizing a focal loss function in the proposed deep learning architecture.
Subsequently, in Section 2, we review the research background, including related studies. Section 3 introduces the proposed method, while Section 4 discusses the results obtained, and the Section 5 provides conclusions and future work directions.

2. Literature Review

2.1. Temporal Convolutional Neural Networks (TCNNs)

Temporal convolutional neural networks (TCNNs) are widely used in image recognition due to their powerful pattern recognition capabilities. Recently, their application has expanded to sequence data prediction, such as weather forecasts, stock markets, and more. Similar to a simple neural network, deep CNNs consist of multiple neurons connected through a hierarchical structure, and the weights between layers are trainable. Unlike other deep network structures, such as fully connected deep networks like Deep Belief Networks (DBN) and Sparse Autoencoder (SAE), CNNs can share weights between neurons of each layer, significantly reducing the network’s weight and preventing it from falling into local optima [28].
A convolutional neural network consists of an input layer, an output layer, and several hidden layers [29]. These layers are generally categorized into three main types: convolution layers, pooling layers, and fully connected layers. A convolutional neural network comprises a wide spectrum of convolution and pooling layers, followed by fully connected layers, meaning one or more fully connected layers are inserted after numerous convolution and pooling layers in the network.
The main unit of a convolutional network is the convolution layer, where most computations happen. This layer comprises an array of ordered neurons, and its parameters consist of a set of learnable filters or kernels. These filters combine with the input data to produce a 2D feature map, which, when stacked along the depth dimension, generates the layer’s output. Neurons in this layer share weights, thus reducing the network’s complexity by keeping the number of parameters low [30]. The convolution layer also comes in other forms, such as the causal convolution layer, which is used in temporal data. Causal convolution represents a specific convolution layer that guarantees the model’s adherence to the sequential order in which the data are processed [31]. Figure 1 presents a stack of causal convolution layers.
The primary structure of a convolutional network is composed of a sequence of convolution layers and pooling layers. The role of the pooling layer is to decrease the spatial dimensions of the feature maps generated by the convolution layer while keeping important information. By doing so, the pooling layer reduces the network’s parameter count, resulting in a reduction in the overall computational intricacy. This layer effectively controls the problem of overfitting. Some common pooling operations include max pooling, average pooling, r Stochastic Pooling [33], spectral pooling [34], spatial pyramid pooling, and multiscale orderless pooling [30,35].
Neurons in the fully connected layer are fully connected to all neurons in the previous layer, resembling a standard neural network. High-level reasoning occurs in this layer. Neurons are disordered spatially and are one-dimensional, meaning there cannot be a fully connected layer immediately following another fully connected layer. Lately, in certain architectural designs, the global average pooling layer has been introduced as a substitution for the fully connected layer [30].

2.2. The Cost-Sensitive Loss Function in Deep Neural Network Learning

A deep neural network uses a loss function to optimize its parameters. Typically, the default loss function is cross-entropy, using the sigmoid function for binary class classification and the SoftMax function for multi-class classification. The regular cross-entropy loss used for classification is mathematically defined as Equation (1):
C E p , y = log P         i f                         y = 1 log 1 p     o t h e r w i s e
The function C E p , y represents the cross-entropy loss function, which is commonly used in classification tasks. Here is a detailed explanation of its parameters:
  • p: The predicted probability that the instance belongs to the positive class (class 1).
  • y: The true label, which can be 0 or 1.
When the true label y = 1 , the loss function becomes log P . This penalizes the model more if p is close to 0, which indicates a wrong prediction with high confidence.
When the true label y = 0 , the loss function becomes log 1 p . This penalizes the model more if p is close to 1, indicating a wrong prediction with high confidence.
In an imbalanced dataset, a large class dominates the cost of a gradient. If C E   ( p t ) = C E   ( p , y ) , α t balances the importance of positive and negative examples. Lin et al. [36] changed the cross-entropy error by adding a modulating factor (1 – pt)γ with an adjustable parameter γ     0 , referred to as a focal loss function. Formally, the focal loss function is expressed as Equation (2), where γ represents a positive scaling factor:
F L p t = α t 1 p t γ log p t         w h e r e     α t = α                   i f                 y = 1 1 α           o t h e r w i s e
where p t , is the model’s predicted probability for the true class. If the true class is y , p t is the predicted probability for class y . α t is a weighting factor for the class, which helps balance the importance of positive versus negative examples. γ is a focusing parameter that adjusts the rate at which easy examples are down-weighted.
  • If y = 1 (positive class): αt = α
  • If y = 0 (negative class): αt = 1 – α
  • α is a scalar factor that adjusts the weight of positive examples. This helps to balance the contribution of positive and negative examples to the loss function.
  • Focusing parameter γ: It adjusts the rate at which easy examples are down-weighted. When γ = 0, the focal loss function becomes equivalent to the standard cross-entropy loss. As γ increases, the focus shifts more towards hard examples, making the model pay more attention to examples where the model’s prediction is incorrect or uncertain.
As mentioned in Equation (2), a combination of cross-entropy and the modulating factor 1 p t γ is used because it leads to better accuracy. In [36], a comparison between the cross-entropy w = 0 and focal loss functions is presented. The term α t 1 p t γ is inversely proportional and depends on the value of p t .
Focal loss functions are emerging as a method of cost-sensitive learning that balances the cross-entropy loss to train hard negative examples, including rare classes [36,37,38]. The authors confirmed the capability of focal loss functions in a detection process by considering imbalanced classes. Focal loss functions are suggested to change the shape and adjust cross-entropy to achieve better classification based on the loss function for dealing with imbalanced classes. Nevertheless, the utility of focal loss functions extends beyond the realm of detection tasks. Regardless of the number of classes and tasks, adjusting learning based on sample difficulty is effective [37].
Focal loss functions can significantly improve the performance of multi-class samples. This aligns with our theoretical hypothesis. The advantage of focal loss functions is dynamically adjusting the weight of mispredicted samples. This function can predict samples that occupy a small proportion of the entire dataset [39].

2.3. Related Works

Different intrusion detection models have tackled the challenge of imbalanced data [10,40,41,42,43,44,45,46,47,48,49,50,51]. These models often rely on data manipulation techniques, including oversampling and undersampling. In 2020, Zhang et al. [14] used a combination of SMOTE and Gaussian Mixture Clustering to balance the data and performed attack detection using convolutional neural networks (CNN). Their experimental results on the UNSW-NB15 and CICIDS2017 datasets demonstrated an accuracy of 96%. Yilmaz et al. [52] employed a Generative Adversarial Network (GAN) and a Multi-Layer Perceptron (MLP) for identifying rare attacks. Results on the UGR’16 dataset indicated that using a GAN for generating artificial data could moderately improve results compared to imbalanced original data. Panigrahi et al. [53] used Supervised Relative Random Sampling (SRRS) to reduce majority class data and employed the C4.5 decision tree algorithm for attack detection. Their results on the NSL-KDD and CICIDS2017 datasets reported accuracies of 99.96% and 99.95%, respectively. In 2022, Ding et al. [2] employed a combination of Generative Adversarial Networks (GANs) and nearest neighbor techniques to handle imbalanced data. Their results across the KDD99, UNSW-NB15, and CICIDS2017 datasets achieved accuracies ranging from 90% to 95%. Huang et al. [54] based their model on a GAN and used convolutional neural networks for classifying attacks. Results on the NSL-KDD, UNSW-NB15, and CICIDS2017 datasets demonstrated accuracies ranging from 82% to 99%. FU et al. [5] used ADASYN, an improved version of SMOTE, to balance the data and performed attack detection using Bidirectional Long Short-Term Memory (BiLSTM). They achieved an accuracy of 90.73% on the NSL-KDD dataset. Yan et al. [55] employed density-based clustering with DPeak for data balancing. Results on the KDD99-2D, CIC-IDS2017, and ISCXID2012 datasets reported accuracies between 79% and 94%. Reem Alshamy et al. [21] used a combination of Random Forest and SMOTE. Their results on the NSL-KDD dataset showed the success of their model with an accuracy of 99.89%. Malik AL-Essa [56] used XGBoost and SMOTE. Similarly, Huang Le et al. [8] also used the XGBoost algorithm. Results on the X-IIoTDS and TON_IoT attack datasets showed an F1 score of around 99%.
In 2022, Cao et al. [57] applied convolutional neural networks and a Gated Recurrent Unit (GRU) to perform attack detection. Results across the UNSW_NB15, NSL-KDD, and CIC-IDS2017 datasets yielded accuracies of 86.25%, 99.69%, and 99.65%. Balyan et al. [58] employed genetic algorithms, Particle Swarm Optimization with Estimation of Distribution (EGA-PSO), and Improved Random Forest (IRF). EGA-PSO was used to increase samples of minority class data. Their evaluation of the NSL-KDD dataset showed an accuracy of 98%. Jung et al. [59] introduced an approach to tackle imbalanced data challenges using SMOTE and a changed neural network that increases the minority class and removes noisy data. Their method showed promising results in evaluations of the PKDD2007 and CSIC2012 datasets. Dar et al. [60] used a Variational Autoencoder and a Multi-Layer Perceptron (MLP) to address imbalanced data. Their evaluation of HDFS data demonstrated an F1 score of 97%. Kayo et al. [61] employed ADASYN for balancing and conducted attack classification using convolutional and Bidirectional GRU (BiGRU) neural networks. They compared their new method with their previous approach [57] and showed that the new method outperformed the previous one. In the UNSW_NB15, NSL-KDD, and CIC-IDS2017 datasets, the new method achieved accuracies of 85.55%, 99.81%, and 99.70%, respectively. In 2022, Tareq et al. [62] used two intelligent network models, namely DenseNet and Inception Time, for cyber-attack detection. They evaluated their models on the ToN-IoT, Edge-IIoT, and UNSW2015 datasets. Their best result was a 99.9% accuracy for Windows 10 with DenseNet. In 2023, Thockchom et al. [63] proposed a group-based learning approach for intrusion detection that outperforms individual classifiers. Their model utilizes lightweight machine learning models and Chi-square feature selection. This method proves effective in imbalanced datasets, reducing the cost of misclassifications, and demonstrates its utility in intrusion detection systems. In another study in 2023, Sarvar et al. [64] presented a machine learning-based anomaly detection approach for smart homes using various classifiers. Their experiments and evaluations on the BoT-IoT dataset resulted in accuracy, recall, and F1 scores of 0.98, 0.96, and 0.96, respectively.
The aforementioned studies reveal that most imbalanced data-focused methods rely on data manipulation, often attempting to augment the number of minority class data through artificial data generation. Data manipulation presents its own set of challenges. In most network traffic datasets, the imbalance is severe, and generating data for minority classes results in increased data volume. This, in turn, leads to issues such as increased computational complexity and model inefficiency in real-world scenarios. On the other hand, generating a large volume of data for the minority class may not ensure data quality, as the outcomes of many algorithms in the SMOTE family depend on the proper determination of an appropriate neighbor count ‘k.’. If the number of neighbors is not correctly determined for SMOTE family algorithms, it can result in the generation of low-quality synthetic data, consequently decreasing the performance of intrusion detection models. Therefore, data manipulation is not an ideal solution for addressing the challenge of imbalance. In this context, considering the above discussions, this article introduces an innovative approach to control and manage data imbalances. The primary objective of this approach is the detection of rare and minority attacks without data manipulation.

3. Proposed Architecture for Deep Binary Intrusion Detection

In this paper, a deep binary architecture named focal causal temporal convolutional neural network (FCTCNN) is proposed for detecting rare attacks in IoT and Industrial IoT environments, aimed at addressing the challenge of imbalanced data. As illustrated in Figure 2, this model defines a deep binary architecture comprising causal temporal convolutional networks. At each level, attacks with similar frequencies are identified and mitigated. One of the advantages of the proposed structure is its reduction in computational and temporal complexity due to the deepening of the architecture. Specifically, rare attacks with minimal sample sizes are detected at the leaf level, while majority attacks are detected at the root level. Consequently, by deepening the architecture, the model’s complexity is reduced and the learning time is significantly decreased. The subsequent sections provide a comprehensive explanation of the proposed model’s stages.
In Figure 2, the proposed deep causal hierarchical architecture is illustrated. This structure is composed of causal convolutional networks (depicted as blue squares in Figure 2), with the constituent layers of these networks shown on the left side of the figure. Briefly, in this method, network traffic data are sorted in descending order based on the number of samples in each class. At the highest level of the architecture, the first type of attack, which has the largest number of samples, along with the Normal class, which typically has the highest number of samples in network traffic data, are processed using the proposed method. These two instances are considered the majority classes in the data, which often leads to inadequate learning with conventional deep learning methods. However, the hierarchical model proposed in this study addresses this issue by defining a deep binary structure at each level.
The proposed network performs the learning process with these two labeled datasets and generates its predictions. At the subsequent level, data from these two classes are removed from the entire dataset, and two additional types of attacks with equal sample sizes are introduced to the model for further learning and classification. At each level, classified data are removed, and the remaining data from the database are presented to the network in pairs according to the sample sizes of each class until all data are classified.
As the volume of data decreases at each level of the model, and majority classes are eliminated at the first level of the tree structure, the number of data points decreases as the depth of the tree structure increases. Consequently, the model trains and predicts more efficiently with smaller datasets. In this model, the levels of the proposed structure vary based on the number of attacks for each database and are automatically adjusted to form a suitable structure for each dataset. The subsequent sections provide a more detailed explanation of the model’s stages.

3.1. Data Preprocessing

In network traffic data, there are many nominal features, such as port names, that machine learning methods cannot directly process. Therefore, a solution is required to convert these data into numerical values without sacrificing model accuracy. In our model, the One-Hot encoding technique is used to transform all nominal data into numerical data. Certainly, as illustrated in Figure 3, distinct variables within each nominal attribute are identified, and for each unique variable, a binary feature is incorporated into the dataset.

3.2. Handling Missing Values

The presence of missing values in network traffic data is another challenge that, if left unaddressed, can reduce the effectiveness of the intrusion detection model. Failing to adopt an appropriate strategy for managing missing data compromises the model’s performance. In this research model, the nearest neighbor technique is used to manage missing data, and the missing values are replaced with the values corresponding to the nearest neighbors based on Euclidean distance. Specifically, for each missing data point, the first Euclidean nearest neighbor is found, and the missing value is replaced with the corresponding value from the complete data.

3.3. Causal Temporal Binary Convolutional Architecture

The causal convolutional network defined in this study, as illustrated in Table 1, is composed of 11 layers and has over 42,000 learnable parameters.
The layers of this network are described as follows:
  • Input Layer: Receives data in a sequential format with dimensions of 1.
  • Second Layer: A 1-dimensional convolutional layer, where 64 filters of size [1 × 5] are defined, with a stride of 1. This means that each filter has a width of 5 units and a height of 1 unit. In practical terms, the filter slides over the input data, which is structured as a 1-dimensional sequence, to detect features of length 5. The stride is set to 1. This parameter controls how much the filter moves (or “strides”) across the input data. With a stride of 1, the filter shifts by one unit at a time. This results in a convolution operation that processes every possible position of the filter over the input sequence.
  • Third Layer: A ReLU activation function layer, which maps negative values to zero using the relation ReLU(x) = max (0, x). This layer passes positive values unchanged to the next layer.
  • Fourth Layer: A normalization layer that normalizes the input data across channels and applies a learnable affine transformation with parameters. Normalizing across channels typically involves adjusting the mean and variance of the data in each channel to ensure that they are standardized. This helps stabilize and accelerate the training process by reducing internal covariate shift. After normalization, the layer applies a learnable affine transformation. This involves two operations: Scaling: Each normalized value is multiplied by a learnable parameter called the scale factor (γ). Shifting: Each normalized value is then shifted by another learnable parameter called the shift factor (β).
This affine transformation allows the network to adjust and fine-tune the normalized data, making the normalization process more flexible and suitable for learning complex patterns. The parameters γ and β are learned during training and help the network recover any potential loss of representational power caused by the normalization.
  • Fifth Layer: A second convolutional layer with 128 filters, where the stride and filter dimensions are consistent with those of the first convolutional layer in the defined architecture.
  • The sixth and seventh layers: These are a ReLU activation layer and a normalization layer, respectively, as previously described.
  • Eighth Layer: This is a global average pooling layer that calculates the average along the 1D input length dimension. It computes the average value across the entire length of the 1D input dimension. Essentially, it takes the mean of all the values in each feature map along the sequence dimension. If the input to this layer is a 1D sequence of length L, the output for each feature map will be a single value representing the average of all L values in that feature map. The GAP layer significantly reduces the dimensionality of the data. By summarizing the entire sequence into a single value per feature map, it transforms a potentially large number of activations into a compact vector. This reduction helps in minimizing computational costs and model complexity. By summarizing the feature maps into a single value per feature map, the GAP layer helps in reducing the risk of overfitting. It does this by eliminating the fine-grained details that might cause the model to memorize the training data rather than generalize well.
  • Ninth Layer: This is a fully connected layer that multiplies the synaptic weights with the input data and computes the sum.
  • Final Layers: The architecture also includes a SoftMax layer and a classification layer that utilizes the focal loss function, which is effective in addressing data imbalances. The focal loss function is used in this layer to address class imbalances by focusing more on hard-to-classify examples and reducing the impact of well-classified examples. The SoftMax layer is used to produce probability distributions for the network’s output. It converts the raw scores (logits) from the previous layer into probabilities by exponentiating the scores and then normalizing them. Specifically, for each class, the SoftMax function computes the probability P(yi) as follows:
P y i = e Z i j e Z j
where Z i represents the score for class i and the denominator is the sum of exponentiated scores for all classes. This ensures that the output probabilities are positive and sum up to 1, making them suitable for classification tasks.
After establishing the causal convolution layers, which act as deep nodes at each level of the proposed hierarchical structure, the learning phase of the hierarchical structure must be executed. To train the network, the data are first divided into training and testing subsets at an 80% and 20% ratio, respectively. Subsequently, the network traffic data are sorted in descending order based on frequency (number of samples per class) as illustrated in Table 2. For the first level of the tree structure, data from each dataset are presented to the network with two labels: Normal and Attack. The model is trained once with the aforementioned data, and all data associated with the Normal class are removed from the dataset.
Following the initial training, the code progresses with a hierarchical approach to address different attack classes. The model is iteratively refined by focusing on different attack categories and removing benign (Normal) samples and classes with fewer samples. The model is retrained and evaluated for each level, with training settings adjusted based on the number of classes and remaining data. Finally, the network’s performance is assessed using ROC metrics and confusion matrices for each attack category, providing a comprehensive view of the model’s effectiveness across various attack levels. This hierarchical approach ensures that the model is designed to manage imbalanced datasets effectively, focusing on detecting less frequent but critical types of attacks.
In more detail, data are classified into two classes at the first level of our architecture: Normal and Non-Normal. Before completing learning and attack prediction in the testing phase, data corresponding to the Normal class are removed from the training dataset. At the second level of the proposed architecture, the first attack, which is larger compared to other attacks, is considered the primary class, and other attacks are treated as secondary attacks for classification. This process continues until the rarest attacks in each dataset are reached. At the final level of the proposed architecture, the two rarest attacks in the dataset are classified. As mentioned, at each level of the deep binary architecture, the detected attack is removed from the list of attacks, which in turn significantly reduces the computational complexity of the model. In our model, the Adam optimization function is used for training the network, and the number of epochs is increased from 50 to 200 with the increasing depth of the network. Additionally, the mini-batch size is reduced from 1024 at the root node to 16 at the leaf level of the proposed architecture. Figure 4 and Figure 5 show a sample learning curve of the proposed network at Level 1 (root node) for classifying data into Normal and Non-Normal (Attack) classes in the TON_IIOT and UNSWNB2015 datasets.

4. Evaluation and Results Analysis

For evaluating the model in this article, precision, recall, accuracy, specificity, average f, Matthews’ correlation coefficient, Cohen’s kappa, time, and ROC curve metrics were used. These metrics were calculated from Equations (4) to (11).
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
T i m e s = e n d t S t a r t t
S p e c i f i c i t y = T N T N + F P
P r e s i c i o n = T P T P + F P
M C C = T P     T N F P     F N T P + F P   T P + F N   T N + F P   T N + F N
K = 2 * ( T P     T N F P     F N ) T P + F P   T P + F N   T N + F P   T N + F N
In the equations above, the variable TP refers to data points that are actually positive and have been correctly predicted as positive by the model. Similarly, the variable TN refers to data points that are actually negative and have been correctly predicted as negative by the model. On the other hand, the variable FP refers to data points that are actually negative but have been incorrectly predicted as positive by the model, while FN refers to data points that are actually positive but have been incorrectly predicted as negative by the model.
The evaluation of models on diverse datasets is available, each with various types of attacks. Considering that the behavior of models can differ from one dataset to another, a comprehensive examination of the model’s performance was conducted using five different datasets, as follows:
TON-IIOT dataset: This dataset includes heterogeneous data sources collected from telemetry datasets of IoT and IIoT sensors, operating system datasets for Windows 7 and 10, etc. [66].
UNSW-NB15 dataset: This dataset consists of four sections, and Section 1 is used in this article. It comprises 47 features, two columns for attack labels, and over 700,000 records. The dataset includes nine types of attacks and a normal class [67].
BoT-IoT dataset: Designed in the Cyber Range laboratory at UNSW Canberra, this dataset simulates a real network environment. It incorporates a blend of normal and botnet traffic, presenting source files in various formats, including original pcap files, generated argus files, and CSV files. This dataset comprises over 72,000,000 records, encompassing attacks like DDoS, DoS, OS and Service Scan, Keylogging, and Exfiltration Data [68].
CICIDS-2017 dataset: This dataset includes six attack types and a normal class, composed of 80 features and two columns for attack labels, totaling 745,423 records. It covers attacks such as SQL injection, brute force, DDoS, SSH-Patator, Port-Scan, and more [69].
NBaIoT dataset: Focused on botnet attacks in the Internet of Things, this dataset is gathered from nine commercial IoT devices. It includes five BASHLITE and five Mirai attacks, featuring 115 features and 701,000 records [70]. In Figure 6, a two-dimensional view of the BoT-IoT and NBaIoT datasets is presented.
For model implementation, MATLAB 2023a was used on a system with nine processors, 32GB RAM, and an Nvidia RTX 3080 GPU. Data rates of 20% for testing and 80% for training were established. Network learning parameters included the Adam optimization function, initial learning rate of 0.01, Max Epoch count ranging from 50 to 200, and minibatch sizes ranging from 1024 to 16.

4.1. Results Analysis in the UNSW-NB2015 Dataset

Considering the imbalance challenge in this study, results for each dataset are reported on classes to assess the model’s performance on rare classes. Table 3 presents the detection results of our model for attack types in comparison with two models [2,71], and the LSTM network in [72] on the UNSW-NB2015 dataset.
UNSW-NB2015 is one of the datasets with a severe data imbalance issue. In this dataset, as indicated in Table 2, four types of attacks, such as Backdoors, have around 500 or fewer data samples, significantly lower than larger attacks. Balancing this data issue through data manipulation and synthetic data generation is unlikely to be very successful and may lead to computational complexity. However, our model, as shown in Table 3, effectively handles the detection of rare attacks without data manipulation.
In the case of the four minority attacks in this dataset, namely ‘Backdoors’, ‘Analysis’, ‘Shellcode’, and ‘Worms’, our deep binary structure has achieved very good results compared to other methods. In the three attacks ‘Analysis’, ‘Shellcode’, and ‘Worms’, our model has achieved 100% results across all metrics, whereas other models have obtained 0% results in these three attack types. In the ‘Analysis’ attack, the LSTM recurrent network [72] has failed in all metrics, indicating that even very deep networks face severe challenges in detecting extremely rare attacks if the data imbalance issue is not addressed. In this attack, a combination of k-nearest neighbors and an adversarial generative network achieved an F1-score of 83.17%, demonstrating that generating synthetic data using deep networks for minority classes can somewhat control the data imbalance challenge. However, this method still showed a weaker recall and average F1-score compared to our proposed deep model. In the ‘Shellcode’ attack, one of the rarest attacks in this dataset with around 223 data samples, other models like the LSTM deep network failed to yield results and completely failed. However, the KNN-GAN, which focuses on generating artificial data with a neural network, ranked second after the proposed model. The ResNet50-Oversample model, despite using oversampling and generating artificial data, still presented results below 20%, indicating that artificial data generation may not be very successful in this context.
Among the attacks in this dataset, our model struggled to succeed only in the ‘Exploits’ and ‘Generic’ attacks compared to other models. Nevertheless, in these two attack types, the results were still very close to other models. Overall, the results in Table 3 indicate that the proposed deep binary architecture based on causal temporal convolutional networks has performed well in detecting attacks across most classes, achieving an over 10% increase in accuracy compared to other methods. Figure 7 presents the ROC curve and confusion matrix of the proposed model in the UNSW-NB2015 dataset.

4.2. Evaluation of Results in the CIC-IDS2017 Dataset

Table 4 compares the results of our model with [63] on the CIC-IDS2017 dataset. In this dataset, the group model proposed in [63] completely failed in detecting two types of attacks. Meanwhile, our proposed model, due to its transformation of multi-class classification into binary classification, achieved an overall precision of over 99% for all attacks. Considering the frequency of attacks in their classification has allowed the model to effectively address the issue of data imbalance. In this dataset, the model achieved an average precision of 99.84% for all attacks, while the group method [63] attained a precision of 67.52% for all attacks. These results indicate that the introduction of the new architecture without data manipulation has led to success in detecting both rare and common attacks, surpassing 30%. Figure 8 presents ROC curve and one confusion matrix of the proposed model in this dataset.

4.3. Results Analysis in the TON-IIOT Dataset

Table 5 reports the results of our model in the heterogeneous Industrial Internet of Things (IIoT) dataset TON-IIOT and the average results across all classes are compared with the model by Tareq et al. [62] in Figure 9. In this dataset, four types of attacks—Injection, DOS, MITM, and Scanning—are rare, and the model achieved an average accuracy of over 97% and recall of over 99% for all four classes. Particularly, for the very rare MITM and Scanning attacks, the model demonstrated 100% accuracy, recall, and F1-scores. These results indicate that the proposed model can effectively control rare attacks across various datasets. The ROC curve diagram and confusion matrix for two rare attacks for the proposed model in this dataset are presented in Figure 10.
Comparing the results with two deep neural networks presented by Tareq et al. [62] shows a 2.3% increase in overall accuracy. Furthermore, in precision, recall, and average F1-score, our model outperformed those of Tareq et al. by 1.66%, 3.66%, and 2.64%, respectively. These comparative results highlight that our model has achieved greater success in this dataset compared to the latest existing methods. The nature of the temporal convolutional network contributes to the model’s success in detecting various types of attacks. The impact of the focal loss function cannot be ignored in this success.

4.4. Results Analysis in the BoT-IoT (UNSW_2018) Dataset

Table 6 presents the results of our model for each type of attack in the BoT-IoT dataset. Figure 11 compares the average results across all attacks using the approach proposed by Sarwar et al. [64]. The results of this evaluation show that the model successfully detected Keylog attacks with 94% accuracy and Data Exfiltration attacks with 50% accuracy, even though these attacks are rare, consisting of 73 and 6 data samples, respectively. Most studies struggle to predict these two types of attacks due to their rarity.
Comparing the results with the methods proposed by Sarwar et al. reveals that, among the deep learning methods used by Sarwar, the artificial neural network (ANN) model outperformed the recurrent neural network (RNN) and Autoencoder. In comparison with the ANN model in this dataset, our model improved accuracy by 1.52%, precision by 1%, and recall and F1-score by 4.60% and 2.65%, respectively. The successful classification of attacks with similar frequency and usage of an appropriate loss function for training the proposed deep neural network, along with the incorporation of temporal convolutional layers, contributes to the model’s success compared to existing methods on various datasets. The ROC curve diagram and confusion matrix for two rare attacks for the proposed model in this dataset are presented in Figure 12.

4.5. Results Analysis in the NBaIoT-2018 Dataset

In the final evaluation category, Table 7 reports the results of our model for detecting different attack types in the NBaIoT-2018 dataset. Figure 13 compares the results with models proposed by Samy et al. [72]. The results of this assessment demonstrate that a well-defined architecture leads to the complete success of our model in handling both common and rare attacks. For instance, the LSTM and BiLSTM methods presented by Samy et al. utterly failed in detecting the Gaf-Tcp attack, yielding 0 in all metrics. In contrast, our cutting-edge deep architecture achieved over 99% accuracy in this attack. While this dataset is not as imbalanced as others in this study, the proposed temporal convolutional binary architecture still provides very promising results. Among the three models, LSTM, BiLSTM, and FCTCNN, LSTM yielded the weakest results. Compared to LSTM, our model increased accuracy by over 19% and recall by over 10%.
In the second rank after our model, BiLSTM exhibited improved accuracy by over 17% and recall by over 10%. In the F1-score metric, our deep neural network outperformed the presented recurrent neural networks in [72] by 14 to 15%. The results across all five datasets indicate the success of our model compared to existing methods.
The results from various experiments highlight that a properly designed deep architecture can outperform data-centric methods that address the imbalance issue using artificial data generation. Overall, the findings in this paper, across five different datasets with various attacks, demonstrate that the proposed cutting-edge binary architecture based on causal temporal convolutional networks can effectively control the challenge of detecting rare attacks. Two sample ROC curve diagrams for our model in the NBaIoT dataset are presented in Figure 14.

4.6. Analysis of Time Complexity across All Data

As stated in the section introducing the proposed method, in our approach, increasing the level of the hierarchical structure leads to a reduction in computational and time complexity. This is because data from previous levels are eliminated at each level in this method, resulting in minimal data available for learning and prediction in the final levels. Consequently, the model can reduce time at the final levels. Figure 15 shows the time charts across five different datasets for each level of the model. These charts clearly demonstrate that the model’s time decreases from the beginning towards the end (the final levels). This reduction occurs because data are eliminated at each level.
The analysis shows that despite the hierarchical structure of the model, the time has not increased due to the reduction in data at each level. Compared to other deep neural networks, our method achieves lower prediction times. Recurrent neural networks like LSTM and BiLSTM have greater time complexity compared to our method, primarily due to the internal structure of their hidden cells, which involves loops and feedback. On the other hand, the KNN algorithm, known as a lazy learning algorithm, does not perform computations during the training phase but only during the evaluation phase. Consequently, its prediction time is longer compared to our model. This evaluation clearly demonstrates that the proposed structure has a favorable complexity compared to other deep neural networks. Table 8 presents the comparative results of average time (seconds) across five datasets.

5. Conclusions

In intrusion detection systems, class imbalances across different classes are a major challenge. The issue of class imbalance in intrusion detection datasets often confounds engineers and researchers. Despite the high accuracy achieved by shallow machine learning and deep neural network models, NIDS models suffer from high false positive rates and lower intrusion detection rates due to imbalanced datasets. Data imbalance in datasets refers to conditions where class distributions are disproportionately represented. This situation arises when a majority class significantly outweighs a minority class. Most existing studies have addressed this challenge using oversampling and undersampling techniques. Although these methods have been somewhat successful, they also introduce problems such as increased data volume and computational complexity. Therefore, this paper presents an innovative deep binary model based on deep convolutional neural networks with focal loss to address the issue of imbalanced data. In this model, multi-class classification is converted into a sequential binary classification problem. Evaluations on five different network traffic flow datasets show that the proposed model effectively addresses the problem of rare attacks and demonstrates the capability to detect them. Our model has not failed in any of the five studied datasets or attacks and has consistently achieved over 99% accuracy in identifying rare attacks. Future work could include the use of wrapper-based feature selection methods based on metaheuristic algorithms to further enhance the results. Additionally, a competitive framework could be introduced where two types of deep architectures compete at each level, with the winning model receiving more data in subsequent levels.

Author Contributions

Conceptualization, M.M. and M.F.; Methodology, M.M.; Software, M.M.; Validation, M.M.; Investigation, M.M., M.F. and J.A.T.; Resources, M.M.; Data curation, M.M.; Writing—original draft, M.M.; Writing—review & editing, M.F. and J.A.T.; Visualization, M.M.; Supervision, M.F.; Project administration, M.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

These data were derived from the following resources available in the public domain: CIC-IDS2017: https://www.unb.ca/cic/datasets/ids-2017.html, accessed on 24 May 2024, BoT-IoT, TON_IoT and UNSW-NB15: https://research.unsw.edu.au/projects/bot-iot-dataset, accessed on 24 May 2024, NBAIOT: https://www.kaggle.com/datasets/mkashifn/nbaiot-dataset, accessed on 24 May 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AcronymsDefinitions
CNNConvolutional neural network
IIoTIndustrial internet of things
IDSIntrusion detection system
FCTCNNFocal causal temporal convolutional neural network
DoSDenial of services
DDoSDistributed denial of services
NIDSNetwork-based intrusion detection system
ROCReceiver operating characteristic
MitMMan in the middle

References

  1. Bagui, S.; Li, K. Resampling imbalanced data for network intrusion detection datasets. J. Big Data 2021, 8, 6. [Google Scholar] [CrossRef]
  2. Ding, H.; Chen, L.; Dong, L.; Fu, Z.; Cui, X. Imbalanced data classification: A KNN and generative adversarial networks-based hybrid approach for intrusion detection. Future Gener. Comput. Syst. 2022, 131, 240–254. [Google Scholar] [CrossRef]
  3. Younisse, R.; Ahmad, A.; Abu Al-Haija, Q. Explaining Intrusion Detection-Based Convolutional Neural Networks Using Shapley Additive Explanations (SHAP). Big Data Cogn. Comput. 2022, 6, 126. [Google Scholar] [CrossRef]
  4. Lee, J.; Park, K. GAN-based imbalanced data intrusion detection system. Pers. Ubiquitous Comput. 2021, 25, 121–128. [Google Scholar] [CrossRef]
  5. Fu, Y.; Du, Y.; Cao, Z.; Li, Q.; Xiang, W. A deep learning model for network intrusion detection with imbalanced data. Electronics 2022, 11, 898. [Google Scholar] [CrossRef]
  6. Rani, M. Gagandeep Effective network intrusion detection by addressing class imbalance with deep neural networks multimedia tools and applications. Multimed. Tools Appl. 2022, 81, 8499–8518. [Google Scholar] [CrossRef]
  7. Cui, J.; Zong, L.; Xie, J.; Tang, M. A novel multi-module integrated intrusion detection system for high-dimensional imbalanced data. Appl. Intell. 2023, 53, 272–288. [Google Scholar] [CrossRef]
  8. Le, T.-T.; Oktian, Y.E.; Kim, H. XGBoost for Imbalanced Multiclass Classification-Based Industrial Internet of Things Intrusion Detection Systems. Sustainability 2022, 14, 8707. [Google Scholar] [CrossRef]
  9. Gan, B.; Chen, Y.; Dong, Q.; Guo, J.; Wang, R. A convolutional neural network intrusion detection method based on data imbalance. J. Supercomput. 2022, 78, 19401–19434. [Google Scholar] [CrossRef]
  10. Ahmed, H.A.; Hameed, A.; Bawany, N.Z. Network intrusion detection using oversampling technique and machine learning algorithms. PeerJ Comput. Sci. 2022, 8, e820. [Google Scholar] [CrossRef]
  11. Damtew, Y.G.; Chen, H. SMMO-CoFS: Synthetic Multi-minority Oversampling with Collaborative Feature Selection for Network Intrusion Detection System. Int. J. Comput. Intell. Syst. 2023, 16, 12. [Google Scholar] [CrossRef]
  12. Sun, Y.; Que, H.; Cai, Q.; Zhao, J.; Li, J.; Kong, Z.; Wang, S. Borderline SMOTE Algorithm and Feature Selection-Based Network Anomalies Detection Strategy. Energies 2022, 15, 4751. [Google Scholar] [CrossRef]
  13. Zhang, H.; Ge, L.; Wang, Z. A High Performance Intrusion Detection System Using LightGBM Based on Oversampling and Undersampling. In Proceedings of the Intelligent Computing Theories and Application: 18th International Conference, ICIC 2022, Xi’an, China, 7–11 August 2022; Proceedings, Part I, 2022. Springer: Berlin/Heidelberg, Germany, 2022; pp. 638–652. [Google Scholar]
  14. Zhang, H.; Huang, L.; Wu, C.Q.; Li, Z. An effective convolutional neural network based on SMOTE and Gaussian mixture model for intrusion detection in imbalanced dataset. Comput. Networks 2020, 177, 107315. [Google Scholar] [CrossRef]
  15. Tan, X.; Su, S.; Huang, Z.; Guo, X.; Zuo, Z.; Sun, X.; Li, L. Wireless sensor networks intrusion detection based on SMOTE and the random forest algorithm. Sensors 2019, 19, 203. [Google Scholar] [CrossRef]
  16. Wu, T.; Fan, H.; Zhu, H.; You, C.; Zhou, H.; Huang, X. Intrusion detection system combined enhanced random forest with SMOTE algorithm. EURASIP J. Adv. Signal Process. 2022, 2022, 39. [Google Scholar] [CrossRef]
  17. ALFRHAN, A.A.; ALHUSAIN, R.H.; Khan, R.U. SMOTE: Class imbalance problem in intrusion detection system. In Proceedings of the 2020 International Conference on Computing and Information Technology (ICCIT-1441), Tabuk, Saudi Arabia, 9–10 September 2020; IEEE: Piscataway, NJ, USA; pp. 1–5. [Google Scholar]
  18. Pimsarn, C.; Boongoen, T.; Iam-On, N.; Naik, N.; Yang, L. Strengthening intrusion detection system for adversarial attacks: Improved handling of imbalance classification problem. Complex Intell. Syst. 2022, 8, 4863–4880. [Google Scholar] [CrossRef]
  19. Massaoudi, M.; Refaat, S.S.; Abu-Rub, H. Intrusion Detection Method Based on SMOTE Transformation for Smart Grid Cybersecurity. In Proceedings of the 2022 3rd International Conference on Smart Grid and Renewable Energy (SGRE), Doha, Qatar, 20–22 March 2022; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar]
  20. Awujoola, O.J.; Ogwueleka, F.N.; Irhebhude, M.E.; Misra, S. Wrapper based approach for network intrusion detection model with combination of dual filtering technique of resample and SMOTE. In Artificial Intelligence for Cyber Security: Methods, Issues and Possible Horizons or Opportunities; Springer: Berlin/Heidelberg, Germany, 2021; pp. 139–167. [Google Scholar]
  21. Alshamy, R.; Ghurab, M.; Othman, S.; Alshami, F. Intrusion detection model for imbalanced dataset using SMOTE and random forest algorithm. In Proceedings of the Advances in Cyber Security: Third International Conference, ACeS 2021, Penang, Malaysia, 24–25 August 2021; Revised Selected Papers 3. Springer: Berlin/Heidelberg, Germany, 2021; pp. 361–378. [Google Scholar]
  22. Ahsan, R.; Shi, W.; Ma, X.; Croft, W.L. A comparative analysis of CGAN-based oversampling for anomaly detection. IET Cyber-Phys. Syst. Theory Appl. 2022, 7, 40–50. [Google Scholar] [CrossRef]
  23. Alabrah, A. A Novel Study: GAN-Based Minority Class Balancing and Machine-Learning-Based Network Intruder Detection Using Chi-Square Feature Selection. Appl. Sci. 2022, 12, 11662. [Google Scholar] [CrossRef]
  24. Priyadarsini, P.I. ABC-BSRF: Artificial Bee colony and borderline-SMOTE RF algorithm for intrusion detection system on data imbalanced problem. In Proceedings of the International Conference on Computational Intelligence and Data Engineering: ICCIDE 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 15–29. [Google Scholar]
  25. Chen, Z.; Zhou, L.; Yu, W. ADASYN− Random Forest Based Intrusion Detection Model. In Proceedings of the 2021 4th International Conference on Signal Processing and Machine Learning, Beijing, China, 18–20 August 2021; pp. 152–159. [Google Scholar]
  26. Ding, S.; Kou, L.; Wu, T. A GAN-based intrusion detection model for 5G enabled future metaverse. Mob. Networks Appl. 2023, 27, 2596–2610. [Google Scholar] [CrossRef]
  27. Kumar, V.; Sinha, D. Synthetic attack data generation model applying generative adversarial network for intrusion detection. Comput. Secur. 2023, 125, 103054. [Google Scholar] [CrossRef]
  28. Hu, Z.; Zhao, Y.; Khushi, M. A survey of forex and stock price prediction using deep learning. Appl. Syst. Innov. 2021, 4, 9. [Google Scholar] [CrossRef]
  29. Sakib, S.; Ahmed, N.; Kabir, A.J.; Ahmed, H. An overview of convolutional neural network: Its architecture and applications. Preprints 2019, 2018110546. [Google Scholar]
  30. Aloysius, N.; Geetha, M. A review on deep convolutional neural networks. In Proceedings of the 2017 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 6–8 April 2017; IEEE: Piscataway, NJ, USA; pp. 0588–0592. [Google Scholar]
  31. Robinson, J.; Kuzdeba, S.; Stankowicz, J.; Carmack, J.M. Dilated causal convolutional model for RF fingerprinting. In Proceedings of the 2020 10th Annual Computing and Communication Workshop and Conference (CCWC), Chennai, India, 6–8 April 2020; IEEE: Piscataway, NJ, USA; pp. 0157–0162. [Google Scholar]
  32. Oord, A.V.D.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. WaveNet: A Generative Model for Raw Audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  33. Zeiler, M.D.; Fergus, R. Stochastic pooling for regularization of deep convolutional neural networks. arXiv 2013, arXiv:1301.3557. [Google Scholar]
  34. Rippel, O.; Snoek, J.; Adams, R.P. Spectral representations for convolutional neural networks. arXiv 2015, arXiv:1506.03767. Available online: https://arxiv.org/abs/1506.03767 (accessed on 24 May 2024).
  35. Gong, Y.; Wang, L.; Guo, R.; Lazebnik, S. Multi-scale orderless pooling of deep convolutional activation features. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part VII 13. Springer: Cham, Switzerland; pp. 392–407. [Google Scholar]
  36. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  37. Tian, X.; Wu, D.; Wang, R.; Cao, X. Focal text: An accurate text detection with focal loss. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA; pp. 2984–2988. [Google Scholar]
  38. Nemoto, K.; Hamaguchi, R.; Imaizumi, T.; Hikosaka, S. Classification of rare building change using cnn with multi-class focal loss. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; IEEE: Piscataway, NJ, USA; pp. 4663–4666. [Google Scholar]
  39. Cheng, Z.; Chai, S. A cyber intrusion detection method based on focal loss neural network. In Proceedings of the 2020 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; IEEE: Piscataway, NJ, USA; pp. 7379–7383. [Google Scholar]
  40. Leevy, J.L.; Khoshgoftaar, T.M. A survey and analysis of intrusion detection models based on CSE-CIC-IDS2018 Big Data. J. Big Data 2020, 7, 104. [Google Scholar] [CrossRef]
  41. Sahu, S.K.; Mohapatra, D.P.; Rout, J.K.; Sahoo, K.S.; Pham, Q.-V.; Dao, N.-N. A LSTM-FCNN based multi-class intrusion detection using scalable framework. Comput. Electr. Eng. 2022, 99, 107720. [Google Scholar] [CrossRef]
  42. Singh, G.; Khare, N. A survey of intrusion detection from the perspective of intrusion datasets and machine learning techniques. Int. J. Comput. Appl. 2022, 44, 659–669. [Google Scholar] [CrossRef]
  43. Zhang, Y.; Liu, Q. On IoT intrusion detection based on data augmentation for enhancing learning on unbalanced samples. Future Gener. Comput. Syst. 2022, 133, 213–227. [Google Scholar] [CrossRef]
  44. Liu, Q.; Wang, D.; Jia, Y.; Luo, S.; Wang, C. A multi-task based deep learning approach for intrusion detection. Knowl. Based Syst. 2022, 238, 107852. [Google Scholar] [CrossRef]
  45. Abedzadeh, N.; Jacobs, M. A Survey in Techniques for Imbalanced Intrusion Detection System Datasets. Int. J. Comput. Syst. Eng. 2023, 17, 9–18. [Google Scholar]
  46. Arık, A.O. A robust Gradient Boosting Model Based on SMOTE and NEAR MISS Methods for Intrusion Detection in Imbalanced Data Sets. Ph.D. Thesis, Işık Üniversitesi, Şile, Türkiye, 2022. [Google Scholar]
  47. Kaissar, A.; Nassif, A.B.; Injadat, M. A Survey on Network Intrusion Detection using Convolutional Neural Network. ITM Web Conf. 2022, 43, 01003. [Google Scholar] [CrossRef]
  48. Vu, L.; Nguyen, Q.U.; Nguyen, D.N.; Hoang, D.T.; Dutkiewicz, E. Deep generative learning models for cloud intrusion detection systems. IEEE Trans. Cybern. 2022, 53, 565–577. [Google Scholar] [CrossRef] [PubMed]
  49. Roy, S.; Li, J.; Choi, B.-J.; Bai, Y. A lightweight supervised intrusion detection mechanism for IoT networks. Future Gener. Comput. Syst. 2022, 127, 276–285. [Google Scholar] [CrossRef]
  50. Dat-Thinh, N.; Xuan-Ninh, H.; Kim-Hung, L. MidSiot: A multistage intrusion detection system for internet of things. Wirel. Commun. Mob. Comput. 2022, 2022, 9173291. [Google Scholar] [CrossRef]
  51. Mokbal, F.M.M.; Wang, D.; Osman, M.; Yang, P.; Alsamhi, S.H. An efficient intrusion detection framework based on embedding feature selection and ensemble learning technique. Int. Arab J. Inf. Technol. 2022, 19, 237–248. [Google Scholar] [CrossRef]
  52. Yilmaz, I.; Masum, R.; Siraj, A. Addressing imbalanced data problem with generative adversarial network for intrusion detection. In Proceedings of the 2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science (IRI), Las Vegas, NV, USA, 11–13 August 2020; IEEE: Piscataway, NJ, USA; pp. 25–30. [Google Scholar]
  53. Panigrahi, R.; Borah, S.; Bhoi, A.K.; Ijaz, M.F.; Pramanik, M.; Kumar, Y.; Jhaveri, R.H. A Consolidated Decision Tree-Based Intrusion Detection System for Binary and Multiclass Imbalanced Datasets. Mathematics 2021, 9, 751. [Google Scholar] [CrossRef]
  54. Huang, S.; Lei, K. IGAN-IDS: An imbalanced generative adversarial network towards intrusion detection system in ad-hoc networks. Ad Hoc Networks 2020, 105, 102177. [Google Scholar] [CrossRef]
  55. Yan, M.; Chen, Y.; Hu, X.; Cheng, D.; Chen, Y.; Du, J. Intrusion detection based on improved density peak clustering for imbalanced data on sensor-cloud systems. J. Syst. Archit. 2021, 118, 102212. [Google Scholar] [CrossRef]
  56. Al-Essa, M.; Appice, A. Dealing with imbalanced data in multi-class network intrusion detection systems using xgboost. In Proceedings of the Machine Learning and Principles and Practice of Knowledge Discovery in Databases: International Workshops of ECML PKDD 2021, Virtual Event, 13–17 September 2021; Proceedings, Part II. Springer: Berlin/Heidelberg, Germany, 2022; pp. 5–21. [Google Scholar]
  57. Cao, B.; Li, C.; Song, Y.; Qin, Y.; Chen, C. Network Intrusion Detection Model Based on CNN and GRU. Appl. Sci. 2022, 12, 4184. [Google Scholar] [CrossRef]
  58. Balyan, A.K.; Ahuja, S.; Lilhore, U.K.; Sharma, S.K.; Manoharan, P.; Algarni, A.D.; Elmannai, H.; Raahemifar, K. A Hybrid Intrusion Detection Model Using EGA-PSO and Improved Random Forest Method. Sensors 2022, 22, 5986. [Google Scholar] [CrossRef]
  59. Jung, I.; Ji, J.; Cho, C. EmSM: Ensemble Mixed Sampling Method for Classifying Imbalanced Intrusion Detection Data. Electronics 2022, 11, 1346. [Google Scholar] [CrossRef]
  60. Lin, Y.-D.; Liu, Z.-Q.; Hwang, R.-H.; Nguyen, V.-L.; Lin, P.-C.; Lai, Y.-C. Machine learning with variational AutoEncoder for imbalanced datasets in intrusion detection. IEEE Access 2022, 10, 15247–15260. [Google Scholar] [CrossRef]
  61. Cao, B.; Li, C.; Song, Y.; Fan, X. Network intrusion detection technology based on convolutional neural network and BiGRU. Comput. Intell. Neurosci. 2022, 2022, 1942847. [Google Scholar] [CrossRef]
  62. Tareq, I.; Elbagoury, B.M.; El-Regaily, S.; El-Horbaty, E.S.M. Analysis of ToN-IoT, UNW-NB15, and Edge-IIoT Datasets Using DL in Cybersecurity for IoT. Appl. Sci. 2022, 12, 9572. [Google Scholar] [CrossRef]
  63. Thockchom, N.; Singh, M.M.; Nandi, U. A novel ensemble learning-based model for network intrusion detection. Complex Intell. Syst. 2023, 9, 5693–5714. [Google Scholar] [CrossRef]
  64. Sarwar, N.; Bajwa, I.S.; Hussain, M.Z.; Ibrahim, M.; Saleem, K. IoT Network Anomaly Detection in Smart Homes Using Machine Learning. IEEE Access 2023, 11, 119462–119480. [Google Scholar] [CrossRef]
  65. Yuk, C.K.; Seo, C.J. Static Analysis and Machine Learning-based Malware Detection System using PE Header Feature Values. Int. J. Innov. Res. Sci. Stud. 2022, 5, 281–288. [Google Scholar] [CrossRef]
  66. Moustafa, N. A new distributed architecture for evaluating AI-based security systems at the edge: Network TON_IoT datasets. Sustain. Cities Soc. 2021, 72, 102994. [Google Scholar] [CrossRef]
  67. Moustafa, N.; Slay, J. UNSW-NB15: A comprehensive data set for network intrusion detection systems (UNSW-NB15 network data set). In Proceedings of the 2015 Military Communications and Information Systems Conference (MilCIS), Canberra, ACT, Australia, 10–12 November 2015; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar]
  68. Ashraf, J.; Keshk, M.; Moustafa, N.; Abdel-Basset, M.; Khurshid, H.; Bakhshi, A.D.; Mostafa, R.R. IoTBoT-IDS: A novel statistical learning-enabled botnet detection framework for protecting networks of smart cities. Sustain. Cities Soc. 2021, 72, 103041. [Google Scholar] [CrossRef]
  69. Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Toward generating a new intrusion detection dataset and intrusion traffic characterization. In Proceedings of the ICISSp, Funchal, Portugal, 22–24 January 2018; pp. 108–116. [Google Scholar]
  70. Meidan, Y.; Bohadana, M.; Mathov, Y.; Mirsky, Y.; Shabtai, A.; Breitenbacher, D.; Elovici, Y. N-Baiot—Network-Based Detection of Iot Botnet Attacks Using Deep Autoencoders; IEEE Pervasive Computing: Piscataway, NJ, USA, 2018; Volume 17, pp. 12–22. [Google Scholar]
  71. Toldinas, J.; Venčkauskas, A.; Damaševičius, R.; Grigaliūnas, Š.; Morkevičius, N.; Baranauskas, E. A novel approach for network intrusion detection using multistage deep learning image recognition. Electronics 2021, 10, 1854. [Google Scholar] [CrossRef]
  72. Samy, A.; Yu, H.; Zhang, H. Fog-based attack detection framework for internet of things using deep learning. IEEE Access 2020, 8, 74571–74585. [Google Scholar] [CrossRef]
Figure 1. A stack of convolutional layers [32].
Figure 1. A stack of convolutional layers [32].
Sensors 24 06335 g001
Figure 2. Conceptual diagram of the proposed FCTCNN model.
Figure 2. Conceptual diagram of the proposed FCTCNN model.
Sensors 24 06335 g002
Figure 3. Transforming nominal data into numeric data using One-Hot encoding [65].
Figure 3. Transforming nominal data into numeric data using One-Hot encoding [65].
Sensors 24 06335 g003
Figure 4. Learning curve at Level 1 in the proposed architecture in the TON_IIOT dataset.
Figure 4. Learning curve at Level 1 in the proposed architecture in the TON_IIOT dataset.
Sensors 24 06335 g004
Figure 5. Learning curve at Level 1 in the proposed architecture in the UNSWNB-2015 dataset.
Figure 5. Learning curve at Level 1 in the proposed architecture in the UNSWNB-2015 dataset.
Sensors 24 06335 g005
Figure 6. A two-dimensional t-NSE representation of the BoT-IoT and NBaIoT datasets.
Figure 6. A two-dimensional t-NSE representation of the BoT-IoT and NBaIoT datasets.
Sensors 24 06335 g006
Figure 7. ROC plot and confusion matrix from the UNSW-NB2015 dataset. (a) ROC curve. (b) Level 9 confusion matrix: Worms and Shellcode attacks.
Figure 7. ROC plot and confusion matrix from the UNSW-NB2015 dataset. (a) ROC curve. (b) Level 9 confusion matrix: Worms and Shellcode attacks.
Sensors 24 06335 g007
Figure 8. Receiver Operating Characteristic (ROC) curve and confusion matrix for the CICIDS-2017 dataset. (a) ROC curve. (b) Level 5 confusion matrix for the SSH-Patator attack.
Figure 8. Receiver Operating Characteristic (ROC) curve and confusion matrix for the CICIDS-2017 dataset. (a) ROC curve. (b) Level 5 confusion matrix for the SSH-Patator attack.
Sensors 24 06335 g008
Figure 9. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by DenseNet, Inception Time and FCTCNN on the Industrial TON-IIOT dataset [62].
Figure 9. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by DenseNet, Inception Time and FCTCNN on the Industrial TON-IIOT dataset [62].
Sensors 24 06335 g009
Figure 10. ROC Curve and confusion matrix in the TON-IIOT dataset. (a) ROC curve. (b) Level 7 confusion matrix: mitm and scanning attacks.
Figure 10. ROC Curve and confusion matrix in the TON-IIOT dataset. (a) ROC curve. (b) Level 7 confusion matrix: mitm and scanning attacks.
Sensors 24 06335 g010
Figure 11. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by ANN, LSTM, AUTOENCODER and FCTCNN on the Industrial BOT-IOT dataset [64].
Figure 11. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by ANN, LSTM, AUTOENCODER and FCTCNN on the Industrial BOT-IOT dataset [64].
Sensors 24 06335 g011
Figure 12. ROC curves for two attack types and confusion matrix for UDP attacks in the BoT-IoT dataset. (a) ROC curves. (b) Level 2 confusion matrix for UDP attacks.
Figure 12. ROC curves for two attack types and confusion matrix for UDP attacks in the BoT-IoT dataset. (a) ROC curves. (b) Level 2 confusion matrix for UDP attacks.
Sensors 24 06335 g012
Figure 13. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by LSTM, BiLSTM and FCTCNN on the Industrial NBaIoT-2018 dataset [72].
Figure 13. Comparison of accuracy, precision, recall, and F1-score for the proposed method, compared to methods by LSTM, BiLSTM and FCTCNN on the Industrial NBaIoT-2018 dataset [72].
Sensors 24 06335 g013
Figure 14. ROC curves and confusion matrix for two attack types in the NBaIoT dataset. (a) ROC curves. (b) Level 2 confusion matrix for the Mir-Syn attack.
Figure 14. ROC curves and confusion matrix for two attack types in the NBaIoT dataset. (a) ROC curves. (b) Level 2 confusion matrix for the Mir-Syn attack.
Sensors 24 06335 g014
Figure 15. Prediction time comparison of the proposed deep hierarchical model across five datasets: (a) UNSW-NB15, (b) CICIDS2017, (c) N_BaIoT2018, (d) BoT-IoT, and (e) TON-IIOT.
Figure 15. Prediction time comparison of the proposed deep hierarchical model across five datasets: (a) UNSW-NB15, (b) CICIDS2017, (c) N_BaIoT2018, (d) BoT-IoT, and (e) TON-IIOT.
Sensors 24 06335 g015
Table 1. The constituent layers of the proposed architecture.
Table 1. The constituent layers of the proposed architecture.
NameTypeActivationsLearnable Properties
1SEQUENCE INPUT
Sequence input with 1 dimension
SEQUENCE INPUT 1 ( C ) × 1 ( B ) × 1 ( T ) -
2CONV1-D_1
64 × 5 convolutions with stride 1 padding “causal”
1-D CONVOLUTION 64 ( C ) × 1 ( B ) × 1 ( T ) W e i g h t s   5 × 1 × 64
Bias 1 × 64
3ReLU_1ReLU 64 ( C ) × 1 ( B ) × 1 ( T ) -
4LAYERNORM_1LAYER NORMALIZATION 64 ( C ) × 1 ( B ) × 1 ( T ) O f f s e t   64 × 1
S c a l e   64 × 1
5CONV1-D_2
128 × 5 convolutions with stride 1 padding “causal”
1-D CONVOLUTION 128 ( C ) × 1 ( B ) × 1 ( T ) W e i g h t s   5 × 64 × 128
Bias 1 × 28
6ReLU_2ReLU 128 ( C ) × 1 ( B ) × 1 ( T ) -
7LAYERNORM_2LAYER NORMALIZATION 128 ( C ) × 1 ( B ) × 1 ( T ) O f f s e t   128 × 1
S c a l e   128 × 1
8GLOBALAVGPOOL1D
1-D Global average Pooling
1-D GLOBAL AVERAGE POOLING 128 ( C ) × 1 ( B ) -
9FC
2 fully connected layer
FULLY CONNECTED 2 ( C ) × 1 ( B ) W e i g h t s   2 × 128
Bias 2 × 1
10SOFTMAXSOFTMAX 2 ( C ) × 1 ( B ) -
11FOCALLOSSFOCAL LOSS LAYER 2 ( C ) × 1 ( B ) -
Table 2. Sorting of attacks in descending order in 5 datasets.
Table 2. Sorting of attacks in descending order in 5 datasets.
NBaIoT-2018ToN_IIoTBoT-IoT (UNSW-2018)CICIDS-2017UNSW-NB2015
CountAttackCountAttackCountAttackCountAttackCountAttack
122,573
105,874
102,195
92,141
81,982
59,718
29,849
29,068
‘mir-syn’
‘gaf-udp’
‘mir-ack’
‘gaf-tcp’
‘gaf-udpplain’
‘gaf-combo’
‘mir-scan’
‘gaf-junk’
4608
3628
1269
612
525
447
15
‘DDos’
‘Password’
‘XSS’
‘injection’
‘Dos’
‘Scanning’
‘Mitm’
1,981,230
1,593,180
73,168
17,914
2474
73
6
‘UDP’
‘TCP’
‘Service_Scan’
‘OS_Fingerprint’
‘HTTP’
‘Keylogging’
‘Data_Exfiltration’
158,930
128,027
7938
5897
1507
21
‘PortScan’
‘DDoS’
‘FTP-Patator’
‘SSH-Patator’
‘Brute Force’
‘Sql Injection’
7522
5409
5051
1759
1167
534
526
223
24
‘Generic’
‘Exploits’
‘Fuzzers’
‘Reconnaissance’
‘DoS’
‘Backdoors’
‘Analysis’
‘Shellcode’
‘Worms’
Table 3. Performance comparison of deep models on the UNSW-NB2015 dataset.
Table 3. Performance comparison of deep models on the UNSW-NB2015 dataset.
Attack NameModelPrecisionRecallF1-ScoreMCCKappaSpecificityAUC
NormalKNN-GAN (2022) [2]-94.3994.03----
ResNet50 (2021) [71]88.775.0481.3083.1083.7390.2197.21
ResNet50-Oversample (2021) [71]92.364.67672.5471.9293.1096.44
LSTM (2020) [72]99.699.599.5493.393.294.210.9
FCTCNN99.8599.7699.8193.9493.9395.330.99
BackdoorKNN-GAN (2022) [2]-49.1753.23----
ResNet50 (2021) [71]2.70.20.370000
ResNet50-Oversample (2021) [71]6.339.310.850000
LSTM (2020) [72]0000000
FCTCNN43.2827.1033.331715.973.040.66
AnalysisKNN-GAN (2022) [2]-71.4083.17----
ResNet50 (2021) [71]0000000
ResNet50-Oversample (2021) [71]2.383.570000
LSTM (2020) [72]0000000
FCTCNN1001001001001001001
FuzzersKNN-GAN (2022) [2]-71.7377.32----
ResNet50 (2021) [71]24.130.426.887.17.933.330.491
ResNet50-Oversample (2021) [71]22.841.529.437.37.934.210.499
LSTM (2020) [72]39.663.348.7210.111.2169.210.51
FCTCNN96.9579.2787.2276.3274.8196.9895.49
ShellcodeKNN-GAN (2022) [2]-81.6441.55----
ResNet50 (2021) [71]32.78.713.740000
ResNet50-Oversample (2021) [71]6.276.211.460000
LSTM (2020) [72]0000000
FCTCNN1001001001001001001
Reconnaissance (Probe)KNN-GAN (2022) [2]-89.7385.53----
ResNet50 (2021) [71]62.970.566.4844.6344.6187.800.736
ResNet50-Oversample (2021) [71]65.660.763.0545.5445.188.210.745
LSTM (2020) [72]64.941.650.7044.6343.4287.530.715
FCTCNN98.4897.4297.9595.1095.0997.870.9936
ExploitsKNN-GAN (2022) [2]-76.9274.41----
ResNet50 (2021) [71]76.386.581.0867.5966.8890.990.9201
ResNet50-Oversample (2021) [71]72.244.354.9061.0160.3989.880.9
LSTM (2020) [72]5071.758.9149.5549.0487.330.83
FCTCNN84.9173.4478.7667.3166.8791.760.9254
DosKNN-GAN (2022) [2]-56.7663.18----
ResNet50 (2021) [71]100000000
ResNet50-Oversample (2021) [71]32.342.936.850053.420.5
LSTM (2020) [72]0000000
FCTCNN79.5953.6664.1044.6342.4087.900.7965
WormsKNN-GAN (2022) [2]-38.4649.26----
ResNet50 (2021) [71]0000000
ResNet50-Oversample (2021) [71]2.484.14.660000
LSTM (2020) [72]0000000
FCTCNN1001001001001001001
GenericKNN-GAN (2022) [2]-99.6498.65----
ResNet50 (2021) [71]989596.4793.9393.7199.10.9892
ResNet50-Oversample (2021) [71]99.393.196.1093.9193.6999.030.9891
LSTM (2020) [72]97.592.394.8292.9292.1098.880.9801
FCTCNN98.7693.6196.1293.9493.9399.400.9907
Table 4. Performance comparison of deep models on the CIC-IDS2017 dataset.
Table 4. Performance comparison of deep models on the CIC-IDS2017 dataset.
Attack NameModelPrecisionRecallF1-ScoreMCCKappaSpecificityAUC
NormalLightweight-ensemble 2023 [63]99.5799.8599.7199.4899.4899.700.9991
FCTCNN99.6199.8099.7199.5099.5099.730.9999
DDosLightweight-ensemble 2023 [63]78.9248.4760.0567.5567.5495.410.981
FCTCNN99.9910099.9999.9899.9899.991
FTP-PatatorLightweight-ensemble 2023 [63]0000000
FCTCNN99.9999.8199.8799.7499.7499.990.999
Port-ScanLightweight-ensemble 2023 [63]99.5999.5299.5699.5699.5699.510.99
FCTCNN99.9999.9999.999.9999.9999.990.999
SSH-PatatorLightweight-ensemble 2023 [63]0000000
FCTCNN10099.9399.9699.9399.931000.9997
Brute-ForceLightweight-ensemble 2023 [63]94.7494.3694.5549.2949.2931.230.9361
FCTCNN99.3210099.6557.5349.7433.330.9474
SQL-InjectionLightweight-ensemble 2023 [63]99.8199.8199.8157.5751.6945.310.9585
FCTCNN10099.3299.6557.5349.7433.330.9474
Table 5. Results of the proposed model in 8 types of attacks in the TON-IIOT dataset, Windows 10 section.
Table 5. Results of the proposed model in 8 types of attacks in the TON-IIOT dataset, Windows 10 section.
Attack NamePrecisionRecallF1-scoreMCCKappaSpecificityAUC
NORMAL99.9999.8699.9199.8099.8099.940.9999
DDOS99.92099.9299.9299.8199.8199.891
Password99.8399.8399.8299.6899.6899.860.9999
XSS99.6610099.8399.6499.6499.621
Injection98.9698.4598.7196.4096.4098.140.9998
DOS90.698.7194.4790.7490.5093.100.9940
MITM 1001001001001001001
Scanning1001001001001001001
Table 6. Results of the proposed model in 8 types of attacks in the BoT-IoT dataset.
Table 6. Results of the proposed model in 8 types of attacks in the BoT-IoT dataset.
Attack NamePrecisionRecallF1-ScoreMCCKappaSpecificityAUC
Normal1001001001001001000.9996
UDP99.9999.9999.9999.9999.9999.990.9999
TCP99.9599.6599.8099.7999.7999.990.9999
Service Scan91.9765.0576.2072.5370.9298.400.9486
OS Fingerprint99.7999.699.6999.6599.6599.970.9954
HTTP99.5899.7999.6892.4092.3790.470.9971
Data Exfiltration50505044.7344.73500.52
Keylogging94.7394.7394.7390.7490.7494.730.978
Table 7. Results of the proposed model in 9 types of attacks in the NBaIoT dataset.
Table 7. Results of the proposed model in 9 types of attacks in the NBaIoT dataset.
Attack NamePrecisionRecallF1-ScoreMCCKappaSpecificityAUC
Normal99.9799.9899.9899.7699.7699.690.9998
Mir-Syn99.9910099.9999.9999.9999.990.99999
Gaf-Udp99.9676.6686.7764.0958.2799.900.8827
Mir-Ack999.9999.9999.9999.9899.9899.980.9999
Gaf-Tcp99.9699.9999.9799.9399.9399.910.9999
Gaf-Udpplain1001001001001001000.9999
Gaf-Combo99.9997.6098.7397.5397.5099.900.9992
Gaf-Junk99.9499.9399.9499.8899.8899.950.9999
Mir-Scan99.9499.9399.9499.8899.8899.950.9999
Table 8. Comparison of average time (s) across five datasets.
Table 8. Comparison of average time (s) across five datasets.
Model/DatasetUnsw-NB2015CIC-IDS2017N_BaIOT_2018BoT-IOTTON_IIOT
FCTCNN0.1621.1132.4042.2660.124
LSTM (2023) [64]1.91.422.632.930.313
ANN (2023) [64]0.1881.232.392.610.298
LSTM (2020) [72]0.1981.482.712.910.327
BiLSTM (2020) [72]0.2011.532.772.810.351
Lightweight-ensemble 2023 [63]0.281.412.382.260.127
KNN-GAN (2022) [2]1.332.422.983.120.571
ResNet50 (2021) [71]0.3210.1212.482.310.233
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Miryahyaei, M.; Fartash, M.; Akbari Torkestani, J. Focal Causal Temporal Convolutional Neural Networks: Advancing IIoT Security with Efficient Detection of Rare Cyber-Attacks. Sensors 2024, 24, 6335. https://doi.org/10.3390/s24196335

AMA Style

Miryahyaei M, Fartash M, Akbari Torkestani J. Focal Causal Temporal Convolutional Neural Networks: Advancing IIoT Security with Efficient Detection of Rare Cyber-Attacks. Sensors. 2024; 24(19):6335. https://doi.org/10.3390/s24196335

Chicago/Turabian Style

Miryahyaei, Meysam, Mehdi Fartash, and Javad Akbari Torkestani. 2024. "Focal Causal Temporal Convolutional Neural Networks: Advancing IIoT Security with Efficient Detection of Rare Cyber-Attacks" Sensors 24, no. 19: 6335. https://doi.org/10.3390/s24196335

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop