Next Article in Journal
A Rapid Beam Pointing Determination and Beam-Pointing Error Analysis Method for a Geostationary Orbiting Microwave Radiometer Antenna in Consideration of Antenna Thermal Distortions
Previous Article in Journal
A Compact Raster Lensless Microscope Based on a Microdisplay
Previous Article in Special Issue
Using Embedded Feature Selection and CNN for Classification on CCD-INID-V1—A New IoT Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Malicious Network Behavior Detection Using Fusion of Packet Captures Files and Business Feature Data

1
School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
School of Computer Science, Beijing University of Posts and Telecommunications, Beijing 100876, China
3
Department of Statistics, Colorado State University, Fort Collins, CO 80523, USA
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(17), 5942; https://doi.org/10.3390/s21175942
Submission received: 20 July 2021 / Revised: 24 August 2021 / Accepted: 29 August 2021 / Published: 3 September 2021
(This article belongs to the Special Issue Sensor Networks Security and Applications)

Abstract

:
Information and communication technologies have essential impacts on people’s life. The real time convenience of the internet greatly facilitates the information transmission and knowledge exchange of users. However, network intruders utilize some communication holes to complete malicious attacks. Some traditional machine learning (ML) methods based on business features and deep learning (DL) methods extracting features automatically are used to identify these malicious behaviors. However, these approaches tend to use only one type of data source, which can result in the loss of some features that can not be mined in the data. In order to address this problem and to improve the precision of malicious behavior detection, this paper proposed a one-dimensional (1D) convolution-based fusion model of packet capture files and business feature data for malicious network behavior detection. Fusion models improve the malicious behavior detection results compared with single ones in some available network traffic and Internet of things (IOT) datasets. The experiments also indicate that early data fusion, feature fusion and decision fusion are all effective in the model. Moreover, this paper also discusses the adaptability of one-dimensional convolution and two-dimensional (2D) convolution to network traffic data.

1. Introduction

Cyber security plays an indispensable role in people’s life. Malicious network behaviors will affect account security, software security, payment security, communication security and so on in daily life [1]. However, with the continuous development of application technology, cyber security threats are becoming increasingly complex. Thus, it is more laborious to identify malicious behaviors on the Internet [2]. Some application systems and frameworks for enhancing information security have also been covered by many researchers [3,4,5]. Some researchers have tried discovering more features in each network application behavior and tried identifying the malicious ones. However, the feature extraction of a single method is always very limited. By assuming that there are different types of data samples which can represent network behaviors in different forms, we can collect dissimilar information from each one. If this information can be combined to build detection models, the identification results may achieve better performance. In this paper, fusion models of packet captures files and business feature data are used to achieve better malicious network behavior detection results.
There are two common approaches for malicious network behavior detection: traditional business methods with pattern setting and ML-based ones [6]. Traditional detection methods based on fixed service characteristics are relatively simple, which always include Deep Packet Inspection (DPI)-based, port-based and statistic-based ones [7,8,9]. However, due to its single mode, the detectable malicious behavior is relatively fixed, and they are less and less used in today’s applications. As the ML-based technology continues to mature in daily application fields such as computer vision (CV), natural language processing (NLP) and intelligent recommendation, malicious behavior detection methods based on ML and DL are used more frequently. Generally, traditional ML-based methods using business features and DL-based methods utilizing raw data collected from traffic include two major methods. As for traditional ML-based methods, which usually consider the features concluded by business experts, they lack generalization ability that makes it difficult to detect some malicious attacks with huge fluctuations. As for DL-based methods, they always extract features automatically. Therefore, some attacks with a set business pattern may not be identified. Moreover, these simple models always address one type of data source for a network behavior sample. If there are over two types of data, some effective features will not be mined while only one type of data is used.
Fusion-based models are widely used in Human Action Recognition (HAR) and achieved good performance [10,11,12]. It frequently happens that HAR uses pictures with human’s action to complete the recognition process. RGB and depth images are two common types of original data for this application. In addition to these vision data, some wearable inertial data outputs from sensors also play an important role in HAR. For each type of data source, it contains relatively unitary information in a certain aspect. Then, fusion models based on DL can solve this problem. Two or more types of data can be fused through early data fusion, feature fusion and decision fusion. Furthermore, researchers show that fusion methods are really effective in HAR. In malicious behavior detection tasks, a few approaches use fusion frameworks to complete this process. Compared with HAR data, the recording method of network access is relatively single. Generally speaking, network traffic and some business feature-based data are two common data sources [13]. This paper conducted these two kinds of data in order to accomplish fusion detection experiments and made some progress. The key contributions of the presented work are the following:
(1)
We proposed a Convolutional Neural Network(CNN)-based fusion model of packet capture (PCAP) files and business feature data for malicious network behavior detection, and experiments results show that the fusion model can improve the detection precision with little increase in time and resource consumption.
(2)
Three fusion approaches, early data fusion, feature fusion and decision fusion, are discussed here, and we find that each method can extract more discriminating and complementary features. On the whole, feature fusion and decision fusion have better performances in the datasets.
(3)
Different from image or video data, traffic network and business feature data lack two-dimensional association information. In this paper, we compared the effect of one-dimensional and two-dimensional convolution on network malicious behavior detection scene. The results show that one-dimensional convolution structures have a better performance.
(4)
We evaluated the proposed fusion methods on different datasets and compared the effects with previous studies. Experimentally, the results illustrated that our methods can adapt to various network data in lots of different scenarios.
The rest of the paper is organized as follows. Section 2 introduces some methods and applications related to this paper. Section 3 depicts the proposed methods and explains some significant concepts. In Section 4, we illustrate the experiment process and results. In addition, we compare the outputs and analyze the results accordingly. Section 5 provides the conclusions of this paper.

2. Related Work

Since few methods use fusion frameworks in network behavior detection, for the purposes of better explaining and understanding, we discuss some fusion models that perform well in other application fields. Furthermore, this section depicts some ML-based behavior detection approaches using PCAP files and business feature data.
Traditional ML-based methods account for a large proportion of current malicious network behavior detection applications. These methods focus on the processed business feature data and make some progress, the results of which are easier trusted by experts. Waskle et al. [14] proposed an approach to develop an efficient intrusion detection system (IDS) by using principal component analysis (PCA) and random forest (RF) algorithm, and the results performed better than the basic traditional ML models such as Naive Bayes and decision tree (DT). Lu et al. [15] combined Synthetic Minority Oversampling Technique (SMOTE) and Edited Nearest Neighbor (ENN) Rule on RF and achieved higher precision, recall and F1-score. Gao et al. [16] proposed an adaptive ensemble-learning model based on DT, RF and K-Nearest Neighbor (KNN). Marteau et al. [17] proposed an ensemble approach composed of random partitioning binary trees named DiFF-RF for malicious behavior detection. Miah et al. [18] proposed a multi-layer classification approach for imbalanced network intrusion based on cluster and RF. Table 1 lists some related literature in the past three years that contain ML-based and DL-based methods in the past three years.
Compared with traditional ML-based network network detection methods, DL-based methods rely less on business experience. DL-based models have a perfect performance on CV, NLP, automatic control(AC) and many other related application fields. In light of their research methodologies, CNN is the most popular model used here. The second half of the Table 1 introduces these methods and their contributions. Some of them show solicitude for data augmentation [22,33], and some focus on the problem of imbalanced data [22,26,27,34]. Moreover, model interpretability [23,29] and feature generation [28,32] are also the focus of attention. Although each method solves the problems in different scenarios, they have consistent goals in detection results. All of them [15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,35] pay close attention to the improvement of detection accuracy and efficiency. In this paper, we also focus on the most basic and important target in intrusion detection system (IDS). The experiment results show that the proposed fusion frameworks make some significant improvements on different datasets.
It is the fusion models of other application fields that inspire us. This paragraph will introduce some related fusion models. Chen et al. [36] made a review of the publicly available datasets on HAR and compared their results. The results show that each sensor modality has its own limitations, and fusion methods can reduce the impacts. After that, some researchers invested their efforts to further verify the effectiveness of fusion methods on HAR application. Dawar and Kehtarnavaz [37] achieved the recognition of the detected actions of interest based on the fusion of depth images and inertial signals, which shows the effectiveness in dealing with realistic continuous action streams. Dawar et al. [38] used the decision fusion method of CNN for depth images and LSTM for inertial signals, and the results indicate the positive impact of decision-level fusion and the data augmentation on recognition accuracies. Through a comprehensive understanding of these methods, we attempt to concentrate on fusion methods for network detection in this paper. PCAP files and business feature data can support us in accomplishing this task.

3. Proposed Methods

In this section, we firstly describe the general process of malicious behavior detection based on the proposed fusion methods. Then, we depict the structures of three fusion models, namely early data fusion model, feature fusion model and decision fusion model. Finally, we introduce the comparation of one-dimensional and two-dimensional convolution structure for malicious behavior detection in our model.

3.1. Framework of Fusion Methods for Malicious Behavior Detection Process

In simple terms, the proposed model aimed to collect more significant information from different types of input data. As shown in Figure 1, the input data are PCAP files and business feature data. PCAP files consist of a series of hexadecimal numbers representing different meanings, which is a little bit difficult for people to understand directly. We used CNN-based model to extract effective features automatically as the basic model before the fusion process. Business feature data comes from experts of network security field and contains a series of business feature data, which is depicted in detail in Section 4.1.2. From the method of generation, they come from machine understanding and human understanding, respectively. The availability of different source approaches make us attempt to fuse both of them and to obtain more effective attributes. Then, when data are ready, we can choose one fusion model and complete the training process. In the fusion process, we utilize one-dimensional and two-dimensional convolution structures to compare their effects on network traffic data. Finally, the classifier will identify the category based on the fusion model results.

3.2. Fusion Models

3.2.1. Early Data Fusion Model

At first, we introduce the directest fusion model, namely early data fusion. Figure 2 depicts the structure of an early fusion model. Why call it “early” fusion process? It happens at the earliest stage of the whole process. After data processing, the PCAP and feature data can fit together as the input data of detection module. We define the PCAP raw data as X 1 R 1 × W 1 and feature data as X 2 R 1 × W 2 , where W denotes the width of vectors. When early data fusion is completed, we can obtain the input data X .
X = [ X 1 , X 2 ] .
As Equation (1) shows, data fusion is a simple and direct process for expanding information of input data. It can be considered as a data augmentation approach for each type of raw data. For model training, there are more contents that can be mined. Experimentally, we design two convolutional layers in the detection model in Figure 2. The current simple CNN-based module can achieve good performance on our datasets; thus, we did not add other more complex structures. Additionally, it is more persuasive to verify the effects of proposed malicious behavior detection fusion methods. Then, the output of the data fusion model Y can be defined as follows:
Y = F c l a s s i f y ( F p o o l i n g × 1 × 3 ( F c o n v × 1 × 25 ( F p o o l i n g × 1 × 3 ( F c o n v × 1 × 25 ( X ) ) ) ) ) ,
where F denotes some manipulation functions; F c o n v × 1 × 25 refers to the convolutional operation with a 25 size one-dimensional convolution kernel; F p o o l i n g × 1 × 3 refers to a pooling operation with 3 size pooling kernel; and F c l a s s i f y conveys a fully connected (FC) operation and an activation function, and then it outputs the classification results. Early data fusion focuses on the basic data augmentation through data connection. It helps the detection model in increasing the amount of input information at the beginning.

3.2.2. Feature Fusion Model

Feature fusion is a deep feature map fusion process. Unlike early data fusion, it is necessary to train two classifiers that have different input data in order to obtain deep feature maps. The structure is described in Figure 3. Similarly, X 1 and X 2 also denote two kinds of input data. Then, both of model M 1 and M 2 are trained, respectively. When models are ready, we define the deep future maps N 1 R H 1 × W 1 and N 1 R H 2 × W 2 from M 1 and M 2 as follows.
N 1 = F p o o l i n g × 1 × 3 ( F c o n v × 1 × 25 ( F p o o l i n g × 1 × 3 ( F c o n v × 1 × 25 ( X 1 ) ) ) ) ,
N 2 = F p o o l i n g × 1 × 2 ( F c o n v × 1 × 2 ( F p o o l i n g × 1 × 2 ( F c o n v × 1 × 2 ( X 2 ) ) ) ) .
Then, the fusion deep feature matrix can be depicted as N f usion .
N fusion = [ N 1 , N 2 ] .
As Equations (3)–(5) depict, feature fusion is completed before the classify layer. N fusion will cross a fully connected layer, and the FC layer will output the final classified results Y .
Y = F c l a s s i f y ( N fusion ) .
Feature fusion occurs at the model training process. It acquires trained information output from two models. It is fused after the completion of their own feature construction of two types of data. Compared with early data fusion, the interference of the two kinds of training data is lesser in the feature fusion process. When the fusion process happens, the model obtains not only the two kinds of data information but also identification abilities of two models.

3.2.3. Decision Fusion Model

Decision fusion occurs at the end of the classification process. Similarly to the feature fusion process, decision fusion contains two training models for PCAP data and business feature data. Figure 4 explains the structure of the decision fusion model. It is built to provide two models with different weights in the decision-making process. Considering that two kinds of input data may have different impacts on the final detection results through the training of two models, we introduced the weight S = [ S 1 1 , S 1 2 , . . . S i j ] of model detection on each category for decision fusion. i { 1 , 2 } refers to two submodels, and j { 1 , 2 , . . . , j } refers to different categories in detection targets. S i j outputs from the last FC layers of two models and represents the detection probability of category j of model i. As Figure 4 shows, decision fusion selects the highest decision score D j as the final precision that is calculated by weight addition. The decision score matrix D can be calculated as follows:
D = P 1 × S 1 + P 2 × S 2 ,
where P 1 and P 2 denote accuracy of M 1 and M 2 in training process. They represent the detection performance of each model.
The final detection results Y = D . We can obtain the final detection output Y.
Y = M a x ( D ) .
Unlike early data fusion and feature fusion, decision fusion artificially deals with the different influence of the two single models’ detection ability instead of automatic model mining.

3.3. Structures of One-Dimensional Convolution and Two-Dimensional Convolution

Due to the popularity of computer vision, two-dimensional convolution has the widest application. It is also widely used in other related fields because of its great achievements. In this work, we attempted to use 2D convolution with 2D filter in order to excavate deep features at first. We expect to find more local features of the data. However, it has been discovered that the 2D feature is not obvious in the network traffic data. Figure 5 shows the performance of 1D convolution and 2D convolution acting on network traffic, respectively.
According to our understanding of network traffic data, our experimental data consist of a series of hexadecimal numbers, which is expounded in Section 4.1. From Figure 5c,d, it can be concluded that 1D filters only consider the relationship between the front and back bits of hexadecimal data of each sample, but 2D filters take the feature aggregation of a square region into account. In terms of professional experience, there are no obvious interactions between the areas of traffic data and the business feature data composed of different kinds of features. Theoretically, 1D convolution will have a better effect on malicious behavior detection task with PCAP files and business feature data. Of course, the experiment results also prove this idea in Section 4.5.

4. Experiments and Results

4.1. Data Description

In this paper, we used four datasets to verify the proposed detection fusion methods. They are NB15 [39], CIC2012 [40], CIC2017 [41] and VPN2016 [42]. There are two kinds of original data in each dataset, namely PCAP files and business feature data. PCAP files are the basic form to store network communication behaviors and contain almost all the original information of a network communication. Business feature data files are collected by business feature collection system for which its features are designed by experts. From the perspective of bearing information, PCAP files utilize more original information in the process of network connection. However, more information content also represents that it is more likely to contain useless messages and have a bad influence on model training. Business feature data pay more attention to the most effective data characteristics that are more precise, explicit and interpretable. In terms of content, they can complement each other. Then, we will introduce the structure of two kinds of data and depict each dataset in detail.

4.1.1. PCAP Files

PCAP files can be translated into a group of hexadecimal numbers. Data segments at specific locations represent different meanings that may explain connection source, connection destination, data length and other connection information. Figure 6 depicts the structure of a PCAP file. Generally, the length of PCAP header of a PCAP file is 24-byte, which contains byte order information, file version number, timestamp accuracy and so on. The content of the packet header is relatively fixed. It always describes the timestamp with an 8-byte, the length of data frame with an 8-byte, namely Caplen, and the length of offline data with an 8-byte. Then, actual transmission data follow the length Caplen defined by packet header.

4.1.2. Business Feature Data

It has been reported that business feature data contain many useful business features. Table 2 describes some types of feature. Flow features record the overall situation of a network connection or session, which mainly includes some of five-tuple information. Base features contain some basic information in people’s cognition, including the duration of a network connection, the number of dropped packets and etc. Content features display some information about transmitting data. Time features portray the important content related to time. Generated features are calculated to achieve a specific purpose, for example, some of them are sorted accordingly with the last time feature to capture similar characteristics of the connection records for each 100 connections that are sequentially ordered.

4.1.3. Datasets Overview

NB15 [39] is generated by the Australian Centre for Cyber Security (ACCS) with PerfectStorm. It contains nine types of network attacks and one type of normal network behavior data. Each category contains a different number of samples from 100 to 20,000, which constitute an unbalanced dataset. In addition, the ratio of training and test data amount is 7:3.
VPN2016 [42] is exposed by University of New Brunswick (UNB). They captured a session over VPN, and the categories are described as follows. TraP2P labels identify file-sharing protocols such as Bittorrent. Chat labels identify instant-messaging applications. VoIP labels represent the Voice over IP label that groups all traffic generated by voice applications. Streaming labels identify multimedia applications that require a continuous and steady stream of data. Email labels are generated using a Thunderbird client and Alice and Bob Gmail accounts. File-transfer labels identify traffic applications for which its main purpose is to send or receive files and documents.
CIC2012 [40] and CIC2017 [41] simulated different types of cyber attacks which are generated by UNB. CIC2017 contains fourteen categories of attack network traffic and one category of normal traffic. CIC2012 contains five attack categories in CIC2017. We attempt to combine CIC2012 with CIC2017 in order to make up a new dataset. Considering that the experiment results have already achieved a perfect performance, we plan to make a new dataset to test the identification ability of fusion models on the data with a certain distribution difference. Finally, this new dataset comprises 15 kinds of data type, including a normal type and other attack categories. There are about 10,000 samples in each category, and the category that numbers the least comprises about 2000 samples. Tt can be concluded from Section 4.3.1 that fusion models also make a good performance on the new one.

4.2. Data Processing

In this paper, we need to deal with two kinds of raw data, PCAP files and business feature data. From the perspective of processing steps, it is more complicated to handle PCAP files than business feature data. As for business feature data, we need to convert characters to numerical variables and normalize the numerical value for each feature. Then, the processed data can feed into the model. We used Figure 7 to explain the process of data processing.
As Figure 7 shows, four steps before the data enters the detection model are data split, data clean, data transfer and data trimming. Firstly, we split PCAP files into flows based on five-tuple, including a source IP address, a source IP port, a destination IP address, a destination IP port and the protocol in use. Then, we filter out some invalid information, namely repetitive flows, empty and interferential flows by data cleaning process. Another step is data transfer that extracts hexadecimal code from PCAP files. Finally, we trim the data to an appropriate size for one dimension CNN or two dimension CNN inputs.
Due to the different length of data, it is impossible to send all data into the model for training. In the experiment, we work out that the average number of packets per flow divided by five-tuple is about two in datasets. Thus, we chose the average length of 784 bytes as the processing length of each stream. If it is longer than 784, it will be intercepted. If it is shorter than 784, it will be filled with zero. The header, transmission control protocol information and payload are concerned here except for IP information. IP information is only used as data split, which ensures that IP information that may have label properties will not be brought into the training data. When processing the PCAP file, we extracted the data of the corresponding location and sent it into the model. In this process, we only care about the data distribution. It does not affect the whole process of data processing whether the data are encrypted or not. As for the division of training set and test data set, we randomly divided the data into a training set and test set according to the ratio of 7:3 on dataset VPN2016 and CIC. Dataset NB15 has been divided into training and test data on the official website and can be used directly. The number of training set and test set of each dataset is depicted in Table 3. After the data processing steps, we can train our models and verify the results.

4.3. Results

In this section, we will compare the effects among the proposed detection fusion models and previous papers’ experiment results. In addition to the proposed fusion model results, we discuss the different detection results between one-dimension-based CNN structure and two-dimension-based one. We calculate the five evaluation indexes to evaluate the effects of the models: Macro-f1, Weighted-f1, Recall, Precision and Accuracy. We use O p r e c , O r e c a l l , O a c c , O m a c r o f 1 and O w e i g h t e d f 1 to denote them. The calculation processed can be described as follows:
O p r e c = 1 N i = 1 N T P T P + F P ,
O r e c a l l = 1 N i = 1 N T P T P + F N ,
O a c c = i = 1 N T P T P + F P ,
O m a c r o f 1 = 1 N i = 1 N 2 × O p r e c × O r e c a l l O p r e c + O r e c a l l ,
O w e i g h t e d f 1 = i = 1 N w i 2 × O p r e c × O r e c a l l O p r e c + O r e c a l l ,
where N is the number of categories in the dataset, TP denotes the number of correctly identified positive samples, TF denotes the number of correctly identified negative samples, FP refers to the number of wrongly identified positive samples, FN refers to the wrongly identified negative samples and w i represents the weight of this category to the total data quantity.
In the experiment, we use the Adam optimizer to optimize the training results. The learning rate is set to 0.001 in order to make our models perform the best, and the batch size is 128 here. We trained about 200 epochs to obtain the best results. Figure 8 shows the changes of F1-score in the training process on VPN2016. The training processes of other datasets are similar to this one. It can be concluded that the improvements of the result become slow after 50 epochs and nearly do not change after 100 epochs.

4.3.1. Overall Performance on Datasets

In this paper, we select four different datasets to compare the performances of each model in order to verify the advancement and rationality. We chose two kinds of baseline results on each dataset. Some recent previous methods training on these datasets are collected, and the simple model without any fusion structure is considered in Table 4, Table 5 and Table 6.
Table 4 shows the experiment results of the proposed models and comparative models on dataset NB15. We compared six previous methods based on DL methods on NB15. It can be summarized from this table that the feature fusion model attains the highest score in each index. Compared with the worse previous models, fusion models can reach over 10% improvement. Furthermore, in contrast to our optimized simple model, fusion models also make some significant improvements. According to the analysis of the results, all fusion methods are effective with a growth rate of 2 to 5%. In addition, we compared some traditional ML-based methods, Logistic Regression (LR), KNN, DT, RF and eXtreme Gradient Boosting (XGBoost). The accuracy of these methods is much lower than the proposed fusion ones.
Table 5 shows the experiment results on dataset VPN2016. The table also shows that fusion models made some improvements. Unlike NB15, the best performance comes from the decision fusion method. Table 6 shows detection results on the mixed dataset with CIC2012 and CIC2017, and the results show that feature fusion has the best performance. It can be concluded that even if the accuracy of a single model is very high, the fusion model can still improve the effect and achieve better results. At the same time, the comparative traditional method results are also depicted in Table 5 and Table 6, which are far less than the results of the fusion ones.
As Tables show, some previous methods have achieved good performance. However, they are not comprehensive enough from the perspective of data understanding. Yang et al. [43] used business features extracted by expert experience to complete the classification tasks. Janarthanan and Zargari [46] proposed a feature selection framework on traffic network. Kim and Chung [49] and Zhou et al. [53] also paid more attention on relatively important features, which may ignore some information that has an impact on the results. This is also the advantage of our proposed fusion models. In addition, some DL-based models [44,45,52] are too simple to achieve good performance. Moreover, some 2D convolution-based CNN [47,48], RNN-based [50] and LSTM-based [51] models may be a little complex for the analysis of network traffic data and can reduce generalization ability. In this paper, the proposed 1D-convolution-based fusion model can balance these problems and achieve ideal results. Different from images, network traffic data and business feature data do not have obvious regional features. Thus, 2D-convolution-based kernels may dig out some features that do not have relevance and have a bad impact on the model. Beyond that, fusion models can combine different information provided by at least two data formats and have stronger recognition ability. We also discuss their advantages in detail in Section 4.4 and Section 4.5.

4.3.2. Details of Each Fusion Method Results

In order to depict the results in some more details, we calculate the evaluation indexes of each category. Figure 9, Figure 10 and Figure 11 reveal the confusion matrices of simple model, early data fusion model, feature fusion model and decision fusion model on dataset NB15, VPN2016, CIC2012 and CIC2017. In Figure 9, there are ten categories with a kind of normal behavior and nine attack behaviors. From the comparison of Figure 9a and the best fusion model in Figure 9c, the accuracy rates of all categories in feature fusion results are improved or flat. The maximum increase in a single category reached 6%. This claimed that the feature fusion model has a positive impact on every subclass. Early data fusion and decision fusion also make sense here, but the effect is weaker than the feature fusion model. Figure 10 portrays six different types results of VPN network behavior. As shown in the figure, two fusion models worked except for the early data fusion one. Moreover, the decision fusion model has the best performance on this data set. Figure 11 displays 15 categories with a benign subclass and other malicious types. Figure 11c describes that the accuracy of all categories made some progress, except for FTP behavior.
For the sake of clarity, we depicted Figure 12, Figure 13 and Figure 14. They compared the effects among simple model and fusion mudels more obviously. They portrayed the results of three datasets: NB15, VPN2016 and the mixed one, CIC. From the figures, we can draw a conclusion that fusion models are better than the simple one on the identification results of each sub category, especially the performance of Feature fusion and Decision fusion models denoted with yellow lines and red lines in the figures.

4.4. Discussion of Fusion Models

From Section 4.3.1 and Section 4.3.2, we introduce the experiment results of the proposed methods. From the performance result point of view, the feature fusion model is the best one that achieves the best result on two datasets, and the decision fusion model achieves the best performance on another dataset. Consequently, we can conclude that, compared with the early fusion method, the late and deep fusion process can obtain a better result.
The early data fusion process occurs at the beginning of the detection process, which combines the raw data directly without any further processing. It can be regarded as adding human business experience to the raw data. The input information increases and provides more dimension features for the training process. From the point of experiments, in the wake of the increasing content of raw data, model has the opportunity to mine more data distribution characteristics. The results show that early data fusion improved on dataset NB15, CIC2012 and CIC2017 but performed worse on dataset VPN2016. Early data fusion is the shallowest fusion process, which means that the composite data only directly stitches the original data of different distributions. It may affect the original good data separability and reduce the cognitive ability of the model. Consequently, it shows better discrimination against some datasets but worse on others. In this case, the feature fusion methods and decision fusion methods are considered here.
The feature fusion process arises after feature generation in the malicious behavior detection model. In this paper, we attempt this process before the last fully connected layer. Before feature-level fusion, both raw datasets go through the same CNN-based network structure. By this means, two different kinds of raw data are more likely to be transformed into a similar feature vector space. Then, stitched data with feature fusion can express the data content preferably. With regard to the decision fusion process, it is not involved in data registration and looks similar to a more independent fusion process. The results show that decision fusion always performs better than data fusion and sometimes performs better than feature fusion. On the whole, later fusion methods are obviously better than the early one, which illustrates that the data trained by models have a more similar data distribution. These later fusion methods can obtain more deeper features and weaken the difference between the original input data.
Above all, fusion models usually play a significant role in promoting detection results. According to different scenes and data, we can choose the most suitable fusion method. We should consider not only accuracy but also efficiency in practical application. Therefore, we also evaluated models’ efficiencies. As for the testing process of malicious behavior detection, each method consumes several milliseconds, which fully meet the efficiency of practical application. In the training process, the original simple model takes 58.86 s to complete a training round with 70,000 pieces of sample. Under the same experimental setting, the training consumptions of data fusion, feature fusion and decision fusion are 62.03 s, 65.55 s and 65.89 s, respectively. In the case of the same resource usage, there is little difference in their computational efficiency. Therefore, we do not have to worry too much about efficiency in the model fusion.

4.5. One-Dimensional and Two-Dimensional Convolution-Based Malicious Behavior Detection Results Comparation

In order to illustrate the rationality of convolution structure and to further verify the difference between 1D convolution and 2D convolution in current malicious behavior detection task, we compared some feature outputs during the training process. Figure 15 shows the visualization of 1D and 2D dimensional convolution feature maps of six different categories in dataset VPN2016. We plan to compare the feature grayscale images output from convolutional layers in the model. Figure 15a describes 1D results, and (b) describes the 2D results. Compared with the images, network traffic has limited association information in each point. Thus, we used the 1D filter and extracted more data characteristics in (a). Figure 15b paid attention to square area features and captured less information that contains fewer pixels here.
We simultaneously tested feature extraction results of the two kinds of convolutions in the training process. Figure 16 depicts the sample distribution converted into a two-dimensional plane on hyperspheres. The points on the circle denote testing samples, and each color refers to a different category. It is obvious that Figure 16a has a smaller intraclass compactness and larger interclass discrepancy after 1D convolution model training, which is consistent with our theoretical inference. Quite evidently, models will be more discriminating in this situation.

5. Conclusions

In this paper, we proposed three different kinds of fusion models based on one-dimensional convolution neural networks for malicious network behavior detection. At the input of each fusion framework, we used packet capture (PCAP) files and business feature data. Then, we extracted distinct and complementary features from the input modalities by applying 1D-CNN-based models and fused these features at more than one stage in our three novel fusion frameworks. The results output from experiments on several datasets proved that the proposed fusion methods are effective on malicious network behavior detection tasks. From the experimental results, feature fusion and decision fusion models have an outstanding performance in terms of detection accuracy, which can improve the results over 5% in the case of almost no increase in computing time. We consider these fusion frameworks to be an important step forward on malicious network behavior detection when there are different kinds of original data for us to utilize.
In the future, we will continue to design some effective methods for malicious network behavior detection and verify their application values in the actual application network environment. Although the proposed models are simple and lightweight, there is still room for improvement to meet real-time requirements. Moreover, we will further study the interpretability of the models in malicious behavior detection tasks, and the visualization of feature maps is a basic work. We will attempt to build a description system of network traffic in order to identify network behavior more intuitively and accurately.

Author Contributions

Conceptualization, M.H. and X.W.; data curation, L.J.; methodology, M.H., K.K. and L.J.; supervision, M.H., X.W. and L.J.; writing—original draft, M.H.; writing—review and editing, M.H., X.W., B.D., X.X. and L.J. All authors have read and agreed to the published version ofthe manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grantno.62071056) and the action plan project of Beijing University of Posts and Telecommunications (No.2020XD-A03-2).

Institutional Review Board Statement

The study did not involve humans or animals.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

The datasets used in this paper are available online [39,40,41,42], and they are also available from the corresponding author upon request.

Acknowledgments

The authors acknowledge ICN and CAD Laboratory of School of Electronic Engineering, Beijing University of Posts and Telecommunications, for the experimental environment. Special thanks are given to Xinlei Wang, Fenfen Li and Yu Zhang for their great help in this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wei, X.; Feng, W.; Wan, S.; Xu, J.; Liu, J.; Lei, Q.; Wang, W. Deep Learning and Distributed Data Storage System in Identity Recognition and Account Security. In Proceedings of the 2020 IEEE 6th International Conference on Computer and Communications (ICCC), Chengdu, China, 11–14 December 2020; pp. 1749–1756. [Google Scholar]
  2. Singh, J.; Singh, J. Detection of malicious software by analyzing the behavioral artifacts using machine learning algorithms. Inf. Softw. Technol. 2020, 121, 106273. [Google Scholar] [CrossRef]
  3. Alrubaiq, A.; Alharbi, T. Developing a Cybersecurity Framework for e-Government Project in the Kingdom of Saudi Arabia. J. Cybersecur. Priv. 2021, 1, 17. [Google Scholar] [CrossRef]
  4. Antunes, M.; Maximiano, M.; Gomes, R.; Pinto, D. Information Security and Cybersecurity Management: A Case Study with SMEs in Portugal. J. Cybersecur. Priv. 2021, 1, 12. [Google Scholar] [CrossRef]
  5. Preuveneers, D.; Joosen, W. Sharing Machine Learning Models as Indicators of Compromise for Cyber Threat Intelligence. J. Cybersecur. Priv. 2021, 1, 8. [Google Scholar] [CrossRef]
  6. Beaugnon, A.; Chifflier, P. Machine Learning for Computer Security Detection Systems: Practical Feedback and Solutions. In Proceedings of the 2018 Intelligence Artificielle et Cybersécurité/Artificial Intelligence and Cybersecurity(C&ESAR), Rennes, France, 19–21 November 2018. [Google Scholar]
  7. Kumar, S.; Dharmapurikar, S.; Yu, F.; Crowley, P.; Turner, J. Algorithms to accelerate multiple regular expressions matching for deep packet inspection. ACM Sigcomm Comput. Commun. Rev. 2006, 36, 339–350. [Google Scholar] [CrossRef] [Green Version]
  8. Finsterbusch, M.; Richter, C.; Rocha, E.; Muller, J.; Hanssgen, K. A Survey of Payload-Based Traffic Classification Approaches. IEEE Commun. Surv. Tutor. 2014, 16, 1135–1156. [Google Scholar] [CrossRef]
  9. Song, W.; Beshley, M.; Przystupa, K.; Beshley, H.; Kochan, O.; Pryslupskyi, A.; Pieniak, D.; Su, J. A software deep packet inspection system for network traffic analysis and anomaly detection. Sensors 2020, 20, 1637. [Google Scholar] [CrossRef] [Green Version]
  10. Wei, H.; Jafari, R.; Kehtarnavaz, N. Fusion of video and inertial sensing for deep learning–based human action recognition. Sensors 2019, 19, 3680. [Google Scholar] [CrossRef] [Green Version]
  11. Wei, H.; Kehtarnavaz, N. Simultaneous utilization of inertial and video sensing for action detection and recognition in continuous action streams. IEEE Sens. J. 2020, 20, 6055–6063. [Google Scholar] [CrossRef]
  12. Wei, H.; Chopada, P.; Kehtarnavaz, N. C-MHAD: Continuous multimodal human action dataset of simultaneous video and inertial sensing. Sensors 2020, 20, 2905. [Google Scholar] [CrossRef]
  13. Drewek-Ossowicka, A.; Pietrołaj, M.; Rumiński, J. A survey of neural networks usage for intrusion detection systems. J. Ambient Intell. Humaniz. Comput. 2020, 12, 1–18. [Google Scholar] [CrossRef]
  14. Waskle, S.; Parashar, L.; Singh, U. Intrusion Detection System Using PCA with Random Forest Approach. In Proceedings of the 2020 International Conference on Electronics and Sustainable Communication Systems (ICESC), Coimbatore, India, 2–4 July 2020; pp. 803–808. [Google Scholar]
  15. Lu, T.; Huang, Y.; Zhao, W.; Zhang, J. The metering automation system based intrusion detection using random forest classifier with smote+ enn. In Proceedings of the 2019 IEEE 7th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 19–20 October 2019; pp. 370–374. [Google Scholar]
  16. Gao, X.; Shan, C.; Hu, C.; Niu, Z.; Liu, Z. An adaptive ensemble machine learning model for intrusion detection. IEEE Access 2019, 7, 82512–82521. [Google Scholar] [CrossRef]
  17. Marteau, P.F. Random Partitioning Forest for Point-Wise and Collective Anomaly Detection—Application to Network Intrusion Detection. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2157–2172. [Google Scholar] [CrossRef]
  18. Miah, M.O.; Khan, S.S.; Shatabda, S.; Farid, D.M. Improving Detection Accuracy for Imbalanced Network Intrusion Classification using Cluster-based Under-sampling with Random Forests. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; pp. 1–5. [Google Scholar]
  19. Khammassi, C.; Krichen, S. A GA-LR wrapper approach for feature selection in network intrusion detection. Comput. Secur. 2017, 70, 255–277. [Google Scholar] [CrossRef]
  20. Halimaa, A.; Sundarakantham, K. Machine learning based intrusion detection system. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; pp. 916–920. [Google Scholar]
  21. Moreira, R.; Rodrigues, L.; Rosa, P.; Silva, F. Improving the network traffic classification using the Packet Vision approach. In Proceedings of the 2019 15th Workshop de Visao Computational (WVC), Sao Paulo, Brazil, 9–11 September 2020; pp. 146–151. [Google Scholar]
  22. Liu, X.; Di, X.; Ding, Q.; Liu, W.; Qi, H.; Li, J.; Yang, H. NADS-RA: Network Anomaly Detection Scheme Based on Feature Representation and Data Augmentation. IEEE Access 2020, 8, 214781–214800. [Google Scholar] [CrossRef]
  23. Lin, Y.; Chang, X. Towards Interpretable Ensemble Learning for Image-based Malware Detection. arXiv 2021, arXiv:2101.04889. [Google Scholar]
  24. Min, E.; Long, J.; Liu, Q.; Cui, J.; Chen, W. TR-IDS: Anomaly-based intrusion detection through text-convolutional neural network and random forest. Secur. Commun. Netw. 2018, 2018, 4943509. [Google Scholar] [CrossRef] [Green Version]
  25. Nguyen, Q.P.; Lim, K.W.; Divakaran, D.M.; Low, K.H.; Chan, M.C. Gee: A gradient-based explainable variational autoencoder for network anomaly detection. In Proceedings of the 2019 IEEE 7th Conference on Communications and Network Security (CNS), Washington, DC, USA, 10–12 June 2019; pp. 91–99. [Google Scholar]
  26. Zheng, W.; Gou, C.; Yan, L.; Mo, S. Learning to Classify: A Flow-Based Relation Network for Encrypted Traffic Classification. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 13–22. [Google Scholar]
  27. He, M.; Wang, X.; Zhou, J.; Xi, Y.; Jin, L.; Wang, X. Deep-Feature-Based Autoencoder Network for Few-Shot Malicious Traffic Detection. Secur. Commun. Netw. 2021, 2021, 6659022. [Google Scholar] [CrossRef]
  28. Marín, G.; Caasas, P.; Capdehourat, G. Deepmal-Deep Learning Models for Malware Traffic Detection and Classification; Haber, P., Lampoltshammer, T., Mayr, M., Plankensteiner, K., Eds.; Science—Analytics and Applications: Wiesbaden, German; pp. 105–112. ISBN 978-3-658-32181-9.
  29. Ahn, S.; Kim, J.; Park, S.Y.; Cho, S. Explaining Deep Learning-based Traffic Classification using a Genetic Algorithm. IEEE Access 2020, 9, 4738–4751. [Google Scholar] [CrossRef]
  30. Yang, Y.; Zheng, K.; Wu, C.; Yang, Y. Improving the classification effectiveness of intrusion detection by using improved conditional variational autoencoder and deep neural network. Sensors 2019, 19, 2528. [Google Scholar] [CrossRef] [Green Version]
  31. Han, L.; Zeng, X.; Song, L. A novel transfer learning based on albert for malicious network traffic classification. Int. J. Innov. Comput. Inf. Control 2020, 16, 2103–2119. [Google Scholar]
  32. He, H.Y.; Yang, Z.G.; Chen, X.N. PERT: Payload Encoding Representation from Transformer for Encrypted Traffic Classification. In Proceedings of the 2020 ITU Kaleidoscope: Industry-Driven Digital Transformation (ITU K), Ha Noi, Vietnam, 7–11 December 2020; pp. 1–8. [Google Scholar]
  33. He, M.; Jin, L.; Wang, X.; Li, Y. Web log classification framework with data augmentation based on GANs. J. China Univ. Posts Telecommun. 2020, 27, 34. [Google Scholar]
  34. Mihailescu, M.E.; Mihai, D.; Carabas, M.; Komisarek, M.; Pawlicki, M.; Hołubowicz, W.; Kozik, R. The Proposition and Evaluation of the RoEduNet-SIMARGL2021 Network Intrusion Detection Dataset. Sensors 2021, 21, 4319. [Google Scholar] [CrossRef]
  35. Ahmad, R.; Wazirali, R.; Bsoul, Q.; Abu-Ain, T.; Abu-Ain, W. Feature-Selection and Mutual-Clustering Approaches to Improve DoS Detection and Maintain WSNs’ Lifetime. Sensors 2021, 21, 4821. [Google Scholar] [CrossRef]
  36. Chen, C.; Jafari, R.; Kehtarnavaz, N. A survey of depth and inertial sensor fusion for human action recognition. Multimed. Tools Appl. 2017, 76, 4405–4425. [Google Scholar] [CrossRef]
  37. Dawar, N.; Kehtarnavaz, N. Action detection and recognition in continuous action streams by deep learning-based sensing fusion. IEEE Sens. J. 2018, 18, 9660–9668. [Google Scholar] [CrossRef]
  38. Dawar, N.; Ostadabbas, S.; Kehtarnavaz, N. Data augmentation in deep learning-based fusion of depth and inertial sensing for action recognition. IEEE Sens. Lett. 2018, 3, 1–4. [Google Scholar] [CrossRef]
  39. UNSW-NB15. 2021. Available online: https://cloudstor.aarnet.edu.au/plus/index.php/s/2DhnLGDdEECo4ys (accessed on 10 July 2021).
  40. Intrusion Detection Evaluation Dataset (ISCXIDS2012). 2021. Available online: https://www.unb.ca/cic/datasets/ids.html (accessed on 10 July 2021).
  41. Intrusion Detection Evaluation Dataset (CIC-IDS2017). 2021. Available online: https://www.unb.ca/cic/datasets/ids-2017.html (accessed on 10 July 2021).
  42. VPN-nonVPN Dataset (ISCXVPN2016). 2021. Available online: https://www.unb.ca/cic/datasets/vpn.html (accessed on 10 July 2021).
  43. Yang, S.; Wu, P.; Guo, H. DualNet: Locate Then Detect Effective Payload with Deep Attention Network. arXiv 2020, arXiv:2010.12171. [Google Scholar]
  44. Khan, F.A.; Gumaei, A.; Derhab, A.; Hussain, A. A novel two-stage deep learning model for efficient network intrusion detection. IEEE Access 2019, 7, 30373–30385. [Google Scholar] [CrossRef]
  45. Vinayakumar, R.; Alazab, M.; Soman, K.; Poornachandran, P.; Al-Nemrat, A.; Venkatraman, S. Deep learning approach for intelligent intrusion detection system. IEEE Access 2019, 7, 41525–41550. [Google Scholar] [CrossRef]
  46. Janarthanan, T.; Zargari, S. Feature selection in UNSW-NB15 and KDDCUP’99 datasets. In Proceedings of the 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE), Edinburgh, UK, 19–21 June 2017; pp. 1881–1886. [Google Scholar]
  47. Guo, L.; Wu, Q.; Liu, S.; Duan, M.; Li, H.; Sun, J. Deep learning-based real-time VPN encrypted traffic identification methods. J. Real-Time Image Process. 2020, 17, 103–114. [Google Scholar] [CrossRef]
  48. Song, M.; Ran, J.; Li, S. Encrypted Traffic Classification Based on Text Convolution Neural Networks. In Proceedings of the 2019 IEEE 7th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 19–20 October 2019; pp. 432–436. [Google Scholar]
  49. Kim, I.; Chung, T.M. Malicious-Traffic Classification Using Deep Learning with Packet Bytes and Arrival Time. In Proceedings of the International Conference on Future Data and Security Engineering, Quy Nhon, Vietnam, 25–27 November 2020; pp. 345–356. [Google Scholar]
  50. Le, T.T.H.; Kim, Y.; Kim, H. Network intrusion detection based on novel feature selection model and various recurrent neural networks. Appl. Sci. 2019, 9, 1392. [Google Scholar] [CrossRef] [Green Version]
  51. Sun, P.; Liu, P.; Li, Q.; Liu, C.; Lu, X.; Hao, R.; Chen, J. DL-IDS: Extracting features using CNN-LSTM hybrid network for intrusion detection system. Secur. Commun. Netw. 2020, 2020, 8890306. [Google Scholar] [CrossRef]
  52. Ferrag, M.A.; Maglaras, L. DeepCoin: A novel deep learning and blockchain-based energy exchange framework for smart grids. IEEE Trans. Eng. Manag. 2019, 67, 1285–1297. [Google Scholar] [CrossRef] [Green Version]
  53. Zhou, Y.; Cheng, G.; Jiang, S.; Dai, M. An efficient intrusion detection system based on feature selection and ensemble classifier. arXiv 2019, arXiv:1904.01352. [Google Scholar] [CrossRef] [Green Version]
Figure 1. This figure depicts the process of malicious behavior detection. It contains two types of raw data collected from network traffic and three fusion methods. Detection results can be improved by using different fusion methods.
Figure 1. This figure depicts the process of malicious behavior detection. It contains two types of raw data collected from network traffic and three fusion methods. Detection results can be improved by using different fusion methods.
Sensors 21 05942 g001
Figure 2. The structure of early data fusion model. The fusion process occurs in the early stages of the detection process, which can be regarded as the direct fusion of raw data information.
Figure 2. The structure of early data fusion model. The fusion process occurs in the early stages of the detection process, which can be regarded as the direct fusion of raw data information.
Sensors 21 05942 g002
Figure 3. The structure of feature fusion. The fusion process occurs after the feature generation of detection model.
Figure 3. The structure of feature fusion. The fusion process occurs after the feature generation of detection model.
Sensors 21 05942 g003
Figure 4. The structure of decision fusion. The fusion process occurs at the end of detection procedure.
Figure 4. The structure of decision fusion. The fusion process occurs at the end of detection procedure.
Sensors 21 05942 g004
Figure 5. The structure of 1D and 2D convolution on network traffic data.
Figure 5. The structure of 1D and 2D convolution on network traffic data.
Sensors 21 05942 g005
Figure 6. The structure of a PCAP file which consists of a PCAP header, a number of data packets with the packet header and packet content.
Figure 6. The structure of a PCAP file which consists of a PCAP header, a number of data packets with the packet header and packet content.
Sensors 21 05942 g006
Figure 7. Data processing steps for PCAP files with data split, data clean, data transfer and data trimming.
Figure 7. Data processing steps for PCAP files with data split, data clean, data transfer and data trimming.
Sensors 21 05942 g007
Figure 8. The changes of F1-score in the training process on VPN2016.
Figure 8. The changes of F1-score in the training process on VPN2016.
Sensors 21 05942 g008
Figure 9. Confusion matrices of detection results on dataset NB15. True labels are on vertical axis and predicted labels are on horizontal axis.
Figure 9. Confusion matrices of detection results on dataset NB15. True labels are on vertical axis and predicted labels are on horizontal axis.
Sensors 21 05942 g009aSensors 21 05942 g009b
Figure 10. Confusion matrices of detection results on dataset VPN2016.
Figure 10. Confusion matrices of detection results on dataset VPN2016.
Sensors 21 05942 g010
Figure 11. Confusion matrices of detection results on dataset CIC.
Figure 11. Confusion matrices of detection results on dataset CIC.
Sensors 21 05942 g011
Figure 12. Classification results of each category on dataset NB15. They are composed of Precision, Recall and F1-score, which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Figure 12. Classification results of each category on dataset NB15. They are composed of Precision, Recall and F1-score, which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Sensors 21 05942 g012
Figure 13. Classification results of each category on dataset VPN2016. They are composed of Precision, Recall and F1-score which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Figure 13. Classification results of each category on dataset VPN2016. They are composed of Precision, Recall and F1-score which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Sensors 21 05942 g013
Figure 14. Classification results of each category on dataset CIC. They are composed of Precision, Recall and F1-score which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Figure 14. Classification results of each category on dataset CIC. They are composed of Precision, Recall and F1-score which are calculated by a simple model with PCAP data, data fusion model, feature fusion model and decision fusion model.
Sensors 21 05942 g014
Figure 15. Visualization of 1D and 2D dimensional convolution feature maps.
Figure 15. Visualization of 1D and 2D dimensional convolution feature maps.
Sensors 21 05942 g015
Figure 16. Experiment results on dataset VPN2016 used 1D convolution and 2D convolution. The interval of each point represents the interval between samples.
Figure 16. Experiment results on dataset VPN2016 used 1D convolution and 2D convolution. The interval of each point represents the interval between samples.
Sensors 21 05942 g016
Table 1. Some related methods in the past three years.
Table 1. Some related methods in the past three years.
Ref.MethodDescriptionType
[15] 2019RFMethod that solves data imbalance.ML
[16] 2019DT, RF, KNNAn adaptive voting algorithm.ML
[17] 2021DiFF-RFRandom partitioning binary trees.ML
[18] 2019Cluster, RFProcess highly imbalanced data.ML
[19] 2017Genetic algorithmImprove accuracy and efficiency.ML
[20] 2019SVMImprove accuracy.ML
[21] 2020CNNOutstanding performance.DL
[22] 2020CNN, GANData augmentation, imbalanced data.DL
[23] 2021CNN, image-basedImprove accuracy, interpretable model.DL
[24] 2018Text-CNN, RFGood classification results.DL
[25] 2019VAE, fingerprintingGood identification representations.DL
[26] 2020AE, end-to-endImbalanced classification.DL
[27] 2021AE, CNNFew-shot malicious traffic detection.DL
[28] 2021CNN, LSTMFeature representations, accuracy.DL
[29] 2020Genetic, DNNImprove accuracy, interpretable.DL
[30] 2019VAE, DNNOutstanding classification performance.DL
[31] 2020TransformerImprove efficiency and accuracy.DL
[32] 2021TransformerThe dynamic word embedding used.DL
[33] 2020GAN, LSTMData augmentation, improve accuracy.DL
Table 2. Features list in business feature data.
Table 2. Features list in business feature data.
Feature TypeFeature Contents
Flow featuresSource port number, Destination port number, Protocol, …
Base featuresRecord total duration, Source to destination time to live, …
Content featuresSource TCP window advertisement, Source TCP sequence number, …
Time featuresRecord start time, Source inter-packet arrival time, …
Generated featuresSome statistical or calculated value
Table 3. Samples number of training set and test set.
Table 3. Samples number of training set and test set.
Dataset NameTypeNumber
Dataset NB15Training set175,341
Test set82,332
Dataset VPN2016Training set125,744
Test set53,890
Dataset CIC2012, CIC2017Training set634,229
Test set271,812
Table 4. Experiment results and comparation of data NB15.
Table 4. Experiment results and comparation of data NB15.
TypeMethodPrecRecallAccF1-MacroF1-Weight
Previous MethodsDualNet, Yang et al. [43]--0.833--
Khan et al. [44]--0.8913--
Vinayakumar et al. [45]0.6230.660.660.596
Yang et al. [30]0.77430.97390.85970.8627
Janarthanan et al. [46]--0.8162--
Khammassi et al. [19]--0.8142--
Comparative MethodsLR0.30890.34830.42250.34190.4697
KNN0.55820.55930.56020.55860.5593
DT0.78960.77560.78800.76800.7902
RF0.77750.77620.77660.77680.7766
XGBoost0.67490.67810.67620.67640.6759
Proposed MethodsSimple (PCAP)0.82780.96530.91830.88220.9228
Simple (Feature)0.61300.81160.79690.65930.8169
Early data fusion0.86970.97600.94100.91320.9440
Feature fusion0.87900.98020.95140.92210.9534
Decision fusion0.86160.97360.93690.90730.9397
Table 5. Experiment results and comparation of data VPN2016.
Table 5. Experiment results and comparation of data VPN2016.
TypeMethodPrecRecallAccF1-MacroF1-Weight
Previous MethodsGuo et al. [47]--0.9292--
Song et al. [48]0.8760.873-0.875
Comparative MethodsLR0.26840.20250.23050.22800.2399
KNN0.68640.67190.68000.67150.6812
DT0.84580.84040.84190.84160.8488
RF0.85260.85210.85520.84910.8557
XGBoost0.80000.84930.81340.80460.8132
Proposed MethodsSimple (PCAP)0.93790.93070.94580.93380.9459
Simple (Feature)0.60110.58420.58430.57660.5702
Early data fusion0.90360.89870.91790.90050.9180
Feature fusion0.95000.94710.95290.94790.9531
Decision fusion0.95510.94990.95940.95190.9596
Table 6. Experiment results and comparation of data CIC2012 and CIC2017.
Table 6. Experiment results and comparation of data CIC2012 and CIC2017.
TypeMethodPrecRecallAccF1-MacroF1-Weight
Previous MethodsKim et al. [49]0.940.94-0.94
Le et al. [50]0.94750.975-0.9708
Sun et al. [51]--0.9844--
Ferrag et al. [52]--0.9823--
Zhou et al. [53]--0.968--
Vinayakumar et al. [45]0.9720.9620.9620.965
Comparative MethodsLR0.11130.02750.04180.03780.0410
KNN0.59950.60310.59700.59690.5970
DT0.95040.94980.94990.94960.9504
RF0.93930.93980.94110.93930.9401
XGBoost0.90650.91110.90120.90390.9067
Proposed MethodsSimple (PCAP)0.98090.97830.98230.97950.9823
Simple (Feature)0.70720.63440.62880.64700.6495
Early data fusion0.98170.98030.98710.98140.9871
Feature fusion0.99240.99150.99260.99190.9926
Decision fusion0.99080.98970.99060.99020.9906
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

He, M.; Wang, X.; Jin, L.; Dai, B.; Kacuila, K.; Xue, X. Malicious Network Behavior Detection Using Fusion of Packet Captures Files and Business Feature Data. Sensors 2021, 21, 5942. https://doi.org/10.3390/s21175942

AMA Style

He M, Wang X, Jin L, Dai B, Kacuila K, Xue X. Malicious Network Behavior Detection Using Fusion of Packet Captures Files and Business Feature Data. Sensors. 2021; 21(17):5942. https://doi.org/10.3390/s21175942

Chicago/Turabian Style

He, Mingshu, Xiaojuan Wang, Lei Jin, Bingying Dai, Kaiwenlv Kacuila, and Xiaosu Xue. 2021. "Malicious Network Behavior Detection Using Fusion of Packet Captures Files and Business Feature Data" Sensors 21, no. 17: 5942. https://doi.org/10.3390/s21175942

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop