Next Article in Journal
Development and Validation of a Multimodal-Based Prognosis and Intervention Prediction Model for COVID-19 Patients in a Multicenter Cohort
Previous Article in Journal
Sliding-Window Normalization to Improve the Performance of Machine-Learning Models for Real-Time Motion Prediction Using Electromyography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Method for Improved Network Traffic Prediction Using Enhanced Deep Reinforcement Learning Algorithm

by
Nagaiah Mohanan Balamurugan
1,*,
Malaiyalathan Adimoolam
2,
Mohammed H. Alsharif
3 and
Peerapong Uthansakul
4,*
1
Department of Computer Science and Engineering, Sri Venkateswara College of Engineering, Sriperumbudur 602117, India
2
Department of Computer Science and Engineering, Saveetha School of Engineering, Saveetha Institute of Medical and Technical Sciences, Thandalam 602105, India
3
Department of Electrical Engineering, College of Electronics and Information Engineering, Sejong University, Seoul 05006, Korea
4
School of Telecommunication Engineering, Suranaree University of Technology, Nakhon Ratchasima 30000, Thailand
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(13), 5006; https://doi.org/10.3390/s22135006
Submission received: 24 May 2022 / Revised: 19 June 2022 / Accepted: 29 June 2022 / Published: 2 July 2022
(This article belongs to the Section Sensor Networks)

Abstract

:
Network data traffic is increasing with expanded networks for various applications, with text, image, audio, and video for inevitable needs. Network traffic pattern identification and analysis of traffic of data content are essential for different needs and different scenarios. Many approaches have been followed, both before and after the introduction of machine and deep learning algorithms as intelligence computation. The network traffic analysis is the process of incarcerating traffic of a network and observing it deeply to predict what the manifestation in traffic of the network is. To enhance the quality of service (QoS) of a network, it is important to estimate the network traffic and analyze its accuracy and precision, as well as the false positive and negative rates, with suitable algorithms. This proposed work is coining a new method using an enhanced deep reinforcement learning (EDRL) algorithm to improve network traffic analysis and prediction. The importance of this proposed work is to contribute towards intelligence-based network traffic prediction and solve network management issues. An experiment was carried out to check the accuracy and precision, as well as the false positive and negative parameters with EDRL. Also, convolutional neural network (CNN) machines and deep learning algorithms have been used to predict the different types of network traffic, which are labeled text-based, video-based, and unencrypted and encrypted data traffic. The EDRL algorithm has outperformed with mean Accuracy (97.20%), mean Precision (97.343%), mean false positive (2.657%) and mean false negative (2.527%) than the CNN algorithm.

1. Introduction

Internet data traffic has been enormously burst out, due to the introduction of big data capable data, along with the invention of speed network components. The resources of such a network are essential, and it have to be utilized for the intended purpose; additionally, it is a very challenging task to monitor and predict data traffic for various reasons. For data traffic prediction, manual processing-based prediction and artificial intelligence-based methods and techniques have already been deployed. Some of the methods and techniques are as follows. (i) The prediction of daily internet traffic using a data mining technique for smart university application [1]; (ii) A low complexity-based boost machine learning algorithm with classification and regression to predict internet data traffic from weak learning to strong learning [2]; (iii) A double exponential predictor [3], based on artificial neural network (ANN), classic time series, and wavelet transform-based predictors; (iv) A deep learning-based prediction [4] for metropolitan area network traffic; (v) A neural network ensemble [5] for internet traffic forecasting.
Network traffic prediction [6] is inevitable, due to the cost of bandwidth, time complexity measurement, data prediction, suspicious traffic identification, and so on. The traffic of network impact is directly proportional to the bandwidth and life span of network and multi-user identification. A recurrent neural network (RNN) was adopted to find network traffic for proactive network management and planning [7]. The RNN prediction formula is as shown in Equation (1).
y i = w i h i + y i 1
where yi is the predicted value at time i, wi is weight of input, hi is hidden layer state at time i, and yi−1 is predicted value at time i − 1 [8]. The GEANT backbone network was tested with the network structure of RNN and optical network parameters with 200 epochs. Yet another work was discussed regarding four different algorithms for predicting network traffic [9]. Those algorithms were RNN, deep learning stacked auto-encoder, multilayer perceptron (MLP), and MLP with back-propagation. Aspects such as adaptive application, bandwidth detection, congestion control, and anomaly detection and admission control network traffic and have been discussed with time series internet traffic prediction.
The application of network traffic analysis and prediction consist of bandwidth monitoring, data analysis, efficient network management for intended users, and so on. Little research has been carried out to measure the bandwidth efficiency. Work regarding bandwidth utilization and forecasting model was discussed for bandwidth utilization with ARIMA and SNMP setups [10]. The computational time was measured, and it was achieved at 83.2%, along with forecast error and standard deviation.
Yet another application-oriented network traffic prediction was performed for measuring the accuracy and timely internet traffic information [4]. In this work, the proposed mechanism has detected regarding network traffic for anomaly detection, admission control, bandwidth allocation, and congestion control with big traffic data and deep architecture model-based internet traffic flow prediction. The novelty has been achieved with special and temporal correlations, as well as the glow data character approach. The training data set was trained in the greedy layer-wise fashion. The dataset is taken from China Unicom for network work utilization. Yet another work surveyed real world network traffic prediction with various machine learning algorithms with a cognitive approach [11]. Here, the applications are coined based on their classifications for threat category, regression for value prediction, and ranking for ordering traffic. The learning algorithms discussed here were neural network, linear time series models, principal component analysis (PCA), linear regression (LR), statistical model, and support vector machine (SVM) for either long-term or short-term predictions. The application’s performance measures have been taken as data availability and system complexity for both local and wide-area networks. From discussed techniques, the applications supported were cellular traffic, optical networks, LTEs network, IP networks, TCP traffic, MPEG, JPEG traffic, Ethernet traffic, and many more.

2. Related Works

The Applications of Network Traffic Analysis

The major applications of network traffic prediction are network management, resource allocation, quality of service (QoS) from the internet service provider (ISP), cyberspace security protections, and malware detection [12]. The ISCX and QUIC public dataset was used here to measure the performance of traffic, with a proposed method called multi-task learning framework. Yet another work was related to the online application of current internet performance measures, as determined by analyzing encrypted packet, virtual private network, and non-VPN traffic using the proposed method, referred to as deep packet. This work had taken file transfer protocol (FTP) and peer-to-peer (P2P) network traffic. The recall performance was measured for the UNB ISCX VPN and non-VPN dataset [13]. The network classifier approach was deployed in hyper-text transfer protocol (HTTP) and session initiation protocol (SIP) [14]. Performance measures such as duration, latency, and traffic volume were measured using RNN and CNN learning algorithms.
In the past decade, more than 40 research works have been introduced that discussed network traffic analysis with manual traffic prediction, machine learning-based network traffic prediction, and deep learning-based traffic prediction. Most of the work used machine and deep learning algorithms to predict network traffic. A work was introduced to predict network traffic using a time series approach with recurrent neural network (RNN) [7]. Also, the variation of RNNs were analyzed using past network traffic dataset. Its performance was measured using the GEANT research and educational network. An experiment was conducted with 200 epochs with a learning rate of 0.01 to 0.5. The performance of variant long short-term memory (LSTM) was better than other variants of RNN.
For network traffic, most cited articles were related to deep learning and machine learning traffic identification algorithms. One work coined a suitable lightweight framework with a deep learning algorithm. These frameworks have penetrated the encrypted traffic, classified the deep full range, and detected the intrusion with two datasets [15]. Yet another application, i.e., the user activity monitoring-based network traffic, was developed using the machine learning algorithm [16]. K-mean and random forest (RF) algorithms have been used to measure the network traffic QoS, accuracy, and real time traffic generated with time bound. A network management-based traffic classification with software defined network (SDN) was tested with a CNN and stacked auto encoder (SAE). This proposed work was used for online traffic service. Recall, accuracy, and precision were measured with a deep learning algorithm [17]. Finally, Table 1 depicts the methods and classes of machine and deep learning algorithms. The CNN prediction formula is given in Equation (2), and it is a neuron calculation for traffic.
y i = b i + i = 1 n w i × x i
where yi is the neuron calculation, wi is the weight matrix of input, xi is input, and bi is the bias of the neuron.
Even though a considerable amount of research has been carried out, accurate predictions and huge traffic predictions remain unclear. So, it is important to carry out inevitable solutions to predict traffic from huge datasets. On the other hand, using CNN for traffic predictions has to be checked, in regard to whether there is less energy consumption and simple infrastructure-based mechanism for traffic prediction of the internet world. Thus, the enhanced reinforcement learning algorithm would be the best to choose. It is important to sort out the aim proposed in this work. The aim of this work is to predict network traffic and measure its related performance measures, such as accuracy and precision, as well as false positive and negative rates, using the EDRL algorithm and comparing its performance with the KNN and CNN algorithms.

3. Materials and Methods

This research work was carried out in the machine learning lab, Saveetha school of Engineering, Saveetha University. The proposed algorithm, i.e., the EDRL algorithm, was compared with CNN. Group 1 was taken as the proposed EDRL and CNN was taken as group 2. For each group, 25 iterative samples are set for experiment, with a dataset size of 54,000. A total of 80% of the pre-test power (G power) was obtained, along with alpha value 0.05. The dataset was collected from ISCXVPN2016 [18] for virtual private network (VPN) and non-VPN. The size of the dataset is almost 15 GB of ARFF file format, with a set of attributes for instance sharing. The dataset is traced with scenarios with different traffics from networks. The dataset is divided into two partitions for testing and training purposes. The training set contains 80%, and the testing set contains 20%.
To understand how EDRL is better than reinforcement learning (RL), deep reinforcement learning (DRL), and deep learning (DL), the following (Table 2) illustrates the relationship between the input and methods, along with the policies and problems for different machine and deep learning algorithms.

3.1. The EDRL Algorithm

EDRL processes input with DL networks to grain the output of accurate network traffic; then, the gained output undergoes reinforcement learning with policy optimization using policy gradient methods, thus achieving an EDRL method with higher accuracy than normal. This method entirely undergoes a multi-layer perceptron neural network for function approximation and various reward functions. The mechanism works after DNN checks and pre-processes the network traffic dataset, using Monte Carlo learning check flow independently and dictionary-based learning to reiterate the RL, along with reward mechanism, achieves fine-tuned accuracy.

3.2. Dataset Pre-Processing

Data pre-processing is the way to process the data for training and testing. Let X be the set of data taken from the dataset and Equation (3), which is called to pre-process the network traffic data for prediction with accuracy and precision.
x = X X m i n X m a x X m i n

3.3. Feature Engineering

To enhance the accuracy, various features are selected and form a feature set, A = a1 to an. These features are based on the parameters of network traffic and types of applications of network traffic. Features such as average segment size, window size, round trip delay time (RTT), variance of packet, actual data size, client port number, and server port numbers were taken to predict network traffic; the features are represented as F in Equation (4), and the description is illustrated in Table 3.
F = f 1 ,   f 2 , f 3 f 7
The classes as www packets, P2P packets, mail packets, database packets, and multimedia packets and its classes are represented as C in Equation (5), and its notation and applications are listed in Table 4. The classes of network c1 to c5 of network traffic are processed with EDRL, with deep penetration of functions f1 to f7, as listed in Table 3. This function can trace the format of network traffic data, such as general browsing data, torrent streaming, SMTP, POP, MIME, IMAP, SQL net, and video storage server of YouTube. The function f3 is checking the variance of packets to check traffic very accurately. RTT checks the delay time occurrence of network traffic. Further, the average size of the segment elaborates on how the network traffic changes from time to time.
C = c 1 , c 2 , c 3 c 5

3.4. The DNN Multi-Layer Perceptron Method

It is a sophisticated model with a mathematical-based complex data processing network. It is the technology built-based model for simulating a neural connection. This is here to represent the recognition of pattern of network traffic dataset by passing input, and it will process through the neural connection with various layers, as shown in Figure 1. Usually, DNN has an input layer and output layer, which are placed between many hidden layers. After the input layer accepts the input as traffic network data, further hidden layers perform the sorting and ordering of data, and finally outputting the layer for getting pre-processed data of traffic network. The main functionality of this DNN is that it handles the unlabelled or unstructured data and introduces multilayer perceptron. Further, it is clear that this input layer has to feed the data of network traffic, both encrypted and unencrypted, including the different functions of f1 to f7 and different classes of c1 to c5. Then, the neighbor of this input layer trains the dataset and finds its threshold values, until deep EDRL getts its corresponding traffic classification clearly. This can be achieved with the help of the output layer of DNN.

3.5. Monte Carlo Learning for Network Traffic Analysis

Monte Carlo learning (MCL) is a Q-learning (QL) method of reinforcement learning (RL), which calculates the policy with agents to do research to find which policy gains more rewards. For traffic to be predicted accurately, reward functions play an inevitable role. Additionally, for the types of network traffic, relative rewards were calculated. For network traffic analysis, the Markov decision process (MDP) was used, and its components are A, P, R, RF, and S [19], where S is set of states in traffic analysis, A is set of actions in traffic analysis, R is reward for traffic analysis from one state to another state, P is set of policy, and, finally, RF is the reduction factor. The working mechanism of QL is that, for traffic analysis, there is a transition of state s1 (s1 ∈ S) to s2 (s2 ∈ S) for carryout action process (a ∈ A) with policy P (s1 to s2, a1), along with reward R (s1 to s2, a1), which could be calculated. This reward calculation will be introducing novelty for traffic prediction with the network dataset. This QL process is illustrated as Equation (6), as follows.
Q L s i , a i =   R F ( max ( Q s i + 1 , a i + 1 +   R s i , a i
where (ai ∈ A) is action set and this QL is the process to achieve maximum network traffic prediction, and this will be an iterative process to achieve the maximum reward with the sum of the reward, and it is expressed as Equation (7), as follows.
Q s i , a i = Δ 1 Q s i + 1 , a i + 1 +   Δ ( R s i , a i + R F ( max Q s i + 1 , a i + 1
Based on Equations (6) and (7) Monte Carlo QL is used to train the network traffic dataset as Equitation (8).
X   t =   Q L X low , X high
For network traffic prediction, the QL-based traffic prediction is coined as Algorithm 1, after calling EDRL.
Algorithm 1: Monte Carlo Learning for Network Traffic Analysis
Pre-requisite: Pre-processed dataset (X) for number of iterations
 Assure: Max (R(X, Ai)
 X = (X−Xmin)/(Xmax−Xmin)
 QL = Xmax − Xmin*Xmax+1 − Xmin+1
 Ri(Si, Si+1) = max(1, Ri(Si, Si+1))
 F = {f1, f2, f3…f7}
 C = {c1, c2, c3…c5}
 For each Si ∈ X do
 Q(si,ai) = RF(max(Q(si+1, ai+1) + R(si, ai)
 Q(si,ai) = (Δ − 1)Q(si+1, ai+1) + Δ(R(si, ai)+RF(max(Q(si+1, ai+1))
 Ri+1(Si, Si+1) = Ri+1(Si, Si+1) +1
 End for
 Max (R(X,Ai)

3.6. Agent—EDRL Traffic Model

EDRL is one of the artificial learning algorithms that computes optimized output, thus enhancing its feedback, along with the deep learning concept. Pre-processed input collected from a dataset with non-encrypted and different application-based network traffics [18] is fed in EDRL algorithm. The dataset is trained in EDRL algorithm and predicts the data as classified 1 to N, and its accuracy has been measured as a parameter with an event-based approach, instead of time series; similarly, the experiment has been carried out with precision and false positive and negative rates. The EDRL architecture is shown in Figure 2. In this work, the network traffic of the encrypted and non-encrypted dataset is used. It consists of VPN- and non-VPN-based patterns. The encrypted dataset of the ARFF file format is essential for tracing network traffic. This ARFF also includes Skype and multimedia network traffic. This work has also traced the network traffic of www packets, P2P packets, mail packets, database packets, and multimedia packets.
This overall EDRL algorithm is coined as Algorithm 2 for network traffic prediction and accuracy measurement.
Algorithm 2: EDRL Algorithm for Network Traffic Prediction
Pre-requisite: pre-processed network traffic data from dataset
 Assure: max (precision), max (accuracy), min (falsepositive), min (falsenegative)
 QL = Xmax − Xmin*Xmax+1 − Xmin+1
 Call feature engineering function
 Call DNN multilayer perceptron method
 Call Monte Carlo learning for network traffic analysis Algorithm 1

3.7. Accuracy and Precision for Network Traffic Analysis

The accuracy of identifying network traffic rests on the closeness of the specific value, while the precision is the measurement of the closeness of the network traffic to each other while checking for network traffic prediction. Equation (9) represents the accuracy for the network traffic data.
Accuracy = TP + TN TP + TN + FP + FN
where TP is true positive and TN is true negative, FP is false positive and FN is false negative. Equation (10) is a formula that is used to measure the precision of network traffic prediction.
Precision = TP TP + FP

3.8. Statistical Analysis

The measured traffic categories’, i.e., accuracy, precision, falsepositive and falsenegative, parameter performance values were collected by repeated experiments for 10 iterations, and these values were recorded. Further, the recorded values of each parameter were tabulated, and comparative graphs were drawn with help of a statistical IBM SPSS tool. The dependent and independent variables are represented to measure the best accuracy-incurring algorithm by comparing EDRL algorithm with CNN and KNN algorithm. Here, the feature of network traffic and classes act as a dependent variables and EDRL and CNN algorithms act as an independent variables; they are used to carry out the experiment to predict precise and accurate network traffic along with the falsepositive and falsenegative parameters.

4. Numerical Results

4.1. Accuracy Comparison

The implemented experiment results have been taken iteration-wise, and the group statistics were carried out with the IBM SPSS tool. For the 10 iterated samples of trained and tested datasets, an independent sample t-test was built. The Group 1 algorithm was taken as EDRL, and the group 2 algorithm was taken as the CNN algorithm. The statistical results’ were observed. Table 5 shows the group statistics of the EDRL and KNN algorithms, as well as EDRL and CNN. The experiment was carried out with 10 iterations of EDRL and compared with the CNN algorithms separately. It is noticed that mean accuracy of EDRL was 97.20%, and the CNN mean accuracy was 93.055%. The standard deviation is comparatively less 1.702% in the EDRL and CNN algorithms (2.298%). Further, the standard error mean was also 0.538% for EDRL, which is comparatively less than the CNN algorithm (0.7268%). The experiment inferred that the validation and trained accuracy increased when the number of iterations was increased for EDRL than for the KNN and CNN algorithms.
Further, the significance value was calculated between the EDRL and CNN algorithms using the SPSS tool comparing the independent t-test as analytics. Table 6 shows the significance value. The inference is that there is a significant difference between EDRL and CNN of 0.306. The inference further claims that there is slight difference between alpha test p = 0.05 with inferred difference.

4.2. Precision Comparison

Table 7 shows the group statistics of the EDRL and CNN algorithms’ precision for 10 iterations of EDRL and compared this with the CNN algorithms separately. It is noticed that mean precision of EDRL was 97.373%, whereas the CNN mean precision was 93.972%. The standard deviation was comparatively less, at 1.5189%, in the EDRL algorithm than the CNN algorithm (2.403%). Further, the standard error mean was 0.4803% for EDRL, which is comparatively less than the CNN algorithm (0.7594%). The inference is that the validation and trained accuracy increased when the number of iterations increased for EDRL, rather than the CNN algorithm.
The significance value was calculated between the EDRL, and the CNN algorithms’ precision, using the SPSS tool, compared the independent t-test as analytics. Table 8 is shows the significance value of the mean precision. The inference is that there is a significant difference between EDRL and CNN of 0.143. The inference further claims that there is slight difference between alpha test p = 0.05, with inferred difference.

4.3. False Positive Comparison

Table 9 shows the group statistics of the EDRL and CNN algorithms’ false positive rates for 10 iterations of EDRL, and they compared with CNN algorithm separately; we noticed that the mean false positive of EDRL was 2.657%, whereas the CNN mean false positive was 6.325%. The standard deviation is comparatively less 1.853% in the EDRL algorithm than the CNN algorithm (2.191%). Further, the standard error mean was 0.581% for EDRL, which was comparatively less than the KNN (0.731%) and CNN (0.693%) algorithms. The inference that the validation and trained accuracy increased when the number of iterations increased for EDRL, rather than the KNN and CNN algorithms.
The significance value calculated between EDRL and CNN algorithms’ false positive using SPSS tool using comparing independent t-test as analytics. Table 10 is showing the significance value of mean false positive. The inference is that there is a significant difference between EDRL and CNN as 0.143. The inference further claims that there is slight difference between alpha test p = 0.05 with inferred difference.

4.4. False Negative Comparison

Table 11 shows the group statistics of EDRL and CNN algorithms’ false negative. For 10 iterations of EDRL and compared with CNN algorithms separately and noticed that mean False negative of EDRL is 2.527% whereas CNN mean False negative is 5.675%. Standard deviation is comparatively less 1.227% in EDRL algorithm than CNN algorithm (1.992%). Further standard error mean is 0.381% for EDRL which is comparatively less than CNN algorithm 0.61643%. Inference that the validation and train Accuracy increases when number of iteration is increases for EDRL than CNN algorithms.
The significance value calculated between EDRL and CNN algorithms’ false negative using SPSS tool using the comparing independent t-test as analytics. Table 12 shows the significance value of the mean false negative. The inference is that there is a significant difference between the ERDL and CNN of 0.143. The inference further claims that there is a slight difference between the alpha test p = 0.05, with inferred difference.

4.5. Accuracy Comparison for EDRL and CNN Algorithms

With the SPSS tool, the mean accuracy of the ERDL and CNN algorithms has been compared, as shown in Figure 3. The graph was generated with SPSS graph builder, with the X axis as the ERDL vs. CNN algorithms and Y axis as the mean accuracy. The standard deviation is set at ±2 with confidence interval of 95%. The mean accuracy of the ERDL algorithm was higher than the CNN algorithm. For the 10 epochs, the mean accuracy was considerably higher for the ERDL than the CNN algorithm, as the sample size increased and error rates decreased with the increasing sample size for two algorithms.

4.6. Precision Comparison for EDRL and CNN Algorithms

With the SPSS tool, the mean precision of the ERDL and CNN algorithms was compared, as shown in Figure 4. The graph was generated with the SPSS graph builder, with the X axis as the ERDL vs. CNN algorithms and Y axis as mean the precision. The standard deviation was set at ±2, and the confidence interval was 95%. The mean precision of the ERDL algorithm was higher than the CNN algorithm. For the 10 epochs, the mean precision was considerably higher for ERDL than the CNN algorithm, as the sample size increased and error rates also decreased, with increasing sample sizes for the two algorithms.

4.7. False Positive Comparison for EDRL and CNN Algorithms

With the SPSS tool, the mean false positive of ERDL and CNN algorithms were compared, as shown in Figure 5. The graph was generated with the SPSS graph builder with the X axis as the ERDL vs. CNN algorithms and the Y axis as mean false positive. The standard deviation was set at ±2, and the confidence interval was 95%. The mean precision of the ERDL algorithm was higher than the CNN algorithm. For the 10 epochs, the mean false positive was considerably higher for ERDL than the CNN algorithm, as sample size increased and error rates decreased, with increasing sample sizes for the two algorithms.

4.8. False Negative Comparison for EDRL and CNN Algorithms

With the SPSS tool, the mean false negative of ERDL and CNN algorithms has been compared, as shown in Figure 6. The graph was generated with the SPSS graph builder, with the X axis as the ERDL vs. CNN algorithms, and the Y axis as the mean false negative. The standard deviation was set at ±2, and the confidence interval was 95%. The mean precision of the ERDL algorithm was higher than the CNN algorithm. For 10 epochs, the mean false negative was considerably higher for the ERDL than the CNN algorithm, and as the sample size increased, and the error rates decreased with increasing sample sizes for the two algorithms.

5. Discussion of Work

From the conducted experiment, the observation was made to infer the mean accuracy, mean precision, mean false positive, and mean false negative for the proposed EDRL algorithm; additionally, these parameters were compared with the CNN algorithm, in order to determine the performance measures of the entire algorithms, while considering the dataset with the data recorded at 54,000. From Figure 3, Figure 4, Figure 5 and Figure 6, it can be observed that the proposed EDRL algorithm outperformed on mean accuracy, mean precision, mean false positive, and mean false negative, when compared to the CNN algorithms performance measures. Initially the dataset contained 54,000. The dataset was collected from ISCXVPN2016 [20] for VPNs and non-VPNs. The size of the dataset is almost 15 GB of ARFF file format, with a set of attributes with instance sharing. A total of 80% of data were taken for training purposes, and 20% data were used for testing purposes.
The EDRL algorithm consumed less storage and computation time than the CNN algorithms, and this reduced consumption is an enhancement of the deep reinforcement learning algorithm EDRL. The network traffic prediction was achieved with a mean accuracy of 97.20% using the EDRL algorithm, which is higher than the CNN algorithm’s mean accuracy of 93.055%. Further, for the network traffic prediction, mean precision was achieved at 97.343% and 93.972%, respectively, for the EDRL and CNN algorithms. The mean false positive measures was achieved at 2.657% and 6.325%, respectively, for the EDRL and CNN algorithms. Finally, the mean false negative was achieved at 2.527% and 5.675%, respectively, for EDRL and CNN algorithms. The confusion matrix of algorithms were plotted to predict the traffic prediction and compared with the trained and testing dataset. When the numbers of iterations were increased, there was linear growth in the accuracy, precision, false positive, and false negative for the EDRL, which were better than that seen with the CNN algorithm. For the standard deviation, the standard error was tabulated for the algorithms. The significant difference was slightly better for the ERDL algorithm than the CNN algorithm. Finally, the performance measuring parameters were also measured with different samples of the dataset for EDRL and CNN, and it was observed that EDRL performed better than the CNN algorithms, in regard to mean accuracy, mean precision, mean false positive, and mean false negative.
The previous studies were measuring prediction accuracy using the KNN and RF [13,20,21,22], and the accuracy was achieved at 72.08% and 90.53%, respectively, with EDONKEY application network traffic. The artificial neural network [23] was used to measure the network traffic, with performance measures regarding the capacity of the network and traffic loss. Further, other studies claimed equivalent accuracy or more or less for traffic prediction, based on the types of network traffic. The application-oriented network traffic achieved comparatively less accuracy than the EDRL for network traffic of Amazon using the ANN algorithm (95.00%) [20,21], compared to 80.00% for EDONKEY traffic and 78.00% for FTP_CONTROL. For the network traffic of FTP and P2P, accuracy and precision were achieved at 94% and 90%, respectively, for the KNN algorithm [13,22]. The CNN-based application identification task accuracy was achieved at 94.00% [23]; when compared to the EDRL algorithm, the accuracy of application traffic classification was less with the UNB ISCX VPN-nonVPN dataset. Another work used SVM for network traffic and classification. The accuracy measure was reached at 94.2% [24]. This work has tested the SVM’s versions, as well, in order to measure the accuracy and precision. A comparative performance measure of various ML and DL algorithms were measured for KNN, RF, neural network (NN), and naïve Bayes (NB) as 79.6%, 84.8%, 84.6%, and 87.6%, respectively. This work has used a real time dataset of the orange platform of Nigerian University [1,25]. Table 13 presents various ML- and DL-based algorithm’s accuracy comparisons with the proposed EDRL algorithm.
The novelty of this work is that the EDRL gained reward-based output, comparatively more than existing the ML and DL algorithms. The reward-based decision-making policy is gained from the EDRL algorithm. Here, the accuracy and precision, as well as false positive and negative rates, were comparatively high, as discussed in Section 4 and Figure 3, Figure 4, Figure 5 and Figure 6.
The factors affecting the network traffic predictions are round trip delay time, application type with dynamic nature, payload of network data, etc. The proposed algorithm EDRL maintains some limitations, if the above factors are to be incorporated. Further, in the future, it is essential to incorporate limitations when enhancing the proposed work for the automation of network traffic predictions, as well.

6. Conclusions and Future Works

This research was carried out to predict the fine-tuned accuracy and precision, as well as the false positive and negative rates, for the network traffic of various types and classes of networks, in order to fill the research gap regarding the lack of algorithms for measures. An EDRL algorithm was coined, in order to get best results for taking parameters. The conducted experiments’ results illustrate that the EDRL algorithm is best with mean accuracy, mean precision, mean false positive, and mean false negative in both the numerical and graphical results.
This work could be used in various applications, such as network traffic prediction applications related to surveillance, sensitive types of traffic predictions, and other commercial applications, in order to monitor application traffic. This work could be extended for the automation of network traffic prediction by introducing an extended algorithm of EDRL for real time network traffic of surveillance and FTP traffic. That would be best, in order to deal with real time network traffic in the future.

Author Contributions

Conceptualization, N.M.B. and M.A.; methodology, M.A. and M.H.A.; software, N.M.B.; validation, M.H.A. and P.U.; formal analysis, N.M.B. and M.H.A.; investigation, P.U.; resources, N.M.B.; writing—original draft preparation, N.M.B. and M.A.; writing—review and editing, M.H.A. and P.U.; visualization, M.A.; supervision, M.A.; project administration, M.H.A.; funding acquisition, P.U. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Suranaree University of Technology (SUT) Research and Development Funds, as well as the Thailand Science Research and Innovation (TSRI).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are thankful to Raju Kannadasan, Department of Electrical and Electronics Engineering, Sri Venkateswara College of Engineering, Sriperumbudur, 602117, India, for his support in organizing this work successfully.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Adekitan, A.I.; Abolade, J.; Shobayo, O. Data mining approach for predicting the daily Internet data traffic of a smart university. J. Big Data 2019, 6, 11. [Google Scholar] [CrossRef]
  2. Alsharif, M.H.; Younes, M.K.; Kim, J. Time Series ARIMA Model for Prediction of Daily and Monthly Average Global Solar Radiation: The Case Study of Seoul, South Korea. Symmetry 2019, 11, 240. [Google Scholar] [CrossRef] [Green Version]
  3. Alsharif, M.H.; Kelechi, A.H.; Yahya, K.; Chaudhry, S.A. Machine Learning Algorithms for Smart Data Analysis in Internet of Things Environment: Taxonomies and Research Trends. Symmetry 2020, 12, 88. [Google Scholar] [CrossRef] [Green Version]
  4. Wang, W.; Bai, Y.; Yu, C.; Gu, Y.; Feng, P.; Wang, X.; Wang, R. A network traffic flow prediction with deep learning approach for large-scale metropolitan area network. In Proceedings of the 2018 IEEE/IFIP Network Operations and Management Symposium (NOMS 2018), Taipei, Taiwan, 23–27 April 2018; pp. 1–9. [Google Scholar] [CrossRef]
  5. Cortez, P.; Rio, M.; Rocha, M.; Sousa, P. Internet Traffic Forecasting using Neural Networks. In Proceedings of the 2006 IEEE International Joint Conference on Neural Network Proceedings, Vancouver, BC, Canada, 16–21 July 2006; pp. 2635–2642. [Google Scholar] [CrossRef] [Green Version]
  6. Narejo, S.; Pasero, E. An Application of Internet Traffic Prediction with Deep Neural Network. In Multidisciplinary Approaches to Neural Computing; Esposito, A., Faudez-Zanuy, M., Morabito, F., Pasero, E., Eds.; Smart Innovation, Systems and Technologies; Springer: Cham, Switzerland, 2018; Volume 69. [Google Scholar] [CrossRef]
  7. Vinayakumar, R.; Soman, K.P.; Poornachandran, P. Applying deep learning approaches for network traffic prediction. In Proceedings of the 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Udupi, India, 13–16 September 2017; pp. 2353–2358. [Google Scholar] [CrossRef]
  8. Park, J.; Yi, D.; Ji, S. Analysis of Recurrent Neural Network and Predictions. Symmetry 2020, 12, 615. [Google Scholar] [CrossRef]
  9. Oliveira, T.P.; Barbar, J.S.; Soares, A.S. Computer network traffic prediction: A comparison between traditional and deep learning neural networks. Int. J. Big Data Intell. 2016, 3, 28. [Google Scholar] [CrossRef]
  10. Yoo, W.; Sim, A. Network bandwidth utilization forecast model on high bandwidth networks. In Proceedings of the 2015 International Conference on Computing, Networking and Communications (ICNC), Garden Grove, CA, USA, 16–19 February 2015; pp. 494–498. [Google Scholar] [CrossRef] [Green Version]
  11. Chen, A.; Law, J.; Aibin, M. A Survey on Traffic Prediction Techniques Using Artificial Intelligence for Communication Networks. Telecom 2021, 2, 518–535. [Google Scholar] [CrossRef]
  12. Rezaei, S.; Liu, X. Multitask learning for network traffic classification. In Proceedings of the International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 3–6 August 2020; pp. 1–9. [Google Scholar]
  13. Lotfollahi, M.; Zade, R.S.H.; Siavoshani, M.J.; Saberian, M. Deep packet: A novel approach for encrypted traffic classification using deep learning. Soft Comput. 2020, 24, 1999–2012. [Google Scholar] [CrossRef] [Green Version]
  14. Lopez-Martin, M.; Carro, B.; Sanchez-Esguevillas, A.; Lloret, J. Network traffic classifier with convolutional and recurrent neural networks for internet of things. IEEE Access 2017, 5, 42–50. [Google Scholar] [CrossRef]
  15. Zeng, Y.; Gu, H.; Wei, W.; Guo, Y. Deep-Full-Range: A deep learning based network encrypted traffic classification and intrusion detection framework. IEEE Access 2019, 7, 182–190. [Google Scholar] [CrossRef]
  16. Labayen, V.; Magana, E.; Morato, D.; Izal, M. Online classification of user activities using machine learning on network traffic. Comput. Netw. 2020, 181, 557–569. [Google Scholar] [CrossRef]
  17. Chang, L.-H.; Lee, T.-H.; Chu, H.-C.; Su, C.-W. Application-based online traffic classification with deep learning models on sdn networks. Adv. Technol. Innov. 2020, 5, 216–229. [Google Scholar] [CrossRef]
  18. Gil, G.D.; Lashkari, A.H.; Mamun, M.; Ghorbani, A.A. Characterization of Encrypted and VPN Traffic Using Time-Related Features. In Proceedings of the 2nd International Conference on Information Systems Security and Privacy (ICISSP 2016), Rome, Italy, 19–21 February 2016; pp. 407–414. [Google Scholar]
  19. Browne, C.B.; Powley, E.; Whitehouse, D.; Lucas, S.M.; Cowling, P.I.; Rohlfshagen, P.; Tavener, S.; Perez, D.; Samothrakis, S.; Colton, S. A survey of Monte Carlo tree search methods. IEEE Trans. Comput. Intell. AI Games 2012, 4, 1–43. [Google Scholar] [CrossRef] [Green Version]
  20. Ahmed, A.A.; Agunsoye, G. A Real-Time Network Traffic Classifier for Online Applications Using Machine Learning. Algorithms 2021, 14, 250. [Google Scholar] [CrossRef]
  21. Troia, S.; Alvizu, R.; Zhou, Y.; Maier, G.; Pattavina, A. Deep Learning-Based Traffic Prediction for Network Optimization. In Proceedings of the 2018 20th International Conference on Transparent Optical Networks (ICTON), Bucharest, Romania, 1–5 July 2018; pp. 1–4. [Google Scholar] [CrossRef] [Green Version]
  22. Pacheco, F.; Exposito, E.; Gineste, M.; Baudoin, C.; Aguilar, J. Towards the deployment of Machine Learning solutions in network traffic classification: A systematic survey. IEEE Commun. Surv. Tutor. 2018, 21, 1988–2014. [Google Scholar] [CrossRef] [Green Version]
  23. Mohammed, A.R.; Mohammed, S.A.; Shirmohammadi, S. Machine Learning and Deep Learning Based Traffic Classification and Prediction in Software Defined Networking. In Proceedings of the 2019 IEEE International Symposium on Measurements & Networking (M&N), Catania, Italy, 8–10 July 2019; pp. 1–6. [Google Scholar] [CrossRef]
  24. Dong, S. Multi class SVM algorithm with active learning for network traffic classification. Expert Syst. Appl. 2021, 176, 114885. [Google Scholar] [CrossRef]
  25. Balamurugan, N.M.; Kannadasan, R.; Alsharif, M.H.; Uthansakul, P. A Novel Forward-Propagation Workflow Assessment Method for Malicious Packet Detection. Sensors 2022, 22, 4167. [Google Scholar] [CrossRef] [PubMed]
Figure 1. DNN multilayer perceptron for network traffic processing.
Figure 1. DNN multilayer perceptron for network traffic processing.
Sensors 22 05006 g001
Figure 2. An EDRL architecture for network traffic analysis.
Figure 2. An EDRL architecture for network traffic analysis.
Sensors 22 05006 g002
Figure 3. Mean accuracy comparison of EDRL and CNN, including error rates.
Figure 3. Mean accuracy comparison of EDRL and CNN, including error rates.
Sensors 22 05006 g003
Figure 4. Precision comparison of EDRL and CNN algorithms with the error mean measure.
Figure 4. Precision comparison of EDRL and CNN algorithms with the error mean measure.
Sensors 22 05006 g004
Figure 5. The false positive comparison between the EDRL and CNN algorithms with stand error bars with SD and confidence interval.
Figure 5. The false positive comparison between the EDRL and CNN algorithms with stand error bars with SD and confidence interval.
Sensors 22 05006 g005
Figure 6. False negative comparison between EDRL and CNN algorithm with stand error bars with SD and confidence interval.
Figure 6. False negative comparison between EDRL and CNN algorithm with stand error bars with SD and confidence interval.
Sensors 22 05006 g006
Table 1. Various classes and methods of machine and deep learning techniques.
Table 1. Various classes and methods of machine and deep learning techniques.
ClassMethodLearning Technique
LSTMDiscriminativeSupervised
CNN
RNN
MLP
Table 2. Data input, policy relationship for types of ML and DL algorithms.
Table 2. Data input, policy relationship for types of ML and DL algorithms.
MethodData Input
Know AnswerPolicy/Problems
Supervised learningLearned output with supervisionLearning reward-based output with supervision
Reinforcement learning methodMaximize reward-based outputFeedback trained maximized reward-based output
Deep learning methodDeep learning-based outputDeep learning and feedback trained-based output
Deep reinforcement learning methodDeep-based maximize reward-based outputDeep and feedback trained-based maximize reward-based output
Enhanced deep reinforcement learning methodAccurate deep-based maximize reward-based outputAccurate output, based on deep and feedback trained maximize reward
Table 3. Network data features and its notation.
Table 3. Network data features and its notation.
Feature IndexNotationFeature Description
f1avg_seg_szAverage size of segment
f2win_szWindow size
f3r_t_tRound Trip delay Time
f4var_packVariance in packets
f5Act_dt_pktActual data packet
f6clt_pnClient port number
f7svr_pnServer port number
Table 4. Classes of network traffic.
Table 4. Classes of network traffic.
Class IndexNotationClass DescriptionApplications
c1www_pktwww packetGeneral browsing data
c2p2p_pktP2P network packetTorrent streaming
c3ml_pktMail service packetSMTP, POP, MIME, IMAP
c4db_pktDatabase packetSQL net
c5mul_pktMultimedia packetVideo storage server YouTube
Table 5. Group statistics for accuracy comparison of EDRL vs. CNN algorithms to measure mean, standard deviation, and standard error mean.
Table 5. Group statistics for accuracy comparison of EDRL vs. CNN algorithms to measure mean, standard deviation, and standard error mean.
Group Statistics
AlgorithmNMeanStd. DeviationStd. Error Mean
AccuracyEDRL1097.2001.711560.538
CNN1093.0552.298350.727
Table 6. Comparison of the independent sample t-test parameters of EDRL and CNN algorithms.
Table 6. Comparison of the independent sample t-test parameters of EDRL and CNN algorithms.
Independent Samples Test
Levene’s Test for Equality of Variancest-Test for Equality of Means
FSig.tdfSig. (2-Tailed)Mean DifferenceStd. Error Difference95% Confidence Interval of the Difference
LowerUpper
AccuracyEqual variances assumed1.1110.3064.519180.0004.0950.906192.1915.999
Equal variances not assumed 4.51916.6340.0004.0950.906192.1806.010
Table 7. Group statistics for precision comparison of EDRL vs. CNN algorithms to measure mean, standard deviation, and standard error mean.
Table 7. Group statistics for precision comparison of EDRL vs. CNN algorithms to measure mean, standard deviation, and standard error mean.
Group Statistics
AlgorithmNMeanStd. DeviationStd. Error Mean
PrecisionEDRL1097.3431.5190.480
CNN1093.9722.4030.760
Table 8. Comparison of independent samples t-test parameters of EDRL and CNN algorithms for precision.
Table 8. Comparison of independent samples t-test parameters of EDRL and CNN algorithms for precision.
Independent Samples Test
Levene’s Test for Equality of Variancest-Test for Equality of Means
FSig.tdfSig. (2-Tailed)Mean DifferenceStd. Error Difference95% Confidence Interval of the Difference
LowerUpper
PrecisionEqual variances assumed2.3510.1434.295180.0003.8610.8991.9725.750
Equal variances not assumed 4.29515.200.0013.8610.8991.9475.780
Table 9. Group statistics for false positive comparison of EDRL vs. CNN algorithms to measure the mean, standard deviation, and standard error mean.
Table 9. Group statistics for false positive comparison of EDRL vs. CNN algorithms to measure the mean, standard deviation, and standard error mean.
Group Statistics
AlgorithmNMeanStd. DeviationStd. Error Mean
False positiveEDRL102.6571.853350.586
CNN106.3252.190630.693
Table 10. Comparison of independent samples t-test parameters of EDRL and CNN algorithms for false positive.
Table 10. Comparison of independent samples t-test parameters of EDRL and CNN algorithms for false positive.
Independent Samples Test
Levene’s Test for Equality of Variancest-Test for Equality of Means
FSig.tdfSig. (2-Tailed)Mean DifferenceStd. Error Difference95% Confidence Interval of the Difference
LowerUpper
False positiveEqual variances assumed0.3720.550−4.042180.001−3.6680.907−5.574−1.762
Equal variances not assumed −4.04217.510.001−3.6680.907−5.578−1.758
Table 11. Group statistics for false negative comparison of EDRL vs. CNN algorithms to measure the mean, standard deviation, and standard error mean.
Table 11. Group statistics for false negative comparison of EDRL vs. CNN algorithms to measure the mean, standard deviation, and standard error mean.
Group Statistics
AlgorithmNMeanStd. DeviationStd. Error Mean
False negativeEDRL102.52701.227340.38812
CNN105.67501.99200.61643
Table 12. Comparison of independent sample t-test parameters of the EDRL and CNN algorithms for the false negative.
Table 12. Comparison of independent sample t-test parameters of the EDRL and CNN algorithms for the false negative.
Independent Samples Test
Levene’s Test for Equality of Variancest-Test for Equality of Means
FSig.tdfSig. (2-Tailed)Mean DifferenceStd. Error Difference95% Confidence Interval of the Difference
LowerUpper
False negativeEqual variances assumed3.1130.095−4.826180.00−3.5980.746−5.164−2.032
Equal variances not assumed −4.82614.870.00−3.5980.746−5.188−2.008
Table 13. Various ML and DL accuracy measure comparisons with the proposed EDRL algorithm.
Table 13. Various ML and DL accuracy measure comparisons with the proposed EDRL algorithm.
Work NameAlgorithm UsedAccuracy
EDONKEY application network traffic [20]KNN and RF72.08% and 90.53%
FTP_CONTROL [20,22]ANN78.00%
The network traffic of FTP and P2P [13,23]KNN94%
The CNN based application identification task [21]CNN94%
Traffic classification was less with UNB ISCX VPN-Non-VPN dataset [24]SVM94.2%
Orange platform of Nigerian University [1]KNN, RF, NN, and NB79.6%, 84.8%, 84.6%, and 87.6%
Internet traffic of different applicationsThe proposed EDRL algorithm97.20%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Balamurugan, N.M.; Adimoolam, M.; Alsharif, M.H.; Uthansakul, P. A Novel Method for Improved Network Traffic Prediction Using Enhanced Deep Reinforcement Learning Algorithm. Sensors 2022, 22, 5006. https://doi.org/10.3390/s22135006

AMA Style

Balamurugan NM, Adimoolam M, Alsharif MH, Uthansakul P. A Novel Method for Improved Network Traffic Prediction Using Enhanced Deep Reinforcement Learning Algorithm. Sensors. 2022; 22(13):5006. https://doi.org/10.3390/s22135006

Chicago/Turabian Style

Balamurugan, Nagaiah Mohanan, Malaiyalathan Adimoolam, Mohammed H. Alsharif, and Peerapong Uthansakul. 2022. "A Novel Method for Improved Network Traffic Prediction Using Enhanced Deep Reinforcement Learning Algorithm" Sensors 22, no. 13: 5006. https://doi.org/10.3390/s22135006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop