Next Article in Journal
Reliable Multicast Based on Congestion-Aware Cache in ICN
Next Article in Special Issue
Modeling and Prediction of Daily Traffic Patterns—WASK and SIX Case Study
Previous Article in Journal
Comparison of the Performance of the Memristor Models in 2D Cellular Nonlinear Network
Previous Article in Special Issue
Modelling and Optimization of Multi-Service Optical Switching Networks with Threshold Management Mechanisms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Classification and Regression Approaches for Optical Network Traffic Prediction

Faculty of Electronics, Wrocław University of Science and Technology, Wybrzeże Wyspiańskiego 27, 50-370 Wrocław, Poland
*
Authors to whom correspondence should be addressed.
Electronics 2021, 10(13), 1578; https://doi.org/10.3390/electronics10131578
Submission received: 27 May 2021 / Revised: 22 June 2021 / Accepted: 25 June 2021 / Published: 30 June 2021
(This article belongs to the Special Issue Telecommunication Networks)

Abstract

:
Rapid growth of network traffic causes the need for the development of new network technologies. Artificial intelligence provides suitable tools to improve currently used network optimization methods. In this paper, we propose a procedure for network traffic prediction. Based on optical networks’ (and other network technologies) characteristics, we focus on the prediction of fixed bitrate levels called traffic levels. We develop and evaluate two approaches based on different supervised machine learning (ML) methods—classification and regression. We examine four different ML models with various selected features. The tested datasets are based on real traffic patterns provided by the Seattle Internet Exchange Point (SIX). Obtained results are analyzed using a new quality metric, which allows researchers to find the best forecasting algorithm in terms of network resources usage and operational costs. Our research shows that regression provides better results than classification in case of all analyzed datasets. Additionally, the final choice of the most appropriate ML algorithm and model should depend on the network operator expectations.

1. Introduction

Quick and global development of network technologies such as the Internet of things, 5G, or cloud computing causes instant growth of endpoint devices. According to the Cisco Annual Internet Report, the number of Internet users will grow from 3.9 billion in 2018 to 5.3 billion in 2023 [1]. Moreover, the recent Nokia report [2] presents and discusses various network traffic trends in 2020, showing a significant growth of network traffic due to the COVID-19 pandemic. To overcome the possible capacity crunch problem in the Internet, network operators build and incessantly improve backbone networks utilizing various optical technologies [3,4]. However, constantly growing network traffic presents new challenges to the network operators. To improve the performance of future optical networks compared to conventional mechanisms currently used in optical networks, a concept of a cognitive optical network [5] has been proposed. In more detail, a cognitive optical network is a network with a cognitive process that can monitor current network conditions and then adjust the network operation to those conditions. The cognitive process, which uses history to improve performance, usually employs machine learning (ML) algorithms [6]. ML techniques, by analyzing and finding dependencies in historical data, e.g., traffic flows, learn and apply gained knowledge to make decisions in the future. More information about the application of ML techniques in optical networks can be found in [7,8,9,10,11].
Knowledge about future traffic in a network allows operators to optimize resource usage, provide better quality of service for users [12], reduce the cost of network operation, or detect anomalies in the traffic dataflow [13]. There are four forecasting types in terms of horizon time [14]: (i) real-time, in case of on-line traffic prediction, (ii) short-term, which forecasts up to several hours of future traffic, (iii) middle-term, predicts a few days traffic, and (iv) long-term, which can predict years of future network traffic. Knowledge obtained from each of the forecasting strategies can be used as valuable information for different network optimization tasks, e.g., traffic flow control, network operational cost reduction, anomalies detection, or physical network expansion. In our work, we focus on a short-term traffic prediction; however, our approach can be successfully applied to all forecasting types.
In this paper, we consider an optical network in which transmission between all physically connected nodes occurs continuously. Traffic flows between pairs of nodes can be modeled as continuous and regular flows, which change in time. We use ML classifiers and regressors to forecast future traffic. As a consequence of everyday user activities, working tasks, etc., some daily and weekly patterns in network traffic can be observed [15]. Because of this, some dependencies and periodicity are created over traffic flows. ML algorithms locate, analyze, and learn those dependencies to forecast future network traffic flows.
A single optical channel supported by a single transceiver can carry a fixed amount of data. As a result, the information required to establish a connection is the number of optical channels required to carry a transition, for instance, the number of 100 Gbps channels in a WDM (wavelength division multiplexing) network. Additionally, most of the transport network technologies such as an Optical Transport Network and various versions of Ethernet are also provisioned in some granularities of the bitrate. Therefore, during traffic prediction we focus on predicting traffic levels rather than the exact traffic volume.
This work is a continuation and extension of our recent paper [16]. The contribution of this paper is threefold. First, we compare two supervised ML approaches: classification, where we classify traffic into traffic levels, and regression, where we first forecast the value of the traffic flow (bitrate in a given time point), and then based on the obtained result, the traffic level is calculated by rounding up the obtained prediction to the closest traffic level. Second, we propose and test for various classifiers and regressors four different ML models based on various features. Last, we introduce a new quality metric, which allows the network operator to compare forecasting algorithms concerning the usage of network resources and operational costs.
The rest of this paper is organized as follows. Section 2 presents related works from a field of study. Next, in Section 3, the dataset creation method is depicted. The network model, ML approach, ML models, and evaluation metrics are described in Section 4. Section 5 shows numerical results and a brief description of them. At the end, final conclusions are presented in Section 6.

2. Related Works

Typically, the problem of forecasting network traffic is formulated as a time series problem. A large majority of works in the field use approaches based on an autoregressive integrated moving average (ARIMA) and its numerous variations, as well as ML techniques [17], to solve this. Authors in [18] compared ARIMA, Holt-Winters, and neural network algorithms for forecasting the amount of traffic in TCP/IP-based networks. In [19], ARIMA and SARMIA models were used for short-term and long-term future traffic volume predictions. As a result, the required bandwidth was reduced by 18.9%. In Reference [20], the allocation of data center traffic with and without traffic prediction was compared. The performance of all traffic allocation algorithms was improved by using ML techniques. In Reference [21], the authors used a graph convolutional generative adversarial networks model to predict burst events in an optical network. The proposed method outperformed the LSTM reference method.
Besides the traffic prediction problem, ML techniques can be successfully employed for other purposes in optical networks. In Reference [22], the authors used ML techniques for the problem of fault localization in optical networks. They successfully localized single-link failures using a Gaussian process classifier trained on data that described the network state upon current and past failure incidents. The presented approach achieves high localization accuracy ranges from 91% to 99%. A similar problem, connected with failure localization, is considered in [23]. Authors presented an ML system for detection and identification of equipment failures in optical networks. They tested several ML methods, a random forest, a neural network with a single hidden layer, and different variants of the support vector machine. As a result, accuracy above 98% was obtained. Authors in [24] used ML classification techniques such as decision tree and naïve Bayes discretization to classify traffic flows into mouse flows (occur frequently but carry a small number of bytes) and elephant flows (occur occasionally but have a huge number of bytes). The paper presented classifiers performance in terms of accuracy and classification speed. Another well-examined issue is estimation of the quality of service. In Reference [25], the authors introduced an alien wavelength performance monitoring technique and ML quality of service estimation for lightpath provisioning of intradomain and interdomain traffic. Obtained results reached up to 95% of prediction accuracy. Authors in [26] proposed an ML regression approach to predict the quality of transmission of an unestablished lightpath. They used a neural network as a base algorithm for prediction. The evaluation was carried out considering the generalized signal-to-noise ratio metric. In Reference [27], the authors presented an intelligent module in the form of an ML application using deep learning modeling. The system described in this publication uses a neutral network for solving the task of proactive network monitoring for the security and protection of computing infrastructures.
Despite the fact that many works have presented promising results, application of ML methods to network problems is still in its early stage [8]. Thus, there is a high demand for exploring the topic of ML usage for solving network problems [7]. Most of the related works implemented the traffic prediction task as a prediction of exact traffic bitrates. In this paper, we formulate the prediction problem as a prediction of fixed traffic levels. Such a concept follows from characteristics of optical networks and other transport network technologies. According to the best of our knowledge, the only paper that focuses on the prediction of traffic levels in optical networks is our previous paper [16], where we presented some preliminary results on this problem. To fill the research gap, in this paper we introduce and examine two supervised ML approaches: classification and regression. We also present four ML models used to test a wide set of various ML algorithms. All experiments were conducted using datasets, which were generated based on real traffic characteristics. The results reported in the next sections prove that the proposed approaches outperform the methods described in [16].

3. Datasets

This section describes the method of dataset generation together with traffic flows used in research.

3.1. Traffic Generator

Datasets used in the research were created with the use of the custom traffic generator. The overall shape of output traffic reflects the real-world traffic based on the time-varying data taken from the Internet Exchange Point in Seattle, Washington (SIX). Traffic statistics for this exchange point are available publicly at its official website (https://www.seattleix.net/ (accessed on 19 December 2019)). The data was gathered for a one month interval, namely, from 24th of October to 19th of December 2019 from rrd traffic files which were uploaded periodically with a granulation time of 5 min per time slot. Every week the recent data was downloaded, read and stored. This data was then applied as an input for the described traffic generator as a general shape of time-varying traffic for a month-span interval.
The generator is an ensemble of smaller requests generators that represent several web services, each with an assigned share [1] and their own properties such as a set of combinations of stochastic processes with assigned individual parameters and contribution scales for each of them. Stochastic processes which are assumed are Poisson process (PP), Poisson Pareto burst process (PPBP) [28], and a constant traffic (CT) with uniformly distributed random offset. Considered web services are:
  • Internet video with a share of 51% of overall bitrate made of two different PPs, PPBP and CT;
  • IP VOD with a share of 22% of overall bitrate made of a single PP;
  • Web data with a share of 18% of overall bitrate made of the different PPs;
  • File sharing with a share of 8% of overall bitrate made of a single CT;
  • Gaming with a share of 1% of overall bitrate made of a single CT.
Such differentiation of traffic characteristics of given web services tries to project the diverse nature of the Internet traffic over time. The overall bitrate of generated requests varies through time depending on the provided traffic characteristics. This traffic is distributed between nodes of a given network topology. In this research, we considered the Euro28 backbone network [29] with 28 nodes, 84 links, and an average link length of 625 km. The distribution of traffic between nodes is inversely proportional to the distance between each pair of nodes. The output traffic is a series of tuples of 756 numbers, representing the volume of traffic for each pair of nodes in each time slot. The expected summed bitrate for each time slot reflects the overall required bitrate for the considered network and keeps the time-varying trends similar to the provided traffic characteristics. The division of traffic between nodes provides a more insightful perspective on the traffic in the network and allows for focusing on particular nodes.

3.2. Datasets

Note that, to measure the fluctuation of various traffic flows obtained from the traffic generator described above, we used the mean absolute percentage error (MAPE) This describes how the values of one function differ from values of the base function, which in this case was the original SIX traffic profile used to generate traffic for particular node pairs, as described in Section 3.1. To calculate MAPE of a single traffic flow from the generated data, as the base function, we used SIX bitrate flows and normalized them to a common range of considered flow bitrate values. In other words, MAPE indicates how the considered traffic flow differs from the SIX traffic. For further experiments, we selected the following 5 datasets with various values of the MAPE parameter:
  • dataset_1–traffic flow between a single nodes pair, MAPE equal to 3.39%,
  • dataset_2–traffic flow between a single nodes pair, MAPE equal to 8.21%,
  • dataset_3–traffic flow between a single nodes pair, MAPE equal to 13.35%,
  • dataset_4–whole traffic incoming to a single network node, MAPE equal to 1.33%,
  • dataset_5–original SIX traffic, MAPE equal to 0%.
Figure 1 presents visual representation of datasets used during experiments. The vertical axis presents normalized bitrate. Because datasets’ bitrates ranges differ from each other, for the sake of visualization, datasets bitrates were normalized to the same range. This clearly shows that traffic flows differ from each other in terms of their fluctuation.

4. Proposed Models and Algorithms

The following section briefly reviews the background concepts of the considered optical network, ML approaches, ML models’ creation, evaluation metrics, and configurations of conducted experiments.

4.1. Network Model

Let n be the number of all nodes in the network. In our work, we assumed that the network is modeled as a directed graph G = N , E , where N represents the set of n physical nodes and E represents the set of links connecting them (reflecting the set of physical links between nodes). The time scale of the network operation was divided into time intervals of the same size (e.g., 5 min). For consecutive time intervals, traffic volumes (bitrates) related to a single pair of nodes or whole traffic going through a single node created continuous and regular flows.

4.2. ML Approach

Usually, the task of forecasting the future traffic in the optical network is to predict the shape of the flow over a selected period of time. Because of that, such a problem is considered as a time series problem that can be solved based on regression solutions. In our work, traffic prediction was realized by an indication of a traffic level in a particular time interval (rather than the exact traffic bitrate). Motivation behind such an approach are optical technologies such as wavelength division multiplexing (WDM) and elastic optical networks (EON) [3,4,30]. In more detail, in optical technologies, each optical channel (i.e., lightpath) can carry a fixed amount of data (fixed bitrate) based on the characteristics of the used transceiver. For example, supposing that a single optical channel supported by a single transceiver can carry 100 Gbps. To serve traffic with bitrates of 110 Gbps or 190 Gbps, regardless of the fact that the bitrates are different, we need to establish two optical channels (two transceivers) for transmission, since we need 200 Gbps to provision each of the considered bitrates. Therefore, we focused on the number of optical channels (defined by provisioned bitrate) that are needed to send data, instead of the exact value of the traffic bitrate. Another motivation behind considering prediction of traffic levels were Ethernet standards, very popular in optical networks and offering transmission of defined bitrate levels such as 100 Gb Ethernet [4].
Figure 2 illustrates the process of defining traffic levels for traffic flows. The blue line represents real bitrate values and the green line traffic levels that correspond to them. Possible classes (determined by a single channel width) are represented by grey horizontal lines. In such cases, traffic forecasting can be carried out by predicting future green line levels. We examined two different supervised ML approaches for defining traffic levels:
1.
Classification approach–a problem is considered as a classification task, where possible classes are determined by traffic levels (multiples of a single channel bitrates). The following classifiers were tested in this approach: linear discriminant analysis (LDA), k neighbors classifier (KNN), Gaussian naïve Bayes (GNB), decision tree (DT).
2.
Regression approach–first, using a selected regression algorithm, the exact value of bitrate in a particular time interval is predicted. Next, based on the obtained results, the traffic levels in time intervals are calculated by rounding up the predicted bitrate to the closest traffic level. The following regressors were tested in this approach: linear regression (LR), Passive Aggressive Regressor (PAR), k neighbors regressor (KNNR), multi-layer perceptron regressor (MLPR).
Figure 2. Traffic levels.
Figure 2. Traffic levels.
Electronics 10 01578 g002
Particular traffic flows change over time, thus the data used in the experiment can be characterized as data streams [31]. Therefore, we did not use a classic k-fold cross validation and training datasets were created based on consecutive iterations (time intervals). In more detail, the whole dataset for the experiment consisted of traffic flows information lasting about 58 days. We took the first 28 days to create a training dataset and the last 30 days to create a testing dataset.

4.3. ML Models

The training dataset was created based on historical traffic flows. Selection of suitable dataset instances’ features is of utmost importance for solving ML problems. Figure 3 presents the autocorrelation of dataset_1 traffic flow. In detail, it determines how much the value of a particular bitrate depends on bitrate values from previous time intervals. The further the distance from zero is, the greater correlation occurs. Vertical axis shows autocorrelation values and horizontal axis contains time intervals. It can be clearly seen that the highest correlations occurred for the closest preceding iterations. Additionally, a meaningful correlation was observed for time intervals that occurred every 12 h (every 144 time intervals, since a single time interval represented 5 min). In addition, positive correlations (greater than zero, occurred every 24 h, which refers to 288 time intervals) were higher than negative correlations. Variability of traffic flows was correlated with time, therefore information about the minute during the day and the day during the week when traffic volume occurred was also essential for predicting tasks.
Let window denote the number of preceding time intervals, of which bitrates are taken as features. We propose four different ML models based on different selection of features:
1.
window = 1, (i.e., only the previous time interval is considered as an input feature);
2.
window = 10, (i.e., the 10 previous time intervals are included in the model);
3.
window = 10, minute, day, (i.e., additionally, the information on the time stamp of the considered time interval is included in the model, the time stamp includes day which denotes the day of week and minute that denotes the number of minutes from the beginning of the day);
4.
window = 10, minute, day, traffic values (bitrates) from previous 24 h, 48 h, …, 336 h (two weeks) time intervals.
The number of features of each instance in the training dataset depends on the selected ML model. Figure 4 depicts the model creation process for the classification task. The feature list includes information about the day, minute and bitrate in the previous time interval (window = 1). Possible traffic levels (classes) are multiples of 100. First window iterations were omitted during the model creation, because there was no information about all of their previous time intervals. During the classification, the classifier took instance feature values and as a result returned the bitrate level for the future time interval.
Figure 5 presents a diagram illustrating the implemented system. At the input, the system took the number of the considered model, historical traffic flows, and list of tested ML algorithms. During the data preprocessing phase, datasets for training and testing were created. The number of instance features depended on the considered model. Datasets were created based on historical traffic flows, as Figure 4 presents. Next, ML algorithms were trained and tested. At the output, the system returned an evaluation metric value for each ML algorithm.

4.4. Quality Metric

Typically, the base metric to compare ML algorithms is accuracy. In the considered problem, this carried information about the percentage number of correctly predicted traffic levels; however, such result analysis omitted information about too low and too high traffic level predictions. This knowledge can be significant for a network operator, because it can result in better resources usage and cost management. For instance, underprediction of the future traffic can result in future allocation of too few optical channels and, as a consequence, the blocking of some traffic, which can trigger the operator having to pay some additional penalties due to SLA agreements. In turn, overprediction of the future traffic has less severe consequences, for instance, such as greater power consumption; however, it does not yield traffic blocking.
Therefore, to gain a better ML algorithms assessment specific for the analyzed data related to network traffic, we propose a new evaluation metric called Traffic Level Prediction Quality (TLPQ), which is calculated based on the key prediction variants. We distinguish the following prediction variants:
a.
predicted traffic level is correct;
b.
predicted traffic level is too low and it differs from correct level by one class, i.e., underprediction by one traffic level;
c.
predicted traffic level is too low and it differs from correct level by more than one class, i.e., underprediction by more than one traffic level;
d.
predicted traffic level is too high and it differs from correct level by one class, i.e., overprediction by one traffic level;
e.
predicted traffic level is too high and it differs from correct level by more than one class, i.e., overprediction by more than one traffic level.
The equation below indicates the way to calculate the TLPQ metric
TLPQ = a + α × d β × b + c
TLPQ varies from 0 to 100, assuming that parameters α and β do not exceed 1. The higher this is, the better performance is. Note that the values of α and β tuning parameters can be freely determined by network operators, based on their expectations and needs. For instance, α = 1 means that overprediction by one traffic level is accepted (it increases the value of the TLPQ). In turn, β = 1 indicates that the underprediction is not acceptable and it leads to a decrease of the TLPQ value.

5. Numerical Results

Experiments were conducted on five datasets described in Section 3 using the ML models and algorithms presented in Section 4. The analyzed ML models and algorithms were implemented in Python, using the scikit-learn ML package [32]. All classifiers’ and regressors’ parameters were left as default. In order to examine performance of the proposed ML methods in various situations that can occur in optical networks, for each analyzed dataset we created three different cases in terms of the number of traffic levels, namely, we considered 7, 12 and 20 traffic levels. The motivation behind this approach was that, based on the traffic bitrates and applied technology (e.g., various bitrates per one optical channel (transceiver)), the number of possible traffic levels can vary in real-world scenarios.

5.1. Accuracy and Root Mean Square Percentage Error

Table 1 reports the accuracy and root mean square percentage error (RMSPE) results obtained for traffic level prediction for all considered regressors and models, based on datasets where traffic was divided into seven traffic levels. Individual models differed in a dataset instances features list. RMSPE indicates by how many percent on average considered regressor prediction differs compared to real value. Such a measure allows researchers to evaluate regressors more precisely. Cells highlighted in green consist of the best results for the considered model and dataset, i.e., for each model (four adjacent rows) and dataset (one column), the result is highlighted in green. Results in blue are the best for the whole dataset (along all models), i.e., one cell in each column is marked in blue.
Based on the presented results, the following general trend can be noticed. For datasets with lower MAPE parameter value (i.e., dataset_4 and dataset_5), the accuracy was quite high—the best results were 92.05% and 96.68%, respectively. In turn, for dataset_3–dataset_5 with higher MAPE values, the accuracy was much lower. The general observation is that the accuracy grows with the increase of the MAPE parameter. By analyzing the results of different models, it is clear that adding to the dataset consecutive features allows regressors to yield better results, i.e., models 3 and 4 with additional features on previous seasonal time stamp outperform models 1 and 2. For dataset_1 to dataset_3 (with traffic between a single nodes pair), the best accuracy for traffic level prediction was achieved using model 4. In turn, for dataset_4 and dataset_5 (with incoming traffic to node and SIX traffic) the best performance was yielded by model 3. In almost every case, the highest accuracy for traffic level prediction was calculated based on the LR regressor. The exception to this rule was seen in case of model 1 for dataset_4 and dataset_5, where the best accuracy for traffic level prediction was calculated based on the MLPR regressor; however, their results were close to the results of the LR regressor.
With the increase in dataset MAPE value, RMSPE of traffic level prediction also increased. This was equal to 2.79% for dataset_5 and reached 15.24% for dataset_3.
Figure 6 and Figure 7 compare the real bitrate with the bitrate predicted by the LR regressor for dataset_2 (Figure 6) and dataset_4 (Figure 7) in cases of traffic with seven possible traffic levels. The blue line represents the real bitrate and the orange line depicts the predicted traffic by the regressor. The dotted line shows possible traffic levels. The charts present data from the first day of tested datasets. They demonstrate that traffic from dataset_2 with a higher MAPE value (Figure 6) was sharper, while traffic from dataset_4 with a lower MAPE value (Figure 7) was rather smooth. In the case of the former, traffic levels were determined based on the real traffic line and the predicted traffic line varied more often. This causes lower accuracy and is reflected in the results presented in Table 1.

5.2. TLPQ Metric

As was stated above, from the point of view of the network operator, analyzing the results of classifiers/regressors considering only accuracy for the traffic level prediction can be insufficient. This overlooks information about the number of instances for which traffic level is underpredicted (based on these predictions, the network operator cannot serve traffic) and overpredicted (based on these predictions, the operator can serve traffic; however, it requires additional network resources). To face this problem, our proposed metric takes into account above-mentioned variables. In our work, we consider three different network operator scenarios, differing from each other by values of the tuning parameters α and β used in the TLPQ (1):
  • Scenario 1 (S_1)–α = 1 and β = 0;
  • Scenario 2 (S_2)–α = 0.7 and β = 0.3;
  • Scenario 3 (S_3)–α = 0.5 and β = 0.5.
Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 present TLPQ values for all scenarios based on data from each dataset. Table 2, Table 3 and Table 4 refer to the regression approach while Table 5, Table 6 and Table 7 refer to the classification approach. Different tables consist of results for datasets with different number of possible traffic levels. Again, cells highlighted in green consist of the best results for the considered model and dataset, i.e., for each model (four adjacent rows) and dataset (one main column), the result is highlighted in green. Results in blue are the best for the whole dataset (among all models) in a particular scenario, i.e., one cell in each column is marked in blue.
Analyzing the results presented in Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7, it is difficult to point out the ML algorithm for which TLPQ values were the best for all datasets, i.e., one algorithm which can be considered as the best regardless of dataset. Regressors that most often obtained the best results were PAR, MLPR, and LR. In the case of the classification approach, the best results were obtained most often by LDA. Other classifiers that achieved the best results in some cases were KNN, GBT, and DT, mainly for dataset_4 and dataset_5; however, their TLPQ values differed slightly from LDA results. Choosing the best model is difficult because for different dataset types and different scenarios the best models are different.
It is hard to define explicitly which value of TLPQ can be considered as good and which as bad. The TLPQ value is strongly correlated with alpha and beta parameters. Based on the results in Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7, it can be seen that the best TLPQ values for a single dataset often vary significantly for different scenarios; nevertheless, they point to the same classifier/regressor, i.e., Table 2 shows that for dataset_1 and model 1, in the case of each scenario, PAR yielded the best results, returning different TLPQ values for all scenarios. Only the values of TLPQ for the same scenarios can be compared. Assessing the results obtained for classification and regression approaches, it is clear that in the case of regression TLPQ values were higher, in detail, calculation of traffic levels based on predicted traffic flow is more accurate than predicting traffic levels directly as the result of classification.

5.3. Classification vs. Regression

Table 8 presents the difference between the best values of TLPQ yielded by regression and classification approaches for all scenarios and all datasets, for instance, value 3.2 shown in the first cell indicates that regression outperforms classification by 3.2 points in terms of the TLPQ metric. The higher the presented value is, the better the performance of regression when compared against classification. The goal of Table 8 is to compare the general performance of the two main applied approaches, namely, regression and classification. Green marks the biggest differences while red marks the smallest differences. It can be seen that in every case the regression approach obtained better results. While increasing the dataset MAPE value, the gap between both approaches also increased and reached 10.6 points for dataset_1 in scenario 1. The difference was higher for datasets with a greater number of possible traffic levels. Exceptions occurred for dataset_2 and dataset_3, where generally the highest difference was for seven possible traffic levels.

5.4. Algorithms Performance Summary

Figure 8, Figure 9 and Figure 10 show the best TLPQ values for all datasets and all scenarios. Figure 8 refers to scenario 1, Figure 9 refers to scenario 2, and Figure 10 refers to scenario 3. The results are summarized in Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7. For each scenario and dataset, the best TPLQ values were obtained for dataset cases with seven traffic levels. For datasets with lower MAPE value parameter, i.e., dataset_4 and dataset_5, similar results were obtained for each traffic level case. With the increase in dataset MAPE value, the difference between TLPQ values for different traffic level cases also increased.
Table 9 consists of regressors and models that obtained the best results in a particular dataset based on the TLPQ value and accuracy (in detail, MLPR 3 means that the best regressor for the considered dataset and metric was MLPR using the model 3). If there are more than one regressor and model in a cell, this means that for different scenarios different regressors and models were considered as the best, i.e., for dataset_3 and TLPQ metric. In all other cases, a single regressor and model was the best for all scenarios. It can be observed that considering different measures, miscellaneous regressors and models can be described as the best. This shows that, even if the regressor is the best in terms of accuracy for the final level, it will not necessarily be the best choice for the operator (considering network resource usage).

6. Conclusions

In this paper, we proposed various ML approaches for traffic level prediction in optical networks based on regression and classification methods using four various ML models with different features. To compare the proposed approaches, we formulated a new aggregate quality measure TLPQ that accounts for information about too low or too high traffic level predictions, which in our opinion can be more informative for network operators. We examined four different classifiers and four different regressors.
Our study showed that in all analyzed cases, the regression approach outperforms the classification approach, i.e., regression provides better results when compared to the results yielded by classification. The TLPQ difference between those two approaches varies from one to ten points depending on the dataset. To recall, using regression, first the traffic bitrate is predicted and then the traffic level is calculated according to the obtained bitrate. In turn, using classification, the traffic level is predicted directly as the result of classification, which causes more frequent wrong predictions.
Comparing the performances of various regressors and models, we noticed that the final selection of the most appropriate regressor and model should follow the network operator’s expectations. Different choices are suitable for the network operator who cannot waste network resources compared to an operator who cares more about the served traffic rather than resource usage. Therefore, the network operator should test various regressors and models to find a configuration guaranteeing the best prediction performance according to their needs. Reported results show that the best model for each dataset cannot be indicated. This proved the no free lunch theorem formulated by Wolpert [33], that there is no one algorithm that suits the best for every problem. Regressors that can be considered as the best among all tested are PAR and MLPR. They usually achieved the best results. In the case of the classification approach, LDA can be pointed to as the classifier which obtained the best results for almost all datasets.
Another observation is that the accuracy as well as TLPQ values are strictly correlated with the dataset MAPE value. Classifiers and regressors obtained better results for datasets characterized by lower MAPE value. Traffic in those datasets rarely changes the level, so the prediction is easier.
In future work, we plan to develop an optimization framework that uses the obtained knowledge about future traffic to improve the routing algorithms applied in the optical network. We also plan to study the possibility of middle-term and long-term prediction using the data stream methods. Moreover, we would like to test our approaches on other datasets generated based on different Internet exchange points.

Author Contributions

Conceptualization, D.S. and K.W.; methodology, D.S. and K.W.; software, D.S. and A.W.; validation, D.S. and K.W.; formal analysis, D.S. and K.W.; resources, D.S., A.W., K.W.; data curation, D.S., A.W.; writing—original draft preparation, D.S., A.W., K.W.; writing—review and editing, D.S., A.W., K.W.; visualization, D.S.; supervision, K.W.; project administration, K.W.; funding acquisition, K.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Science Centre, Poland under Grant 2017/27/B/ST7/00888.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARIMAAutoregressive Integrated Moving Average
CTConstant Traffic
DTDecision Tree
EONElastic Optical Network
GNBGaussian Naive Bayes
KNNK Neighbors Classifier
KNNRK Neighbors Regressor
LDALinear Discriminant Analysis
LRLinear Regression
MAPEMean Absolute Percentage Error
MLMachine Learning
MLPRMulti-Layer Perceptron Regressor
PARPassive Aggressive Regressor
PPPoisson Process
PPBPPoisson Pareto Burst Process
RMSPERoot Mean Squared Percentage Error
S_1Scenario 1
S_2Scenario 2
S_3Scenario 3
SIXSeattle Internet Exchange Point
TLPQTraffic Level Prediction Quality
WDMWavelength Division Multiplexing

References

  1. CISCO. Cisco Annual Internet Report (2018–2023); CISCO: San Jose, CA, USA, 2020. [Google Scholar]
  2. Nokia. Deepfield Network Intelligence Report Networks in 2020; Nokia: Espoo, Finland, 2020. [Google Scholar]
  3. Walkowiak, K. Modeling and Optimization of Cloud-Ready and Content-Oriented Networks; Studies in Systems, Decision and Control; Springer: Cham, Switzerland, 2016; Volume 56. [Google Scholar]
  4. Mukherjee, B.; Tomkos, I.; Tornatore, M.; Winzer, P.; Zhao, Y. Springer Handbook of Optical Networks; Springer: Cham, Switzerland, 2020. [Google Scholar]
  5. Chan, V.W.S. Cognitive optical networks. In Proceedings of the IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018. [Google Scholar] [CrossRef]
  6. Chan, V.W.S.; Jang, E. Cognitive all-optical fiber network architecture. In Proceedings of the International Conference on Transparent Optical Networks (ICTON), Girona, Spain, 2–6 July 2017. [Google Scholar] [CrossRef]
  7. Boutaba, R.; Salahuddin, M.A.; Limam, N.; Ayoubi, S.; Shahriar, N.; Estrada-Solano, F.; Caicedo, O.M. A comprehensive survey on machine learning for networking: Evolution, applications and research opportunities. J. Internet Serv. Appl. 2018, 9, 16. [Google Scholar] [CrossRef] [Green Version]
  8. Mata, J.; de Miguel, I.; Duran, R.J.; Merayo, N.; Singh, S.K.; Jukan, A.; Chamania, M. Artificial intelligence (AI) methods in optical networks: A comprehensive survey. Opt. Switch. Netw. 2018, 28, 43–57. [Google Scholar] [CrossRef]
  9. Musumeci, F.; Rottondi, C.; Nag, A.; Macaluso, I.; Zibar, D.; Ruffini, M.; Tornatore, M. An Overview on Application of Machine Learning Techniques in Optical Networks. IEEE Commun. Surv. Tutor. 2019, 21, 1383–1408. [Google Scholar] [CrossRef] [Green Version]
  10. Rafique, D.; Velasco, L. Machine Learning for Network Automation: Overview, Architecture, and Applications [Invited Tutorial]. J. Opt. Commun. Netw. 2018, 10, D126–D143. [Google Scholar] [CrossRef] [Green Version]
  11. Khan, F.N.; Lu, C.; Lau, A.P.T. Optical Performance Monitoring in Fiber-Optic Networks Enabled by Machine Learning Techniques. In Proceedings of the Optical Fiber Communications Conference and Exposition (OFC), San Diego, CA, USA, 11–15 March 2018. [Google Scholar]
  12. Alarcon-Aquino, V.; Barria, J.A. Multiresolution FIR neural-network-based learning algorithm applied to network traffic prediction. IEEE Trans. Syst. Man Cybern. Part C 2006, 36, 208–220. [Google Scholar] [CrossRef] [Green Version]
  13. Krishnamurthy, B.; Sen, S.; Zhang, Y.; Chen, Y. Sketch-based change detection: Methods, evaluation, and applications. In Proceedings of the 3rd ACM SIGCOMM Conference on Internet Measurement, Miami Beach, FL, USA, 27–29 October 2003. [Google Scholar] [CrossRef]
  14. Ding, X.; Canu, S.; Denoeux, T.; Rue, T.; Pernant, F. Neural network based models for forecasting. In Proceedings of the Applied Decision Technologies Conference, San Antonio, TX, USA, 4–8 June 1995; pp. 243–252. [Google Scholar]
  15. Argon, O.; Shavitt, Y.; Weinsberg, U. Inferring the periodicity in large-scale Internet measurements. In Proceedings of the IEEE INFOCOM, Turin, Italy, 14–19 April 2013. [Google Scholar] [CrossRef]
  16. Szostak, D.; Walkowiak, K.; Włodarczyk, A. Short-term Traffic Forecasting in Optical Network using Linear Discriminant Analysis Machine Learning Classifier. In Proceedings of the International Conference on Transparent Optical Networks (ICTON), Bari, Italy, 19–23 July 2020. [Google Scholar] [CrossRef]
  17. Rzym, G.; Boryło, P.; Chołda, P. A Time-Efficient Shrinkage Algorithm for Fourier-Based Prediction Enabling Proactive Optimization in Software Defined Networks. Int. J. Commun. Syst. 2019, 33, e4448. [Google Scholar] [CrossRef]
  18. Cortez, P.; Rio, M.; Rocha, M.; Sousa, P. Multi-scale Internet traffic forecasting using neural networks and time series methods. Expert Syst. 2012, 29, 143–155. [Google Scholar] [CrossRef] [Green Version]
  19. Otoshi, T.; Ohsita, Y.; Murata, M.; Takahashi, Y.; Ishibashi, K.; Shiomoto, K. Traffic prediction for dynamic traffic engineering. Comput. Netw. 2015, 85, 36–50. [Google Scholar] [CrossRef]
  20. Aibin, M. Traffic prediction based on machine learning for elastic optical networks. Opt. Switch. Netw. 2018, 30, 33–39. [Google Scholar] [CrossRef]
  21. Vinchoff, C.; Chung, N.; Gordon, T.; Lyford, L.; Aibin, M. Traffic Prediction in Optical Networks Using Graph Convolutional Generative Adversarial Networks. In Proceedings of the International Conference on Transparent Optical Networks (ICTON), Bari, Italy, 19–23 July 2020; pp. 3–6. [Google Scholar] [CrossRef]
  22. Panayiotou, T.; Chatzis, S.P.; Ellinas, G. Leveraging Statistical Machine Learning to Address Failure Localization in Optical Networks. J. Opt. Commun. Netw. 2018, 10, 162–173. [Google Scholar] [CrossRef]
  23. Shahkarami, S.; Musumeci, F.; Cugini, F.; Tornatore, M. Machine Learning-Based Soft-Failure Detection and Identification in Optical Networks. In Proceedings of the Optical Fiber Communications Conference and Exposition (OFC), San Diego, CA, USA, 11–15 March 2018. [Google Scholar]
  24. Wang, L.; Wang, X.; Tornatore, M.; Kim, K.J.; Kim, S.M.; Kim, D.U.; Han, K.E.; Mukherjee, B. Scheduling with Machine-Learning-Based Flow Detection for Packet-Switched Optical Data Center Networks. J. Opt. Commun. Netw. 2018, 10, 365–375. [Google Scholar] [CrossRef]
  25. Proietti, R.; Chen, X.; Zhang, K.; Liu, G.; Shamsabardeh, M.; Castro, A.; Velasco, L.; Zhu, Z.; Yoo, S.J.B. Experimental Demonstration of Machine-Learning-Aided QoT Estimation in Multi-Domain Elastic Optical Networks with Alien Wavelengths. J. Opt. Commun. Netw. 2018, 11, A1–A10. [Google Scholar] [CrossRef]
  26. Khan, I.; Bilal, M.; Siddiqui, M.; Khan, M.; Ahmad, A.; Shahzad, M.; Curri, V. QoT Estimation for Light-path Provisioning in Un-Seen Optical Networks using Machine Learning. In Proceedings of the International Conference on Transparent Optical Networks (ICTON), Bari, Italy, 19–23 July 2020. [Google Scholar] [CrossRef]
  27. Nguyen, G.; Dlugolinsky, S.; Tran, V.; Lopez Garcia, A. Deep learning for proactive network monitoring and security protection. IEEE Access 2020, 8, 19696–19716. [Google Scholar] [CrossRef]
  28. Zukerman, M.; Neame, T.D.; Addie, R.G. Internet traffic modeling and future technology implications. In Proceedings of the IEEE INFOCOM 2003, San Francisco, CA, USA, 30 March–3 April 2003. [Google Scholar] [CrossRef] [Green Version]
  29. Walkowiak, K.; Klinkowski, M.; Lechowicz, P. Dynamic Routing in Spectrally Spatially Flexible Optical Networks with Back-to-Back Regeneration. J. Opt. Commun. Netw. 2018, 10, 523–534. [Google Scholar] [CrossRef]
  30. Simmons, J. Optical Network Design and Planning, 2nd ed.; Springer: Cham, Switzerland, 2014. [Google Scholar]
  31. Bifet, A.; Morales, G.D.F. Big Data Stream Learning with SAMOA. In Proceedings of the IEEE International Conference on Data Mining Workshop, Shenzhen, China, 14 December 2014. [Google Scholar] [CrossRef]
  32. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  33. Wolpert, D.H. The Supervised learning no-free-lunch theorems. In Proceedings of the 6th Online World Conference on Soft Computing in Industrial Applications, Online. 10–24 September 2001. [Google Scholar] [CrossRef]
Figure 1. Traffic flows.
Figure 1. Traffic flows.
Electronics 10 01578 g001
Figure 3. Autocorrelation.
Figure 3. Autocorrelation.
Electronics 10 01578 g003
Figure 4. ML model creation example, features: day, minute, window = 1, classification.
Figure 4. ML model creation example, features: day, minute, window = 1, classification.
Electronics 10 01578 g004
Figure 5. System diagram.
Figure 5. System diagram.
Electronics 10 01578 g005
Figure 6. Real and predicted traffic for dataset_2, LR regressor.
Figure 6. Real and predicted traffic for dataset_2, LR regressor.
Electronics 10 01578 g006
Figure 7. Real and predicted traffic for dataset_4, LR regressor.
Figure 7. Real and predicted traffic for dataset_4, LR regressor.
Electronics 10 01578 g007
Figure 8. Scenario 1 (S_1) the best TLPQ values.
Figure 8. Scenario 1 (S_1) the best TLPQ values.
Electronics 10 01578 g008
Figure 9. Scenario 2 (S_2) the best TLPQ values.
Figure 9. Scenario 2 (S_2) the best TLPQ values.
Electronics 10 01578 g009
Figure 10. Scenario 3 (S_3) the best TLPQ values.
Figure 10. Scenario 3 (S_3) the best TLPQ values.
Electronics 10 01578 g010
Table 1. Regression approach, accuracy and RMSPE for datasets with 7 possible traffic levels.
Table 1. Regression approach, accuracy and RMSPE for datasets with 7 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelRegAccuracyRMSPEAccuracyRMSPEAccuracyRMSPEAccuracyRMSPEAccuracyRMSPE
1LR84.55%6.60%68.55%11.14%62.55%15.57%91.77%3.80%95.54%3.28%
PAR84.23%6.71%60.04%11.66%34.15%24.73%91.71%3.75%95.37%3.39%
KNNR83.44%6.80%66.55%11.51%60.50%16.28%91.21%3.89%95.17%3.35%
MLPR84.50%6.57%68.36%11.08%61.86%15.24%91.83%3.76%95.59%3.24%
2LR84.81%6.54%69.43%10.90%63.66%15.21%92.04%3.69%96.64%2.81%
PAR83.18%6.47%30.17%19.03%61.05%16.28%87.88%4.71%95.27%3.30%
KNNR83.24%6.79%65.23%11.76%59.67%16.23%90.15%4.07%94.79%3.49%
MLPR84.60%6.33%69.19%10.92%63.40%15.14%91.12%3.76%95.49%3.19%
3LR84.96%6.51%69.43%10.90%63.73%15.20%92.05%3.68%96.68%2.79%
PAR74.33%7.65%64.60%12.33%63.14%15.42%91.49%3.79%89.23%4.57%
KNNR83.05%6.85%64.10%12.06%57.92%16.80%90.25%4.04%94.79%3.49%
MLPR83.97%6.89%69.21%10.75%63.54%15.50%91.53%3.86%95.35%3.23%
4LR85.47%6.41%70.00%10.83%64.38%15.02%91.85%3.78%96.50%2.90%
PAR83.69%6.43%24.01%21.35%60.02%15.88%88.93%4.51%90.43%5.09%
KNNR79.81%7.48%61.44%12.89%55.56%17.48%83.40%5.52%84.03%6.31%
MLPR84.07%6.86%69.51%11.04%63.33%15.73%90.42%4.05%91.38%4.77%
Table 2. Regression, TLPQ values for datasets with 7 possible traffic levels.
Table 2. Regression, TLPQ values for datasets with 7 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelRegS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LR92.187.584.483.474.167.879.868.661.295.993.491.897.896.595.6
PAR93.388.685.564.752.744.786.870.359.394.892.390.798.497.196.1
KNNR91.486.483.181.771.765.181.169.461.695.292.690.897.596.095.0
MLPR92.187.584.482.573.166.874.963.555.996.093.591.997.896.595.6
2LR92.287.784.684.275.068.980.769.962.695.993.591.998.397.296.6
PAR86.681.678.290.372.160.086.775.367.698.895.192.798.196.795.7
KNNR91.086.082.781.971.564.578.266.258.294.391.489.497.095.494.4
MLPR89.284.681.585.075.769.681.070.162.893.390.688.898.597.296.3
3LR92.487.984.884.275.169.080.769.962.795.993.591.998.397.396.6
PAR75.267.462.392.081.574.582.671.664.395.292.690.989.386.083.9
KNNR91.186.082.681.470.763.678.065.657.394.491.589.597.095.494.4
MLPR95.690.887.681.672.466.283.472.665.396.794.192.497.796.395.4
4LR93.088.685.785.176.270.281.270.663.596.393.992.298.397.296.5
PAR87.682.779.488.468.955.981.469.661.796.192.890.698.195.293.3
KNNR89.683.679.581.770.362.675.662.553.891.886.983.594.189.386.1
MLPR94.990.287.085.576.470.385.074.166.894.691.789.898.796.194.3
Table 3. Regression, TLPQ values for datasets with 12 possible traffic levels.
Table 3. Regression, TLPQ values for datasets with 12 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelRegS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LR86.177.972.573.359.249.868.653.142.792.988.585.796.494.292.8
PAR83.274.969.340.120.98.172.457.347.369.259.953.796.594.392.9
KNNR84.976.170.271.256.246.366.050.139.592.587.984.896.093.692.0
MLPR86.178.072.772.158.248.968.753.142.892.788.485.596.494.292.8
2LR86.778.773.474.660.951.869.554.244.093.189.086.397.496.095.1
PAR88.080.174.876.863.154.055.738.827.594.690.287.383.077.974.5
KNNR84.575.469.371.156.046.065.248.837.991.386.382.996.594.392.8
MLPR88.680.675.276.362.653.571.155.945.892.588.385.597.094.993.4
3LR86.678.773.474.761.051.969.554.244.093.189.186.497.496.095.0
PAR91.182.777.179.463.452.722.3−2.5−19.096.374.459.796.794.693.3
KNNR85.276.470.570.955.845.664.147.336.291.486.583.296.494.292.7
MLPR85.677.872.773.960.251.172.457.247.195.390.887.898.095.694.1
4LR87.679.974.875.862.353.369.854.444.193.889.786.997.696.295.2
PAR80.071.565.884.269.659.820.9−3.2−19.394.588.985.194.891.088.4
KNNR82.871.864.469.553.743.161.144.232.989.280.975.390.683.678.9
MLPR83.174.769.172.158.449.371.055.745.587.081.978.597.094.693.0
Table 4. Regression, TLPQ values for datasets with 20 possible traffic levels.
Table 4. Regression, TLPQ values for datasets with 20 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelRegS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LR76.162.954.154.236.925.447.529.717.886.378.072.593.289.186.3
PAR84.869.459.151.033.121.239.629.322.491.382.877.175.167.362.2
KNNR74.460.551.250.833.321.645.128.116.885.176.270.291.586.683.3
MLPR75.462.253.352.234.422.547.329.517.586.478.272.793.188.986.1
2LR76.563.554.855.938.627.148.530.718.986.778.873.595.392.490.4
PAR78.065.056.328.722.017.540.820.67.291.782.776.877.069.965.2
KNNR73.359.049.550.432.921.243.625.513.584.675.569.592.488.185.3
MLPR81.168.059.258.242.632.247.229.217.188.279.774.189.985.782.9
3LR76.463.354.656.038.727.248.530.818.986.678.773.595.392.390.4
PAR86.971.861.853.640.131.19.7−17.8−36.253.338.929.286.381.478.2
KNNR74.260.250.950.232.821.242.324.713.084.575.469.292.588.285.3
MLPR71.858.749.956.139.027.649.031.720.284.676.470.988.884.581.5
4LR78.065.056.456.939.828.448.931.119.388.080.174.795.592.690.6
PAR54.939.028.456.439.428.128.77.7−6.318.7−5.7−22.098.285.877.6
KNNR68.152.341.747.430.118.540.622.810.977.463.353.882.269.961.6
MLPR76.362.853.856.739.528.135.113.4−1.187.077.270.691.286.383.0
Table 5. Classification, TLPQ values for datasets with 7 possible traffic levels.
Table 5. Classification, TLPQ values for datasets with 7 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelClassS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LDA91.687.084.084.074.768.479.868.761.393.791.089.197.094.893.3
KNN91.186.082.679.268.461.278.966.758.694.992.190.397.796.295.2
GNB91.987.384.383.273.867.678.967.960.594.091.389.597.295.394.0
DT90.184.981.581.972.265.880.068.861.493.089.787.496.894.993.7
2LDA91.887.284.284.274.768.580.469.462.193.891.189.397.595.694.4
KNN90.585.181.580.369.562.274.661.753.194.090.988.896.895.194.0
GNB89.982.978.278.165.557.272.558.449.089.783.579.492.086.983.5
DT89.282.878.678.466.057.772.958.348.794.391.188.996.494.593.3
3LDA92.387.784.784.975.569.280.669.662.393.190.188.297.495.193.6
KNN90.485.181.579.568.360.873.960.551.694.191.189.196.895.194.0
GNB90.183.178.578.065.457.072.558.348.990.184.180.192.087.083.7
DT88.682.177.777.665.357.172.658.148.594.190.788.396.594.493.0
4LDA92.487.484.085.175.468.980.669.562.193.590.388.196.292.990.7
KNN89.283.078.879.066.858.670.556.146.691.786.683.294.189.285.9
GNB90.483.679.178.165.156.572.858.148.290.885.381.792.787.483.9
DT88.281.376.778.866.358.074.260.050.692.788.385.395.192.490.5
Table 6. Classification, TLPQ values for datasets with 12 possible traffic levels.
Table 6. Classification, TLPQ values for datasets with 12 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelClassS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LDA86.278.072.573.659.650.368.352.642.192.788.285.396.093.792.1
KNN84.775.769.764.348.337.760.643.832.592.587.884.796.393.892.1
GNB85.877.772.272.458.348.966.250.540.192.688.385.495.993.692.0
DT82.973.867.671.056.747.167.251.541.189.183.479.794.791.589.4
2LDA86.378.172.774.660.851.668.953.342.992.988.685.797.095.394.2
KNN82.873.266.864.448.537.958.040.528.991.185.982.596.494.092.5
GNB79.968.560.965.149.639.357.640.328.785.176.370.488.881.977.3
DT79.768.460.964.247.937.157.140.128.890.384.680.795.693.291.6
3LDA86.478.172.674.660.851.569.153.443.092.487.984.995.492.991.3
KNN83.874.468.163.847.536.756.638.927.190.985.982.596.394.092.4
GNB80.469.261.665.349.839.557.440.328.885.476.771.088.681.677.0
DT80.369.161.765.249.038.357.240.128.890.584.680.796.293.892.3
4LDA86.477.671.774.760.551.067.651.741.091.885.982.093.389.486.8
KNN81.069.561.863.046.034.753.335.123.089.080.474.690.383.278.4
GNB83.072.265.066.050.640.457.340.429.187.579.073.488.081.076.4
DT79.066.858.763.247.236.557.540.729.588.080.575.694.891.589.3
Table 7. Classification, TLPQ values for datasets with 20 possible traffic levels.
Table 7. Classification, TLPQ values for datasets with 20 possible traffic levels.
--DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
ModelClassS_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3S_1S_2S_3
1LDA76.262.954.053.936.825.346.928.916.886.378.072.592.287.884.9
KNN69.654.845.044.025.212.737.919.06.483.774.368.191.486.483.0
GNB75.762.453.652.734.923.145.827.214.986.177.872.392.187.784.7
DT70.155.545.950.332.620.845.427.315.278.067.360.189.183.179.1
2LDA76.363.154.254.537.225.746.929.217.486.678.673.393.890.688.5
KNN66.851.641.542.023.210.735.616.23.382.873.166.591.687.083.9
GNB66.751.541.343.726.515.037.519.77.874.661.452.681.471.064.0
DT66.650.840.242.124.813.334.617.45.981.370.363.091.486.883.7
3LDA76.262.854.054.237.025.546.829.317.786.077.672.192.789.086.6
KNN67.352.041.842.323.410.834.014.71.882.772.966.391.586.983.9
GNB67.352.242.143.526.114.637.019.17.174.861.853.281.270.763.8
DT67.151.140.542.325.013.536.119.07.681.670.863.691.386.483.2
4LDA75.561.652.253.636.725.445.127.716.084.373.766.692.086.382.5
KNN62.746.235.239.220.58.032.512.9−0.175.560.851.080.868.460.1
GNB69.454.344.144.026.915.635.618.26.577.164.055.382.571.864.7
DT68.152.542.043.125.914.435.918.77.278.866.257.888.281.276.6
Table 8. Difference of TLPQ value between regression and classification approaches.
Table 8. Difference of TLPQ value between regression and classification approaches.
-DATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
-Scenario 1 (S_1)
73.26.96.13.91.0
124.79.53.43.41.0
201.63.82.15.14.4
-Scenario 2 (S_2)
73.06.05.63.01.1
124.68.83.92.20.9
208.85.42.44.22.0
-Scenario 3 (S_3)
72.95.45.32.41.5
124.58.24.32.21.0
207.56.54.73.82.1
Table 9. The best regressors for datasets based on different measures.
Table 9. The best regressors for datasets based on different measures.
LevelsMeasureDATASET_1DATASET_2DATASET_3DATASET_4DATASET_5
7TLPQMLPR 3PAR 3PAR 1/PAR3PAR 2MLPR 4/LR 3
AccuracyLR 4LR 4LR 4LR 3LR 3
12TLPQPAR 3PAR 4MLPR 3/PAR 1PAR 3/MLPR 3MLPR 3/LR 4
AccuracyLR 4LR 4LR 3LR 3LR 2
20TLPQPAR 3MLPR 2MLPR 3/PAR 1PAR 2/PAR 1PAR 4/LR 4
AccuracyLR 4LR 4LR 4LR 2LR 2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Szostak, D.; Włodarczyk, A.; Walkowiak, K. Machine Learning Classification and Regression Approaches for Optical Network Traffic Prediction. Electronics 2021, 10, 1578. https://doi.org/10.3390/electronics10131578

AMA Style

Szostak D, Włodarczyk A, Walkowiak K. Machine Learning Classification and Regression Approaches for Optical Network Traffic Prediction. Electronics. 2021; 10(13):1578. https://doi.org/10.3390/electronics10131578

Chicago/Turabian Style

Szostak, Daniel, Adam Włodarczyk, and Krzysztof Walkowiak. 2021. "Machine Learning Classification and Regression Approaches for Optical Network Traffic Prediction" Electronics 10, no. 13: 1578. https://doi.org/10.3390/electronics10131578

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop