Next Article in Journal
YOLOv7-3D: A Monocular 3D Traffic Object Detection Method from a Roadside Perspective
Previous Article in Journal
Perceptions on Drinking Water and Wastewater in a Local Area in Western Romania
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces

1
State Key Laboratory of Optical Fiber and Cable Manufacture Technology, China Telecom Research Institute, Beijing 100032, China
2
School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(20), 11398; https://doi.org/10.3390/app132011398
Submission received: 21 September 2023 / Revised: 8 October 2023 / Accepted: 12 October 2023 / Published: 17 October 2023
(This article belongs to the Section Electrical, Electronics and Communications Engineering)

Abstract

:
With the continuous emergence of the concept of 6G, the rapid development of industrial application scenarios, and the rise in demand for dedicated line services, there has been a strong and diverse demand for network capacity. The introduction of fine-granularity FlexE technology provides a new approach to addressing the diversification of Quality of Service (QoS) and the resource waste caused by large granularity (5 Gbps). However, the reduction in granularity leads to a significant increase in the number of slots, posing a serious challenge to deterministic latency assurance, especially in scenarios with dynamic fluctuations in service bandwidth. Sudden bandwidth variations can lead to service interruptions and fluctuations, resulting in uncontrolled network blocking rates. To address this issue, this paper proposes a traffic-driven proactive FlexE slot orchestration algorithm called PMFAN-GDSO. Based on this algorithm, bandwidth variations are calculated in advance based on the bandwidth change provided via the traffic prediction algorithm before service fluctuations occur. This avoids the time-consuming slot orchestration calculations and service abrupt changes. The real-world network datasets indicate that, based on our proposed algorithm, significant time savings of up to 46.8% are achieved compared to non-proactive prediction.

1. Introduction

In the context of advancing 6G technologies, the demands on optical communication networks have become more intricate and diverse [1]. The emergence of increasingly complex application scenarios in vertical industries, coupled with the rise of high-value dedicated line services, has led to a surge in network carrying demands [2]. This encompasses requirements for small bandwidth, high isolation, low-latency determinism, and a strong emphasis on high security and reliability. To meet the burgeoning need for fine-granularity resource allocation beyond conventional coarse-grained provisioning, a pivotal shift in the network architecture and protocols is imperative. This is particularly crucial for scenarios requiring dynamic and flexible resource management, such as ultra-reliable low-latency communications (URLLC) and the massive machine-type communications (mMTC) envisioned in the 5G paradigm.
In response to this exigency, the introduction of FlexE (Flexible Ethernet) technology emerges as a transformative solution [3]. FlexE is a pivotal advancement in optical networking that enables the subdivision of Ethernet connections into smaller, flexible sub-connections, each with an independently configurable bandwidth and characteristics [4]. This granular resource allocation empowers network operators to efficiently utilize the physical network infrastructure, thereby accommodating the diverse traffic demands expected in the 6G landscape [5].
FlexE operates at the physical layer of the network stack, enabling the dynamic partitioning of Ethernet links into sub-links [6,7]. These sub-links, referred to as “FlexE groups”, are configurable with respect to their bandwidth allocation, providing a level of flexibility unprecedented in traditional Ethernet architectures [8]. The key innovation lies in the ability to allocate and manage bandwidth in increments as small as 10 Mbps, allowing for fine-grained control over network resources.
Moreover, FlexE introduces the concept of “slots”, allowing for the precise allocation of bandwidth resources within a FlexE group. Each slot corresponds to a discrete unit of time during which a specific amount of bandwidth is available for the data transmission. This temporal granularity enhances the adaptability of the network to varying traffic patterns, ensuring the efficient utilization of resources in scenarios characterized by dynamic and bursty traffic [9].
Upon the arrival of a service within a network slicing, ensuring Quality of Service (QoS) for distinct services mandates the proper allocation of route paths contingent upon the extant network status, coupled with the provision of corresponding time slot resources along each link [10]. The transition from a granularity of 5 Gbps to 10 Mbps engenders the introduction of a pioneering fine-granularity unit (FGU) sublayer within the FlexE frame structure. This introduction, in tandem with the further segmentation and reutilization of FGUs for 5 Gbps time slots, precipitates an exponential augmentation in the number of time slots managed per link. The reduction in granularity accentuates the salience of the non-deterministic delay induced by time slot conflicts, rendering it non-negligible. Additionally, the dynamic fluctuation in service bandwidth presents a formidable challenge for expeditious and dependable time slot reallocation. The surpassing of the prescribed service bandwidth thresholds may necessitate not only the assignment of slot resources, but could also exert an influence on routing determinations, potentially instigating momentary service interruptions and fluctuations. In recent years, traffic prediction technology has garnered widespread application in improving network decision-making along with the variation in the dynamic traffic trends. Empowered via the a priori knowledge afforded by traffic prediction, networks are aptly poised to proactively extend resource computation and bandwidth reservation for the network service. Ensuring QoS for latency-sensitive services is contingent upon the imperative design of a time slot scheduling apparatus tailored for 10 Mbps fine-granularity slices, alongside a time slot reallocation mechanism underpinned by traffic prediction.
In this study, we address fine-granularity flexible ethernet slot orchestration for dynamic service scenarios based on proactive multi-flow assisted service awareness. We propose a proactive multi-flow assisted attention-based neural network (PMFAN) for high-accuracy service traffic prediction. Building upon service-awareness information, we further present a genetic algorithm-based deterministic slot orchestration algorithm (GDSO) to support end-to-end low-latency transmission, resolving the slot scheduling challenge posed by dynamic traffic requirements. We evaluate the proposed traffic prediction algorithm and slot orchestration mechanism using the real-world traffic datasets and network data. The results demonstrate that a proactive priori knowledge provided by traffic prediction significantly reduces the computation time by up to 46.8%.
The rest of the paper is organized as follows. Section 2 provides a review of related work on FlexE and network traffic prediction. Section 3 formulates the problem. Section 4 presents the PMFAN prediction algorithm and GDSO Orchestration algorithm. The simulation results are represented in Section 5, and Section 6 concludes this paper.

2. Related Works

FlexE is an interface technology for the bearer network to realize service isolation and network slicing. Since ITU-T standard organizations have accelerated the FlexE standard process [11], it has developed rapidly in recent years. Eira et al. [4] provided a solution for decoupling the interface rates between routers and transport devices from the actual data flows and evaluated the trade-off between a transport box’s complexity and its ability to utilize light paths effectively, offering insights into the impact of FlexE use cases on the router port efficiency, transport box provisioning, and DWDM layer capacity in DCI contexts. D. Koulougli et al. [12] explored optimized routing in complex multi-layer, multi-domain IP-optical networks using a hierarchical PCE and FlexE technology. It formulates optimization problems that consider QoS, privacy, and FlexE constraints and introduces novel algorithms for efficient routing and client assignment. P. Zhu et al. [13] addressed the security concerns in next-gen RAN transport, focusing on eavesdropping attacks at the physical layer. A cross-layer approach using FlexE and WDM is proposed for enhanced security. Various attack levels are considered, and a trade-off between resource efficiency and security is explored. The numerical results demonstrate the effectiveness of the defense strategies. D. Koulougli et al. [14] addressed routing optimization in multi-layer multi-domain networks, emphasizing inter-layer and inter-domain coordination. It introduces a hierarchical path computation engine (PCE) that leverages FlexE technology to enhance the network performance by linking the IP and optical domains, presenting an efficient algorithm for routing and FlexE assignment and achieving a substantial performance improvement of an 87% optimal throughput. And, in [12], the authors investigate the FlexE Traffic Restoration (FTR) problem that aims to maintain high network utilization via the fast recovery of FlexE clients with the minimum cost using the spare capacity in the already deployed PHYs. H. Liang et al. [15] addressed the integration of flexible Ethernet (FlexE) and elastic optical networks (EONs) in FlexE-over-EONs scenarios, specifically focusing on the FlexE-aware architecture. It introduces mixed integer linear programming (MILP) and integer linear programming (ILP) models for single-hop and multi-hop scenarios, respectively, and presents highly time-efficient approximation algorithms that provide solutions closely approaching the optimal ones in large-scale planning. Based on [12,14], D. Koulougli et al. [16] investigated the FlexE Traffic Restoration (FTR) problem that aims to maintain high network utilization via the fast recovery of FlexE clients with the minimum cost using the spare capacity in the already deployed PHYs. And, based on [13], P. Zhu et al. [17] introduced a cross-layer security design for FlexE over WDM networks, specifically addressing eavesdropping threats at the physical layer. The approach combines universal Hashing-based FlexE data block permutation with parallel fiber transmission to enhance security. The study evaluates different attack levels, balancing resource efficiency and security, and demonstrates the effectiveness of the proposed cross-layer defense strategies. M. Wu et al. [18] investigated cross-layer restoration (CLR) in FlexE-over-EONs, specifically addressing temporary outages in FlexE switches. Extensive simulations confirm the effectiveness of these CLR strategies, highlighting the potential of FlexE and elastic optical networks for efficient restoration in scenarios involving FlexE switch outages. Recently, Gu, R., et al. [19] established a model of routing embedded timeslot scheduling for the routing of fine-granularity slices and timeslot scheduling problems in SPN-based FlexE interfaces, for which a deterministic timeslot allocation mechanism supporting end-to-end low-latency transmission is proposed.
Currently, most optical network resource scheduling algorithms focus on improving resource utilization based on the rules, with a limited utilization of knowledge information within the network, such as historical traffic data and historical decision data. This leads to a lack of learning tailored to the deployment environment in network decision-making. As a result, some studies are dedicated to predicting key information affecting the scheduling of optical network resources to enhance resource allocation. This paper aims to utilize traffic prediction technology to anticipate the future operational state of FlexE clients, enabling the proactive reservation of slot resources. However, network traffic exhibits highly nonlinear and bursty characteristics, posing significant challenges for accurate traffic prediction. Additionally, the rapid fluctuations in traffic, in contrast to long-term time steps, present substantial challenges for traffic prediction, particularly in the context of FlexE service calendar switching that necessitates recalculations.
Traffic prediction methods can be categorized into two types: classical models and deep learning methods. Classical models exhibit good interpretability and perform well in linear prediction. Auto Regressive Integrated Moving Average (ARIMA) combines autoregressive, differencing, and moving average components to capture the linear dependencies and trends in the data [20]. However, ARIMA struggles to provide accurate predictions for the nonlinear part of the traffic data in the present network.
With the increase in the data and computability, deep learning methods have been widely applied in network traffic prediction. Recurrent Neural Networks (RNN) and their variants are a class of deep models widely applied in time series forecasting due to their ability to retain information from previous sequences. Hallas et al. applied RNN to network traffic prediction [21]. Trinh et al. utilized LSTM for traffic prediction using the real-world datasets [22]. To capture longer-term dependencies, Hochreiter et al. introduced Long Short-Term Memory (LSTM) by modifying memory cells to preserve long-term information [23]. Lazaris et al. demonstrated that LSTM outperforms the ARIMA model by approximately 30% in accurately predicting the link throughput [24]. Zhang et al. proposed LSTM-based Network Traffic Prediction (LNTP), a hybrid optimized model for end-to-end network traffic prediction [25].
The aforementioned deep learning methods primarily focus on utilizing the temporal features of individual flows. Vinchoff et al. utilized Graph Convolutional Networks (GCN) to improve the prediction accuracy [26,27]. Lin et al. combined MGCN (Multi-Graph Convolutional Network) with LSTM for wireless traffic prediction [28]. Zhang et al. utilized densely connected Convolutional Neural Networks (CNN) to capture the spatial dependencies of cell traffic [29]. Huang et al. experimentally verified that CNN is suitable for extracting inter-node correlations, while RNN is effective in capturing temporal features [30]. These methods achieved a 70% to 80% accuracy in various forecasting tasks, outperforming CNN and 3DCNN. Li et al. proposed LA-ResNet, which combines residual networks with RNN and incorporates attention mechanisms to assist in traffic prediction [31]. Cui et al. [32] employed Convolutional Long Short-Term Memory (ConvLSTM) to integrate convolutional layers into the LSTM cells, enabling the combination of spatial–temporal features. However, these works mainly focus on utilizing spatial or topological information to assist in wireless traffic prediction without considering the correlation between services, which could further improve the accuracy of traffic prediction.
Several studies have explored the application of Generalized Nets (GNs) in network traffic prediction, demonstrating its effectiveness in capturing the dynamics of network traffic [33]. Smith et al. [34] proposed a GN-based network traffic prediction model, leveraging GNs to describe network elements and their interactions. The model demonstrated high accuracy and robustness in predicting network traffic fluctuations. Li and Wang [35] introduced a GN-based approach for network traffic prediction, combining time series analysis and machine learning techniques. They employed GNs to describe the dynamic changes in network traffic and applied time series analysis to the GN node and connection attributes. The integration of GNs and machine learning algorithms improved the accuracy and stability of traffic prediction. In addition, Chen et al. [36] proposed a fusion model that combines GNs with deep learning for network traffic prediction. GNs were utilized to describe network elements, and deep learning algorithms were applied to the GN node and connection attributes. The fusion model exhibited a superior performance in capturing the complex patterns and features of network traffic. In summary, GNs have shown promising applications in network traffic prediction. By describing and simulating network elements and their relationships, GNs effectively capture the dynamic changes and fluctuations in network traffic. Furthermore, the integration of GNs with other techniques, such as time series analysis and deep learning, enhances the accuracy and reliability of traffic prediction. Future research can further explore the applications of GNs in network traffic prediction and develop more efficient and robust prediction models.

3. Problem Formulation

In this section, we analyze and formally describe the Flexible Ethernet slot orchestration problem based on traffic prediction.

3.1. FlexE Architecture

The service data flow of traditional Ethernet technology based on IEEE 802.3 is first encapsulated into MAC frames at the MAC layer and connected to the PHY layer through the RS Layer (Reconciliation Sublayer, physical coordination sublayer), and the PCS (Physical Coding Sublayer) in the PHY layer completes the 64B/66B encoding work; the PMA (Physical Medium Attachment) sub-layer completes the serial-to-parallel conversion, clock synthesis, and recovery work; and finally, the PMD (Physical Medium Dependent, physical media related) sub-layer completes the interface work for various actual physical media.
As depicted in Figure 1, the FlexE architecture consists of the FlexE Client, FlexE Shim, and FlexE Group. Firstly, the FlexE Client represents various standard service interfaces, it mainly represents the data flow of MAC carried in the FlexE network. The main function is to convert the data stream into 66bit code blocks after 64B/66B encoding and transmit it to the FlexE Shim layer. Secondly, the FlexE Shim is to implement the core function of FlexE technology between the MAC and the PHY defined via IEEE 802.3, which completes the mapping from the Client to the Group based on the calendar mechanism. At last, the FlexE Group is essentially a collection of multi-layered PHY, which carries the 66-bit data code blocks mapped via FlexE Shim from the FlexE Client layer and also supports by carrying multiple FlexE Instances.

3.2. Network Model

A network-based FlexE interface can be represented via a directed graph G = ( V , E , L , B ) , where V = { v 1 , v 2 , , v N V } represents the set of nodes, | V | = N V indicates that the number of the nodes is N V , and E = { e i j | v i , v j V } represents the set of links in the network. Also, the adjacent nodes are connected by two links in opposite directions; for example, e i j denotes the link from node i to node j , and e i j denotes the link from node j to node i . The link length sets are defined as L = { l e i j | v i , v j V } , and B = { b e i j | v i , v j V } represents the link bandwidth sets in the FlexE.

3.3. Service Model

For any service r k ( r k R ) in the network connection request sets, we represent it via a triple r k = ( s k , d k , B k ) , where s k denotes the source node, d k denotes the destination node, and B k denotes the bandwidth required by the service, respectively. Also, for the service r k , we define the path as P r k = { e ij | i , j 1 , N v ) } ( P r k P ) , where P is composed of K pre-computed paths P r k . In addition, we assume that the service is transmitted from the source node to the destination node with the intermediate nodes, only considering the forwarding method based on slot crossing.
Each service in the FlexE is an Ethernet traffic flow based on the MAC data rate, and all services are represented as a set B = { B k | k = 1 , 2 , , n } . In order to accurately predict the traffic demand, our goal is to calculate the most likely data matrix volume at time slot k based on the historical data of the previous T s slots. The problem formulation is shown as follows:
B k , i ˜ = argmax B k , i ( B k , i | B k , i T s , , B k , i 1 )
where B k , i T s , , B k , i 1 are the historical observations of traffic flow at previous T s time slots.

3.4. Problem Formulation

The slot occupancy of the FlexE frames transmitted on the link e i j is represented by an array U e i j = [ u 1 , u 2 , , u N f r a m e T S m a x ] , where each binary bit of the array represents the occupancy of the slot of the corresponding frame, where the value “0” means that the slot is available in the idle state, and the value “1” means the slot has been occupied, and N f r a m e T S m a x denotes the total number of slots in each FlexE frame. In addition, we assume that there are already some background services in the network occupying the slot resources.
Importantly, the current cellular network faces challenges in propagating the high-capacity data with improved speed, QoS, latency, and efficient HO and mobility management [37]. And ref. [38] proposed an overall prediction method of QoE parameters of users and telecommunication networks based on the QoS indicators’ values prediction, discussed four normalization techniques, and finally proposed a normalization method of the index scale. However, the problem presented in this paper is oriented to scenarios with high requirements for low latency, and delay is one of the most important factors in QoS. Therefore, we decided to set the optimization objective to minimize the end-to-end delay, where the objective function is defined as follows:
M i n e i j P r k τ e i j prop + v j V τ v j f o r w a r d + e i j P r k τ e i j s c h e
where τ e i j prop indicates the propagation delay of service r k from s k to d k , τ v j f o r w a r d is the forwarding delay generated by the service at the intermediate nodes in the path, and τ e i j s c h e is the scheduling delay caused by the service due to slot allocation.
The above objective function should be minimized under the following constraints.
Routing Continuity Constraint: (3) means there is a routing continuity constraint between s k to d k , where φ p , l r k is a binary variable that is equal to 1 if the P r k occupies e p l and is equal to 0 for otherwise.
e p l P r k φ p , l r k e l q P r k φ l , q r k = { 1 , p = s k 0 , o t h e r s 1 , q = d k , p , q V , p q , l V
Bandwidth Continuity Constraint: (4) means the service r k occupies the same size of bandwidth resources on different links e p l and e l q in the route.
b p , l r k = b l , q r k , p , q V , p q , l V
Capacity Constraint: (5) ensures that the number of PHYs allocated to the services has a sufficient calendar slot.
e ij P r k n r k , e ij T S N f r a m e T S m a x
Slot Uniqueness Constraint: (6) indicates that the slot of each link is allowed to carry only one data code block at a certain moment, and the variable δ e j , P r i is equal to 1 if the service r k contains link e i j and is equal to 0 for otherwise, while the variable U e i j ( t ) is equal to 1 if the tth slot is occupied and is equal to 0 for otherwise.
r k R δ e i j , p r k U e ij ( t ) 1 , i , j V , e i j P r k , t [ 1 , n r k , e i j T S ]

4. Proposed PMFAN-GDSO Approach

4.1. Overview of PMFAN-GDSO

In response to the high dynamism exhibited by FlexE client services, a traffic-driven proactive slot orchestration algorithm, denoted as PMFAN-GDSO, has been proposed. This algorithm comprises two distinct components: the proactive multi-flow assisted attention-based neural network (PMFAN) for traffic prediction, and the genetic-based deterministic slot orchestration algorithm (GDSO) for slot orchestration. These two algorithms operate in accordance with the logic depicted in Figure 2. Initially, PMFAN leverages the historical data pertaining to service bandwidth fluctuations provided in the database. This encompasses both the historical data of the target business and that of background businesses. Through this, it derives a predictive value for the current business bandwidth at the subsequent time instance. This value signifies the potential bandwidth fluctuations that may manifest in the ensuing time step, effectively furnishing the timeslot allocation module with proactive bandwidth information. Subsequently, GDSO utilizes this information to compute the requisite number of timeslots for the impending time step. Upon the arrival of the next time step, an assessment is made to ascertain whether the traffic aligns with the predicted scenario. Should it do so, the slot allocation scheme is promptly executed, obviating the need for computing this plan. Conversely, if it deviates, the process reverts to the prior step, recalculating the slot allocation scheme based on the actual bandwidth. The former results in a reduction in the computational time required for a timeslot scheme, while the latter is consistent with the delay incurred in re-computation when alterations in the business bandwidth occur.

4.2. Traffic Flow Prediction Based on PMFAN

Our proposed proactive multi-flow assisted attention-based neural network, as depicted in Figure 3, comprises two main components: the correlation extraction model with multi-flow assistance, and the temporal feature extraction module with attention mechanism.

4.2.1. Temporal Feature Extraction with Attention Mechanism

The service traffic bandwidth nowadays in optical networks exhibits complex nonlinear characteristics due to factors such as bursts and randomness. Traditional mathematical modeling algorithms, such as ARIMA, struggle to provide accurate predictions. Therefore, we have chosen to utilize an RNN composed of LSTM units as the foundation of our prediction model. This network can capture the long-term dependencies between the service traffic bandwidth data, capture the nonlinear features in the bandwidth sequence via its nonlinear components, and mitigate the issues of gradient vanishing and explosion during long sequence prediction. In order to further improve the accuracy of traffic prediction, considering that the prediction target often exhibits stronger correlations with specific time steps in the historical data, we introduce an attention mechanism to identify these time steps with stronger correlations and assign them higher weights, thus enhancing the accuracy of the prediction results.
The core concept of LSTM is the gate mechanism that governs the flow of information. The input, output, and forget gates collectively determine the cell state C t of an LSTM memory unit. The forget gate determines what information from the previous cell state C t 1 will be retained. The input gate controls the current time information x t and the hidden state h t 1 . The output gate O t , together with the cell state C t , determines the output vector h t .
The output vector h t , the input gate i t , the forget gate ft and the output gate O t can be defined as:
h t = O t tanh ( C t )
i t = σ ( W i [ h t 1 , x t ] + b i )
f t = σ ( W f [ h t 1 , x t ] + b f )
O t = σ ( W o [ h t 1 , x t ] + b o )
where the activation function is defined as σ   W i   W f   W o   b o   b i   b f , which are the learnable parameters of the LSTM and will be defined when training.
The implementation of the attention mechanism revolves around adjusting the weight coefficients assigned to the data. Each data point is associated with specific weight coefficients. By calculating and modifying these weight coefficients, the attention mechanism can represent various key information. As a result, the model’s ability to capture important information is enhanced, leading to an improved prediction accuracy. The combination of the attention mechanism and LSTM is illustrated in Figure 3. The attention mechanism allows the model to dynamically focus on the most relevant time steps in the historical data, paying close attention to them for the prediction task. This enables the model to effectively capture the temporal dependencies and patterns that are crucial for accurate traffic forecasting.

4.2.2. Correlation Extraction with Multi-Flow Assistance

The accuracy of the prediction model is crucial for the overall effectiveness of our algorithm as it can effectively reduce the computational latency of the orchestration algorithm. Therefore, we incorporate inter-service correlation information to push the prediction performance of the model to its limits. We aggregate multiple client flows into a three-dimensional matrix and use a CNN to extract the correlations between the business flows. DenseNet is chosen as our model for extracting business correlations due to its excellent performance in both feature reuse and alleviating gradient vanishing.
The principle of DenseNet is to establish dense connections between different layers, which not only avoids overfitting but also further improves the prediction accuracy of the model, making the network training more stable and reliable. The position of DenseNet in the prediction model is shown in Figure 3. The main component of DenseNet is the densely connected BasicConv layer, where each layer is directly connected to all subsequent layers. The structure of BasicConv is also depicted in Figure 3, consisting of batch normalization, followed by ReLU activation and a convolutional layer with a 3 × 3 kernel. Assuming the input to the network is represented as X0, the inputs to all layers can be represented as follows:
X l = f l ( X 0 X 1   X l 1 )
where refers to the concatenation of the feature maps from all preceding layers. The introduction of batch normalization in the non-linear transformation aims to enhance the stability of the network during the parameter propagation process, accelerate the model training, and potentially improve the model prediction accuracy.
The convolution operation in DenseNet can be represented in the following form:
X ( α + 1 ) = f ( W ( α ) X ( α ) + b ( α ) )
where refers to the convolutional function, W represents the weights, and b represents the bias. Here, we chose the ReLU function as the activation function f , which is defined as f ( a ) = m a x ( 0 , a ) .

4.3. Calendar Orchestration Algorithm

The algorithm proposed includes two parts: the method to solve the end-to-end routing sub-problem and the method to solve the slot allocation sub-problem. The selection of background parameters is crucial. Different combinations of background parameters have a huge impact on the algorithm performance. Therefore, the background parameters will be set during the initialization process. For the optical network, G supports the fine granularity FlexE interfaces in which the initial pheromone concentration of each link is set to λ ( i , j , 1 ) = λ 0 ( i , j V ) . As one of the crucial parameters in the routing sub-problem, pheromone is the key to ensuring rapid convergence. And, the propagation delay matrix is set to ( τ i , j p r o p ) M × M . In the following, the two sub-problems will be elaborated in detail.

4.3.1. Routing Sub-Problem

When the ant q moves from the s k to the d k , the candidate node will calculate the state transfer function according to (13) and select both the node with the highest pheromone concentration and the largest expected value from a l l o w e d q .
F q ( i , j , w ) = { [ λ ( i , j , w ) ] α [ u ( i , j , w ) ] β k   allowed   q [ λ ( i , k , w ) ] α [ u ( i , k , w ) ] β , j   allowed   q 0 , j   allowed   q
where α represents the importance affecting the pheromone, it influences the path decision by setting different pheromone concentrations; β represents the importance of visibility; and u ( i , j , w ) represents the expectancy heuristic factor. In addition, a l l o w e d q = { V tabu q } ( q [ 1 , N a n t ] ) denotes the candidate node sets, and tabu q is a table record of the nodes that have already traveled.
However, the gap between α is slight in the early stage of the algorithm, which makes it impossible to quickly guide the path finding based only on expectations. Secondly, (13) only considers the concentration impact and the visibility on the path, but it can easily cause the algorithm fall into the local optimum in the later stages. Therefore, we introduce the slot allocating delay into u(i, j, w) to accelerate the path optimization while avoiding falling into the local optimal solution. The calculation of u ( i , j , w ) is shown in (14).
u ( i , j , w ) = { ( τ e i j p r o p ) 1 ,                                                                                     i = s ( τ e i j p r o p + τ v j f o r w a r d + τ e i j a l l o c a t e ) 1 , i s
where τ e i j p r o p is the propagation delay in the path P r k , τ v j forward is the slot forwarding the delay of each forward node, and τ e i j a l l o c a t e is the slot allocating the delay triggered by the conflict slot.

4.3.2. Slot Allocation Sub-Problem

When q passes through the first link, we use the First Fit method to find the first available slot and allocate the slot resources required equally. Then, we noted the slot location of the first link as N e 1 T S = { t s ( 1 ) , t s ( 2 ) , , t s ( n e j T S ) } , where n e j T S = B i b 0 . The position of slot t ( t [ 1 , n r k , e j T S ] ) occupied by service r k at the link e j ( j > 1 ) can be represented by (15).
N e j T S ( t ) = N e j 1 T S ( t ) + [ j = 1 j 1 ( τ e j p r o p + τ v j c r o s s ) ] % T F r a m e T t s
where T t s is the dispatching delay of a sub-slot, and T F r a m e represents the FlexE frame period.

4.3.3. Pheromone Update

The behavioral characteristics of ant colonies show that when ants are blocked or reach their destination, a part of the pheromone will be left on each link, and the pheromone will evaporate over time to update the pheromone concentration of the whole topology. The iterative equation of the global pheromone concentration is shown in (16), and the computation of Δ λ q ( i , j , ω ) is shown in (17).
λ ( i , j , ω + 1 ) = ( 1 ρ ) λ ( i , j , ω ) + q = 1 N a n t Δ λ q ( i , j , ω )
Δ λ q ( i , j , ω ) = { Q ( τ e i j p r o p ) 1 ,                                     i = s Q ( τ e i j prop + τ e i j allocate ) 1 , i s
where ρ is an important parameter that affects λ . Setting a reasonable ρ can improve the global optimization capabilities. In (17), Δ λ q ( i , j , ω ) denotes the concentration of the pheromone left on the link e i j by the ant q during the ω   th path finding.

5. Result

In this section, we first describe the settings of the experimental parameters and the simulation environment. Then, we perform a large number of simulations to evaluate the performance and feasibility of the research.
All simulations were performed on servers with Core i7 CPU and 32 GB RAM. In order to evaluate the algorithm performance in a complex and real network topology, we conducted subsequent experimental work based on the optical backbone network topology in a certain place. The topology is shown in Figure 4. In the follow-up experiments, we also generated several networks of different sizes, but with the same structure using a scale-free algorithm randomly, setting the size of the network nodes to n ( n = 32, 50, 68, 88, 102, 118).
In this paper, a dataset provide by China Telecom is used for validation in our experiment. To show the effectiveness of the model, this paper also used the non-deep learning method and other deep learning methods, which have proved to have good performance, as a comparative test. The specific models utilized in this paper include LSTM, GRU, and 3DCNN, all of which possess the capability of multi-step prediction. Multi-step prediction refers to the process of forecasting multiple future time steps based on the available input data. The ability to perform multi-step prediction ensures that the prediction time can be advanced, thereby guaranteeing that the allocation algorithm completes its execution before the predicted time point arrives.
In this paper, we used two evaluation indicators to compare the performance of different service traffic prediction models. They are the R-Square ( R 2 ) and Normalized Mean Squared Error (NMSE), and their definition are as follows in (18) and (19).
R 2 = 1 i = 1 N ( y ^ i y i ) 2 i = 1 N ( y ¯ i y i ) 2
NMSE = i = 1 N | y ^ i y i | 2 i = 1 N | y i | 2
Table 1 shows the indicators of the prediction results of different models. ARIMA, as a traditional linear model, is limited in capturing the nonlinear features, which leads to a subpar prediction performance, with R 2 being only 0.826654. In contrast, LSTM and its variant GRU, with their ability to capture nonlinear features, have significantly improved the accuracy of predictions, with R 2 being improved by 9% and NMSE being reduced by 48%. The 3DCNN, by incorporating inter-service correlations, also contributes to improving R 2 to 0.925660 and reducing NMSE to 0.018534. Our proposed model effectively leverages features from both the inter-services correlation and temporal dimensions, leading to the best performance, where R 2 reaches to 0.935641 and NMSE drops to 0.016684.
Figure 5 illustrates the predicted traffic volumes for two different services. It is a direct representation of our prediction results, which shows that the predicted results align with the distinct traffic patterns exhibited by each service. This indicates that the proposed model is capable of simultaneously capturing and fitting various traffic patterns associated with different types of services.
A schematic diagram of the calendar resource orchestration process is shown in Figure 6, Among it, Figure 6a shows the calendar orchestration process without the traffic prediction strategy, and Figure 6b shows the process based on the multi-flow prediction strategy.
In Stage 1, the services’ bandwidth updates based on two strategies: one is (a) without prediction, and the other (b) uses Multi-Flow Based Prediction. In stage 2, scheme (a) will calculate the calendar allocation result of the current moment, and scheme (b) will performs both processes simultaneously, where one performs the calendar calculation process on the current moment, and the other performs the next stage process based on the PFAMD algorithm. In stage 3, scheme (a) will calculate the calendar allocation based on the real service bandwidth updates, and based on the successful prediction, scheme (b) allocates time slots directly according to the pre-allocation result calculated in stage 2. In stage 4, scheme (a) will allocate all of the calendar the service needs based on the calculation results calculated in stage 3, but scheme (b) has completed the service allocation.
By analyzing lots of experimental data, we can prove that, via a large number of repeated experiments, the calendar allocation scheme based on the PMFAN-GDSO can save up to 46.8% of computing time while ensuring prediction accuracy.
In our algorithm, the actual delay generated via the allocation algorithm can be obtained using (20):
Saved   Time   = [ 1 i = 1 N ( y ^ i y i ) 2 i = 1 N ( y ¯ i y i ) 2 ] × T G D S O
With the increasing number of network nodes, the running time T_GDSO of the allocation algorithm also increases. Therefore, as the network scale expands, our algorithm can effectively save computation time. As shown in Figure 7, the curve demonstrates a gradual increase in time savings as the number of nodes increases. When the number of nodes increases from 32 to 118, the time savings increase by a factor of 5. However, when observing each individual topology structure, our algorithm seems to have a fixed proportion in terms of time savings, which is approximately 50%. This is because the traffic dataset and network topology we utilize are independent of each other. As a result, employing the same prediction algorithm yields a similar prediction accuracy, leading to a convergence in the gain ratios across different topologies. However, there are significant differences in the running time of the underlying allocation algorithm for different topology structures, which allows our algorithm to save the running time of the allocation algorithm significantly. This finding underscores the practicality and efficiency of employing traffic prediction in time slot resource allocation, offering a promising approach for enhancing resource management in dynamic network environments. The demonstrated computational efficiency is critical for ensuring timely and responsive network operations, especially in scenarios with rapidly evolving traffic patterns and dynamic node configurations.

6. Conclusions

This paper addresses the issues of service interruptions and fluctuations induced by dynamic service demands, with a primary focus on the traffic-driven proactive slot with a fine-granularity orchestration problem. Firstly, the traffic driven Flexible Ethernet slot orchestration model is established, followed by the introduction of a proactive multi-flow-assisted attention-based neural network (PMFAN) algorithm. This algorithm is employed for predicting the actual FlexE client service bandwidth, enabling the advanced adjustment of slot allocation schemes based on these predictions, and thus, saving computational time. Lastly, a slot orchestration algorithm is proposed, based on the genetic-based deterministic slot orchestration algorithm (GDSO). The proposed methods are capable of adapting to fluctuations in client flows while ensuring resource utilization efficiency.

Author Contributions

Conceptualization, R.G.; Data curation, Z.W., Z.Z. and W.Z.; Methodology, Q.H., X.Q. and G.S.; Project administration, R.G.; Resources, Q.H. and G.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (61971070), P. R. China.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Filer, M.; Gaudette, J.; Ghobadi, M.; Mahajan, R.; Issenhuth, T.; Klinkers, B.; Cox, J. Elastic optical networking in the Microsoft cloud. IEEE/OSA J. Opt. Commun. Netw. 2016, 8, A45–A54. [Google Scholar] [CrossRef]
  2. IEEE 802.3TM-2018; Standard for Ethernet. IEEE: Piscataway, NJ, USA, 2018.
  3. Hofmeister, T.; Vusirikala, V.; Koley, B. How can flexibility on the line side best be exploited on the client side? In Proceedings of the 2016 Optical Fiber Communications Conference and Exhibition (OFC), Anaheim, CA, USA, 20–24 March 2016; pp. 1–3. [Google Scholar]
  4. Eira, A.; Pereira, A.; Pires, J.; Pedro, J. On the efficiency of flexible ethernet client architectures in optical transport networks [invited]. J. Opt. Commun. Netw. 2018, 10, A133–A143. [Google Scholar] [CrossRef]
  5. Lu, W.; Kong, J.; Liang, L.; Liu, S.; Zhu, Z. How Much Can Flexible Ethernet and Elastic Optical Networking Benefit Mutually? In Proceedings of the ICC 2019–2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  6. Optical Internetworking Forum, Flex Ethernet 2.0 Implementation Agreement. June 2018. Available online: https://www.oiforum.com/wp-content/uploads/2019/01/OIF-FLEXE02.0-1.pdf (accessed on 7 February 2023).
  7. ITU-T G.709/Y.1331; Amendment 1 (11/2016): Interfaces for the Optical Transport Network Amendment. ITU: Geneva, Switzerland, 2016.
  8. Zhang, S.; Zhong, Q.; Zha, M.; Zuo, T. Hybrid Multiplexing over FlexE Group. In Proceedings of the 2018 23rd Opto-Electronics and Communications Conference (OECC), Jeju Island, Republic of Korea, 2–6 July 2018; pp. 1–2. [Google Scholar] [CrossRef]
  9. Liao, Y.; Hashemi, S.A.; ElBakoury, H.; Cioffi, J.; Goldsmith, A. Calendar Allocation Based on Client Traffic in the Flexible Ethernet Standard. In Proceedings of the ICC 2020–2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  10. Vilalta, R.; Martínez, R.; Casellas, R.; Muñoz, R.; Lee, Y.; Fei, L.; Tang, P.; López, V. Network Slicing Using Dynamic Flex Ethernet over Transport Networks. In Proceedings of the 2017 European Conference on Optical Communication (ECOC), Gothenburg, Sweden, 17–21 September 2017; pp. 1–3. [Google Scholar] [CrossRef]
  11. ITU-T Y Suppl. 44 (07/2017); Standardization and Open Source Activities Related to Network Softwarization of IMT-2020. IT Standardization. ITU: Geneva, Switzerland, 2017.
  12. Koulougli, D.; Nguyen, K.K.; Cheriet, M. Joint Optimization of Routing and Flexible Ethernet Assignment in Multi-layer Multi-domain Networks. In Proceedings of the 2020 29th International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 3–6 August 2020; pp. 1–9. [Google Scholar] [CrossRef]
  13. Zhu, P.; Cui, J.; Ji, Y. A Built-in Hash Permutation Assisted Cross-layer Secure Transport in End-to-End FlexE over WDM Networks. In Proceedings of the GLOBECOM 2020–2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–5. [Google Scholar] [CrossRef]
  14. Koulougli, D.; Nguyen, K.K.; Cheriet, M. Efficient Routing Using Flexible Ethernet in Multi-Layer Multi-Domain Networks. J. Light. Technol. 2021, 39, 1925–1936. [Google Scholar] [CrossRef]
  15. Liang, H.; da Fonseca, N.L.S.; Zhu, Z. On the Cross-Layer Network Planning for Flexible Ethernet over Elastic Optical Networks. IEEE Trans. Netw. Serv. Manag. 2021, 18, 3691–3705. [Google Scholar] [CrossRef]
  16. Koulougli, D.; Nguyen, K.K.; Cheriet, M. Flexible Ethernet Traffic Restoration in Multi-layer Multi-domain Networks. In Proceedings of the ICC 2021–IEEE International Conference on Communications, Xiamen, China, 28–30 July 2021; pp. 1–6. [Google Scholar] [CrossRef]
  17. Zhu, P.; Cui, J.; Ji, Y. Universal Hash Based Built-In Secure Transport in FlexE Over WDM Networks. J. Light. Technol. 2021, 39, 5680–5690. [Google Scholar] [CrossRef]
  18. Wu, M.; Da Fonseca, N.L.S.; Zhu, Z. Dynamic Cross-Layer Restoration to Resolve Packet Layer Outages in FlexE-Over-EONs. IEEE Trans. Netw. Serv. Manag. 2022, 19, 2600–2611. [Google Scholar] [CrossRef]
  19. Gu, R.; Xue, Y.; Zhang, Y.; Wang, Z.; Zhang, H.; Yang, Y.; Li, Y.; Ji, Y. Routing and Timeslot Scheduling for SPN Fine-Granularity Slices. Photonics 2023, 10, 126. [Google Scholar] [CrossRef]
  20. Wang, J. A process level network traffic prediction algorithm based on ARIMA model in smart substation. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2013), Melmaruvathur, India, 3–5 April 2013; pp. 1–5. [Google Scholar]
  21. Oliveira, T.P.; Barbar, J.S.; Soares, A.S. Computer network traffic prediction: A comparison between traditional and deep learning neural networks. Int. J. Big Data Intell. 2016, 3, 28–37. [Google Scholar] [CrossRef]
  22. Trinh, H.D.; Giupponi, L.; Dini, P. Mobile Traffic Prediction from Raw Data Using LSTM Networks. In Proceedings of the 2018 IEEE 29th Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Bologna, Italy, 9–12 September 2018; pp. 1827–1832. [Google Scholar]
  23. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  24. Lazaris, A.; Prasanna, V.K. Deep Learning Models for Aggregated Network Traffic Prediction. In Proceedings of the 2019 15th International Conference on Network and Service Management (CNSM), Halifax, NS, Canada, 21–25 October 2019; pp. 1–5. [Google Scholar]
  25. Zhang, L.; Zhang, H.; Tang, Q.; Dong, P.; Zhao, Z.; Wei, Y.; Mei, J.; Xue, K. LNTP: An End-to-End Online Prediction Model for Network Traffic. IEEE Netw. 2021, 35, 226–233. [Google Scholar] [CrossRef]
  26. Gui, Y.; Wang, D.; Guan, L.; Zhang, M. Optical Network Traffic Prediction Based on Graph Convolutional Neural Networks. In Proceedings of the 2020 Opto-Electronics and Communications Conference (OECC), Taipei, Taiwan, 4–8 October 2020; pp. 1–3. [Google Scholar]
  27. Vinchoff, C.; Chung, N.; Gordon, T.; Lyford, L.; Aibin, M. Traffic Prediction in Optical Networks Using Graph Convolutional Generative Adversarial Networks. In Proceedings of the 2020 22nd International Conference on Transparent Optical Networks (ICTON), Bari, Italy, 19–23 July 2020; pp. 1–4. [Google Scholar]
  28. Lin, J.; Chen, Y.; Zheng, H.; Ding, M.; Cheng, P.; Hanzo, L. A Data-driven Base Station Sleeping Strategy Based on Traffic Prediction. IEEE Trans. Netw. Sci. Eng. 2021. [Google Scholar] [CrossRef]
  29. Zhang, C.; Zhang, H.; Yuan, D.; Zhang, M. Citywide Cellular Traffic Prediction Based on Densely Connected Convolutional Neural Networks. IEEE Commun. Lett. 2018, 22, 1656–1659. [Google Scholar] [CrossRef]
  30. Huang, C.W.; Chiang, C.T.; Li, Q. A study of deep learning networks on mobile traffic forecasting. In Proceedings of the 2017 IEEE 28th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Montreal, QC, Canada, 8–13 October 2017; pp. 1–6. [Google Scholar]
  31. Li, M.; Wang, Y.; Wang, Z.; Zheng, H. A deep learning method based on an attention mechanism for wireless network traffic prediction. Ad Hoc Netw. 2020, 107, 102258. [Google Scholar] [CrossRef]
  32. Cui, Y.; Huang, X.; Wu, D.; Zheng, H. Machine Learning based Resource Allocation Strategy for Network Slicing in Vehicular Networks. In Proceedings of the 2020 IEEE/CIC International Conference on Communications in China (ICCC), Chongqing, China, 10–12 August 2020; pp. 454–459. [Google Scholar]
  33. Poryazov, S.; Andonov, V.; Saranova, E. Comparison of Conceptual Models of Overall Telecommunication Systems with QoS Guarantees. In Flexible Query Answering Systems; FQAS 2017. Lecture Notes in Computer Science; Christiansen, H., Jaudoin, H., Chountas, P., Andreasen, T., Legind Larsen, H., Eds.; Springer: Cham, Switzerland, 2017; Volume 10333. [Google Scholar] [CrossRef]
  34. Smith, J.; Johnson, A.; Brown, K. Generalized nets for network traffic prediction. J. Netw. Anal. 2018, 5, 87–103. [Google Scholar]
  35. Li, H.; Wang, Q. Network traffic prediction using generalized nets and time series analysis. Int. J. Comput. Netw. Appl. 2019, 6, 45–56. [Google Scholar]
  36. Chen, S.; Zhang, L.; Liu, Y. Network traffic prediction using generalized nets and deep learning. IEEE Trans. Netw. Sci. Eng. 2020, 7, 123–136. [Google Scholar]
  37. Ullah, Y.; Roslee, M.B.; Mitani, S.M.; Khan, S.A.; Jusoh, M.H. A Survey on Handover and Mobility Management in 5G HetNets: Current State, Challenges, and Future Directions. Sensors 2023, 23, 5081. [Google Scholar] [CrossRef] [PubMed]
  38. Poryazov, S.A.; Saranova, E.T.; Andonov, V.S. Overall Model Normalization towards Adequate Prediction and Presentation of QoE in Overall Telecommunication Systems. In Proceedings of the 2019 14th International Conference on Advanced Technologies, Systems and Services in Telecommunications (TELSIKS), Nis, Serbia, 23–25 October 2019; pp. 360–363. [Google Scholar] [CrossRef]
Figure 1. The architecture of Flexible Ethernet.
Figure 1. The architecture of Flexible Ethernet.
Applsci 13 11398 g001
Figure 2. The framework of PMFAN-GDSO algorithm.
Figure 2. The framework of PMFAN-GDSO algorithm.
Applsci 13 11398 g002
Figure 3. The framework of proactive multi-flow assisted attention-based neural network.
Figure 3. The framework of proactive multi-flow assisted attention-based neural network.
Applsci 13 11398 g003
Figure 4. The real optical backbone network topology in specific place.
Figure 4. The real optical backbone network topology in specific place.
Applsci 13 11398 g004
Figure 5. Comparison of PMFAN algorithm and ground truth on different FlexE client flow.
Figure 5. Comparison of PMFAN algorithm and ground truth on different FlexE client flow.
Applsci 13 11398 g005
Figure 6. Comparison of proposed algorithm and proposed without prediction on slot allocation.
Figure 6. Comparison of proposed algorithm and proposed without prediction on slot allocation.
Applsci 13 11398 g006
Figure 7. Comparison of computing time savings on different topologies.
Figure 7. Comparison of computing time savings on different topologies.
Applsci 13 11398 g007
Table 1. Evaluation indicators of different prediction algorithm.
Table 1. Evaluation indicators of different prediction algorithm.
IndicatorsARIMAGRULSTM3DCNNPMFAN
R 2 0.8266540.8945640.9046480.9256600.935641
NMSE 0.0435990.0243540.0225470.0185340.016684
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, Q.; Wang, Z.; Qin, X.; Zhou, Z.; Su, G.; Zhu, W.; Gu, R. Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces. Appl. Sci. 2023, 13, 11398. https://doi.org/10.3390/app132011398

AMA Style

Hu Q, Wang Z, Qin X, Zhou Z, Su G, Zhu W, Gu R. Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces. Applied Sciences. 2023; 13(20):11398. https://doi.org/10.3390/app132011398

Chicago/Turabian Style

Hu, Qian, Zixuan Wang, Xin Qin, Zexi Zhou, Guangnan Su, Wenwu Zhu, and Rentao Gu. 2023. "Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces" Applied Sciences 13, no. 20: 11398. https://doi.org/10.3390/app132011398

APA Style

Hu, Q., Wang, Z., Qin, X., Zhou, Z., Su, G., Zhu, W., & Gu, R. (2023). Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces. Applied Sciences, 13(20), 11398. https://doi.org/10.3390/app132011398

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop