Next Article in Journal
Bidirectional Propagation Properties of Partially Coherent Laguerre–Gaussian Beams in Atmospheric Turbulence and Plasma
Previous Article in Journal
Optimization Design and Simulation of Coin-Slot-Type Anti-Resonant Fiber Structure for 2 μm Transmission
Previous Article in Special Issue
DeepChaos+: Signal Detection Quality Enhancement of High-Speed DP-16QAM Optical Fiber Communication Based on Chaos Masking Technique with Deep Generative Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Artificial Neural Networks to Evaluate the Capacity and Cost of Multi-Fiber Optical Backbone Networks

1
Department of Electrical and Computer Engineering, Instituto Superior Técnico, Universidade de Lisboa, Avenida Rovisco Pais 1, 1049-001 Lisboa, Portugal
2
Department of Electrical and Computer Engineering and Instituto de Telecomunicações, Instituto Superior Técnico, Universidade de Lisboa, Avenida Rovisco Pais 1, 1049-001 Lisboa, Portugal
*
Author to whom correspondence should be addressed.
Photonics 2024, 11(12), 1110; https://doi.org/10.3390/photonics11121110 (registering DOI)
Submission received: 14 October 2024 / Revised: 4 November 2024 / Accepted: 22 November 2024 / Published: 24 November 2024
(This article belongs to the Special Issue Machine Learning Applied to Optical Communication Systems)

Abstract

:
A possible solution to address the enormous increase in traffic demands faced by network operators is to rely on multi-fiber optical backbone networks. These networks use multiple optical fibers between adjacent nodes, and, when properly designed, they are capable of handling petabits of data per second (Pbit/s). In this paper, an artificial neural network (ANN) model is investigated to estimate both the capacity and cost of a multi-fiber optical network. Furthermore, a fiber assignment algorithm is also proposed to complement the network design, enabling the generation of datasets for training and testing of the developed ANN model. The model consists of three layers, including one hidden layer with 50 hidden units. The results show that for a large network, such as one with 100 nodes, the model can estimate performance metrics with an average relative error of less than 0.4% for capacity and 4% for cost, while achieving a computation time nearly 800 times faster than the heuristic approach used in network simulation. Additionally, the network capacity is around 5 Pbit/s.

1. Introduction

In recent years, data traffic has increased significantly, a trend expected to continue due to the growth of applications and services that require high bandwidth and generate large amounts of data. Examples include video streaming services, cloud computing, machine-to-machine applications, online gaming, and the adoption of emerging technologies like 5G and beyond and advanced artificial intelligence applications [1]. This evolving scenario places special requirements on the backbones of network operators, which could experience traffic flows between their nodes reaching tens of Tb/s in the medium term, and even up to hundreds of Tb/s in the long term [1]. This situation presents a significant challenge for the design of future optical networks, particularly their backbone segments.
Optical networks are communication infrastructures, owned by telecommunication operators (telcos) or internet companies (e.g., Google, Microsoft, Meta), that utilize light for transmission, processing, and routing information and rely on optical fibers as their transmission medium. A fundamental technology in the field of optical networking is Wavelength Division Multiplexing (WDM). WDM allows the simultaneous transmission of multiple optical signals (also designated as optical channels) on the same optical fiber, with each channel using a different wavelength. The number of optical channels that can be transmitted over an optical fiber is limited to about 100 when using the traditional C-band, restricting the maximum WDM transmission capacity to well below 100 Tb/s for significant distances [2]. To greatly increase the number of optical channels to cope with the enormous growth in bandwidth demand, one can rely on space division multiplexing (SDM) techniques. This approach can be implemented using a multi-fiber (MF) solution, i.e., multiple standard single-mode fiber pairs per link instead of just one, as it is typical, or, alternatively, advanced fibers such as multicore fibers or few-mode fibers, with both solutions still operating in the C-band [2]. By relying on these solutions, it is feasible to design petabit-class optical networks, which are networks capable of handling data at speeds reaching or exceeding one petabit per second (Pb/s) [3].
For designing MF networks, it is crucial to define, in addition to the traditional routing and wavelength assignment solutions, a strategy for allocating fibers to the network, specifically, a fiber assignment strategy. In [4], two approaches were proposed to optimize network capacity by adding extra fibers. In the first approach, fibers were added to links supporting the maximum number of traffic demands, while in the second, fibers were added to links exhibiting the highest number of adjacent demands. Furthermore, in [5], the idea is to add extra fibers to links that are responsible for blocking traffic demands due to spectrum exhaustion, with the goal of minimizing the number of fibers added.
Network capacity is a key performance metric in optical networks. This capacity can be defined as the maximum amount of data that the entire network can handle per unit of time, and it is closely related to the concept of channel capacity introduced by Claude Shannon in 1948 [6]. The estimation of network capacity is a challenging task because it depends not only on physical layer aspects related to optical fibers and other optical devices but also on networking aspects such as physical and logical topology, routing, as well as wavelength and modulation assignment. Consequently, it suffers from the hurdle of long computation times, especially when dealing with large-scale networks. Although the problem of predicting optical network capacity has been the focus of many studies, (see [7,8,9,10]), to the best of the authors’ knowledge, none of the published research has relied on machine learning (ML) techniques for this purpose, despite these techniques being widely used in the context of optical networks to address other problems [11,12,13]. The closest study is reported in [14], where a routing and wavelength assignment (RWA) problem is treated using ML techniques by transforming it into a multi-classification problem, which is then solved using logistic regression and deep neural network techniques. However, the network capacity estimation problem, although also involving RWA calculations, is more general than this. Furthermore, the complexity of the problem for MF networks is even higher due to the necessity of using fiber assignment techniques.
In this paper, we investigate the utilization of an ML solution, specifically an artificial neural network (ANN) model [12], to estimate both the capacity and cost metrics of an MF-based optical network capable of handling Pb/s of data, with the cost being defined as the total length of optical fiber required in the network. The goal is to determine whether it is possible to significantly speed up the computations of these two metrics in comparison with heuristic methods, while still achieving accurate results.
To generate the large sets of synthetic data needed to train and test the model, we used a tool previously developed by the authors [10]. This tool not only generates random network topologies that aim to mimic real optical backbone networks but also performs routing and fiber assignment operations on these networks using heuristics developed specifically for this purpose, including the fiber assignment algorithm that is described in this work, which is a crucial component of our methodology.
The rest of the paper is organized as follows: Section 2 reviews important aspects of network modeling and random network generation and explains how both network capacity and cost can be computed. It also describes the fiber assignment algorithm proposed here for allocating fibers in MF networks. Section 3 details the ANN model introduced in this work. Section 4 presents some simulation results and, finally, Section 5 summarizes and concludes the paper.

2. Network Aspects and Dataset Generation

2.1. Network Modeling

In an abstract way, an optical network can be described as an undirected weighted graph G V , E , with V = { v 1 ,   ,   v N } denoting a set of nodes and E = { e 1 ,   ,   e K } denoting a set of links, where N = V is the number of nodes and K = E is the number of links. In transparent optical networks, all node functionalities take place in the optical domain, and the nodes are built using reconfigurable optical add–drop multiplexers (ROADMs), which are responsible for switching optical channels between different fibers, among other functions. Interconnection between these elements and client equipment is achieved using transponders, which are devices responsible for mapping client signals into optical channels. Meanwhile, an optical link represents a physical interconnection between two nodes, implemented using a pair of optical fibers, along with optical amplifiers spaced appropriately to compensate for fiber losses. Note that in the case of MF networks, multiple pairs of fibers are used instead. Furthermore, each optical fiber supports WDM signals, meaning it carries a specific number of optical channels. Each link v i , v j E is characterized by three attributes: l i , j , the link length in kilometers between the nodes v i and v j ; n f i , j , the number of optical fiber pairs in the link; and u i , j , the link capacity measured in terms of the number of optical channels denoted as N c h . In this work, we assume that fiber transmission takes place in the extended C-band, which has a bandwidth of 4800 GHz, enabling the support of N c h , m a x = 75   channels, with a channel spacing of 64 GHz corresponding to a baud rate of 64 Gbaud.
In addition to N and K , other important parameters of the graph G are the node degree δ G , the network diameter d G , and the algebraic connectivity a G . δ G defines the number of links connected to a given node, d G is the length of the longest shortest path between any two nodes, and a G is the second smallest eigenvalue of the graph’s Laplacian matrix [15].
In the context of ANNs, it is necessary to have very large datasets for training and testing purposes. To achieve this, it is useful to be able to generate numerous network topologies, which can be performed through random graphs designed to adequately describe the characteristics of real-world optical networks. In [10], we described a tool that we developed to generate random networks appropriate for describing optical backbone networks. The tool is based on a modified Waxman model and can generate networks that are resilient to single-link failures. In a simplified way, this model works by dividing a two-dimensional (2D) square plane with area A = L 2 (L is the side length of the plane) into a set of regions. In these regions, N nodes are randomly placed, and then the nodes are interconnected with links according to the Waxman probability, which is characterized by the α and β parameters, both in the range 0 , 1 .

2.2. Routing, Fiber Assignment, Capacity, and Cost

Network capacity refers to the maximum amount of data that the network can theoretically handle per unit of time, typically measured in bits per second (bit/s). This metric depends on many network parameters, including the physical topology defined by the graph G V , E and the logical topology, which describes the way how the information flows between all the network nodes. The logical topology is defined by the traffic matrix T = [ t s , d ], where each entry t s , d represents a traffic demand, or in other terms, the volume of traffic flowing from a source node s to a destination node d , with s , d V . For each traffic demand t s , d , it is necessary to find a path in the graph G V , E , between node s and node d . This is the role of the routing process. The routing process can be implemented using rigorous mathematical techniques, such as integer linear programing (ILP), or heuristics, as an alternative [16]. As ILPs become computationally infeasible for large-scale networks, we have to rely on heuristics in this work, as the analysis of such networks is paramount.
When the number of channels N c h per fiber is limited, as in this work, the routing process is known as constrained routing and can lead to traffic demand blocking whenever no channels (wavelengths) are available on one or more links of the path. To overcome such a limitation, one can add more pairs of optical fibers per link as needed, as it is the case for MF networks. This leads to a new process referred to as unconstrained routing plus fiber assignment. This process can be implemented using the heuristics proposed in [10]. In a simplified way, the heuristic method first computes the shortest paths between each pair of nodes in G V , E using the Dijkstra algorithm, with distance as the metric. Traffic demands between node pairs are then prioritized according to a specific sorting strategy and routed along their designated paths. Each path is assigned a wavelength using a first-fit strategy, thereby forming an optical channel. Finally, an optical fiber is allocated to each channel using Algorithm 1, which is described below.
To generate the datasets required to train and test the ANN, we have applied the referred heuristics to the randomly generated networks using the modified Waxman model assuming a uniform traffic demand between all the network node pairs, which can be defined as
t s , d = 1 s d 0 s = d .
Taking into account the traffic matrix T = [ t s , d ] of size N × N , and assuming that u i , j = , we can apply unconstrained routing to each network graph G V , E to compute the list of established paths P = π s , d , with the path π s , d having the length l π s , d = i , j l i , j . Additionally, we compute the link wavelength matrix, W = w i , j , which is also a N × N matrix, where w i , j is the list of all the wavelengths λ k present in the link i , j ,   i . e .   , w i , j = λ k . As referred to before, fiber assignment is a central process in MF networks. To implement this process, we propose Algorithm 1, which allows us to obtain the fiber matrix N F = n f i , j , representing the number of fibers per link, taking into account that the maximum number of wavelengths per link is N λ , m a x = N c h , m a x .
Algorithm 1: Fiber Assignment
Input:  graph   G V , E ,   wavelength   matrix   W = w i , j ,   number   of   wavelengths   N λ , m a x .
Output:   fiber   matrix   N F = n f i , j .
1: Initialize   N F , with   n f i , j = 0 , i , j E .
2:for  each   pair   of   nodes   i , j in W  do
3: if  G   has   an   edge   i , j   then
4: if  there   are   no   wavelengths   used   in   i , j ,   i . e . ,   w i , j = 0  then
5: n f i , j     1                                                                                 : At least one fiber is required
6: else
7: normalized   wavelengths       w i , j   mapped   into   the   range   1   to   N λ , m a x
num_fibres ← maximum number of wavelengths repetitions in normalized wavelengths
n f i , j     num_fibres
8: end if
9: else
10: n f i , j     0                                                                                                                 : Case there is no edge (i,j)
11: end if
12:end for
13:return   N F
Note that with the unconstrained routing, the number of wavelengths in each link is not limited, so the value assigned to a given λ k can be any natural number, in contrast to constraint routing, where it is bounded by N λ , m a x . In the algorithm, to determine the number of fibers needed in each link, the maximum number of “repeated wavelengths” in that link must be determined. A wavelength is considered a “repeated wavelength” when its value modulo N λ , m a x (where the modulo operation returns the remainder after division) is equal to that of another wavelength also present in that link. For instance, if N λ , m a x is 75, then wavelengths 1 and 76 are “repeated” because 76 modulo 75 equals 1. This implies that both wavelengths would occupy the same channel in a link, hence they are “repeated”. This concept is crucial in determining the number of fibers needed for a link, ensuring that each “repeated” wavelength has its own fiber. Finding the maximum count of “repeated wavelengths” will ensure that there are enough fibers to accommodate all the wavelengths, thus assuring that there are no channels with the same wavelength on the same fiber.
By knowing the length of the path π s , d ,   l π s , d , it is possible to compute its maximum capacity value, C π s , d , also denoted as the Shannon capacity, measured in bits per second. This calculation uses the optical reach values of the path (see Table 2 of [10]), where optical reach is defined as the maximum length of the path for which a certain value of the capacity can be achieved assuming a baud rate of 64 Gbaud. Furthermore, after obtaining the capacity of all the established paths, one arrives at the network capacity, which is given by
C n e t = s , d C ( π s , d ) .
Another important metric is the network’s cost. The overall cost of an optical network is the sum of the costs of all nodes and links, with node costs primarily driven by transponders and link costs by optical amplifiers. It is reasonable to assume that, in optical backbone networks, link costs are the dominant contributors to the network costs. As a result, these costs are predominantly determined by fiber length, since this parameter defines the number of optical amplifiers required [17]. In this case, the network cost is given by
Λ n e t = i , j l i , j × n f i , j .

3. Neural Network Design

An artificial neural network is a network of units, also called neurons, which are organized in multiple layers, including an input layer, a variable number of hidden layers, and an output layer. These layers operate in a fully connected way, meaning that each neuron of a given layer is connected to all the neurons of the next layer. Each neuron has a variable weight per input, denoted as ω m , i , with m defining the neuron position in a layer and i its input, which are summed together along with a bias term b m . The result of this operation is then passed through an activation function to obtain the output of that neuron. The activation function used in this study for the hidden layers is the ReLU (Rectified Linear Unit) function, which is given by
g x = max 0 ,   x
while for the output layer we have the linear activation function, that is,
g x = x .
Note that both activation functions are commonly used in regression problems, such as the one we are considering here [18].
The training of neural networks consists of determining the values for all Ω = ω m , i matrices and bias vectors B = b m that minimize a given loss function with a given iterative method (optimizer algorithm). For the training process, it is necessary to randomly generate a large number of datasets using the procedures described previously. Each dataset includes an array of inputs X = x 1 ,   x 2 x n , called features, and an array of outputs Y = y 1 ,   y 2 , obtained by network simulation, called labels. The features include the number of nodes, the number of links, the network diameter, the algebraic connectivity, and quantities such as the maximum, minimum, average, and variance of both link length and node degree. Furthermore, the labels include the network capacity y 1 = C n e t , given by (2), and the network cost y 2 = Λ n e t , given by (3).
In the training process, each dataset is split into a training set (the data used to determine the model’s parameters), a validation set (used to make an unbiased evaluation of the model’s performance during training), and a test set (used to assess the model’s performance after the training is complete). Before the data are split into these three sets, they need to be pre-processed and shuffled. Data pre-processing consists of preparing the data to make them more suitable for the training process.
The loss function is used to measure the difference between the value predicted by the ANN and the actual value obtained by simulation. In other words, it measures the error associated with the model’s predictions. For regression problems, the mean squared error (MSE) is commonly used as the loss function [18]. MSE can be expressed as follows:
M S E = 1 M i = 1 M y ^ i y i 2  
where M is the number of data values being considered, y ^ i are the estimated values, and y i are the actual values.
The optimizer algorithm is the method that determines how the weight matrices and bias vectors are updated during the training process. Common optimizers include the Stochastic Gradient Descent (SGD) and the Adaptive Momentum Estimation (Adam), with the former being used in this work. The updating of the network parameters requires the computation of the gradient of the loss function, a task performed by the backpropagation algorithm [19]. An important parameter related to the optimizer is the learning rate. This parameter determines the magnitude of the updates applied to the weights and biases during each iteration. Another important parameter is the batch size, which refers to the size of subsets into which the training data are divided. The dropout regularization can also be used to prevent overfitting, which occurs when a model learns the training data too closely but fails to make accurate predictions on the testing data.
A key aspect of training an ANN is optimizing the hyperparameters. Hyperparameters are the variables that configure how the model learns from the data. This includes the number of hidden layers, the number of units in each hidden layer, the learning rate, the batch size, and dropout regularization. During training, various hyperparameter combinations are tested to achieve the best performance on the validation set. This operation is called hyperparameter tuning.
In this work, the tuning operation is performed using the R 2 score metric, which is defined as
R 2 = 1 i = 0 M y i y ^ i 2   i = 0 M y i y ¯ i 2  
where y i is the actual value, y ^ i is the predicted value, y ¯ is the mean of the actual values, and M is the number of data values being considered. The R 2 score will take values between 0 and 1, where a value of 1 indicates that the model fits the data perfectly and 0 that it does not fit the data at all. That means that the closer the values are to 1, the better the model is performing [20].
From the hyperparameter tuning process, the ANN’s structure was defined (see Figure 1). The model that achieved the best performance on the validation set has one hidden layer with 50 hidden units ( m = 50 ), considering the number of features equal to 12 ( n = 12 ). The learning rate was optimized to 0.1, the batch size was set to 64, and no dropout regularization was needed.
This model structure and learning rate resulted in relatively high R 2 scores: 0.9994 for y 1 and 0.9962 for y 2 . Additionally, it has a relatively low number of trained parameters (the total number of weights and biases), with 752 parameters, which represents a good balance between model complexity and performance. To build and optimize the ANN model, we used the PyTorch 2.2 framework [21].

4. Simulation Results and Discussion

To train the ANN model, a set of 8480 networks was used. These networks were generated with the tool described in [10] considering a 2D square plane with side lengths varying from 1000 km to 5000 km in increments of 1000 km, the number of regions in the plane set to 4, the number of nodes varying from 5 to 100, the number of links varying from 5 to 231, and an average node degree varying from 2 to around 5. The Waxman parameters chosen were α = β = 0.4. Furthermore, the maximum number of channels per links was set to N c h , m a x   = 75.
Once the model is trained, the final step is to evaluate its performance through testing. For this purpose, a dataset of 1440 random networks was generated under the same conditions as those used to train the model. The network simulation took around 1 h and 16 min for the entire dataset, while the prediction time for the ANN model was just 11 milliseconds.
The mean relative errors for this test dataset, as defined by (8), are as follows: 2.47% for the network capacity ( y 1 ) and 5.29% for the total fiber cost (network cost) ( y 2 ) predictions. Figure 2 shows the scatter plot of the relative errors against the number of nodes for both outputs. Each blue dot represents the relative error (RE) for each individual network in the set, given by
R E = y i y ^ i y i
with y i being the value determined from the simulation solution and y ^ i the prediction made with the ANN model.
It was also shown that for the total network capacity (Figure 2a), 89.45% of the examples have a relative error below 5%, and 96.67% of the examples have a relative error below 10%. In the case of the total fiber cost (Figure 2b), 87.02% of the examples have a relative error below 10%, and 94.24% of the examples have a relative error below 15%. It can be seen that the model tends to perform better on networks with a higher number of nodes, while its performance is more irregular on networks with fewer nodes. A possible explanation for this behavior is that smaller networks might exhibit more variability in their features as well as in the relationships between features and labels, which makes it more challenging for the model to learn stable patterns that are crucial for making accurate predictions. This irregular performance of smaller-scale networks is particularly evident for the label “network cost” when the number of nodes is 10 or fewer. On the other hand, larger networks could be more homogeneous, exhibiting more uniform and consistent patterns that the model can learn and predict more effectively.
In order to analyze how the ANN model behaves with testing datasets that have a number of nodes outside the training range, we generated 3920 additional networks under the same conditions as the previous sets, but with the number of nodes ranging from 5 to 200. Generating this set took 55 h and 31 min, while the ANN model predicted the corresponding set in only 79 milliseconds. Figure 3 shows a scatter plot comparing the relative errors as a function of the number of nodes for this set of networks. The plots in Figure 3 show that the results are identical to those of Figure 2 when the number of nodes ranges from 5 to 100. However, outside this range, the model’s performance becomes unreliable, although it still performs quite well for up to about 115 nodes. The cause of this irregular behavior differs from that observed in small-scale networks and is due to the model being trained on a specific data range (number of nodes ranging from 5 to 100), and extrapolating beyond this range can lead to less reliable predictions.
The capability of a model to perform well in the presence of unseen data, which it was not trained on, is known as Out-of-Distribution (OOD) generalization [22]. This is a challenge that conventional supervised learning methods (such as ANNs) often find difficult to handle effectively as these types of models fundamentally assume that the training and test datasets originate from the same distribution. Note that addressing the OOD generalization problem is an active area of research in the field of ML [22].
Table 1 compares the results predicted by the ANN model for the total network capacity ( y ^ 1 ) and total fiber cost ( y ^ 2 ) with the corresponding results obtained by applying a heuristic approach to different random networks, using the tool described in [10], as well as Algorithm 1 for the fiber assignment task. These results show that the ANN models tend to have a good performance in the generated networks within this range of nodes, with the relative errors generally being low. Furthermore, the prediction times with the ANN are always significantly faster than the computation times obtained with the network simulation tool. For example, for a network with 100 nodes, the prediction time is about 17.1 milliseconds, whereas the computation time is about 13.5 s. This means that the ANN model is roughly 800 times faster than the heuristic approach, while achieving low relative errors of about 0.4% for network capacity and about 4% for network cost.
Remarkably, the network capacity for a number of nodes greater than or equal to 50 nodes exceeds 1 Pb/s, reaching about 5 Pb/s for 100 nodes. However, this comes at the cost of significantly increasing the required optical fiber length in the network, as this work is based on the MF paradigm, where additional fibers are added whenever a link reaches its maximum supported number of optical wavelengths.
A key point in the analysis is understanding how the ANN model performs on real optical network topologies, despite being trained on synthetic data generated from random networks. To address this point, Table 2 provides results for four real reference networks: COST239 ( N = 11 , K = 26 ,   l ¯ = 462.6   km ) [23], DTAG ( N = 14 ,     K = 23 ,     l ¯ = 236.5   km ) [9], NSFNET ( N = 14 ,     K = 21 ,     l ¯ = 1211.3   km ) [23], and UBN ( N = 24 ,     K = 43 ,     l ¯ = 993.2   km ) [23], with l ¯ being the average link length.
The results show that the ANN model predicts both outputs with low relative errors in the majority of cases and achieves computation times approximately 10 times faster than the heuristic method. However, there are instances where higher relative errors have been observed, with two cases exceeding a 10% relative error for the network cost: the NSFNET and UBN cases. Interestingly, these two cases correspond to the networks with larger average link lengths. An explanatory hypothesis for this behavior is that these networks exhibit significant variability in their features, making it more difficult for the ANN model to accurately capture the relationships between features and labels, a trend similar to the one shown in Figure 2b for networks with fewer than 40 nodes.

5. Conclusions

In this paper, the problem of estimating the capacity and cost of multi-fiber optical networks was addressed, using for this purpose an ANN model. These networks, by using multiple fiber pairs per link, can achieve very high network capacities, even in the order of petabits per second.
To generate the datasets required to train and test the ANN, we applied an appropriate heuristic that relies on a fiber assignment algorithm which was also proposed in the context of this work.
The implemented model was an ANN with 12 inputs (parameters related to the physical topology of the optical network), 2 outputs, and 1 hidden layer. The outputs correspond to two metrics: the network capacity, measured in Tbit/s, and the network cost, quantified by the total length of optical fiber deployed in the network, measured in km.
The ANN was trained with a number of nodes varying between 5 and 100, and it was extensively tested within the same range. The results showed good performance with a mean relative error of 2.47% and 5.29% for the first and second metric, respectively. The ANN model also showed significantly faster performance compared to the heuristic method, with the ANN predictions never taking more than a few tens of milliseconds, while the network simulation could take up to tens of seconds to reach the results in larger networks.
Remarkably, the network capacity for 50 or more nodes exceeds 1 Pb/s, reaching about 5 Pb/s for 100 nodes. However, this comes at the cost of a significant increase in the length of the total optical fiber required in the network.

Author Contributions

Conceptualization, A.F. and J.P.; methodology, A.F. and J.P.; software, A.F.; validation, A.F.; formal analysis, A.F. and J.P.; investigation, A.F. and J.P.; writing—original draft preparation, J.P.; writing—review and editing, J.P.; visualization, A.F.; supervision, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ruiz, M.; Hernández, J.A.; Quagliotti, M.; Salas, E.H.; Riccardi, E.; Rafel, A.; Velasco, L.; de Dios, O.G. Network traffic analysis under emerging beyond-5G scenarios for multi-band optical technology adoption. J. Opt. Commun. Netw. 2023, 15, F36–F47. [Google Scholar] [CrossRef]
  2. Winzer, P.J. The future of communications is massively parallel. J. Opt. Commun. Netw. 2023, 15, 783–787. [Google Scholar] [CrossRef]
  3. Furukawa, H.; Luís, R.S. Petabit-class optical networks based on spatial-division multiplexing technologies. In Proceedings of the 2020 International Conference on Optical Network Design and Modelling (ONDM), Barcelona, Spain, 18–21 May 2020. [Google Scholar] [CrossRef]
  4. Parker, M.C.; Wright, P.; Lord, A. Multiple fiber, flexgrid elastic optical network design using MaxEnt optimization. J. Opt. Commun. Netw. 2015, 7, B194–B201. [Google Scholar] [CrossRef]
  5. Lopez, V.; Zhu, B.; Moniz, D.; Costa, N.; Pedro, J.; Xu, X. Optimized design and challenges for C&L band optical line systems. J. Lightw. Technol. 2020, 38, 1080–1091. [Google Scholar] [CrossRef]
  6. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  7. Vincent, R.J.; Ives, D.J.; Savory, S.J. Scalable capacity estimation for nonlinear elastic all-optical core networks. J. Lightw. Technol. 2019, 37, 5380–5391. [Google Scholar] [CrossRef]
  8. Matzner, R.; Semrau, D.; Luo, R.; Zervas, G.; Bayvel, P. Making intelligent topology design choices: Understanding structural and physical property performance implications in optical networks. J. Opt. Commun. Netw. 2021, 13, D53–D67. [Google Scholar] [CrossRef]
  9. Ives, D.J.; Bayvel, P.; Savory, S.J. Routing, modulation, spectrum and launch power assignment to maximize the traffic throughput of a nonlinear optical mesh network. Photon. Netw. Commun. 2015, 29, 244–256. [Google Scholar] [CrossRef]
  10. Freitas, A.; Pires, J. Investigating the impact of topology and physical impairments on the capacity of an optical backbone network. Photonics 2024, 11, 342. [Google Scholar] [CrossRef]
  11. Musumeci, F.; Rottondi, C.; Nag, A.; Macaluso, I.; Zibar, D.; Ruffini, M.; Tornatore, M. An overview on application of machine learning techniques in optical networks. IEEE Commun. Surv. Tuts. 2019, 21, 1383–1408. [Google Scholar] [CrossRef]
  12. Gu, R.; Yang, Z.; Ji, Y. Machine learning for intelligent optical networks: A comprehensive survey. arXiv 2020, arXiv:2003.05290v1. [Google Scholar] [CrossRef]
  13. Morais, R.M.; Pedro, J. Machine learning models for estimating quality of transmission in DWDM Networks. J. Opt. Commun. Netw. 2018, 10, D84–D99. [Google Scholar] [CrossRef]
  14. Martín, I.; Troia, S.; Hernández, J.A.; Rodríguez, A.; Musumeci, F.; Maier, G.; Alvizu, R.; de Dios, O.G. Machine learning-based routing and wavelength assignment in software-defined optical networks. IEEE Trans. Netw. Serv. Manag. 2019, 16, 871–883. [Google Scholar] [CrossRef]
  15. Châtelain, B.; Bélanger, M.P.; Tremblay, C.; Gagnon, F.; Plant, D.V. Topological wavelength usage estimation in transparent wide area networks. J. Opt. Commun. Netw. 2009, 1, 196–203. [Google Scholar] [CrossRef]
  16. Santos, J.; Pedro, J.; Monteiro, P.; Pires, J. Optimized routing and buffer design for optical transport networks based on virtual concatenation. J. Opt. Commun. Netw. 2011, 3, 725–738. [Google Scholar] [CrossRef]
  17. Çetinkaya, E.K.; Alenazi, M.J.F.; Cheng, Y.; Peck, A.M.; Sterbenz, P.G. A comparative analysis of geometric graph models for modelling backbone networks. Opt. Switch. Netw. 2014, 14, 95–106. [Google Scholar] [CrossRef]
  18. Chugh, S.; Ghosh, S.; Gulistan, A.; Rahman, B. Machine learning regression approach to the nanophotonic waveguide analyses. J. Lightw. Technol. 2019, 37, 6080–6089. [Google Scholar] [CrossRef]
  19. Le Cun, Y. Theoretical framework for back-propagation. In Proceedings of the 1988 Connectionist Models Summer School; Morgan Kaufman: Pittsburg, PA, USA, 1988; pp. 21–28. [Google Scholar]
  20. Coefficient of Determination, R-Squared. Available online: https://www.ncl.ac.uk/webtemplate/ask-assets/external/maths-resources/statistics/regression-and-correlation/coefficient-of-determination-r-squared.html (accessed on 7 July 2024).
  21. Pytorch. Available online: www.archive.org (accessed on 20 April 2024).
  22. Liu, J.; Shen, Z.; He, Z.Y.; Zhang, X.; Xu, R.; Yu, H.; Cui, P. Towards out-of-distribution generalization: A survey. arXiv 2023, arXiv:2108.13624. [Google Scholar] [CrossRef]
  23. Pires, J.J.O. On the capacity of optical backbone networks. Network 2024, 4, 114–132. [Google Scholar] [CrossRef]
Figure 1. Model of the ANN network with 1 hidden layer.
Figure 1. Model of the ANN network with 1 hidden layer.
Photonics 11 01110 g001
Figure 2. Relative errors as a function of the number of nodes for both outputs of the ANN model (N ranging from 5 to 100): (a) total network capacity; (b) total fiber cost.
Figure 2. Relative errors as a function of the number of nodes for both outputs of the ANN model (N ranging from 5 to 100): (a) total network capacity; (b) total fiber cost.
Photonics 11 01110 g002
Figure 3. Relative errors as a function of the number of nodes for both outputs of the ANN model (N ranging from 5 to 200): (a) total network capacity; (b) total fiber cost.
Figure 3. Relative errors as a function of the number of nodes for both outputs of the ANN model (N ranging from 5 to 200): (a) total network capacity; (b) total fiber cost.
Photonics 11 01110 g003
Table 1. Accuracy of ANN prediction: y 1 : capacity; y 2 : cost.
Table 1. Accuracy of ANN prediction: y 1 : capacity; y 2 : cost.
N y 1 [Tb/s] y ^ 1   [Tb/s]RE (%) y 2 [103 km] y ^ 2 [103 km]RE (%)
1048.045.45.4424.4724.081.59
20303.2317.94.8614.7114.004.85
30708.0705.20.3927.0526.621.57
40803.0820.12.13122.27123.210.77
501244.21214.42.40231.63257.3611.1
601837.21937.35.45267.02247.977.14
703432.83393.91.13128.44131.792.61
803185.03197.80.40626.00597.004.63
905898.65864.00.59189.51194.542.66
1005394.65376.50.34718.09690.023.91
Table 2. Accuracy of DNN predictions in reference networks. y 1 : capacity; y 2 : cost.
Table 2. Accuracy of DNN predictions in reference networks. y 1 : capacity; y 2 : cost.
Network y 1 [Tb/s] y ^ 1   [Tb/s]RE (%) y 2 [103 km] y ^ 2 [103 km] RE (%)
COST23981.282.82.0124.0623.074.11
DTAG147.4145.91.0110.8810.950.69
NSFNET98.0104.16.1845.3938.6314.87
UBN272.8269.91.1085.42101.4218.73
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Freitas, A.; Pires, J. Using Artificial Neural Networks to Evaluate the Capacity and Cost of Multi-Fiber Optical Backbone Networks. Photonics 2024, 11, 1110. https://doi.org/10.3390/photonics11121110

AMA Style

Freitas A, Pires J. Using Artificial Neural Networks to Evaluate the Capacity and Cost of Multi-Fiber Optical Backbone Networks. Photonics. 2024; 11(12):1110. https://doi.org/10.3390/photonics11121110

Chicago/Turabian Style

Freitas, Alexandre, and João Pires. 2024. "Using Artificial Neural Networks to Evaluate the Capacity and Cost of Multi-Fiber Optical Backbone Networks" Photonics 11, no. 12: 1110. https://doi.org/10.3390/photonics11121110

APA Style

Freitas, A., & Pires, J. (2024). Using Artificial Neural Networks to Evaluate the Capacity and Cost of Multi-Fiber Optical Backbone Networks. Photonics, 11(12), 1110. https://doi.org/10.3390/photonics11121110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop