2.1. Experimental Setup
Figure 1a shows the implemented network structure. It was composed of nodes 1, 2 and 3 connected in a linear, unidirectional structure using two 19-core homogeneous single-mode MCFs of 30 km and 10 km. Each node was equipped with a 7-core joint circuit switch as well as an emulated 10-core joint packet switch.
Figure 1b shows the resource allocation map. The network was used to transport a 10 Tb/s cSSC composed of six polarization-division multiplexed 64-ary quadrature amplitude modulated (PDM-64QAM) signals over a total bandwidth of 150 GHz and occupying 7 spatial channels. In addition, we transmitted a 1 Tb/s line rate pSSC composed of ten 10 Gb/s on-off keying (OOK) signals over a bandwidth of approximately 1 THz and occupying 10 spatial channels.
Figure 1c shows the three-dimensional structure of the pSSC. Each carrier transported packets composed by a preample and payload. One of the carriers in one of the spatial channels carried the packet header.
For this demonstration, we assumed 4 transmission scenarios, as shown in
Figure 1a. In scenarios A and B, the cSSC was transmitted from node 1 (add) to node 2 (drop) and from node 2 (add) to node 3 (drop), respectively. In scenario C, the cSSC was transmitted from node 1 (add) with multicasting at node 2 and drop at node 3. Simultaneously, the network ran scenario D, establishing layer 3 ethernet connections between the 3 nodes with the pSSC.
Figure 2a shows a simplified diagram of the experimental setup used to test the network structure shown in
Figure 1a. Due to limitations in our laboratory, we implemented the proposed network by constructing two SDM-ROADMS with one acting as nodes 1 (east side) and 3 (west side) simultaneously, and the other acting as node 2. The nodes were connected via 19-core MCFs, whose profiles are shown in
Figure 2b. Between nodes 1 and 2 we used a 30 km fiber with cores positioned in a two ring configuration and a center core. Each core was surrounded by an index trench which lead to inter-core crosstalk between −40 dB and −45 dB. This fiber was spliced to a multi-core erbium-doped fiber amplifier (MC-EDFA), which acted as a booster on the east line side of node 1. Detailed descriptions of the fiber and amplifier may be found in reference [
10]. The link between nodes 2 and 3 was performed using a 10 km, 19-core MCF whose cores were also surrounded by index trenches to reduce inter-core crosstalk and were positioned in an hexagonal arrangement. This lead to crosstalk values between −30 dB and −35 dB. A detailed description of this fiber is included in reference [
11].
Figure 2b shows the profiles of the 30 km and 10 km 19-core MCFs used in this experiment and the selection of cores used to transport the cSSC and the pSSC. We selected these cores to reduce the crosstalk within each spatial super channel but made no attempt to reduce crosstalk between different spatial super channels. The nodes were connected to the 19-core MCFs using 3D-waveguides to act as SDM multiplexers and demultiplexers. To handle the 7-core 10 Tb/s cSSC, the corresponding outputs of the 19-core SDM demultiplexers were multiplexed into 7-core MCFs by free space couplers (FSC). Seven erbium doped fiber amplifiers (EDFAs) followed by variable optical attenuators (VOAs) were used to compensate for the loss of the FSCs. The outputs of the FSCs were connected to 7-core joint circuit switches, further detailed in
Figure 2c. These were composed of four 7-core MCF-pigtailed collimators aligned in a cross with a circular plate at the center. The plates were servo steered into one of 4 positions containing (A) a mirror for add/drop; (B) a half-mirror for multicasting; (C) a hole for through switching; and (D) a termination. The east sides of the switches were connected to the ROADMs’ east line sides via 7-core MCFs whose outputs were demultiplexed using a 3D waveguide coupler in the case of node 1, and a FSC in the case of node 2. In the case of node 1, the coupler outputs were multiplexed into the 19-core MC-EDFA by the east side 19-core SDM multiplexer. In the case of node 2, we used discrete EDFAs followed by VOAs preceding the 19-core SDM multiplexer. The add and drop from the joint circuit switches took the form of 7-core cSSCs carried by 7-core MCFs.
Figure 3a,b show the experimental setups used to produce and receive the 10 Tb/s cSSC, respectively. The output of a wideband frequency comb generator [
12] was filtered with a band-pass filter (BPF) to select 6 carriers with a frequency spacing of 25 GHz, centered between 1549.32 nm and 1550.32 nm. The selected carriers were amplified by an EDFA followed by a VOA for power control and split into even and odd carriers by an optical processor (OP). Each group of carriers was again amplified and sent to a dual-polarization dual-parallel Mach–Zehnder modulator (DP-MZM). The modulators were driven by arbitrary waveform generators (AWGs) to produce 24.5 Gbaud PDM-64QAM signals. Even and odd signals were combined at a 2 × 2 coupler with one of its outputs acting as the test signal. The other output of the coupler was amplified at a high power EDFA and split into 6 replicas. These acted as dummy spatial channels and were decorrelated and multiplexed with the test channel at a 3D waveguide SDM multiplexer. The combined throughput of the cSSC was 12.348 Tb/s which corresponded to a net throughput of 10.29 Tb/s, assuming a 20% overhead to support soft-decision forward error correction [
13]. The receiver was composed of an SDM demultiplexer followed by a manual switch to select the spatial channel under analysis. The selected spatial channel was then pre-amplified by an EDFA followed by a VOA and filtered with a BPF to limit the ASE power around the WDM channel under analysis. The resulting signal was detected by a polarization diverse coherent receiver whose outputs were digitized by a real-time oscilloscope operating at 80 GS/s. Digital signal processing was performed offline, and performance was estimated by direct error counting on 5 × 10 μs traces.
To handle the 1 Tb/s line rate pSSC, the 10 corresponding outputs of the west side 19-core SDM demultiplexers were directed to emulated 10-core joint packet switches, which are further detailed in
Figure 4a. Due to the limited amount of equipment in our laboratory, we were only able to implement the setup for the processing of one of the spatial channels composing the 10-core pSSC. As such, it was necessary to process each spatial channel in turn. The selection of the spatial channel to be processed was performed using a manual 1:10 switch. The selected spatial channel was directed to a 2 × 2 electro-absorption switch (EASW), which routed the incoming multi-wavelength packets either to the other line side or to a 100G packet transponder. The EASW is described in detail in reference [
14]. The add packets from the 100G packet transponder were directed by the EASW to a 1:10 power splitter to emulate the pSSC as a set of 10 replicas of the same 100G signal. The control of the switch was performed by a switch control module which took the label of one of the received spatial channels or the signals from the 100G transponder. Within the switch, burst-mode EDFAs [
15] were used to compensate for the EASW loss. Finally, the use of dispersion compensation fiber (DCF) was required to compensate the dispersion introduced on the transmission MCF. We note that this setup relied on the assumption that the inbound spatial channels were temporally aligned, which was possible due to the use of homogeneous MCFs. This point will be further discussed in the following sub-section.
Figure 4a also shows the internal structure of the 100G packet transponder. An electronic mapping and demapping circuit converted 10GEthernet packets into 10 × 10 Gb/s signals which were modulated onto 10 lightwaves by Mach-Zehnder modulators (MZM). The wavelengths of the lightwaves modulated by the MZMs were selected from the ITU-T grid with 100 GHz spacing and wavelength multiplexed by WDM multiplexers composed of arrayed waveguide gratings. The inverse operation was also carried out to take an incoming multi-wavelength packet and demultiplex it into 10 separate 10 Gb/s signals which were then directly detected by a set of 10 photodetectors. The resulting electrical signals were mapped into a 10GEthernet signal. One of the 10 Gb/s signals also carried the packet header to control the EASW. A detailed description of this system is available in reference [
15].
Figure 4b shows a trace of one of the 10 Gb/s signals when produced from a 64 byte 10GEthernet packet. The packet was composed of a preamble, used for packet detection and synchronization, followed by the payload. Before and after the packet, there was a guard time of 24 ns, set to align packets with different arrival times. The test traffic was generated by a commercial 10GEthernet network tester.
2.2. Analysis of the Inter-Core Skew
Although we used homogeneous MCFs, it is well known that even in such fibers, the ICS is not negligible [
16]. It results from limitations in the fiber fabrication process as well as environmental factors [
6] and may be decomposed into a static component and a random time varying component [
17]. It has been shown that static ICS is generally small enough to be compensated using periodically positioned static optical delay lines, even though such an effort may prove to be unfeasible for complex networks. On the other hand, dynamic ICS requires either potentially costly dynamically adjusted optical delay lines or buffering mechanisms to handle the delays between the signals arriving at the receivers or switches from different spatial channels. Since such mechanisms are generally implemented in packet-based telecommunications systems to handle other phenomena, such as delays in electrical components, the transition to SDM systems may require simply a readjustment to cope with MCF skew. This point was addressed in reference [
7]. There, it was shown that the bandwidth utilization efficiency (BUE) in such networks may be limited by both static and dynamic components of the ICS.
Here, we evaluated both the static and dynamic ICS of the 30 km MCF, as this is the most limiting link. For this purpose, we used an experimental setup similar to that in reference [
17]. We transmitted time-aligned packets on the relevant MCF cores using the pSSC transmitter setup shown in
Figure 4a. The delay between packets was then measured at the output of the 30 km MCF using an oscilloscope connected to the outputs of the SDM demultiplexer.
Figure 5a shows the static ICS of the cores used for the transmission of the pSSC with respect to core 2 which was considered to be the reference. It is shown that the maximum absolute static ICS was below 8.2 ns. However, a guard time mechanism to support transmission of the pSSC would have to support the maximum difference between ICS values, corresponding to 13.8 ns.
Figure 5b shows the dynamic skew for the same combinations of cores estimated over a period of 7 h. This skew was not corrected to remove the impact of the jitter introduced by the measurement apparatus. Nevertheless, it shows that, over the measurement period, the dynamic ICS was within a range of approximately 160 ps which is an order of magnitude below the static ICS. These results are higher than previous measurements, potentially due to the impact of electronic jitter on the measurement system. However, the approximate ratio between the static and dynamic components of ICS was similar to the results reported in reference [
18]. As such, we may assume that the ICS impact was negligible in our system. In addition, we showed in the previous sub-section that the pSSC packets had a 24.8 ns guard time which is sufficient to account for static ICS without further compensation mechanisms.
We note that the use of 48 ns of guard time for each packet severely reduced the BUE of the proposed system. In fact, the maximum effective throughput of the pSSC was little over 600 Gb/s, which is far from the line data-rate of 1 Tb/s. However, the guard times could potentially be reduced in future iterations of our testbed by (i) improving the clock recovery sub-system used by the switch control as well as the electronic packet mapper and demapper, or (ii) by introducing static ICS compensation or MCFs with very low ICS. Such techniques will be required in future networks that accomodate pSSCs.
2.3. Network Performance
Figure 6 shows the Q-factor estimates for each of the spatial channels comprising the cSSC when implementing scenarios A to C. In the case of scenario A, the performance varied between spatial channels due to the different crosstalk levels imposed by the 30 km MCF onto different cores. This variation was less pronounced in scenario B, where the crosstalk levels on all relevant cores of the 10 km MCF were similar, though substantially higher. Hence, the performance was substantially lower with respect to scenario A. In the case of scenario C, the spatial channels dropped in node 2 had a performance similar to scenario A but with an approximately 1 dB penalty, mainly due to the impact of crosstalk introduced by the joint circuit switch in node 1. Finally, the spatial channels produced in node 1 and dropped at node 3 after multicasting at node 2 had the worst performance, dominated by the crosstalk introduced by the 10 km MCF. Nevertheless, in all cases, the performance exceeded the required forward error correction (FEC) threshold, despite the limited operational margin. This margin could be improved by (i) improving the crosstalk characteristics of the MCF used to connect nodes 2 and 3 and (ii) improving the isolation properties of the joint circuit switch.
Figure 7 shows the performance estimates for the spatial channels comprising the pSSC, assuming the three cases comprising scenario D. These cases are the transmission of packets from node 1 to node 2, the transmission of packets from node 1 to node 3 with forwarding at node 2, and the transmission of packets from node 2 to node 3. For each estimate, we used 10
8 packets. When transmitting packets from node 1 to node 2, we did not record frame errors with either 1518 byte or 9600 byte packets; hence, we can assume that the frame error rate (FER) was below 10
−7. In contrast, the link from node 2 to node 3 was affected by the transient response of the non-burst-mode amplifiers in node 2 as well as electrical issues at the EASW switch in node 2. The impact of power bursts in the amplifiers was particularly strong with 9600 byte packets due to the longer length, resulting in the worst performance. For this reason, we observed FERs up to 10
−5 on the link from node 1 to node 3 via node 2, and 10
−6 on the link from node 2 to node 3. Nevertheless, all FER values were well below the required 10
−4 requirement from ITU-T.