Next Article in Journal
Polarization-Mode Transformation of the Light Field during Diffraction on Amplitude Binary Gratings
Previous Article in Journal
SuperNANO: Enabling Nanoscale Laser Anti-Counterfeiting Marking and Precision Cutting with Super-Resolution Imaging
Previous Article in Special Issue
Design and Experimental Demonstration of an Atmospheric Turbulence Simulation System for Free-Space Optical Communication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Mitigating Atmosphere Turbulence Fading by Relay Selections in Free-Space Optical Communication Systems with Multi-Transceivers

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Changchun Vocational Institute of Technology, Information Technology Branch, Changchun 130033, China
*
Author to whom correspondence should be addressed.
Photonics 2024, 11(9), 847; https://doi.org/10.3390/photonics11090847
Submission received: 6 August 2024 / Revised: 21 August 2024 / Accepted: 26 August 2024 / Published: 6 September 2024
(This article belongs to the Special Issue Recent Advances in Optical Turbulence)

Abstract

:
In free-space optical communication (FSOC) systems, atmospheric turbulence can bring about power fluctuations in receiver ends, restricting channel capacity. Relay techniques can divide a long FSOC link into several short links to mitigate the fading events caused by atmospheric turbulence. This paper proposes a Reinforcement Learning-based Relay Selection (RLRS) method based on Deep Q-Network (DQN) in a FSOC system with multiple transceivers, whose aim is to enhance the average channel capacity of the system. Malaga turbulence is studied in this paper. The presence of handover loss is also considered. The relay nodes serve in decode-and-forward (DF). Simulation results demonstrate that the RLRS algorithm outperforms the conventional greedy algorithm, which implies that the RLRS algorithm may be utilized in practical FSOC systems.

1. Introduction

Due to the advantages of high bandwidth, large capacity, and no frequency authorization, free-space optical communication (FSOC) technology is highly favored by various communication transmission systems [1,2,3]. However, in ground transmission scenarios, it suffers from atmospheric turbulence fading. The atmospheric turbulence can introduce severe events including wavefront aberration, power scintillation, and beam wander, which degrade the channel capacity of FSOC systems. To address this limitation, relay technology can be applied in FSOC systems, enhancing system performance by dividing long links into several short links [4,5,6], where short links have gentler conditions. In addition, relay technology can also eliminate the requirement for unobstructed transmission between the transmitting and receiving nodes [7,8,9].
Ref. [10] proposes a relay selection algorithm that minimizes the bit error rate (BER) of the system and analyzes the performance of various strategies under different atmospheric turbulence conditions, pointing errors, relay numbers, and distance configurations. Additionally, Ref. [11] presents a link scheduling algorithm to improve bandwidth utilization and reduce the number of idle FSOC links. Ref. [12] proposes a distributed relay selection algorithm, where each relay node transmits data only if its signal-to-noise ratio (SNR) exceeds a threshold value, thereby minimizing the system’s BER. In Ref. [13], a relay selection scheme is proposed to improve diversity gain when the buffer data of the relay node is limited. Reference [14] indicates that by optimally configuring the number and spatial positioning of relay nodes, the bit error rate (BER) can be significantly reduced, thereby enhancing communication efficiency and reliability in turbid water environments. However, as the number of relays increases, traditional model-based relay selection methods often lead to a significant increase in selection complexity and implementation delay. In recent years, with the advancement of machine learning technologies, more researchers have sought to optimize relay selection through machine learning approaches. For instance, references [15,16] applied deep learning to relay resource allocation in FSO systems, successfully overcoming the reliance on system models inherent in traditional algorithms, significantly reducing the complexity of channel capacity analysis, and maximizing capacity. This demonstrates the substantial potential of deep learning in handling unknown and complex models. Furthermore, references [17,18] proposed supervised learning-based methods for relay selection in complex network environments, offering new solutions suitable for dense networks and latency-sensitive scenarios. However, the performance of supervised learning models is highly dependent on the quality and quantity of training data; if the data quality is poor, the model’s performance may degrade significantly. Reference [19] introduced a deep reinforcement learning (DRL)-based method that does not rely on any prior knowledge of channel statistical information, allowing for optimized power allocation under unknown channel models and fading distributions. Similarly, references [20,21] also employed reinforcement learning, modeling the relay selection process as a Markov Decision Process (MDP), targeting outage probability and mutual information as objective functions, and proposed a relay selection scheme based on deep Q networks (DQNs). Ref. [22] considered the handover loss between different nodes and proposed a DQN-based relay selection algorithm to maximize the average channel capacity.
In summary, reinforcement learning maximizes cumulative rewards by optimizing long-term returns, whereas supervised learning focuses only on the optimal choice in the current time slot, without considering the impact on future slots. In the context of relay selection in this paper, the presence of switching losses means that the choice made in each time slot will affect the channel capacity in subsequent slots. Therefore, we propose a Reinforcement Learning-based Relay Selection (RLRS) algorithm based on a Dueling DQN structure, which aims to mitigate the harm on the channel capacity caused by turbulence (the turbulence is modeled as Malaga turbulence). The handover loss of switching between different relay nodes is considered. Unlike Ref. [21], which models the channel as an MDP process, meaning that the channel information at each moment is related to the previous, this paper assumes that the channels in different time slots are independent, and the relay node selected for each time slot is modeled as an MDP process. Different from Ref. [22], which considers the scenario of a single transmitting node, this paper assumes multiple transmitting nodes, which means that different transmitting nodes need to select different relay nodes, thereby increasing the complexity of the relay selection algorithm. Therefore, the main innovations of this paper are as follows:
  • Based on the Dueling DQN structure, the RLRS algorithm is proposed to maximize the average channel capacity, thus mitigating the degradation caused by atmospheric turbulence;
  • The average channel capacity expression of the decode-and-forward (DF) mode in an FSOC relay system with multiple transceiver nodes is considered and the handover loss is derived;
  • In the implementation of the RLRS algorithm, the actions are encoded in a multi-digit format, and a reward function with a penalty term is designed based on whether the multi-digit actions are repeated.
The structure of this paper is as follows. Section 2 describes the system model and problem presentation. Section 3 details the specific process of the RLRS algorithm. In Section 4, the simulation results of the RLRS algorithm are given and compared with the conventional greedy algorithm. Conclusions are drawn in Section 5.

2. System Model and Problem Formulation

2.1. System Model

The system model of this paper is shown in Figure 1. In a multi-transceiver FSOC system with relays, there are P transmitting terminals (TX-1, …, TX-P), P receiving terminals (RX-1, …, RX-P), and N relay terminals (RLY-1, …, RLY-N). The red line represents the laser link. At any time, all the transmitting terminals will choose a different relay terminal as the relay node, which means that there are P parallel independent links in each time slot. In this paper, it is assumed that the relay terminals employ the DF method, and that all transmitting terminals and relay terminals adopt On–Off Keying (OOK) modulation in the NRZ format, chosen for its simple transceiver structure and balanced 0/1 signal characteristic, which simplify clock data recovery at the receiver. The Gaussian noise power between the transmit-relay and relay-receive terminals is σ r 2 and σ s 2 , respectively.

2.2. Malaga Turbulence and Pointing Errors

Before analyzing the probability density function (PDF) of channel fluctuations, considering pointing errors, it is essential to select an appropriate turbulence model. The log-normal, lognormal-Rician, and gamma-gamma models are among the most prominent [23]. However, the log-normal model is limited to scenarios with weak irradiance fluctuations. Although the lognormal-Rician model aligns well with experimental data, it lacks a closed-form solution for its integral and has inherent convergence issues. Alternatively, the gamma-gamma model is proposed as a practical substitute for the lognormal-Rician model, owing to its more manageable mathematical formulation. It is worth noting that Ref. [24] introduces the Malaga turbulence (M distribution), applicable to both plane and spherical waves across all turbulence conditions, ranging from weak to extremely strong in the saturation regime. This distribution unifies the log-normal and gamma-gamma models into a single closed-form expression, encompassing various models proposed for atmospheric optical communications. Consequently, the gamma-gamma and log-normal distributions are special cases of the Malaga model under different turbulence conditions [25]. The Mdistribution PDF of irradiance is expressed as follows:
f h h = A k = 1 β a k h α + k 2 1 K α k 2 α β h γ β + Ω
A 2 α α / 2 γ 1 + α / 2 Γ ( α ) γ β γ β + Ω β + α / 2 a k β 1 k 1 ( γ β + Ω ) 1 k / 2 ( k 1 ) ! Ω γ k 1 α β k / 2
where α , β , γ , Ω are the Malaga channel parameters. In Equation (1), Kν(·) is the modified Bessel function of the second kind and order ν.
Let h k i , j represent the channel gain between the i-th ( 1 i P ) transmitting terminal and the j-th ( 1 j N ) relay terminal in the k-th time slot. Similarly, let g k j , i represent the channel gain between the j-th ( 1 j N ) relay terminal and the i-th ( 1 i P ) receiving terminal in the k-th time slot. In this paper, atmospheric attenuation, turbulence fluctuation, and pointing errors are considered. Then, using the Malaga model, the PDF of the channel fluctuation with pointing errors can be expressed as follows:
f h ( h ) = ξ 2 A 2 h k = 1 β a k α β γ β + Ω α + k 2 G 1 , 3 3 , 0 α β γ β + Ω h h k i , j g k j , i h l A 0     ξ 2 + 1 ξ 2 , α , k
where G 1 , 3 3 , 0 | represents the Meijer’G function, h l is the path loss which obtains turbulence attenuation, geometric loss, and both transmission and reception losses. ξ 2 and A 0 represent the pointing error parameter.

2.3. Effect of Handoff Loss

Similarly to Ref. [21], τ is defined as the handoff loss during the handover of adjacent nodes. However, Ref. [21] only considers the case of a single transmitting terminal, and all relay nodes always point to the transmitting terminal and the receiving terminal, thus eliminating the problem of relay node-handover direction. In this paper, the issue of relay node-pointing switches needs to be additionally considered. As mentioned above, in the kth time slot, it is assumed that the relay node R R i k is selected for the i-th transmitting terminal to the i-th receiving terminal. Similarly, in the (k + 1)-th time slot, the relay node RX-i is selected for the i-th transmitting terminal to the i-th receiving terminal. Then, in the (k + 1)-th time slot, the handover loss τ i k + 1 from TX-i to R R i k + 1 and then to RX-i link should be calculated as follows:
τ i k + 1 = max R R i k + 1 R R i k τ , R R i k + 1 R R i ' k + 1 τ
where R R i ' k + 1 represents the direction of the R R i k + 1 relay node before the (k + 1)-th time. Therefore, considering the handover loss, the channel capacity in the k-th time slot should be expressed as follows:
C k = i = 1 P C k i = i = 1 P 1 τ i k C k i
where C k i is the channel capacity of the link from TX-i to RX-i through the relay node in the k-th time slot under the premise of considering the handover loss. C k i is the channel capacity of the link from TX-i to RX-i through the relay node in the k-th time slot without considering the handover loss. The relationship between the two is C k i = 1 τ i k C k i . Consider that in the k-th time slot, if the relay node R R i k is selected for the link from the i-th transmitting terminal to the i-th receiving terminal, the expression of C k i is
C k i = min log 2 1 + η P h k i , R R i k 2 / σ r 2 , log 2 1 + η P g k i , R R i k 2 / σ d 2
The objective of this study is to maximize the average channel capacity in M time slots, namely
max Ω M 1 M k = 0 M 1 C k         s . t .1 R R i k i = 1 , P k = 0 , , M 1 N      
where Ω M represents the set of relay nodes selected in M time slots, that is, Ω M = R R i k i = 1 , P ; k = 0 , , M 1 . From Equation (7), the system needs to carefully select relay nodes in each time slot because each selection will affect the handover loss in subsequent time slots, further impacting the final channel capacity. In the actual systems, the study cannot predict the channel state of subsequent time slots, but fortunately, the study can leverage reinforcement learning techniques to optimize the relay selection in each time slot by maximizing the characteristics of long-term cumulative reward functions.
As mentioned above, the primary objective of this paper is to maximize the average channel capacity of M time slots. Therefore, the cumulative channel capacity with a discount factor starting from time k needs to be defined:
G k = t = k γ t k C k
where γ represents the discount factor, which ranges between 0 and 1. Therefore, the problem of maximizing the average channel capacity in M time slots can be transformed into the form of Equation (9)
max Ω M E G k         s . t .1 R i k i = 1 , P k = 0 , , M 1 N      

3. The RLRS Algorithm

In reinforcement learning, the problem to be solved is typically modeled as an MDP. The MDP generally consists of four elements S , A , R , P : the state space S , the action space A , the reward function at the current time R , and the transition probability P . The next state is only determined by the current state and the current action and is independent of previous states. In the relay selection model of this paper, since the transition probability is unknown, it can be modeled as an incomplete MDP, consisting of three elements S , A , R . Correspondingly, S k , A k , R k are defined as the state, action, and immediate reward functions at the k-th time slot, respectively. In the model of this paper, the specific contents of S k , A k , R k are as follows:
State S k : In any k-th time slot, state S k includes two parts: the pointing I k (including 2P + N elements) of all nodes in the (k-1)th time slot and the channel gain of the S-R link and R-D link in all current time slots (including 2PN elements).
Action A k : In any k-th time slot, action A k represents the sequence number (including P elements) of the relay nodes selected for all transmitting nodes. There are N ! / N P ! possibilities in the action space. With N ! / N P ! N P , we can write any action A k into the form of a P-bit N-ary number, that is, A k = A k 1 , A k 2 , , A k P , A k i 0 , N 1 .
Immediate reward function R k : When action A k is performed under state S k , an immediate reward function R k will be obtained. This reward function is utilized to indicate whether action A k performed in the current state is beneficial.
Since action A k is written in the form of a P-bit N-ary number, the action space is expanded from N ! / N P ! dimension to N P dimension. In this case, there will be situations where multiple transmitting nodes select the same relay. To avoid this, we define R k as follows:
R k = 1 , i f   t h e r e   i s   t h e   s a m e   e l e m e n t   i n   A k C k ,   e l s e
In reinforcement learning, the agent maximizes the cumulative reward function with a discount factor by learning the optimal strategy. In order to evaluate the performance of the strategy, the Q-value is typically introduced as a metric, defined as follows: Q ( s , a ) = E G t s = S t , a = A t .When the state and action spaces are small enough, the common reinforcement learning approach is the Q-table method. By listing all actions for each state, the Q-value table of each action-–state pair is obtained, and the action with the largest Q-value is selected as the optimal strategy. However, in this paper, the state space is continuous, and the action space is also very large, making it impossible to list all Q-tables. Therefore, a DQN is adopted. In a DQN, a neural network is used to simulate the action value function, allowing it to handle each action value function in high-dimensional state space and output the maximum value function.
Inspired by Ref. [16], this paper also adopts a Dueling-DQN network structure and proposes an RLRS algorithm. However, the action space, state space, and reward function in this paper differ from those in Ref. [16]. The overall structure of this paper’s RLRS algorithm is shown in Figure 2. The RLRS algorithm includes two neural networks—one is an online neural network, the other a target neural network, both with the same network structure. The input layer, hidden layer, and output layer structures of the RLRS algorithm differ slightly from those of conventional DQNs. In the RLRS structure, the layer before the output layer is divided into two parts: the state value part V s and the action advantage part A s , a . They are expressed in Equation (11).
V ( s ) = V ( s ω , ϑ ) ,   A ( s , a ) = A ( s , a ω , ζ )
where ω represents the weight threshold of the other layers except the last layer in the online network. ϑ and ζ , respectively, represent the weight threshold of the state-value part and the action-advantage-value part in the online network. Since the target network has the same structure as the online network, the corresponding parameters ω , ζ , and ϑ in the target network can be defined. Similarly, the parameters of the online network can be completely determined by ω , ζ , and ϑ .
The input layer of the RLRS algorithm has 2PN + 2P + N neurons, corresponding to the dimensions of the state space, and the output layer has N elements, corresponding to the dimensions of the expanded action space. As mentioned earlier, the output of the online network represents the Q-value of the optimal state–action pair in the current strategy, which can be expressed as follows:
Q ( s , a ) = Q ( s , a ω , ζ , ϑ ) = V ( s ω , ϑ ) + A ( s , a ω , ζ ) 1 | A | a A s , a ω , ζ
where | | represents the number of elements in . Similarly, Q ( s , a ω , ζ , ϑ ) represents the output of the target network when the input is s and the selection action is a . In any k-th time slot, in the process of selecting an action A k , a ε - greedy approach is adopted, that is, the network will arbitrarily select an action in the state space with the probability of ε . Otherwise, it will select the action with the maximum Q-value, as shown in Equation (13).
A k = argmax A k Q ( s , a ω , ζ , ϑ ) s = S t , a = A k , 1 ε   p r o b a b i l i t y A k S , ε   p r o b a b i l i t y  
In the k-th time slot, based on the state S k , after selecting A k , the immediate reward R k of k-th time slot will be obtained, and then it enters the next state S k + 1 . At this point, it yields a set of empirical values S k , A k , R k , S k + 1 , which can be put into experience pool D . The experience pool has a first-in, first-out structure and can store D elements. If the experience pool D is not full, the experience value S k , A k , R k , S k + 1 will be directly stored in the experience pool; if the experience pool is already full, the earliest stored experience will be removed, and the new experience value will be stored in the experience pool D .
L ( ω , ζ , ϑ ) 1 X l r l + γ max A l + 1 Q S l + 1 , A l + 1 ω , ζ , ϑ Q S l , A l ω , ζ , ϑ 2
The online network is updated first,
ω ω ς ω L ζ ζ ς ζ L ϑ ϑ ς ϑ L
where ω , ζ , and ϑ represent the gradient values of the loss function L ( ω , ζ , ϑ ) with respect to ω ζ , and ϑ , respectively. ς represents the learning rate of the online network. After updating the online network, the target network is subsequently updated. Assuming the learning rate of the target network is λ , the network parameters are updated as follows:
ω λ ω + ( 1 λ ) ω ζ λ ζ + ( 1 λ ) ζ ϑ λ ϑ + ( 1 λ ) ϑ
The pseudocode diagram of the proposed RLRS algorithm is depicted in Algorithm 1.
Algorithm 1. The pseudocode diagram of the proposed RLRS algorithm.
Input: The FSOC system simulator and its parameters.
Output: Optimal action A k of each time slot.
1:Initialize experience replay memory D .
2:Initialize ω , ζ and ϑ with random weights and initialize ω , ζ and ϑ .
3:Initialize the minibatch size with X.
4:FOR episode in 1 , 2 , 3 ,  DO
5:     Observe the environment initial state S 0 .
6:     FOR  k = 1 to K DO
7:        Select a relay selection action A k by Equation (11), and execute action A k .
8:        IF there is same element in A k
9:          Calculate immediate reward R k = 1 .
10:         ELSE
11:          Calculate immediate reward R k = C k .
12:        Obtain next state S k + 1 and store transition data S k , A k , R k , S k + 1 in replay memory D .
13:        IF  D is full
14:        Sample a random minibatch of X sets of transition data from D .
15:        Update the online network by (13).
16:        Update the target network by (14).
17:      END FOR
18:END FOR

4. Simulation Results

In this section, the performance simulation of the RLRS algorithm is presented, and the simulation parameters are shown in Table 1. As a comparison, a greedy algorithm is utilized as a reference to demonstrate the advantages and effectiveness of the proposed RLRS algorithm.
According to the above parameters and pseudo-code, the following simulation is obtained. Figure 2a,b shows the cumulative reward curve and the loss function curve for a time slot number of 10. Figure 2a shows the training results across different episodes. The training process can be roughly divided into three parts: the experience pool storage part, the parameter update part, and the training completion part. Before the experience pool is full, the cumulative reward curve fluctuates without a clear upward trend. Once the experience pool is full, the system extracts experiences from the pool and utilizes the mini-batch method for parameter updates. At this point, the cumulative reward curve shows an overall upward trend. After the 400th episode, the cumulative reward value tends to stabilize, and the stable cumulative reward value is superior to the performance of the greedy algorithm. The loss function curve in Figure 2b shows a periodic trend of decline, indicating the convergence of the algorithm.
Figure 3a,b shows the cumulative reward curve and the loss function curve for a time slot number of 50. Overall, the curve trends in Figure 3 are almost identical to those in Figure 2. This also demonstrates that the proposed RLRS algorithm in this paper can converge under different time slot conditions and outperforms the greedy algorithm.
Figure 4 presents the average channel capacity curves under different handover loss conditions. Figure 4a,b describes the cases of 10 time slots and 50 time slots. As can be seen from Figure 4a,b, the proposed RLRS algorithm in this paper consistently outperforms the greedy algorithm under various handover loss conditions. Additionally, as the handover loss increases, the gap between the two algorithms gradually widens, with the RLRS algorithm providing at least a 2.47% improvement in channel capacity. This is because the conventional greedy algorithm only selects the relay node with the largest channel gain without considering handover loss. As a result, when the handover loss increases, its capacity decreases linearly. In contrast, the proposed RLRS algorithm in this paper takes both handover loss and channel gain into account, so as to maximize the long-term rewards of relay selection in each time slot.
Figure 5 shows the average capacity of the RLRS algorithm and the traditional greedy algorithm when fog’s influence is considered. According to the Kruise formula, the attenuation due to fog (dB/km) is defined as follows:
α atm ( V ) = 3.912 V λ 550 ( nm ) q
where λ is the wavelength (in nm), V represents the visibility (in km). q is a parameter that depends on the particle size distribution of scattering particles, given as follows:
q = 1.6 V 50 km 1.3 6 km < V < 50 km 0.585 V 1 / 3 V < 6 km ;
This paper selected the wavelength as 1550 nm. The visibility is set to 1 km, 2 km, 5 km, 10 km, 50 km, respectively. As shown in Figure 5a, which illustrates Malaga turbulence, the average capacity increases with visibility. This is due to the fact that greater visibility results in reduced attenuation. Figure 5b illustrates Gamma-Gamma turbulence with α = 8.43 , β = 6.92 , where the Malaga turbulence can reduce to Gamma-Gamma turbulence with ρ 1 and Ω ' = 1 . As can be seen from Figure 5a,b, the RLRS algorithm always performs better than the traditional greedy algorithm. Besides the situation of weak turbulence, which has a larger average capacity in both the RLRS and greedy cases, Figure 5a,b also show that the RLRS algorithm performs more effectively under strong turbulence. This is because strong turbulence can cause significant variations in channel gains across multiple potential links. Consequently, during relay selection, the difference in system performance between high-quality and low-quality relays becomes more pronounced, leading to a more noticeable improvement in system performance with the RLRS algorithm.

5. Conclusions

This paper proposes a relay selection algorithm to mitigate the capacity degradation caused by atmospheric turbulence, where there are N transmitting nodes (and N receiving nodes) and P relay nodes. We also consider the handover loss caused by the handover process of different relay nodes. The RLRS algorithm is proposed based on a Dueling DQN structure to maximize the average channel capacity. In the implementation of the RLRS algorithm, the actions are encoded in a multi-digit format, and a reward function with a penalty term is designed based on whether the multi-digit actions are repeated. Finally, through simulation comparison, the RLRS algorithm proposed in this paper is shown to be superior to the greedy comparison algorithm, which can enhance the channel capacity by at least 2.47%, with the gap increasing as handover loss rises. Moreover, under stronger turbulence, the RLRS algorithm demonstrates a noticeable improvement in channel capacity, further validating its effectiveness in combating the capacity loss induced by turbulence. Therefore, this study not only provides a reference design scheme for the relay FSOC system to further fight against turbulence, but also demonstrates the potential of reinforcement learning in improving channel capacity, which offers new insights and directions for future high-performance communication system design.

Author Contributions

Conceptualization, X.S. and Z.L.; Methodology, X.S.; Software, Z.L.; Formal analysis, X.S.; Writing—original draft, X.S.; Writing—review and editing, Z.L.; Supervision, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Al-Kinani, A.; Wang, C.X.; Zhou, L.; Zhang, W. Optical Wireless Communication Channel Measurements and Models. IEEE Commun. Surv. Tutor. 2018, 20, 1939–1962. [Google Scholar] [CrossRef]
  2. Wang, T.; Lin, P.; Dong, F. Progress and Prospect of Space Laser Communication Technology. Strateg. Study CAE 2020, 22, 92–99. (In Chinese) [Google Scholar] [CrossRef]
  3. Khalighi, M.A.; Uysal, M. Survey on free space optical communication: A communication theory perspective. IEEE Commun. Surv. Tutor. 2014, 16, 2231–2258. [Google Scholar] [CrossRef]
  4. Safari, M.; Uysal, M. Relay-assisted free-space optical communication. IEEE Trans. Wireless Commun. 2008, 7, 5441–5449. [Google Scholar] [CrossRef]
  5. Chatzidiamantis, N.; Michalopoulos, D.; Kriezis, E.; Karagiannidis, G.; Schober, R. Relay Selection Protocols for relay-assisted Free Space Optical systems. Opt. Commun. Netw. 2013, 5, 92–103. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Zhang, P.; Yu, H. Research on polar laser communication based on relay system. Opt. Commun. Technol. 2021, 45, 56–58. (In Chinese) [Google Scholar]
  7. Boluda-Ruiz, R.; Garcia-Zambrana, A.; Castillo-Vazquez, B. Ergodic capacity analysis of decode-and-forward relay-assisted FSO systems over alpha–Mu fading channels considering pointing errors. IEEE Photonics J. 2016, 8, 1–11. [Google Scholar] [CrossRef]
  8. Abou-Rjeily, C.; Noun, Z.; Castillo-Vazquez, B. Impact of inter-relay cooperation on the performance of FSO systems with any number of relays. IEEE Trans. Wirel. Commun. 2016, 15, 3796–3809. [Google Scholar] [CrossRef]
  9. Zhang, R.; Zhang, W.; Zhang, X. Research status and development trend of high earth orbit satellite laser relay links. Laser Optoelectron. Prog. 2021, 58, 9–21. (In Chinese) [Google Scholar]
  10. Taher, M.; Abaza, M.; Fedawy, M. Relay selection schemes for FSO communications over turbulent channels. Appl. Sci. 2019, 9, 1281. [Google Scholar] [CrossRef]
  11. Tan, Y.; Liu, Y.; Guo, L.; Han, P. Joint relay selection and link scheduling in cooperative free-space optical system. Opt. Eng. 2016, 55, 111604. [Google Scholar] [CrossRef]
  12. Ben, H. Round, Robin, Centralized and distributed relay selection for free space optical communications. Wirel. Pers. Commun. 2019, 108, 51–66. [Google Scholar] [CrossRef]
  13. Abou-Rjeily, C. Improved Buffer-Aided Selective relaying for free space optical cooperative communications. IEEE Trans. Wirel. Commun. 2022, 21, 6877–6889. [Google Scholar] [CrossRef]
  14. Palitharathna, S.; Godaliyadda, I.; Herath, R. Relay-assisted optical wireless communications in turbid water. In Proceedings of the International Conference on Under Water Networks and Systems, Shenhen, China, 3–5 December 2018; pp. 1–5. [Google Scholar] [CrossRef]
  15. Gao, Z.; Eisen, M.; Ribeiro, A. Resource Allocation via Model-Free Deep Learning in Free Space Optical Communications. IEEE Trans. Commun. 2021, 70, 920–934. [Google Scholar] [CrossRef]
  16. Le-Tran, M.; Vu, H.; Kim, S. Performance Analysis of Optical Backhauled Cooperative NOMA Visible Light Communication. IEEE Trans. Veh. Technol. 2021, 70, 12932–12945. [Google Scholar] [CrossRef]
  17. Dang, H.; Liang, Y.V.; Wei, L. Enabling Relay Selection in Cooperative Networks by Supervised Machine Learning. In Proceedings of the Eighth International Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC), Harbin, China, 19–21 July 2018; pp. 1459–1463. [Google Scholar] [CrossRef]
  18. Dang, S.; Tang, J.; Li, J.; Wen, M.; Abdullah, S. Combined relay selection enabled by supervised machine learning. IEEE Trans. Veh. Technol. 2021, 70, 3938–3943. [Google Scholar] [CrossRef]
  19. Kapsis, T.; Lyras, K.; Panagopoulos, D. Optimal Power Allocation in Optical GEO Satellite Downlinks Using Model-Free Deep Learning Algorithms. Electronics 2024, 13, 647. [Google Scholar] [CrossRef]
  20. Su, Y.; Li, W.; Gui, M. Optimal Cooperative Relaying and Power Control for IoUT Networks with Reinforcement Learning. IEEE Internet Things J. 2021, 8, 791–801. [Google Scholar] [CrossRef]
  21. Su, Y.; Lu, X.; Zhao, Y.; Huang, L.; Du, X. Cooperative communications with relay selection based on deep reinforcement learning in wireless sensor networks. IEEE Sens. J. 2019, 19, 9561–9569. [Google Scholar] [CrossRef]
  22. Gao, S.; Li, Y.; Geng, T. Deep reinforcement learning-based relay selection algorithm in free-space optical cooperative communications. Appl. Sci. 2022, 12, 4881. [Google Scholar] [CrossRef]
  23. Jurado-Navas, A.; Garrido-Balsells, J.; Paris, J.; Castillo-Vázquez, M. Impact of pointing errors on the performance of generalized atmospheric optical channels. Opt. Express. 2012, 20, 12550–12562. [Google Scholar] [CrossRef] [PubMed]
  24. Jurado-Navas, A.; Garrido-Balsells, J.; Paris, J.; Puerta-Notario, A. A unifying statistical model for atmospheric optical scintillation. In Numerical Simulations of Physical and Engineering Processes; Intech: Rijeka, Croatia, 2011; Volume 181, pp. 181–205. [Google Scholar] [CrossRef]
  25. Ansari, I.S.; Yilmaz, F.; Alouini, M.S. Performance analysis of free-space optical links over Málaga turbulence channels with pointing errors. IEEE Trans. Wirel. Commun. 2015, 15, 91–102. [Google Scholar] [CrossRef]
Figure 1. Diagram of a multi-transceiver FSOC system with relays.
Figure 1. Diagram of a multi-transceiver FSOC system with relays.
Photonics 11 00847 g001
Figure 2. Performance of an RLRS algorithm in 10 time slots. (a) Curves of cumulative reward. (b) Curve of loss function.
Figure 2. Performance of an RLRS algorithm in 10 time slots. (a) Curves of cumulative reward. (b) Curve of loss function.
Photonics 11 00847 g002
Figure 3. Performance of RLRS algorithm in 50 time slots. (a) Curves of cumulative reward. (b) Curve of loss function.
Figure 3. Performance of RLRS algorithm in 50 time slots. (a) Curves of cumulative reward. (b) Curve of loss function.
Photonics 11 00847 g003
Figure 4. Average channel capacity versus different handover loss. (a) 10 time slots. (b) 50 time slots.
Figure 4. Average channel capacity versus different handover loss. (a) 10 time slots. (b) 50 time slots.
Photonics 11 00847 g004
Figure 5. Average capacity of RLRS and greedy algorithm under Malaga and Gamma-Gamma turbulence with fog. (a) Average capacity under Malaga turbulence with fog. (b) Average capacity under Gamma-Gamma turbulence with fog.
Figure 5. Average capacity of RLRS and greedy algorithm under Malaga and Gamma-Gamma turbulence with fog. (a) Average capacity under Malaga turbulence with fog. (b) Average capacity under Gamma-Gamma turbulence with fog.
Photonics 11 00847 g005
Table 1. Simulation parameters.
Table 1. Simulation parameters.
Parameter NameValue
Number of relay nodes N 4
Number of transmitting nodes P 2
Number of time slot M 5, 10, 50
Unit handover loss τ 0.05–0.25
Responsiveness of detector η 0.9
Channel parameters α , β , ρ , Ω , A 0 , ρ 5.97, 4.39, 0.596, 1, 0.0032, 6.25
Normalized power P t / σ r , P t / σ s 10
Discount factor γ 0.9
Online learning rate ς 0.001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

San, X.; Liu, Z.; Wang, Y. Research on Mitigating Atmosphere Turbulence Fading by Relay Selections in Free-Space Optical Communication Systems with Multi-Transceivers. Photonics 2024, 11, 847. https://doi.org/10.3390/photonics11090847

AMA Style

San X, Liu Z, Wang Y. Research on Mitigating Atmosphere Turbulence Fading by Relay Selections in Free-Space Optical Communication Systems with Multi-Transceivers. Photonics. 2024; 11(9):847. https://doi.org/10.3390/photonics11090847

Chicago/Turabian Style

San, Xiaogang, Zuoyu Liu, and Ying Wang. 2024. "Research on Mitigating Atmosphere Turbulence Fading by Relay Selections in Free-Space Optical Communication Systems with Multi-Transceivers" Photonics 11, no. 9: 847. https://doi.org/10.3390/photonics11090847

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop