Skip to Content
SensorsSensors
  • Article
  • Open Access

9 April 2024

Deep-Reinforcement-Learning-Based Joint Energy Replenishment and Data Collection Scheme for WRSN

,
,
and
1
Yunnan Key Laboratory of Computer Technology Applications, Kunming University of Science and Technology, Kunming 650500, China
2
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.

Abstract

With the emergence of wireless rechargeable sensor networks (WRSNs), the possibility of wirelessly recharging nodes using mobile charging vehicles (MCVs) has become a reality. However, existing approaches overlook the effective integration of node energy replenishment and mobile data collection processes. In this paper, we propose a joint energy replenishment and data collection scheme (D-JERDG) for WRSNs based on deep reinforcement learning. By capitalizing on the high mobility of unmanned aerial vehicles (UAVs), D-JERDG enables continuous visits to the cluster head nodes in each cluster, facilitating data collection and range-based charging. First, D-JERDG utilizes the K-means algorithm to partition the network into multiple clusters, and a cluster head selection algorithm is proposed based on an improved dynamic routing protocol, which elects cluster head nodes based on the remaining energy and geographical location of the cluster member nodes. Afterward, the simulated annealing (SA) algorithm determines the shortest flight path. Subsequently, the DRL model multiobjective deep deterministic policy gradient (MODDPG) is employed to control and optimize the UAV instantaneous heading and speed, effectively planning UAV hover points. By redesigning the reward function, joint optimization of multiple objectives such as node death rate, UAV throughput, and average flight energy consumption is achieved. Extensive simulation results show that the proposed D-JERDG achieves joint optimization of multiple objectives and exhibits significant advantages over the baseline in terms of throughput, time utilization, and charging cost, among other indicators.

1. Introduction

Benefiting from the advancements in wireless power transfer (WPT) technology [1,2], wireless rechargeable sensor networks (WRSNs) equipped with wireless charging systems have emerged as an effective solution for addressing node energy constraints [3,4]. Unlike traditional approaches [5,6,7], WRSN fundamentally overcomes the predicament of nodes relying solely on battery power. It provides a promising solution for the sustainable energy replenishment of nodes [8,9,10]. WRSN normally includes one or more mobile charging vehicles (MCVs) and a base station (BS) for the battery replacement of MCVs. MCV can move autonomously and is equipped with a wireless charging device to replenish energy for nodes by wireless charging. However, in remote and harsh environments, e.g., farmlands, forests, or disaster areas, it becomes challenging for ground vehicles to enter designated areas and perform close-range wireless charging [11,12,13]. Utilizing UAVs as aerial mobile vehicles to provide remote services to ground devices is considered to bring significant benefits to WRSN [14]. UAV possesses excellent maneuverability, coverage capabilities, and relatively low operating costs. It can perform various tasks within the network coverage, such as routing, communication, data collection, and energy supply [15,16,17,18]. In UAV-assisted wireless rechargeable sensor networks (UAV-WRSNs), UAVs can serve both as energy transmitters and data transceivers. By continuously accessing and exchanging data with nodes within their communication range, they can forward the data to the BS. This eliminates the energy consumption caused by multihop transmissions between nodes. Additionally, UAVs provide range-based charging to ensure simultaneous data collection and energy replenishment.
Although UAV offers new possibilities for the future development of WRSN, more research is needed to consider the unified process of node energy replenishment and data collection in UAV-WRSN. There are still unresolved issues in UAV-WRSN. For example, in a one-on-one node service scenario, UAV must hover multiple times, increasing hover time and energy consumption, leading to subsequent node data overflow or energy depletion. Moreover, target node selection is crucial to enable UAV to cover more nodes with each hover. In practical environments, there is a significant difference in energy consumption among nodes, and some nodes may run out of energy earlier due to heavy loads. Therefore, balancing node energy consumption and solving the problem of energy imbalance among nodes are equally important. Based on the above issues, this paper proposes D-JERDG, which integrates node joint energy replenishment and data collection using deep reinforcement learning.
The main contributions of this study can be summarized as follows:
  • First, we consider deploying UAV in delay-tolerant WRSN and combining wireless power transfer (WPT) with wireless information transfer (WIT) technologies. To achieve the unified process of sensor node energy replenishment and mobile data collection, we propose a deep-reinforcement-learning-based method called D-JERDG for UAV-WRSN.
  • We introduce a cluster head selection algorithm based on an improved dynamic routing protocol to minimize the number of UAVs hovering and balance node energy consumption. Based on the obtained cluster head visiting sequence, we employ the simulated annealing algorithm [19] to approximate the optimal solution for the traveling salesman problem (TSP), thereby reducing the UAV flight distance. We then employ the DRL model MODDPG [20] and design a multiobjective optimized reward function to control the UAV instantaneous speed and heading, aiming to minimize the node death rate and the UAV’s average energy consumption.
  • Simulation results are conducted to evaluate the feasibility and effectiveness of D-JERDG. The results demonstrate that D-JERDG outperforms existing algorithms in node death rate, time utilization, and charging cost.
The remaining parts of this paper are as follows: Section 2 introduces related work, Section 3 describes the system model and problem formulation of UAV-WRSN, Section 4 presents the cluster head selection algorithm based on an improved dynamic routing protocol and the MODDPG algorithm in D-JERDG, Section 5 validates the effectiveness and feasibility of D-JERDG through comparative experiments, and Section 6 summarizes the paper and discusses future research directions.

3. System Model and Problem Formulation

In this section, we present the system model and problem formulation of UAV-assisted wireless rechargeable sensor networks (UAV-WRSNs). For the sake of clarity, Table 2 lists the symbols used in this paper.
Table 2. Notation and definition.

3.1. System Model

3.1.1. Network Model

As shown in Figure 1, the network is deployed in a two-dimensional area and consists of a UAV, a base station (BS), and randomly distributed sensor nodes. The UAV is assumed to have computing and wireless transmission capabilities, be equipped with a sufficiently charged battery, and be able to obtain its position through the Global Positioning System (GPS). The BS has ample energy and communication capabilities, enabling direct wireless communication with the UAV. The sensor nodes can monitor their remaining energy and data fusion capabilities. Cluster member (CM) nodes send data packets to cluster head (CH) nodes, which aggregate the data and store them in a data buffer. The UAV takes off from the BS and follows a “fly-hover” pattern to visit each CH node in a predetermined order continuously. The CH nodes consume energy to upload data packets to the UAV, while the UAV provides charging to the nodes within its coverage range using radio frequency transmission technology.
Figure 1. Network model of the UAV-WRSN.

3.1.2. Sensor Node Model

The network deploys k sensor nodes with fixed and known positions. Consider the set of nodes in the network as S x { s 1 , s 2 , s 3 , , s k } . The CH nodes are denoted as c h i { c h 1 , c h 2 , c h 3 , , c h m } , and the CM nodes are denoted as c m j { c m 1 , c m 2 , c m 3 , , c m h } . Assuming that node batteries can be charged within a very short time compared with the data collection time by the UAV, we can neglect the charging time. Regarding the energy consumption of nodes in sleep mode, since this portion of energy consumption is very small and can be considered negligible, we have not included it as a significant influencing factor, and therefore, it has not been incorporated into the node energy consumption. Nodes only consume energy when receiving and transmitting data. The data buffer length of the CH node c h i at t is denoted as D c h i ( t ) = [ 0 , Q m a x ] . Based on the literature [43], assuming that, at any time t, the energy consumption of the CM node c m j sending a unit data packet (where the unit data packet size f is 1024 kb) to the CH node c h i can be given as follows:
E j c m ( t ) = e t j = 1 h f j , i
where e t indicates the energy consumption of transmitting unit data, the energy consumption of the CH node c h i can be divided into two parts: reception energy and transmission energy:
E i c h ( t ) = j = 1 h f j , i i = 1 m e r + E c h i t ( t )
where e r indicates the energy consumption of receiving unit data, and the remaining energy of the CM node c m j and the CH node c h i at t can be respectively defined as follows:
R E j c m ( t ) = E j c m ( t 1 ) E j c m ( t )
R E i c h ( t ) = E i c h ( t 1 ) E i c h ( t )

3.1.3. UAV Model

The UAV maintains a fixed flight altitude of H. It acquires the positions of all nodes before taking off from the base station and maintains communication with the base station throughout the operation. At t, the three-dimensional coordinates of the UAV can be represented as [ x u ( t ) , y u ( t ) , H ] . The distance between the UAV and the node plays a crucial role in determining the link quality and the ability of the node to upload data. In our scenario, we assume that the UAV has limited communication and charging ranges, denoted as R d (maximum data transmission radius) and R c (maximum charging radius), respectively. Δ d u , c h i t a r represents the distance between the UAV and the target CH node if Δ d u , c h i t a r R d ; the UAV enters the hovering state to collect data from the target CH node and charge all nodes within the charging range. The instantaneous speed v u ( t ) and the instantaneous heading angle θ u ( t ) describe the flight control of the UAV. For safety reasons, v u ( t ) must be within the minimum and maximum speed limits.
V min v u ( t ) V max
This paper employs a rotor-wing UAV, and the energy consumption of the UAV can be divided into propulsion energy consumption and communication energy consumption. The propulsion energy consumption is further divided into flight energy consumption and hovering energy consumption [20]. When the UAV has an instantaneous speed v u ( t ) , the propulsion energy consumption model of the UAV can be given as follows:
E u p r o ( v u ( t ) ) = P 0 ( 1 + 3 v u ( t ) 2 U tip 2 ) + P f ( 1 + v u ( t ) 4 4 v 0 4 v u ( t ) 2 2 v 0 2 ) 1 2 + 1 2 d 0 ρ air s rotor A rotor v u ( t ) 3
where P 0 and P f are constants that represent the blade profile power and induced power in hovering status, respectively; U t i p indicates the tip speed of the rotor blade; v 0 denotes the mean rotor-induced speed; and d 0 , ρ a i r , s r o t o r , and A r o t o r , respectively, denote the fuselage drag ratio, air density, rotor solidity, and rotor disc area. At t, UAV speed is v u ( t ) , and the flight energy consumption is expressed as E u f l y = E u p r o ( v u ( t ) ) . It is worth noting that the UAV is in hovering status with flight speed v u ( t ) = 0 and hovering energy consumption E h o v e r = E u p r o ( v u ( t ) = 0 ) . To facilitate the expression, we divide the time domain T into n time steps, where each time step is denoted as t 0 , t 1 , t 2 , t n ; thus, the total flight energy consumption of the UAV at each time step t n is given as follows:
E u t o t a l ( t n ) = 0 t n E u p r o ( v u ( t ) ) d t
the UAV’s average flight energy consumption is given as follows:
E u a v e = E u t o t a l ( T ) T

3.1.4. Transmission Model

The wireless communication link between the UAV and the nodes mainly consists of a line-of-sight (LoS) link and a non-line-of-sight (NLoS) link, similar to many existing works [20,39,40]. Let the coordinate of the target CH node c h i t a r be [ x c h i t a r , y c h i t a r , 0 ] ; the free path loss of the LoS link between the UAV and c h i t a r at t is modeled as follows:
h u , c h i t a r ( t ) = γ 0 Δ d u , c h i t a r ( t ) 2 = γ 0 H 2 + ( x u ( t ) x c h i t a r ) 2 + ( y u ( t ) y c h i t a r ) 2
where the channel’s power gain at a reference distance of Δ d u , c h i t a r = 1 m is denoted by γ 0 , and Δ d u , c h i t a r ( t ) denotes the Euclidean distance between the UAV and c h i t a r at t, where the path loss of the NLoS link is given as η N L o s γ 0 Δ d u , c h i t a r ( t ) 2 , and η N L o s is the attenuation coefficient of the NLoS links.
At t, the LoS link probability between the UAV and c h i t a r is modeled as follows:
P u , c h i t a r L o S ( θ c h i t a r ( t ) ) = 1 1 + α exp ( β ( θ u , c h i t a r ( t ) α ) )
where α and β are constant sigmoid parameters that depend on the signal propagation environment and propagation distance. θ c h i t a r ( t ) is the elevation angle of the UAV and c h i t a r in degree; it is given as θ c h i t a r ( t ) = 180 π sin 1 ( H Δ d u , c h i t a r ) . The NLoS link probability is given as follows:
P u , c h i t a r N L o S ( θ c h i t a r ( t ) ) = 1 P c h i t a r L o S ( θ c h i t a r ( t ) )
let g u , c h i t a r ( t ) be the wireless channel gain between the UAV and c h i t a r ; it is modeled as follows:
g u , c h i t a r ( t ) = ( P u , c h i t a r L o S ( θ u , c h i t a r ( t ) ) + η N L o s P u , c h i t a r N L o S ( θ u , c h i t a r ( t ) ) ) h u , c h i t a r ( t )
in this paper, we assume that the uplink and downlink channels are approximately equal. As a result, the channel power gain between the UAV and c h i t a r is given as follows:
h u , c h i t a r ( t ) g u , c h i t a r ( t ) = ( P u , c h i t a r L o S ( θ u , c h i t a r ( t ) ) + η N L o s P u , c h i t a r N L o S ( θ u , c h i t a r ( t ) ) ) h u , c h i t a r ( t )
Assuming that the UAV establishes a communication link with c h i t a r at t and c h i t a r maintains a constant transmit power P t to upload data to the UAV, according to Shannon’s formula [44], the data transmission rate between the UAV and c h i t a r can be given as follows:
R c h i t a r t ( t ) = B log 2 ( 1 + P t g u , c h i t a r ( t ) 2 σ 2 )
where B and σ 2 represent the channel bandwidth and noise power, respectively, and P t denotes the transmit power of the node; hovering time (upload data time) is given as follows:
t h o v e r = D c h i t a r ( t ) R c h i t a r t ( t )
the energy consumption for c h i t a r to upload data at time t can be given as follows:
E c h i t a r t ( t ) = P t t h o v e r

3.2. Problem Formulation

In the proposed D-JERDG, we consider the UAV capabilities of energy replenishment and data collection. The UAV departs from the BS and moves according to a planned flight path. It is responsible for collecting data generated by all cluster heads in the network. After completing a mission cycle, the UAV returns to the BS to recharge and transmit the collected data back to the BS. The flight path consists of the hovering positions and the CH node access sequence. First, to improve data collection efficiency, the UAV needs to serve as many nodes as possible within the maximum flight time T and collect data. Second, the charging strategy for the target CH node is crucial to minimize the node death rate and flight costs. To ensure that high-energy-consuming nodes and remote area nodes receive energy replenishment while reducing ineffective flight time, the UAV flight distance and speed need to be controlled.
In summary, the overall objective of D-JERDG is to minimize the UAV flight energy consumption and maximize its throughput while minimizing the node death rate. This problem is a multiobjective optimization problem, and the objective functions can be formulated as follows:
P 1 : min ( N d , E u a v e , L u t o t a l )
The constraint condition can be formulated as follows:
0 n = 0 N t n T
Δ d u , c h i t a r < R d
Δ d u , c m j < R c
N d i = 0 , R E i ( t ) = 0 1 , R E i ( t ) > 0
v min v u ( t ) v max
0 < θ u ( t ) 2 π
0 x u ( t ) L
0 y u ( t ) L
In objective functions, N d i ( i = ( 1 , 2 , , k ) ) is denoted as the number of dead nodes, L u t o t a l represents the total flight distance, and E u a v e represents the UAV’s average flight energy consumption. In the constraint condition, Formula (18) represents the UAV with a maximum flight time of no more than T. Formula (19) represents the maximum data transmission range of the node into the UAV to establish a communication link with the UAV. Formula (20) indicates that nodes can harvest energy from the UAV within the maximum charging range of the UAV. Formula (21) represents the energy state. If R E i ( t ) = 0 , node i is considered a dead node. If R E i ( t ) < 0 , node i is in a normal state. Formulas (22) and (23) represent the size limit of the UAV’s speed and heading angle. Formulas (24) and (25) represent that the UAV’s flight range does not extend beyond the two-dimensional area.

4. A Joint Energy Replenishment and Data Collection Algorithm Based on Deep Reinforcement Learning

In this section, we presented the effective implementation of D-JERDG in UAV-WRSN. First, we introduced a UAV-WRSN with a random topology structure. Then, we described our proposed cluster head selection algorithm. Using the obtained set of CH nodes, we applied the simulated annealing algorithm to solve the traveling salesman problem (TSP). Afterward, we provided the state space, action space, and reward function of the MODDPG algorithm.

4.1. Cluster Head Selection Algorithm

Referring to [39], this network studied in this paper is a UAV-WRSN that uses a dynamic routing protocol. Figure 2 illustrates the generated clusters and data flow. Compared with static routing protocols, dynamic routing protocols can evenly consume the energy of nodes. However, in a traditional LEACH routing protocol [45], the CM nodes determine whether to become CH nodes based on randomly generated thresholds. For each CM node c m j , w j t h is given as follows:
w j t h = p 1 p ( r mod 1 p ) , if j c m j 0 , otherwise
where p is the expected percentage of CH nodes, and r is the current number of rounds. This method lacks consideration for the remaining energy and location of the nodes. We have redesigned the CH node selection mechanism based on these two factors. Due to insufficient energy information before the network operation, we utilize the K-means algorithm to partition all the data into m clusters and identify m centroids within each cluster. The overall optimization objective of the K-means algorithm is to minimize the sum of distances between each point and its respective centroid within the clusters. Based on the distances between nodes and centroids, we have determined m CH nodes. As a result, when the UAV reaches a CH node, it can cover more nodes, thereby improving the charging efficiency.
Figure 2. Illustration of clusters and data flow.
At each round, the cluster member nodes at each cluster elect CH nodes based on their own energy status and location information to balance the energy consumption of the nodes. Dynamic routing protocols are used within each cluster. We stipulate no communication link between the sensor nodes and BS. CM nodes use CH nodes as the next hop for communication, while CH nodes collect data sent by member nodes and transmit them to the UAV. (If a CH node has nonzero data storage after the UAV completes a flight cycle, it will continue to transmit data as a CM node in the next mission.) The weight of CM nodes can be obtained using the following formula:
w c m j = κ E max R E j c m ( t ) R E a v e ( t ) + ( 1 κ ) d i , j d a v e
where R E a v e ( t ) represents the average remaining energy of CM nodes, d a v e represents the average distance between CM nodes and CH nodes, d i , j represents the distance between the CH node c h i and the CM node c m j , and κ is a weighting coefficient. The weight value w j represents the priority or importance of CM; among CM nodes, the one with the highest w j will be selected as the CH node in the next mission round. The CH node will then join the cluster as a CM node. Based on the obtained coordinates of the CH nodes, this paper utilizes the SA algorithm to solve the TSP. The pseudocode for the SA algorithm is presented in Algorithm 1. The specific algorithmic process is shown in Figure 3.
Figure 3. Illustration of the algorithm process.

4.2. UAV Control Algorithm Based on MODDPG

In contrast with reinforcement learning algorithms based on an action value, our scenario involves a continuous action space for controlling the instantaneous speed and heading angle of the UAV. This necessitates a distinct modeling and solution approach. To tackle this continuous action space, we propose utilizing the MODDPG model for policy learning. The algorithmic specifics of MODDPG are elaborated in Algorithm 2.
Algorithm 1 Solve TSP by using simulated annealing algorithm
Input: 
The coordinates of CH nodes l i
Output: 
The sequence of CH nodes l n e w
  1:
Initialize maximum iteration temperature K 0 = 50 , 000 , the sequence of CH nodes l i , and annealing coefficient q = 0.98 .
  2:
K K 0 .
  3:
while  K > K 0  do
  4:
   Swap the order of any two nodes to generate a new visiting sequence l n e w .
  5:
   Calculating l n e w , denoted as d f .
  6:
   if  d f > 0  then
  7:
     if  e x p ( d f K 0 ) > r a n d ( 0 , 1 )  then
  8:
         l f l n e w
  9:
     end if
10:
   else
11:
      K = K × q
12:
   end if
13:
end while
Algorithm 2 MODDPG
Input: 
State space s r l
Output: 
The action of UAV a r l
  1:
Initialize Actor network parameters θ μ , and Target Actor network parameters θ μ , θ μ θ μ .
  2:
Initialize Critic network parameters θ Q , and Target Critic network parameters θ Q , θ Q θ Q .
  3:
Initialize experience Replay Buffer.
  4:
for  episode = 0 to  M do
  5:
   Initialize s r l .
  6:
   for time step = t 1 , t 2 , , t n  do
  7:
     repeat
  8:
        With probability of ϵ choose an action a r l = c l i p ( μ ( s r l | Q μ ) + ϵ , a l o w , a h i g h ) .
  9:
        Perform action a r l and observe reward r r l and next state s r l .
10:
        Store the transition ( s r l , a r l , r r l , s r l ) from P.
11:
        if P ≥ Batch size then
12:
           Randomly sample Mini batch transitions ( s r l , a r l , r r l , s r l ) from P.
13:
           Compute y r l (19).
14:
           Update Critic network by minimizing the critic loss (20).
15:
           Update Actor network by maximizing the actor loss (21).
16:
           Soft Update Target Network Parameters.
17:
            θ Q τ θ Q + ( 1 τ ) θ Q
            θ μ τ θ μ + ( 1 τ ) θ μ
18:
        end if
19:
     until  n = 0 N t n T
20:
   end for
21:
end for
  • State Space: The state space set is defined as follows:
    s r l ( t ) = d u , c h i t a r x ( t ) , d u , c h i t a r y ( t ) , x u ( t ) , y u ( t ) , N f ( t ) , N d ( t ) , i 1 , m
    Specifically, with a state space size of 6, d u , c h i t a r x and d u , c h i t a r y represent the relative distances in the horizontal and vertical coordinates between the UAV and the target CH node i, and N f ( t ) records the cumulative number of times that the UAV has continuously exceeded the restricted area by time t.
  • Action Space: At each t, the action of the UAV is defined as follows:
    θ u ( t ) ( 0 , 2 π ] : Instantaneous heading angle of the UAV along the horizontal direction at t.
    v u ( t ) [ 0 , v m a x ] : Instantaneous speed of the UAV at t.
    The action of the UAV is defined as a r l = [ v u ( t ) , θ u ( t ) ]
  • Reward Function: At any time step t n , the UAV receives a reward r r l . To ensure the throughput of the UAV, the UAV receives an immediate reward r 0 whenever it successfully establishes communication with the target node c h i t a r . The reward function is defined as follows:
    r r l = r 0 ( t ) + r 1 ( t ) , if Δ d u , c h i t a r ( t ) R d r 1 ( t ) ,   otherwise
    r 0 ( t ) = r s e r v e + η 1 λ cov e r ( t ) + η 2 R c h i t a r t r a n ( t )
    r 1 ( t ) = η 2 Δ d u , c h i t a r ( t ) η 3 ( E u p r o ( t ) + N d ( t ) )
    Specifically, the reward function is divided into two parts, r 0 and r 1 , where r s e r v e represents the reward for the UAV establishing a connection with CH nodes; λ c o v e r indicates the number of nodes covered by the UAV; and η 1 , η 2 , and η 3 are reward weight factors for the optimization objectives. r 1 is based on the UAV actions at each time step and includes factors such as the distance between the UAV and CH nodes, the energy consumption of the UAV flight, and the number of dead nodes N d ( t ) .
The MODDPG network framework is shown in Figure 4; it consists of four networks, with the main network and target network each comprising two networks: the actor network, critic network, target actor network, and target critic network. The main network and target network share the same network structure, where the actor network outputs UAV actions, and the critic network evaluates these actions to update the actor network. The MODDPG algorithm optimizes the UAV instantaneous speed and heading angles, ultimately learning the optimal policy.
Figure 4. MODDPG network framework.
The calculation of the target value y r l is as follows:
y r l = r r l + γ Q ( s r l , μ ( s r l ; θ μ ) ; θ Q )
The critic network is trained using gradient descent to minimize the loss between the predicted value Q ( s r l , a r l ; θ Q ) and the target value y r l .
L o s s = 1 M ( y r l Q ( s r l , a r l ; θ Q ) ) 2
Update the actor network using a gradient-based policy algorithm:
θ μ J = 1 M a r l Q ( s r l , a r l ; θ Q ) | s = s r l , a = μ ( s r l ) θ μ μ ( s r l ; θ μ ) | s r l
where J represents the total discounted cumulative reward. The specific algorithm flow is described as follows: Initialize the replay buffer, critic network, and actor network parameters θ Q and θ μ , as well as the target critic network and target actor network parameters θ Q and θ μ (lines 1–3). Perform the exploration phase for the UAV in lines 4–10, and generate an action a r l = c l i p ( μ ( s r l | Q μ ) + ϵ ) from the actor network. This process is repeated until the maximum task time is reached, and the experiences are stored in the replay buffer. Updating network parameters, when the experience replay buffer P is full, start training the networks to update actor network and critic network parameters (lines 11–15). Update the critic network parameters by minimizing the loss function (line 14), and update the actor network parameters using the deterministic policy gradient (line 15). Finally, use the soft update technique to update the parameters of the target actor network, controlling the update frequency to converge faster toward the optimal policy (lines 17–21).

5. Experimental Results

In this section, we conducted simulation results to validate the effectiveness of D-JERDG. Specifically, we provided numerical results and analyzed the convergence of D-JERDG by adjusting the number of CM nodes, CH nodes, and maximum charging radius. We compared D-JERDG with MODDPG and the random method. We analyzed the differences between the algorithms from five aspects: node death rate, data overflow rate, UAV throughput, average flight energy consumption, and time utilization. By comparing these metrics, we were able to evaluate and highlight the differences and advantages of D-JERDG over the other algorithms in terms of efficiency and performance.
  • D-JERDG: D-JERDG uses the K-means clustering algorithm to divide nodes into multiple clusters, designing the CH node selection mechanism based on an improved dynamic routing protocol. The SA algorithm determines the CH node access sequence, and the UAV flight actions are controlled by inputting the UAV and network states into the DRL model MODDPG.
  • MODDPG (multiobjective deep deterministic policy gradient) [20]: This method selects the node with the highest data volume as the target node. By observing the relative positions of the UAV and the target node, as well as the node states, a DRL model is established to control the UAV learning of all node positions. Through accessing the nodes, data collection and range-based charging are achieved.
  • Random: The method is based on the MODDPG [20] definition, with the difference being that the random algorithm randomly selects a node as the target node and uses the DRL model to optimize the UAV’s flight trajectory.

5.1. Simulation Settings

We consider the UAV taking off from the BS, the maximum flight time T to be set at 10 min. The sensor nodes are randomly distributed within a 400 × 400 m 2 square two-dimensional area. The UAV flies at an altitude of 10 m, with a maximum data transmission radius R d = 10 m and a maximum charging radius R c = 30 m. The maximum flight speed of the UAV is set to 20 m/s [20]. The coordinates of the nodes are randomly generated and remain fixed. The nodes’ maximum data storage capacity is Q m a x = 100 Mb [40]. The total energy of the nodes is set to E m a x = 800 J [40], and their initial energy is randomly generated within the range of [0 J, 800 J] [40]. The transmission energy consumption for CM nodes is e t = 0.4 J [43], and the reception energy consumption for CH nodes is e r = 0.5 J [43]. Table 3 and Table 4, respectively, show the network parameters of MODDPG and the main environmental parameters, as specified in reference [20]. The operating system environment of the simulation experiment is TensorFlow 2.5.0, TensorLayer 2.2.5, and Python 3.7 on a Linux server with four NVIDIA 3090 GPUs.
Table 3. Network parameters.
Table 4. Main simulation parameters.

5.2. Evaluation Metrics

We introduce five key performance metrics to analyze and compare the optimization effects of D-JERDG on UAV-WRSN in terms of system charging efficiency, data collection efficiency, and UAV flight energy consumption.
  • Node death rate: The node death rate is determined by the proportion of dead nodes, where node i is considered dead when R E i ( t ) = 0 . A lower node death rate indicates a higher system charging efficiency. This metric determines which charging strategy can maintain the maximum number of surviving nodes in the experiment.
  • Node data overflow rate: The node data overflow rate is determined by the proportion of nodes with data storage exceeding the maximum capacity, represented as D i ( t ) = 100 . This metric measures the effectiveness of UAV data collection and the clustering algorithm.
  • UAV throughput: The UAV throughput is the ratio of the number of nodes accessed by the UAV to the total number of nodes within the maximum flight time. It measures the overall efficiency of the UAV-WRSN system.
  • Average flight energy consumption: The average flight energy consumption is defined as the ratio of the total energy consumption during UAV flight to the maximum flight time. It is a key indicator for evaluating the energy consumption of the UAV-WRSN system.
  • Time utilization rate: The time utilization rate represents the ratio of the hover time to the flight time of the UAV during a flight mission. A lower time utilization indicates a higher proportion of time occupied by UAV flights, serving as a metric for evaluating the data collection efficiency of the UAV. The time utilization rate R t u is given as follows:
    R t u = t h o v e r t f l y ( T = t h o v e r + t f l y )

5.3. Convergence and Stability

After D-JERDG is implemented, we first test the convergence and stability of the proposed model. The episode was set as 400, and the model of the number of different clusters, the number of nodes in each cluster, and the maximum charging radius of the UAV were trained. “C20N200R30” means that there are 20 CH nodes and 180 CM nodes, the maximum charging radius of UAV is R c = 30 m, and the other sections in this chapter follow the same pattern. Then the functional relationship between episode and reward is shown in Figure 5. As can be seen from Figure 5, the reward of most models increased rapidly before 150 episodes. After 200 episodes, the value of rewards becomes stable and oscillates around a certain value until the training ends. The results indicate that the trained D-JERDG algorithm successfully converges and can make high-reward decisions in a dynamically changing network state. Figure 5a shows the reward curves under different numbers of CH nodes and CM nodes, ranging from “C20N200R30” to “C40N400R30” experimental scenarios. Figure 5b displays the reward curves for different numbers of CH nodes, with the experimental scenarios ranging from “C15N200R30” to “C30N200R30”. Figure 5c illustrates the reward curves under different UAV charging radii, with the experimental scenarios ranging from “C20N200R15” to “C20N200R30”.
Figure 5. Learning curve with the episode = 400.
Additionally, Figure 6 displays the loss curve and temporal difference error (TD error) curve of network training for the three algorithms, with the experimental scenario being “C20N200R30”. In Figure 6a, the curve trends of the three algorithms are mostly consistent before 150 episodes. After that, due to the increased UAV throughput in D-JERDG, some fluctuations occur. However, overall, the loss curves of the three algorithms converge before 400 episodes. In Figure 6b, we show the TD error curve of network training for the three algorithms. Overall, all three algorithms can converge before 400 iterations. However, before 100 episodes, D-JERDG has a slower convergence rate compared with the other two algorithms. Because we expect D-JERDG to learn the positions of more nodes in the early training episodes, we have implemented a strategy where the set of CH nodes initialized using the K-means algorithm is not fixed. The purpose of this approach is to ensure that D-JERDG can learn the positions of as many nodes as possible. By flexibly adjusting the CH node set, D-JERDG can better adapt to the varying node distributions in different environments, enhancing its performance and training effectiveness. After 100 episodes, as the rewards for D-JERDG start to increase and gradually converge, it also exhibits convergence in the TD error curve.
Figure 6. Learning curve with the episode = 400.

5.4. Performance Comparison

Specifically, without loss of generality, we conducted experiments with different numbers of nodes and varying UAV flight speeds. The number of nodes was set from “C10N100R30” to “C40N400R30”, with a maximum data transmission radius of R d = 10 m, a maximum charging radius of R c = 30 m, and a maximum flight speed ranging from 15 m/s to 20 m/s. With an increasing number of nodes, both the energy consumption and data volume of CH nodes increase within the same period. A larger flight speed allows the UAV to reach nodes faster, but it also increases energy consumption. On the other hand, a smaller speed can save energy but may sacrifice some throughput and result in a certain number of dead nodes, thereby affecting UAV throughput, the number of dead nodes, and time utilization. Additionally, as the number of nodes increases, in our proposed approach, the UAV flight time and distance are influenced by the increasing number of CH nodes, which also affects flight energy consumption. Based on these hypotheses, we conducted comparative experiments.

5.4.1. Node Death Rate

In this section, we compared the performances of different algorithms in terms of node death rate. Specifically, in the MODDPG experiment, we designated the node with the least remaining energy as the target node. The experimental results are shown in Figure 7.
Figure 7. Experimental results of node death rate.
Figure 7a demonstrated a clear disparity in node death rates between D-JERDG and Random. Initially, MODDPG exhibited lower node death rates than D-JERDG for up to 200 nodes. This can be attributed to the advantage of the greedy-based node charging strategy in scenarios with a small number of nodes and low density. However, after reaching 200 nodes, D-JERDG demonstrated a lower node death rate than MODDPG, with a declining trend beyond 300 nodes. When the number of nodes is 250, compared with the MODDPG and random, D-JERDG can decrease the node death rate by about 17 % and 44 % , respectively. This improvement can be attributed to the clustering algorithm employed in D-JERDG, where several nodes closest to the CH node are selected as CM nodes. This approach ensures that the UAV covers more nodes during each hover, enhancing charging efficiency. The effectiveness of the proposed CH node selection mechanism was also validated, as it increased the likelihood of selecting CM nodes farther from the current CH node as new CH nodes. This enabled energy replenishment and reduced the node death rate of remote area nodes, making D-JERDG more suitable for large-scale networks aiming to maintain network connectivity. From Figure 7b, we examined the variations in node death rates across different algorithms while considering different UAV flight speeds, with 200 nodes in the experiment. Overall, increasing the UAV flight speed resulted in reduced fluctuations in the node death rate, and D-JERDG consistently outperformed MODDPG and random by maintaining lower node death rates. When the speed of the UAV was 15 to 20, compared with MODDPG and random, the average decline rate of D-JEGDG in terms of node death rate was about 10 % and 45.5 % , respectively. Furthermore, as the flight speed increased, the node death rate exhibited a significant decrease. This highlights the advantageous performance of D-JERDG in terms of node death rate when compared with the other algorithms.

5.4.2. Node Data Overflow Rate

In this section, we analyze the differences in node data overflow rates among different algorithms. In the MODDPG experiment, we designate the node with the highest data volume as the target node, and the experimental results are shown in Figure 8. In Figure 8a,b, with an increasing number of nodes, compared with MODDPG and random, the average decline rate of D-JEGDG in terms of node overflow rate is about 46.4 % and 48.2 % , respectively. With an increasing speed of the UAV, it is about 35.1 % and 46 % , respectively. This is because the dynamic routing protocol periodically selects CM nodes as CH nodes, allowing CH nodes to gather data from CM nodes and reduce node data overflow. However, for MODDPG and random, the one-to-one data collection approach leads to lower efficiency and inevitable node data overflow.
Figure 8. Experimental results of node data overflow rate.

5.4.3. UAV Throughput

In this section, we compare the performances of different algorithms in terms of UAV throughput by changing the number of CH nodes and CM nodes in the network. The experimental results are shown in Figure 9. In Figure 9a, we observe that as the number of nodes increases, all algorithms experience a decrease in UAV throughput. However, D-JERDG consistently achieves the highest throughput among the algorithms. Compared with MODDPG and random, the average growth rate of D-JEGDG in terms of UAV throughput is about 79.3 % and 43 % , respectively. In D-JERDG, the use of a K-means-based clustering algorithm for CH node selection introduces randomness, allowing the UAV to quickly learn the positions of all nodes by obtaining different node coordinates in each training round. This enables D-JERDG to achieve a higher node access count within the same flight time. At 250 nodes, D-JERDG reaches a node access count of 50, indicating that it completes two rounds of CH node traversal. Beyond 300 nodes, D-JERDG throughput reaches a plateau and remains competitive with MODDPG. This can be attributed to the increased number of CH nodes requiring the UAV to spend more time hovering for data collection, leading to longer flight times and a decrease in node access count. In Figure 9b, we analyze the impact of different UAV flight speeds on throughput with a fixed number of 200 nodes. Overall, compared with MODDPG and random, D-JERDG outperforms MODDPG and random algorithms, and the throughput significantly increases as the flight speed rises. The average growth rate of D-JEGDG in terms of UAV throughput is about 55.3 % and 42.4 % , respectively.
Figure 9. Experimental results of UAV throughput.

5.4.4. Average Flight Energy Consumption

In this section, we analyze the differences in average UAV flight energy consumption among three algorithms. The experimental results are depicted in Figure 10. In Figure 10a, we observe that the UAV energy consumption in D-JERDG initially increases and then decreases as the number of nodes increases. Before reaching 200 nodes, the larger distances between CH nodes prompt the UAV to adapt its flight speed to minimize the flight time and efficiently collect data, resulting in an increasing energy consumption curve. However, after 200 nodes, as the number and density of CH nodes increase, the distances between them become shorter. Compared with MODDPG and random, the average decline rate of D-JEGDG in terms of average flight energy consumption is about 4.2 % and 4 % , respectively. When the number of nodes is 350, D-JERDG can decrease the average flight energy consumption by about 10 % and 9.8 % , respectively. As a result, the UAV adjusts its flight speed by reducing it, which leads to longer flight times but decreases the overall UAV energy consumption. Compared with MODDPG and random, in Figure 10b, we examine the impact of different flight speeds on average energy consumption under the three algorithms. Overall, there is no significant difference among the algorithms (D-JERDG, MODDPG, and random) in terms of average energy consumption. However, as the flight speed increases, the average energy consumption increases as well. Notably, D-JERDG consistently maintains a lower average UAV flight energy consumption compared with MODDPG and random algorithms under different flight speed conditions.
Figure 10. Experimental results of average flight energy consumption.

5.4.5. Time Utilization Rate

In this section, we compare the differences in time utilization rates between the D-JERDG and MODDPG algorithms. The experimental results are presented in Figure 11. In Figure 11a, we observe that the time utilization of D-JERDG initially increases and then decreases as the number of CH nodes varies. This trend aligns with the analysis discussed earlier. In general, D-JERDG exhibits a similar pattern to MODDPG, and after 300 nodes, their time utilization becomes comparable. When the number of nodes is 150 to 350, the average growth rate of D-JEGDG in terms of time utilization rates is about 23 % . The reason behind this behavior lies in the D-JERDG method, which utilizes the SA algorithm to adjust the sequence of CH node visits, thereby reducing the UAV flight distance and time. Before reaching 300 nodes, with a smaller network size, the total flight time includes longer hover times. At 200 nodes, the time utilization reaches its peak, which aligns with the trend observed in the average UAV energy consumption curve. However, beyond 300 nodes, as the number of CH nodes increases, the flight time also increases. With the total flight time remaining constant, the time utilization decreases accordingly. In Figure 11b, we examine the impact of UAV flight speed on time utilization under both algorithms. Notably, D-JERDG consistently achieves higher time utilization compared with MODDPG, maintaining the highest level throughout the different flight speeds. As the speed of UAV increases, the average growth rate of D-JEGDG in terms of time utilization rates is about 36 % .
Figure 11. Experimental results of time utilization rate.

6. Conclusions

This study proposes a DRL-based method called D-JERDG for joint data collection and energy replenishment in UAV-WRSN. The main problems D-JERDG addresses are individual nodes’ low data collection efficiency and the imbalance in node energy consumption. D-JERDG optimizes the network to tackle these issues by considering node inefficiency, UAV flight energy consumption, and UAV throughput. The clustering algorithm based on K-means and an improved dynamic routing protocol is used to cluster the nodes in the network. CH nodes are selected based on the remaining energy and geographical locations of the nodes within the clusters, effectively adapting to the dynamic nature of node energy consumption. Furthermore, a simulated annealing algorithm is employed to determine the visiting sequence, and the DRL model MODDPG is introduced to control the UAV for node servicing. Through extensive simulation results, the evaluation metrics of node inefficiency, UAV throughput, and average flight energy consumption are used to assess the performance of D-JERDG. The results demonstrate that D-JERDG achieves joint optimization of multiple objectives. It outperforms the existing MODDPG approach by significantly reducing node inefficiency, saving flight costs, and improving data collection efficiency. Moreover, multiple research studies suggest that multiagent systems have significant advantages in accomplishing complex tasks. Therefore, Multi-UAV-WRSN is expected to be a primary research focus in the future.

Author Contributions

Conceptualization, J.L.; Data curation, J.L. and Z.D.; Funding acquisition, Y.F.; Methodology, J.L. Project administration, Y.F. and N.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant No. 62062047, the major scientific and technological projects in Yunnan Province under Grant No. 202202AD080006, and the Yuxi Normal University under Grant No. 202105AG070010.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to it is not permitted.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kurs, A.; Karalis, A.; Moffatt, R.; Joannopoulos, J.D.; Fisher, P.; Soljacic, M. Wireless power transfer via strongly coupled magnetic resonances. Science 2007, 317, 83–86. [Google Scholar] [CrossRef] [PubMed]
  2. Kurs, A.; Moffatt, R.; Soljačić, M. Simultaneous mid-range power transfer to multiple devices. Appl. Phys. Lett. 2010, 96, 044102. [Google Scholar] [CrossRef]
  3. Yang, W.; Lin, C.; Dai, H.; Wang, P.; Ren, J.; Wang, L.; Wu, G.; Zhang, Q. Robust Wireless Rechargeable Sensor Networks. IEEE/ACM Trans. Netw. 2023, 31, 949–964. [Google Scholar] [CrossRef]
  4. Rodić, A.; Mester, G. Virtual WRSN—Modeling and simulation of wireless robot-sensor networked systems. In Proceedings of the IEEE 8th International Symposium on Intelligent Systems and Informatics, Subotica, Serbia, 10–11 September 2010; pp. 115–120. [Google Scholar] [CrossRef]
  5. Deng, Q.; Ouyang, Y.; Tian, S.; Ran, R.; Gui, J.; Sekiya, H. Early Wake-Up Ahead Node for Fast Code Dissemination in Wireless Sensor Networks. IEEE Trans. Veh. Technol. 2021, 70, 3877–3890. [Google Scholar] [CrossRef]
  6. Schurgers, C.; Srivastava, M. Energy efficient routing in wireless sensor networks. In Proceedings of the 2001 MILCOM Proceedings Communications for Network-Centric Operations: Creating the Information Force (Cat. No.01CH37277), McLean, VA, USA, 28–31 October 2001; Volume 1, pp. 357–361. [Google Scholar] [CrossRef]
  7. Goyal, N.; Dave, M.; Verma, A.K. Data aggregation in underwater wireless sensor network: Recent approaches and issues. J. King Saud Univ.-Comput. Inf. Sci. 2019, 31, 275–286. [Google Scholar] [CrossRef]
  8. Dai, H.; Zhang, Y.; Wang, X.; Liu, A.X.; Chen, G. Omnidirectional Chargability with Directional Antennas. IEEE Trans. Mob. Comput. 2023, 23, 4483–4500. [Google Scholar] [CrossRef]
  9. Dai, H.; Wang, X.; Lin, X.; Gu, R.; Shi, S.; Liu, Y.; Dou, W.; Chen, G. Placing Wireless Chargers With Limited Mobility. IEEE Trans. Mob. Comput. 2023, 22, 3589–3603. [Google Scholar] [CrossRef]
  10. Zeng, Y.; Zhang, R.; Lim, T.J. Wireless communications with unmanned aerial vehicles: Opportunities and challenges. IEEE Commun. Mag. 2016, 54, 36–42. [Google Scholar] [CrossRef]
  11. Pedditi, R.B.; Debasis, K. Energy Efficient Routing Protocol for an IoT-Based WSN System to Detect Forest Fires. Appl. Sci. 2023, 13, 3026. [Google Scholar] [CrossRef]
  12. da Silva, R.I.; Del Duca Almeida, V.; Poersch, A.M.; Nogueira, J.M.S. Wireless sensor network for disaster management. In Proceedings of the 2010 IEEE Network Operations and Management Symposium—NOMS 2010, Osaka, Japan, 19–23 April 2010; pp. 870–873. [Google Scholar] [CrossRef]
  13. Yang, J.; Wang, X.; Li, Z.; Yang, P.; Luo, X.; Zhang, K.; Zhang, S.; Chen, L. Path planning of unmanned aerial vehicles for farmland information monitoring based on WSN. In Proceedings of the 2016 12th World Congress on Intelligent Control and Automation (WCICA), Guilin, China, 12–15 June 2016; pp. 2834–2838. [Google Scholar] [CrossRef]
  14. Liu, Y.; Pan, H.; Sun, G.; Wang, A. Scheduling Optimization of Charging UAV in Wireless Rechargeable Sensor Networks. In Proceedings of the 2021 IEEE Symposium on Computers and Communications (ISCC), Athens, Greece, 5–8 September 2021; pp. 1–7. [Google Scholar] [CrossRef]
  15. Park, S.Y.; Jeong, D.; Shin, C.S.; Lee, H. DroneNet+: Adaptive Route Recovery Using Path Stitching of UAVs in Ad-Hoc Networks. In Proceedings of the GLOBECOM 2017—2017 IEEE Global Communications Conference, Singapore, 4–8 December 2017; pp. 1–7. [Google Scholar] [CrossRef]
  16. Wang, Y.; Gao, Z.; Zhang, J.; Cao, X.; Zheng, D.; Gao, Y.; Ng, D.W.K.; Renzo, M.D. Trajectory Design for UAV-Based Internet of Things Data Collection: A Deep Reinforcement Learning Approach. IEEE Internet Things J. 2022, 9, 3899–3912. [Google Scholar] [CrossRef]
  17. Wu, Q.; Zeng, Y.; Zhang, R. Joint Trajectory and Communication Design for Multi-UAV Enabled Wireless Networks. IEEE Trans. Wirel. Commun. 2018, 17, 2109–2121. [Google Scholar] [CrossRef]
  18. Baek, J.; Han, S.I.; Han, Y. Energy-Efficient UAV Routing for Wireless Sensor Networks. IEEE Trans. Veh. Technol. 2020, 69, 1741–1750. [Google Scholar] [CrossRef]
  19. Delahaye, D.; Chaimatanan, S.; Mongeau, M. Simulated Annealing: From Basics to Applications. In Handbook of Metaheuristics; Gendreau, M., Potvin, J.Y., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 1–35. [Google Scholar] [CrossRef]
  20. Yu, Y.; Tang, J.; Huang, J.; Zhang, X.; So, D.K.C.; Wong, K.K. Multi-Objective Optimization for UAV-Assisted Wireless Powered IoT Networks Based on Extended DDPG Algorithm. IEEE Trans. Commun. 2021, 69, 6361–6374. [Google Scholar] [CrossRef]
  21. Li, J.; Sun, G.; Wang, A.; Lei, M.; Liang, S.; Kang, H.; Liu, Y. A many-objective optimization charging scheme for wireless rechargeable sensor networks via mobile charging vehicles. Comput. Netw. 2022, 215, 109196. [Google Scholar] [CrossRef]
  22. Dong, Y.; Li, S.; Bao, G.; Wang, C. An Efficient Combined Charging Strategy for Large-Scale Wireless Rechargeable Sensor Networks. IEEE Sens. J. 2020, 20, 10306–10315. [Google Scholar] [CrossRef]
  23. Zhu, X.; Shen, L.; Yum, T.S.P. Hausdorff Clustering and Minimum Energy Routing for Wireless Sensor Networks. IEEE Trans. Veh. Technol. 2009, 58, 990–997. [Google Scholar] [CrossRef]
  24. Chand, S.; Singh, S.; Kumar, B. Heterogeneous HEED protocol for wireless sensor networks. Wirel. Pers. Commun. 2014, 77, 2117–2139. [Google Scholar] [CrossRef]
  25. Wu, Q.; Sun, P.; Boukerche, A. A Novel Joint Data Gathering and Wireless Charging Scheme for Sustainable Wireless Sensor Networks. In Proceedings of the ICC 2020—2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  26. Li, B.; Xiao, X.; Tang, S.; Ning, W. An Energy Efficiency-oriented Routing Protocol for Wireless Rechargeable Sensor Networks. In Proceedings of the 2021 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2021; pp. 1–5. [Google Scholar] [CrossRef]
  27. Han, G.; Yang, X.; Liu, L.; Zhang, W. A Joint Energy Replenishment and Data Collection Algorithm in Wireless Rechargeable Sensor Networks. IEEE Internet Things J. 2018, 5, 2596–2604. [Google Scholar] [CrossRef]
  28. Zhao, M.; Li, J.; Yang, Y. A Framework of Joint Mobile Energy Replenishment and Data Gathering in Wireless Rechargeable Sensor Networks. IEEE Trans. Mob. Comput. 2014, 13, 2689–2705. [Google Scholar] [CrossRef]
  29. Sah, D.K.; Amgoth, T. Renewable energy harvesting schemes in wireless sensor networks: A Survey. Inf. Fusion 2020, 63, 223–247. [Google Scholar] [CrossRef]
  30. Qi, F.; Zhu, X.; Mang, G.; Kadoch, M.; Li, W. UAV Network and IoT in the Sky for Future Smart Cities. IEEE Netw. 2019, 33, 96–101. [Google Scholar] [CrossRef]
  31. Xu, J.; Zeng, Y.; Zhang, R. UAV-Enabled Wireless Power Transfer: Trajectory Design and Energy Optimization. IEEE Trans. Wirel. Commun. 2018, 17, 5092–5106. [Google Scholar] [CrossRef]
  32. Liu, Y.; Pan, H.; Sun, G.; Wang, A.; Li, J.; Liang, S. Joint Scheduling and Trajectory Optimization of Charging UAV in Wireless Rechargeable Sensor Networks. IEEE Internet Things J. 2022, 9, 11796–11813. [Google Scholar] [CrossRef]
  33. Wu, P.; Xiao, F.; Sha, C.; Huang, H.; Sun, L. Trajectory Optimization for UAVs’ Efficient Charging in Wireless Rechargeable Sensor Networks. IEEE Trans. Veh. Technol. 2020, 69, 4207–4220. [Google Scholar] [CrossRef]
  34. Zhao, C.; Wang, Y.; Zhang, X.; Chen, S.; Wu, C.; Teo, K.L. UAV Dispatch Planning for a Wireless Rechargeable Sensor Network for Bridge Monitoring. IEEE Trans. Sustain. Comput. 2023, 8, 293–309. [Google Scholar] [CrossRef]
  35. Baek, J.; Han, S.I.; Han, Y. Optimal UAV Route in Wireless Charging Sensor Networks. IEEE Internet Things J. 2020, 7, 1327–1335. [Google Scholar] [CrossRef]
  36. Lin, C.; Guo, C.; Du, W.; Deng, J.; Wang, L.; Wu, G. Maximizing Energy Efficiency of Period-Area Coverage with UAVs for Wireless Rechargeable Sensor Networks. In Proceedings of the 2019 16th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON), Boston, MA, USA, 10–13 June 2019; pp. 1–9. [Google Scholar] [CrossRef]
  37. Hu, H.; Xiong, K.; Qu, G.; Ni, Q.; Fan, P.; Letaief, K.B. AoI-Minimal Trajectory Planning and Data Collection in UAV-Assisted Wireless Powered IoT Networks. IEEE Internet Things J. 2021, 8, 1211–1223. [Google Scholar] [CrossRef]
  38. Bouhamed, O.; Ghazzai, H.; Besbes, H.; Massoud, Y. A UAV-Assisted Data Collection for Wireless Sensor Networks: Autonomous Navigation and Scheduling. IEEE Access 2020, 8, 110446–110460. [Google Scholar] [CrossRef]
  39. Liu, R.; Qu, Z.; Huang, G.; Dong, M.; Wang, T.; Zhang, S.; Liu, A. DRL-UTPS: DRL-Based Trajectory Planning for Unmanned Aerial Vehicles for Data Collection in Dynamic IoT Network. IEEE Trans. Intell. Veh. 2023, 8, 1204–1218. [Google Scholar] [CrossRef]
  40. Li, K.; Ni, W.; Dressler, F. LSTM-Characterized Deep Reinforcement Learning for Continuous Flight Control and Resource Allocation in UAV-Assisted Sensor Network. IEEE Internet Things J. 2022, 9, 4179–4189. [Google Scholar] [CrossRef]
  41. Shan, T.; Wang, Y.; Zhao, C.; Li, Y.; Zhang, G.; Zhu, Q. Multi-UAV WRSN charging path planning based on improved heed and IA-DRL. Comput. Commun. 2023, 203, 77–88. [Google Scholar] [CrossRef]
  42. Liu, N.; Zhang, J.; Luo, C.; Cao, J.; Hong, Y.; Chen, Z.; Chen, T. Dynamic Charging Strategy Optimization for UAV-Assisted Wireless Rechargeable Sensor Networks Based On Deep Q-network. IEEE Internet Things J. 2023, 11. [Google Scholar] [CrossRef]
  43. Wang, Y.; Feng, Y.; Liu, M.; Liu, N. Dynamic Spatiotemporal Charging Scheduling Based on Deep Reinforcement Learning for WRSN. J. Softw. 2023, 35, 1485–1501. [Google Scholar]
  44. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  45. Heinzelman, W.; Chandrakasan, A.; Balakrishnan, H. Energy-efficient communication protocol for wireless microsensor networks. In Proceedings of the 33rd Annual Hawaii International Conference on System Sciences, Maui, Hawaii, 4–7 January 2000; Volume 2, p. 10. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.