Next Article in Journal
Sollerman Hand Function Sub-Test “Write with a Pen”: A Computer-Vision-Based Approach in Rehabilitation Assessment
Previous Article in Journal
Research on Diesel Engine Fault Status Identification Method Based on Synchro Squeezing S-Transform and Vision Transformer
Previous Article in Special Issue
FedDdrl: Federated Double Deep Reinforcement Learning for Heterogeneous IoT with Adaptive Early Client Termination and Local Epoch Adjustment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mobility-Aware Resource Allocation in IoRT Network for Post-Disaster Communications with Parameterized Reinforcement Learning

1
Department of Electrical and Electronic Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, Sungai Long Campus, Kajang 43000, Malaysia
2
Department of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Lembah Pantai, Kuala Lumpur 50603, Malaysia
3
Resilient ICT Research Center, Network Research Institute, National Institute of Information and Communications Technology (NICT), Tokyo 184-8795, Japan
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6448; https://doi.org/10.3390/s23146448
Submission received: 10 March 2023 / Revised: 4 May 2023 / Accepted: 15 June 2023 / Published: 17 July 2023
(This article belongs to the Special Issue Internet of Things, Sensing and Cloud Computing)

Abstract

:
Natural disasters, including earthquakes, floods, landslides, tsunamis, wildfires, and hurricanes, have become more common in recent years due to rapid climate change. For Post-Disaster Management (PDM), authorities deploy various types of user equipment (UE) for the search and rescue operation, for example, search and rescue robots, drones, medical robots, smartphones, etc., via the Internet of Robotic Things (IoRT) supported by cellular 4G/LTE/5G and beyond or other wireless technologies. For uninterrupted communication services, movable and deployable resource units (MDRUs) have been utilized where the base stations are damaged due to the disaster. In addition, power optimization of the networks by satisfying the quality of service (QoS) of each UE is a crucial challenge because of the electricity crisis after the disaster. In order to optimize the energy efficiency, UE throughput, and serving cell (SC) throughput by considering the stationary as well as movable UE without knowing the environmental priori knowledge in MDRUs aided two-tier heterogeneous networks (HetsNets) of IoRT, the optimization problem has been formulated based on emitting power allocation and user association combinedly in this article. This optimization problem is nonconvex and NP-hard where parameterized (discrete: user association and continuous: power allocation) action space is deployed. The new model-free hybrid action space-based algorithm called multi-pass deep Q network (MP-DQN) is developed to optimize this complex problem. Simulations results demonstrate that the proposed MP-DQN outperforms the parameterized deep Q network (P-DQN) approach, which is well known for solving parameterized action space, DQN, as well as traditional algorithms in terms of reward, average energy efficiency, UE throughput, and SC throughput for motionless as well as moveable UE.

1. Introduction

Due to rapid climate change, natural catastrophes, including earthquakes, floods, landslides, tsunamis, wildfires, and hurricanes, have frequently occurred worldwide, directly affecting humanity by direct and secondary deaths of human, economic, and environmental losses [1,2]. Recently, authorities have deployed various types of robots and drones such as unmanned ground vehicles (UGVs), unmanned aerial vehicles (UAVs), unmanned underwater vehicles (UUVs), mobile robots, health care robots, etc. that can be defined as user equipment (UE) for post-disaster management (PDM) because they can be dispatched to locations which cannot be accessed or too risky to be accessed by human rescuers after a disaster has occurred and yet work nonstop [3,4,5,6]. Furthermore, the Internet of Things (IoT) and robotic technologies have recently been combined in order to expand the functional capabilities of these robots, commonly called the Internet of Robotic Things (IoRT) [7,8,9,10]. The communication between IoRT devices can be provided by 4G/LTE/5G and beyond cellular communication, which can be the heterogeneous network (HetNet) [11]. Residents in the affected area are unable to express their demands and circumstances when regular IoRT/IoT networks are substantially compromised due to the disaster [12,13]. However, victims frequently require essential services, for example, food, water, medical assistance, and shelter, which must be rapidly arranged within 72 h after the disaster to save lives and mitigate losses. As a result, the immediate requirement is for rapid and effective post-disaster network rebuilding [13].
In a post-disaster scenario, Movable and Deployable Resource Units (MDRUs) developed by Tohoku University (TU), Japan, and Nippon Telegraph and Telephone (NTT) can be adopted as SBSs to restore the network coverage and capacity due to the rapid deployment, flexibility, interoperability, and resilience [14]. On the other hand, the UAV-aided cellular network has been regarded as a crucial solution for PDM; however, UAVs can support a maximum of one hour due to the power limitation [15]. Due to the electric power unavailability for a long time in disaster-affected areas, van-type MDRU has around seven hours of battery life, similar to the battery backup of the small base station (SBS), was conducted the field test for a comprehensive solution to satisfy the demands of UEs in disaster areas [16]. In addition, the authors of [17] recommended deploying MDRUs to provide communication services where few SBSs have been interrupted due to a small disaster; in contrast, others are in working condition to reconstruct the whole cellular network. Furthermore, MDRU has been deployed to provide communication services and process the big data for minimizing the latency that is important to find injured people, animals, and damaged infrastructure in the disaster-affected area [18]. It was also used to build heterogeneous wireless IoT networks to sense, exchange, and monitor natural disasters by humanitarian organizations [19]. The authors [20] reconstructed the communication network deploying MDRU in the disaster area and found the best results in small coverage and dense area. Furthermore, the intelligent post-disaster network was developed using big crowd data. The authors deployed MDRUs connected with multiple still-alive base stations by using the virtual vertex [21]. The author in [22] recommended building a resilient IoT network by deploying MDRU, which is connected to a backbone network. In summary, all still-alive SBSs and MDRUs (replacement of damaged SBSs) are generally associated with MBS through wireless backhaul connection to handle the vast UE data generated for PDM with the quality of service (QoS). The whole system can be called two-tier HetNet, as illustrated in Figure 1.
The mobility of UEs is one of the critical points to collecting the data in the disaster-affected area for PDM, which impacts channel conditions, path loss, shadow effect, etc., and is a more realistic phenomenon. Our impression is that only a few studies have been conducted based on the mobility of UEs. In [23], we developed a twin delayed deep deterministic policy gradient (TD3) based power allocation algorithm considering UE mobility in one tier IoRT network; however, the major limitation of that research is UE association was not considered. In [24], power allocation optimization is conducted by convex optimization. However, most of the formulated problems, for example, dynamic PA, maximization of the coverage area, traffic offloading, traffic load balancing with user association, maximization of sum rate, etc., are strongly nonconvex as well as nondeterministic polynomial-time hardness (NP-hard) [25]. In this research, we investigate optimizing the energy efficiency and throughput of UE as well as serving cell (SC) of the MDRU-aided two-tier HetNet scenario by ensuring the QoS of mobility-aware UEs where user association and power allocation for each UE have been considered without knowing the environmental priori knowledge. Hence, this optimization problem is strongly nonconvex as well as NP-hard.
Deep Reinforcement Learning (DRL) algorithms (one of the most potent AI algorithms) can handle nonconvex and NP-hard optimization problems [25,26] by leveraging the power of deep neural networks to learn a policy that maps states to actions. The reinforcement learning framework provides a way to learn this policy by trial and error through interaction with the environment illustrated in Figure 2. By learning from experience, the agent can gradually improve its performance and find suitable solutions to complex optimization problems. Consequently, DRL has been applied in wireless communication, robots, computer vision, IoT, IoRT, etc. [27]. According to the action space, DRL is classified as discrete action space algorithms, for example, Deep Q network (DQN), Double DQN (DDQN), Rainbow DQN, dueling DQN, etc., continuous action space algorithms; for instance, Deep Deterministic Policy Gradient (DDPG), Twin delayed DDPG (TD3), Distributed Distributional DDPG (D4PG), Soft Actor-Critic (SAC), etc., that are based on policy gradient and hybrid action space algorithms, such as Q-PAMDP, PA-DDPG, Parametrized DQN (P-DQN), Multi-Pass DQN (MP-DQN), etc., that can handle discrete-continuous combinedly [28].
In [29], authors investigated a combined strategy for power allocation, which is considered continuous action and user association that is discrete action to improve downlink energy efficiency while ensuring QoS of stationary UEs under standard backhaul connection in HetNet by implementing the hybrid action space-based DRL called P-DQN. Furthermore, the architecture of P-DQN is similar to DDPG. However, discrete action is produced from the Q network, while continuous action is generated from the actor network. In P-DQN, the joint action parameter vector for all whole actions at a one-time step is taken as the input of the Q network. As a result, the gradients concerning all action parameters are calculated, which generates false gradients. In [30], the authors developed MP-DQN (similar architecture as P-DQN) and tested it in well-known Robot Soccer Goal and Half Field Offense games. They forwarded the continuous action parameter with a standard basis vector to the Q-network. Consequently, it mitigated the effects of false gradients of P-DQN, converged the complex optimization problem, and outperformed P-DQN regarding data efficiency and converged policy performance [28,30]. After that, researchers are applying MP-DQN to solve hybrid action space-based optimization problems in various fields from robotics to communications. MP-DQN was implemented in the Golf simulator to find the best action that consisted of shot selection with height, spin, and speed [31]. In [32], the authors implemented MP-DQN to build a decision tree method for the imbalanced binary classification where the continuous attributes represented the discrete action, and the threshold values were continuous action. For intelligent traffic signal control development, MP-DQN was deployed in [33], considering selecting traffic lights (red, amber, and green) with on–off time intervals. In [34], the authors applied the MP-DQN in an actual robot setup to primitive action (translation, rotation, and insertion) with the end-effector velocity as well as the contact force limits. In wireless communication, MP-DQN has been implemented for task scheduling of the Radio Access Network [35] and joint task offloading and resource allocation in the Non-orthogonal multiple access (NoMA) system [36].
This paper explores MDRU aided two-tier HetNet scenario considering the UE mobility for post-disaster communication. It aims to optimize resource allocation by deploying the parameterized DRL called MP-DQN without knowing the environmental priori knowledge. The main contributions of this article are summarized below:
  • We investigate UE association and power allocation for maximizing the energy efficiency of downlink in MDRUs-aided two-tier HetNet for post-disaster communications by considering the backhaul links of MDRUs/SBSs with MSB where UE association as discrete action space and power allocation as continuous action space combinedly called parameterized action space of DRL has been considered when UEs are stationary.
  • Mobility-aware resource allocation (UE association and power allocation) has been formulated for parameterized DRL to optimize the energy efficiency, the throughput of SBSs/MDRUs, and the throughput of UEs in MDRUs aided two ties HetNet.
  • Model-free and parameterized action space-based MP-DQN algorithm, which utilizes several concurrent batch processing to provide action parameters to the Q network, has been proposed to maximize the energy efficiency, the throughput of SBSs/MDRUs and throughput of UEs of MDRUs aided HetNet.
Note that the proposed framework improves network robustness, which is one of the goals of the ASEAN IVO project titled “Context-Aware Disaster Mitigation using Mobile Edge Computing and Wireless Mesh Network”.

2. Related Work

TU, Japan, and NTT are conducting research continuously to improve the MDRU performance in terms of connectivity, serviceability, and coordination during PDM. They deployed the channel allocation algorithm in MDRU and conducted the test successfully in the Philippines and Nepal [37]. Due to the limited power after the disaster, emitting power optimization of rapidly deployable resource units by satisfying the demand of UEs (search and rescue robots, drones, smartphones, etc.) has paid great attention [38]. In [39], authors proposed radio access control based on DRL for selecting the van type MDRUs/relay and optimizing the power of MDRUs. In [40], authors investigated spectrum and energy-efficient methods for providing communication services to UE of MDRU-based networks. The authors [41] analyzed the problem of UAV deployment as MDRU in both standalone deployment scenarios to support fixed SBSs already in place where SBSs are damaged due to malfunction or disaster in HetNeT. In addition, they considered that UAVs were connected with remaining SBS or MBS by wireless backhauls, which was essential to serve the UEs by fulfilling their demands.
In [24], we implemented DRL, consisting of two value-based networks for energy-efficient radio resource allocation in IoRT that outperformed the DQN [42], where the UE demand and status (active or sleep) of BS are considered as state and action is to estimate the status of each BS. In addition, emitting power of active BS to serve UEs was optimized by a convex optimizer. In [43], value-based distributed DRL has been proposed to find user association and resource allocation by ensuring UE QoS. After that, the simulation results were improved by implementing the D3QN consisting of DDQN and dueling architecture in [44], where the degree of satisfaction of UEs was state space, and the selection of BS and transmission channels combinedly were action space. However, emitting power of BS can be adequately optimized when emitting power is conserved as continuous action of the DRL algorithm. In [23], continuous action-based DRL algorithms, TD3, have been applied to estimate the optimal emitting power of BS in the IoRT network by considering the interfering multiple access channel (IMAC). In [45], the authors developed a novel DRL based on DDPG to optimize the joint issue user association and power allocation of BS in HetNet that achieved the load balance and improved the energy efficiency of the network. In [46], a transfer learning algorithm based on DDPG has been developed to optimize the power allocation and ensure user association in HetNet. However, user association is discrete and power allocation is continuous. Hence, to solve the joint optimization problem combined with user association and power allocation, a hybrid action space-based DRL algorithm is the most suitable. In [29], the author formulated the problem of combined user association and power allocation, where user association considers as the discrete action and power allocation is expressed as the continuous action. In addition, P-DQN has been implemented to maximize energy efficiency by satisfying the QoS of each UE. Simulation results of P-DQN outperformed compared to DQN in terms of overall efficiency by satisfying the QoS of stationary UEs.
For PDM, UEs need to move in the vicinity to collect the appropriate information about victims that movement directly affects communication channel quality and throughput. This critical phenomenon has not been taken into account by many academics. The UE mobility model in non-orthogonal multiple access (NOMA), where each UE moved from one point to another with varied directions and speeds, was taken into consideration by the authors [47]. Due to UE mobility, the authors [48] suggested a conventional dynamic power allocation (DPA) method that took the channel circumstances into account and asserted that UE mobility significantly influences NOMA’s performance, particularly for downlink throughput. The authors of [49] created a power control method for a wireless network where UE mobility causes changes in the communication channel. In [23], two widespread UE mobility models, (a) modified Gauss–Markov and (b) random walk, have been investigated to maximize the sum rate in dynamic power allocation problems where the TD3-based DRL algorithm has been implemented; however, user association was not considered. In [50], the authors implemented Genetic Algorithm (GA) to allocate the UEs worked to share the information after the disaster in overlapping areas among the appropriate MDRUs. The proposed GA algorithm outperformed greedy and random algorithms as well as the nearest MDRU in terms of latency and QoS. In order to maximize the expected achievable rate of UE in ultra-dense networks, the authors [51] developed a matching game algorithm, where mobility-aware user association was considered by minimizing the handovers number. The authors [52] deployed the DRL algorithm to estimate the transmit timing, routing as well as power allocation for UEs from MDRU deployed in disaster areas where UE mobility, channel states, and energy harvesting were considered.

3. System Model

In this section, we consider two-tier HetNet that consists of one MBS with M active SBSs and N deployed MDRUs (replacement of damaged SBSs due to disaster) where M = { 1 , 2 , , M } and N = { 1 , 2 , , N } are the sets of active SBSs and deployed MDRUs [17,20,50]. The total SCs for PDM is K = M + N where K = { 1 , 2 , , K } is the set of SCs that serve U UEs considering U = { 1 , 2 , , U } is the set of UEs. In addition, we assume two different bands that are 6 GHz and millimeter wave bands for MBS to SBSs/MDRUs (tier 1) and SCs to UEs (tier 2), respectively. As a result, interference between tiers is not available in this network. For tier 1 downlinks, the antenna array of MBS is larger than the total number of SCs. Furthermore, orthogonal frequency division multiple access (OFDMA) is deployed to communicate from SBSs/MDRUs to UEs where the total subchannels number is N s u b . To collect the data and survey the disaster-affected area, UEs need to move from one place to another. Hence, the mobility model of UE has to be considered for PDM. Modified Gauss–Markov is the well-known mobility model of UE, especially for robots and drones that are considered in our research.

3.1. Modified Gauss-Markov Mobility Model

The Modified Gauss–Markov (MGM) mobility model improves past approaches by including temporal dependence. Here, the speed and direction of a UE are updated in line with the recorded values of earlier time periods. The degree of randomness used in calculating these two figures can also be changed based on the features of the simulated wireless network. The MGM mobility paradigm is not stateless since the memory of past actions are retained. Nonetheless, the UE mobility continues to be distinct from that of other mobile terminals linked to the same network [47,53]. According to Figure 3, UE mobility makes possible u t h UE to move randomly with average velocity that is indicated as Δ α u ( t 1 , t ) and v u ( t 1 , t ) , u t h UE are x u ( t )   y u ( t ) t are presented below:
x u ( t ) = x u ( t 1 ) + v u ( t 1 , t ) cos ( α u ( t 1 , t ) ) Δ t ,
y u ( t ) = y u ( t 1 ) + v u ( t 1 , t ) sing ( α u ( t 1 , t ) ) Δ t ,
α u ( t ) = α u ( t 1 ) + Δ α u ( t 1 , t ) ,
where x u ( t 1 ) , y u ( t 1 ) , and α u ( t 1 , t ) are the x-axis, y-axis, and direction of u t h UE at t 1 time slot. The distance traveled by u t h within Δ t can be illustrated by
d u ( t 1 , t ) = ( x u ( t 1 ) x u ( t ) ) 2 + ( y u ( t 1 ) y u ( t ) ) 2 .
The distance between k t h SC and u t h UE at t time slot is presented as
d k , u ( t ) = ( x k x u ( t ) ) 2 + ( y k y u ( t ) ) 2 ,
where x k and y k are the coordinates of k t h SC.

3.2. Network Model

Even though each SBS/MDRU are using OFDMA to serve UEs that construct a cluster of UEs, each UE can only be connected to a single SC. Let’s consider k t h serving cell serves to u t h UE by F r frequency subchannel. Here, c k u ( t ) = 0 , 1 is represented as the status of user association where c k u ( t ) = 1 denotes the u t h UE is associated with k t h SC and c k u ( t ) = 0 otherwise. After that, the set of UEs in the cluster k is assumed by C k ( t ) = { u : c k u ( t ) = 1 ,   u ϵ U } . The SBS serves the u t h UE can be illustrated by S u ( t ) = { k : c k u ( t ) = 1 ,   m ϵ M } where | S u ( t ) | is one. The set of active SC is at t time slot K a c t i v e ( t ) = { k | | C k ( t ) | > 0 } . The channel gain between from k t h SC to u t h UE can be defined as
g k , u , f ( t ) = h k , u , f ( t ) 2 ,
where h k , u , f ( t ) is the channel coefficient when subchannel frequency is f . The signal to interference plus noise ratio (SINR) from k t h serving cell to u t h UE can be illustrated as follows:
S I N R u f ( t ) = k = 1 K c k , u ( t ) g k , u , f ( t ) p k , u , f ( t ) σ 2 + I u , f ( t ) ,
where p k , u , f ( t ) is the allocated power of k t h SC for u t h UE which must be satisfied the 0 u C k f F R p k , u , f ( t ) P S C k , m a x . P S C k , m a x is the maximum emitting power from k t h SC. The observed interference and noise power by u t h UE is I u , f ( t ) and σ 2 respectively. To ensure no intra-cluster interference in each cluster, we investigate the case in which each SC allots orthogonal subchannels to various UEs within its serving area. Every UE can receive a minimum of one subchannel to transmit the data for data transmission when the cluster size does not exceed the total number of sub-channels. When intra-cluster interference is absent, just inter-cluster interference makes up the interference term I u , f which may be represented as
I u , f ( t ) = w C S u f F u F w g k , u , f ( t ) p k , w , f ( t ) .
The spectral efficiency of the u t h UE is illustrated as
ρ u ( t ) = f F k log 2 ( 1 + S I N R u , f ( t ) ) .
The user sum-rate for the k t h SC is calculated as
ρ k S C ( t ) = u C k ρ u ( t ) = u = 1 U c k , u ( t ) ρ u ( t ) .
The summation of data transmission power and the operating power that is defined as the minimum power requirement to maintain the SC active is defined total consumed power of our network that can be represented as
P t o t a l ( t ) = | K a c t i v e ( t ) | . P o , S C + k K u C k f F k p k , u , f ( t ) .
where P o , S C is the operational power of SC. Detailed notation descriptions are summarized in Table 1.
We strive for a way that results in optimizing user association and emitting power allocation to maximize the energy efficiency expressed as the achievable sum rate per utilized power in our assumed network by considering the QoS guarantee, and wireless backhaul link capacity constraints without knowing the environmental priori knowledge. The problem can be formulated as
c k u ( t ) , p k , u , f ( t ) m a x t = 0 t = T 1 P t o t a l ( t ) u = 1 U ρ u ( t ) ,
Subject   to   C 1 : k c k , u ( t ) = 1 , c k , u ( t ) 0 , 1 , k K , u ϵ U ,
C 2 : 0 u C k f F R p k , u , f ( t ) P S C k , m a x , k K , u ϵ U ,
C 3 : ρ u ( t ) v u , u ϵ U ,
C 4 : C k ( t ) C k m a x , k M ,
C 5 : ρ k S C ( t ) D k S C , m M .
Each UE is presumed to be serviced by a single SC in C 1 in (12b), and the transmit power limit at the k t h SC is discussed in C 2 in (12c), where P S C k , m a x is the maximum power that is used at the k t h SC. C 3 in (12d), where υ u is the capacity threshold for u t h UE denotes the QoS requirement for each UE. The cluster size limitation in (12e) is C 4 , and the maximum number of users in k cluster is | C k | m a x . To prevent intra-cluster interference, this makes sure that UEs in the same cluster are given distinct subchannels. D k S C is the highest feasible downlink data rate for k t h SC, while C 5 in (12f) is the backhaul connection capacity restriction.
By identifying the best user associations as well as power distribution, which is often a difficult task with a variety of unknowns and hybrid unknown spaces (discrete clustering and continuous power) in the network, the technique in (12a) aims to maximize energy efficiency. Additionally, the optimization issue in (12a) involves a one-shot situation at a certain time instant that must be reassessed as the network advances until the following time instant. We are consequently driven to deploy MP-DQN approaches to address the issues.

4. Deep Reinforcement Learning for Parameterized Action Space

In this section, we illustrate the DRL which can handle the parameterized action space for identifying optimal user association (discrete action) as well as emitting power allocation (continuous action) of SC by satisfying the QoS. The parameterized action space [54] combined with discrete and continuous action space represented as A d and A j respectively is illustrated as A = { ( j , z j ) | z j A j   f o r   a l l   j A d } , where a ( t ) = ( j , z j ) is the hybrid action. A discrete action j has been chosen from the discrete action set A d = j 1 , j 2 , j 3 , , j J = { c k u ( t ) : c k u ( t ) = 0 , 1 , k K , u ϵ U } . The continuous action parameters for that specific discrete action j is z j = p U E ( t ) = [ p 1 U E ( t ) , p 2 U E ( t ) , , p u U E ( t ) ] , where p u U E ( t ) = p S u , u , f ( t ) f : f F u for downlink data transmission in all sub channels allocated to u t h UE. Furthermore, z j Z , where Z is the set of continuous actions considering all possible discrete action. According to [55], parameterized action MDP (PAMDP) is presented as < S , P , A , R , γ > . Here, S represents the state space, the Markov probability of transition is illustrated as P , the parameterized action space is denoted by A , the reward is defined R and the discount faction is γ [ 0 , 1 ] . At the t t h timeslot, the agent observes the state of environment s ( t ) S and chooses suitable parameterized action a ( t ) A based on its policy π . After applying the chosen parameterized action, the immediate reward r ( s ( t ) , a ( t ) ) is received with next state of environment s ( t + 1 ) ~ P ( s ( t + 1 ) | s ( t ) , a ( t ) ) .
To solve the non-convex, the NP-hard and joint optimization problem consists of selecting the user association and allocating the transmitted power of MDRU-aided two-tier HetNet discussed in Section 3 by parametrized DRL, state, action, reward, and experience replay are described below:
State: The data rate of each UE at t t h timeslot has been generated from SINR that is calculated considering the user association, emitting power allocation, channel gain, interference, and noise poser observed by UE in that specific time slot. Hence, the set of data rate for all UE has been assumed as the state at t t h timeslot for DRL agent.
s ( t ) = ρ 1 ( t ) , , ρ u ( t ) .
Action: In this optimization problem, discreate (identification of UE association) and continuous (emitting power for each UE from SBS) action spaces at t t h timeslot have been combinedly considered as follows:
a ( t ) = c U E ( t ) , p U E ( t ) ,
where c U E ( t ) = c k u ( t ) , k = 1 : M , u = 1 : U with c k u ( t ) = 0 , 1 , k K , u ϵ U is denoted for UE association with SC. When c k u ( t ) = 1 , it means u t h UE is associated with k t h SC for that specific time slot and otherwise c k u ( t ) = 0 . After ensuring the UE association, SC is allocation power to that UE at t t h timeslot. The vector of power allocation from SCs at t t h timeslot is defined as p U E ( t ) = [ p 1 U E ( t ) , p 2 U E ( t ) , , p u U E ( t ) ] .
Reward: The maximization of the overall energy efficiency according to the Equation (12a) is the prime goal of this research by satisfying the QoS of every UE and the constraint capacity of backhaul link of each SBS. Therefore, the reward r ( t ) at t t h time slot is illustrated as:
(a)
Reward function one (RFO) [29]:
r ( s ( t ) , a ( t ) ) = r ( s ( t ) , a ( t ) ) i f   ρ k S C R k S C , k K r ( s ( t ) , a ( t ) ) r t h i f   ρ k S C > R k S C   f o r   s o m e   k K ,
where r ( s ( t ) , a ( t ) ) λ 1 Z α 1 ( t ) λ 2 Z α 2 ( t ) with α 1 ( t ) = 1 P T u = 1 U ρ u ( t ) that is the energy efficiency of system and α 2 ( t ) = u = 1 U ( ρ u ( t ) ν u ) 2 which is the penalty term is deployed for discouraging the agent to take the actions, for example, the capacity of every UE huge diverges from the threshold of QoS and Z α 1 ( t ) and Z α 2 ( t ) are the Z-scores of α 1 ( t ) and α 2 ( t ) , respectively. In addition, r t h is the threshold value is deployed to mitigate the likelihood of violating the backhaul capacity constraint.
(b)
Reward function two (RFT):
r ( s t , a t ) = λ 1 Z α 1 ( t ) + λ 2 u = 1 U R Q o S _ U E ( t ) + λ 3 m = 1 M R Q O S _ b a c k h a u l l i n k ( t ) ,
α 1 ( t ) = 1 P T u = 1 U ρ u ( t ) , R Q O S U E ( t ) = i f   ρ u v u , r = 1   e l s e   r = 1 , R Q O S b a c k h a u l l i n k ( t ) = i f   ρ k S C D k S C , r = 1   e l s e   r = 1 .
Here, λ 1 , λ 2 , λ 3 are non-negative weights of the corresponding terms and range from 0 to 1.
Experience replay: It is a DRL strategy that utilizes replay memory to record the agent’s experiences at each time step in a data set that is pooled over several episodes. After that, a minibatch of experience is selected randomly from the experience replay that is utilized for training. This process solves the problem of autocorrelation leading to unstable training.
Furthermore, three well-known DRL including our proposed method called MP-DQN which can handle parameterized action space are discussed below.

4.1. Deep Q Netwrok

One of the most well-known DRL algorithms is DQN [56], which is value-based and utilized for discrete action space only. The goal of traditional DQN is to find optimized the action by maximizing the action value function Q ( s , a ) as follows:
Q ( s , a ) E k = 0 γ k r ( s ( t + k ) , a ( t + k ) ) | s ( t ) = s , a ( t ) = a .
The maximization of (17) is equivalent to the Bellman equation and can be described as
y ( t ) = r ( t ) + γ m a x a ( t + 1 ) Q ( s ( t + 1 ) , a ( t + 1 ) ; θ ) ,
where y ( t ) represents the optimized value of Q. The loss function is represented as
L = ( y ( t ) Q ( s ( t ) , a ( t ) ; θ ) ) 2 ,
which mitigates the correlation between current value Q ( s ( t ) , a ( t ) ; θ ) and target value y ( t ) . In addition, the traditional DQN can be deployed for continuous action space when it is converted into a finite set of discrete action spaces by discretizing the process. Furthermore, DQN can also be utilized for parameterized action space by converting from continuous to discrete action space that concatenates with existing discrete action space. When continuous action has conducted the quantization to reverse discrete action, many action values are generated and those action values may round off. Consequently, the complexity of the DQN exponentially rises with the size of the action space, resulting in very massive power consumption and a delay in convergence speed. To overcome those issues, P-DQN has been deployed to handle the parameterized action space-based optimization problem [28].

4.2. Paramataized Deep Q Learing

P-DQN [57] is a DRL algorithm that handles hybrid (discrete-continuous) action spaces combined without relaxation or approximation. The structure of P-DQN is similar to DDPG, which describes a deterministic function that takes the state as input and produces continuous parameters of each discrete action. After that, generated continuous action parameters are concatenated with the state, which is utilized as input to the Q network to generate the Q values. Finally, the optimal function chooses the best discrete action from generated Q values. Let’s consider one actor parameter network z j ( s ; θ ) with weight θ and one actor network Q ( s , j , z j ; w ) with weight w . Furthermore, the weights θ has been estimated by optimizing the expected function of the action-value that are described as E [ Q ( s , j , z j ( j ; θ ) ; w ) ] . And the weight w has been determined by optimizing the mean squared error E [ ( y ( t ) [ Q ( s ( t ) , a ( t ) ; w ) ] ) 2 ] , where a ( t ) = ( j , z j ) and the target value is described as
y ( t ) = r ( t ) + γ   max j A d Q ( s ( t + 1 ) , j , z j ( s ( t + 1 ) ; θ ) ; ω ) .
In addition, the loss function of the actor parameter and the actor network can be presented as follows:
L x ( θ ) = 1 N t = 1 N Q ( s ( t ) , j , z j ( s ( t ) ; θ ) ; ω ) ,
L Q ( ω ) = 1 N t = 1 N ( y ( t ) Q ( s ( t ) , j , z j ( s ( t ) ; θ ) ; ω ) ) 2 .
Furthermore, the weights θ and ω are updated according to
θ θ α a , p 𝛻 θ L x ( θ ) ,
ω ω α a 𝛻 ω L x ( ω ) ,
where α a , p and α a are the learning rate for the actor parameter and actor network.
Even if P-DQN can converge and the impact is excellent, there is still room for improvement in the theory behind discrete and continuous action selection. Updates to any action’s continuous action parameter will affect all actions’ Q values, not just the Q value linked to the action parameter [28,30].

4.3. Multi Pass Deep Q Learing

The issue of excessive parameterization of P-DQN is resolved by MP-DQN [30] by employing multiple concurrent batch processing to provide action parameters to the Q network. Without altering the P-DQN structure, MP-DQN isolates the continuous parameters and inputs each one into the Q network individually. They executed a forward pass once for each discreate action j where the state s and action parameter vector z e j are concatenated as input and E j represents the j dimensional standard basis vector. Hence, the joint parameter vector is represented as Z E j = ( 0 , , 0 , z j , 0 , , 0 ) where each z i , i j is set to zero. As a consequence, the impact of network weights is negated for unassociated action parameters z j from the input layer where all false gradients are set to zero. Furthermore, Q is only depended on associated z j where
Q ( s , j , Z E j ) Q ( s , j , z j ) .
To forecast all Q values, c forward passes are necessary as opposed to just one. To perform the multi pass, the capacity of parallel minibatch processing by PyTorch or Tensorflow library can be deployed. A multi-pass with j actions is processed in the same manner as a minibatch of size j :
( Q ( s , . , Z E 1 ; θ Q ) . . . Q ( s , . , Z E J ; θ Q ) ) = ( Q 11 Q 1 j Q j 1 Q J J ) ,
where the Q-value for action b produced on the a t h pass is Q a b . Furthermore, the diagonal elements Q a a is pivotal and deployed in the final output Q a Q a a as shown in Figure 4. According to [33], MP-DQN makes it easier to choose the best hybrid action by reducing the impact of a single discrete action on other continuous action parameters.

5. Performance Evaluations

In this section, we utilize TensorFlow 1.14.0 on Spyder IDE 3.3.6 in an 11th Gen inter-core i7, 16 GB RAM, and RTX 3060 laptop GPU to demonstrate the simulation scenario. In addition, a HetNet has been considered which consists of one MBS with 100 antennas and 20 beamforming groups, three SBSs connected with MBS through backhaul transmission model [58], and five UEs presented in Figure 5a. We consider the non-line-of-sight path-loss model for urban MBSs and SBSs [59] and slow Rayleigh fading channels h ~ C N ( 0 , 1 ) . We followed the same system configuration as [29] for ensuring a fair comparison, tabulated in Table 2. All simulation results have been standardized by using the Z-score.
We compare the proposed MP-DQN presented in Algorithm 1 for the optimization problem, where UE association and emitting power allocation of SBSs/MDRUs have been considered jointly as the hybrid action space with two DRL based algorithms P-DQN and DQN. For DQN, continuous action space is converted into discrete by quantization process with P S C m , m a x 10 L , where L is the discrete power levels ( L = 5 is considered in our simulation). In addition, the simulation results of the proposed method are compared to a well-known method called Nearest SBS/MDRU with Random Power. Each UE is connected to the nearby SC, which generates random power to serve every UE in its SC’s cluster by fulfilling the conditions (1) the total power for all UE must be less or equal to the maximum power and (2) the total sum rate cannot exceed backhaul capacity of each SBS/MDRU. Furthermore, we consider the size of replay memory is 20,000, mini batch is 128, and the discount factor is 0.95 for all DRL algorithms. The total episodes for MP-DQN and P-DQN are 2000 while each episode has 50 timesteps. However, 3200 episodes are considered to simulate DQN. It takes more episodes to converge in hybrid action space-based optimization problem. In addition, other hyperparameters for MP-DQN, P-DQN, and DQN are tabulated in Table 3.
Algorithm 1: Multi pass DQN (MP-DQN) Algorithm.
Input: Probability distribution ξ , mini batch size B , exploration parameter ε , learning rates { α a , α a , p } .
Initialization: actor weights ω , ω and actor parameter weights ( θ , θ )
For t = 1, 2, 3, T do
   Estimate the action parameters z j ( s ( t ) ; θ ( t ) ) by actor network
   Choose the action a ( t ) = ( j , z j ) based on the ε greedy policy:
       a ( t ) = r a d o m   s a m p l e   a c c o r d i n g   t o   p r o b a b i l i t y   d i s t r i b u t i o n   ξ , w i t h   ε ( j , z j ) : j = a r g m a x c A d Q ( s ( t ) , j , z e j ; ω ) , w i t h   ( 1 ε )
   Execute action a ( t ) , receive immediate reward r ( s ( t ) , a ( t ) ) and next state s ( t + 1 )
   Save the experience ( s ( t ) , a ( t ) , r ( t ) , s ( t + 1 ) ) into replay memory
   Select mini batch size B randomly from the replay memory
   Define the target y(t) by
        y ( t ) = r ( t ) + γ max j A d Q ( s ( t + 1 ) , j , z j ( s ( t + 1 ) ; θ ) ; ω )
   Select the diagonal element from ( Q 11 Q 1 c Q c 1 Q c c )
   Choose the best action j by argmax from diagonal elements
   Use the ( y ( t ) , s ( t ) , a ( t ) ) to estimate the gradients 𝛻 ω L x ( ω ) and 𝛻 θ L x ( θ )
Update the weights parameters ω , ω , θ , θ

5.1. Simulation Results for Stationary UEs

We illustrate the average normalized results versus step over 500 realizations of proposed MP-DQN, P-DQN, and DQN algorithms during the training session. Figure 5b presents the average normalized reward of proposed MP-DQN, P-DQN and traditional DQN. Due to the complexity of the traditional DQN for discretization issues and the size of action space, the average reward is not properly converged. In P-DQN, the results are perfectly converged and saturated after-time steps. The final value of the average normalized reward is around 0.91. In comparison, the results of the proposed MP-DQN have converged perfectly but are a saturated bit later than P-DQN. The saturated value of the proposed method is around 1.25, which is clearly best compared with P-DQN and DQN algorithms.
We compare the average normalized test results of our proposed MP-DQN method with P-DQN, DQN, and Nearest SBS+ Random Power in Figure 6a,b and Figure 7 by considering the total time steps with 100 realizations. In Figure 6a, the average standardized test reward has been shown for all methods where maximum average results (around 1.26) for all timesteps are generated by our proposed method MP-DQN. The second height test reward is produced from P-DQN while the nearest SBS with random power for the UE method gives the worst results. In addition, the average normalized energy efficiency for test sessions has been depicted in Figure 6b for all discussed methods. The energy efficiency of our proposed method is approximately 9.89%, 94.7%, and 160.44% better than P-DQN, traditional DQN, and distance-based association methods, respectively during the whole test period. In addition, the average UE throughput by the normalized process has been illustrated in Figure 7 for all methods. According to Figure 7, the average normalized system throughput of our proposed method MP-DQN is approximately 4.27 which is 12.36%, 44.74%, and 19.607% better results compared to P-DQN, DQN, and the nearest distance with random power allocation algorithms. The summary of test results for all methods including the proposed method is presented in Table 4.

5.2. Simulation Results Considering UE’s Mobility

Due to its ability to remember previous activities, the GM mobility model is not stateless. It is appropriate for moveable UEs such as robots, cars, UGV, etc. We have illustrated average standardized reward, emergency efficiency, UE throughput, and SBS/MDRU throughput in Figure 8, Figure 9, Figure 10 and Figure 11, respectively. In addition, simulation results based on RFO and RFT have been illustrated in Figure Xa and Figure Xb, respectively, where X is within 8 to 11. Each UE’s average velocity has been considered 10 km/h with a random direction. In Figure 8a, the average standardized rewards mean is 0.6, 0.98, 1.48, and 1.61 from Nearest SBS + random power, DQN, PDQN, and MPDQN, respectively, based on the RFO. When the simulation is run with the RFT, the MPDQN generates (mean) 1.12, while PDQN and DQN produce 1.08 and 0.74, respectively illustrated in Figure 8b. Average standardized emergency efficiencies for all algorithms have been illustrated in Figure 9a,b according to RFO and RFT, respectively. In Figure 9a, MPDQN gives 5.03 while PDQN, DQN, and Nearest SBS + random power produce 4.81, 3.52, and 3.13, respectively. Furthermore, the MPDQN and PDQN generate almost similar energy efficiency that is better than the DQN and Nearest SBS + random power illustrated in Figure 9b.
For evaluating the IoRT network, the QoS of UE is the crucial parameter that directly depends on the downlink throughput of UE in each time slot. In Figure 10, we have depicted the average standardized UE throughput. When we have utilized the RFO, the means of average standardized UE throughputs are 2.92, 2.85, and 2.97 from MPDQN, PDQN, and DQN, respectively. However, the Nearest SBS + Random power generates 3.08, as shown in Figure 10a. We have illustrated the simulation results using the RFT in Figure 10b. The mean of average standardized UE throughput is 3.05 (similar to Figure 10a) by Nearest SBS + Random power, while DRL-based algorithms generate better results. Hence, the design of an appropriate reward function is the key factor in DRL-based problem formulation. The proposed method (MPDQN) gives 3.91, which is the best UE throughput compared to PDQN (3.44) and DQN (3.40). Another key factor of two-tier HeNet is the backhaul connection from MBS to SBS/MDRU, which depends on the throughput of SBS/MDRU illustrated in Figure 11. The proposed method with the RFT outperforms others that are clearly shown in Figure 11.
In Figure 12, the mean of average standardized UE throughput has been presented concerning the velocity range from 10 Km/h to 60 Km/h. With RFO, the average standardized UE throughput means approximately 3.10, 2.95, 2.92, and 2.84 from Nearest SBS + random power, DQN, MPDQN, and PDQN, respectively for all velocities until 60 Km/hour. In contrast, those are around 3.10, 3.40, 3.44, and 3.93 for Nearest SBS + random power, DQN, PDQN, and MPDQN, respectively, when adopting the RFT. The results are varied for the Nearest SBS + random power method due to the random power allocations. In our simulations, the discrete action (user association) selects the SC, and the continuous action allocates the power from SBS according to user association in every time step. As a result, the increment of velocity does not impact the simulation results. It is shown that the proposed method with the proposed reward function RFT gives a better result compared to others (Nearest SBS + random power, DQN, and PDQN).
The proposed reward function RFT consists of three main factors of a two-tier IoRT network (i) the energy efficiency, (ii) the QoS of UE, and (iii) the QoS of SBS/MDRU, while the original reward function RFO mainly depends on the average standardized energy efficiency and throughput of UE. As a result, DRL algorithms with the proposed reward function produce better results in contrast to DRL algorithms with the original reward function. The proposed method (MP-DQN) performs better than other algorithms due to the solution of excessive parameterization of P-DQN. In summary, MP-DQN with the proposed reward function RFT outperforms PDQN, DQN, and Nearest SBS + random power in reward, average energy efficiency, average system throughput, and average SBS throughput for various velocities of UE.

6. Conclusions

For the PDM, authorities deploy various UE such as UGVs, UAVs, UUVs, health care robots, and smartphones via IoRT to collect information in affected areas, where wireless network, especially 4G/LTE/5G and beyond, works as a backbone. Few SBSs of HetNet can be damaged due to the disaster. Hence, the deployment of MDRU to replace malfunctioning SBS is well-established nowadays. In addition, the electric power crisis is a big challenge for PDM. Therefore, power optimization of HetNet by satisfying all UE demands has paid great attention to research. In this article, we have examined UE association and power allocation of SBS/MDRU to optimize the energy efficiency, UE throughput, and SC throughput of the downlink without knowing the environmental priori knowledge while taking into account the backhaul link and QoS guarantee for stationary and movable UE in MDRU aided two-tier HetNet, which are nonconvex, NP-hard, as well as a hybrid action space problem. We have proposed MP-DQN, which is model-free as well as hybrid action space-based DRL algorithm. The simulation results of the proposed method (MP-DQN) have been compared with two DRL-based algorithms (P-DQN and DQN) and the nearest distance-based SBS with random allocation power. During the whole test period considering the stationary UE, our suggested method’s energy efficiency was around 9.89%, 94.7%, and 160.44% better than P-DQN, standard DQN, and distance-based association approaches, respectively. When the problem formulation by considering the modified Gauss–Markov UE mobility model has been investigated, we have proposed a new reward function RFT that is dependent on (i) the average standardized energy efficiency, (ii) the QoS of UE, and (iii) the QoS of SC; however, the original reward function RFO consists of average standardized energy efficiency and throughput of UE. Hence, DRL algorithms with the RFT are superior outcomes to those with the RFO. At various velocities, MP-DQN with the RFT outperforms PDQN, DQN, and Closest SBS + random power regarding reward, average energy efficiency, average system throughput, and average SC throughput.

Author Contributions

Conceptualization, H.K.; Investigation, H.K., M.-L.T. and Y.C.C.; Methodology, H.K., M.-L.T. and Y.C.C.; Resources, C.-O.C. and Y.O.; Writing—original draft, H.K., M.-L.T. and Y.C.C.; Software, H.K.; Funding acquisition, M.-L.T.; Writing—review and editing, H.K., C.-O.C. and Y.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Universiti Tunku Abdul Rahman (UTAR), Malaysia, under UTAR Research Fund (UTARRF) (IPSR/RMC/UTARRF/2021C1/T05). The ASEAN IVO (http://www.nict.go.jp/en/asean_ivo/index.html) project, “Context-Aware Disaster Mitigation using Mobile Edge Computing and Wireless Mesh Network”, was also involved in the production of the contents of this work.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lawry, L.; Burkle, F.M. Measuring the true human cost of natural disasters. Disaster Med. Public Health Prep. 2008, 2, 208–210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Shimada, G. The impact of climate-change-related disasters on Africa’s economic growth, agriculture, and conflicts: Can hu-manitarian aid and food assistance offset the damage? Int. J. Environ. Res. Public Health 2022, 19, 467. [Google Scholar] [CrossRef]
  3. Kamegawa, T.; Akiyama, T.; Sakai, S.; Fujii, K.; Une, K.; Ou, E.; Matsumura, Y.; Kishutani, T.; Nose, E.; Yoshizaki, Y.; et al. Development of a separable search-and-rescue robot composed of a mobile robot and a snake robot. Adv. Robot. 2020, 34, 132–139. [Google Scholar] [CrossRef]
  4. Vera-Ortega, P.; Vázquez-Martín, R.; Fernandez-Lozano, J.J.; García-Cerezo, A.; Mandow, A. Enabling Remote Responder Bio-Signal Monitoring in a Cooperative Human–Robot Architecture for Search and Rescue. Sensors 2023, 23, 49. [Google Scholar] [CrossRef] [PubMed]
  5. Paravisi, M.; Santos, D.H.; Jorge, V.; Heck, G.; Gonçalves, L.M.; Amory, A. Unmanned Surface Vehicle Simulator with Realistic Environmental Disturbances. Sensors 2019, 19, 1068. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. AlAli, Z.T.; Alabady, S.A. A survey of disaster management and SAR operations using sensors and supporting techniques. Int. J. Disaster Risk Reduct. 2022, 82, 103295. [Google Scholar] [CrossRef]
  7. Lee, M.-F.R.; Chien, T.-W. Artificial intelligence and Internet of Things for robotic disaster response. In Proceedings of the 2020 International Conference on Advanced Robotics and Intelligent Systems (ARIS), Taipei, Taiwan, 19–21 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  8. Kamilaris, A.; Botteghi, N. The penetration of Internet of Things in robotics: Towards a web of robotic things. J. Ambient. Intell. Smart Environ. 2020, 12, 491–512. [Google Scholar] [CrossRef]
  9. Villa, D.; Song, X.; Heim, M.; Li, L. Internet of Robotic Things: Current Technologies, Applications, Challenges and Future Directions. arXiv 2021, arXiv:2101.06256. [Google Scholar]
  10. Ray, P.P. Internet of robotic things: Concept, technologies, and challenges. IEEE Access 2016, 4, 9489–9500. [Google Scholar] [CrossRef]
  11. Vermesan, O.; Bahr, R.; Ottella, M.; Serrano, M.; Karlsen, T.; Wahlstrøm, T.; Sand, H.E.; Ashwathnarayan, M.; Gamba, M.T. Internet of Robotic Things Intelligent Connectivity and Platforms. Front. Robot. AI 2020, 7, 104. [Google Scholar] [CrossRef]
  12. Rengaraju, P.; Sethuramalingam, K.; Lung, C.H. Providing internet access for post-disaster communications using balloon networks. In Proceedings of the 18th ACM Symposium on Performance Evaluation of Wireless Ad Hoc, Sensor & Ubiquitous Networks, Alicante, Spain, 22–26 November 2021; pp. 111–117. [Google Scholar]
  13. Panda, K.G.; Das, S.; Sen, D.; Arif, W. Design and Deployment of UAV-Aided Post-Disaster Emergency Network. IEEE Access 2019, 7, 102985–102999. [Google Scholar] [CrossRef]
  14. Sakano, T.; Fadlullah, Z.M.; Ngo, T.; Nishiyama, H.; Nakazawa, M.; Adachi, F.; Kato, N.; Takahara, A.; Kumagai, T.; Kasahara, H.; et al. Disaster-resilient networking: A new vision based on movable and deployable resource units. IEEE Netw. 2013, 27, 40–46. [Google Scholar] [CrossRef]
  15. Zeng, Y.; Zhang, R.; Lim, T.J. Wireless communications with unmanned aerial vehicles: Opportunities and challenges. IEEE Commun. Mag. 2016, 54, 36–42. [Google Scholar] [CrossRef] [Green Version]
  16. Sakano, T.; Kotabe, S.; Komukai, T.; Kumagai, T.; Shimizu, Y.; Takahara, A.; Ngo, T.; Md Fadlullah, Z.; Nishiyama, H.; Kato, N. Bringing movable and deployable networks to disaster areas: Development and field test of MDRU. IEEE Netw. 2016, 30, 86–91. [Google Scholar] [CrossRef]
  17. Matracia, M.; Saeed, N.; Kishk, M.A.; Alouini, M.-S. Post-Disaster Communications: Enabling Technologies, Architectures, and Open Challenges. IEEE Open J. Commun. Soc. 2022, 3, 1177–1205. [Google Scholar] [CrossRef]
  18. Wang, J.; Sato, K.; Guo, S.; Chen, W.; Wu, J. Big Data Processing With Minimal Delay and Guaranteed Data Resolution in Disaster Areas. IEEE Trans. Veh. Technol. 2018, 68, 3833–3842. [Google Scholar] [CrossRef]
  19. Porte, J.; Briones, A.; Maso, J.M.; Pares, C.; Zaballos, A.; Pijoan, J.L. Heterogeneous wireless IoT architecture for natural disaster monitorization. EURASIP J. Wirel. Commun. Netw. 2020, 2020, 184. [Google Scholar] [CrossRef]
  20. Wang, Y. Models and Algorithms for Efficient Data Processing in Fog Computing Supported Disaster Areas. Ph.D. Dissertation, University of Aizu, Aizuwakamatsu, Japan, 2019. [Google Scholar]
  21. Wang, X.; Jiang, F.; Zhong, L.; Ji, Y.; Yamada, S.; Takano, K.; Xue, G. Intelligent Post-Disaster Networking by Exploiting Crowd Big Data. IEEE Netw. 2020, 34, 49–55. [Google Scholar] [CrossRef]
  22. Xu, K.; Qu, Y.; Yang, K. A tutorial on the internet of things: From a heterogeneous network integration perspective. IEEE Netw. 2016, 30, 102–108. [Google Scholar] [CrossRef] [Green Version]
  23. Kabir, H.; Tham, M.-L.; Chang, Y.C. Twin Delayed DDPG based Dynamic Power Allocation for Mobility in IoRT. J. Commun. Softw. Syst. 2023, 19, 19–29. [Google Scholar] [CrossRef]
  24. Kabir, H.; Tham, M.-L.; Chang, Y.C. DRL based Energy-Efficient Radio Resource Allocation Algorithm in Internet of Robotic Things. In Proceedings of the 2022 IEEE Symposium on Wireless Technology & Applications (ISWTA), Kuala Lumpur, Malaysia, 17–18 August 2022; pp. 104–109. [Google Scholar]
  25. Nguyen, H.T.; Nguyen, M.T.; Do, H.T.; Hua, H.T.; Nguyen, C.V. DRL-based intelligent resource allo-cation for diverse QoS in 5G and toward 6G vehicular networks: A comprehensive survey. Wirel. Commun. Mob. Comput. 2021, 2021, 5051328. [Google Scholar] [CrossRef]
  26. Abbasi, M.; Shahraki, A.; Piran, J.; Taherkordi, A. Deep Reinforcement Learning for QoS provisioning at the MAC layer: A Survey. Eng. Appl. Artif. Intell. 2021, 102, 104234. [Google Scholar] [CrossRef]
  27. Xiong, Z.; Zhang, Y.; Niyato, D.; Deng, R.; Wang, P.; Wang, L.-C. Deep reinforcement learning for mobile 5G and beyond: Fundamentals, applications, and challenges. IEEE Veh. Technol. Mag. 2019, 14, 44–52. [Google Scholar] [CrossRef]
  28. Zhu, J.; Wu, F.; Zhao, J. An Overview of the Action Space for Deep Reinforcement Learning. In Proceedings of the 2021 4th International Conference on Algorithms, Computing and Artificial Intelligence, Sanya, China, 22–24 December 2021; pp. 1–10. [Google Scholar] [CrossRef]
  29. Hsieh, C.-K.; Chan, K.-L.; Chien, F.-T. Energy-Efficient Power Allocation and User Association in Heterogeneous Networks with Deep Reinforcement Learning. Appl. Sci. 2021, 11, 4135. [Google Scholar] [CrossRef]
  30. Bester, C.J.; James, S.D.; Konidaris, G.D. Multi-pass Q-networks for deep reinforcement learning with parameterised action spaces. arXiv 2019, arXiv:1905.04388. [Google Scholar]
  31. Omstedt, F. A deep reinforcement learning approach to the problem of golf using an agent limited by human data. In Degree Project in Computer Science and Engineering; Kth Royal Institute of Technology: Stockholm, Sweden, 2020. [Google Scholar]
  32. Wen, G.; Wu, K. Building decision tree for imbalanced classification via deep reinforcement learning. Proc. Mach. Learn. Res. 2021, 157, 1645–1659. [Google Scholar]
  33. Bouktif, S.; Cheniki, A.; Ouni, A. Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning. Sensors 2021, 21, 2302. [Google Scholar] [CrossRef]
  34. Zhang, X.; Jin, S.; Wang, C.; Zhu, X.; Tomizuka, M. Learning insertion primitives with discrete-continuous hybrid action space for robotic assembly tasks. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 9881–9887. [Google Scholar] [CrossRef]
  35. Yan, Y.; Du, K.; Wang, L.; Niu, H.; Wen, X. MP-DQN Based Task Scheduling for RAN QoS Fluctuation Minimizing in Public Clouds. In Proceedings of the 2022 IEEE International Conference on Communications Workshops, ICC Workshops, Seoul, Republic of Korea, 16–20 May 2022; pp. 878–884. [Google Scholar] [CrossRef]
  36. Guo, L.; Jia, J.; Chen, J.; Du, A.; Wang, X. Joint Task Offloading and Resource Allocation in STAR-RIS assisted NOMA System. In Proceedings of the 2022 IEEE 96th Vehicular Technology Conference, VTC2022-Fall, London, UK, 26–29 September 2022; pp. 1–5. [Google Scholar] [CrossRef]
  37. Shimizu, Y.; Suzuki, Y.; Sasazawa, R.; Kawamoto, Y.; Nishiyama, H.; Kato, N.; Yamamoto, A.; Kotabe, S. Development of Movable and Deployable ICT Resource Unit (MDRU) and its Overseas Activities. J. Disaster Res. 2019, 14, 363–374. [Google Scholar] [CrossRef]
  38. Khan, A.; Mukhtar, M.; Ullah, F.; Bilal, M.; Kwak, K.-S. EA-RDSP: Energy Aware Rapidly Deployable Wireless Ad hoc System for Post Disaster Management. Comput. Mater. Contin. 2021, 69, 1725–1746. [Google Scholar] [CrossRef]
  39. Zhou, H.; Wang, X.; Umehira, M.; Chen, X.; Wu, C.; Ji, Y. Deep reinforcement learning based access control for disaster response networks. In Proceedings of the GLOBECOM 2020—2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  40. Ngo, T.; Nishiyama, H.; Kato, N.; Sakano, T.; Takahara, A. A Spectrum- and Energy-Efficient Scheme for Improving the Utilization of MDRU-Based Disaster Resilient Networks. IEEE Trans. Veh. Technol. 2014, 63, 2027–2037. [Google Scholar] [CrossRef]
  41. Wang, H.; Zhao, H.; Wu, W.; Xiong, J.; Ma, D.; Wei, J. Deployment algorithms of flying base stations: 5G and beyond with UAVs. IEEE Internet Things J. 2019, 6, 10009–10027. [Google Scholar] [CrossRef]
  42. Xu, Z.; Wang, Y.; Tang, J.; Wang, J.; Gursoy, M.C. A deep reinforcement learning based framework for power-efficient resource allocation in cloud RANs. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–6. [Google Scholar]
  43. Zhao, N.; Liang, Y.-C.; Niyato, D.; Pei, Y.; Jiang, Y. Deep reinforcement learning for user association and resource allocation in heterogeneous networks. In Proceedings of the 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  44. Zhao, N.; Liang, Y.-C.; Niyato, D.; Pei, Y.; Wu, M.; Jiang, Y. Deep Reinforcement Learning for User Association and Resource Allocation in Heterogeneous Cellular Networks. IEEE Trans. Wirel. Commun. 2019, 18, 5141–5152. [Google Scholar] [CrossRef]
  45. Li, Z.; Wen, X.; Lu, Z.; Jing, W. A General DRL-based Optimization Framework of User Association and Power Control for HetNet. In Proceedings of the 2021 IEEE 32nd Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Helsinki, Finland, 13–16 September 2021; pp. 1141–1147. [Google Scholar]
  46. Li, Z.; Wen, X.; Lu, Z.; Jing, W. A DDPG-based Transfer Learning Optimization Framework form User Association and Power Control in HetNet. In Proceedings of the 2022 IEEE International Conference on Communications Workshops (ICC 750 Workshops), Seoul, Republic of Korea, 16–20 May 2022; pp. 343–348. [Google Scholar]
  47. Narottama, B.; Shin, S.Y. Dynamic power allocation for non-orthogonal multiple access with user mobility. In Proceedings of the 2019 IEEE 10th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON), Vancouver, BC, Canada, 17–19 October 2019; pp. 0442–0446. [Google Scholar] [CrossRef]
  48. Masaracchia, A.; Nguyen, V.-L.; Nguyen, M. The impact of user mobility into non-orthogonal multiple access (NOMA) transmission systems. EAI Endorsed Trans. Ind. Netw. Intell. Syst. 2020, 7, e5. [Google Scholar] [CrossRef]
  49. Neely, M.J.; Modiano, E.; Rohrs, C.E. Dynamic power allocation and routing for time varying wireless networks. In Proceedings of the IEEE INFOCOM 2003. Twenty-Second Annual Joint Conference of the IEEE Computer and Communications Societies 756 (IEEE Cat. No. 03CH37428), San Francisco, CA, USA, 30 March–3 April 2003; Volume 1, pp. 745–755. [Google Scholar]
  50. Wang, Y.; Meyer, M.C.; Wang, J. Base Station Allocation for Users with Overlapping Coverage in Wirelessly Networked Disaster Areas. In Proceedings of the 2019 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Fukuoka, Japan, 5–8 August 2019; pp. 954–960. [Google Scholar] [CrossRef]
  51. Zhang, B.; Liu, S.; Yu, J.-L.; Han, Z. A Learning Aided Long-Term User Association Scheme for Ultra-Dense Networks. IEEE Trans. Veh. Technol. 2021, 71, 820–830. [Google Scholar] [CrossRef]
  52. Zhou, H.; Wang, X.; Umehira, M.; Chen, X.; Wu, C.; Ji, Y. Wireless Access Control in Edge-Aided Disaster Response: A Deep Reinforcement Learning-Based Approach. IEEE Access 2021, 9, 46600–46611. [Google Scholar] [CrossRef]
  53. Bai, F.; Helmy, A. A survey of mobility models. In Wireless Ad hoc Networks; University of Southern California: Los Angeles, CA, USA, 2004; Volume 206, p. 147. [Google Scholar]
  54. Hausknecht, M.; Stone, P. Deep reinforcement learning in parameterized action space. arXiv 2015, arXiv:1511.04143. [Google Scholar]
  55. Masson, W.; Ranchod, P.; Konidaris, G. Reinforcement learning with parameterized actions. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. No. 1. [Google Scholar]
  56. François-Lavet, V.; Henderson, P.; Islam, R.; Bellemare, M.G.; Pineau, J. An introduction to deep renforcement learning. Found. Trends® Mach. Learn. 2018, 11, 219–354. [Google Scholar] [CrossRef] [Green Version]
  57. Xiong, J.; Wang, Q.; Yang, Z.; Sun, P.; Han, L.; Zheng, Y.; Fu, H.; Zhang, T.; Liu, J.; Liu, H. Parametrized deep Q-networks learning: Reinforcement learning with discrete-continuous hybrid action space. arXiv 2018, arXiv:1810.06394. [Google Scholar]
  58. Wang, N.; Hossain, E.; Bhargava, V.K. Joint Downlink Cell Association and Bandwidth Allocation for Wireless Backhauling in Two-Tier HetNets with Large-Scale Antenna Arrays. IEEE Trans. Wirel. Commun. 2016, 15, 3251–3268. [Google Scholar] [CrossRef] [Green Version]
  59. 3rd Generation Partnership Project (3GPP). Further Advancements for E-UTRA Physical Layer Aspects (Release 9); 3rd Generation Partnership Project (3GPP): Sofia Technology Park, France, 2016. [Google Scholar]
Figure 1. MDRU aided wireless communication after disaster.
Figure 1. MDRU aided wireless communication after disaster.
Sensors 23 06448 g001
Figure 2. Architecture of deep reinforcement learning (DRL).
Figure 2. Architecture of deep reinforcement learning (DRL).
Sensors 23 06448 g002
Figure 3. UE Mobility model for UE with random direction and average velocity.
Figure 3. UE Mobility model for UE with random direction and average velocity.
Sensors 23 06448 g003
Figure 4. Architecture of multi pass deep Q learning (MP-DQN).
Figure 4. Architecture of multi pass deep Q learning (MP-DQN).
Sensors 23 06448 g004
Figure 5. (a) Network geometry of HetNet based IoRT network and (b) Average reward (normalized) of proposed method (MP-DQN), P-DQN and DQN during the training session.
Figure 5. (a) Network geometry of HetNet based IoRT network and (b) Average reward (normalized) of proposed method (MP-DQN), P-DQN and DQN during the training session.
Sensors 23 06448 g005
Figure 6. (a) Average reward (normalized) and (b) Average energy efficiency (normalized) of proposed method (MP-DQN), P-DQN, and DQN during the test session.
Figure 6. (a) Average reward (normalized) and (b) Average energy efficiency (normalized) of proposed method (MP-DQN), P-DQN, and DQN during the test session.
Sensors 23 06448 g006
Figure 7. The average normalized system throughput proposed method (MP-DQN), P-DQN, and DQN during the test session.
Figure 7. The average normalized system throughput proposed method (MP-DQN), P-DQN, and DQN during the test session.
Sensors 23 06448 g007
Figure 8. Average standardized reward based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Figure 8. Average standardized reward based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Sensors 23 06448 g008
Figure 9. Average standardized energy efficiency based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Figure 9. Average standardized energy efficiency based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Sensors 23 06448 g009
Figure 10. Average standardized UE throughput based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Figure 10. Average standardized UE throughput based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Sensors 23 06448 g010
Figure 11. Average standardized SBS throughput based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Figure 11. Average standardized SBS throughput based on (a) RFO and (b) RFT from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Sensors 23 06448 g011
Figure 12. The mean of average standardized UE throughput based with proposed and old reward function from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Figure 12. The mean of average standardized UE throughput based with proposed and old reward function from proposed method (MP-DQN), P-DQN, and DQN and Nearest SBS + random power.
Sensors 23 06448 g012
Table 1. Notation summary.
Table 1. Notation summary.
NotationDefinition
M , N , K   a n d   U set of SBSs, MRDUs, SCs and UEs
M, N, K and UTotal number of SBSs, MDRUs, SCs, and UEs
F w Set of subchannels allocated to u t h UE
S k ( t ) The SC serving the u t h UE at time slot t
B s u b Subchannel bandwidth
N T The number of antennas on MBS
P t o t a l ( t ) Total consumed power by active SCs
g k , u , f ( t ) The channel gain from k t h SC to u t h UE with f t h subchannel at time slot t
h k , u , f ( t ) The channel coefficient from k t h SC to u t h UE in f t h subchannel at time slot t
P S C k , m a x The maximum power available of k t h SC
p k , u , f ( t ) Emitting power from k t h SC to u t h UE in f t h subchannel at time slot t
| K a c t i v e ( t ) | Total quantity of active SCs at time slot t
σ 2 Noise power
I u , f ( t ) Interference observed by u t h UE in subchannel f at time slot t
C k ( t ) The set of UEs in cluster k at time slot t
c k u ( t ) Link indicator between k t h SC and u t h UE at time slot t
S I N R u f ( t ) SINR for u t h UE in the f t h subchannel at time slot t
υ u Capacity threshold for u t h UE
D k S C Maximum downlink data rate for k t h SC
Table 2. Simulation Parameters.
Table 2. Simulation Parameters.
ParameterValue
Carrier frequency2 GHz
Subchannel bandwidth15 kHz
Number of subchannels3
Number of subchannels per user1
MBS antenna array size100
MBS beamforming group size20
The radius of the entire network500 m
Number of SBS2
Number of MDRU1
Number of UE5
SINR threshold of UE1 for each UE
Rayleigh channel coefficient h ~ C N ( 0 , 1 )
Path loss model from MBS to SBS 19.77 + 3.91 × l o g 10 d k in dB and d k in km
Path loss model for SBS to UEs 30.53 + 36.71 × l o g 10 d k , u , t in dB and d k , u , t in km at time t
Noise power spectral density174 dBm/Hz
Maximum transmit power of SBS24 dBm
Maximum cluster size3
Transmit power of MBS43 dBm
Operational power of SBS0.5 W
Operational power of MBS130W
Table 3. The hyperparameter of MP-DQN.
Table 3. The hyperparameter of MP-DQN.
ParametersMP-DQN
Q Network
MP-DQN
Actor
P-DQN
Q Network
P-DQN
Actor
DQN
Learning rate10−410−510−510−510−3
Exploratione-greedyOrnstein-Uhlenbeck noisee-greedyOrnstein Uhlenbeck noisee-greedy
Number of Outputs | A d | U . | A d | | A d | U . | A d | | A d | * L U
Hidden layerReLu, 1024
ReLu, 512
ReLu, 256
ReLu, 1024
ReLu, 512
ReLu, 256
ReLu, 512
ReLu, 128

ReLu, 256
ReLu, 512
ReLu, 128
Relu 128Relu 128Relu 16 Relu 16
Number of Inputs U + U . | A d | U U + U . | A d | U U
Table 4. The average normalized value of implemented methods.
Table 4. The average normalized value of implemented methods.
MethodsAverage RewardAverage Energy EfficiencyAverage UE Throughput
MP-DQN (Proposed)1.266.834.27
P-DQN0.926.213.80
Nearest SBS+ Random Power−0.752.623.57
DQN0.1183.512.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kabir, H.; Tham, M.-L.; Chang, Y.C.; Chow, C.-O.; Owada, Y. Mobility-Aware Resource Allocation in IoRT Network for Post-Disaster Communications with Parameterized Reinforcement Learning. Sensors 2023, 23, 6448. https://doi.org/10.3390/s23146448

AMA Style

Kabir H, Tham M-L, Chang YC, Chow C-O, Owada Y. Mobility-Aware Resource Allocation in IoRT Network for Post-Disaster Communications with Parameterized Reinforcement Learning. Sensors. 2023; 23(14):6448. https://doi.org/10.3390/s23146448

Chicago/Turabian Style

Kabir, Homayun, Mau-Luen Tham, Yoong Choon Chang, Chee-Onn Chow, and Yasunori Owada. 2023. "Mobility-Aware Resource Allocation in IoRT Network for Post-Disaster Communications with Parameterized Reinforcement Learning" Sensors 23, no. 14: 6448. https://doi.org/10.3390/s23146448

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop