Next Article in Journal
An Automatic Classification System for Environmental Sound in Smart Cities
Previous Article in Journal
A Survey of Image-Based Fault Monitoring in Additive Manufacturing: Recent Developments and Future Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Optimization of Bandwidth and Power Allocation in Uplink Systems with Deep Reinforcement Learning

1
School of Information and Communication Engineering, Beijing University of Posts and Telecommunications (BUPT), Beijing 100876, China
2
School of Artificial Intelligence, Beijing University of Posts and Telecommunications (BUPT), Beijing 100876, China
3
Key Laboratory of Dynamic Cognitive System of Electromagnetic Spectrum Space, College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics (NUAA), Nanjing 211106, China
4
School of Cyberspace Science and Technology, Beijing Institute of Technology, Beijing 100081, China
5
Shaanxi Key Laboratory of Information Communication Network and Security, School of Communications and Information Engineering, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6822; https://doi.org/10.3390/s23156822
Submission received: 2 July 2023 / Revised: 28 July 2023 / Accepted: 28 July 2023 / Published: 31 July 2023
(This article belongs to the Topic Machine Learning in Communication Systems and Networks)

Abstract

:
Wireless resource utilizations are the focus of future communication, which are used constantly to alleviate the communication quality problem caused by the explosive interference with increasing users, especially the inter-cell interference in the multi-cell multi-user systems. To tackle this interference and improve the resource utilization rate, we proposed a joint-priority-based reinforcement learning (JPRL) approach to jointly optimize the bandwidth and transmit power allocation. This method aims to maximize the average throughput of the system while suppressing the co-channel interference and guaranteeing the quality of service (QoS) constraint. Specifically, we de-coupled the joint problem into two sub-problems, i.e., the bandwidth assignment and power allocation sub-problems. The multi-agent double deep Q network (MADDQN) was developed to solve the bandwidth allocation sub-problem for each user and the prioritized multi-agent deep deterministic policy gradient (P-MADDPG) algorithm by deploying a prioritized replay buffer that is designed to handle the transmit power allocation sub-problem. Numerical results show that the proposed JPRL method could accelerate model training and outperform the alternative methods in terms of throughput. For example, the average throughput was approximately 10.4–15.5% better than the homogeneous-learning-based benchmarks, and about 17.3% higher than the genetic algorithm.

1. Introduction

The fifth generation (5G) and beyond fifth generation (B5G) era is boosting a mega growth in the number of mobile devices [1], thereby resulting in explosive increasing demand that prompts people to explore new technologies to ease the demand strains. Recently, the large-scale dense network is gradually developing as a trend for the next-generation communication networks [2,3] due to its advantages traffic capacity and diversified services [4]. The densification of the network [5] is one of the key features of the 5G wireless network architecture, which not only contributes to increasing the system capacity of 5G networks, but also is closely related to user experience enhancement. As an important technique for improving the efficiency and quality of communications, dense networks still suffer from extremely complex interference problems [6]. In the dense multi-cell multi-user system, explosive rising users in different cells have an interplay due to the reuse of resources, which leads to increased co-channel interference and scarce resources. Furthermore, it is not conducive to deliver high throughput and a good quality of service (QoS) [7]. As a result, reasonable radio resource management [8] is imperative to improve the performance of future communications.
As pointed out in [9,10], whether resource allocation is rational or not determines the throughput performance of the system. Consider the multi-cell multi-user system where multiple resources (e.g., the bandwidth and transmitted power) are allocated to each user. As one user who interferes other users improving individual throughput causes serious interference, the coordination of resource allocation can avoid this situation efficiently. Therefore, bandwidth assignment and power allocation are the essential components of radio resource management, which can effectively suppress co-channel interference and conserve frequency resources. For the challenges of bandwidth and power allocations in the multi-cell multi-user network, a variety of methods have been proposed to increase throughput. Xu et al. [11] improved the throughput by selecting mobile relay and assigning subcarriers in the existence of various interferences. Liu et al. [12] increased the throughput by means of fast power allocation while guaranteeing stringent latency and reliability. The authors in [13] proposed a metaheuristic algorithm to solve the power control problem, which relied on discrete power allocation schemes. For the network cost problem of the large-scale heterogeneous system, Cao et al. [14] improved the network coverage using an adaptive seagull algorithm. In addition, various joint allocation methods have been proposed to maximize the rate, energy efficiency, and spectral efficiency [15,16,17]. The above-mentioned research works are based on traditional methods, such as the genetic algorithm [18], game theory [19], water-filling method [20], graph theory [21], and so on. These approaches are usually able to achieve the goals for different optimizations and application scenarios. Nevertheless, all of them experience dilemmas in exponentially growing the search space for the large-scale system, which are unsuitable for addressing high-dimensional joint optimization problems.
Reinforcement learning (RL) has been an efficient tool to solve optimization problems with a large number of data. It relies on uncharted exploitation with available samples for good reward feedback, which has been widely applied in large-scale scenarios [22,23]. Han et al. [24] proposed a State-Action-Reward-State-Action (SARSA) algorithm for power control to improve throughput. By taking advantage of machine learning, deep RL (DRL) is more effective for multi-user systems with large action spaces, which speeds the training process. The deep Q network (DQN) combines deep neural networks with Q-learning to approximate the value function with the help of maximizing the Q value [25], which has been deployed in many studies [26,27,28]. In [26], the authors developed a DQN-based method to allocate resource blocks in order to reduce the collision ratio and improve the throughput. Instead of directly using the maximum Q value, the double DQN (DDQN) selects the action by de-coupling the maximum Q value, which can avoid the overestimation of the Q value and speed up the convergence. Iqbal et al. [29] designed a DDQN method for power allocation to minimize the total power consumption. Nevertheless, many optimization variables, such as power allocation, are continuous in practice and are not applicable to the DQN and DDQN due to the discrete nature of actions. Furthermore, although the DQN and DDQN can transform continuous ranges into actions with different discrete granularities, they are impractical because of the limited granularity. For problems with infinite choices (e.g., power allocation), continuous action-selection-based algorithms such as the deep deterministic policy gradient (DDPG) [30] can overcome the disadvantages of discretization. Meng et al. [31] customized a DDPG to maximize the sum rate in a downlink cellular communication system. The authors in [32] optimized the long-term throughput using the adjusted DDPG extended from the DDPG, which is valid for two absolutely different action spaces. However, a centralized method such as the above works is feasible but inefficient and unsuitable for large-scale systems [33]. Multi-agent DRL (MADRL) is an advanced RL method that can outperform the single agent in resource allocation, especially in the multi-cell multi-user system [34,35]. In [36], a joint resource allocation problem is settled by a MADRL relying on the independent Q-learning method [37]. Similarly, Tian et al. [38] presented a DDPG-based MADRL method to allocate the channel and power by optimizing the QoS in vehicular networks.
Though MADRL contributes a great progress in the filed of joint resource allocation, it still continues to have the following limitations typically: (1) It generally ignoring the importance of the transition replay in sampling a mini-batch. In the traditional MADRL, since the complex communication environments usually contain a large amount of information, uniform experience replay leads to poor stability and the slow convergence of neural networks; (2) It weakens the interconnectivity between agents, especially in the system where the agent plays a direct role with the other agents (for example, an agent promotes individually and hinders others). Therefore, the traditional MADRL, which uses a distributed training process to explore solutions, is unsuitable for finding the action characteristics of each agent; and (3) It is not realistic to simplify the channel with a free-space propagation model, since some test scenarios are neglected in different channel models [39], including the urban macro-cell (UMa), rural macro0cell (RMa), and rural micro-cell (RMi) in IMT-2020.
Inspired by the success of DRL and the above research, the joint-priority-based RL (JPRL) method has been proposed to maximize the average throughput, which considers the co-channel interference between different cells. Unlike the traditional DRL algorithm that optimizes multiple variables, we selected different algorithms to optimize variables according to the problem property and deployed a distributed learning and centralized training framework. The main contributions of this paper are summarized as follows:
  • We proposed a joint bandwidth and power allocation framework based on the JPRL method to maximize the average throughput of the uplink large-scale system, which considered the co-channel interference between different cells with the assurance of the QoS. For the joint optimization problem, since the bandwidth assignment is a discrete problem, while the power allocation is continuous, we decomposed the joint problem into two sub-problems and used different algorithms to solve them.
  • We proposed a priority experience replay mechanism for power allocation. By analyzing the characteristics of the optimization sub-problems, the proposed experience replay mechanism was applied to a multi-agent DDPG (MADDPG), which was named the prioritized MADDPG (P-MADDPG), which trained valuable experiences to improve the throughput in the training process, thereby surpassing the issue of infinite power action space.
  • The proposed JPRL method is shown in Figure 1. It consists of a multi-agent DDQN (MADDQN) algorithm and the P-MADDPG algorithm, where MADDQN was developed to solve the bandwidth assignment sub-problem, and the P-MADDPG was employed to solve the transmit power allocation. Besides, both the MADDQN and P-MADDPG used a centralized training framework with a joint action-value function.
The remainder of this paper is organized as follows. Section 2 introduces the system model and optimization problem. The proposed JPRL method is described in Section 3. Section 4 demonstrates the simulation results, and the conclusions are presented in Section 5.

2. System Model and Problem Formulation

2.1. System Model

Consider a large-scale uplink multi-cell multi-user network, where M single-antenna base stations (BSs) are collected by the set M = { 1 , 2 , , M } are deployed at the center of M cells, respectively. Assume that there are N users collecting by the set L m = { l m , 1 , l m , 2 , , l m , N } in each cell m, where l m , n denotes the index of the n-th user in the m-th BS. The total users of the considered system are collected by the set K = { 1 , 2 , , K } , where K = M N . The total bandwidth of the considered system is denoted as W and is divided into three widths, which are collected by the set B = { B i } = { 15 kHz , 30 kHz , 60 kHz } , where i { 1 , 2 , 3 } [40]. Let X i = { 1 , 2 , , X i } denote the set of the sub-bands of the width i of the bandwidth, where X i is the total number of allocated bandwidth of width i.
Since users in different cells would occupy the same frequency band when transmitting their uplink signals, there exists interference between these users. This interference is called co-channel interference [41]. In this paper, each cell occupies the same frequency band and serves the same number of users N. For each user l m , n , some users in the neighboring cells can cause co-channel interference. In other words, users in the same cell can use different frequency band sub-carriers, and, thus, each user is subject to co-channel interference from users in other cells. Let M = { l m , n | m M , m m } denote the set of interfering users. Thus, these users from different cells belonging to the set M will interfere with user l m , n . The channel gain between user l m , n and BS m at the slot t is represented by the following: g ( d l m , n , m ) = h l m , n [ β ( d l m , n , m ) ] 1 2 , where β ( d l m , n , m ) = 10 P L l m , n + σ β z β 10 is the large scale fading corresponding to the distance d l m , n , m between user l m , n , BS m, P L l m , n is the path loss of user l m , n , σ β is the standard deviation of shadow fading, z β N ( 0 , 1 ) is a Gaussian random variable, and h l m , n CN ( 0 , σ h 2 ) is the small-scale fading with variance σ h 2 . Then, the power of co-channel interference on user l m , n is expressed as follows:
I l m , n = m M g d l m , n , m , t p l m , n ,
where p l m , n denotes the transmit power for user l m , n .
The signal y l m , n received by BS m from user l m , n can be written as
y l n , m = x l n , m + I l m , n + n l m , n ,
where x l m , n = b l m , n | g d l m , n , m , t | p l m , n denotes the transmitted signal by user l m , n , b l m , n is the transmitted symbol from user l m , n to BS m, and n 0 CN 0 , σ l m , n 2 is the additive white complex Gaussian noise. As a result, the received signal-to-interference-plus-noise ratio (SINR) at BS m of user l m , n is given by
ξ l m , n p l m , n , B i , l m , n = p l m , n g ( d l m , n , m , t ) σ l m , n 2 + I l m , n ,
where σ l m , n 2 = n f B i , l m , n indicates the variance of the Gaussian white noise, and n f is the power spectral density of noise. p l m , n is the power vector that includes the power of user l m , n and its interfering users, and B i , l m , n is the i-th width of the bandwidth allocated to the user l m , n . Then, by considering the normalized rate [42], the achievable throughput of user l m , n at BS m is
T H l m , n = log 2 1 + ξ l m , n p l m , n , B i , l m , n .

2.2. Problem Formulation

This paper mainly focuses on maximizing the average throughput of the considered large-scale multi-cell multi-user system subject to QoS of all users by jointly optimizing the transmit power and bandwidth allocation of all the users. Denote the average throughput of all the users by T H ¯ ; then, the joint resource allocation problem is formulated as follows:
P 1 : max p l m , n , B i , l m , n T H ¯ 1 K m = 1 M n = 1 N T H l m , n s . t . C 1 : P min p l m , n P max , l m , n L m , m M , C 2 : i = 1 3 B i X i W , C 3 : T H l m , n T H th , l m , n L m , m M ,
where P min and P max are the minimum and maximum transmit power of each user, respectively. Constraint C1 limits the transmit power budget per user; C2 indicates that the allocated bandwidth cannot exceed the total bandwidth of the system; and C3 ensures the QoS of each user. T H th denotes the required minimum throughput. Note that p l m , n and B i , l m , n are the decision variables associated with user l m , n , where p l m , n is the allocated power of the user l m , n , and B i , l m , n denotes the bandwidth assigned to the user l m , n of width i. This paper aims at obtaining better throughput by jointly optimizing the two variables.
Problem P1 is non-convex; it is difficult to solve using traditional methods due to the high computational complexity. Furthermore, owing to the intricacy of the co-channel interference relationship in large-scale systems and the interaction between users in different cells, it is challenging to find the effective solution for joint transmission power and bandwidth allocation directly. To tackle these challenges, we proposed the JPRL method, which is excellent for the multi-cell multi-user system. In the proposed method, the MADDQN algorithm was used to allocate the bandwidth, and the P-MADDPG algorithm was developed to optimize the transmit power.

3. JPRL-Based Joint Resource Allocation Approach

The detailed structure of the joint uplink bandwidth and transmit power allocation is shown in Figure 2. Joint resource allocation often optimizes multiple variables consistently. However, for the problem of the joint allocation of the bandwidth and transmit power, there exist infinite combinations of joint assignment schemes that are influenced by the users interactions, thereby leading to unfortunate performance. In addition, the bandwidth assignment with limited choices is a discrete assignment scheme, rather than the continuous range such as for the power allocation. Thus, we de-coupled problem P1 into two sub-problems and designed an efficient JPRL method to solve the joint resource allocation problem in the considered large-scale multi-cell multi-user system. Specifically, the MADDQN algorithm was developed to solve the bandwidth allocation sub-problem with a discrete action space, and the P-MADDPG algorithm was designed to solve the transmit power allocation subproblem in the continuous domain. This resource assignment procedure satisfies the decentralized partially observable Markov decision process. Therefore, the proposed JPRL based on the RL method employed each user as an agent to model the optimization, which could solve large-scale resource allocation while meeting QoS constraints.
The RL can be described as a stochastic game, which is defined by a tuple K , S , A , R , P , where K is the set of agents, and S and A denote the set of states and the joint actions the space of all agents, respectively. The R is the reward function, and P is the state transition probability. The game is generally concerned with the interaction between the environment and one or more agents in a series of iterations. In each iteration, the agent observes the environmental state S to take action from action space A . Thenm the agent receives an immediate reward R t to reflect the quality of this iteration and observes a new state to the next step. Our goal was to maximize of the long-term rewards over various iterations. The details of the proposed framework are illustrated as follows.
  • Agent: All users K.
  • State space: The state s k ( t ) of agent k is denoted as its co-channel interference, and the global environment state is thus defined as a set including the state of all agents, i.e.,
    S t = { s 1 ( t ) , , s k ( t ) , , s K ( t ) } , = { I l 1 , 1 t , , I l 1 , n t , , I l M , N t } .
  • Actions space: The actions of each agent consist of the bandwidth and power allocation and can be expressed as
    A t = { a 1 b t , a 1 p t , , a K b t , a K p t } ,
    where A t b = { a 1 b t , , a K b t } is defined as the bandwidth allocation, and A t p = { a 1 p t , , a K p t } is defined as the power allocation of all agents.
  • Reward function: Since the whole performance is influenced by all users in the considered system, the sparse reward is a serious issue. Inspired by the entire long-term evaluation mechanism, in the learning process, previous lessons are indicative of the current learning. Therefore, a novel reward function is defined as
    R t = T H ¯ t T H ˜ t , τ c ,
    where T H ¯ t denotes the average throughput of the current step t, τ denotes the moving step, and T H ˜ t , τ = 1 τ τ = 1 τ T H ¯ t τ + 1 is the moving average of T H ¯ t . c is a non-negative value. Especially, c = 0 if constraint C3 of Problem P1 is satisfied for all users; otherwise, c > 0 . Unlike the typical reward functions that evaluate the single-step target by setting a threshold, the proposed reward function employs a long short-term criterion that varies autonomously as the performance over time, which allows agents to perform more stable exploration in the multi-cell multi-user system.
In the proposed JPRL method, we developed a distributed learning and centralized training framework, as shown in Figure 3, which promised to explore the entire action space fully and encourage each agent to leverage the experience of other agents. Specifically, all agents are guided by the harmonized loss feedback value of the MADDQN and P-MADDPG when learning the bandwidth and power individually. The details of the proposed JPRL method are given as follows, its structure is illustrated in Algorithm 1, and the flow chart is shown in Figure 4. In the learning phase, the state of each agent is input into the the MADDQN and P-MADDPG algorithms synchronously, and then each agent individually performs the bandwidth allocation and power allocation actions. Based on the actions, rewards and new states are generated and stored in the replay buffers of the two algorithms. Note that the reward is calculated by Equation (8), which corresponds to all actions of the bandwidth and power. In the training phase, the values in the buffer are randomly selected to compute correlation values to guide the intelligence in the direction of increasing throughput. The details are described as follows.

3.1. Bandwidth Allocation of MADDQN

Bandwidth allocation is a non-convex problem with discrete space; there are finite choices. The size of the action space grows exponentially with the number of users. Therefore, a MADDQN algorithm with centralized training was presented to achieve sufficient exploration of the actions, which had good performance in large-scale discrete action spaces.
A MADDQN model consists of a target Q network and an evaluated Q network, which creates a copy of neural network for the two networks, respectively. For multiple agents, an arbitrary agent taking actions to improve its performance could lead to the degradation of the overall performance as the agents are interacting with each other. Therefore, the effect of mutual synergy between agents cannot be ignored. A centralized training architecture, to this end, denotes a joint action-state function Q sum b that composes the action-state functions from different agents to promote cooperation between agents. The concrete formula is defined as
Q sum b S t , A t b = k = 1 K Q k b ( s k ( t ) , a k b ( t ) ω ) ,
where ω is the parameter of the evaluated Q network. Q k b is the k-th user’s action-state function based on its own state. In the training phase, the joint action-state function is used for back propagation to promote cooperation, and a mini-batch sample is randomly sampled from the replay memory D 1 that stores the states, actions, next states, and rewards of all the agents (note that all the agents have the same reward value) to minimize the loss function, which is written as
L = E S t , A t b , S t + 1 , R t D 1 [ Y t Q sum b S t , A t b ω 2 ] ,
where E [ ] denotes the mathematical expectation and
Y t = R t + γ 1 argmax A Q sum b S t + 1 , A t + 1 b ω ;
γ 1 is the discount ratio. For each agent, the soft updating is given by
ω η ω + ( 1 η ) ω , η ( 0 , 1 ) ,
where ω are the parameters (including the weights and biases) of target Q network.
In the multi-cell multi-user system, the MADDQN model of agent k chooses the bandwidth assignment action according to its own state s k ( t ) in step t. Note that the agents can share their past training process (state, the influence based on training). Then, all the agents are centralized trained to minimize the loss value by Q sum b .

3.2. P-MADDPG-Based Uplink Power Allocation

For power allocation, a huge action space is not helpful for exploitation. In addition, although the discrete DRL algorithms can quantize power, they ignore the diversity of power choices. To this end, a novel P-MADDPG algorithm was proposed to solve the transmit power allocation subproblem. This is an enhancement of the DDPG with a prioritized replay buffer. In contrast to the power quantization, the P-MADDPG directly outputs the power of all the users in a continuous domain with infinite choices. Furthermore, by applying the prioritized replay buffer, it is more sensitive to the negative effect of the bad actions than the general MADDPG algorithms.
Similar to DDPG, an actor-critic architecture [43] applies for learning and training; both the actor and critic networks of each agent contain two identical neural networks, which are named the online network and target network, respectively. For a multi-agent system, the actor network of agent k outputs the power allocation under the current state through a policy π , i.e., a k p ( t ) = π ( s k ( t ) ) . However, the inherent exploration–exploitation dilemma in the DRL is prevalent for an inflexible action policy. By taking advantage of the DQN, it is balanced by a stochastic noise whose function is similar to the ϵ g r e e d y mechanism. Consequently, the actions of all agents are written as
A t p = [ π S t ω μ + Σ t ] P min P max ,
where ω u is the weight of the actor network, and Σ t follows a Gaussian distribution N ( 0 , ϱ ) ; ϱ is the variance of Gaussian noise and decreases linearly to zero as the iteration proceeds. Similarly, applying the individual action-value function to each agent ignores the features of others, which reduces the learning stability and weakens agent interaction. To this end, the critic network uses the joint action-value function Q sum p S t , A t to evaluate all actions. The specific Q sum p is defined as
Q sum p S t , A t = E R t , S t D 2 [ R t + γ 2 Q sum p S t + 1 , π ( S t + 1 ) ] ,
where D 2 is the experience replay buffer, and γ 2 ( 0 , 1 ] is a discount factor. According to the deterministic policy gradient theorem, the action-value function Q sum p is used to update the actor parameters ω μ in the direction of increasing the cumulative discounted reward with D samples of a mini-batch, that is
ω μ π E π [ ω μ Q s u m p ( S , A ω Q ) S = S t , A = π ( S t ω μ ) ] , = E π [ ω μ Q s u m p ( S , A ω Q ) S = S t , A = π ( S t ) ω μ π ( S ω μ ) S = S t ] , = 1 D k a k p ( t ) Q sum p S t , A t p ω Q | ω μ π s k ( t ) ω μ | s k ( t ) ,
where ω Q is the weight of critic network.
A common method for training neural networks is to randomly and uniformly sample mini-batches from the buffer D2, which often results in a high probability of selecting bad actions among the vast combinations of different actions, thereby lowering performance. This method is inefficient and poorly helpful for guiding the networks to update in the correct direction. Considering the transition samples of all agents, we designed the P-MADDPG algorithm to enhance the MADDPG by customizing a prioritized experience replay technique, where the more important transition samples have a higher probability of being replayed to participate in network updating. Specifically, in each step t, the transition samples of all agents are measured by the corresponding importance denoted by V t , which is combined with S t , A t p , R t , and S t + 1 to form a tuple S t , A t p , R t , S t + 1 , V t being stored in D 2 . Similar to the MADDQN, the goal of P-MADDPG updating is to minimize the magnitude between the joint Q-value and target joint Q-value, i.e., joint temporal-difference (JTD) error. The transitions with the large JTD error contain more information and are more necessary to the update of neural networks. Thus, the JTD error is a reasonable proxy measure of important value, and V t is written as
V t = | Y t Q sum p S t , A t p ω Q | .
However, in the sampling process, initially stored transition samples with small JTDs may not be sampled to replay if the sampling only relies on the importance. This can result in over-fitting, since the system lacks the sampling diversity of transitions. To avoid the issue, a probability associated with the importance is assigned to each transition sample, which can overcome the above issues effectively. The probability of the arbitrary transition sample φ at the step t is expressed as
P t φ = V t φ α φ = 1 | D 2 | ( V t φ ) α ,
where α [ 0 , 1 ] is a contribution factor that controls the impact of importance. In particular, α = 0 means that all samples are equally distributed, i.e., no contribution is made according to importance (uniform sampling). Original samples are equally probability-distributed in the replay buffer, but prioritized experience replay introduces bias, since it changes the original distribution by assigning different probabilities to the transitions. The compensation weight is thus introduced to correct this bias, which is expressed as
λ t φ = 1 D 1 P t φ β ,
where β [ 0 , 1 ] is a hyperparameter, which regulates the degree of bias compensation. In particular, there is no compensation for non-uniform probabilities P t φ if β = 0 ; there ispartial compensation if 0 < β < 1 ; and thre is full compensation if β = 1 . As a result, The loss of a mini-batch φ after weight compensation is rewritten as
L = E S t , A t p , S t + 1 , R t , V t φ D 2 λ t φ V t φ 2 .
Algorithm 1 JPRL methodfor joint bandwidth and power allocation
Initialize
  Initialize the network parameters in MADDQN and P-MADDPG respectively, ω , ω Q ;
  Initialize the replay buffer D 1 and the prioritized replay buffer D 2 , | D 1 | , | D 2 | ;
  Initialize a sum tree for D 2 , α , β .
Excute:
1:for episode i = 1 , , I  do
2:   Receive initial observation state of all agents K, and input s k ( t ) to agent k.
3:   Initialize the actions of all agents.
4:   for step t = 1 , , T i  do
5:     for agent k = 1 , , K  do
6:        if random number ζ < ϵ t  then
7:              Randomly choose a k b ( t ) from bandwidth allocation action space.
8:        else
9:                 a k b ( t ) = a r g m a x a k b ( t ) Q k b s k ( t ) , a k b ( t ) ω .
10:        end if
11:        Choose power allocation a k p ( t ) = [ π s k ( t ) + σ t ] P min P m a x .
12:        Execute actions a k b ( t ) , a k p ( t ) and observe next state s k ( t + 1 ) .
13:     end for
14:     Calculate reward with all agents’ actions by Equation (8).
15:     Store transition with bandwidth allocation S t , A t b , S t + 1 , R t in D 1 .
16:     Store transition φ with power allocation S t , A t p , S t + 1 , R t , V t φ in D 2 .
17:     if Both D 1 and D 2 are full then
18:        Sample a mini-batch of transition from D 1 .
19:        Sample a mini-batch of transition from D 2 according to sample importance.
20:        Compute the action-value function of MADDQN and P-MADDPQ according to Equations (9) and (14), respectively.
21:        Update evaluated Q network of MADDQN by Equation (10).
22:        Update actor online network by Equation (15).
23:        Update critic online network by Equation (19).
24:        Update the MADDQN and P-MADDPG networks by soft updating.
25:     end if
26:   end for
27:end for
For a mini-batch with D samples, directly traversing the experience buffer D 2 for sampling requires D times, and the complexity is intolerable. To tackle this matter, a sum-tree frame is designed for D 2 , where the sample φ is stored with the sampling probability P t φ . As shown in Figure 2, the structure is a binary tree with a root node at the top, and there are only two child nodes for each node of the upper level. For the leaf nodes at the bottom, the tuple S t , A t p , R t , S t + 1 , V t φ of transition φ is stored with its probability according to Equation (16). Note that the value of each node is the sum of its child nodes’ value. We divided the value of the root node (the sum of the probabilities of all samples) into D segments, which have an equal interval. In each interval, a random value, which is no more than the range of the interval generated to backtrack the leaf node from top to bottom. The specific backtracking rules are listed as follows, until the leaf node is selected, if the random value is less than or equal to the value of the left child node, and we continue backtracking from left child node; otherwise, we continue backtracking from the right child node and calculates the difference between this value and the value of the left child node as the basis for the next backtracking. Then, the critic and actor networks are updated by the selected transition samples. The proposed JPRL method is summarized in Algorithm 1.

3.3. Time Analysis of the Proposed JPRL Method

We analyzed the time complexity of our proposed JPRL method. In Algorithm 1, let I be the total number of training episodes and T i be the training steps in the episode i. Therefore, the total amount of training iterations implies the time complexity, that is O i = 1 I i T i . For each iteration, the computational efficiency is subjected to the the size of the neural network, i.e., the number of parameters. According to [44], the time complexity for a fully connected layer is O l = 1 L c l c l 1 , where l is the fully connected layer and c l denotes the number of neural units in layer l. In the JPRL method, each agent utilizes two algorithms to output the bandwidth and power actions. Note that the two algorithms are run simultaneously. Thus, the time complexity is c = O max l = 1 L { MADDQN , P MADDPG } c l c l 1 . The total time complexity of the JRPL method is O c i = 1 I i T i .

4. Simulations

In this section, we evaluate the performance of the proposed JPRL method. First of all, the simulation setup is portrayed. Then, the experience results are discussed in terms of the convergence, learning rate analysis, and performance comparison. Lastly, the performance of our proposed method compared to different models is exhibited.

4.1. Setup

Parameter Setting of Environment: We set the location of seven base stations in the cell center, and four users wererandomly distributed in each cell. The uplink user power was limited to P min = 40   dBm and P max = 23   dBm [40]. The total bandwidth of the system was W = 20   MHz . The minimum throughput requirement of all the users was T H th = 0.15   bit / s , and the power spectral density n f was 174   dBm / Hz .
The size of the cells and channel model change according to different scenarios [39], which are referenced from the test scenior in the 3GPP protocol, such as UMa, RMa, RMi. By default, the outsider scenario of the non-line-of-sight of the RMa was selected to evaluate the proposed method. The RMa stipulates the radius of cell r, and the pathloss is defined as
P L l m , n = max P L l m , n , 1 , P L l m , n , 2 ,
where P L l m , n , 1 and P L l m , n , 2 denote the line-of-sight and non-line-of-sight pathloss, respectively, which are written as
P L l m , n , 1 = P L l m , n , 11 , 10 m < d h < d BP , P L l m , n , 12 , d BP < d h < 5 km ,
where
P L l m , n , 11 = min 0.03 h b ε , 10 lg d s min 0.044 h b ε , 14.77 + 0.002 lg h b d s + 20 lg 40 π d s f c ,
P L l m , n , 12 = P L l m , n , 11 + 40 lg ( d s 2 π h a 1 h a 2 f c / v ) ,
and
P L l m , n , 2 = 161.4 7.11 lg ( l w ) + 7.5 lg ( h b ) 24.37 3.7 h b 2 h a 1 2 lg ( h a 1 ) + 43.42 3.1 lg ( h a 1 ) lg d s 3 + 20 lg ( f c ) 10.24 lg ( 11.75 h a 2 ) 2 + 4.97 .
Here, d s = d h 2 + h a 1 h a 2 2 and d h = d l m , n , m denote the straight distance and horizontal distance between BS and user respectively, where h a 1 and h a 2 are the heights of the antenna in the BS and user, respectively. h b is the building height, l w is the average width of the road, and ε is the excitation factor. For the long distance line-of-light pathloss P L l m , n , 12 , f c is the central frequency, and v denotes the propagation velocity. These parameter settings are listed in Table 1. In this paper, the five benchmarks were considered:
(1)
DDQN and DDPG: The existing DDQN for bandwidth assignment and the DDPG for allocating the power. The architecture with a one-layer fully connected network was used in the DDQN, and the DDPG deployed two-layer fully connected networks in the actor and critic networks. Both of them adopted the uniform sampling-based experience replay.
(2)
DDQN andP-DDPG: The settings were the same as (1), except that the DDPG used the prioritized experience replay.
(3)
MADDQN and MADDPG(ct): We treated each user as an agent and deployed the DDQN and DDPG on each agent. Centralized training was adopted.
(4)
MADDQN and MADDPG(dt): The MADDQN and MADDPG with distributed training. Note that each agents had the exclusive reward and loss.
(5)
Genetic algorithm (GA): The GA framework in the DEAP was used to realize this benchmark [45]. The bandwidth and power allocation schemes were encoded into the chromosome of each individual, which is the action sequence about the bandwidth and power allocation of all the users. We set the population size to 200. The crossover rate and mutation rate were set as 0.8 and 0.05, respectively.
Table 1. Environmental parameters.
Table 1. Environmental parameters.
ParametersValuesDescription
M7The number of cells
N4The number of users per cell
P min −40 dBm The minimum transmitting power
P max 23 dBm The maximum transmitting power
W2 GHz The total bandwidth
T H th 0.15 bit / s The minimum throughput constraint
n f 174   dBm / Hz The power spectral density of noise
r866 m The radius of cells
h b 10 m The average height of building
h a 1 10 m The antenna height of BS
h a 2 1.5 m The antenna height of user
ε 1.72Excitation factor
f c 1 GHz The center frequency
v 3 × 10 8 m / s The propagation velocity
l w 20 m The average road width
Note that the GA depends on the fitness rather than the learning-based reward; thus it is appropriate to compare the results after final convergence instead of comparing the entire optimization process with the learning-based approach.
Hyperparameter Setting of JPRL: The JPRL method contains an MADDQN algorithm and a P-MADDPG algorithm. There are the same size of the experience buffer for the two algorithms, which are set to | D 1 | = | D 2 | = 10000 . The learning rate, including the MADDQN, actor network, and critic network of the P-MADDPG was set as 0.0001. Furthermore, we set the hyperparameters of the prioritized replay buffer in the P-MADDPG D 2 as α = 1 and β = 0.1 . In the training phase, both the MADDQN and P-MADDPG used the Adam optimizer to optimize the loss function. The sampling batch size was D = 128 , and the reward discount factor was γ 1 = γ 2 = 0.89 . The system began to train the neural network when the memory buffer was full, and it updated the neural parameters at one-step frequency after training. Besides, we set the number of episodes to I = 500 . Note that every episode did not have fixed steps. To determine whether an episode was completed, a done flag was designed, where the done flag was true if the reward R increased by 200 steps; otherwise, it was false (the learning of this episode was not finished). the other parameters of each neural network are listed in Table 2.
All experiences were operated by a computer with the 12-th Gen Intel(R) Core(TM) i7-12700F @2.10 GHz, 16-GB RAM. The simulation results were presented using Numpy 1.21.5 and Tensorflow 2.3.0 on the Python 3.6 platform.

4.2. Results

The setting of the learning rate has a profound impact on the learning of the distribution scheme of the proposed method, which determines the ability to explore action space. Specifically, higher learning rates are detrimental to the exploration of the action space, as well as to the updating of network parameters in large systems with large action spaces. Moreover, in large systems with large action spaces, a lower learning rate implies finer-grained exploration, which does not mean that better actions can be explored, since having more actions in a large action space degrades performance. Thus, it is necessary to study the setting of the learning rate in a multi-cell multi-user system. Firstly, Figure 5 compares the loss values of the multiple networks under different learning rates. In order to view the variation and performance clearly, the loss values within the first 3000 steps after training are given. Figure 5a–c imply an interaction between the MADDQN and P-MADDPG in the proposed method. It is worth noting that the curve values in Figure 5b show a clear loss reduction in the P-MADDPG with a lower learning rate. It reveals the fact that the MADDQN exploring bandwidth influenced the P-MADDPG training. However, as shown in Figure 5c, a decrease in loss value did not signify an increase in throughput, and it may have also been trapped in sub-optimality. As a result, we set the learning rate of the MADDQN to 0.0001 to achieve a high throughput and fast convergence speed of the P-MADDPG.
Figure 6 illustrates the loss values and throughput of the actor and critic networks in the P-MADDPG at different learning rates. For the learning rates of the actor network, the proposed JPRL achieved the best in terms of the loss value and throughput whenthe learning rate was 0.0001. The loss curves of the MADDQN in Figure 6a show a slight increase after 1000 steps, and a similar trend appears in Figure 6d. The reason is that the power actions selected from the P-MADDPG affected the training process of the MADDQN. As shown in Figure 6b,c, it is noticed that, the smaller the learning rate, the better the performance, since the larger learning rate may skip various actions within the infinite action space. Finally, from Figure 6d–f, in the large action spaces, the critic network with a higher learning rate converged faster but converged to a worse value. The reason is that a larger learning rate of the critic network implies a more coarse-grained exploration, which is prone to learning sub-optimality. As a result, when the learning rate of the actor and critic networks were set to 0.0001, our method could jump out of the local optimal.
With respect to recording the reward for every 200 steps, Figure 7 plots the reward values of the proposed method and benchmarks; the benchmarks included the DDQN and DDPG, DDQN and P-DDPG, MADDQN and MADDPG based on the centralized training (ct) and decentralized training (dt). In the process of early random exploration (before the buffers are full), rewards decrease to negative values. The reason is that there are users whose throughput does not meet the QoS requirement. As the system begins to train, all five curves have a sharp augment. After a period of training, the moving average of the average throughput T H ˜ t , τ will be close to the average throughput T H ¯ t , e.g., the reward is close to 0, which indicates that the methods fall into a local optimal or converge to an optimal. It is seen that the curve of the MADDQN and MADDPG(dt) swung more than that of the MADDQN and MADDPG(ct). As a result, Figure 7 indicates that the JPRL method has an excellent ability to jump out of sub-optimal conditions and obtain good feedback.
Figure 8 illustrates the average throughput of the different methods after 500 episodes. In the random exploration stage, the throughput is unstable and relatively small because of the impacts of Gaussian noise and the randomly selected actions. All methods are prone to get stuck in the local optimum during the learning process, and there is a small fluctuation for the average throughput because of the existence of the Gaussian noise. Since a small change in power of any user may cause a large variation for co-channel interference, the benchmarks fall into the local optimum easily and are difficult to jump out of it. We can also see that the joint method MADDQN and MADDPG(dt) was extremely unstable, since the distributed training favored the individual performance of the agent at the expense of the overall performance. In other words, an agent, which follows its own wishes while neglecting the other characteristics for increasing power, will increase interference and decrease throughput. It was observed that the proposed JPRL outperformed the other methods in terms of throughput, since it explored the action spaces fully.
Figure 9 depicts a comparison of the average throughput for the six methods versus the cell number M. It should be observed that the average throughput decreased as the number of cells M increased. This is because fewer cells mean less interference from users l m , n , which leads to a lower amount of co-channel interference. Obviously, it can be seen that the RL-based approach was far superior to the GA, which is because the GA fell into the local optimum easily. We also see that the proposed JPRL had a steeper curve than the others, since it had better exploration in the small action spaces as cells decreased. Therefore, the JPRL method could achieve the high throughput.
As shown in Figure 10, we further tested the average throughput of the proposed JPRL under some different channel models, including the RMa, RMi, and UMa. The average throughput of the users for the urban environment (UMa model) is generally less than that of the users in rural scenarios (RMa and RMi models). This is because severe interference is caused by a lot of users in a small range. It can be seen that the JPRL method is universally applicable to different environments.

5. Conclusions

This paper mainly studied the resource allocation to maximize the throughput by jointly optimizing the bandwidth assignment and power allocation subject to the QoS constraint for the multi-cell multi-user uplink system. According to the variable attributes of the joint resource allocation problem, we proposed a JPRL method to decouple the optimization problem into two sub-problems, where the MADDQN was used to allocate bandwidth, and the P-MADDPG assigned uplink power with the given importance of transition. In order to compare the loss value and learning performance of the different networks with various learning rates, we set the appropriate parameters for the proposed JPRL method and analyzed the impact of the different learning rates. Furthermore, we evaluated the reward value and throughput of the proposed JPRL method against other existing methods. the simulation results showed that our approach can (1) obtain a better performance and be more applicable to the complex environments than other alternative methods (e.g., the average throughput was approximately 10.4–15.5% better than the average throughput of the benchmarks.) and (2) be universally applicable to other large-scale scenarios.
It is worth noting that, for simplicity, the single antenna system was used in this work. As for multi-antenna systems such as MIMO, the impact of more complex channel matrices caused by multiple antennas on user interference needs to be considered. In future work, the multiple antennas, the users’ trajectory, and cloud computing will be taken into consideration in multi-cell systems to facilitate communication–computing integration. By considering the interference corresponding to the complex channel matrix, the optimization is relevant to the compromised performance of the computing delay and energy consumption, which is based on the resource allocation and task offloading under various constraints, such as QoS constraints and offloading decisions. Moreover, multi-dimensional and deep analysis will be researched to validate the system tradeoff.

Author Contributions

Conceptualization, C.Z. and T.L.; methodology, P.H.; software, Z.L.; validation, C.Z., T.L. and P.H.; formal analysis, J.Z.; investigation, Y.R.; resources, T.L.; data curation, C.Z.; writing—original draft preparation, C.Z.; writing—review and editing, T.L.; visualization, C.Z.; supervision, T.L.; project administration, Z.L.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant number 62271068 and 61827801, the Beijing Natural Science Foundation under Grant number L222046, and the Basic Scientific Research Project under Grant NS2022046.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, G.; Cai, B.; Xie, W. Research on 5G Wireless Networks and Evolution. In Proceedings of the IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), Chengdu, China, 4–6 August 2021; pp. 1–5. [Google Scholar]
  2. Shah, A.F.M.S.; Qasim, A.N.; Karabulut, M.A.; Ilhan, H.; Islam, M.B. Survey and Performance Evaluation of Multiple Access Schemes for Next-Generation Wireless Communication Systems. IEEE Access 2021, 9, 113428–113442. [Google Scholar]
  3. Song, H.J.; Lee, N. Terahertz Communications: Challenges in the Next Decade. IEEE Trans. Terahertz. Sci. Technol. 2022, 12, 105–117. [Google Scholar]
  4. Vaezi, M.; Azari, A.; Khosravirad, S.R.; Shirvanimoghaddam, M.; Azari, M.M.; Chasaki, D.; Popovski, P. Cellular, Wide-Area, and Non-Terrestrial IoT: A Survey on 5G Advances and the Road Toward 6G. IEEE Commun. Surv. Tutor. 2022, 24, 1117–1174. [Google Scholar]
  5. Sharma, S.K.; Wang, X. Toward Massive Machine Type Communications in Ultra-Dense Cellular IoT Networks: Current Issues and Machine Learning-Assisted Solutions. IEEE Commun. Surv. Tutor. 2020, 22, 426–471. [Google Scholar]
  6. Mughees, A.; Tahir, M.; Sheikh, M.A.; Ahad, A. Energy-Efficient Ultra-Dense 5G Networks: Recent Advances, Taxonomy and Future Research Directions. IEEE Access 2021, 9, 147692–147716. [Google Scholar]
  7. Mardian, R.D.; Suryanegara, M.; Ramli, K. Measuring Quality of Service (QoS) and Quality of Experience (QoE) on 5G Technology: A Review. In Proceedings of the IEEE International Conference on Innovative Research and Development (ICIRD 2019), Jakarta, Indonesia, 28–30 June 2019; pp. 1–6. [Google Scholar]
  8. Xu, Y.; Gui, G.; Gacanin, H.; Adachi, F. A Survey on Resource Allocation for 5G Heterogeneous Networks: Current Research, Future Trends, and Challenges. IEEE Commun. Surv. Tutor. 2021, 23, 668–695. [Google Scholar]
  9. Lin, M.; Zhao, Y. Artificial intelligence-empowered resource management for future wireless communications: A survey. China Commun. 2020, 17, 58–77. [Google Scholar]
  10. Bossy, B.; Kryszkiewicz, P.; Bogucka, H. Energy-Efficient OFDM Radio Resource Allocation Optimization with Computational Awareness: A Survey. IEEE Access 2022, 10, 94100–94132. [Google Scholar]
  11. Xu, J.; Niu, H.; Zhao, T.; Gao, X.; Ye, J.; Liang, C.; Liu, Z.; Liang, J. Robust Optimal Power Control and Subcarrier Allocation in Uplink OFDMA Network With Assistance of Mobile Relay. IEEE Access 2021, 9, 57475–57485. [Google Scholar]
  12. Liu, R.; Yu, G.; Yuan, J.; Li, G.Y. Resource Management for Millimeter-Wave Ultra-Reliable and Low-Latency Communications. IEEE Trans. Commun. 2021, 69, 1094–1108. [Google Scholar]
  13. Sun, Q.; Wu, H.; Petrosian, O. Optimal Power Allocation Based on Metaheuristic Algorithms in Wireless Network. Mathematics 2022, 10, 3336. [Google Scholar]
  14. Cao, L.; Wang, Z.; Wang, Z.; Wang, X.; Yue, Y. An Energy-Saving and Efficient Deployment Strategy for Heterogeneous Wireless Sensor Networks Based on Improved Seagull Optimization Algorithm. Biomimetics 2023, 8, 231. [Google Scholar]
  15. Zeng, M.; Nguyen, N.P.; Dobre, O.A.; Ding, Z.; Poor, H.V. Spectral-and energy-efficient resource allocation for multi-carrier uplink NOMA systems. IEEE Trans. Veh. Technol. 2019, 68, 9293–9296. [Google Scholar]
  16. Sharif, Z.; Jung, L.T.; Razzak, I.; Alazab, M. Adaptive and Priority-Based Resource Allocation for Efficient Resources Utilization in Mobile-Edge Computing. IEEE Internet Things J. 2023, 10, 3079–3093. [Google Scholar]
  17. Xue, J.; An, Y. Joint Task Offloading and Resource Allocation for Multi-Task Multi-Server NOMA-MEC Networks. IEEE Access 2021, 9, 16152–16163. [Google Scholar]
  18. Brahmi, I.; Koubaa, H.; Zarai, F. Genetic Algorithm based Resource Allocation for V2X Communications. In Proceedings of the International Conference on Communications and Networking, ComNet, Hammamet, Tunisia, 27–30 October 2020; pp. 1–5. [Google Scholar]
  19. Dun, H.; Ye, F.; Jiao, S.; Li, Y.; Jiang, T. The Distributed Resource Allocation for D2D Communication with Game Theory. In Proceedings of the 2019 IEEE APS Topical Conference on Antennas and Propagation in Wireless Communications (APWC), Granada, Spain, 9–13 September 2019; pp. 104–108. [Google Scholar]
  20. Li, M.; Peng, T.; Wu, H. Power Allocation to Achieve Maximum Throughput in Multi-radio Multi-channel Mesh Network. In Proceedings of the IEEE 6th International Conference on Computer and Communications (ICCC), Chengdu, China, 11–14 December 2020; pp. 293–297. [Google Scholar]
  21. Wang, C.; Yan, F. Graph Theory based Resource Allocation Algorithm in Terahertz Communication Networks. In Proceedings of the 2021 IEEE International Conference on Information Networking, Jeju Island, Republic of Korea, 13–16 January 2021; pp. 304–308. [Google Scholar]
  22. Xiong, Z.; Zhang, Y.; Niyato, D.; Deng, R.; Wang, P.; Wang, L.C. Deep Reinforcement Learning for Mobile 5G and Beyond: Fundamentals, Applications, and Challenges. IEEE Trans. Veh. Technol. 2019, 14, 44–52. [Google Scholar]
  23. Du, Z.; Deng, Y.; Guo, W.; Nallanathan, A.; Wu, Q. Green Deep Reinforcement Learning for Radio Resource Management: Architecture, Algorithm Compression, and Challenges. IEEE Trans. Veh. Technol. 2021, 16, 29–39. [Google Scholar]
  24. Han, K.; Ye, C. Power Control Research for Device-to-Device Wireless Network Underlying Reinforcement Learning. In Proceedings of the Global Conference on Robotics, Artificial Intelligence and Information Technology (GCRAIT), Chicago, IL, USA, 30–31 July 2022; pp. 351–354. [Google Scholar]
  25. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar]
  26. Liu, J.; Ma, X.; Han, W.; Wang, L. Resource Allocation in OFDMA Networks with Deep Reinforcement Learning. In Proceedings of the IEEE 8th International Conference on Information, communication and networks (ICICN), Xi’an, China, 17–20 August 2020; pp. 111–117. [Google Scholar]
  27. Guan, X.; Lv, T.; Lin, Z.; Huang, P.; Zeng, J. D2D-Assisted Multi-User Cooperative Partial Offloading in MEC Based on Deep Reinforcement Learning. Sensors 2022, 22, 7004. [Google Scholar]
  28. Rahman, G.M.S.; Dang, T.; Ahmed, M. Deep reinforcement learning based computation offloading and resource allocation for low-latency fog radio access networks. Intell. Converg. Netw. 2020, 1, 243–257. [Google Scholar]
  29. Iqbal, A.; Tham, M.L.; Chang, Y.C. Double Deep Q-Network for Power Allocation in Cloud Radio Access Network. In Proceedings of the 2020 IEEE 3rd International Conference on Computer and Communication Engineering Technology (CCET), Beijing, China, 14–16 August 2020; pp. 272–277. [Google Scholar]
  30. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  31. Meng, F.; Chen, P.; Wu, L.; Cheng, J. Power Allocation in Multi-User Cellular Networks: Deep Reinforcement Learning Approaches. IEEE Trans. Wirel. Commun. 2020, 19, 6255–6267. [Google Scholar]
  32. Zheng, K.; Jia, X.; Chi, K.; Liu, X. DDPG-Based Joint Time and Energy Management in Ambient Backscatter-Assisted Hybrid Underlay CRNs. IEEE Trans. Comm. 2023, 71, 441–456. [Google Scholar]
  33. Yue, Y.; Cao, L.; Lu, D.; Hu, Z.; Xu, M.; Wang, S.; Li, B.; Ding, H. Review and empirical analysis of sparrow search algorithm. Artif. Intell. Rev. 2023, 1–53. [Google Scholar] [CrossRef]
  34. Zhang, K.; Yang, Z.; Başar, T. Multi-agent reinforcement learning: A selective overview of theories and algorithms. In Handbook of Reinforcement Learning and Control; Springer: Cham, Switzerland, 2021; pp. 321–384. [Google Scholar]
  35. Rosenberger, J.; Urlaub, M.; Schramm, D. Multi-agent reinforcement learning for intelligent resource allocation in IIoT networks. In Proceedings of the 2021 IEEE Global Conference Artificial Intelligence and Internet of Things (GCAIoT), Dubai, United Arab Emirates, 12–16 December 2021; pp. 118–119. [Google Scholar]
  36. Hu, J.; Wang, X.; Li, D.; Xu, Y. Multi-agent DRL-Based Resource Allocation in Downlink Multi-cell OFDMA System. In Proceedings of the 2020 International Conference on Wireless Communications and Signal Processing (IWCSP), Nanjing, China, 21–23 October 2020; pp. 257–262. [Google Scholar]
  37. Jiang, M.; Hai, T.; Pan, Z.; Wang, H.; Jia, Y.; Deng, C. Multi-Agent Deep Reinforcement Learning for Multi-Object Tracker. IEEE Access 2019, 7, 32400–32407. [Google Scholar]
  38. Tian, J.; Liu, Q.; Zhang, H.; Wu, D. Multiagent Deep-Reinforcement-Learning-Based Resource Allocation for Heterogeneous QoS Guarantees for Vehicular Networks. IEEE Internet Things J. 2022, 9, 1683–1695. [Google Scholar]
  39. Zhu, Q.; Wang, C.X.; Hua, B.; Mao, K.; Jiang, S.; Yao, M. 3GPP TR 38.901 channel model. In The Wiley 5G Ref: The Essential 5G Reference Online; Wiley Press: Hoboken, NJ, USA, 2021; pp. 1–35. [Google Scholar]
  40. Morais, D.H.; Morais, D.H. 5G NR Overview and Physical Layer. In Key 5G Physical Layer Technologies: Enabling Mobile and Fixed Wireless Access; Springer: Cham, Switzerland, 2022; pp. 233–297. [Google Scholar]
  41. Modak, K.; Rahman, S. Multi-cell Interference Management in In-band D2D Communication under LTE-A Network. In Proceedings of the 2021 International Conference on Computing, Electronics & Communications Engineering (ICCECE), Virtual, 16–17 August 2021; pp. 13–18. [Google Scholar]
  42. Jia, R.; Liu, L.; Zheng, X.; Yang, Y.; Wang, S.; Huang, P.; Lv, T. Multi-Agent Deep Reinforcement Learning for Uplink Power Control in Multi-Cell Systems. In Proceedings of the 2022 IEEE International Conference on Communications Workshops (ICC Workshops), Seoul, Republic of Korea, 16–20 May 2022; pp. 324–330. [Google Scholar]
  43. Haarnoja, T.; Zhou, A.; Hartikainen, K.; Tucker, G.; Ha, S.; Tan, J.; Kumar, V.; Zhu, H.; Gupta, A.; Abbeel, P.; et al. Soft actor-critic algorithms and applications. arXiv 2018, arXiv:1812.05905. [Google Scholar]
  44. Liu, Y.; Wang, H.; Peng, M.; Guan, J.; Wang, Y. An Incentive Mechanism for Privacy-Preserving Crowdsensing via Deep Reinforcement Learning. IEEE Internet Things J. 2021, 8, 8616–8631. [Google Scholar]
  45. Fortin, F.A.; De Rainville, F.M.; Gardner, M.A.G.; Parizeau, M.; Gagné, C. DEAP: Evolutionary algorithms made easy. J. Mach. Learn. Res. 2012, 13, 2171–2175. [Google Scholar]
Figure 1. Joint bandwidth and power allocations scheme.
Figure 1. Joint bandwidth and power allocations scheme.
Sensors 23 06822 g001
Figure 2. System model of the JPRL-based bandwidth and power allocations.
Figure 2. System model of the JPRL-based bandwidth and power allocations.
Sensors 23 06822 g002
Figure 3. Framework of centralized training.
Figure 3. Framework of centralized training.
Sensors 23 06822 g003
Figure 4. Flow chart of the JPRL method.
Figure 4. Flow chart of the JPRL method.
Sensors 23 06822 g004
Figure 5. The loss value and throughput under different learning rates of MADDQN. The learning rates of both actor networks and critic networks of P-MADDPG are set to 0.0001, and the loss value was extracted at 3000 steps after the beginning of network training. (a) MADDQN loss function value. (b) P-MADDPG loss function value. (c) Average throughput.
Figure 5. The loss value and throughput under different learning rates of MADDQN. The learning rates of both actor networks and critic networks of P-MADDPG are set to 0.0001, and the loss value was extracted at 3000 steps after the beginning of network training. (a) MADDQN loss function value. (b) P-MADDPG loss function value. (c) Average throughput.
Sensors 23 06822 g005
Figure 6. The loss function value and throughput of the two networks of P-MADDPG under different learning rates. (ac) show the effect of variable learning rates on actor network, and (df) are the exhibitions within the changing learning rates of critic network. Note that the learning rates of other networks were set as default when a network varied in learning rate.
Figure 6. The loss function value and throughput of the two networks of P-MADDPG under different learning rates. (ac) show the effect of variable learning rates on actor network, and (df) are the exhibitions within the changing learning rates of critic network. Note that the learning rates of other networks were set as default when a network varied in learning rate.
Sensors 23 06822 g006
Figure 7. Reward comparison of different methods as the number of steps increased.
Figure 7. Reward comparison of different methods as the number of steps increased.
Sensors 23 06822 g007
Figure 8. Average throughput comparison of different methods as the number of episodes increased.
Figure 8. Average throughput comparison of different methods as the number of episodes increased.
Sensors 23 06822 g008
Figure 9. Comparison of average throughput for the different methods versus the number of cells.
Figure 9. Comparison of average throughput for the different methods versus the number of cells.
Sensors 23 06822 g009
Figure 10. Comparison of average throughput for the different channel models versus the number of episodes.
Figure 10. Comparison of average throughput for the different channel models versus the number of episodes.
Sensors 23 06822 g010
Table 2. Network parameters.
Table 2. Network parameters.
NetworkNeural UnitsActivationOptimizer
MADDQN64sigmoidAdam Optimizer
Actor Network of P-MADDPG32, 16tanhAdam Optimizer
Critic Network of P-MADDPG32, 16ReLuAdam Optimizer
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, C.; Lv, T.; Huang, P.; Lin, Z.; Zeng, J.; Ren, Y. Joint Optimization of Bandwidth and Power Allocation in Uplink Systems with Deep Reinforcement Learning. Sensors 2023, 23, 6822. https://doi.org/10.3390/s23156822

AMA Style

Zhang C, Lv T, Huang P, Lin Z, Zeng J, Ren Y. Joint Optimization of Bandwidth and Power Allocation in Uplink Systems with Deep Reinforcement Learning. Sensors. 2023; 23(15):6822. https://doi.org/10.3390/s23156822

Chicago/Turabian Style

Zhang, Chongli, Tiejun Lv, Pingmu Huang, Zhipeng Lin, Jie Zeng, and Yuan Ren. 2023. "Joint Optimization of Bandwidth and Power Allocation in Uplink Systems with Deep Reinforcement Learning" Sensors 23, no. 15: 6822. https://doi.org/10.3390/s23156822

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop