1. Introduction
With the rapid development of the Internet of Things (IoT), where all the physical and smart things such as machines, vehicles, and devices are connected through wireless networks to exchange information about themselves and their surroundings, it is a challenge to fulfil user requirements with the required system performance. IoT can enhance our life, improve our daily experience, and cooperate to support various industrial applications by creating an efficient platform for connecting all devices [
1,
2]. The future of IoT is a bright one because a large amount of IoT devices can communicate and merge to enhance our daily life. For example, the nest smart home—this is an application provided for users on their phones to monitor their devices at their home. This application can provide the user with the optimum required security and convenience. Additionally, the IoT system supports traffic management and vehicle behaviors, and it also enables vehicle-to-vehicle and vehicle-to-person communication to improve transportation safety and traffic management. Moreover, in factories and farms, instruments can communicate and collaborate to enhance the performance of the farms and factories’ operations. Many other IoT application scenarios improve and facilitate our daily life [
3].
Access networks are responsible for connecting and merging all IoT devices to a centralized infrastructure or access point using either wireless or wired links [
4]. Wired links are not suitable for all IoT applications, especially those deployed in remote locations and vehicular applications due to cost ineffectiveness and lack of scalability or mobility. On the other hand, they can provide high reliability, high rates, and short delay. Wireless access technologies can either be using capillary networks, which are local networks that use short-range radio-access technologies to provide local connectivity (e.g., wireless local area networks, Zigbee, Wi-Fi, Bluetooth, etc.) [
5], or using cellular technology such as the global system for mobile communication (2G/GSM), universal mobile telecommunications system (3G/UMTS), Long-Term Evolution (4G/LTE), fifth-generation (5G) mobile network, etc. The wireless capillary is considered as a low-power and lossy network [
6]; additionally, it is relatively low cost and scalable. However, cellular networks are considered reliable and support mobility, which makes cellular IoT a promising technology for several applications [
7]. Moreover, the evolving fifth-generation (5G) cellular wireless networks provide higher data rates, enhance end-user quality of experience (QoE), reduce end-to-end latency, and lower energy consumption [
8]. Furthermore, 5G networks are expected to massively expand today’s IoT, which can boost cellular operations, IoT security, and network challenges and drive the Internet future to the edge [
9].
Fifth-generation (5G) technologies enable machine-to-machine (M2M), device-to-device (D2D), and device-to-everything (D2E) communication, IoT, and Internet of Vehicles (IoV). This kind of communication will require an enormous demand in future communication systems, which will be fast and will include more connected devices that are normally supported in combined networks. In this work, we have multiple IoT devices and cellular users (CUEs) sharing the same spectrum, transmitting and receiving data to IoT gateways and base stations (BSs), respectively. We investigate the effect of the interference occurring due to the number of transmitting IoT devices (IoTDs) on base stations (BSs) and the number of cellular users (CUEs) on the IoT gateway (IoTG) and find an adequate solution to decrease interference using an optimization strategy to enhance the IoT and cellular network performance in terms of reliability and efficiency under different environmental conditions. These findings are used to identify, based on different constraints and requirements, the suitable and acceptable distance between CUE and IoTG and IoTD and BS that can decrease interference and optimize system performance for the whole network in a changing environment. The proposed approach is unique in the sense that it provides an efficient way to find the most adequate distance for CUE–IoTG and IoTD–BS. This method is evaluated based on the performance model we derive and its simulation. The remainder of this paper is organized as follows.
Section 2 discusses the relevance of this research with other work reported. The system model and the problem formulation are presented in
Section 3. Simulation results produced by MATLAB and discussions are presented in
Section 4. Finally, the paper is concluded in
Section 5.
2. Related Work
IoT devices’ deployment over cellular networks has major technical challenges at the architecture, operational, and economic levels [
10,
11]. Furthermore, satisfying the IoT application requirements along with their associated communication technologies [
12] and selecting an efficient service for optimal management of both energy and quality of service (QoS) [
13,
14] are other challenges to be considered. Thus, different techniques and methods have been studied extensively to show how to enhance the QoS, the communication between IoT devices, and increase the energy efficiency [
15]. To meet the requirements of high energy efficiency and large system capacity for the fifth-generation Internet of Things (IoT), the authors of [
16] considered the energy-efficient design of a multipair decode-and-forward relay-based IoT network to obtain an accurate and efficient system performance. Moreover, to enhance the transmission efficiency as well as the reception reliability, cooperative relaying to the IoT network was introduced in [
17] and a cooperative IoT protocol was proposed. Additionally, for IoT system enhancement, ref. [
18] presented multiband cooperative spectrum sensing and a resource allocation framework for IoT in cognitive 5G networks to enhance the system QoS in terms of data rate, latency, reliability, economic price, and environmental cost. On the other hand, ref. [
19] defined effective throughput and effective amount of information to enhance transmission efficiency and reliability.
Concerning the energy consumption for IoT, different techniques and methods have been presented to solve this issue—for example, ref. [
20] proposed a power measurement utility for reinforcement learning (PMU-RL) algorithm to dynamically adjust the resource utilization of heterogeneous platforms in order to minimize power consumption, and [
21] described an energy-centric and context-aware optimization framework that accounts for the energy impact of the fundamental functionalities of an IoT system. Moreover, a new architecture has been presented in [
22] to allocate the IoT network resource and minimize the overall energy consumption of the pico relay base stations. On the other hand, the problem of how to achieve a tradeoff between the QoS provisioning and the energy efficiency for the industrial IoT systems was proposed in [
23], to achieve the objective of balancing the outage performance and the network lifetime. Furthermore, an energy- and spectrum-efficient IoT network for 5G systems was presented in [
24] to increase the IoT network utilization. Furthermore, it has been proven that the energy consumption of a network can be minimized via joint power control and time allocation [
25] or via joint transmission time and power assignment [
26]. Additionally, the paper [
27] addressed the optimized scheduling of the monitoring role between the embedded devices in IoT networks, to minimize energy consumption and communication overhead of monitoring, for each node. Furthermore, an efficient suboptimal algorithm was proposed [
28] to solve the energy efficiency maximization problem and achieve better energy efficiency performance for IoT devices. Moreover, increasing the energy efficiency fairness among heterogeneous mobile devices in interference channels will become a crucial issue in future wireless networks that can be solved using the deep learning approach [
29]. However, an elastic cell-zooming algorithm based on the quality of service and traffic loads of end-users was proposed in [
30], performed by adaptively adjusting the transmission power of small cells in order to reduce energy consumption for IoT.
Moreover, decreasing the energy consumption for decreasing interference in IoT network is a key issue, depending on the connectivity between IoT devices. Nash non-cooperative power game was an effective method implemented in [
31] to decrease the interference between IoT users which largely enhanced the edge IoT user throughput. The sum rate of the IoT cellular network can be maximized by a joint sub-band assignment and power allocation optimization [
32] and low-complexity, low-cost IoT communication is enabled by applying a multi-rate filtered orthogonal frequency division multiplexing system (F-OFDM) [
33]. Additionally, the mitigation of the IoT interference to LTE-A is an important issue presented in [
34] to improve accuracy of and eliminate IoT interference in LTE-A systems.
Although there have been many reports establishing the possibility of improving the system performance of IoT networks by using different proper communication protocols or data processing algorithms, there is a lack of information concerning how to identify the suitable approximate distance between CUE-IoTG and IoTD-BS under different conditions that can decrease the interference and ensure the best system performance for both IoT and cellular networks. In this paper, the main concern is the identification of the conditions for establishing the minimum appropriate distance between CUE-IoTG and IoTD-BS to decrease interference at the BS and IoTG. The proposed approach is based on the development of analytical and simulation models for the proposed approach and the assessment of its performance in reliability, throughput, and energy efficiency on both networks. The suggested approach shows how to achieve optimum performance by adapting the distance between CUE-IoTG and IoTD-BS to enhance their system performance. However, the following
Table 1 gives a summary of the comparison between the relevant research literature and the proposed model in this work.
3. System Model and Problem Formulation
The analytical models of the suitable distance between IoTD-BS and CUE-IoTG, system outage probability, energy efficiency, and throughput in the context of IoT and cellular networks are established in this section. Based on this model, the interference between IoTD-BS and CUE-IoTG can be controlled and decreased to optimize and enhance the system performance for both networks. Given a heterogeneous cellular network consisting of
N CUEs connected to the BS at the center and a cluster of
K IoTDs connected to the IoTG, for any CUE-BS, the goal of the optimization proposed in this work is achieved by maximizing energy efficiency (
EE), maximizing the end-to-end throughput (
Sth), and minimizing the system outage probability (
Pout) based on controlling the interference between CUE-IoTG and IoTD-BS; i.e.,:
where
Poutij,
EEij, and
Sthij are the total system outage probability, the total energy efficiency, and the overall system throughput, respectively, of the
i-th path between a CUE-BS and the
j-th path between IoTD and IoTG;
IIB,
ICG,
IIBmax, and
ICGmax are the interference and the maximum acceptable interference between IoTD and BS and CUE and IoTG, respectively.
The transmission scheme in the context of cellular and IoT networks is identified in
Figure 1, including direct CUE-BS and direct IoTD-IoTG. The IoTG acts as the data condenser of all the IoTDs and it sends data to the BS by using the available radio resources from the CUEs. In this work, we intend to examine and compare cellular and IoT networks’ performance in energy efficiency (
EE), throughput (
Sth), and system outage probability (
Pout) and to optimize both network performance in different environmental conditions.
A narrowband Rayleigh fading with additive white Gaussian noise (AWGN) [
19,
35,
36] and resulting propagation path loss are assumed in the considered system. The channel fading coefficient is assumed to be statistically mutually independent for different links. The signals received at the BS and the IoTG due to the transmissions are given by [
35]:
where
PC,
dCB, and
hCB are the transmission power, transmission distance, and the channel coefficient, respectively, of the CUE–BS link;
PI,
dIG, and
hIG are the transmission power, transmission distance, and the channel coefficient, respectively, of the IoTD–IoTG link; α is the path loss exponent;
X and
Y are the transmitted symbols of CUE–BS and IoTD–IoTG links, respectively, with unit power;
nCB and
nIG are the independent and identically distributed (i.i.d.) additive white Gaussian noise (AWGN) of CUE–BS and IoTD–IoTG links, respectively, with variance
No/2 per dimension, where
No is the thermal noise power spectral density per Hertz. The corresponding signal-to-interference ratio (
SINR) at the BS and the IoTG are represented by:
where
B is the system bandwidth in Hertz;
γCB and
γIG are the path loss model of CUE–BS and IoTD–IoTG links, respectively. The path loss model between any two nodes
i and
j at a distance
dij can be expressed as [
35]:
where
γo is the path loss constant. The interference between IoTD and BS (
IIB) and CUE and IoTG (
ICG) can be given by:
where
hIB and
hCG are the channel coefficients of IoTD–BS and CUE–IoTG links, respectively;
γIB and
γCG represent the path loss model of IoTD–BS and CUE–IoTG links, respectively. We assume that both destinations have certain QoS constraints that need to be satisfied to correctly decode the received signals from the CUE and the IoTD. The QoS constraints can be satisfied if, and only if,
SINRCB ≥
SINRthCB and
SINRIG ≥
SINRthIG, where
SINRthCB and
SINRthIG are the threshold
SINR values at the CUE–BS and the IoTD–IoTG links, respectively. Based on this assumption,
IIBmax and
ICGmax can be expressed as:
3.1. Minimization of Outage Probability under the Interference Control Model
An outage probability (
Pout) is defined as the probability that the
SINR at the receiver falls below the required threshold
β for CUE–BS links and
η for IoTD–IoTG links and which allows error-free decoding. The outage probability of the single-hop transmission is given by [
35,
36].
Assuming
βNoB <<
PC |
hCB|
2 γCB and
ηNoB <<
PI |
hIG|
2 γIG, Equations (11) and (12) can be written as:
Then, the overall system outage probability can be expressed as:
The main objective of the performance optimization of the proposed model is to minimize the total overall system outage probability under different environmental conditions.
In the formulated optimization problem, C1 is the constraint that the interference between IoTD and BS (
IIB) must be lower than the maximum acceptable interference between IoTD and BS (
IIBmax). C2 denotes the constraints that the interference between CUE and IoTG (
ICG) must be lower than the maximum acceptable interference between CUE and IoTG (
ICGmax). To find the solution to the optimization problem specified in (17), we now investigate the first-order optimality conditions. The Lagrangian of the optimization problem can be calculated as:
where
λIB and
λCG consider the non-negative Lagrangian multipliers for C1 and C2, respectively. By taking the derivative of Equation (18) with respect to
dIB and
dCG, we find the optimal solution to (17) as:
Equation (19) shows the value of
λIB that satisfies the constraint of the optimization problem.
Equation (20) shows the value of
λCG that satisfies the constraint of the optimization problem. By taking the derivative of (18) with respect to
λIB and
λCG, we find that:
Then, from the first-order optimality conditions, we can derive the following two propositions. Proposition 1: this defines the optimal required distance between IoTD and BS, and Proposition 2, which defines the optimal required distance between CUE and IoTG. The two propositions aim to decrease the interference and enhance the system performance through maximizing QoS. Thus, to achieve the system model goal, IoTD should be at a distance equal to
dIB from the BS and CUE should be at a distance equal to
dCG from the IoTG, which is given by:
3.2. Maximization of Energy Efficiency under the Interference Control Model
Energy efficiency (
EE) can be defined as the ratio between the overall system throughput and the total transmission power.
EE for the proposed model can be formulated as:
where
RCB and
RIG are the throughput of CUE–BS and IoTD–IoTG links, respectively;
Po is the internal circuitry power consumption of the CUE–BS and IoTD–IoTG links.
RCB and
RIG can be given by:
Then,
EE can be expressed as:
We can formulate the
EE maximization problem by specifying a maximum interference level in a similar way to the method expressed in (17), which can be stated as:
To find the solution to the optimization problem specified in (27), we now investigate the first-order optimality conditions. The Lagrangian of the optimization problem can be calculated as:
where
ξIB and
ξCG consider the non-negative Lagrangian multipliers for C1 and C2, respectively. By taking the derivative of Equation (28) with respect to
dIB and
dCG, we find the optimal solution to (27) as:
Equations (29) and (30) show the values of
ξIB and
ξCG that satisfy the constraint of the optimization problem, respectively. By taking the derivative of (28) with respect to
ξIB and
ξCG, we find that:
Therefore,
where
dIB and
dCG represent the optimum required distances between IoTD and BS and CUE and IoTG, respectively, to reduce the interference and enhance the system performance in terms of energy efficiency.
3.3. Maximization of Overall System Throughput under the Interference Control Model
The overall system throughput can be defined as the sum of the data rates that are delivered to all terminals in a network. The throughput is measured in bit/s or bps. The overall system throughput of the proposed model can be stated as:
In a similar way to the method expressed in (17), we can express the overall system throughput maximization problem by:
To solve the optimization problem specified in (27), we will investigate the first-order optimality conditions. The Lagrangian of this optimization problem can be found as:
where
ΩIB and
ΩCG consider the non-negative Lagrangian multipliers for C1 and C2, respectively. This can be obtained by taking the derivative of Equation (36) with respect to
dIB and
dCG to maximize the overall system throughput with interference control as follows:
Equations (37) and (38) show the values of
ΩIB and
ΩCG that satisfy the constraint of the optimization problem, respectively. By taking the derivative of Equation (36) with respect to
ΩIB and
ΩCG,
we respectively conclude that:
where
dIB and
dCG represent the optimum required distance between IoTD and BS and CUE and IoTG, respectively, to reduce the interference and enhance the system performance in terms of throughput.
The related results of the proposed approach will be shown in
Section 4.
4. Numerical Results and Discussion
In this section, the performance of the proposed approach through MATLAB simulation is examined in terms of reliability, system outage probability, energy efficiency, and overall system throughput and is compared with previous related work [
15,
18]. The conditions for adapting the optimal required distance through the analysis of the results obtained are revealed. The network settings for the simulation are listed in
Table 2.
Figure 2 and
Figure 3 depict the optimum required distance between CUE and IoTG and IoTD and BS, respectively, which enhances the system performance and decreases the interference. Assume that P
C and P
I are 23 dBm and
SINRCBth and
SINRIGth are 10 dB. As we can see from
Figure 2, based on our assumption and scenario, the optimum required distance between IoTD and BS (
dIB) must be approximately equal to 1.8 times the transmission distance between CUE and BS (
dCB)—for example, when
dCB = 300 m,
dIB must be at least 552 m to satisfy the system requirement, decrease interference, and enhance the system performance. The same result is obtained as shown in
Figure 3 to enhance the IoT system performance; the optimum required distance between CUE and IoTG (
dCG) must be approximately equal to 1.8 times the transmission distance between IoTD and IoTG (
dIG)—for example, when
dIG = 198 m,
dCG must be at least 354 m. The obtained results can help in designing the IoT architecture and to determine the installation location for IoTD.
The optimum required distance between IoTD and BS and CUE and IoTG is examined again for different scenarios but against
SINRthCB and
SINRthIG, as shown in
Figure 4 and
Figure 5. In
Figure 4, we assume that
dCB is 100, 200, and 300 m and
SINRthCB varied from −10 to 20 dB. As we can observe from
Figure 4, each
SINRthCB required a specific distance between IoTD and BS to satisfy the system requirements and enhance the system performance—for example, when
dCB = 200 m and
SINRthCB = 0 dB, the optimum required distance between IoTD and BS is 200 m, while when
dCB = 200 m and
SINRthCB = 15 dB, the optimum required distance between IoTD and BS is 486 m. In
Figure 5, we assume that
dIG is 40, 80, and 120 m and
SINRthIG varied from −10 to 20 dB. The optimum required distance between CUE and IoTG will be chosen based on the required
SINRthIG, as demonstrated in
Figure 5—for example, when
dIG = 120 m and
SINRthIG = 0 dB, the optimum required distance between CUE and IoTG is 200 m; however, when
dIG = 120 m and
SINRthCB = 16 dB, the optimum required distance between IoTD and BS is 301 m. These two figures show that an adaptive smart system should be held based on the system requirements and environmental conditions to enhance the performance of the CUE and IoT networks.
Figure 6 and
Figure 7 demonstrate the effect of CUE transmission power (
PC) and IoTD transmission power (
PI) on overall energy efficiency (
EE) and overall system throughput (
Sth) for the proposed model and compare it with the model proposed in [
15], assuming that
dCB = 200 m,
dIG = 80 m, and
SINRthCB =
SINRthIG = 0 dB. As shown in
Figure 6, the increase in
PC and
PI leads to higher
EE in the proposed model as compared to the schemes previously proposed in [
15], whereas after a certain threshold, the
EE reaches a peak and additional increase in maximum transmit power brings no enhancement in
EE but rather a decrease in
EE, which means that for this scenario and based on the transmission distance and SINR between CUE and BS and IoTD and IoTG, the best CUE transmission power (
PC) and IoTD transmission power (
PI) is 6 dBm for both to obtain the maximum
EE. It is also noticed that the proposed optimization model outperforms the other previously proposed schemes and increases the overall
EE. The same performance is obtained in
Figure 7; the overall system throughput of the proposed model performs well comparing with the model proposed in [
15]. It is also mentioned that the overall system throughput of the optimized proposed model for this transmission distance and SINR between CUE and BS and IoTD and IoTG increases when increasing
PC and
PI, which means that maximizing the throughput of the system may lead to the decrease in the
EE.
We can conclude from the results obtained in
Figure 6 and
Figure 7 that each transmission distance and each SINR between CUE and BS and IoTD and IoTG leads to determining the required transmission power of CUE and IoTD to achieve the maximum
EE with the best overall system throughput. Furthermore, these two figures demonstrate the effectiveness and the accuracy of the proposed model when comparing it with the other model in [
15].
To better clarify the results obtained in
Figure 6 and
Figure 7, we plotted
Figure 8 to present the EE versus the overall system throughput. For a low throughput, EE increases with the throughput. However,
EE decreases with throughput when it is larger. It means that there is a tradeoff between the
EE and throughput, namely, increasing the throughput may decrease the
EE. In practical systems, for obtaining the maximum EE with the best system throughput, the two performances can be jointly considered in order to attain the required desirable tradeoff between the two metrics under different conditions. The system performance is evaluated in terms of reliability and overall system outage probability versus the number of CUEs and IoTDs, where
PC =
PI = 23 dBm, and
SINRthCB =
SINRthIG = 20 dB.
Figure 9 demonstrates that increasing the number of CUEs and IoTDs leads to higher reliability for the proposed model and the model proposed in [
18], which means that both models aim to enhance the system reliability using different techniques regardless of the number of transmissions. The technique presented in [
18] was based on determining a minimum number of channels to be sensed by each IoT node in a multiband approach to enhance system reliability and increase EE, while our proposed model is based on adapting the optimum required distance between CUE and IoTG and the system performance is evaluated in terms of reliability and overall system outage probability versus the number of CUEs and IoTD, where
PC =
PI = 23 dBm, and
SINRthCB =
SINRthIG = 20 dB.
The system performance is evaluated in terms of reliability and overall system outage probability versus the number of CUEs and IoTDs, where
PC =
PI = 23 dBm, and
SINRthCB =
SINRthIG = 20 dB.
Figure 9 demonstrates that increasing the number of CUEs and IoTDs leads to higher reliability for the proposed model and the model proposed in [
18], which means that both models aim to enhance the system reliability using different techniques, whatever the number of transmissions is. The technique presented in [
18] was based on determining a minimum number of channels to be sensed by each IoT node in a multiband approach to enhance system reliability and increase
EE, while our proposed model is based on adapting the optimum required distance between CUE and IoTG and IoTD and BS based on the transmission distance between CUE and BS and IoTD and IoTG to decrease the interference, enhance system reliability, and increase
EE and system throughput. It is also mentioned that our proposed framework outperforms the model proposed in [
18]. The same performance is obtained when evaluating the system performance in terms of overall system outage probability, as shown in
Figure 10.
Figure 10 depicts that increasing the number of sender nodes decreases the system outage probability for both models, and our proposed model has better performance than the model proposed in [
18]. This result is due to the high link reliability for any CUE–BS and IoTD–IoTG, which also reflects the results obtained in
Figure 9.
Finally, we examine the overall system throughput and energy efficiency for the different number of CUEs and IoTDs, assuming
PC = PI = 23 dBm and
SINRthCB = SINRthIG = 20 dB, and the distances between the CUE and BS and IoTD and IoTG links are randomly chosen and vary between 10 and 500 m.
Figure 11 depicts the overall system throughput of the proposed model and the model proposed in [
18] versus the number of CUEs and IoTDs. As it is shown, the overall system throughput increases when the number of CUEs and IoTDs increases for both models, as increasing the number of sender nodes increases the number of transmitted packets, which increases the overall system throughput. This also reflects the result obtained in
Figure 9; higher reliability means higher system throughput. It is worth mentioning that the proposed model has a much higher overall system throughput than the model proposed in [
18]; this is due to the suggested proposed model being based on adapting the distances between IoTD and BS and CUE and IoTG, which decreases the interference between IoTD and BS and CUE and IoTG and thus improves the system performance.
Figure 12 depicts the energy efficiency of the proposed model and the model proposed in [
18] by counting on all possible transmission distances between CUE-BS and IoTD-IoTG links. As illustrated in
Figure 12 the proposed model in [
18] has lower energy efficiency than the proposed model, additionally, it shows that for both models
EE gradually increases when the number of CUE and IoTD increases. As mentioned previously, increasing the number of sender nodes increases the number of transmitted packets which increases the overall system throughput and then increases
EE. The result obtained in this figure and
Figure 9 and
Figure 11 show the effectiveness and the accuracy of the proposed model compared with other previously proposed models.
A summarized analysis for the performance of the proposed model can be given based on the above discussions and the results presented in
Figure 2,
Figure 3,
Figure 4,
Figure 5,
Figure 6,
Figure 7,
Figure 8,
Figure 9,
Figure 10,
Figure 11 and
Figure 12. Firstly, based on our analysis, there are several factors affecting system throughput, energy efficiency, reliability, and outage probability. The idea of the proposed approach is an adaptive way to adjust the distance between IoTD-BS (
dIB) and CUE-IoTG (
dCG), which decreases the interference that occurs at BS and IoTG and affects the system performance. (
dIB) and (
dCG) will be adapted based on different environmental conditions such as path loss exponent, (
dCB) and (
dIG), and the required QoS to enhance the system performance.
Secondly, in the evaluation of the QoS performance of the proposed model, we demonstrate a clear advantage of the proposed approach over the other models. Specifically, an important performance measure for QoS is reliability and also system outage probability, which is affected by the transmission power, as decreasing the transmission power affects the system outage probability; on the other hand, if the transmission power increases, the interference at the BS and IoTG increases due to the different number of CUEs and IoTDs. As a result of the overall impact of this factor on throughput and energy efficiency, a desirable tradeoff between them must be jointly evaluated to show improvement in the system performance to maintain the QoS required.
Clearly, based on the requirement of a particular application, to achieve the best system performance in terms of throughput, energy efficiency, outage probability, and system reliability, proper distance between CUE-IoTG and IoTD-BS should be adapted for the given transmission conditions, such as overall distance, dCB, dIG, channel quality in terms of α, and the transmission power of CUE (PC) and IoTD (PI). Our work provides effective guidance for deciding when and how the CUEs and IoTDs should be communicating and be placed. For achieving high throughput and energy efficiency as the main QoS requirement, the proposed approach is recommended for the given transmission conditions. Therefore, based on our results, the enhancement of cellular and IoT network performances can be achieved by adaptively choosing the appropriate distance between CUE-IoTG and IoTD-BS under different network and transmission conditions.