Next Article in Journal
Evolving Paradigm of Prothrombin Time Diagnostics with Its Growing Clinical Relevance towards Cardio-Compromised and COVID-19 Affected Population
Next Article in Special Issue
Managing the Cloud Continuum: Lessons Learnt from a Real Fog-to-Cloud Deployment
Previous Article in Journal
Multiplexed Affinity Measurements of Extracellular Vesicles Binding Kinetics
Previous Article in Special Issue
Matching IoT Devices to the Fog Service Providers: A Mechanism Design Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Prioritized Task Distribution Considering Opportunistic Fog Computing Nodes

School of Computer Engineering, Hanshin University, Osan 18101, Korea
Sensors 2021, 21(8), 2635; https://doi.org/10.3390/s21082635
Submission received: 21 March 2021 / Revised: 4 April 2021 / Accepted: 7 April 2021 / Published: 9 April 2021

Abstract

:
As service latency and core network load relates to performance issues in the conventional cloud-based computing environment, the fog computing system has gained a lot of interest. However, since the load can be concentrated on specific fog computing nodes because of spatial and temporal service characteristics, performance degradation can occur, resulting in quality of service (QoS) degradation, especially for delay-sensitive services. Therefore, this paper proposes a prioritized task distribution scheme, which considers static as well as opportunistic fog computing nodes according to their mobility feature. Based on the requirements of offloaded tasks, the proposed scheme supports delay sensitive task processing at the static fog node and delay in-sensitive tasks by means of opportunistic fog nodes for task distribution. To assess the performance of the proposed scheme, we develop an analytic model for the service response delay. Extensive simulation results are given to validate the analytic model and to show the performance of the proposed scheme, compared to the conventional schemes in terms of service response delay and outage probability.

1. Introduction

In the modern world, lots of smart devices, including smartphones, wearable devices, factory facilities, and vehicles, have been equipped with various sensors and are connected between devices. This realizes the internet of things (IoT) networks, where information is collected and shared among the connected devices and also accelerates the commercialization of IoT applications, such as smart factory, smart home, and smart environment [1,2].
Since IoT applications have their own service requirements according to their characteristics, each service should be processed and provided to guarantee the requirements. For example, video-on-demand (VoD), online gaming, augmented reality (AR), and virtual reality (VR) are typically delay-sensitive services that need low latency. On the other hand, a large portion of data traffic can tolerate relatively long latency, such as pushing of contents to the edge, log and backup services, and tenant delivery [3,4,5]. According to the various service requirements, efficient resource utilization to process the requests of services should be considered.
Since IoT devices have limited resources in terms of computing and energy, the cloud service has been provided to process the data generated from lots of IoT devices owing to the flexible utilization of computing resources and support of high volume with fast scalability [6]. However, the physical distance between the cloud servers and IoT devices results in long latency and consumes high bandwidth of the core network. In addition, the load can be concentrated to the cloud according to the number of IoT devices. In order to address these challenges, the concept of fog computing is introduced, where computing resources are moved close to the IoT devices to distribute the load of the cloud, minimize latency for IoT services, and reduce core network resource usage [7,8].
In the fog computing environment, the requests of IoT devices usually can be offloaded to the fog computing nodes (FNs) co-located with the access point (AP) such as the base station (BS) where IoT devices are connected [9,10]. This physical proximity can provide low latency. However, a bottleneck can occur when the number of requests of IoT devices increase. To handle this problem, there have been works on the task distribution of IoT devices considering the co-work with other FNs and collaboration between the cloud and FNs [9,11,12,13,14], as well as context awareness [15,16].
Recently, task distribution considering not only static FNs but also mobile FNs has gained increasing attention [17,18,19,20,21,22,23]. Owing to the intermittent availability of mobile FNs, they are considered as the concept of opportunistic FNs (OFNs) [18]. Research coverage on OFNs has been extended to include smart phones, vehicles, and unmanned aerial vehicles (UAVs) [21].
As mentioned above, OFNs are not always available because of mobility and intermittent connectivity. In addition, communication and computing latency can be additionally considered. This means that the service requirements of delay-sensitive services cannot be guaranteed through offloading with the OFNs [18]. In order to handle this problem, this paper introduces the prioritized task distribution scheme, which utilizes OFNs when they are available only for delay-tolerant tasks and static FNs for delay-sensitive tasks. To evaluate the performance of the proposed scheme, we develop the analytical model for response delay. Extensive simulation results validate the analytic model and demonstrate that the proposed scheme has lower response delay while maintaining the outage probability at a low level compared to that of the conventional schemes.
The key contribution of this paper is two-fold: (1) this paper develops an analytic model based on Markov chain of the proposed and conventional schemes for the response delay; and (2) by means of extensive simulation works, this paper demonstrates the performance of the proposed and conventional schemes under various environments, which can be a valuable design reference for OFN-based architecture.
The remainder of this paper is organized as follows. After related work is reviewed in Section 2, system models for the proposed and conventional schemes are given in Section 3. Simulation results and concluding remarks are described in Section 4 and Section 5, respectively.

2. Related Work

In cloud and fog architecture, the task distribution of IoT devices has been discussed a lot [9,11,12,13,14,15,16]. Task distribution schemes were provided by associating tasks into suitable FNs to minimize the service response delay considering communications and processing procedures [9,11]. In addition, collaborative computing between cloud and fog (i.e., load sharing) was utilized to achieve better delay performance by means of optimal task splitting [12,13]. Yi et al. [14] presented a different role between fog and cloud. For example, local and regional task can be processed on fog to provide timely feedback, such as emergency cases and computational-intensive task can be scheduled on the cloud. Kayes et al. [15] reviewed the previous context aware access control approaches and provided general requirements with challenging issues to provide context-awareness of the fog-based access control. Moreover, a fog-based, context-aware access control scheme was proposed [16], which provides the benefits of a unified data model and its associated access and privacy control policies to reduce the administrative and processing overheads. Although these works did not consider the OFNs, their efforts became the groundwork for works on OFN-based task distribution.
There have been lots of studies on distribution of load utilizing OFNs with different objectives [17,18,19,20,21,22,23]. Minimizing service latency has been one of the major issues [17,22]. A dynamic task allocation scheme was proposed utilizing both static FN and OFN to optimize the service latency and quality loss rate [17]. Although service interruption due to the mobility of OFN was mentioned, it does not consider the differentiation of delay-sensitive flows, which can result in a long response delay for flows due to the repetitive resource re-allocation. Wang et al. [22] introduced a model with parked and moving vehicles (i.e., FNs and OFNs) to minimize the average system response time. Since it also allocates the request to OFNs without differentiation, the performance requirements of delay-sensitive services cannot be guaranteed due to the mobility of OFNs even though the average performance can be improved. In addition, various issues on OFN have been covered as follows. Fernando et al. [18] reviewed motivations and identified the requirements to enable OFN applications. Then, they provided a model of OFNs to support IoT applications, especially for hazardous and volatile events. Ning et al. [19] introduced an energy-efficient scheduling scheme. It schedules the task to the static FN and OFN in a cooperative manner to minimize the energy consumption of network access devices within the delay constraint. Liu et al. [20] formulated a task scheduling decision problem based on task dependency requirements to reduce the average completion time. Zhou et al. [21] investigated a computation resource allocation problem for the task assignment to optimize long-term network delay performance. To motivate OFNs for resource sharing, they utilized a contract-based incentive mechanism. Liu et al. [23] analyzed the utility-based task distribution model according to the mobility of OFNs. They focused on the temporal and spatial characteristics of the relationships between OFNs without considerations on FNs.
Unlike the aforementioned works, this paper mainly considers the differentiation of service requests and distribution according to the availability of OFN.

3. System Model

3.1. System Architecture

Figure 1 depicts the system architecture, where IoT devices offload tasks to FN and wait for a response. As shown in Figure 1, the offloading can be processed by either OFN or FN when OFN is available (or not) due to mobility. This paper assumes that OFN notifies its events of entering and leaving and reports current computing status to AP, based on the existing cellular registration mechanism [17].
As mentioned above, since there are various requirements, depending on the IoT applications, this paper divides request flows into high priority (HP) flows (which require delay constraint) and low priority (LP) flows (which are delay tolerant). The criteria of this differentiation can be changed according to the network status and operator’s policy.
Offloading to OFN enables the load distribution of FN. This means the response of FN can be reduced. However, flows processed by OFN can have longer response delay than that of FN because the offloading to OFN needs an additional delivery procedure and the computing resource of OFN cannot always be guaranteed.
Therefore, load distribution should be performed considering the flow differentiation and features of FN and OFN, as described above.

3.2. System Model of the Proposed Scheme

To develop an analytical model, this paper considers M/M/1 queuing models for fog computing architecture [10,17], where the request flows of HP and LP follow Poisson distribution with rates λHP and λLP. In addition, since the task sizes of the request flows are assumed to follow the exponential distribution, the service times of FN and OFN also follow the exponential distribution with mean of 1/μF and 1/μOF, respectively [9]. Figure 2 shows the Markov chain model of FN and OFN in the proposed scheme. In state (i, n, j), i represents the status of availability for OFN, n means the serving node to process the request, and j denotes the number of requests currently served by n. Each request can be served by either FN (F) or OFN (O), where CF and CO are the capacities of FN and OFN, respectively. Note that each FN (OFN) can have different values of CF (CO) depending on its own capability. Status A means that the offloading to OFN is available because it is located in the coverage of AP. On the other hand, OFN is unavailable at status U. The sojourn time of OFN at status A and U follows the exponential distribution with rates 1/η and 1/ξ, respectively [23,24,25]. In Figure 2, LP flows are only be offloaded to OFN at status A; to distributed the load of FN and HP, flows are processed by FN to reduce the latency. At status U, FN processes both HP and LP in a round-robin fashion because OFN is not available. This paper assumes that the requests from IoT devices within the range of AP are offloaded to FN or OFN connected directly with AP. Collaborative offloading with other APs and cloud [9,11] and prioritized processing even at status U, such as using priority queue allocation [26], will be one of our future works.
The transition rates of FN in Figure 2 can be obtained as follows.
p ( A , F , j ; U , F , j ) = η ( 0 j C F ) p ( U , F , j ; A , F , j ) = ξ ( 0 j C F ) p ( A , F , j ; A , F , j 1 ) = μ F ( 0 j C F ) p ( A , F , j ; A , F , j + 1 ) = λ H P ( 0 j C F ) p ( U , F , j ; U , F , j 1 ) = μ F ( 0 j C F ) p ( U , F , j ; U , F , j + 1 ) = λ H P + λ L P ( 0 j C F ) p ( A , F , j ; A , O , j + 1 ) = λ L P ( j = 0 ) p ( A , O , j ; A , O , j + 1 ) = λ L P ( 1 j C O ) p ( A , O , j ; A , O , j 1 ) = μ O F ( 1 j C O ) p ( A , F , j ; A , O , j + 1 ) = λ L P ( j = 1 ) p ( A , O , j ; A , F , j 1 ) = μ O F ( j = 1 )
In order to find out steady state probability πi,n,j, the balance equations can be calculated as follows:
( 1 ) i = A , n = F , j = 0 , ( λ H P + λ L P ) π i , n , j = μ F π i , n , j + 1 + ξ π U , n , j + μ O F π i , O , j + 1 ( 2 ) i = A , n = F , 0 < j < C F , ( λ H P + η + μ F ) π i , n , j = λ H P π i , n , j 1 + μ F π i , n , j + 1 + ξ π U , n , j ( 3 ) i = A , n = F , j = C F , ( η + μ F ) π i , n , j = λ H P π i , n , j 1 + ξ π U , n , j ( 4 ) i = U , n = F , j = 0 , ( ξ + λ H P + λ L P ) π i , n , j = μ F π i , n , j + 1 + η π A , n , j ( 5 ) i = U , n = F , 0 < j < C F , ( ξ + λ H P + λ L P + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + μ F π i , n , j + 1 + η π A , n , j ( 6 ) i = U , n = F , j = C F , ( ξ + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + η π A , n , j ( 7 ) i = A , n = O , j = 1 , ( μ O F + λ L P ) π i , n , j = μ O F π i , n , j + 1 + λ L P π U , F , j 1 ( 8 ) i = A , n = O , 1 < j < C O F , ( λ L P + μ O F ) π i , n , j = ( λ L P ) π i , n , j 1 + μ O F π i , n , j + 1
Because of the complexity of closed-forms for πi,n,j, this paper utilizes an iterative algorithm to obtain πi,n,j [26]. To get the response delay of HP flow, the average number of HP requests (NH) in FN can be given by:
N H = i = A , U j = 0 C F j π i , F , j
The average number of LP requests (NL) can also be given by:
N L = j = 0 C F j π U , O , j + j = 1 C O j π A , O , j
In addition, by considering each status of FN, the effective request arrival rate of HP requests (λeH) can be calculated as follows:
λ e H = j = 0 C F λ H P π A , F , j + j = 0 C F ( λ H P + λ L P ) π U , F , j
In the same way, the effective request arrival rate of LP requests (λeL) can be calculated as follows:
λ e L = j = 1 C O λ L P π A , O , j + j = 0 C F ( λ H P + λ L P ) π U , F , j
Then, by means of Little’s law [27], the average response delay of HP (WHP) and LP (WLP) flows can be obtained by (7) and (8), respectively:
W H P = N H λ e H
W L P = N L λ e L
Even though the proposed scheme preferentially handles the HP flow, the response delay cannot guarantee the required delay constraint if the amount of incoming requests increases continuously or the incoming requests are concentrated instantly. Therefore, the outage probability to show the QoS degradation, wherein the request cannot get a response with application delay constraint, will be analyzed in the next chapter.
On the other hand, if the delay constraint is always guaranteed when the amount of incoming request is small, the distribution to the OFN can be preferred or not based on the network policy. Even in this situation, this paper utilizes the OFN as shown in Figure 2 because making the best use of OFN is efficient for scalability [28].

3.3. System Model of the Conventional Scheme

Compared to the proposed scheme where incoming requests are classified into HP and LP flows and processed based on this classification as shown in Figure 2, the conventional scheme offloads the incoming request to FN and OFN without flow differentiation to make the best use of available resources [21]. This means all the incoming requests can be processed evenly by both FN and OFN if OFN is available or by only FN if OFN is unavailable, as shown in Figure 3. Especially when OFN is available, since the incoming requests can be distributed to FN and OFN, the proportions of the incoming requests to FN and OFN are set to α and β, respectively (i.e., α + β = 1). Since the differentiation is not considered in the conventional scheme, the response delay for the incoming requests depends on the proportions (α, β), irrespective of the delay requirement.
The transition rates of FN in Figure 3 can be obtained as follows:
p ( A , F , j ; U , F , j ) = η ( 0 j C F ) p ( U , F , j ; A , F , j ) = ξ ( 0 j C F ) p ( A , F , j ; A , F , j 1 ) = μ F ( 0 j C F ) p ( A , F , j ; A , F , j + 1 ) = ( λ H P + λ L P ) α ( 0 j C F ) p ( U , F , j ; U , F , j 1 ) = μ F ( 0 j C F ) p ( U , F , j ; U , F , j + 1 ) = λ H P + λ L P ( 0 j C F ) p ( A , F , j ; A , O , j + 1 ) = ( λ H P + λ L P ) β ( j = 0 ) p ( A , O , j ; A , F , j 1 ) = μ O F ( j = 1 ) p ( A , O , j ; A , O , j + 1 ) = ( λ H P + λ L P ) β ( 1 j C O ) p ( A , O , j ; A , O , j 1 ) = μ O F ( 1 j C O )
The difference between Equations (1) and (6) is the transition probability from state j to j + 1 at status A. In order to find out steady state probability πi,n,j, the balance equations can be calculated as follows:
( 1 ) i = A , n = F , j = 0 , ( λ H P + λ L P ) π i , n , j = μ F π i , n , j + 1 + ξ π U , n , j + μ O F π i , O , j + 1 ( 2 ) i = A , n = F , 0 < j < C F , ( λ H P + λ L P + η + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + μ F π i , n , j + 1 + ξ π U , n , j ( 3 ) i = A , n = F , j = C F , ( η + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + ξ π U , n , j ( 4 ) i = U , n = F , j = 0 , ( ξ + λ H P + λ L P ) π i , n , j = μ F π i , n , j + 1 + η π A , n , j ( 5 ) i = U , n = F , 0 < j < C F , ( ξ + λ H P + λ L P + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + μ F π i , n , j + 1 + η π A , n , j ( 6 ) i = U , n = F , j = C F , ( ξ + μ F ) π i , n , j = ( λ H P + λ L P ) π i , n , j 1 + η π A , n , j ( 7 ) i = A , n = O , j = 1 , β ( λ H P + λ L P ) π i , n , j = μ O F π i , n , j + 1 + β ( λ H P + λ L P ) π i , F , j 1 ( 8 ) i = A , n = O , 0 < j < C O , β ( λ H P + λ L P ) π i , n , j = β ( λ H P + λ L P ) π i , n , j 1 + μ O F π i , n , j + 1
As mentioned above, πi,n,j can be obtained using an iterative algorithm. In addition, by means of Equations from (3) to (8), the average response delay of HP and LP flows can also be calculated.
In addition, the basic scenario without OFN can be modelled only considering status A in Figure 3.

4. Performance Analysis

In this section, we evaluate the performance of the proposed scheme compared with the conventional scheme without differentiation (NoDiff) [21] and basic scenario without OFN. For numerical analysis, the average service time of FN is assumed to be 1 ms. Since OFNs generally have limited capacity compared to FNs [29], this paper assumes that the average size of the total capacity of FN (CF) and OFN (CO) is set as 20 and 10, respectively. For NoDiff, the requests are evenly distributed to FN and OFN by AP when OFN is available (i.e., both α and β are set to 1/2). To verify the analytical results marked as (A), event-driven simulations based on MATLAB R2018a are conducted and the simulation results are marked as (S) in the following figures. In the simulations, this paper assumes that CF and CO follow uniform distribution from 17 to 23 and from 7 to 13, respectively. Arrival times of events with CF and CO are drawn by generating 50,000 random numbers according to the distribution, and then the response delay and outage probability are computed.

4.1. Response Delay

Figure 4 shows the response delay according to the LP flow arrival rate when the HP flow arrival rate is 0.3. Both η and ξ are set to 1/3 in Figure 4a and η and ξ are 2/3 and 1/3, respectively, in Figure 4b. First of all, as shown in Figure 4, simulation results are almost consistent with analytic results in all simulation settings. In Figure 4a, the NoDiff scheme has a higher response delay compared to that of HP flows in the proposed scheme because it makes the best use of FN and OFN without the differentiation when OFN is available to be utilized. In addition, the response delay of NoDiff scheme is higher than that of the basic scheme when LP flow arrival rate is low. This is because NoDiff offloads the incoming request to OFN, which has relatively low capacity when OFN is available even though the load of FN is not high. Note that this effect can be higher if the latency between FN and OFN is considered, although they are not included in this paper. Therefore, it can be noticed that it is required to differentiate the requests for the delay performance with efficient available resource utilization. For example, the response delay of the proposed scheme is about 49% and 62% shorter than that of NoDiff and the basic scheme, respectively, when LP flow arrival rate is 0.5 (i.e., fifth x-axis point in Figure 4a). On the other hand, LP flows in the proposed scheme have a higher response delay compared to those of NoDiff because LP flows are processed only using OFNs when OFNs are available. Figure 4b shows a similar trend with Figure 4a. However, the difference of the response delay between the proposed and NoDiff schemes does not increase according to the LP flow arrival rate, compared to that in Figure 4a. This is because the period when FN processes all the requests by itself increases owing to the lower sojourn time than that of Figure 4a. Comparing Figure 4b with Figure 4a, the response delay of LP flows becomes improved while HP flows of the proposed scheme has the lowest response delay among all schemes. From the results, in order to prevent LP flows from starvation, the appropriate criteria for the differentiation between HP and LP flows should be determined based on the network status such as OFN sojourn time, delay constraint, and the amount of incoming requests. The optimal solution to find the criteria will be one of our future works.
Figure 5 shows the response delay according to the ratio of HP flow arrival rate to LP flow arrival rate when the LP flow arrival rate is 0.5. Both η and ξ are set to 1/3 in Figure 5a and η and ξ are 2/3 and 1/3, respectively in Figure 5b. Figure 5 also shows similar trend with Figure 4 because the proposed scheme performs differentiated processing of HP flow requests. In Figure 5b, it can be noted that the effects of the proposed scheme can be reduced according to the ratio of HP and LP flow arrival rate. This also means that the impact of the differentiation between HP and LP flows becomes smaller. For example, the differences in response delay for the HP flows in the proposed scheme from that of NoDiff and LP flows in the proposed scheme are 12.9 ms and 21.5 ms when the ratio is 0.3, and 0.6 ms and 0.9 ms when the ratio is 1, respectively. This is because in the proposed scheme, the amount of HP flow requests to FN increases under the capacity constraint and the amount of LP flow requests processed by OFN is reduced owing to the lower sojourn time than that of Figure 5a.

4.2. Outage Probability

Figure 6 shows the outage probability (i.e., the ratio of the number of HP flow requests which do not satisfy the delay constraints to the total number of HP flow requests). Both η and ξ are set to 1/3 in Figure 6a and η and ξ are 2/3 and 1/3, respectively in Figure 6b. The delay constraint is assumed to be 30 ms. From Figure 6, outage probability increases according to the LP flow arrival rate because the system has a capacity constraint. However, the proposed scheme can have lower outage probability, which means a higher QoS satisfaction ratio compared to the conventional schemes because it preferentially handles the HP flow requests. As explained above, since the response delay of the proposed scheme becomes higher with increasing η, the outage probability also increases when comparing Figure 6a with Figure 6b. However, the proposed scheme still has lower outage probability compared to the conventional schemes because of the flow differentiation. Maintaining the outage probability below the specific value is an important performance metric from a network operator’s perspective.

5. Conclusions

In this paper, a prioritized task distribution scheme considering opportunistic fog computing nodes in the fog computing environment is proposed. The proposed scheme differentiates incoming flow requests into delay-sensitive and delay-insensitive flows. Then, delay-sensitive flows can be processed by static FN to support the delay requirement. On the other hand, the proposed scheme makes the best use of OFN for delay-insensitive flows to reduce the load of the static FN. Numerical and simulation results show that the proposed scheme can provide lower service delay for the delay-sensitive flows, compared to the conventional schemes, while maintaining the outage probability at a low level. In our future work, experiments considering the real environment with commercial IoT devices and mobile computing node to distribute tasks will be performed.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (MSIT) (No. 2020R1G1A1100493).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Roy, S.S.; Puthal, D.; Sharma, S.; Mohanty, S.P.; Zomaya, A.Y. Building a sustainable Internet of Things: Energy-efficient routing using low-power sensors will meet the need. IEEE Consum. Electron. Mag. 2018, 7, 42–49. [Google Scholar] [CrossRef]
  2. Yao, J.; Ansari, N. QoS-Aware Fog Resource Provisioning and Mobile Device Power Control in IoT Networks. IEEE Trans. Netw. Serv. Manag. 2019, 16, 167–175. [Google Scholar] [CrossRef]
  3. Si, P.; He, Y.; Yao, H.; Yang, R.; Zhang, Y. Deve: Offloading Delay-Tolerant Data Traffic to Connected Vehicle Networks. IEEE Trans. Veh. Technol. 2016, 65, 3941–3953. [Google Scholar] [CrossRef]
  4. Li, M.; Si, P.; Zhang, Y. Delay-Tolerant Data Traffic to Software-Defined Vehicular Networks with Mobile Edge Computing in Smart City. IEEE Trans. Veh. Technol. 2018, 67, 9073–9086. [Google Scholar] [CrossRef]
  5. Pereira, P.P.; Casaca, A.; Rodrigues, J.J.P.C.; Soares, V.N.G.J.; Triay, J.; Cervello’-Pastor, C. From delay-tolerant networks to vehicular delay-tolerant networks. IEEE Commun. Surv. Tutor. 2012, 14, 1166–1182. [Google Scholar] [CrossRef]
  6. Truong, H.-L.; Dustdar, S. Principles for engineering IoT cloud systems. IEEE Cloud Comput. 2015, 2, 68–76. [Google Scholar] [CrossRef]
  7. Liu, Y.; Fieldsend, J.E.; Min, G. A framework of fog computing: Architecture, challenges, and optimization. IEEE Access 2017, 5, 25445–25454. [Google Scholar] [CrossRef]
  8. Dastjerdi, A.V.; Buyya, R. Fog Computing: Helping the Internet of Things Realize Its Potential. IEEE Comput. 2016, 49, 112–116. [Google Scholar] [CrossRef]
  9. Fan, Q.; Ansari, N. Towards Workload Balancing in Fog Computing Empowered IoT. IEEE Trans. Netw. Sci. Eng. 2018, 7, 253–262. [Google Scholar] [CrossRef]
  10. Iorga, M.; Feldman, L.; Barton, R.; Martin, M.J.; Goren, N.; Mahmoudi, C. Fog Computing Conceptual Model; NIST: Gaithersburg, MA, USA, 2018.
  11. Sun, X.; Ansari, N. Latency Aware Workload Offloading in the Cloudlet Network. IEEE Commun. Lett. 2017, 21, 1–22. [Google Scholar] [CrossRef]
  12. Ren, J.; Yu, G.; He, Y.; Li, Y. Collaborative Cloud and Edge Computing for Latency Minimization. IEEE Trans. Veh. Technol. 2019, 68, 5031–5044. [Google Scholar] [CrossRef]
  13. Yousefpour, A.; Ishigaki, G.; Jue, J.P. Fog Computing: Towards Minimizing Delay in the Internet of Things. In Proceedings of the 2017 IEEE 1st International Conference on Edge Computing, Honolulu, HI, USA, 25–30 June 2017. [Google Scholar]
  14. Yi, S.; Li, C.; Li, Q. A Survey of Fog Computing: Concepts, Applications and Issues. In Proceedings of the ACM Mobidata, Hangzhou, China, 21 June 2015; pp. 37–42. [Google Scholar]
  15. Kayes, A.S.M.; Rahayu, W.; Watters, P.; Alazab, M.; Dillon, T. Achieving Security Scalability and Flexibility Using Fog-Based Context-Aware Access Control. Future Gener. Comput. Syst. 2020, 107, 307–323. [Google Scholar] [CrossRef]
  16. Kayes, A.S.M.; Kalaria, R.; Sarker, I.H.; Islam, M.S.; Watters, P.A.; Ng, A.; Hammoudeh, M.; Badsha, S.; Kumara, I. A Survey of Context-Aware Access Control Mechanisms for Cloud and Fog Networks: Taxonomy and Open Research Issues. Sensors 2020, 20, 2464. [Google Scholar] [CrossRef]
  17. Zhu, C.; Tao, J.; Paster, G.; Xiao, Y.; Ji, Y.; Zhou, Q.; Li, Y.; Yla-Jaaski, A. Folo: Latency and Quality Optimized Task Allocation in Vehicular Fog Computing. IEEE Internet Things J. 2019, 6, 4150–4161. [Google Scholar] [CrossRef] [Green Version]
  18. Fernando, N.; Loke, S.W.; Avazpour, I.; Chen, F.; Abkenar, A.B.; Ibrahim, A. Opportunistic Fog for IoT: Challenges and Opportunities. IEEE Internet Things J. 2019, 6, 8897–8910. [Google Scholar] [CrossRef]
  19. Ning, Z.; Juang, J.; Wang, X.; Rodrigues, J.J.P.C.; Guo, L. Mobile Edge Computing-Enabled Internet of Vehicles: Toward Energy-Efficient Scheduling. IEEE Netw. 2019, 33, 198–205. [Google Scholar] [CrossRef]
  20. Liu, Y.; Wang, W.; Zhao, Q.; Du, S.; Zhou, A.; Ma, X.; Yang, F. Dependency-Aware Task Scheduling in Vehicular Edge Computing. IEEE Internet Things J. 2020, 7, 4961–4971. [Google Scholar] [CrossRef]
  21. Zhou, Z.; Liu, P.; Feng, J.; Zhang, Y.; Mumtaz, S.; Rodriguez, J. Computation Resource Allocation and Task Assignment Optimization in Vehicular Fog Computing: A Contract-Matching Approach. IEEE Trans. Veh. Technol. 2019, 68, 3113–3125. [Google Scholar] [CrossRef]
  22. Wang, X.; Ning, Z.; Wang, L. Offloading in Internet of Vehicles: A Fog-enabled Real-time Traffic Management System. IEEE Trans. Ind. Inform. 2018, 14, 4568–4578. [Google Scholar] [CrossRef]
  23. Liu, Y.; Wang, W.; Ma, Y.; Yang, Z.; Yu, F. Distributed Task Offloading in Heterogeneous Vehicular Crowd Sensing. Sensors 2016, 16, 1090. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, D.; Liu, Z.; Wang, X.; Lan, Y. Mobility-Aware Task Offloading and Migration Schemes in Fog Computing Networks. IEEE Access 2019, 7, 43356–43368. [Google Scholar] [CrossRef]
  25. Lee, J.; Lee, G.; Pack, S. Pseudonyms in IPv6 ITS Communications: Use of Pseudonyms, Performance Degradation, and Optimal Pseudonyms Change. Int. J. Distrib. Sens. Netw. 2015, 11, 1–7. [Google Scholar] [CrossRef]
  26. Kim, Y.; Ko, H.; Pack, S.; Lee, W.; Shen, X. Mobility-Aware Call Admission Control Algorithm with Handoff Queue in Mobile Hotspots. IEEE Trans. Veh. Technol. 2013, 62, 3903–3912. [Google Scholar] [CrossRef] [Green Version]
  27. Kleinrock, L. Queueing Systems: Theory; Wiley: New York, NY, USA, 1975; Volume I. [Google Scholar]
  28. McCauley, J.; Panda, A.; Krishnamurthy, A.; Shenker, S. Thoughts on Load Distribution and the Role of Programmable Switches. ACM SIGCOMM Comput. Commun. Rev. 2019, 49, 18–23. [Google Scholar] [CrossRef]
  29. Zhao, L.; Yang, K.; Tan, Z.; Li, X.; Sharma, S.; Liu, Z. A Novel Cost Optimization Strategy for SDN-Enabled UAV-Assisted Vehicular Computation Offloading. IEEE Trans. Intell. Transp. Syst. 2020, 1–11. [Google Scholar] [CrossRef]
Figure 1. System architecture.
Figure 1. System architecture.
Sensors 21 02635 g001
Figure 2. Markov chain model of the proposed scheme: (a) OFN; (b) FN.
Figure 2. Markov chain model of the proposed scheme: (a) OFN; (b) FN.
Sensors 21 02635 g002
Figure 3. Markov chain model of the conventional scheme: (a) OFN; (b) FN.
Figure 3. Markov chain model of the conventional scheme: (a) OFN; (b) FN.
Sensors 21 02635 g003
Figure 4. Response delay according to the LP flow arrival rate (A: analytic results, S: simulation results): (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Figure 4. Response delay according to the LP flow arrival rate (A: analytic results, S: simulation results): (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Sensors 21 02635 g004
Figure 5. Response delay according to the ratio of HP flow arrival rate to LP flow arrival rate (A: analytic results, S: simulation results): (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Figure 5. Response delay according to the ratio of HP flow arrival rate to LP flow arrival rate (A: analytic results, S: simulation results): (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Sensors 21 02635 g005
Figure 6. Outage probability of HP flow requests: (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Figure 6. Outage probability of HP flow requests: (a) η = 1/3, ξ = 1/3, (b) η = 2/3, ξ = 1/3.
Sensors 21 02635 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kyung, Y. Prioritized Task Distribution Considering Opportunistic Fog Computing Nodes. Sensors 2021, 21, 2635. https://doi.org/10.3390/s21082635

AMA Style

Kyung Y. Prioritized Task Distribution Considering Opportunistic Fog Computing Nodes. Sensors. 2021; 21(8):2635. https://doi.org/10.3390/s21082635

Chicago/Turabian Style

Kyung, Yeunwoong. 2021. "Prioritized Task Distribution Considering Opportunistic Fog Computing Nodes" Sensors 21, no. 8: 2635. https://doi.org/10.3390/s21082635

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop