1. Importance of Task Scheduling for Hybrid Internet of Things and Cloud Environment
To improve production processes and produce higher quality products at lower prices, the manufacturing process incorporates ample information, analytic capabilities, high-performance computation, and the commercial Internet of Things (IoT) into production systems and sectors [
1]. Nowadays, factories, gadgets, machines, and procedures are linked, monitored, and optimized to increase production and efficiency. Today, the use of these systems in manufacturing is a critical facilitator for tackling issues with scheduling, order fulfillment, transportation, and supply management. When sensors are used to monitor manufacturing lines in real-time, they save the performance data and then analyze the data (founded on the observations) to generate the planned or anticipated production schedule. For the increasing computational demands of big data cloud computing (CC) applications, a smarter approach is required. The big data and cloud computing applications are widely used in digital society. The combined form of application is utilized in the business sector. By providing scalable and malleable computing devices/resources that can be utilized on a pay-as-you-go basis, cloud computing presents a substantial increase in the number of Internet of Things (IoT) applications. Many Big-Data-related tasks for these IoT apps are performed in the cloud as a service to the cloud, and their results are then shared amongst many devices. Big-Data-related tasks manage all types of datasets and analyze and evaluate the complete structure of data. Big data are implemented in IoT application by the transfer of data among wearable sensors, smart devices. However, another obvious difficulty is estimating how many devices will be needed to carry out a cloud-based set of extremely large-scale operations.
The public cloud is gaining popularity as a new form of distributed computation in the digital age. By 2019, the cloud provider will have processed over 70% of the calculations. Cloud computing offers consumers effective, secure computation and communication services, but it has driven significant economic growth in various industries [
2,
3,
4,
5]. The interconnection of particles is becoming more extensive and healthier, the devices connected to networks are exploding, and many devices at the edge require a low delay and high serviceability due to the fast expansion of connected phones, smart systems, and other industrial IoT [
5].
Although cloud technology has a large-scale parallelization design, some problems, such as low resource utilization and an imbalanced load, still exist and significantly impact task efficiency. The rise of the cloud environment is crucial at this time because large-scale data transport uses up a lot of bandwidth and adds complexity to datacenters; because of the various terminal requirements, the necessities for assets are also diverse, and the long-distance communication system produces a significant delay, making it impossible to satisfy time-sensitive customer requirements.
The most fundamental issue with cloud computing task scheduling is matching requirements with available resources. In IoT environments, the user requirements vary, and the performance of the cloud computing integrated services node is noticeably worse than that of the combined services node. Therefore, it is essential to address the problem of effectively allocating the resources in cloud computing and allocating the resources based on user needs.
This research looked at intelligent reconfiguring and planning for a smart industrial system with advanced technologies. The goal is to reduce the entire cost of all tasks related to tardiness. The research enables real-time optimization and intelligent rearrangement judgment. Q-learning, the most prominent algorithm in reinforcement learning, calculates reinforcement for states and actions. Q-results learning is affected by both the initial state and the subsequent actions taken. Where there is a finite set of states and actions for a reinforcement learning problem, Q-learning is employed.
Schedules can be made in a variety of ways, and the genetic algorithm is among them. By calculating fitness costs, this method can evaluate different schedules and choose the one that provides the best overall outcome. Then, people can find more effective routines by employing crossover, mutation, and elitism selections.
The main contributions are listed below:
An optimal scheduling model founded on deep learning is utilized to assess and process different tasks to reduce energy costs and end-to-end delay.
A regional exploratory search strategy enhances the optimization algorithm’s ability to use data and avoid becoming stuck in local optimization.
Accurate trace data and the CloudSim tools were used to compare DLA-BDTSS to other popular task allocation methods. DLA-BDTSS’s rapid convergence made it an excellent choice for scheduling tasks involving big data. The task is scheduled and processed by means of Q-learning, which can be developed as the effective scheduling agent for all potential users to complete every task that needs to be fulfilled. From the collected task, the GA encodes the task and processes the task to obtain a better result. With a 34s execution time and a fitness value evaluation of 76.8 percent, DLA-BDTSS achieved an 8.43% improvement in results.
In this study, we propose a deep learning algorithm for scheduling big data tasks in the cloud and the Internet of Things. The results of the study showed that DLA-BDTSS was superior to the most common competing methods. It is helpful for scheduling big data tasks because it achieves higher percentage improvement in the outcomes and converges more quickly than competing methods.
The later part of the research is listed below: In
Section 2, we will cover the history and results of the task scheduling models. In
Section 3, we develop, test, and deploy the Deep Learning Algorithm for Big Data Task Scheduling System (DLA-BDTSS) that we propose. In
Section 4, we show the results of our analysis of the proposed system’s software.
Section 5 collects the report’s final thoughts and findings from the system.
2. Background Study
There are two types of standard task scheduling techniques: exact scheduling techniques and approximate scheduling techniques. By searching for the complete optimal solution, the optimal solution on a global scale is found by the same methods. The high computational complexity makes precise practices ineffective for dealing with large-scale planning issues. Instead of searching every possible solution, approximation scheduling methods explore a large number of possible paths based on a small number of predetermined strategies. Therefore, approximate procedures are preferable in dealing with complex scheduling issues because they generate practical solutions more quickly and with less computing complexity. Estimation techniques, however, cannot promise the best answer for a specific scheduling problem.
Some academics have suggested reinforcement learning techniques for work planning in a service-oriented setting. A reinforcement learning technique was presented [
6] to address the scheduling issue in distributed systems while considering the heterogeneity and placement of units within the grids to achieve a faster completion time. A multi-agent paradigm for optimization utilizing metaheuristics is suggested. All individuals share data and cooperate to allow the agent to change their actions depending on experiences obtained through engaging with other agencies and their surroundings [
7]. The number of process apps and virtualization software were employed as state variables [
8] in multi-agent deep-Q networks, along with the maximum completion time and cost, to help schedule multiple workflows in the public cloud.
Wang et al. [
9] expanded the deep-Q system to handle the task scheduling choices of numerous network edges to monitor the development of multi-workflows over infrastructure-as-a-service clouds. The issue of scheduling algorithms in cloud technology has drawn attention for some time. However, fog computing has lately been proposed, and studies on different cloud computing features have been published. Still, only a few tasks are scheduled. Yang et al. [
10] present Real-Time Dynamic Max min (RTDM) dynamic task scheduling, which accounts for makespan, load balancing, and overall waiting time. The main goal of the Expense-Makespan Aware Scheduling Heuristic Model proposed by [
11] is to distinguish between the benefits of employing cloud services and infrastructure and the efficiency of application code.
Fellir et al. [
12] presented a system that uses a decision tree to handle tasks while taking into account changing environments. Every sensor’s decision tree algorithm categorizes seasonal chores and determines which ones should be used in which order. The investigation of enhanced heuristic algorithm-based hierarchical capacity scheduling for IoT technology is thorough. Many scientists have previously investigated ways to strengthen task scheduling weaknesses through the use of intelligent algorithms for example genetic algorithms (GA) and particle swarm optimizations (PSO) [
13,
14], which have produced excellent results on a variety of levels (such as makespan and planning time).
A new planning technique was also introduced in [
15], founded on the Q-learning process, which is among the most effective reinforcement learning techniques. An adaptive Cauchy mutation-based cuckoo search approach is utilized in [
16]. Designing an optimization strategy is based on the life of a cuckoo. Therefore, this guideline for optimization strategies helps to find the best way to carry out the task with reliable and beneficial results. To create a system-performance-sensitive planner [
17], considering the framework of the IoT, one examines and analyzes K-means, principal component analysis, and fuzzy C-means clustering (FCM) methods. Sefati et al. [
18] proposed a dynamic scheduling technique for cloud computing that prioritized requests from the IoT while considering homogenous and heterogeneous servers.
Hasan et al. [
19] propose a scheduling algorithm strategy for time-sensitive data streaming systems on homogeneous clouds that reschedules such tasks and provides enough resources to reduce their waiting time. Pal S. [
20] created the IoT-SCOM model and optimise it to select the optimal deployment option with the shortest guaranteed delay. The experimental results show that the IoT-SCOM approach outperforms established methods and the stochastic optimization algorithm in terms of reliability and performance for the difficulty of data-intensive delivery component installation in the edge-cloud environment.
After carefully examining the studies mentioned above, it was discovered that they have advantages and disadvantages in several areas: (1) Numerous studies, such as [
21,
22,
23], contributed to the study of optimization techniques and provided adequate exposure to fog computing theory and platforms. (2) Previous research, such as [
24,
25,
26], considered optimizing the duration and power usage caused by cloud computing. Nonetheless, they failed to consider the issue of device loading and latency in terms of activities and resource types.
Pal. S. et al. [
27] discussed a system design for the optimal sequence of the Johnson Scheduling algorithm. This sequence represents service times. A multi-server queueing model with finite capacity diminishes waiting time and channel capacity, improving job scheduling.
Mukherjee, D et al. [
28] created the Adaptive Scheduling Algorithm Based Task Loading (ASA-TL) algorithm, which is used to store tasks from cloud data centers on digital devices. According to the results of the experiment, ASA-TL appears to have the best scores for response time, data center processing time, and overall cost.
M. Goudarzi et al. [
29] organized recent research on IoT system scheduling in fog computing. The current literature is examined, research gaps are identified, and future directions are described in light of their new categorizations. Saha, S. et al. [
30] describe a Task Scheduling algorithm based on a genetic algorithm that employs a queueing model to reduce the amount of time people must wait and the length of the queue.
Sing, R. et al. [
31] presented a multi-objective optimization technique that concurrently minimizes execution time, cost, and power usage by enhancing their respective parameters for optimal performance using evolutionary techniques. Better performance can be obtained by balancing cloud and fog nodes, as demonstrated by simulation results using various cloud and fog nodes.
In order to evaluate the efficacy of the proposed framework, Shresthamali, S. et al. [
32] focused on designs to simulate both single-task and dual-task systems. The results show that their Multi-Objective Reinforcement Learning (MORL) algorithms can learn superior policies while incurring lower learning costs and effectively balancing competing goals during execution. Pal. S [
20] and Mukherjee D. [
33] have also discussed optimization techniques for better results in cloud-IoT-edge environments.
Table 1 deals with the contributions and limitations of this background work in tabular format.
Gap Analysis
The literature review above indicates that meta-heuristics and deep learning have been thoroughly researched for production planning. Additionally, collaborative scheduling and reconfiguration optimization have become increasingly popular in recent years. Research has not yet addressed smart planning and reconfiguration that take dynamic work arrival into account. Methods for creating task schedules are driven by one or more goals; thus, it is important to compare them to gauge how effective they are. In this sense, the following critical metrics can be taken into account in scheduling techniques in distributed applications, such as cloud and fog environments:
Makespan: The overall amount of time required to process a group of tasks through to completion. However, in scheduling algorithms, makespan is useless because it is hard to anticipate when a collection of tasks is completed because a new task stream is always being added.
Efficiency is measured as the ratio of runtime to overall makespan.
Throughput: Throughput is the number of units of information that a system can process in a given amount of time.
Waiting period: the interval between task submissions and implementation, as well as the interval during which a task waits for occurrences or other supplies.
Execution time: the length of time a process runs and uses its resources.
Response time: This is the amount of time between when a task is submitted and when it is finished. It can also be computed as the total of the waiting period and the processing time.
Cost: the overall amount paid for the use of resources as well as supplemental costs such as energy costs.
The primary motivation for the creation of scheduling algorithms was the desire to both maximize efficiency and ensure fair distribution of available resources. Scheduling is the process of allocating available resources to the requests that have been received. Schedulers are a general class of algorithms.
In this paper, we propose a Deep Learning Algorithm for Big Data Task Scheduling System (DLA-BDTSS) that can be implemented in scheduling systems that operate in the cloud or over the Internet of Things. The research concluded that DLA-BDTSS is superior to several widely used alternatives. When compared to other methods, it converged faster and produced better results, making it a good choice for scheduling big data tasks.
3. Deep Learning Algorithm for Big Data Task Scheduling System
To simulate task planning and reconfiguration systems, this section designed reward, react, and state elements for planning and reconfiguring agents. When one of the two agents needs to take action, the agent creates that action based on aspects of the current environment. A reward is given to the agency to change its parameters when the action is completed.
In
Figure 1, we see the overall layout of the DLA-BDTSS system. Elements of the architecture include the Internet of Things units, cloud nodes, and fog nodes. The system’s generated tasks are queued using the deep learning model in accordance with the advised scheduling strategies. Emulating a network required the development of both reconfiguration and programming agents. A few reasons why autonomous design is a good idea are listed below. Firstly, it can ease training burdens by reducing the dimensionality of state features and action domains. The cloud, a fog node, and the final users are the foundation of
Figure 1. Fog nodes receive cloud data, which are then sent to the devices in the field. It is easy to see the information flowing between the cloud and the fog node. There can be both an actual queue and a scheduled queue for tasks to be completed in a fog node. The information is provided in its final form for consumers.
Additionally, planning and reconfiguring decision-making durations are similar if a unified agent is built. The agents must go through several epochs to understand how to decrease the frequency of reconfiguring. Additionally, the separate designation is scalable and can include other decision-makers, including a logistics provider.
3.1. System Architecture
Processing occurs in an information hub on a connected home, a smart router, port, converter, host system, etc., in the cloud environment. The connected home in the information hub represents the connection from the smart router to the host system in the cloud environment. The connected home in the cloud environment is the connection between the router and the host system. All the devices in the cloud environment are given to the connected home. The power and the smog nodes are given to the connected home. The complete connection in the information hub is dealt with by the connected hub. Due to the inadequate processing power, some fog nodes collaborate locally, and simultaneously link to cloud endpoints recognized as virtual cloud servers to meet mobile users’ demands. They assume that a fog brokerage, node, and fog nodes are all included in the network. Fog nodes talk to phone devices immediately. The fog brokers, who are in charge of analyzing, estimating, and then scheduling all activities to be carried out in the cloud service, receive all demands from phone devices right away. Since fog brokers are located near fog devices, data transfer between them takes time and is eliminated.
Figure 2A presents the DLA-BDTSS flow diagram for your viewing pleasure. Data analytics and related services are part of the cloud services offered from the beginning stages onward. The actual queue and the scheduled queue in the fog node work together to ensure that the task is placed in the appropriate queue. The final users are provided with the planned task in the manner of a smartphone.
Step-wise DLA-BDTSS Algorithm: The DLA-BDTSS method, built into fog brokers, seeks to create an ideal task execution schedule that maximizes time and money efficiency to secure the study’s validity. First, a message is sent by a smartphone network (step 1), and the connected fog node responds. The fog broker immediately receives this request (step 2). Step 3 divides each task into a group of activities that will be handled in a distributed network. Step 4 calculates the expected number of commands and resource usage. The fog broker manages all project and node data and uses a scheduling scheme (step 5) to determine a suitable task allocation. After the output, tasks are sent to the appropriate cloud networks and systems (step 6). Every node is in charge of completing all tasks allocated to it (step 7), after which it delivers the results directly to the fog brokerage (step 8). The response is given to the smartphone network via the fog layer the user is connected to when the outcome of the task is merged (step 9) after all tasks have been executed (step 10).
As shown in
Figure 2B, the DLA-BDTSS pipeline has been broken down and its process mapped out. This system includes fog nodes, cloud nodes, base stations, and mobile user devices. The following is a description of the scheduler’s process for the proposed plan:(1) The data from the mobile are sent as a request to the fog node.(2) This forms the fog node, and the task is forwarded to the base station. (3) The task from the base station is estimated, scheduled, and given to the cloud nodes. (4) The cloud node executes the task and then sends the result in a particular time. The process is initiated by users, and contact is maintained between cloud and fog nodes. The base station collects the forwarded task from the fog node. The base station collects the task, decomposes the task, estimates, and sends the obtained task in the form of a combined result. The base station combines resultsby decomposing the task, and schedules and sendsthe task. The combined result in the base station is implemented in particular time. The final combined result is calculated and distributed to users. The DLA-BDTSS architecture with the smartphone network sends the request to the fog node. The task is routed to the base station from the fog node. The task is transmitted to and communicated with via the base station. The cloud node receives the decomposed task from the base station. The task is carried out in the cloud node, and the results are executed. The architecture stage is followed by the reconfiguration agent.
3.2. Reconfiguration Agent
When the reconfiguring judgment unit chooses to rearrange, the reconfiguring agent (RA) generates a reconfiguring activity based on recent state characteristics and receives the benefit of the previous reconfiguring activity as well as the benefits of earlier transactions at each T step. The reconfiguration agent is used to rearrange the reconfiguring activity. In order to rearrange the reconfiguring operation, the reconfiguration agent is utilized. The creation of reward, training, and state elements is described below.
3.2.1. Reward
Remember that the goal of the investigated problem is to lower the overall tardiness cost for all tasks in the network. Every reconfiguration measure should contribute to lowering the overall cost of delay. In other words, the increase in the overall cost of timeliness throughout this configuration stage must be minimized for every reconfiguration process. A deep learning agent aims to maximize the cumulative reward for an episode. As a result, every reconfiguration phase’s compensation is specified as the opposite of the tardy cost that was just implemented per second at that step.
The timeliness cost is produced during a phase using buffer tasks and completed tasks. Let each finished task be converted during a reconfiguration process. When a present reconfiguration phase is finished, the function is reset.
and
are the startingand end durationof the presentreconfiguration stage, respectively. Consequently, the time step
reconfigure reward is shown in Equation (1).
The initial and final reconfiguration stages are expressed as and . The cumulative rewards in each episode are rewarded and given for the reconfiguration judgement. The reconfiguration judgement is described in the section below.
3.2.2. Reconfiguration Judgment
When the first device completes a task, the reconfiguration judgment is triggered to determine whether to rearrange. The RA makes a reconfiguration decision only when it chooses to do so. The frequency of human reconfiguration is decreased by the reconfiguration judgment system. If not, it will take many events for the RA to learn how to reduce the frequency of reconfiguration.
Let
represent the existing production mode and
represent its buffer. The present system time is indicated by
,
. The
stands for the work x’s current area tardy cost, where current cost refers to the moment in time
. The current cost
is expressed in Equation (2).
At the three main circumstances, the reconfiguration judgement unit makes a decision to reconfigure. The current time is denoted as , and the length of the queue is denoted as . The lower and higher cost function is denoted as and .
The cost function’s pictorial representation is expressed in
Figure 3. The final result is obtained from the currentcomputation cost of the task and the initial and final task values. The cost function is expressed below:
Case 1: when is blank.
Case 2: When
is smaller than the
xth% of and there are no outstanding tasks in
. Here,
is a list and
is the mean of the present unit delay cost of
y. The mean delay is expressed in Equation (3).
The mean cost is expressed as , and the total number of samples available is denoted as N.
Case 3: When N more tasks have been handled in
and
is below the y% of
. Every state characteristic is an array, so take note of it. The maximum, lowest, mean, and standard error are computed for every state attribute to reflect the properties of the underlying data. As a result, the state vector has a total dimension of 4 × 4 = 16. For state characteristics, min-max normalizing is employed. State feature
’s counterpart,
, is expressed in Equation (4).
The minimum and maximum functions are represented as . The group of all without resampling during the first ten periods is represented by in this case. For the resource , the max-min algorithm gives higher priority to the larger task than to the smaller tasks. In max-min, several smaller tasks can run parallel to the larger ones. The overall timeframe is determined by the completion of the longest job here. RA is a reward earned for an episode that reduces the cost delay. Human configuration is reduced as a result of the reconfiguration judgment. After the device reconfiguration stage is completed, deep learning uses an efficient scheduling agent to schedule the task. The goal of the scheduling agent is to choose, for each new task, the most suitable one from the pool of available options for resources and services.
3.3. Deep Learning Method
To create a satisfying real-time timetable for all upcoming tasks over the period, it is sometimes necessary to identify one or more good planning rules to employ throughout the entire plan-making process. Shortest Processing Time (SPT), Longest Processing Time (LPT), Shortest Queue Time (SQT), and Longest Queue Time (LQT) are examples of standard scheduling criteria. This study develops an effective scheduling agent to select an appropriate service among all potential resources for every task that arrives. The scheduling strategy is implemented using the deep learning system [
18]. Deep learning is a popular approach to machine learning that was developed with the purpose of teaching computers to perform tasks that are second nature to humans. The methods of deep learning are incorporated into the scheduling agent in order to partition the
Q-values into a variety of activities and services that are determined by the task. Specifically, the deep
Q-learning approach is used to resolve the problem under consideration. The
Q-learning method is used to observe the
Q-value with the particular action and services. The actions are given as the tasks which are given as the reward to serve the function. The scheduling agent totally depends on the
Q-value and the actions that focus on the services. The observed data are completely focused on the
Q-value and the action. The complete stages of
Q-value, action, service function, and the reward are called the scheduling agent.
The deep-learning-based scheduling model is depicted in
Figure 4. The system is used to analyze and process different tasks and produce an optimized task management model to lower the energy cost and end-to-end delay.
Scheduling Agent
It expresses the agency as an equation that can be optimized by representing it coherently and educating it. The loss function measures how far expectations and actual objectives diverge. A model might predict that delivering a task to its prospective services provides excellent value, even if providing the task to a different service yields higher rewards. The difference among the projected value and the desired value should be close. Equation (5) shows the cost function.
The initial and the current cost function are denoted as . The Q value is expressed as Q. The cost function is denoted as , and the number of data is represented as .
First, it involves selecting a provider for the actual project and adding it to that service’s backlog. The prize and the new state of the scheme are then visible. It determines the maximum goal Q-value derived from the findings and then reduces it. Finally, it calculates the goal value by multiplying the present reward by the reduced target Q-value.
There is a significant discrepancy between the specific value and the forecast value, so it is risky to use the same system to generate both. Therefore, it employs the framework comprising the target system and prediction system for instructing throughout the learning phase. With some values locked in place, the architecture of the predictive network is transplanted into the target environment. Consistent learning is achieved by periodically copying variables from the forecast network into the target network. The scheduling agent makes it possible to monitor the state of operations and services in relation to assigned jobs. In the serve function, the Q-value is rewarded. A target and prediction system are types of systems encountered during the learning phase. The task-scheduling mechanism is then provided with the normative values.
3.4. Task Scheduling Mechanism
The configuration of the edge server includes several virtual machines (VMs), indicated by . The computational capability of these VMs, shown by , is diverse. The scheduling algorithm chooses how tasks should be scheduled, including which VM to allocate each task to. The first task kept in the queue is placed into the awaiting slot that was left when a task was planned, which causes it to leave the awaiting slot. It is anticipated that just one VM would handle every task and that all of the VM’s computational resources would be used. Each VM’s anticipated processing period is unknown until it is executed.
A task’s system response, when assigned to a VM, comprises both the VM completion time and the waiting period in the awaiting slot. One can calculate task
x’s completion time in VM y as follows in Equation (6).
The launch time (
) of the actual project is the arrival date if the VM is not currently performing any tasks (
); otherwise, the task starts when the VM becomes available, which means all previous activities are done. Let
and
stand for the opening time and end time of task x on VM y, respectively. As a result, the beginning time is dependent on the completion times of all previous tasks as shown:
, and the task x completion time is given by Equation (7).
where
is how long it takes VM y to complete task
x, and the starting time is denoted as
. The waiting period and the completion time make up the reaction duration of task x on VM
y, andare given in Equation (8).
where
is the duration of task x’s waiting period on VM y, the task completion time is denoted as
. If the mission is finished instantly, there is no time delay; if not, the waiting period is defined as the interval between the task’s initial onset and its completion. The waiting time is denoted in Equation (9).
The task’s quality of experience is assessed using internet speed. The task starting time is expressed as
, and the delay time is expressed as
. The work satisfaction level, which is calculated as the proportion of the predicted delay and the reaction time, is the quality of experience (QoE) definition for every task. Equation (10) is used to indicate the task’s overall satisfaction of the work and the cost computation was carried out on VM y.
where
is anticipated delay,
y denotes the computational cost. It is evident that the ratio increases as work satisfaction increases. The task starting point and the ending point calculate the QoE. The edge server with the VM measures the computational capability and schedules the task for each VM. After task scheduling to VM, each task is given to an evolutionary algorithm called a genetic algorithm.
3.5. Evolutionary Algorithm
The DLA-BDTSS, built on an evolutionary process called genetic algorithm (GA), is suggested in order to address work scheduling issues in cloud-fog processing environments as follows.
3.5.1. Encoding of Chromosomes
In evolutionary algorithms, a person defined by a chromosome stands in for a task-scheduling system. An n-dimensional array matching y chromosomes is being used to encode chromosomally. Every gene has a meaning of a numeral x in the band [1; k] with k being the set of nodes, indicating that the associated task is given to the cluster designated x. The sequence identity of a gene is also the sequence identity of a task. For instance, if a system of three nodes in the cloud-fog executes a series of 10 tasks, one scheduling algorithm option is as follows:
Node tasks are denoted as . The chromosome that expresses the above solution will be . Node 1 performs the collection of tasks {2,5,9}, node 2 is tasked with processing the collection of tasks {1,3,8,10}, and node 3 is in charge of the collection of tasks {4,6,7}.
This method of chromosomal encoding was chosen because it allows for flexible genetic processes such as crossovers and mutations to produce new organisms that explore the space of potential solutions while inheriting good gene sections from their parents. Because goals, such as makespan and overall cost, are unaffected by tasks performed in a different sequence on a single machine, the order in which tasks are executed is ignored. The GA encodes the chromosomes, and each task is scheduled based on the processing of the collected task. The population is initialized after scheduling.
3.5.2. Initialize the Population
The starting population is the collection of all individuals used in a GA to determine the best course of action. Considering an N-person population, these N people are completely random in order to find many locations in the solution space and to guarantee the variety of the community in the first iteration. The next generation is formed by selecting some people from the starting population and performing some procedures on them. The upcoming generation is initialized, and the fitness evaluation is implemented in the next section of the fitness function.
3.5.3. Fitness Function
A person’s supremacy in the population is determined by their fitness function. A high-fitness person stands in for a high-quality solution. The utility function F estimates each person’s fitness value. The ability of a person to survive or perish through every generation depends on their fitness level. After the fitness function is initialized, the crossover operators, selection strategy, and mutation operator are initialized.
3.5.4. Genetic Operators
To produce offspring that inherit healthy genetics from their parents, two crossover operations are used with chromosomally encoded integers. In the process, two crossing nodes are randomly chosen, the first parent trades the middle genetic segment with the step-parent, and the other genes are left unaltered, resulting in the creation of a new person.
Figure 5 depicts the analysis of the DLA-BDTSS crossover model. The two parents are taken into account, and the crossover offspring chromosome is depicted above. The parental choice has a notable impact on the algorithm’s effectiveness in the population-wide crossover procedure. Every individual has a crossover rate. Each member of the community is assigned a probability of becoming the first parent, and the next parent in the community is chosen to participate in the crossover procedure using the roulette wheel method. High-quality participants with higher fitness values are much more probable to be select as parents in this type of selection, increasing the likelihood that beneficial gene segments will be passed down to the next generation.
The selection method is needed to choose individuals who will make up a community for the following generation. Spontaneous selection begins just after the crossing process. The fitness worth of the child is determined and compared to that of the mother; if the child is superior, the individual is retained to create the community of the following generation. Otherwise, the parent is kept in the following generation and the child is destroyed.
Whenever the genetic mutations of the less capable people are not removed as quickly and they still donate to exploring new territories on the search window, this selection retains the uniqueness of the populace over centuries. At the same time, great gene sections are not found in large numbers of people in a population.
A new population of N people is formed as a result of the processes of crossover and spontaneous choice. At a certain frequency of N people, everyone participates in one-point alteration. The position of the gene mutation was chosen at random and replaced with a new number to assign the task to a separate network.
Mutations improve the shortcomings of crossover and mutation by identifying the best solution when people avoid the extremes or flee them to investigate other regions of the candidate solutions. The array is matched with the chromosome after the matching process in encoding chromosomes. The population is created for the first time following the matching process. The fitness function determines a person’s authority. The genetic operators employ the crossover operator, the selection strategy, and the mutation operator. After initialization and processing, evaluation and validation take place.
3.6. Evaluation and Validation
The method for evaluating the tasks and validating the solutions is discussed below.
3.6.1. Applicative Results
It first builds a practical scenario to assess the suggested scheduler. It created a cloud layer of two fog networks to meet user demands. In the scenario, the fog node was set up to develop schedules for task devices, five of which were carried out in a public cloud. Applying the proposed approach in the timeframe, slots t1 and t2 produced the planned queuing results.
Each agent is assigned important tasks to deliver the critical ones and create the scheduled queue. Important information is used to show how important the activities are. This includes task arrival priorities, dependencies, progress, and the resources needed to finish.
are the tasks that have arrived. The number of the agencies is denoted as . The asset “types” in the instance are Type A, Type B, Type C, and Type D on our fog-cloud system, and the number of components available for each type. The suggested model’s planned queue and the worldwide queue of employees predicated on the first-come first-served method are shown as follows:
Tasks are in the universal queue depending on FCFS (applying wait period).
The queue that was planned at time t1.
Only and are available in the first time frame since their condition is Yes, and in our suggested model that equals Yes = 1 and No = 0. All tasks are prepared to be completed during the next time slot because their state is Yes, which in the suggested model equals Yes = 1. The queue was planned based on the proposed model for the period.
As seen in the practical scenario, the most critical work was completed first due to the application of the suggested model, which takes into account the task’s importance, duration, data processing, condition, and resource requirements. This results in better resource utilization. This review investigates the possibility of deep learning to provide informed recommendations regarding working groups and reconfiguration when scheduling activities with dynamical task delivery. To improve simulation results, the DLA-BDTSS system’s IoT nodes and big data analytics module with edge and fog layers employ an innovative task scheduling system with optimization models. The system is made up of fog nodes, cloud nodes, base stations, and mobile user devices. The user initiates the process, and interaction occurs between cloud and fog nodes. The results are totaled and reported to customers.
3.6.2. Task Scheduling
The following is the scheduling algorithm procedure:
Set up settings, such as metrics for techniques, virtualization software, and tasks.
The integration of resources and the classification of tasks into three groups.
Start with N explosions; each one symbolizes a solution.
Choose the resources needed to optimize.
Determine the fitness feature value for fireworks and the number of sparkles and explosion ranges.
Develop a technique to detect explosion radii and calculate the present cost.
Choose an optimization parameter to undergo a random evolution for task scheduling.
Select N iterations to represent the future testing population.
Step 4 (mentioned in
Section 3.1. System Architecture, Step-wise DLA-BDTSS Algorithm) continues if the maximal amount of repetitions is not achieved; otherwise, it ends.
In this section, the DLA-BDTSS system is designed with a deep learning and scheduling model. The DLA-BDTSS system’s results are theoretically evaluated in this section, and the findings are shown in the following section. Each stage of the DLA-BDTSS system is carried out in stages, such as initialization and task scheduling, and the evaluation is carried out in the proposed section.
4. Simulation Outcomes and Comparisons
We ran a series of numerical simulations in the trial report’s cloud-fog architecture to evaluate the performance of the DLA-BDTSS system in scheduling a collection of tasks. The costs of processing energy and resource utilization differ between fog and edge devices. Every node has a distinct processing capability, which is indicated by its operating rate and distinguished by its central processing unit (CPU), storage, and bandwidth utilization costs. In the cloud, edge devices such as routers, bridges, and workstations have limited processing capabilities.
In the cloud layer, tasks are handled by servers or virtualization software in high-speed data centers. As a result, cloud nodes interpret things significantly more quickly than fog nodes. However, using virtualized resources is more costly than using them in the cloud. Dollars (USD, $), a unit of cash used in simulations to represent actual money, are used to determine these expenditures.
The fitness level evaluation of the DLA-BDTSS system is represented in
Figure 6. The simulation evaluations of the DLA-BDTSS system are performed by varying the iteration size from minimum to maximum, and the outcomes are evaluated. As the iteration count increases, the individual fitness level of the system also increases. The existing PSO and GA exhibit poor performance, and the DLA-BDTSS system with deep learning and big data for task scheduling produces higher simulation outcomes.
Figure 7 shows the results of the cost study for the DLA-BDTSS system. After retrieving the number of tasks from the big data models, the overall cost of the computation may be determined. The energy cost rises proportionally with the number of operations performed. Overall system performance is enhanced by the DLA-BDTSS system’s combination of deep learning, hybrid cloud, fog, and IoT modules with big data analytics modules. With varying job sizes, the system’s efficiency improves while the cost does not.
The execution time analysis of the DLA-BDTSS is analyzed, and the results are tabulated in
Table 2. The simulation outcomes are measured by the variation in the tasks, which is varied from 10 to 100 at the same time. The proposed DLA-BDTSS with a deep learning algorithm and big data analytics module enhances task scheduling and hence produces a lower execution time than the other existing models. The cost function is directly linked to execution time, and the big data module with deep learning reduces the computation time and hence the execution time.
The software outcome evaluations of the DLA-BDTSS system are analyzed and plotted in
Figure 8. The balancing function is based on the cloud and fog nodes. The queries and the task are implemented based functions implemented. The scheduling of tasks is focused on the balancing function. The balancing function is computed using the edge and fog layers. The total energy cost and the makespan of the system are calculated based on the number of tasks. Depending on the number of tasks, the total cost and the life span of the proposed method are given. The balancing function is plotted against the total cost and the life span. The big data analytics module helps generate enough samples for the simulation evaluations. As the number of functions increases, the total energy cost of the system also increases, and the makespan of the system decreases. The makespan time should be kept to a minimum to generate faster results.
Figure 9 depicts the execution time analysis computed and plotted. With a step count of ten tasks, the number of tasks for the simulation outcomes varied from low to high. The execution time is calculated as the difference between the time it takes to generate a task and the time it obtains to complete it in the cloud and at edge nodes. The DLA-BDTSS system, which includes big data analytics and an optimized model, improves task scheduling and management. In terms of execution time, the hybrid edge and fog modules with IoT nodes produce better simulation results.
The fitness value evaluation of the DLA-BDTSS system is represented in
Table 3. The different iterations are considered for the simulation analysis, the outcomes are evaluated for the DLA-BDTSS and the results are compared with the existing models, such as PSO and GA. As the number of iterations increases, so does the fitness value of the DLA-BDTSS system. The fitness function is directly linked to the cost function and indirectly related to the execution time. The DLA-BDTSS exhibits higher outcomes with the hybrid cloud system with IoT and a big data analytics module.
The DLA-BDTSS system is designed and evaluated in this section. The DLA-BDTSS system with IoT nodes and a big data analytics module with an edge and fog layer enhances the simulation outcomes with a higher task scheduling system with optimization models. The performance measurement of DLA-BDTSS is based on the fitness value, cost computation, execution time, and software outcome evaluation.
5. Conclusions and Findings
This paper investigated intelligent planning and realignment for task scheduling with dynamical task delivery, utilizing deep learning to offer informed decisions for task scenarios and reconfiguring. System architecture for thoughtful planning and reconfiguring in a smart factory is provided as a solution to the issue. To reduce the overall timeliness cost for all tasks, a quantitative model is developed. Additionally, a deep learning system is suggested with a schedule and reconfigurable agents. The two services are the target audience for the agency’s activities, features, and incentives.
A deep learning model is also used to propose solutions to dynamic sequencing and reconfiguration with new work delivery. We can bridge the gap between traditional classroom instruction and digital learning environments by combining well-designed deep learning, instant decision making after training, and extensive use of performance data. These advantages highlight the value of applying extensive knowledge to address careful planning and reconfiguration issues in a real-world production environment. The findings of the study can be used to create a real-time, self-optimizing, decentralized platform for task scheduling in industrial automation systems.
The main bottleneck for cloud computing systems is task scheduling. A dependable task-scheduling method is required to improve system performance. CPU memory, execution time, and execution cost are all major concerns for today’s task scheduling algorithms. As a solution to the aforementioned issues, the Deep Learning Algorithm for Big Data Task Scheduling System (DLA-BDTSS) is proposed. DLA-BDTSS improved outcomes by 8.43% with an execution time of 34 s and a fitness value evaluation of 76.8%.
The outcomes of DLA-BDTSS are improved to 8.43%with an execution time of 34 s and fitness value evaluation of 76.8%. More research is needed to determine the optimal production parameters, for example the optimal number of terminals, vehicles, and workers. In order to create more efficient states and activities, or to supply the network with characteristics and actions based on the results of training and testing conducted with big data analytics, more study of the system’s architecture is required. In the future of the Internet of Things and cloud computing, an optimal scheduling algorithm should be designed to optimize CPU utilization, throughput, latency, and energy efficiency. It is standard practice in IoT applications to have numerous sensor nodes conduct the same operation repeatedly, which can increase the sensing cost and lower the network lifetime.