Next Article in Journal
Design and Experiment of Dry-Ice Cleaning Mechanical Arm for Insulators in Substation
Next Article in Special Issue
Optimal Design and Control of MMC STATCOM for Improving Power Quality Indicators
Previous Article in Journal
Multi-Criteria Efficiency Analysis of Using Waste-Based Fuel Mixtures in the Power Industries of China, Japan, and Russia
Previous Article in Special Issue
Comparison of Various Mother Wavelets for Fault Classification in Electrical Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Managing Energy Plus Performance in Data Centers and Battery-Based Devices Using an Online Non-Clairvoyant Speed-Bounded Multiprocessor Scheduling

1
Department of Computer Science and Engineering, Amity School of Engineering and Technology, Amity University Uttar Pradesh, Lucknow Campus, Lucknow 226010, India
2
Department of Electrical and Computer Engineering, Hawassa University, Hawassa P.O. Box 05, Ethiopia
3
Power System Planning Division, Rajasthan Rajya Vidhyut Prasaran Nigam Ltd., Jaipur 302005, India
4
Department of Electrical Power Engineering, Faculty of Mechanical and Electrical Engineering, Tishreen University, 2230 Lattakia, Syria
5
Department of Electrical Power Engineering, Dresden University, 01069 Dresden, Germany
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(7), 2459; https://doi.org/10.3390/app10072459
Submission received: 5 March 2020 / Revised: 22 March 2020 / Accepted: 30 March 2020 / Published: 3 April 2020

Abstract

:
An efficient scheduling reduces the time required to process the jobs, and energy management decreases the service cost as well as increases the lifetime of a battery. A balanced trade-off between the energy consumed and processing time gives an ideal objective for scheduling jobs in data centers and battery based devices. An online multiprocessor scheduling multiprocessor with bounded speed (MBS) is proposed in this paper. The objective of MBS is to minimize the importance-based flow time plus energy (IbFt+E), wherein the jobs arrive over time and the job’s sizes are known only at completion time. Every processor can execute at a different speed, to reduce the energy consumption. MBS is using the tradition power function and bounded speed model. The functioning of MBS is evaluated by utilizing potential function analysis against an offline adversary. For processors m ≥ 2, MBS is O(1)-competitive. The working of a set of jobs is simulated to compare MBS with the best known non-clairvoyant scheduling. The comparative analysis shows that the MBS outperforms other algorithms. The competitiveness of MBS is the least to date.

1. Introduction

There are number of server farms equipped with hundreds of processors. The cost of energy used for cooling and running a machine for around three years surpasses the hardware cost of the machine [1]. Consequently, the major integrated chips manufacturers such as Intel and AMD are producing the dynamic speed scaling (DSS) enabled multiprocessor/multi-core machine and software such as Intel’s SpeedStep [2], which support the operating system in managing the energy by varying the execution speed of processors. A founder chip maker Tilera forecasted that the numbers of processors/cores will be doubled every eighteen months [3], which will increase the energy demand to a great extent. Data centers consume 1.5% of total electricity usage in United States [4]. To avoid such critical circumstances, the current issue in the scheduling is to attain the good quality of service by generating an optimal schedule of jobs and to save the energy consumption, which is a conflicting and complicated problem [5].
The power P consumed by a processor running at speed s is s V 2 , where V is a voltage [6]. The traditional power function is P = s α ( α 2 for CMOS based chips [7,8]). There are two types of speed models: the first unbounded speed model, in which the processor’s speed range is, i.e., [ 0 ,   ) ; the second bounded speed model, in which the speed of a processor can range from zero to some maximum speed, i.e., [ 0 ,   η ] . This DSS plays a vital role in energy management, where in a processor can regulate its speed to save energy. A few qualities of service metrics are slowdown, throughput, makespan, flow time and weighted flow time. At low speed, the processor finishes jobs slower and save energy, whereas at high speed, the processor finishes jobs faster but consumes more energy, as shown in Figure 1. To get a better quality of service and low energy consumption the objective should be to minimize the sum of flow time and energy; in case, if the importance or priority is attached, the objective should be to minimize the sum of importance-based flow time and energy. The objective of minimizing the IbFt+Ehas a natural explanation, as it can be considered in monetary terms [9].
In the multiprocessor systems, there is a requirement of three different policies: the first policy is job selection, which decides the next job to be executed on every processor; the second policy is speed scaling, which decides every processor’s execution speed at all time; the third policy is job assignment, which indicates that to which processor the new job should be assigned. In the c-competitive online scheduling algorithm, for each input the cost received is less than or equal to c times the cost of optimal offline algorithm [9]. Unlike non-clairvoyant scheduling, the size of job is unknown at arrival time, such as in UNIX operating system where jobs arrive with no information of processing requirement. Unlike online modes, in the offline mode, the whole job progression is known in advance. No online algorithm can attain a constant competitiveness with equal maximum speed to optimal offline algorithm [10].
Motwani et al. [10] commenced the study of the non-clairvoyant scheduling algorithms. Yao et al. inducted the theoretical study of speed scaling scheduling algorithm [11]. Yao et al. proposed an algorithm average rate heuristic (AVR) with a competitive ratio at most 2 α 1 α α using the traditional power function. Koren et al. [12] presented an optimal online scheduling algorithm D o v e r for a overloaded uniprocessor system with competitive ratio- ( 1 ( 1 + k ) 2 ) for the objective of minimizing the throughput, where k is the importance ratio. The competitiveness of shortest remaining processing time (SRPT) for multiprocessor system is O ( m i n ( l o g ( m n ) ,   l o g   σ ) ) , where m is number of processors, n is total number of jobs and σ represents the ratio of minimum to maximum job size [13]. Kalyanasundaram et al. [14] presented the idea of resource augmentation. If the resources are augmented and, ( 2 + Δ ) -speed p processors are used then the competitive ratio of Equi-partition lies between 2 3 ( 1 + Δ ) and ( 2 + 4 Δ ) [15]. Multilevel feedback queue, a randomized algorithm with n jobs is O ( l o g   n ) -competitive [16,17]. The first algorithm with non trivial guarantee is O ( log 2 σ ) -competitive [18], where σ is the ratio of minimum to maximum job size. There are different algorithms proposed with different objectives over a span of time [19,20,21,22,23,24,25,26,27].
Chen et al. [19] proposed algorithms with different approximation bounds for processors with/without constraints on the maximum processor speed. The concept of merging dual objective of energy used and total flow time into single objective of energy used plus total flow time is proposed by Albers et al. [20]. Bansal et al. [21] proposed an algorithm, which uses highest density first (HDF) for the job selection with a traditional power function. Lam et al. [22] proposed a multiprocessor algorithm for homogeneous processors in which job assignment policy is a variant of round robin, the job selection. Random dispatching can provide ( 1 + Δ ) -speed O ( 1 Δ 3 ) -competitive non-migratory algorithm [23]. Chan et al. [24] proposed an O ( 1 ) -competitive algorithm using sleep management for the objective of minimizing the flow time plus energy. Albers et al. [25] studied an offline problem in polynomial time and proposed a fully combitorial algorithm that relies on repeated maximum flow computation. Gupta et al. [26] proved that highest density first, weighted shortest elapsed time first and weighted late arrival processor sharing are not O ( 1 ) -speed O ( 1 ) -competitive for the objective of minimizing the weighted flow time even in fixed variable speed processors for heterogeneous multiprocessor setting. Chan et al. [27] studied an online clairvoyant sleep management algorithm scheduling with arrival-time-alignment (SATA) which is ( 1 + Δ ) -speed O ( 1 Δ 2 ) -competitive for the objective of minimizing the flow time plus energy. For a detailed survey refer to [28,29,30,31,32,33,34].
In this paper, the problem of online non-clairvoyant (ON-C) DSS scheduling is studied and an algorithm multiprocessor with bounded speed (MBS) is proposed with an objective to minimize the IbFt+E. On the basis of potential function analysis MBS is O(1)- competitive. The notations used in this paper are mentioned in the Table 1.
The organization of the paper is as follows. In Section 2, some related non-clairvoyant algorithms are explained and their competitive values are compared to the proposed algorithm MBS. Section 3 presents the preliminary definition and information for the proposed work. In Section 4, the proposed algorithm, its flow chart and potential function analysis is presented. The processing of a set of jobs are simulated using MBS and the best identified algorithm to observe the working of MBS. Section 6 provides the conclusion and future scope of the work.

2. Related Work

Gupta et al. [35] gave an online clairvoyant scheduling algorithm GKP (proposed by Gupta, Krishnaswamy and Pruhs) for the objective of minimizing the weighted flow time plus energy. Under the traditional power function, GKP is O ( α 2 ) -competitive without a resource augmentation for power heterogeneous processors. GKP uses highest density first (HDF) for the selection of jobs on each processor; the speed of any processor scales such that the power of a processor is the fractional weight of unfinished jobs; jobs are assigned in such a way that it gives the least increase in the projected future weighted flow time. Gupta et al. [35] used a local competitiveness analysis to prove their work. Fox et al. [36] considered the problem of scheduling the parallelizable jobs in the non-clairvoyant speed scaling settings for the objective of minimizing the weighted flow time plus energy and they used the potential function analysis to prove it. Fox et al. presented weighted latest arrival processor sharing with energy (WLAPS+E), which schedules the late arrival jobs and every job use the same number of machines proportioned by the job weight. WLAPS+E spares some machines to save the energy. WLAPS+E is ( 1 + 6 Δ ) -speed ( 5 Δ 2 ) -competitive, where 0 < Δ 1 6 . Thang [37] studied the online clairvoyant scheduling problem for the objective of minimizing the weighted flow time plus energy in the unbounded speed model and using the traditional power function. Thang gave an algorithm (ALGThang) on unrelated machines and proved that ALGThang is 8 ( 1 + α l n α ) -competitive. In AlGThang, the speed of any processor depends on the total weight of pending jobs on that machine, and any new job is assigned to a processor that minimizes the total weighted flow time.
Im et al. [38] proposed an ON-C scheduling algorithm SelfishMigrate-Energy (SM-E) for the objective of minimizing the weighted flow time plus energy for the unrelated machines. Using the traditional power function SM-E is O ( α 2 ) -competitive. In SM-E, a virtual queue is maintained on every processor where the new or migrated jobs are added at tail; the jobs migrate selfishly until equilibrium is gained. Im et al. simulates sequential best response (SBR) dynamics and they migrates each job to the machine that is provided by the Nash equilibrium. The scheduling policy applied on every processor is a variant of weighted round robin (WRR), wherein the larger speed is allotted to jobs residing at the tail of the queue (like Latest Arrival Processor Sharing (LAPS) and Weighted Latest Arrival Processor Sharing (WLAPS)). Bell et al. [39] proposed an online deterministic clairvoyant algorithm dual-classified round robin (DCRR) for the multiprocessor system using the traditional power function. The motive of ( 2 4 α ( l o g α P + α α 2 α 1 ) ) -competitive DCRR is to schedule the jobs so that they can be completed within deadlines using minimum energy, i.e., the objective is to maximize the throughput and energy consumption. In DCRR, the sizes and the maximum densities (= size/(deadline – release time)) of jobs are known and the classification of jobs depends on the size and the maximum density both. The competitive ratio of DCRR is high, as it considers the jobs with deadlines and using a variation of round robin with the speed scaling.
Azar et al. [40] gave an ON-C scheduling algorithm NC-PAR (Non-Clairvoyant for Parallel Machine) for the identical parallel machines, wherein the job migration is not permitted. Using traditional function NC-PAR is ( α + 1 α 1 ) -competitive for the objective of minimizing the weighted flow time plus energy in unbounded speed model. In NC-PAR a global queue of unassigned jobs is maintained in First In First Out (FIFO) order. A new job is assigned to a machine, when a machine becomes free. In NC-PAR jobs are having uniform density (i.e., w e i g h t / s i z e = 1 ) and the jobs are not immediately allotted to the processors at release time. The speed of a processor using NC-PAR is based on the total remaining weight of the active jobs. In non-clairvoyant model with known arbitrary weights no results are known [40].
An ON-C multiprocessor speed scaling scheduling algorithm MBS is proposed and studied against an offline adversary with an objective of minimizing IbFt+E. The speed of a processor using MBS is proportional to the sum of importance of all active jobs on that processor. In MBS, the processor’s maximum speed can be ( 1 + Δ 3 m ) η (i.e., the range of speed is from zero to ( 1 + Δ 3 m ) η ), whereas the processor’s maximum speed using Opt (Optimal algorithm) is η , where m is number of processors and 0 < Δ ( 3 α ) 1 a constant. In MBS, a new job is assigned to an idle processor (if available) or to a processor having the minimum sum of the ratio of importance and executed size for all jobs on that processor; the policy for job selection is weighted/importance-based round robin, and each active job receives the processor speed equal to the ratio of its importance to the total importance of jobs on that processor. In this paper, the performance of MBS is analysed using a competitive analysis, i.e., the worst-case comparison of MBS and optimal offline scheduling algorithm. MBS is ( 1 + Δ 3 m ) -speed, ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) = O ( 1 ) competitive, i.e., the value for competitive ratio c for m = 2 , α = 2 is 2.442; for m = 2 , α = 3 is 2.399; the detailed results for different values of m, Δ = ( 3 α ) 1 and α = 2   &   3 is shown in Table 2. The comparison of results is given along with the summary of results in Table 3.
On the basis of the values mentioned in the Table 2, it can be observed that in proposed algorithm MBS if the number of processor increases then the speed ratio and competitive ratio increases. The data mentioned in Table 3 describe the competitive values of different scheduling algorithm. Some clairvoyant and non-clairvoyant algorithms competitive ratio are considered at α = 2 , α = 3 . The lower competitive value represents the better algorithm. The value of competitiveness is least for the proposed algorithm MBS.

3. Definitions and Notations

An ON-C job scheduling on a multiprocessor using speed bounded setting is considered, where the jobs arrive over time, the job’s importance/weight are known at release time and the size of a job is revealed only after the job’s completion. Processor’s speed using Opt can vary dynamically from 0 to the maximum speed η i.e., [ 0 ,   η ] . The nature of jobs is sequential as well as unrestricted pre-emption is permitted without penalty. The traditional power function P o w e r   P = s p e e d α is considered, where α > 1 a fixed constant. If s is the processor’s speed then a processor executes s unit of work per unit time. An active job j has release time lesser than the current time t, and it is not completely executed. The flow time F ( j ) of job j is the time duration since j released and until it is completed. The total importance-based flow time F is j I i m p ( j ) F ( j ) . Amortized analysis is used for algorithms where an occasional operation is very slow, but most of the other operations are faster. In amortized analysis, we analyse a sequence of operations and guarantee a worst case average time which is lower than the worst case time of a particular expensive operation.

4. Methodology

In this study, the amortized potential function analysis of the objective is used to examine the performance of the proposed algorithm. Amortized analysis is a worst-case analysis of a sequence of operations—to obtain a tighter bound on the overall or average cost per operation in the sequence than is obtained by separately analyzing each operation in the sequence. The amortized potential method, in which we derive a potential function characterizing the amount of extra work we can do in each step. This potential either increases or decreases with each successive operation, but cannot be negative. The objective of study is to minimize the total IbFt+E, denoted by G = F + E. It reflects that the target is to minimize the quality of service and energy consumed. The input to the problem is the set of jobs I. A scheduler generates the schedule S of jobs in I. The total energy consumption E for the scheduling is 0 s ( t ) α   d t . Let Opt be an optimal offline algorithm such that for any job sequence I, IbFt+E F O p t ( I ) + E O p t ( I ) of Opt is minimized among all schedule of I. The notations used in MBS are mentioned in the Table 1. Any online algorithm ALG is said to be c-competitive for c ≥ 1, if for all job sequences I and any input the cost incurred is never greater than c times the cost of optimal offline algorithm Opt, and the following inequality is satisfied:
( F A L G ( I ) + E A L G ( I ) ) c · ( F O p t ( I ) + E O p t ( I ) )
The traditional power function is utilized to simulate the working of the proposed algorithm and compare the effectiveness by comparing with the available best known algorithm. The jobs are taken of different sizes and the arrival of jobs is considered in different scenario to critically examine the performance of the proposed algorithm. Different parameters (such as IbFt, IbFt+E, speed of processor and speed growth) are considered to evaluate the algorithm.

5. An O ( 1 ) -Competitive Algorithm

An ON-C multiprocessor scheduling algorithm multiprocessor with bounded speed (MBS) is explained in this section. The performance of MBS is observed by using potential function analysis, i.e., the worst-case comparison of MBS with an offline adversary Opt. The competitiveness of MBS is O ( 1 ) with an objective to minimize the IbFt+E for m processors with the highest speed ( 1 + Δ 3 m ) η .

5.1. Multiprocessor with Bounded Speed Algorithm: MBS

At time t, the processing speed of u adjusts to s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( ( i m p u a ( t ) Ϯ ) 1 α ,   η ) , where 0 < Δ ( 1 3 α ) , Ϯ 1 and α 2 are constants. The importance i m p ( j ) of a job is uninformed and acknowledged only at release time r ( j ) . The policies considered for the multiprocessor scheduling MBS are as follows:
Job selection policy: The importance-based/weighted round robin is used on every processor.
Job assignment policy: a newly arrived job is allotted to an idle processor (if available) or to a processor having the minimum sum of the ratio of importance to the executed size for all jobs on that processor (i.e., m i n   f = 1 n u a ( i m p u ( j f ) e x s u ( j f ) ) ).
Speed scaling policy: The speed of every processor is scaled on the bases of the total importance of active jobs on that processor. Every active job j i on u obtains the fraction of speed:
p r o c e s s o r s   s p e e d ( i m p o r t a n c e   o f j i   t o t a l   i m p o r t a n c e   o f   a l l   a c t i v e   j o b s   o n   t h a t   p r o c e s s o r )
i.e., s u a · ( i m p u ( j i ) k = 1 n u a i m p u ( j k ) ) or s u a · ( i m p u ( j i ) i m p u a ) . The speed of any processor gets adjusted (re-evaluated) on alteration in total importance of active jobs on that processor. MBS is compared against an optimal offline algorithm Opt, using potential function analysis. The principal result of this study is stated in Theorem 1. The Algorithm 1 of MBS is given next and the flow chart for MBS is given in Figure 2.
Algorithm 1: MBS (Multiprocessor with Bounded Speed)
Input: total m number of processors { u 1 , ,   u k , ,   u m } , n a NoAJ { j 1 , ,   j i ,   ,   j n a } and the importance of all n a active jobs { imp ( j 1 ) , ,   imp ( j i ) , ,   imp (   j n a ) } .
Output: number of jobs allocated to every processor, the speed of all processors, at any time and execution speed share of each active job.
Repeat until all processors become idle:
1. If any job j i arrives
2. if m n a
3. allocate job j i to a idle processor u
4. otherwise, when m < n a
5. allocate job j i to a processor u with m i n   f = 1 n ua ( imp u ( j f ) exs u ( j f ) )
6. imp ua = imp ua + imp u ( j i )
7. s ua = ( 1 + Δ 3 m ) · m i n ( ( imp ua Ϯ ) 1 α ,   η ) , where 0 < Δ ( 1 3 α ) and Ϯ > 1 is a constant value
8. Otherwise, if any job j i completes on any processor u and other active jobs are available for execution on that processor then
9. imp ua = imp ua imp u ( j i )
10. s ua = ( 1 + Δ 3 m ) · m i n ( ( imp ua Ϯ ) 1 α ,   η ) , where 0 < Δ ( 1 3 α ) and Ϯ 1 is a constant value
11. the speed received by any job j i , which is executing on a processor u, is s ua · ( imp u ( j i ) imp ua )
12. otherwise, processors continue to execute remaining jobs
Theorem 1.
When using more than two processors ( i . e . ,   m 2 ) and each processor has the permitted maximum speed ( 1 + Δ 3 m ) η , MBS is c-competitive for the objective of minimizing the IbFt+E, where c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) = O ( 1 ) and 0 < Δ 1 3 α .

5.2. Necessary Conditions to be Fulfilled

A potential function is needed to calculate the c-competitiveness of an algorithm. An algorithm is called c-competitive if at any time t, the sum of augmentation in the objective cost of algorithm and the modification in the value of potential is at the most c times the augmentation in the objective cost of the optimal adversary algorithm. A potential function Φ ( t ) is required to demonstrate that MBS is c-competitive. A c-competitive algorithm should satisfy the conditions:
Boundary Condition: The value of potential function is zero before the release of any job and after the completion of all jobs.
Job Arrival and Completion Condition: The value of potential function remains same on arrival or completion of a job.
Running Condition: At time when the above condition do not exist, the sum of the (rate of change) RoC of G a and the RoC of Φ is at the most c times the RoC of G o .
d G a ( t ) d t + γ · d Φ d t c · d G o ( t ) d t ,   where   γ > 0 .

5.3. Potential Function Φ ( t )

An active job j is lagging, if ( p w k a ( j ,   t ) p w k o ( j ,   t ) ) > 0 . Since t is the instantaneous time, this factor is dropped from the rest of the analysis. For any processor u, let L G u = { j 1 ,   j 2 ,   ,   j l g u } be a group of lagging jobs using MBS and these jobs are managed in the ascending order of latest time (when any job gets changed into lagging job). L G = u = 1 m L G u is a set of all lagging jobs on all m processors. Further, i m p l g u = i = 1 l g u i m p u ( j i ) is the sum of the importance of lagging jobs on a processor u. Following this, i m p l g = u = 1 m i m p l g u is the sum of the importance of lagging jobs on all m processors. Our potential function Φ ( t ) for IbFt+E is the addition of all potential values of m processors.
Φ ( t ) = u = 1 m Φ u ( t )
Φ u ( t ) = { i = 1 l g u ( ( k = 1 i i m p u ( j k ) ) 1 2 δ ) · ω i i f   k = 1 i i m p u ( j k )   η 1 2 δ i = 1 l g u ( 1 1 δ ) · ( k = 1 i i m p u ( j k ) · η 1 ) · ω i o t h e r w i s e
Where ω i = m a x { 0 ,   ( p w k a ( j i ,   t ) p w k o ( j i ,   t ) ) } δ = 1 2 α   and
( k = 1 i i m p u ( j k ) ) 1 2 δ   and   ( 1 1 δ ) · ( k = 1 i i m p u ( j k ) · η 1 )
are the coefficients ci of ji on processor u
MBS is analyzed per machine basis. Firstly, the verification of boundary condition: the value of Φ is zero after finishing of all jobs and prior to release of any job on any processor. There will be no active job on any processor in both situations. Therefore, the boundary condition is true. Secondly, the verification of arrival and completion condition: at time t, on release of a new job j i in I, j i without execution is appended at end of I. ω i is zero as p w k a ( j i ,   t ) p w k o ( j i ,   t ) = 0 . The coefficient of all other jobs does not change and Φ remains unchanged. At the time of completion of a job j i , ω i becomes zero and other coefficients of lagging jobs either remains unchanged or decreases, so, Φ does not increase. Thus the arrival and completion criteria holds true. The third and last criterion to confirm is running condition, with no job arrival or completion.
According to previous discussion, for any processor u, let d G u a d t = i m p u a + s u a α and d G u o d t = i m p u o + s u o α be the alteration of IbFt+E in an infinitesimal period of time [ t ,   t + d t ] by MBS and Opt, respectively. The alteration of Φ because of Opt and MBS in an infinitesimal period of time [ t ,   t + d t ] by u is d Φ u o d t and d Φ u a d t , respectively. The whole alteration in Φ because of Opt and MBS in infinitesimal period of time [ t ,   t + d t ] by u is d Φ u d t = d Φ u o d t + d Φ u a d t . As this is multiprocessor system therefore to bound the RoC of Φ by Opt and MBS, the analysis is divided in two cases based on n a and m , and then every case is further divided in three sub cases depending on whether i m p u a > η α and i m p l u > η α , afterwards each sub case is further divided in two sub cases depending on i m p l g > ( i m p a ( 3 3 + Δ ) · i m p a ) and i m p l g ( i m p a ( 3 3 + Δ ) · i m p a ) , where 0 < Δ < 1 , μ = 3 3 + Δ . The potential analysis is done on individual processor basis, the reason behind it is that all the processors will not face the same case at the same time; rather different processors may face same or different cases.
Lemma 1.
For the positive real numbers x, y, A and B, if x 1 + y 1 = 1 holds then [2]:
x 1 · A x + y 1 · B y A · B
Lemma 2.
If n a m and i m p l g u η α
(a) d Φ u o d t s u o α α + ( 1 2 δ ) · i m p l g u ; (b) d Φ u a d t ( s u a · i m p l g u 1 2 δ )
Proof. 
If n a m then every processor executes not more than one job, i.e., every job is processed on individual processor.
(a) It is required to upper-bound d Φ u o d t for a processor u. To calculate the upper-bound, the worst-case is considered which occurs if Opt executes a job on u with the largest coefficient c l g u = i m p l g u 1 2 δ . At this time, ω i increases at the rate of s u o (because of Opt on u). The count of lagging jobs on some u may be only one.
d Φ u o d t c l g u · s u o i m p l g u 1 2 δ · s u o
Using Young’s inequality, Lemma 1 (Equation (6)) in (7) such that A = s u o , B = ( i m p l g u ) 1 2 δ , x = α and y = 1 1 2 δ we have:
d Φ u o d t s u o α α + ( 1 2 δ ) · i m p l g u
(b) Next, it is required to upper-bound d Φ u a d t for a processor u. To compute the upper-bound, consider that a lagging job j i on u is executed at the rate of ( s u a · i m p u ( j i ) k = 1 n u a i m p u ( j k ) ) or ( s u a · i m p u ( j i ) i m p u a ) , therefore, the change in ω i is at the rate of ( s u a · i m p u ( j i ) i m p u a ) .
d Φ u a d t = i = 1 l g u ( ( k = 1 i i m p u ( j k ) ) 1 2 δ ) · ( s u a · i m p u ( j i ) i m p u a )
As only one job executes on a processor, therefore i m p u ( j i ) i m p u a = 1 and l g u = i = 1 ,
d Φ u a d t = ( i m p l g u 1 2 δ ) · ( s u a ) d Φ u a d t = ( s u a · i m p l g u 1 2 δ )
Lemma 3.
If n a m and i m p l g u > η α
(a) d Φ u o d t ( 1 1 δ ) · i m p l g u ; (b) d Φ u a d t = ( 1 + Δ 3 m ) ( 1 δ ) · i m p l g u
Proof. 
If n a m then every processor executes not more than one job, i.e., every job is processed on individual processor.
(a) It is required to upper-bound d Φ u o d t for a processor u. To calculate the upper-bound, the worst-case is considered which occurs if Opt executes a job on u with the largest coefficient c l g u = ( 1 1 δ ) · i m p l g u · η 1 . At this time, ω i increases at the rate of s u o (because of Opt on u) where s u o η . The count of lagging jobs on any u may be only one.
d Φ u o d t c l g u · s u o c l g u · η = ( 1 1 δ ) · i m p l g u · η 1 · η d Φ u o d t ( 1 1 δ ) · i m p l g u
(b) Next, it is required to upper-bound d Φ u a d t for a processor u. To compute the upper-bound, consider that a lagging job j i on u is executed at the rate of ( s u a · i m p u ( j i ) k = 1 n u a i m p u ( j k ) ) or ( s u a · i m p u ( j i ) i m p u a ) , therefore the change in ω i is at the rate of ( s u a · i m p u ( j i ) i m p u a ) . i m p u a i m p g l u > η α , s u a = ( 1 + Δ 3 m ) · η
d Φ u a d t = i = 1 l g u ( 1 1 δ ) · ( k = 1 i i m p u ( j k ) · η 1 ) · ( s u a · i m p u ( j i ) i m p u a )
As only one job executes on a processor, therefore i m p u ( j i ) i m p u a = 1 and l g u = i = 1 ,
d Φ u a d t = ( 1 1 δ ) · ( i m p l g u · η 1 ) · ( s u a ) = ( 1 1 δ ) · ( s u a · i m p l g u · η 1 ) = ( 1 1 δ ) · ( ( 1 + Δ 3 m ) · η · i m p l g u · η 1 )
d Φ u a d t = ( 1 + Δ 3 m ) ( 1 δ ) · i m p l g u
Lemma 4.
If n a > m and i m p l g u η α
(a) d Φ u o d t s u o α α + ( 1 2 δ ) · i m p l g u ; (b) d Φ u a d t   s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
Proof. If n a > m then:
(a) It is required to upper-bound d Φ u o d t for a processor u. To calculate the upper-bound, the worst-case is considered which occurs if Opt is executing a job on u with the largest coefficient c l g u = i m p l g u 1 2 δ . At this time, ω i increases at the rate of s u o (because of Opt on u).
d Φ u o d t c l g u · s u o = i m p l g u 1 2 δ · s u o
Using Young’s inequality, Lemma 1 (Equation (6)) in (12) such that A = s u o , B = i m p l g u 1 2 δ , x = α and y = 1 1 2 δ we have:
d Φ u o d t s u o α α + ( 1 2 δ ) · i m p l g u
(b) Next, it is required to upper-bound d Φ u a d t for a processor u, to compute the upper-bound consider that a lagging job j i on u is executed at the rate of ( s u a · i m p u ( j i ) k = 1 n u a i m p u ( j k ) ) or ( s u a · i m p u ( j i ) i m p u a ) , therefore the change in ω i is at the rate of ( s u a · i m p u ( j i ) i m p u a ) . To make the discussion straightforward, let h u i = k = 1 i i m p u ( j k ) , h u 0 = 0 , h u l g u = i m p l g u and i m p u ( j i ) = h u i h u i 1 . (by using Equation (3):
d Φ u a d t = i = 1 l g u ( ( k = 1 i i m p u ( j k ) ) 1 2 δ ) · ( s u a · i m p u ( j i ) i m p u a ) =   s u a i m p u a i = 1 l g u ( ( h u i ) 1 2 δ ) · ( h u i h u i 1 )   s u a i m p u a i = 1 l g u h u i 1 h u i f 1 2 δ   d f   s u a i m p u a 0 h u l g u f 1 2 δ   d f =   s u a i m p u a · h u l g u 2 2 δ ( 2 2 δ ) =   s u a i m p u a · i m p l g u 2 2 δ ( 2 2 δ )
d Φ u a d t   s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
Lemma 5.
If n a > m and   i m p l g u > η α
(a) d Φ u o d t ( 1 1 δ ) · i m p l g u ; (b) d Φ u a d t ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a )
Proof. 
If n a > m then:
(a) It is required to upper-bound d Φ uo dt for a processor u. To calculate the upper-bound, the worst-case is considered which occurs if Opt executes a job on u with the largest coefficient c lg u = ( 1 1 δ ) · imp lg u · η 1 (as imp ua   imp lg u > η α ). At this time, ω i increases at the rate of s u o (because of Opt on u).
d Φ u o d t c l g u · s u o = ( 1 1 δ ) · i m p l g u · η 1 · s u o   ( 1 1 δ ) · i m p l g u · η 1 · η { s u o η }
d Φ u o d t ( 1 1 δ ) · i m p l g u
(b) Next, it is required to upper-bound d Φ u a d t for a processor u. To compute the upper-bound, consider that a lagging job j i on u is executed at the rate of ( s u a · i m p u ( j i ) k = 1 n u a i m p u ( j k ) ) or ( s u a · i m p u ( j i ) i m p u a ) , therefore the change in ω i is at the rate of ( s u a · i m p u ( j i ) i m p u a ) . To make the discussion uncomplicated, let h u i = k = 1 i i m p u ( j k ) , h u 0 = 0 , h u l g u = i m p l g u > η α , i m p u a   i m p l g u > η α and i m p u ( j i ) = h u i h u i 1 . Let z < l u be the largest integer such that h u z η α . (using Equation (3)):
d Φ u a d t = i = 1 l g u c i · ( s u a · i m p u ( j i ) i m p u a ) = (   s u a i m p u a ) · ( i = 1 z ( i m p u ( j i ) · ( h u i ) 1 2 δ ) + i = z + 1 l g u ( ( 1 1 δ ) · ( i m p u ( j i ) · h u i · η 1 ) ) ) (   s u a i m p u a ) · ( ( 0 h u z f 1 2 δ d f ) + ( ( 1 1 δ ) · η 1 · ( h u z h u l g u f d f ) ) ) = (   s u a i m p u a ) · ( h u z 2 2 δ ( 2 2 δ ) + h u l g u 2 h u z 2 ( 2 2 δ ) · η ) = (   s u a i m p u a ) · ( h u z 2 ( 2 2 δ ) h u z 1 α + h u l g u 2 h u z 2 ( 2 2 δ ) · η )   ( 1 + Δ 3 m ) · η i m p u a · ( h u z 2 ( 2 2 δ ) η + h u l g u 2 h u z 2 ( 2 2 δ ) · η ) { h u z η α   } =   ( 1 + Δ 3 m ) i m p u a · h u l g u 2 ( 2 2 δ ) =   ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a )
d Φ u a d t ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a )
Lemma 6. At all time t, when Φ does not comprise discrete alteration dG ua dt + γ · d Φ u dt c · dG uo dt , where c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) . Assume that γ = 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) .
Proof. 
The analysis is divided in two cases based on n a > m or n a m , and then each case is again alienated in three sub-cases depending on whether i m p u a > η α or i m p u a η α and i m p l g u > η α or i m p l g u η α , afterwards each sub-case is again alienated in two sub-cases depending on whether i m p l g u   > ( i m p u a ( 3 3 + Δ ) · i m p u a ) or i m p l g u   ( i m p u a ( 3 3 + Δ ) · i m p u a ) , where 0 < μ = ( 3 3 + Δ ) < 1 and Δ = ( 1 3 α ) . As a job in MBS which is not lagging must be an active job in Opt,
i m p u o   i m p u a i m p l g u i m p u a ( i m p u a μ · i m p u a ) μ   i m p u a i m p u a   i m p u o   μ
μ = ( 3 3 + Δ )
γ = 1 16 · ( 1 + ( 1 + Δ 3 m ) α )
c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α )
Case I: When n a m and i m p u a η α , since i m p l g u i m p u a we have i m p l g u η α , and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) · i m p u a 1 α .
(a) If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(using Equations (8) and (9))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u ) ( s u a · i m p l g u 1 2 δ )
(by using Equations (1) and (21))
d G u a d t + γ · d Φ u d t
( i m p u a + s u a α + γ · ( ( s u o α α + ( 1 2 δ ) · i m p l g u ) ( s u a · i m p l g u 1 2 δ ) ) )
= ( i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u γ · ( 1 + Δ 3 m ) · i m p u a 1 α · i m p l g u 1 2 δ )
( γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α ) · i m p u a + γ · ( 1 2 δ ) · i m p u a γ · ( 1 + Δ 3 m ) · i m p u a 1 α · ( ( 1 ( 3 3 + Δ ) ) · i m p u a ) 1 2 δ )
= ( γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) γ · ( 1 + Δ 3 m ) · ( Δ 3 + Δ ) 1 2 δ ) )
γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( Δ 3 + Δ ) 1 2 δ )
γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( Δ 3 + Δ ) )
= γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 3 3 + Δ ) )
γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ )  
= γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) )   (by using Equation (19))
γ α · s u o α + i m p u o   μ · ( 17 16 · ( 1 + ( 1 + Δ 3 m ) α ) )   (by using Equation (17))
= γ α · s u o α + i m p u o · ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) )   (by using Equation (18))
γ α · s u o α + i m p u o   · ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) )   (by using Equation (20))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u o   · c
Since c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) and γ = 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) , we have
c = ( 9 8 + 3 Δ 8 ) · 16 γ γ c = 1 18 + 6 Δ < 1 γ c < 1 γ < c
Since   γ < c   and   α > 1 1 > 1 α γ α < c
(by using Equation (23) in Equation (22))
d G u a d t + γ · d Φ u d t c · s u o α + c · i m p u o   = c · ( s u o α + i m p u o   ) = c · d G u o d t
Hence the running condition is fulfilled for n a m , i m p u a η α , i m p l g u η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(b) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 .
(by using Equation (8))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u )
(by using Equations (1) and (24))
d G u a d t + γ · d Φ u d t i m p u a + s u a α + γ · ( s u o α α + ( 1 2 δ ) · i m p l g u )
= i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u
γ α · s u o α + i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 2 δ ) · i m p u a
= γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) )
γ α · s u o α + i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) ) (by using Equation (17))
γ α · s u o α + ( 1 μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + γ ) ) · i m p u o
= γ α · s u o α + ( 1 μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) ) · i m p u o (by using Equation (19))
= γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o
c · s u o α + c · i m p u o (by using Equations (18) and (23))
d G u a d t + γ · d Φ u d t c · ( s u o α + w u o )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is satisfied for n a m , i m p u a η α , i m p l g u η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
Case II: When n a m , i m p u a > η α , i m p l g u η α , and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) η .
(a) If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(by using Equations (8) and (9))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u ) ( s u a · i m p l g u 1 2 δ )
(by using Equations (1) and (25))
d G u a d t + γ · d Φ u d t
i m p u a + s u a α + γ · ( ( s u o α α + ( 1 2 δ ) · i m p l g u ) ( s u a · i m p l g u 1 2 δ ) )
= i m p u a + ( 1 + Δ 3 m ) α · η α + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u γ · ( 1 + Δ 3 m ) · η · i m p l g u 1 2 δ
i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p u a γ · ( 1 + Δ 3 m ) · i m p l g u 2 δ · i m p l g u 1 2 δ
= γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) ) · i m p u a γ · ( 1 + Δ 3 m ) · i m p l g u
γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + γ ) · i m p u a γ · ( 1 + Δ 3 m ) · ( 1 ( 3 3 + Δ ) ) · i m p u a
γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( Δ 3 + Δ ) ) · i m p u a
= γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + γ · ( 3 3 + Δ ) ) · i m p u a
γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + γ ) · i m p u a
γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o   μ (by using Equations (17) and (19))
= γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o   (by using Equation (18))
γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o   (by using Equations (20) and (23))
c · s u o α + c · i m p u o  
= c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is fulfilled for n a m , i m p u a > η α , i m p l g u η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(b) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 . (by using Equation (7))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u )
(by using Equations (1) and (26))
d G u a d t + γ · d Φ u d t i m p u a + s u a α + γ · ( s u o α α + ( 1 2 δ ) · i m p l g u )
= i m p u a + ( 1 + Δ 3 m ) α · η α + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u
i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u
γ α · s u o α + i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 2 δ ) · i m p u a
= γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) )
γ α · s u o α +   i m p u o μ · ( 1 + ( 1 + Δ 3 m ) α + γ ) (by using Equation (17))
= γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) ·   i m p u o μ (by using Equation (19))
= γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o
c · s u o α + c · i m p u o (by using Equations (20) and (23))
= c · ( s u o α + i m p u o )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is satisfied for n a m , i m p u a > η α , i m p l g u η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
Case III: When n a m , i m p u a > η α , i m p l g u > η α , and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) η   .
(a) If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(by using Equations (10) and (11))
d Φ u d t ( 1 1 δ ) · i m p l g u ( 1 + Δ 3 m ) ( 1 δ )   · i m p l g u
(by using Equations (1) and (27))
d G u a d t + γ · d Φ u d t  
i m p u a + s u a α + γ · ( ( 1 1 δ ) · i m p l g u ( 1 + Δ 3 m ) ( 1 δ )   · i m p l g u )
= i m p u a + ( 1 + Δ 3 m ) α · η α + γ · ( 1 1 δ ) · i m p l g u γ · ( 1 + Δ 3 m ) ( 1 δ )   · i m p l g u
i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 1 δ ) · i m p u a γ · ( 1 + Δ 3 m ) ( 1 δ ) · ( i m p u a ( 3 3 + Δ ) · i m p u a )
= i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 1 δ ) γ · ( 1 1 δ ) · ( 1 + Δ 3 m ) · ( Δ 3 + Δ ) )
i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 1 δ ) γ · ( 1 1 δ ) · ( Δ 3 + Δ ) )
= i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 1 δ ) · ( 3 3 + Δ ) )
i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) · ( 1 1 δ ) ) (by using Equations (17) and (19))
d G u a d t + γ · d Φ u d t i m p u o   μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) · ( 2 α 2 α 1 ) )
Since   α > 1 2 α 2 α 1 = 2 α 1 + 1 2 α 1 = 1 + 1 2 α 1 < 2 1 < ( 2 α 2 α 1 ) < 2
(by using Equations (29) and (28))
d G u a d t + γ · d Φ u d t i m p u o   μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + 2 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) · i m p u o     ( by   using   Equation   ( 18 ) ) = c · i m p u o     ( by   using   Equation   ( 20 ) ) c · ( s u o α + i m p u o   ) d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is fulfilled for n a m , i m p u a > η α , i m p l g u > η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(b) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 .
(by using Equation (10))
d Φ u d t ( 1 1 δ ) · i m p l g u
(by using Equations (1) and (30))
d G u a d t + γ · d Φ u d t   i m p u a + s u a α + γ · ( 1 1 δ ) · i m p l g u = i m p u a + ( 1 + Δ 3 m ) α · η α + γ · ( 1 1 δ ) · i m p l g u   i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 1 δ ) · i m p u a = i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 2 α 2 α 1 ) ) i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 2 γ )   ( by   using   Equations   ( 17 )   and   ( 29 ) ) = i m p u o   μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + 2 16 · ( 1 + ( 1 + Δ 3 m ) α ) )   ( by   using   Equation   ( 19 ) ) = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) · i m p u o     ( by   using   Equation   ( 18 ) ) = c · i m p u o     ( by   using   Equation   ( 20 ) ) c · ( s u o α + i m p u o   ) d G u a d t + γ · d Φ u d t   c · d G u o d t
Hence the running condition is satisfied if n a m ,   i m p u a > η α , i m p l g u > η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , for c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
Case IV: When n a > m and i m p u a η α , since i m p l g u i m p u a we have i m p l g u η α and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) i m p u a 1 α   .
If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(by using Equations (13) and (14))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u ) ( s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a ) )
(by using Equations (1) and (31))
d G u a d t + γ · d Φ u d t  
  i m p u a + s u a α + γ · ( (   s u o α α + ( 1 2 δ ) · i m p l g u ) (   s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a ) ) )
= i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u γ ·   ( 1 + Δ 3 m ) i m p u a 1 α ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
  γ α · s u o α + ( 1 + ( 1 + Δ 3 m ) α ) · i m p u a + γ · i m p u a γ ·   ( 1 + Δ 3 m ) i m p u a 1 α ( 2 2 δ ) · ( (   i m p u a ( 3 3 + Δ ) · i m p u a ) 2 2 δ i m p u a )
= γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( 1 + Δ 3 m ) ( 2 2 δ ) · ( Δ 3 + Δ ) 2 2 δ )
  γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( Δ 3 + Δ ) 2 2 δ ( 2 2 δ ) )
d G u a d t + γ · d Φ u d t     γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · ( Δ 3 + Δ ) 2 ( 2 2 δ ) )
Since   α > 1 2 α 2 α 1 = 2 α 1 + 1 2 α 1 = 1 + 1 2 α 1 > 1 ( 1 2 2 δ ) = α 2 α 1 = 1 2 · ( 2 α 2 α 1 ) > 1 2
(by using Equations (32) and (33))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ 2 · ( Δ 3 + Δ ) 2 ) = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ( 1 1 2 · ( Δ 3 + Δ ) 2 ) ) = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ( 2 Δ 2 + 11 Δ + 18 2 Δ 2 + 12 Δ + 18 ) ) γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ )
(by using Equations (17) and (19))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + i m p u o   μ · ( 17 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  (by using Equation (18))
d G u a d t + γ · d Φ u d t γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  
d G u a d t + γ · d Φ u d t c · s u o α + c · i m p u o   (by using Equations (20) and (23))
d G u a d t + γ · d Φ u d t = c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is fulfilled for n a > m , i m p u a η α , i m p l g u η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(a) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then total RoC of Φ because of Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 .
(by using Equation (13))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u )
(by using Equations (1) and (34))
d G u a d t + γ · d Φ u d t i m p u a + s u a α + γ · ( s u o α α + ( 1 2 δ ) · i m p l g u ) = i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u γ α · s u o α + i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 2 δ ) · i m p u a = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 2 δ ) ) γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ )
(by using Equations (17) and (19))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + i m p u o   μ · ( 17 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
d G u a d t + γ · d Φ u d t γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  
d G u a d t + γ · d Φ u d t c · s u o α + c · i m p u o   (by using Equations (20) and (23))
d G u a d t + γ · d Φ u d t = c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is satisfied for n a > m , i m p u a η α , i m p l g u η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
Case V: When n a > m and i m p u a > η α , i m p l g u η α , and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) η   .
(a) If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then the total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(by using Equations (13) and (14))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u ) (   s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a ) )
(by using Equations (1) and (35))
d G u a d t + γ · d Φ u d t
  i m p u a + s u a α + γ · ( (   s u o α α + ( 1 2 δ ) · i m p l g u ) (   s u a ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a ) ) )
= i m p u a + ( 1 + Δ 3 m ) α · η α + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u γ · ( 1 + Δ 3 m ) η ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
  i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ α · s u o α + γ · ( 1 2 δ ) · i m p u a γ · ( 1 + Δ 3 m ) η ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
  γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ) γ · ( 1 + Δ 3 m ) i m p l g u 1 α ( 2 2 δ ) · ( i m p l g u 2 2 δ i m p u a )
  γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ) γ · 1 ( 2 2 δ ) · ( 1 ( 3 3 + Δ ) ) 2 · i m p u a 2   i m p u a
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ · 1 ( 2 2 δ ) · ( Δ 3 + Δ ) 2 )
(by using Equations (36) and (33))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ γ 2 · ( Δ 3 + Δ ) 2 ) = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ( 1 1 2 · ( Δ 3 + Δ ) 2 ) ) = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ ( 2 Δ 2 + 11 Δ + 18 2 Δ 2 + 12 Δ + 18 ) ) γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ )
(by using Equations (17) and (19))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + i m p u o   μ · ( 17 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
d G u a d t + γ · d Φ u d t γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  
d G u a d t + γ · d Φ u d t c · s u o α + c · i m p u o   (by using Equations (20) and (23))
d G u a d t + γ · d Φ u d t = c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is fulfilled for n a > m , i m p u a > η α , i m p l g u η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(a) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then total RoC of Φ due to Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 .
(by using Equation (13))
d Φ u d t ( s u o α α + ( 1 2 δ ) · i m p l g u )
(by using Equations (1) and (37))
d G u a d t + γ · d Φ u d t i m p u a + s u a α + γ · ( s u o α α + ( 1 2 δ ) · i m p l g u )
d G u a d t + γ · d Φ u d t = i m p u a + ( 1 + Δ 3 m ) α · η α + γ α · s u o α + γ · ( 1 2 δ ) · i m p l g u
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · i m p u a
d G u a d t + γ · d Φ u d t = γ α · s u o α + i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ )
(by using Equations (17) and (19))
d G u a d t + γ · d Φ u d t γ α · s u o α + i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 1 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + i m p u o   μ · ( 17 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
d G u a d t + γ · d Φ u d t = γ α · s u o α + ( 17 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
d G u a d t + γ · d Φ u d t γ α · s u o α + ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  
d G u a d t + γ · d Φ u d t c · s u o α + c · i m p u o   (by using Equations (20) and (23))
d G u a d t + γ · d Φ u d t = c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is satisfied for n a > m , i m p u a η α , i m p l g u η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
Case VI: When n a > m and i m p u a > η α , i m p l g u > η α , and s u a ( t ) = ( 1 + Δ 3 m ) · m i n ( i m p u a 1 α ,   η ) = ( 1 + Δ 3 m ) η .
(a) If i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) then total RoC of Φ because of Opt and MBS is d Φ u d t = d Φ u o d t + d Φ u a d t .
(by using Equations (15) and (16))
d Φ u d t ( ( 1 1 δ ) · i m p l g u ) ( ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a ) )
(by using Equations (1) and (38))
d G u a d t + γ · d Φ u d t
i m p u a + s u a α + γ · ( (   ( 1 1 δ ) · i m p l g u ) ( ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a ) ) )
i m p u a + ( 1 + Δ 3 m ) α · η α + γ · ( 1 1 δ ) · i m p u a γ · ( 1 + Δ 3 m ) ( 2 2 δ ) · ( i m p l g u 2 i m p u a )  
i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 1 δ ) · i m p u a γ · 1 ( 2 2 δ ) · ( (   i m p u a ( 3 3 + Δ ) · i m p u a ) 2 i m p u a )
= i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 1 δ ) · i m p u a γ · ( Δ 3 + Δ ) 2 ( 2 2 δ ) · i m p u a  
= ( 1 + ( 1 + Δ 3 m ) α + γ · ( 2 α 2 α 1 ) γ · ( Δ 3 + Δ ) 2 ( 2 2 δ ) ) · i m p u a
( 1 + ( 1 + Δ 3 m ) α + 2 γ γ 2 · ( Δ 3 + Δ ) 2 ) · i m p u a (by using Equations (29) and (33))
= ( 1 + ( 1 + Δ 3 m ) α + γ · ( 2 1 2 · ( Δ 3 + Δ ) 2 ) ) · i m p u a
= ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 + 2 Δ 2 + 11 Δ + 18 2 Δ 2 + 12 Δ + 18 ) ) · i m p u a
( 1 + ( 1 + Δ 3 m ) α + 2 γ ) · i m p u a
i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 2 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) (by using Equations (17) and (19))
= i m p u o   μ · ( 18 16 · ( 1 + ( 1 + Δ 3 m ) α ) )
= ( 18 16 · ( 1 + Δ 3 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o (by using Equation (18))
= ( ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) ) · i m p u o  
= c · i m p u o   (by using Equation (20))
c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is fulfilled for n a > m , i m p u a > η α , i m p l g u > η α , i m p l g u > ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
(a) If i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) then total RoC of Φ due to Opt and MBS depends on d Φ u o d t since d Φ u a d t 0 .
(by using Equations (15))
d Φ u d t ( ( 1 1 δ ) · i m p l g u )
(by using Equations (1) and (39))
d G u a d t + γ · d Φ u d t i m p u a + s u a α + γ · ( ( 1 1 δ ) · i m p l g u )
d G u a d t + γ · d Φ u d t = i m p u a + ( 1 + Δ 3 m ) α · η α + γ · ( 1 1 δ ) · i m p l g u  
d G u a d t + γ · d Φ u d t i m p u a + ( 1 + Δ 3 m ) α · i m p u a + γ · ( 1 1 δ ) · i m p u a
d G u a d t + γ · d Φ u d t i m p u a ( 1 + ( 1 + Δ 3 m ) α + γ · ( 1 1 δ ) )
d G u a d t + γ · d Φ u d t = i m p u a · ( 1 + ( 1 + Δ 3 m ) α + γ · ( 2 α 2 α 1 ) )
d G u a d t + γ · d Φ u d t i m p u o   μ · ( 1 + ( 1 + Δ 3 m ) α + 2 γ ) (by using Equations (17) and (29))
d G u a d t + γ · d Φ u d t = i m p u o   μ · ( ( 1 + ( 1 + Δ 3 m ) α ) + 2 16 · ( 1 + ( 1 + Δ 3 m ) α ) ) (by using Equation (19))
d G u a d t + γ · d Φ u d t = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) · i m p u o   (by using Equation (18))
d G u a d t + γ · d Φ u d t = c · i m p u o   (by using Equation (20))
d G u a d t + γ · d Φ u d t c · ( s u o α + i m p u o   )
d G u a d t + γ · d Φ u d t c · d G u o d t
Hence the running condition is satisfied for n a > m , i m p u a > η α , i m p l g u > η α , i m p l g u ( i m p u a ( 3 3 + Δ ) · i m p u a ) , c = ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) .
The analysis of all cases and sub cases in Lemma 6 prove that the first condition, running condition is fulfilled. Aggregating the discourse about all conditions job arrival and completion condition, boundary condition and Lemma 6, it is concluded that Theorem 1 follows. The competitive values of related algorithms and MBS on α = 2   and   3 are shown in Table 3. Among all online clairvoyant and ON-C scheduling algorithms, the competitiveness of MBS is least, which reflects that the MBS outperforms other algorithms.

6. Illustrative Example

To observe the performance of MBS, a group of four processors and a set of seven jobs are considered. The best known result in the online non-clairvoyant scheduling algorithms is provided by the Azar et al. [40] in NC-PAR. NC-PAR is a super-constant lower bound on the competitive ratio of any deterministic algorithm even for fractional flow-time in the case of uniform densities. The processing of jobs using algorithms MBS and NC-PAR [40] is simulated and the results are stated in Table 4 as well as in Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11. The jobs arrived along with their importance but the size of jobs was computed on the completion of jobs. The response time (Rt) is the time interval between the starting time of execution and arrival time of a job. The turnaround time is the time duration between completion time and arrival time of a job. Most of the jobs using MBS have lesser turnaround time than using NC-PAR. The Rt of the jobs using MBS is better than NC-PAR. In Figure 3 and Figure 4, the allocation and execution sequence of jobs on four processors is depicted with the help of triangles and rectangles using NC-PAR and MBS, respectively. As per the Figure 3 and Figure 4, the importance of the jobs in NC-PAR increased with time where as in MBS the importance remains constant during the life time of the jobs. It is clearly evident from the Figure 3 and Figure 4 that on any processor using NC-PAR at a time only one job has been executed, whereas using MBS the processor has been shared by more than one job. The hardware specifications are mentioned in the Table 5.
Figure 5 and Figure 6 present the speed of different processors and combined speed of all processors with respect to time using MBS and NC-PAR, respectively. As per the graphs of Figure 5, the speed of a processor using MBS goes high initially but later it reduces and most of the time the speed of processors using MBS is constant, but when processors executes jobs using NC-PAR the speed of processors have heavy fluctuations, which shows that some extra energy may be needed for such frequent fluctuation in NC-PAR. The graphs of the Figure 6 shows that the combined speed of processors using NC-PAR increased and decreased linearly whereas using MBS it increased and decreased stepwise. The count of local maxima and minima in the speed growth graphs (Figure 7) of NC-PAR is more than MBS. Therefore, not only individual processor’s speed but also the combined speed of all the processors is reflecting the heavy fluctuation in NC-PAR and varying-constant mixed behaviour of MBS.
In this simulation analysis the traditional power function is used and the value of α is 2. The processors are having the maximum limit of speed which is considered 3.6. The value of Δ = ( 3 α ) 1 is considered for the analysis. The power consumed is square of the speed, i.e., proportional to the speed this fact can be viewed by comparing the graphs of Figure 5 and Figure 9. Figure 8, shows that initially MBS consumed more power but power consumption decreased with respect to increase in time, whereas in case of NC-PAR there is no fix pattern, but power consumption is higher most of the time than in MBS.
The graphs of Figure 10 demonstrate the objective of the algorithm (important based flow time plus energy). It reveals that except one processor P1, all other processor have lesser objective value, when these processors executed jobs by using MBS than NC-PAR. The combined objective of all processor is given in the Figure 11, which strengthen the previous observation of Figure 10 (the objective values using MBS is lesser than using NC-PAR). It can be concluded from the different observations and the Figure 11, that the algorithm MBS performs better than NC-PAR.

7. Conclusions and Future Work

To date, the problem of ON-C scheduling algorithms with an objective to minimize the IbFt+E for multiprocessor setting is studied less extensively. A scheduling algorithm multiprocessor with bounded speed (MBS) is proposed, which uses importance-based/weighted round robin (WRR) for job selection. MBS extends the theoretical study of an ON-C multiprocessor DSS scheduling problem with an objective to minimize the IbFt+E using the bounded speed model, where every processor’s maximum speed using MBS is ( 1 + Δ 3 m ) η and using offline adversary Opt is η . The speed of any processor changes if there is a variation in the total importance of jobs on that processor. The competitiveness of MBS is ( 9 8 + 3 Δ 8 ) · ( 1 + ( 1 + Δ 3 m ) α ) = O ( 1 ) against an offline adversary, using the potential function analysis and traditional power function. The performance of MBS is compared with best known algorithm NC-PAR [40]. A set of jobs and processors are used to simulate the working of MBS and NC-PAR. The average turnaround and response time of jobs, when they are executed by using MBS is lesser than NC-PAR. The speed scaling strategy and power consumption in MBS is better than NC-PAR. For all processors at any time, MBS provides the lesser value of the sum of important-based flow time and energy consumed than NC-PAR. Competitiveness of NC-PAR is 3 for α = 2 and 3.5 for α = 3 , whereas the value of competitive ratio c of MBS for Δ = ( 3 α ) 1 , m = 2 and α = 2 is 2.442; for Δ = ( 3 α ) 1 , m = 2 and α = 3 is 2.399; for Δ = ( 3 α ) 1 , m > 2 , α = 2 is 2.375 < c < 2.442 ; for Δ = ( 3 α ) 1 , m > 2 , α = 3 is 2.333 < c < 2.399 . These results demonstrate that the scheduling algorithm MBS outperforms other algorithms. The competitive value of MBS is least to date. Before these outcomes, there were no results acknowledged for the multi-processor machines in the ON-C model with identified importance, even for unit importance jobs [40]. The further enhancement of this study will be to implement the MBS in real environment. One open problem is to achieve a reasonably less competitive algorithm than MBS. In this study, author considers non-migratory and sequential jobs and this work may be extended to find a scheduling for migratory and non-sequential jobs. Other factors (such as memory requirement) may also be considered for analysis in future extension.

Author Contributions

All authors have worked on this manuscript together. Writing—original draft, P.S.; writing—review and editing, B.K., O.P.M., H.H.A. and G.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Belady, C. In the Data Center, Power and Cooling Costs More Than the IT Equipment It Supports, Electronics Cooling Magazine. Available online: http://www.electronics-cooling.com/2007/02/in-the-data-center-power-and-cooling-costs-more-than-the-it-equipment-it-supports/ (accessed on 10 January 2020).
  2. Chan, H.L.; Edmonds, J.; Lam, T.W.; Lee, L.K.; Marchetti-Spaccamela, A.; Prush, K. Non-clairvoyant speed scaling for flow and energy. Algorithmica 2011, 61, 507–517. [Google Scholar] [CrossRef] [Green Version]
  3. Merritt, R. CPU Designers’ Debate Multi-Core Future, EE Times. 2 June 2008. Available online: http://www.eetimes.com/document.asp?doc_id=1167932 (accessed on 15 January 2020).
  4. U.S. Environmental Protection Agency, EPA Report on Server and Data Center Energy Efficiency. Available online: https://www.energystar.gov/ia/partners/prod_development/downloads/EPA_Report_Exec_Summary_Final.pdf (accessed on 14 January 2020).
  5. Singh, P.; Wolde-Gabriel, B. Executed-time Round Robin: EtRR an online non-clairvoyant scheduling on speed bounded processor with energy management. J. King Saud Univ. Comput. Inf. Sci. 2016, 29, 74–84. [Google Scholar] [CrossRef] [Green Version]
  6. Bansal, N.; Chan, H.L.; Pruhs, K. Speed scaling with an arbitrary power function. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms, New York, NY, USA, 4–6 January 2009; pp. 693–701. [Google Scholar]
  7. Bansal, N.; Kimbrel, T.; Pruhs, K. Dynamic speed scaling to manage energy and temperature. J. ACM 2007, 54, 1–39. [Google Scholar] [CrossRef]
  8. Singh, P.; Khan, B.; Vidyarthi, A.; Haes Alhelou, H.; Siano, P. Energy-aware online non-clairvoyant scheduling using speed scaling with arbitrary power function. Appl. Sci. 2019, 9, 1467. [Google Scholar] [CrossRef] [Green Version]
  9. Lam, T.W.; Lee, L.K.; To, I.K.K.; Wong, P.W.H. Nonmigratory multiprocessor scheduling for response time and energy. IEEE Trans. Parallel Distrib. Syst. 2008, 19, 1–13. [Google Scholar]
  10. Motwani, R.; Phillips, S.; Torng, E. Nonclairvoyant scheduling. Theor. Comput. Sci. 1994, 30, 17–47. [Google Scholar] [CrossRef] [Green Version]
  11. Yao, F.; Demers, A.; Shenker, S. A scheduling model for reduced CPU energy. In Proceedings of the Annual Symposium on Foundations of Computer Science, Berkeley, CA, USA, 23–25 October 1995; pp. 374–382. [Google Scholar]
  12. Koren, G.; Shasha, D. Dover: An optimal on-line scheduling algorithm for overloaded uniprocessor real-time systems. SIAM J. Comput. 1995, 24, 318–339. [Google Scholar] [CrossRef]
  13. Leonardi, S.; Raz, D. Approximating total flow time on parallel machines. In Proceedings of the ACM Symposium on Theory of Computing, El Paso, TX, USA, 4–6 May 1997; pp. 110–119. [Google Scholar]
  14. Kalyanasundaram, B.; Pruhs, K. Speed is as powerful as clairvoyant. J. ACM 2000, 47, 617–643. [Google Scholar] [CrossRef]
  15. Edmonds, J. Scheduling in the dark. Theor. Comput. Sci. 2000, 235, 109–141. [Google Scholar] [CrossRef] [Green Version]
  16. Kalyanasundaram, B.; Pruhs, K. Minimizing flow time nonclairvoyantly. J. ACM 2003, 50, 551–567. [Google Scholar] [CrossRef]
  17. Becchetti, L.; Leonardi, S. Nonclairvoyant scheduling to minimize the total flow time on single and parallel machines. J. ACM 2004, 51, 517–539. [Google Scholar] [CrossRef] [Green Version]
  18. Chekuri, C.; Goel, A.; Khanna, S.; Kumar, A. Multiprocessor scheduling to minimize flow time with epsilon resource augmentation. In Proceedings of the 36th Annual ACM Symposium on Theory of Computing, Chicago, IL, USA, 13–15 June 2004; pp. 363–372. [Google Scholar]
  19. Chen, J.J.; Hsu, H.R.; Chuang, K.H.; Yang, C.L.; Pang, A.C.; Kuo, T.W. Multiprocessor energy efficient scheduling with task migration considerations. In Proceedings of the 16th Euromicro Conference on Real-Time Systems, Catania, Italy, 2 July 2004; pp. 101–108. [Google Scholar]
  20. Albers, S.; Fujiwara, H. Energy-efficient algorithms for flow time minimization. ACM Trans. Algorithms 2007, 3, 49. [Google Scholar] [CrossRef]
  21. Bansal, N.; Pruhs, K.; Stein, C. Speed scaling for weighted flow time. In Proceedings of the 18th Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; pp. 805–813. [Google Scholar]
  22. Lam, T.W.; Lee, L.K.; To, I.K.K.; Wong, P.W.H. Competitive non-migratory scheduling for flow time and energy. In Proceedings of the 20th ACM Symposium on Parallelism in Algorithms and Architectures, Munich, Germany, 14–16 June 2008; pp. 256–264. [Google Scholar]
  23. Chadha, J.; Garg, N.; Kumar, A.; Muralidhara, V. A competitive algorithm for minimizing weighted flow time on unrelated processors with speed augmentation. In Proceedings of the Annual ACM Symposium on Theory of Computing, Bethesda, MD, USA, 31 May–2 June 2009; pp. 679–684. [Google Scholar]
  24. Chan, S.H.; Lam, T.W.; Lee, L.K.; Liu, C.M.; Ting, H.F. Sleep management on multiple processors for energy and flow time. In Proceedings of the 38th International Colloquium on Automata, Languages and Programming, Zurich, Switzerland, 4–8 July 2011; pp. 219–231. [Google Scholar]
  25. Albers, S.; Antoniadis, A.; Greiner, G. On multi-processor speed scaling with migration. In Proceedings of the 23rd Annual ACM Symposium on Parallelism in Algorithms and Architectures, San Jose, CA, USA, 4–6 June 2011; pp. 279–288. [Google Scholar]
  26. Gupta, A.; Im, S.; Krishnaswamy, R.; Moseley, B.; Pruhs, K. Scheduling heterogeneous processors isn’t as easy as you think. In Proceedings of the 23rd Annual ACM-SIAM Symposium on Discrete Algorithms, Kyoto, Japan, 17–19 January 2012; pp. 1242–1253. [Google Scholar]
  27. Chan, S.H.; Lam, T.W.; Lee, L.K.; Zhu, J. Nonclairvoyant sleep management and flow-time scheduling on multiple processors. In Proceedings of the 25th Annual ACM Symposium on Parallelism in Algorithms and Architectures, Montreal, QC, Canada, 23–25 July 2013; pp. 261–270. [Google Scholar]
  28. Lawler, E.L.; Lenstra, J.K.; Kan, A.R.; Shmoys, D.B. Sequencing and Scheduling: Algorithms and Complexity. In Handbooks in Operations Research and Management Science; Elsevier: Amsterdam, The Netherlands, 1993; Volume 4, pp. 445–522. [Google Scholar]
  29. Hall, H. Approximation algorithm for scheduling. In Approximation Algorithm for NP-Hard Problems; Hochbaum, D.S., Ed.; PWS Publishing Company: Boston, MA, USA, 1997; pp. 1–45. [Google Scholar]
  30. Sgall, J. On-line scheduling. In Online Algorithms, The State of the Art; Fiat, A., Woeginger, G.J., Eds.; Springer: Berlin/Heidelberg, Germany, 1998; pp. 196–231. [Google Scholar]
  31. Karger, D.; Stein, C.; Wein, J. Scheduling Algorithms, CRC Handbook of Theoretical Computer Science; CRC Press: Boca Raton, FL, USA, 1999. [Google Scholar]
  32. Irani, S.; Pruhs, K. Algorithmic problems in power management. ACM SIGACT News 2005, 36, 63–76. [Google Scholar] [CrossRef] [Green Version]
  33. Albers, S. Energy efficient algorithms. Commun. ACM 2010, 53, 86–96. [Google Scholar] [CrossRef] [Green Version]
  34. Albers, S. Algorithms for dynamic speed scaling. In Proceedings of the 28th International Symposium of Theoretical Aspects of Computer Science, Dortmund, Germany, 10–12 March 2011; pp. 1–11. [Google Scholar]
  35. Gupta, A.; Krishnaswamy, R.; Pruhs, K. Scalably scheduling power-heterogeneous processors. In Proceedings of the 37th International Colloquium on Automata, Languages and Programming, Bordeaux, France, 5–10 July 2010; pp. 312–323. [Google Scholar]
  36. Fox, K.; Im, S.; Moseley, B. Energy efficient scheduling of parallelizable jobs. In Proceedings of the 24th Annual Symposium on Discrete Algorithms, New Orleans, LA, USA, 6–8 January 2013; pp. 948–957. [Google Scholar]
  37. Thang, N.K. Lagrangian duality in online scheduling with resource augmentation and speed scaling. In Proceedings of the 21st European Symposium on Algorithms, Sophia Antipolis, France, 2–4 September 2013; pp. 755–766. [Google Scholar]
  38. Im, S.; Kulkarni, J.; Munagala, K.; Pruhs, K. SelfishMigrate: A scalable algorithm for non-clairvoyantly scheduling heterogeneous processors. In Proceedings of the 55th IEEE Annual Symposium on Foundations of Computer Science, Philadelphia, PA, USA, 18–21 October 2014; pp. 531–540. [Google Scholar]
  39. Bell, P.C.; Wong, P.W.H. Multiprocessor speed scaling for jobs with arbitrary sizes and deadlines. J. Comb. Optim. 2015, 29, 739–749. [Google Scholar] [CrossRef] [Green Version]
  40. Azar, Y.; Devanue, N.R.; Huang, Z.; Panighari, D. Speed scaling in the non-clairvoyant model. In Proceedings of the 27th Annual ACM Symposium on Parallelism in Algorithms and Architectures, Portland, OR, USA, 13–15 June 2015; pp. 133–142. [Google Scholar]
Figure 1. Performance and speed curve.
Figure 1. Performance and speed curve.
Applsci 10 02459 g001
Figure 2. Flow chart of the MBS scheduling algorithm.
Figure 2. Flow chart of the MBS scheduling algorithm.
Applsci 10 02459 g002
Figure 3. Scheduling of jobs using NC-PAR.
Figure 3. Scheduling of jobs using NC-PAR.
Applsci 10 02459 g003
Figure 4. Scheduling of jobs using MBS.
Figure 4. Scheduling of jobs using MBS.
Applsci 10 02459 g004
Figure 5. Speed of processors using MBS and NC-PAR.
Figure 5. Speed of processors using MBS and NC-PAR.
Applsci 10 02459 g005
Figure 6. Combined speed of all processors using MBS and NC-PAR.
Figure 6. Combined speed of all processors using MBS and NC-PAR.
Applsci 10 02459 g006
Figure 7. Growth of combined speed of all processors using MBS and NC-PAR.
Figure 7. Growth of combined speed of all processors using MBS and NC-PAR.
Applsci 10 02459 g007
Figure 8. Total power consumed by all processors using MBS and NC-PAR.
Figure 8. Total power consumed by all processors using MBS and NC-PAR.
Applsci 10 02459 g008
Figure 9. Power consumed by processors using MBS and NC-PAR.
Figure 9. Power consumed by processors using MBS and NC-PAR.
Applsci 10 02459 g009
Figure 10. Importance-based flow time + energy consumed using MBS and NC-PAR.
Figure 10. Importance-based flow time + energy consumed using MBS and NC-PAR.
Applsci 10 02459 g010
Figure 11. Total importance-based flow time + energy consumed using MBS and NC-PAR.
Figure 11. Total importance-based flow time + energy consumed using MBS and NC-PAR.
Applsci 10 02459 g011
Table 1. Notations used.
Table 1. Notations used.
NotationsMeaning
tCurrent time
jA job
uA processor
r ( j ) or r j Release/arrival time of a job j
p ( j ) Processing requirement (size) of a job j
m Number of processors
l g u On a processor u, the count of lagging jobs, at time t
η Maximum speed of a processor using Opt
PPower of a processor at speed s
s ( t ) or s At time t, speed of some processor
α A constant, commonly believed that its value is 2 or 3
Δ A constant, its value depends on the value of α
I, SA set of jobs and their schedule, respectively
p w k ( j ,   t ) , p w k a ( j ,   t ) and p w k o ( j ,   t ) Remaining/pending work of a job j at time t, using MBS and Opt, respectively
F ( j ) Flow time of a job j
FTotal importance-based flow time
i m p u j ( t ) or i m p u ( j ) Importance/weight of a job j, at time t on a processor u
i m p u a ( t ) or i m p u a and i m p u o ( t ) or i m p u o Importance of all active jobs using MBS and Opt at time t on a processor u, respectively
i m p l g ( t ) or i m p l g and i m p l g u ( t ) or i m p l g u Total importance of lagging jobs, at time t on all m processors and on a processor u, respectively
n a ( t ) or n a and n o ( t ) or n o Total number of active jobs (NoAJ) in MBS and Opt at time t on all m processors, respectively
n u a ( t ) or n u a and n u o ( t ) or n u o NoAJ in MBS and Opt at time t on a processor u, respectively
s u a ( t ) or s u a and s u o ( t ) or s u o Speed of a processor u for MBS and Opt at time t, respectively
i m p ( t ) Total importance of all active jobs n a , at time t
E Energy consumed by processors
GTotal IbFt+E
c Competitiveness
μ A constant ( 0 < μ < 1 ), its value depends on the value of Δ
G a ( t ) or G a and G o ( t ) or G o IbFt+E acquired till time t by the MBS and Opt, respectively
d G a ( t ) d t or d G a d t and d G o ( t ) d t or d G o d t Rate of change (RoC) of G a due to MBS and G o due to Opt at time t, respectively
G u a ( t ) or G u a and G u o ( t ) or G u o IbFt+E acquired on a processor u till time t by the MBS and Opt, respectively
d G u a ( t ) d t or d G u a d t and d G u o ( t ) d t or d G u o d t RoC of G a due to MBS and Opt at time t on a processor u, respectively
γ A constant (> 0)
c i Coefficient of a job j i at time t
ω i Difference of pending work of a job j i using MBS and Opt at time t
δ A constant depends on α , its value is ( 1 2 α )
L G u A set of lagging jobs using MBS on a processor u
L G A set of all lagging jobs using MBS on all m processors
Φ ( t ) or Φ Total potential value of all m processors at time t
Φ u ( t ) or Φ u Potential value of a processor u at time t
d Φ o d t and d Φ a d t RoC of Φ due to Opt and MBS, respectively
d Φ d t RoC of Φ due to Opt and MBS
d Φ u o d t and d Φ u a d t RoC of Φ due to Opt and MBS on a processor u, respectively
d Φ u d t RoC of Φ due to Opt and MBS on a processor u
Table 2. Results of multiprocessor with bounded speed (MBS).
Table 2. Results of multiprocessor with bounded speed (MBS).
Number of Processors (m) α = 2 α = 3
Speed Ratio ( s r ) Competitive Ratio (c)Speed Ratio ( s r ) Competitive Ratio (c)
21.027782.441891.018522.39936
41.013892.408221.009262.36604
81.006942.391551.004632.34961
161.003472.383261.002312.34145
641.000862.377061.000582.33536
1281.000432.376031.000292.33435
5121.000102.375261.000072.33359
10241.000052.375131.000032.37513
40961.000012.375031.000012.33336
11,2641.0000042.375011.0000032.33334
Note: the speed ratio s r = M a x i m u m   s p e e d   o f   a   p r o c e s s o r   u s i n g   M B S M a x i m u m   s p e e d   o f   a   p r o c e s s o r   u s i n g   O p t .
Table 3. Summary of Results.
Table 3. Summary of Results.
MultiprocessorCompetitiveness for Weighted Flow Time + EnergyModelling Criteria
AlgorithmsGeneral α = 2 α = 3 (Bounded (BS)/Unbounded Speed(US)) (Clairvoyant (C)/Non-Clairvoyant (NC)
GKP [35] α 2 49US, C
WLAPS+E [36] 5 Δ 2 (where 0 < Δ 1 6 )180180US, NC
ALGThang [37] 8 ( 1 + α l n α ) 31.08529.85US, C
SM-E [38] α 2 49US, NC
DCRR [39] 2 4 α ( l o g α P + α α 2 α 1 )
w h e r e   P   i s   t h e   r a t i o   b e t w e e n   t h e   m a x i m u m   a n d   m i n i m u m   j o b   s i z e
>2048>442368US, C
NC-PAR [40] ( α + 1 α 1 ) 33.5US, NC
MBS [This Paper] ( 9 8 + 3 Δ 8 ) ( 1 + ( 1 + Δ 3 m ) α )
(where 0 < Δ ( 3 α ) 1 )
2.4422.399BS, NC
Table 4. Job details and execution data using MBS and NC-PAR.
Table 4. Job details and execution data using MBS and NC-PAR.
Job DetailsMBS [This Paper]NC-PAR [40]
JobSizeImportanceArrival TimeCompletion TimeResponse TimeTurnaround TimeCompletion TimeResponse TimeTurnaround Time
J135811401314013
J2641022402223021
J3155410061208
J4831163002433027
J545572902229622
J617482301523715
J75661032022431433
Average Values23.143017.71425.2863.85719.857
Table 5. Hardware specifications.
Table 5. Hardware specifications.
Simulation ParametersValues
CPUIntel(R) Core(TM) i5-4210U CPU @ 1.70 GHz
RAM4.00 GB RAM
Hard Drive1.0 TB
Operating SystemRed Hat Linux 6.1
KernelLinux kernel version 2.2.12

Share and Cite

MDPI and ACS Style

Singh, P.; Khan, B.; Mahela, O.P.; Haes Alhelou, H.; Hayek, G. Managing Energy Plus Performance in Data Centers and Battery-Based Devices Using an Online Non-Clairvoyant Speed-Bounded Multiprocessor Scheduling. Appl. Sci. 2020, 10, 2459. https://doi.org/10.3390/app10072459

AMA Style

Singh P, Khan B, Mahela OP, Haes Alhelou H, Hayek G. Managing Energy Plus Performance in Data Centers and Battery-Based Devices Using an Online Non-Clairvoyant Speed-Bounded Multiprocessor Scheduling. Applied Sciences. 2020; 10(7):2459. https://doi.org/10.3390/app10072459

Chicago/Turabian Style

Singh, Pawan, Baseem Khan, Om Prakash Mahela, Hassan Haes Alhelou, and Ghassan Hayek. 2020. "Managing Energy Plus Performance in Data Centers and Battery-Based Devices Using an Online Non-Clairvoyant Speed-Bounded Multiprocessor Scheduling" Applied Sciences 10, no. 7: 2459. https://doi.org/10.3390/app10072459

APA Style

Singh, P., Khan, B., Mahela, O. P., Haes Alhelou, H., & Hayek, G. (2020). Managing Energy Plus Performance in Data Centers and Battery-Based Devices Using an Online Non-Clairvoyant Speed-Bounded Multiprocessor Scheduling. Applied Sciences, 10(7), 2459. https://doi.org/10.3390/app10072459

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop