1. Introduction
The public sector faces mounting pressure to transform its operations and service delivery mechanisms in an era marked by rapid technological advancements and evolving societal demands. The need for agility, responsiveness, and efficiency within government agencies has never been more pressing [
1]. Central to addressing these challenges is the development and implementation of a robust dynamic workload management system (DWMS) tailored to the unique demands of the public sector.
Traditional public administration systems often struggle to adapt to changing circumstances, leading to inefficiencies, delays, and suboptimal resource allocation [
2]. In addition, they lack independent performance assessments, while reformation approaches must be adapted to evolving circumstances [
3]. In this context, a DWMS emerges as a critical tool for enhancing the performance of public sector organizations by optimizing work distribution, resource allocation, and service delivery processes in real time. The current paper analyzes the conceptualization, design, and practical implementation of such a system, aiming to bridge the gap between theory and practice, as well as seeks to explore the development of a DWMS as a strategic imperative for the public sector. The significance of this study lies in its potential to transform how government agencies operate, enabling them to respond swiftly to emerging challenges, allocate resources judiciously, and enhance overall service quality. By providing a comprehensive understanding of DWMS design and implementation in the public sector, this research contributes to the body of knowledge on public administration and management. Workload distribution was always the “Achilles heel” of any system related to human resource management. The current situation is characterized by an overabundance of public sector employees, many without the necessary training. At the same time, it is not uncommon for individuals to become overwhelmed by the high volume of work, leading to delays in completing several tasks. Consequently, the quality of these tasks may suffer as they are rushed to meet deadlines. Highly skilled employees are overloaded with tasks that cannot be handled in their daily schedule, resulting in overtime and physical and mental problems [
4,
5]. On the other hand, underutilized human resources, due to a subjective assessment of lack of specialization, do not undertake jobs, resulting in the operation of two-speed services and the gradual cessation of all activities.
The existing literature reveals a significant gap in research regarding the development and implementation of DWMSs in the public sector. Although related concepts and theories have been explored, two critical aspects remain unaddressed. First, existing management systems often rely on performance indicators and assessments based on supervisors’ opinions rather than productivity factors. This oversight limits the objectivity and effectiveness of workload management. Second, there is a lack of a holistic approach to modeling the potential efficiency of public bodies based on employee efficiency. This gap hinders the ability to comprehensively evaluate and enhance organizational performance. Furthermore, the selection of the ANFIS algorithm as the candidate for producing the Capacity Factor for each employee was based on pre-analysis with actual data and is supported by a robust theoretical framework. Thus, there is a clear need for comprehensive analysis and practical guidance on designing and implementing DWMSs that address these gaps. The current investigation sheds light on various aspects of the problem by highlighting task boundary management.
The paper is organized into several sections, including a literature review of relevant concepts and theories, a discussion of the critical components of a DWMS, and case studies illustrating successful implementations. It concludes by offering insights into the potential future developments in dynamic workload management in the public sector.
In summary, current research promotes a dynamic workload management system’s pivotal role in enhancing public sector organizations’ efficiency and effectiveness. By addressing the challenges and opportunities inherent in such a system, we aim to provide a comprehensive resource for policymakers, directors, and researchers invested in the future of public administration.
2. Related Work and Contributions
Workload management in the public sector has not been extensively reported, although researchers have attempted to outline the proposed topic. A quite exciting approach was initiated by Michalopoulos et al. [
6], where a four-level factor profile was utilized to produce the correlation between the hard skills and the Capacity Factor (CF) to determine each employee’s efficiency. This initial deployment was further deployed in the research by Giotopoulos et al. [
7], where results illustrated the significant impact of work experience in the private sector compared to the public sector on the value of the Capacity Factor. ANFIS was selected due to its ability to learn and represent complex nonlinear relationships effectively. While the deployment of neural networks has demonstrated remarkable predictive capabilities, a large amount of information in the public domain remains unexploited. Therefore Theodorakopoulos et al. [
8,
9] showed the use of big data and machine learning for predicting valuable information in many fields of modern business that could be applied to any public body. The study of Adams et al. [
10] suggests that for sustainability reporting to become widespread in the public sector, it may require mandatory adoption or a shift toward competitive resource allocation based on sustainability performance. However, these conclusions are subject to potential limitations, and future research should explore the role of regulatory environment congruence in sustainability performance. The importance of intangible assets (IAs) and intellectual capital (IC) was also highlighted by Boj et al. [
11], who introduced a methodology using an analytic network process (ANP). Regarding productivity in the public sector, the use and usefulness of corresponding performance measures are considered vital [
12]. Arnaboldi et al. [
13] proposed leveraging complexity theory to address the intricacies of public sector performance management, taking into account performance management, exploring the applicability of complexity theory to public service managers’ everyday tasks, and conducting a holistic evaluation including public managers. Gunarsih et al. [
14] emphasized the importance of comprehensive performance measurement by integrating the balanced scorecard (BSC) method along with system dynamics (SD) modeling to capture complex interactions. On the other hand, Bruhn et al. [
15] introduced a multi-level approach to understanding frontline interactions within the public sector and demonstrated the value of applying conversation analysis methods to investigate how policies and rules are applied, negotiated, and reshaped during those interactions. Indicators such as allocation, distribution, and stabilization have been analyzed, leading to defined measurements of public sector performance (PSP) and efficiency (PSE), giving an advantage to smaller public bodies [
16].
There is still an ongoing debate regarding measuring productivity in public services. This debate suggests that we are currently in a phase characterized by testing various approaches, making it challenging to draw meaningful comparisons.
When discussing and assessing the performance of public services, it is crucial to clearly distinguish between various aspects of public service performance, such as productivity, efficiency, and effectiveness, and how working time flexibility affects those factors. Integrating working time flexibility into this strategy can contribute positively to productivity but has various impacts due to the range of measures implemented. The manner of implementation and collective influence affect working conditions, and work–life balance and other relevant factors are emphasized [
17]. A measurable dimension that will determine the efficiency of a civil servant is still missing. Instead, government outputs typically encompass complex social outcomes that are challenging to precisely delineate and often exhibit multidimensional and interconnected characteristics [
18].
A challenge that arises when considering the private sector when comparing public and private sector efficiency is the achievement of complete comparability, which would allow for an adequate evaluation of each. Even a cursory analysis reveals that the public and private sectors are not interchangeable. Their objectives diverge significantly, with the private sector primarily focused on profit generation. At the same time, the public sector seeks economic gains and the attainment of social benefits, with a primary mission of ensuring the public’s well-being. Private sector projects are primarily driven by the pursuit of economic benefits: often with limited attention to social and environmental concerns. However, in contemporary times, many companies are gradually shifting their mindsets and are striving to integrate social responsibility alongside profit generation. Conversely, public projects may prioritize social benefits over economic gains [
19].
The relationship between outcomes or outputs, as the literature refers to them, and inputs or efforts determines efficiency. While this relationship may seem straightforward, practical implementation often proves otherwise. Identifying and measuring inputs and outputs in the public sector is generally a challenging endeavor.
How is productivity related to working hours? Definitely, increasing working time does not correspond to a proportional increase in productivity due to fatigue, which is the most crucial parameter in the equation [
20,
21,
22]. On the other hand, factors such as wages, work arrangements, job content, IT skills, working conditions, health, stress, and job satisfaction significantly contribute to employee productivity [
23]. In contrast, incentives for companies to adopt and expand flexible working time arrangements, like flextime and working time accounts, can improve morale, individual performance, and overall company productivity and sustainability [
24].
An interesting approach elucidates a path to augmenting production efficiency by implementing a work-sharing methodology. Workforce reduction is achieved by meticulously assessing cycle times per head, as organized by a process [
25].
While it is relatively straightforward to measure inputs in terms of physical units, such as the number of employees or hours worked, or in financial terms, defining and quantifying outputs presents more significant challenges. This is due to the diverse perspectives that consumers may have, whether they are viewed as end-users or representatives of society at large.
Furthermore, additional complexities arise when it comes to defining and measuring the outcomes of public services. External factors such as individual behavior, culture, and social norms can significantly influence the final results, making it a multifaceted and intricate process [
26].
In analogous investigations aimed at forecasting student performance through prior academic attainments, algorithms including naive Bayes, ID3, C4.5, and SVM were utilized, with particular emphasis on their applicability and analyses concerning students. The evaluation of these algorithms centered on metrics such as accuracy and error rates [
27].
Aligned with the prevailing tendency towards incorporating AI in human resource management, Chowdhury et al. [
28] undertook a systematic review of AI research in HRM and delineated key themes such as AI applications, collective intelligence, and AI drivers and barriers. While the existing literature predominantly concentrates on AI applications and their associated advantages, a research gap exists regarding collective intelligence, AI transparency, and ethical considerations. The paper introduces an AI capability framework for organizations and suggests research priorities, including validating the framework, assessing the impact of AI transparency on productivity, and devising knowledge management strategies for fostering AI–employee collaboration. It underscores the necessity for empirical studies to comprehensively evaluate the effects of AI adoption.Use of machine learning combined with metrics such as mean absolute error, mean squared error, and R-squared have been widely used lately for evaluating employee performance [
29]. Alsheref et al. [
30] proposed automated model that can predict employee attrition based on different predictive analytical techniques such as random forest, gradient boosting, and neural networks, while Arslankaya, S., focused on employee labor loss, demonstrating the superiority of ANFIS over pure fuzzy logic [
31]. In all cases, it was evident that there is a need for a recruitment system that can use artificial intelligence during the hiring procedure to quantify the objectives set by the HR department [
32] and move into a more personalized HRM system [
33] that provides insight into workload performance of the personnel. The incorporation of artificial intelligence (AI) and machine learning (ML) into business process management (BPM) within organizations and enterprises holds promise for achieving these goals and enhancing performance, innovation procedures, and competitive edges [
34].
In continuation of the research defining efficiency to include the measurement and evaluation of workloads, Casner, S.M., and Gore [
35] made a quite remarkable approach by highlighting factors such as speed, accuracy, and task analysis and tried to illustrate the performance relation to the quantification of the workloads of pilots. A load-balancing algorithm could analyze and manage the load distributed on any public service in the same way that protocols operate in the IT world. Compared to static ones, dynamic algorithms gather and combine the load information to make decisions for load balancing. Sharifian et al. [
36] proposed an approach that involves predicting server utilization and then correcting this prediction using real-time feedback on response time, while Xu and Wang [
37] introduced a modified round-robin algorithm designed to enhance web server performance, particularly during periods of transient overload. In the same way, Diao, Y., and Shwartz [
38] illustrated the development of autonomic systems for IT service management, aiming to improve service quality while optimizing costs. They employ automated data-driven methodologies encompassing workload management through feedback controllers, workforce management using simulation optimization, and service event management with machine learning models. Real-world examples from a large IT services delivery environment validate the effectiveness of these approaches. The impact of workloads on HR was highlighted by Razzaghzadeh et al., who introduced a novel load-balancing algorithm designed for expert clouds that emphasized both load distribution and efficient task allocation based on a mathematical model. It leverages nature-inspired colorful ants to rank and distinguish the capabilities of human resources (HRs) within a tree-structured site framework. Tasks and HRs are labeled and allocated by super-peer levels using Poisson and exponential distribution probability functions. The proposed method enhances throughput, reduces tardiness, and outperforms existing techniques when tested in a distributed cloud environment [
39]. In many research works, multitasking has been considered in order to reduce and effectively cope with workloads. Bellur et al. [
40] deployed cognitive theories to understand multitasking’s effects in educational settings, emphasizing students’ technology use. They investigated multitasking’s impact on college GPAs and distinguished multitasking efficacy and additional study time as covariates. The study revealed that multitasking during class negatively affects GPA, surpassing the influence of study time. Rubinstein et al. [
41] also indicated that, regardless of task type, people experienced time loss when transitioning from one task to another. Additionally, the duration of these time costs escalated with the intricacy of the tasks, resulting in notably longer switching times between more complex tasks. Moreover, the time costs were higher when individuals shifted to less familiar tasks. In theory and experimentation, significant advancements have been achieved concerning how cognitive control affects immediate task performance. Increased cognitive control requirements during encoding consistently lead to a deceleration in performance and an uptick in error rates according to Reynolds [
42] and Meier and Rey-Mermet [
43], while Muhmenthaler and Meier proved in their research that task switching consistently impaired memory across all experiments conducted [
44]. Mark et al. [
45] pointed out in their study that task interruptions “cost” an average of 23 min and 15 s to get back to the task. On average, individuals who frequently switch tasks experience a loss of focus for just tens of minutes each time. Consequently, their efficiency is reduced by approximately 20–40% [
41].
However, the aforementioned approaches lack a reliable management system for evaluating employee productivity. Traditional evaluation systems, characterized by subjectivity from experts and bureaucratic processes, fall short in providing adequate performance appraisals. This inadequacy underscores the urgent need for transformative strategies in public administration, especially in the context of the ongoing digital transformation era. This research emphasizes the integration of objective efficiency assessment mechanisms, moving away from traditional systems that rely on domain experts and their inherent subjectivity. Introducing a time variant factor that measures the execution of each task by a central HRMS is an innovative approach. Therefore, the novel contribution of this study extends further by utilizing the capabilities of employees within each public body to achieve a balanced workload network. Merely assessing and classifying employees is insufficient without a strategic framework that demonstrates how to effectively deploy those results.
3. Methodology
3.1. Tasks and Capacity Factor
Every country has laws defining the number of and duration of time spent on breaks for each employee. In Greece, if the continuous daily working hours exceed 2 h, public servants are entitled to a minimum 15 min break, during which they can temporarily leave their workstations. This break cannot be accumulated at the start or end of the workday, as per the adjustments made by Law 4808 of 2021, Article 56, and the 540/2021 decision of the Supreme Court, while the required skill set for a public servant is based on decision No. 540/2021 from the Greek Council of State according to Council Directive 90/270/EEC.
We base our hypothesis on the fact that each employee possesses a unique set of complex skills that influence their ability to complete tasks within a specified time frame, referred to as the “Time Factor”. As each employee is bound to a specific profile, indicated by Michalopoulos et al. [
6], each individual can be treated as a separate node with designated skills for its properties, leading to a unique Time Factor. As such, for every employee or node, a specific Time Factor is assigned based on their skills and is represented by factors K1 through K4 according to Michalopoulos et al. [
6].
The selected factors are
K1: academic skills (number of Bachelor’s degrees, availability of Master’s degree, certification from the National School of Public Administration, and PhD diploma);
K2: working experience in the public sector (number of years, with a maximum number of 35, plus type of responsibility);
K3: working experience in the private sector (number of years, with a maximum number of 35, plus type of responsibility);
K4: Age in number of years within the range of 20–67.
For each employee with profile
, a task with ID
is assigned.
Tasks are categorized according to complexity into
, indicating a proportional increase in their execution times, and
is considered the task with minimum complexity. Therefore,
= task weight x, which indicates the complexity of each task as measured in a specified time unit
, which can be mins, hours, etc., in terms of the metric.
When a task is allocated to employee
, the associated timer
is triggered and continues until the task is completed.
For as long as the timer runs, each profile is assigned with various tasks, but all of them will be dependent on the base . So for a given number of samples n, the Time Factor is denoted as the mean value of all timers of tasks proportional to .
Therefore, for any given profile
and task
,
In continuation of the previous assumptions, each employee profile has a unique capability to accomplish tasks per time unit TU; this is designated as the Capacity Factor. Therefore,
So from the assumptions above, it is deducted that:
, where
is the skill identification for each employee according to Michalopoulos et al. [
6].
Figure 1 illustrates the procedure, which involves a multi-step approach to regulate load control in a network using four factors,
through
, as inputs. These steps are outlined and explained as follows: Input to ANFIS: The four factors
,
,
, and
are utilized as inputs to the adaptive neuro-fuzzy inference system (ANFIS). The role of the ANFIS here is to process these input factors and produce an output termed the Time Factor. Derivation of Capacity Factor: Once the Time Factor is obtained from the ANFIS, it serves as an intermediary variable for the calculation of the Capacity Factor. The Capacity Factor is a critical metric that quantifies the efficiency of the node within each public body. Load control mechanism: Each node within the network utilizes the derived Capacity Factor to regulate its load. This process is part of a load control mechanism, wherein the nodes adjust their operational loads based on the Capacity Factor to optimize overall network performance. The Capacity Factor (CF) is mathematically dependent on the four initial factors (K1 to K4). This relationship can be expressed as a function CF = f(K1,K2,K3,K4), indicating that the Capacity Factor is a resultant function of the inputs K1, K2, K3, and K4.
The public body now consists of a multi-core “equivalent to IT based platform”, where each employee (
) is a potential CPU “running” at a distinct Capacity Factor. For the rest of the study, each employee will be denoted as
(node
) such that:
In the current research, the following primary tasks (
Table 1) for areas of interest were identified:
Table 2 provides the durations, in minutes, for draft tender design (
) and international tender design (
) from 15 distinct profile samples according to Giotopoulos et al. (2023) [
7].
is always the baseline for Time Factor calculations on any given profile
. The Time Factor
Table 2 presented below showcases fifteen sample data records received from the system. It delineates the variations in the time taken to accomplish each sample task.
and
indicate the samples taken for Tasks
and
, respectively.
3.3. Load Control and Loadability
Effective load management across the interconnected nodes is fundamental in every interconnected network that handles traffic. The task acceptance limit (TAL) is introduced as the rate, in
, at which each node accepts and executes successful tasks.
is defined as hourly, daily, weekly, or monthly (in minutes) depending on the time framework imposed by the public service supervisor and is the remaining time for the execution of tasks in the queue. The rate is adjusted automatically per minute by the load control function. Its maximum value is set based on the Capacity Factor.
So for a period of
, where 1
and 1
(
) are accomplished by
, it can be deducted that:
What is important is that the daily time framework is different for each country depending on national laws. For eight hours of work,
is
mins from (
10):
In real-time systems, instantaneous monitoring of processor load is conducted at regular intervals during minuscule time segments. However, the proposed solution deviates from this real-time IT-based monitoring approach since supervisors’ allocated time for a task queue is task-specific and predetermined. This signifies that in our envisaged scenario, the node load (
) shall be computed by aggregating the weight of each task within the node queue, factoring in the predefined remaining time (
) for all the tasks in the queue, as shown in
Figure 2 below. In addition,
is defined as the total time of all tasks the node currently possesses in the buffer queue.
where
k is the total number of tasks
, and
is the remaining time in minutes.
is the result of
, while
is updated every minute. The following Algorithm 1 illustrates the Node Load calculation.
Algorithm 1 Node Buffer |
- 1:
REPEAT - 2:
Initial - 3:
/*Calculate Node Load*/ - 4:
For each minute , is updated and decreases by . - 5:
If is executed within the same time period, proceed to next up to the end of - 6:
UNTIL the end of timer .
|
So in the case of and minutes, all tasks will be executed in a single cycle, leaving the remaining minutes either idle or ready for task reception.
We introduce the term
(calculated maximum task acceptance limit), which estimates the rate of
tasks, which the system accepts when each node operates at maximum load
.
is the maximum estimated number of tasks, based on minimum task weight
, that can be processed per time unit (Giotopoulos et al. [
7]) when the exchange operates at maximum task load (loadability).
So what represents the maximum load capacity for each individual node?
We define loadability as the upper limit for the task load each node Ni can process. The limit is set by (
1) and (
2) as described above, real time delays, and risk of task deployment due to employee performance; the limit is expressed as a percentage of the total available load.
Therefore, taking into account that every node reaches its maximum operational load without degradation at MaxContinousTime per (ValidBreak + MaxContinousTime), from (
1) + (
2) + (
16):
From a theoretical perspective, it is evident that achieving an optimized distribution of the task workload is essential for reaching the highest level of efficiency in any organizational setting. Ideally, each employee should dedicate approximately 88.89% of their total working time to tasks. When the workload surpasses this upper threshold, it inevitably leads to delays and inefficiencies emanating from the employees.
Load control is a vital mechanism that ensures each node within a system remains within a designated protected load zone, preventing it from being overwhelmed by excessive load beyond its handling capacity. The objective is to guarantee successful task throughput even under conditions in which the load surpasses the predefined limit that the node can effectively manage.
In the absence of a load control function safeguarding the system, throughput would exhibit a sharp decline in efficiency at an early overload stage. To sustain optimal task throughput, it becomes imperative to regulate the system’s load by redistributing tasks appropriately or rejecting tasks proportional to the system’s current load. So what is the connection between stress level due to overload and employee performance?
From
Figure 3 above, according to Corbett et al. (2015), it becomes clear that as the level of stress becomes too high, performance decreases [
46]. As the task-related load increases, stress on every employee also increases proportionally. Similar to our case, there is a direct dependency between employee performance and load distribution.
Tasks are never executed by following a strict FIFO queue simply because no such mechanism can be applied when working in a dynamic working environment. A dedicated buffer will be kept for all incoming tasks, but the one executed each time will be priority dependent. There are distinct priority scales,
, with
indicating the highest degree. Thus, as shown in
Table 3 for every task with
k,
:
Although the list of tasks cannot be considered to be large enough to waste time during indexing, a binary search will be deployed. The worst-case scenario for a binary search occurs when Tki is not in the list and the correct insertion point needs to be determined. In this case, for a total number of n tasks, the worst-case time complexity is , which is still quite efficient for large lists compared to a linear search ().
3.9. Apache Spark
Apache Spark, developed at UC Berkeley’s AMPLab [
47,
48], is a robust platform for processing large-scale data. It boasts a hybrid framework that seamlessly integrates batch and stream processing capabilities. Unlike Hadoop’s MapReduce engine, Spark performs excellently because of its innovative features [
49].
Spark’s greatest strength in batch processing lies in its utilization of in-memory computation. Unlike MapReduce, which frequently reads from and writes to the disk, Spark primarily operates within memory, significantly enhancing processing speed. This advantage is further amplified by Spark’s holistic optimization techniques, which analyze and optimize entire sets of tasks preemptively. This optimization is facilitated by directed acyclic graphs (DAGs), which represent the operations and data relationships within Spark. Spark employs resilient distributed datasets (RDDs), which are read-only data structures that are maintained in memory to ensure fault tolerance without constant disk writes to support in-memory computation [
50].
Apache Spark boasts several notable features. Its remarkable speed, which is a hundred times faster than Hadoop and ten times faster than disk access, is particularly noteworthy. Additionally, Spark offers exceptional usability by supporting multiple programming languages like Java, Scala, R, and Python, allowing developers to leverage familiar languages for parallel application development [
51].
Furthermore, Spark facilitates advanced analytics beyond simple maps and reduces operations, including SQL queries, data streaming, machine learning, and graph algorithms. Its versatility extends to deployment options, as it can run on various platforms such as Apache Hadoop YARN, Mesos, EC2, Kubernetes, or in a standalone cluster mode in the cloud. It integrates with diverse data sources like HDFS, Cassandra, and HBase.
In-memory computing is a pivotal feature of Spark that enables iterative machine learning algorithms and rapid querying and streaming analyses by storing data in server RAM for quick access. Spark’s real-time stream processing capabilities, fault tolerance, and scalability make it a versatile and powerful tool for data-intensive applications [
47].
The architecture of Apache Spark is structured around a controller node, which hosts a driver program responsible for initiating an application’s main program. This driver program can either be code authored by the user or, in the case of an interactive shell, the shell itself. Its primary function is to create the Spark context, which serves as a gateway to all functionalities within Apache Spark. The Spark context collaborates with the cluster manager, who oversees various job executions.
Both the Spark context and the driver program collectively manage the execution of tasks within the cluster. Initially, the cluster manager handles resource allocation, dividing the job into multiple tasks distributed to the worker or agent nodes. Upon creation, resilient distributed datasets (RDDs) within the Spark context can be allocated across different agent nodes and cached for optimization purposes [
51].
The agent nodes assume responsibility for executing the tasks assigned to them by the cluster manager and, subsequently, return the results to the Spark context. Executors are integral components of the architecture and perform the actual task execution. Their lifespan coincides with that of Spark itself.
The number of worker nodes can be increased to enhance system performance, allowing jobs to be further subdivided into logical portions, thereby optimizing resource utilization and task execution. This scalable architecture ensures efficient processing of large-scale data tasks within Apache Spark.