Next Article in Journal
A Smartcard-Based User-Controlled Single Sign-On for Privacy Preservation in 5G-IoT Telemedicine Systems
Next Article in Special Issue
Custom Outlier Detection for Electrical Energy Consumption Data Applied in Case of Demand Response in Block of Buildings
Previous Article in Journal
Thermionic Electron Beam Current and Accelerating Voltage Controller for Gas Ion Sources
Previous Article in Special Issue
Design, Implementation, and Configuration of Laser Systems for Vehicle Detection and Classification in Real Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Heating Homes with Servers: Workload Scheduling for Heat Reuse in Distributed Data Centers

by
Marcel Antal
1,
Andrei-Alexandru Cristea
2,
Victor-Alexandru Pădurean
1,
Tudor Cioara
1,*,
Ionut Anghel
1,
Claudia Antal (Pop)
1,
Ioan Salomie
1 and
Nicolas Saintherant
3
1
Computer Science Department, Technical University of Cluj-Napoca, Memorandumului 28, 400114 Cluj-Napoca, Romania
2
Physics Department, Merton College, Merton St, Oxford OX1 4JD, UK
3
Qarnot Computing, 40–42 Rue Barbès, 92120 Montrouge, France
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(8), 2879; https://doi.org/10.3390/s21082879
Submission received: 26 March 2021 / Revised: 16 April 2021 / Accepted: 16 April 2021 / Published: 20 April 2021

Abstract

:
Data centers consume lots of energy to execute their computational workload and generate heat that is mostly wasted. In this paper, we address this problem by considering heat reuse in the case of a distributed data center that features IT equipment (i.e., servers) installed in residential homes to be used as a primary source of heat. We propose a workload scheduling solution for distributed data centers based on a constraint satisfaction model to optimally allocate workload on servers to reach and maintain the desired home temperature setpoint by reusing residual heat. We have defined two models to correlate the heat demand with the amount of workload to be executed by the servers: a mathematical model derived from thermodynamic laws calibrated with monitored data and a machine learning model able to predict the amount of workload to be executed by a server to reach a desired ambient temperature setpoint. The proposed solution was validated using the monitored data of an operational distributed data center. The server heat and power demand mathematical model achieve a correlation accuracy of 11.98% while in the case of machine learning models, the best correlation accuracy of 4.74% is obtained for a Gradient Boosting Regressor algorithm. Also, our solution manages to distribute the workload so that the temperature setpoint is met in a reasonable time, while the server power demand is accurately following the heat demand.

1. Introduction

Data centers (DCs) are consuming about 2–3% of the total electrical energy generated worldwide, thus, they are becoming a global problem. The high energy demand is used not only for the DCs’ primary business objective, which is to execute their client’s workload, but also to maintain temperature conditions for the safe operation of IT equipment. As result, dealing with excess heat has become an expensive process that is negatively affecting the DCs’ profit and sustainability [1]. The continuous hardware upgrades of computing resources that are increasing the power density of the processor make the cooling processes even more complex. This generated even higher energy demands for cooling systems in an effort to remove the heat produced by the computing resources. Studies show that DCs in 2019 consumed over 1900 MW of energy, while the associated heat generated is sufficient to heat around 2 million households need 20 GJ of heating on average [2].
In current DCs, the cooling processes are executed continuously to remove the heat generated by the computing resources and transfer it to a heat exchanger that uses air or liquid coolant [3]. This is rather inefficient, as the energy is consumed twice: first by the computing resources to execute the clients’ workload and second by the cooling system to dissipate the accumulated heat. To address this issue, new research directions have emerged lately aiming to reuse the otherwise wasted heat of the DCs in nearby district heat grids [4,5,6].
Despite these recent efforts, only a few DCs are effectively reusing the generated heat and only a fraction of the excess heat is being recovered [7,8]. There are several reasons making heat reuse difficult. First, the DCs should be in an urban agglomeration benefiting from policies, operations, and infrastructure that enable the smart distribution of thermal energy. Few cities are offering the needed conditions for DCs to re-use their heat. A positive example is Stockholm where almost 10 percent of the city’s heating needs are assured by using heat recovery [9,10]. Second is the relatively low quality of the recovered heat and the losses that occur when it is transported over long distances. To cope with this problem, systems were designed to transfer the absorbed heat at a higher temperature, making it suitable for long-distance transportation. Heat pumps are used to increase the temperature of the recovered heat to make it more marketable. With the help of heat pumps, the heat generated by servers at around 40 degrees Celsius can be transferred to heat water to around 80 degrees Celsius, suited for long-distance transportation in the nearby residences. At the same time, studies have shown that the coolant flow rate and server room outlet temperatures are important factors for the system’s overall efficiency. The third is the concern regarding the exploitation of DCs’ thermal flexibility of the server rooms in safe conditions for the IT equipment’s operation [11]. The equipment can overheat leading to malfunctions if the temperature within the server room increases to generate more heat. The formation of hot spots needs to be prevented using complex management strategies and accurate simulations of the thermodynamic processes within the DC are used to avoid dangerous hot spots [12,13].
In this paper, we address the problem of DCs heat reuse from a novel perspective while considering the case of distributed DCs. In such a design, the IT equipment is not deployed in a server room but is distributed and deployed into buildings and used to provide heat for the tenants while executing the workload (see Figure 1).
The aforementioned design can be more energy efficient because it will also reuse the electricity that is normally used for space heating for executing the workload. The IT equipment is used as a primary source of heat being deployed in the building’s rooms, which eliminates the costs associated with the cooling processes. The building network connection is used to get the delay-tolerant workload to be executed.
In summary, the paper provides the following contributions:
  • Definition of the thermal aware workload scheduling for the distributed DC case as a constraint satisfaction problem aiming to meet the workload service level agreements and at the same time meet the heat demand of the tenants.
  • Definition of a thermodynamic model to accurately estimate the heat demand needed to be generated by the IT equipment and the workload to be allocated for execution to meet the temperature setpoint defined by the tenant.
  • Development of a machine learning-based model for learning and correlating the heat demand with monitored data related to the actual temperature in the room, the temperature of the heat generated by the IT equipment, and temperature setpoint.
  • The heat models and workload scheduling solution were tested and validated considering the characteristics and actual monitored data from an operational distributed DC, with the results being promising in terms of meeting the heat demand and heat model’s correlation accuracy.
The rest of the paper is structured as follows: Section 2 presents the related work on DC heat reuse and workload scheduling models; Section 3 defines the thermal aware workload scheduling in a distributed DC, Section 4 describes the models defined to determine the heat demand and correlation between the workload to be executed and the temperature of the generated heat, Section 5 presents experimental results for using test data from an operation distributed DC, while Section 6 concludes the paper.

2. Related Work

Most approaches in the literature address the heat reuse of the common type of DC in which all the IT equipment is hosted in one building (i.e., a server room) and features a support infrastructure for power and cooling management [6,14,15]. Several heat reuse options are proposed such as for district heating, [7,8,9] hot water grid [16,17], or nearby office buildings [4,5,18]. The heat reuse policy and infrastructure are well established in Nordic European countries [6,9,10], thus the DC potential for heat reuse is usually analyzed considering this use case. The efficient heat reuse will provide new revenue streams for DCs but at the same time, several research challenges still need to be faced such as the low-grade waste heat generated by the servers, especially in the case of the air-cooled DCs and the high investment costs [1,19]. These investments usually address the deployment and installation of heat pumps that are used for raising the quality of the heat [20,21,22]. District heating is seen as one of the most promising alternatives for residual heat recycling. The DCs’ waste heat has the potential of replacing natural gas-based heat, bringing considerable cost savings and a lower carbon footprint to local communities [19,23].
In the area of DCs’ heat reuse, two main research topics have a strict relation to this paper’s objective and contributions: modeling and simulating the thermal characteristics of the DC and the thermal aware workload scheduling.
The first topic addresses the development of models to study the thermodynamic processes inside a DC and to determine the heat generation, transfer, and reuse characteristics [4,24,25,26]. The thermodynamics impose limits on both the maximum allowable temperature of the microprocessors and the coefficient of performance of the heat pumps [27,28]. Setting higher temperature setpoints in the server room is proposed and used to improve the quality of the recovered heat [4,29]. In this case, accurate thermal models of the server room are developed to predict the temperature variations, detect the formation of hot spots which may lead to equipment malfunctioning, and evaluate alternatives in cooling system configurations [3,11]. Computational Fluid Dynamics (CFD) models of the server room are used to run simulations for studying the interactions between servers and cooling units and their effect on the heat and temperature distribution [5,13,30]. The server heat generation and dissipation rates are analyzed and used to set the recommended temperature values for inlet air into the server room [15,31]. The CFD models of the IT server room are used to analyze the supply air temperature of the cooling system units and the inlet air temperature to find the allowed range of temperature for not damaging the equipment and activating the cooling systems [11]. As a result, higher temperature setpoints can be used for short periods while techniques such as pre- or post-cooling may be used for thermal profile adaptation [4,32]. Even though the CFD simulations provide accurate temperature predictions, they are computationally expensive [33]. Tradeoffs should be done among the accuracy of the simulation, execution time, and resource overheads. Mathematical models and machine-leaning-based approaches are used to address such tradeoffs [4,29,34] with varying levels of success. Mathematical models of the temperature evolution in a server room are presented in [35,36] addressing the thermal behavior concerning heat generation, circulation, and air-cooling system using Navier–Stokes equations expressing thermal laws or by using fast approximate solvers [37,38].
The combination of thermodynamics processes simulations with machine learning techniques offers promising results for determining a set of parameters empirically from monitored data [5,39,40]. In [39], a thermal forecasting model is defined and used for predicting temperatures surrounding servers in data centers. Continuous streams of temperature and airflow measurements are collected for obtaining online predictions with real-time sensor measurements. In [40], a fast converging solution is proposed using both a feed-forward network and a dynamic recurrent artificial neural network. The neural networks learn incrementally, using the incoming stream of data samples. Adaptiveness is presented as an essential feature, as the model can learn the characteristics of a server room with minimal training, and then it may continuously adapt to new data fed without retraining [41]. In [5], a heat reuse model is defined which combines the simulation of the thermodynamic processes in a server room with deep learning processes. Multi-Layer Perceptron neural networks are used for predicting the hot air temperature distribution in the server room. Some models use a set of parameters from the server room that are relevant for the thermodynamics processes and use machine learning to predict their evolution. Gradient boosting decision trees, artificial neural networks, or deep learning models are used to predict the server room temperature [42,43]. Finally, Grammatical Evolution techniques [44] and Environmentally Opportunistic Computing [45] are used for analyzing server and inlet air temperatures and predicting the temperatures, in conjunction with thermal models of DCs. The models should reflect the physical nature of the system, rather than fitting the data purely mathematically. This is enforced using rules for a model’s generation expressed grammars written in Backus–Naur form [46].
Thermal aware workload scheduling algorithms for heat reuse are derived from scheduling algorithms developed to minimize the cooling system energy consumption [12,47]. The main goal of these scheduling algorithms is to distribute the workload in a data center to maintain a low ambient temperature and avoid hotspot formation [47,48]. In the case of heat reuse, the workload scheduling aims to increase the efficiency of heat pump operation and to meet the heat demand of the district heating network [49]. They rely on an optimization problem, defined either reactively or proactively, whose complexity is highly dependent on the representation of the thermodynamics processes and the correlations considered among workload and power and heat demand [50,51]. Workload placement strategies considered are based on zones discretization, minimize the heat recirculation, and prioritize the servers for task allocation by observing hot airflow within the DC [47,48,51]. Scheduling methodologies common in DCs such as first-come-first-serve or backfilling do not usually consider the thermal perspective [52,53]. Machine learning-based models are proposed to infer scheduling policies with thermal features. Server room temperature prediction approaches using machine learning are proposed in conjunction with scheduling algorithms to avoid thermal stress and hotspot formation [54]. In [4], thermal aware workload scheduling is proposed to adapt the DC heat generation to the district heating demand and maximize the waste heat reuse. Neural networks are used to learn the heat generation and heat distribution in the server room. Thermal aware scheduling may consider different heuristics such as tasks and servers’ classification in hot or cold thermal prediction models, node ranking based on heat generation features, etc. [48,51,55]. In [56], an optimization problem is proposed for thermal scheduling considering the optimal setpoints for the workload distribution and the temperature in the server room. Heat flow models are proposed for determining temperatures in case of a thermal aware workload scheduling policy, while a heat recirculation matrix is used to define the thermal influences between the servers [57,58]. Algorithms are prosed to allocate workload on multiprocessors while minimizing the makespan and temperature constraints [59,60]. They aim to reduce the chip temperature while meeting the workload SLA, with the optimization problem being usually modeled as a mixed-integer linear program. Thermal aware task scheduling approaches to adjust CPU frequency based on Dynamic Voltage Frequency Scaling [61] are used to manage the energy demand and heat generation [62,63,64]. In [65], the authors build a steady and dynamical thermal interaction model of the DC. Based on these models, a task assignment and frequency optimization are performed in the first optimization stage, while the second stage uses a model predictive control (MPC) to represent the optimization problems that aim to minimize the cooling system power demand. In [12], a thermal aware consolidation mechanism is defined using a heat recirculation matrix and a set of bio-inspired algorithms that minimize overall DC energy consumption. Finally, in [56], the scheduling optimization problem is defined by considering the energy footprint reduction with thermal exchanges while incorporating both temperature and workload constraints.
After analyzing the existing state of the art, we did not find any relevant literature approach that addresses the thermal aware workload scheduling in the case of distributed DCs while considering the IT equipment as a primary source of heat. Several papers advocate the Data Furnace as the method of heating residential homes by deploying IT equipment in their premises [66,67]. Nevertheless, they are at the stage of ideas promoting some of its advantages such as a smaller carbon footprint or a reduced total cost of ownership per server without offering an actual scheduling solution.
In our paper, we address the identified knowledge gap in the literature by proposing a thermal-aware workload scheduling solution for distributed DCs. Our approach can consider both the service level agreements constraints of the workload to be executed and the heat demand of the residential home’s tenants. To accurately estimate the heat to be generated by the IT equipment for meeting the heat demand levels, we define thermodynamic and machine learning models. They evaluate the heat demand based on the temperature of the heat generated by the IT equipment to determine the workload to be allocated for execution such that the temperature setpoint defined by the tenant is meet. The scheduling algorithm and heat models are evaluated considering relevant data sets from an operational distributed DC showing promising results.

3. Thermal Aware Workload Scheduling

The distributed DC should be modeled as a collection of N isolated IT equipment (i.e., servers or micro data centers) deployed in residential buildings. Each one uses the building internet network and is deployed in a room that needs to be heated:
D C D i s t r i b u t e d = { < r o o m k , s e r v e r k > } , k = 1 , N .
The IT equipment offers a direct heat source, converting the electrical energy consumed for workload execution into thermal energy (i.e., a computing heater). The thermal energy is dissipated through radiators in the surrounding air, heating the room as conventional heaters to maintain the thermal comfort of the inhabitants. The objective, in this case, is to schedule the workload tasks to be executed by the IT equipment deployed in the room k to bring and maintain the ambient temperature T R o o m k close to a set point T s e t p o i n t k temperature desired by the dwellers:
r o o m k : < T R o o m k , T s e t p o i n t k > .
The workload scheduling should consider the dockerized tasks and Service Level Agreements constraints in terms of computational resources to be allocated and execution time deadline. Also, the number of task migration should be kept as low as possible to minimize its impact on task SLA.
We defined the workload to be scheduled for execution over an interval [ 0 ] as a set of M tasks:
W o r k l o a d = { T a s k j , j = 1 , M } .
Each task specifies the computational resources that need to be allocated, estimated execution time, and deadline according to the Service Level Agreements:
T a s k : < C P U , R A M , H D D , t e x e c u t i o n , t d e a d l i n e > ,       t e x e c u t i o n , t d e a d l i n e [ 0 ] .
The rooms should be heated by reusing residual heat generated by the servers. For each server, the computational resources, the idle, and maximum power consumption are specified:
S e r v e r : < C P U , R A M , H D D , P I D L E , P M A X > .
To allocate tasks on servers to be executed during the interval [ 0 ] , a scheduling matrix W s c h e d u l i n g is defined. The matrix stores the starting time for tasks execution and their allocation on specific servers:
W s c h e d u l i n g R N × M , W s c h e d u l i n g [ S e r v e r k ] [ T a s k j ] = t s t a r t k j .
If there is a task j scheduled to the executed-on server k, then:
0 < t s t a r t k j < T
otherwise, the t s t a r t k j = 0 .
The subset of tasks W a l o c a t i o n k scheduled to be executed to each server k , from the N locations where the DC IT equipment is distributed, is determined using the task scheduling matrix:
W a l o c a t i o n k = { T a s k j | W s c h e d u l i n g ( k ) ( j ) > 0 , j { 1 M } , k { 1 N } } .
In the process of determining the scheduling matrix, several constraints need to be met. The first one is referring to the execution time of a scheduled task that is not allowed to exceed the specified execution deadline:
t s t a r t k j + t e x e c u t i o n j t d e a d l i n e j .
The second set of constraints are referring to the relation between the total computational resources requested by the tasks scheduled for execution on a server and the server’s available resources. The computational resources allocated to the tasks scheduled on a server k should be less than the total resources of that server:
t [ 0 ] , C P U k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M C P U j ( t )
t [ 0 ] , R A M k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M R A M j ( t )
t [ 0 ] , H D D k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M H D D j ( t )   .
Considering the resources allocated to tasks, the computational resources utilization levels for a server k are computed as:
μ s e r v e r k ( t ) = { μ C P U k ( t ) , μ R A M k ( t ) , μ H D D k ( t ) }
μ C P U k ( t ) = j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M C P U j ( t ) C P U k ( t )
μ R A M k ( t ) = j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M R A M j ( t ) R A M k ( t )
μ H D D k ( t ) = j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M H D D j ( t ) H D D k ( t )   .
The power consumed by the server is determined using the utilization ratio and the power characteristics of the servers as:
P s e r v e r k ( t ) = f p o w e r ( μ C P U k ( t ) , μ R A M k ( t ) , μ H D D k ( t ) , P i d l e k , P M A X k ) .
For the function f p o w e r , we used a linear model to compute the power considering the idle and maximum power of the server k and the CPU utilization level [68]:
P s e r v e r k ( t ) = P i d l e k + μ C P U k ( t ) ( P M A X k P i d l e k ) .
The energy consumed by the s e r v e r k over the time interval [0…∇] can be computed as an integral of the power over this time window:
E s e r v e r k = 0 P s e r v e r k ( t ) d t   .
Among the computational resources considered for a server, the main heat source is the CPU, which is responsible for 30% up to 65% of the total heat dissipated and also has the highest temperature. Other resources have less impact with varying proportions, thus we denoted them with parameters that can be determined empirically. According to the law of energy conservation, the electrical energy consumed by the servers can be transformed into heat that is dissipated in the room where the servers reside over the interval [0…∇]. This heat can be split according to the n c computational resources of the server, with each having been assigned a weight ω according to the proportion of the thermal energy generated:
Q s e r v e r k ( ) = i = 1 n c ω ( i ) E c o m p o n e n t k ( i )   .
Considering the tasks allocated over the interval [0…∇], the function f s c h e d u l e : R M R n c estimates for each server, the energy consumed by each component, which can be used to estimate the server heat generation.
E c o m p o n e n t k = f s c h e d u l e ( W a l o c a t i o n k )
The heat dissipation in each room leads to a temperature modification according to a function f Q that estimates the ambient temperature T R o o m k modification over the interval [0…∇]:
T R o o m k ( ) = f Q k ( T R o o m k ( 0 ) , Q s e r v e r k ( ) ) .
In our case, the workload scheduling for a distributed DC aims to allocate the workload tasks on the IT servers to optimally generate the heat according to the resident’s demand. The room temperature setpoints need to be reached as fast as possible and are kept constant over the rest of the time interval, with no task migrations, in each of the N locations where the IT equipment is distributed.
The thermal aware workload scheduling problem is modeled as a constraint satisfaction problem to determine the optimal workload scheduling matrix W s c h e d u l i n g such that the set-points temperatures in each of the N rooms are met (see Algorithm 1). Its solving process involves nonlinear programming [69] because of the non-linearities of the objective function and the workload scheduling function f s c h e d u l e and the continuous values of the W s c h e d u l i n g matrix. Also, it is an NP-hard problem [70], thus an approximation algorithm is needed to determine a solution.
Algorithm 1 Thermal aware workload scheduling
Input: D C D i s t r i b u t e d , N the number of distributed servers, M the number of workload tasks, T R o o m the temperatures in the rooms where the servers are distributed, T s e t p o i n t the desired temperature in the rooms.
Output: W s c h e d u l i n g matrix keeping the tasks scheduling on servers
                                  O p t i m i z a t i o n G o a l : d e t e r m i n e W s c h e d u l i n g t o M I N ( k = 1 N d i s t a n c e ( T R o o m k , T s e t p o i n t k ) )
Considering the following constraints for each server k = 1 N
C 1 : W a l o c a t i o n k = { T a s k j | j { 1 M } a n d W s c h e d u l i n g ( k ) ( j ) > 0 }
C 2 : Q g e n e r a t e d k ( T ) = i = 1 n c ω ( i ) f s c h e d u l e ( W a l o c a t i o n k )
C 3 : T R o o m k ( ) = f Q k ( T R o o m k ( 0 ) , Q s e r v e r k ( ) )
C 4 : t s t a r t k j + t e x e c u t i o n j t d e a d l i n e j
C 5 : t [ 0 T ] , C P U k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M C P U j ( t )
C 6 : t [ 0 T ] , R A M k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M R A M j ( t )
C 7 : t [ 0 T ] , H D D k ( t ) j = 1 t s t a r t k j t t s t a r t k j + t e x e c u t i o n j M H D D j ( t )
The main challenge is the unknown nature of the f Q k function that may feature complex representations, making the utilization of optimization algorithms derived from stochastic gradient descent unfeasible [71]. To address this, we split the thermal aware scheduling problem into two subproblems (see Figure 2). For the first subproblem involving estimates for each room k, the heat that needs to be generated by the server to make the transition from T R o o m k ( 0 ) to T s e t p o i n t k . The second one determines the workload scheduling matrix W s c h e d u l i n g such that a large enough workload is scheduled for execution on each server to generate heat to meet the demand. The latter problem can be solved using an adaptation of the Multiple Knapsack Problem [72].
In the case of the first subproblem, we aimed to approximate the heat demand Q D e m a n d k needed over the interval [0…∇] to make the transition from T R o o m k ( 0 ) to T s e t p o i n t k and then keep the temperature constant at the set point level. It can be solved by defining a function f Q 1 that estimates the amount of heat needed to make the temperature transition:
Q d e m a n d k ( ) = f Q 1 ( T R o o m k ( 0 ) , Q i n i t i a l k , T s e t p o i n t k ) .
In the case of the second subproblem, the goal is now to determine the workload scheduling matrix W S c h e d u l i n g so that the heat generated by each server matches closely the heat demand of the corresponding room:
M I N ( k = 1 N d i s t a n c e ( Q d e m a n d k , Q g e n e r a t e d k ) )   .
The optimization still involves nonlinear programming, but the functions are not defined as black-box models. As result, it can be solved using approximation algorithms used to determine solutions for the Multiple Knapsack Problem [72]. Using this time discretization over the interval [0…∇], the scheduling problem can be reduced to an N × ∇ knapsack problem, where the N × ∇ knapsacks volumes correspond to the values of the heat demand to be generated on each of the ∇ intervals in each of the N rooms, while the items being packed are the workload tasks to be deployed on servers and executed.

4. Heat Demand Estimation

The thermal aware workload scheduling model described in the previous section needs accurate estimations of the heat demand to be generated by the IT equipment such that the temperature set point in the room set by the residents is met. In this section, we propose two models for implementing the f Q 1 function used to determine the head demand estimations.

4.1. Server Heat Transfer Model

We defined a thermodynamic model of the server calibrated with measurements. To ease the representation, we assumed that the room temperature T R o o m k is constant during the workload scheduling period [ 0 ] . This means that all the power generated by the server Q s e r v e r is dissipated in the surrounding environment by rising to the ceiling, while cold air from the floor passes through the computing element radiator, leading to an energy loss Q l o s s of the room.
The server power consumption for executing the allocated tasks, the temperature of the heat generated ( T s e r v e r ), and room temperature are linked as follows [73]:
P s e r v e r = Q s e r v e r = c s e r v e r × T s e r v e r t + f a i r × c a × ε s e r v e r × ( T s e r v e r T R o o m )
where c s e r v e r is the server heat capacity, f a i r represents the airflow over the server surface, c a i r is the specific heat capacity of air, and ε s e r v e r is the thermal server efficiency. Thermal efficiency is defined as the ratio of real to maximum power transfer between the server’s body and the airflow and can be experimentally determined by measuring the temperature of emerging airflow denoted as T e x .
ε s e r v e r = P s e r v e r P s e r v e r m a x = f a i r × c a i r × ( T e x T R O O M ) f a i r × c a i r × ( T s e r v e r T R O O M ) = T e x T R O O M T s e r v e r T R O O M
The changes in the server power P s e r v e r influence the temperature of the heat generated. We assumed that the power is changed linearly with respect to time with the ratio R :
P s e r v e r ( t ) = R   × t + P s e r v e r ( 0 ) .
We defined the transient state of server-generated heat temperature T s e r v e r over the time interval in which P s e r v e r changes, while the equilibrium state was defined to be when the P s e r v e r reaches a constant value. Relation (25) can be used to define both the transient and the equilibrium regimes of the computational equipment. The transient temperature is defined in Equation (28) as a solution of (25), as a function of the time coordinate, while the solution in the equilibrium regime is given by Equation (29).
T s e r v e r t r a n s i e n t = [ A × e t ( C s e r v e r f a i r × c a × ε s e r v e r ) + R × t f a i r × c a × ε s e r v e r ] R × C s e r v e r ( f a i r × c a × ε s e r v e r ) 2 + T R o o m + P s e r v e r   ( 0 ) f a i r × c a × ε s e r v e r
T s e r v e r e q u i l i b r i u m = B × e t ( C s e r v e r f a i r × c a × ε s e r v e r ) + T R o o m + P s e r v e r final f a i r × c a × ε s e r v e r
In Equations (28) and (29), the constants A and B are fixed by the initial conditions in each case. The equilibrium regime defined in Equation (29) tends asymptotically to a state independent of time due to the rapid exponential decay of the first term, eventually reaching a state depending only on the power workload and predicting the behavior of T s e r v e r with respect to the change of P s e r v e r . The server temperature after the equilibrium temperature setting can be estimated with Equation (30), derived from (29) when the first term decays to zero.
T s e r v e r f i n a l   T R O O M + P s e r v e r f i n a l f a i r × c a × ε s e r v e r
The parameters used in defining the server as a heater model are detailed in Table 1, mentioning which parameter should be measured, and which are tuned experimentally for a particular model. The parameters from Table 1 that are determined experimentally can be computed while considering a set of measurements from the physical server room, allowing the model to be fitted to the exact configuration of the real-world system.
Finally, after reaching the equilibrium temperature, the assumption that T R o o m is constant can be relaxed, as the computational equipment will dissipate heat in the room, leading to a temperature change. Considering the equations developed above, the heat per unit of time demanded by the servers overtime to pass the transient regime and transition the temperature T S e r v e r to a temperature T s e r v e r e q u i l i b r i u m close to T s e r v e r f i n a l can be computed as the server demand over the transient regime:
Q d e m a n d ( t ) = P s e r v e r ( t ) = R × t + P s e r v e r ( 0 ) .
However, to transition the room temperature from T R O O M to T s e t p o i n t , we considered that the server has reached a state sufficiently close to the asymptotic equilibrium and the relation between T s e r v e r f i n a l and T R O O M is approximately linear, so for small changes and fluctuations, the difference between the two temperatures is largely unchanged. Using simple thermodynamic considerations once again, it can then be stated that:
( P s e r v e r f i n a l P l o s s ) t = M × c a × ( T s e t p o i n t T R O O M )
which is an expression of the energy required to heat the room from T R O O M to T s e t p o i n t in time t after the server has reached its equilibrium temperature, M is the mass of the air in the room, computed by multiplying the room volume V by the air density ρ a i r , and P l o s s is the energy lost per unit time by the room due to air drafts and imperfect thermal insulation. Thus, from (32) we can deduce the expression:
P s e r v e r f i n a l = P l o s s + M × c a × ( T s e t p o i n t T R O O M )   t .
We can now obtain the more useful expression for Q D e m a n d ( t ) , given T s e t p o i n t for the air in the room, knowing that Q D e m a n d ( ) = P s e r v e r f i n a l and that the increase in power is linear:
Q D e m a n d ( t ) = P l o s s + M × c a × ( T s e t p o i n t T R O O M )   t P s e r v e r ( 0 ) × t + P s e r v e r ( 0 ) , t [ 0 ]
It should be noted that the time in which the room heats up, t , is a parameter that can be chosen freely, and should not be made too short, as this will cause the server to overheat i.e., exceed the recommended functioning temperature of approximately 35 °C. The limiting minimum size t is easily deducible from previous expressions. Finally, if the room is to be heated, P s e r v e r f i n a l must be larger than P l o s s , otherwise, the room will not receive any net heat and will not increase its temperature.

4.2. Machine Learning-Based Model

The machine learning model aims to infer and correlate the heat demand that needs to be generated by the server executing the allocated workload out of monitored data (see Figure 3). We considered that the room temperature is collected using heat sensors placed at a certain distance from the server. The heatsink temperature is collected using the server’s embedded temperature sensor, whereas the power consumption is obtained using a wattmeter.
The machine learning model takes as inputs the actual room temperature T R o o m ( 0 ) , the initial power consumption P s e r v e r ( 0 ) , the desired room temperature, T s e t p o i n t , and provides as output the heat demand that the server should generate by executing workload, to reach the desired temperature, T R o o m ( ) .
All data needs to be smoothed so that outliers are eliminated. We aimed to detect areas of interest, defined by a high correlation coefficient between the heatsink and ambient temperature, limited between two local peaks of the temperatures and power. Thus, we used the Pearson Correlation Coefficient, as defined in Equation (35), to measure the linear correlation between two datasets. Only relevant samples with values larger than a threshold were considered for further processing, namely data where the room temperature, server temperature, and power consumed by the server present similar patterns.
r T h e a t s i n k , T R O O M w i n d o w = c o v ( T s e r v e r , T R O O M ) σ T s e r v e r σ T R O O M > t h r e s h o l d
As the area of interest still may contain data that might be irrelevant for our final purpose, it needed further processing. Only the initial and final temperatures were selected from the scheduling interval [ 0 ] . They represent the relevant information for predicting the power consumption, given the temperature. The initial temperatures can be computed as the mean temperature before a sudden change, whereas the final ones can be computed as the mean temperature after stabilization. The same goes for power, consequently obtaining the four needed values:
[ < T R o o m ( t 1 ) , P s e r v e r ( t 1 ) , T s e r v e r ( t 1 ) > P s e r v e r ( t 1 + ) < T R o o m ( t 2 ) , P s e r v e r ( t 2 ) , T s e r v e r ( t 2 ) > P s e r v e r ( t 2 + ) < T R o o m ( t n ) , P s e r v e r ( t n ) , T s e r v e r ( t n ) > P s e r v e r ( t n + ) ]
where t 1 , t 2 , t n [ 0 ] , are timestamps of the data acquisition.
Several models were implemented and used to learn the behavior of the function f Q 1 used to estimate the heat demand. Table 2 presents the description of the models and their configuration determined empirically on the test data.

5. Evaluation Results

To evaluate the thermal aware workload scheduling and the heat demand estimation models, we considered a test case distributed DC [74] composed of a main DC and a set of edge sites (see Figure 4). Each edge site hosts a workload distribution node, the QBox, and a set of server heater nodes, the QRads, that provide heat by executing the workload provided by the QBox. The QRad is a server-heater with no moving parts and three motherboards equipped with CPUs that execute workload and dissipate the heat in the surrounding environment, leading to a temperature rise. Each QRad node consumes about 400 W of electrical energy and generates an amount of roughly 400 W of heat, depending mainly on the CPU model. The workload that runs on the QRad is composed of tasks that use full CPU resources, such as 3D animation or financial risk computing.
A sensors-based monitoring infrastructure was used to acquire relevant data regarding the Ambiental temperature T R o o m , power consumption of the QRad ( P s e r v e r ) and temperature of the server ( T s e r v e r ) . The data acquired spans over several months, being recorded at a granularity of 10 s. Both the characteristics of the QRad heaters and monitored data acquired by the installed infrastructure are the inputs of our study.
However, the monitored data used to model the thermal behavior of the QRad heaters must be extracted for larger intervals that show suggestive temperature and power changes. Thus, a pipeline of data pre-processing operations was employed to extract the relevant data samples (see Figure 5).
The acquired data were smoothed using an exponential weighted moving average window with the span of 90 data points corresponding to a time window of 15 min. A mean filter with a window span of 60 data points, corresponding to a time window of 10 min, was applied to ignore sudden fluctuations. To find the exact times when the ambient ( T R o o m ) and the server temperatures ( T s e r v e r ) rise or fall together, Pearson’s r coefficient was used. This was applied on a rolling window with the span of 360 data points corresponding to a time interval of 1 h to compute the correlation. The samples corresponding to intervals where the scores exceeded 0.5 were considered. Finally, local maximum and minimum peaks were found over window sizes of 180 data points corresponding to 30 min time intervals. Only samples that started at a local minimum and ended at a local maximum or vice-versa were taken into consideration. The process involved manual inspection of the selected data and fine fitting so that the included regions could also contain the power instant change, which sometimes takes place before a local peak.
Figure 6 shows a relevant data sample obtained by filtering, smoothing, and applying Pearson’s coefficient. The length of the time interval is determined between the start of the temperature change and the end of the temperature change. Analyzing the data set, we determined for each sequence to record most temperature changes and to reach a steady-state situation. Since most of the samples show that the server power change is similar to a linear function, only the final server power consumption was considered ( P S e r v e r ( t n + ) ) . This can be computed based on the initial state parameters at the time t n : the ambient temperature at the time T R o o m ( t n ) , the server temperature, T s e r v e r ( t n ) , and the server power demand P s e r v e r ( t n ) .
Firstly, we considered the server heat transfer model detailed in Section 4.1. The model was calibrated by fitting some of its parameters on the actual data gathered from the QRad heaters. The heat model parameters from Table 1 (i.e., the server heat capacity C s e r v e r , the server thermal efficiency, ε s e r v e r ) have to be computed to fit best on monitored data, while the airflow over the server (i.e., f a i r ) has to also be estimated.
The fitting process consists of feeding a trace of processed data in the form T R O O M ( t n ) , P s e r v e r ( t n ) , T s e r v e r ( t n ) P S e r v e r ( t n + T ) for each QRad server CPU to an optimizer that can compute the model parameters using gradient descent-based algorithms [71]. The processed data samples were split into 80% training data for model fitting, while 20% was used to validate the model. The fitting process is performed iteratively, at each step a sample of training data is read, and the variables C s e r v e r , ε s e r v e r and f a i r were varied from a set of initial values until they fit best the data. The sequential fitting was done using a custom script from the SciPy library, namely the scipy.optimize method [75], applied on each data sample, recording values of C s e r v e r , ε s e r v e r and f a i r from the fit process and using them as initial guesses for the next sample’s fit. Repeating this operation gave us progressively refined values of the server’s parameters.
After having determined the characteristics for each QRad server, we predicted the change in power workload on the testing data. This was done by fixing C s e r v e r , ε s e r v e r , and f a i r to the values obtained, reading T s e r v e r and T R o o m from the data and finally obtaining our prediction for the change in P s e r v e r , which we would compare to the one in the data. A prediction chart is shown in Figure 7, illustrating as a linear function the power change for a server heater (depicted in green line) to increase the temperature (depicted in blue line) to match the requested temperature as closely as possible (depicted in red line), over a time interval of 20 min.
The model was evaluated on a set of scenarios, with both fitting processes described above, achieving the average prediction accuracy from Table 3.
Secondly, we evaluated the machine learning-based heat model presented in Section 4.2. A set of several machine learning algorithms were implemented to model the QRad server heat generation. The Linear and Polynomial Regression, Random Forest Regression, Support Vector Regression, and K Neighbors Regression were implemented using Python’s SciKitLearn [76] library. The Multi-Layer Perceptron was implemented using Keras [77] and Tensorflow [78] while for the Gradient Boosting Regression, the XGBoost’s [79] was used.
The dataset processed using the defined pipeline of operations can be split into train and test subsets, with the proportions of 0.8 and 0.2 of the initial data. The models were validated using 5-fold cross-validation and were evaluated by computing the Root Mean Square Error (RMSE), Root Mean Percentage Square Error (RMSPE), the coefficient of determination R2 (R Squared), the error mean, the error standard deviation, and MAPE. The average results obtained are listed in Table 4.
As Table 4 shows, considering the MAPE and the RMSPE, the best results are obtained by the Gradient Boosting Regression (GBR). These are the most relevant metrics for prediction accuracy, considering the percentages of the errors. They are showing that the GBR model can predict the heater power demand with less than 5% error, corresponding to less than 10 W of power. However, by analyzing the mean and the standard deviation on the error besides the RMSPE and the MAPE, the Random Forrest Regression (RFR) gives better results, showing a small error distribution considering the mean error. Finally, the R2 metric for the two models is closest to 1, with values of 0.94 for GBR and 0.92 for RFR. Looking at all calculated metrics, we considered that the GBR gives the best results considering the datasets collected from the QRad heaters and is most suitable for being used in a thermal aware workload scheduling algorithm.
Thirdly, we evaluated the thermal aware workload scheduling solution presented in Section 3. The heating requirements for a room can be determined by several models used in the industry to determine the recommended size of the heaters [80]. The main factors that influence the required power of the heaters in a room are the volume of the room and the caloric coefficient of the room. The latter has a value between 40 and 70 kcal / m 3 and is influenced by the thermal insulation of the room, the number of exterior walls, the number of windows, and their size and type. The power needed to heat a room according to industry standards can be computed as:
P R o o m R e c c o m e n d e d = V R O O M C R O O M c a l c W c a l
where V R O O M is the volume of the room expressed in cubic meters m 3 , C R O O M c a l is the caloric coefficient of the room expressed in kcal m 3 , and c W c a l is the conversion factor from kcal to W having an approximate value of 1.163 W / mcal .
We determined the required heating power for rooms with different configurations and we assessed the number of QRad heaters needed to be installed (see Table 5). We considered that the heat losses from the room are much smaller than the heat generation ratio.
From the electrical energy consumption perspective, a heating system based on electrical heater radiators would consume the same amount of electricity as the server heaters for generating the same amount of heat. This is a result of the server design having no moving components, thus according to the law of energy conservation, most of the electrical energy consumed by the servers is converted into thermal energy and dissipated as heat in the room. Finally, as the server design is like a standard heater, they are installed in the room in the same positions as standard heaters, requiring no ventilation system for heat recirculation.
To illustrate our case study, we have considered the power and thermal behavior of 4 QRad server heaters containing three motherboards, each with an Intel® Core™ i7-6950X CPU with 10 cores [81], totaling 12 CPU with 120 cores located in a room with a volume of 30 cubic meters, that require roughly 1.8 kWh for heating.
The server’s initial temperature is 29 °C and their total power demand is 540 W, while each of the four servers only has one of the three CPUs active with one core active out of the 10 total cores. The temperature and power status of the server over a time interval of 15 min is displayed in Figure 8, showing a steady-state condition in the room. The desired server temperature is 34.5 degrees Celsius, requiring the heater to generate thermal energy to increase the room temperature.
Figure 9 shows the power demand prediction using the GBR model for total heat demand estimation for the 4 QRad server heaters to transition the room temperature to the set-point temperature. Each of the 12 CPUs’ estimated power demand is roughly 115 W, meaning that the 4 QRads will generate approximately 1.4 kW of heat.
Based on the power demand given by the GBR model for heat demand estimation, the thermal aware workload scheduling algorithm aims to schedule workload for execution so that the CPU usage leads to the required power demand. We considered a set of synthetic tasks running in Docker, each requiring 1 active CPU core processing near 100% and 1 GB of RAM. The task allocation result after solving the optimization problem from Section 3 is shown in Figure 10 left. For each of the 4 servers and their 3 CPUs, the plan activates 8 cores to run at a time interval of 0–100 s and 9 cores to run in the time interval of 100–800 s. Thus, in total 108 cores will be active during the heating period, leading to a total of 1488 W, shown in Figure 10-right, a value close to the predicted power demand of 1400 W estimated by the GBR model.
Finally, we use the data logs to estimate the thermal behavior of the 4 servers for the load computed by the thermal aware scheduling algorithm while considering the predicted load of the GBR model for heat demand estimation. As Figure 11-right shows, the server power demand is close to the predicted heat demand, being able to match it with 97% accuracy. As a result of the power demand and workload execution, the servers dissipate heat in the room, leading to the temperature evolution from Figure 11-left. The blue dotted line shows the room temperature evolution because of simulating the installed QRad behavior with the power demand depicted in Figure 11-right. The green dotted line depicts the room monitored temperature extracted from the data logs, showing a temperature increase close to the simulated QRad behavior and matching the set-point temperature after 800 s.

6. Conclusions

In this paper, we consider the case of distributed DCs and associated problems related to heat reuse when the servers are installed in residential homes are used as a primary source of heat. We propose a workload scheduling solution based on constraint satisfaction to allocate workload on severs for reaching and maintaining the desired temperature set-point in residential homes by reusing their residual heat. Two models were defined to correlate the heat demand with the amount of workload to be executed by the servers: a mathematical model derived from thermodynamic laws calibrated with monitored data and a machine learning model able to predict the amount of workload to be executed by a server to reach a desired temperature set point. The results obtained considering monitored data from an operation distributed DC are promising. The workload scheduling solution can distribute the workload so that the temperature setpoints are meet in a reasonable time, while the server heat and power demand correlation models achieve good accuracy levels.

Author Contributions

Conceptualization, T.C., I.S. and M.A.; methodology, C.A. and M.A. software, M.A., A.-A.C. and V.-A.P.; validation, N.S., A.-A.C. and V.-A.P.; formal analysis, T.C., A.-A.C. and V.-A.P.; investigation, I.A. and I.S.; resources, T.C. and M.A.; data curation, C.A. and N.S.; writing—original draft preparation, T.C., M.A., A.-A.C. and V.-A.P.; writing—review and editing, I.A., C.A. and I.S.; visualization, I.A. and C.A.; supervision, T.C. and I.S.; project administration, T.C.; funding acquisition, T.C. and M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been conducted within the CATALYST project Grant number 768739, co-funded by the European Commission as part of the H2020 Framework Programme (H2020-EE-2016-2017) and it was partially supported by a grant of the Romanian Ministry of Education and Research, CNCS–UEFISCDI, project number PN-III-P1-1.1-PD-2019-0154, within PNCDI III.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pärssinen, M.; Wahlroos, M.; Manner, J.; Syri, S. Waste heat from data centers: An investment analysis. Sustain. Cities Soc. 2019, 44, 428–444. [Google Scholar] [CrossRef]
  2. Bashroush, R. Data Center Energy Use Goes Up and Up and Up, January 2020, Uptime Institute. Available online: https://journal.uptimeinstitute.com/data-center-energy-use-goes-up-and-up/ (accessed on 23 March 2021).
  3. Capozzoli, A.; Primiceri, G. Cooling Systems in Data Centers: State of Art and Emerging Technologies. Energy Procedia 2015, 83, 484–493. [Google Scholar] [CrossRef] [Green Version]
  4. Antal, M.; Cioara, T.; Anghel, I.; Pop, C.; Salomie, I. Transforming Data Centers in Active Thermal Energy Players in Nearby Neighborhoods. Sustainability 2018, 10, 939. [Google Scholar] [CrossRef] [Green Version]
  5. Antal, M.; Cioara, T.; Anghel, I.; Gorzenski, R.; Januszewski, R.; Oleksiak, A.; Piatek, W.; Pop, C.; Salomie, I.; Szeliga, W. Reuse of Data Center Waste Heat in Nearby Neighborhoods: A Neural Networks-Based Prediction Model. Energies 2019, 12, 814. [Google Scholar] [CrossRef] [Green Version]
  6. Wahlroos, M.; Pärssinen, M.; Rinne, S.; Syri, S.; Manner, J. Future views on waste heat utilization—Case of data centers in Northern Europe. Renew. Sustain. Energy Rev. 2018, 82 Pt 2, 1749–1764. [Google Scholar] [CrossRef]
  7. Nielsen, S.; Hansen, K.; Lund, R.; Moreno, D. Unconventional Excess Heat Sources for District Heating in a National Energy System Context. Energies 2020, 13, 5068. [Google Scholar] [CrossRef]
  8. Wahlroos, M.; Pärssinen, M.; Manner, J.; Syri, S. Utilizing data center waste heat in district heating—Impacts on energy efficiency and prospects for low-temperature district heating networks. Energy 2017, 140 Pt 1, 1228–1238. [Google Scholar] [CrossRef]
  9. Su, C.; Dalgren, J.; Palm, B. High-resolution mapping of the clean heat sources for district heating in Stockholm City. Energy Convers. Manag. 2021, 235, 113983. [Google Scholar] [CrossRef]
  10. Taking the Next Steps: Stockholm, the Circular City, DatacenterDynamics. April 2020. Available online: https://www.datacenterdynamics.com/en/analysis/taking-next-steps-stockholm-circular-city/ (accessed on 23 March 2021).
  11. Cho, J.; Park, B.; Jeong, Y. Thermal Performance Evaluation of a Data Center Cooling System under Fault Conditions. Energies 2019, 12, 2996. [Google Scholar] [CrossRef] [Green Version]
  12. Marcel, A.; Cristian, P.; Eugen, P.; Claudia, P.; Cioara, T.; Anghel, I.; Ioan, S. Thermal aware workload consolidation in cloud data centers. In Proceedings of the 2016 IEEE 12th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 8–10 September 2016; pp. 377–384. [Google Scholar] [CrossRef]
  13. Cho, J.; Woo, J.; Park, B.; Lim, T. A Comparative CFD Study of Two Air Distribution Systems with Hot Aisle Containment in High-Density Data Centers. Energies 2020, 13, 6147. [Google Scholar] [CrossRef]
  14. Koronen, C.; Åhman, M.; Nilsson, L.J. Data centres in future European energy systems—Energy efficiency, integration and policy. Energy Effic. 2020, 13, 129–144. [Google Scholar] [CrossRef] [Green Version]
  15. Silva-Llanca, L.; del Valle, M.; Ortega, A.; Díaz, A.J. Cooling Effectiveness of a Data Center Room under Overhead Airflow via Entropy Generation Assessment in Transient Scenarios. Entropy 2019, 21, 98. [Google Scholar] [CrossRef] [Green Version]
  16. Antal, M.; Pop, C.; Cioara, T.; Anghel, I.; Salomie, I.; Pop, F. A system of systems approach for data centers optimization and integration into smart energy grids. Future Gener. Comput. Syst. 2020, 105, 948–963. [Google Scholar] [CrossRef]
  17. Oltmanns, J.; Sauerwein, D.; Dammel, F.; Stephan, P.; Kuhn, C. Potential for waste heat utilization of hot-water-cooled data centers: A case study. Energy Sci. Eng. 2020, 8, 1793–1810. [Google Scholar] [CrossRef]
  18. Swinhoe, D. Switch Datacenters to Heat Homes and Offices Using Residual Server Heat, Data Center Dynamics. 13 January 2021. Available online: https://www.datacenterdynamics.com/en/news/switch-datacenters-heat-homes-and-offices-using-residual-server-heat/ (accessed on 23 March 2021).
  19. Wheatcroft, E.; Wynn, H.; Lygnerud, K.; Bonvicini, G.; Leonte, D. The Role of Low Temperature Waste Heat Recovery in Achieving 2050 Goals: A Policy Positioning Paper. Energies 2020, 13, 2107. [Google Scholar] [CrossRef] [Green Version]
  20. Jirinec, J.; Rot, D. The Control System for Heating of Small Buildings with Heat Recovery unit and Heat Pump. In Proceedings of the 2020 21st International Scientific Conference on Electric Power Engineering (EPE), Prague, Czech Republic, 19–21 October 2020; pp. 1–4. [Google Scholar] [CrossRef]
  21. Härtel, P.; Ghosh, D. Modelling heat pump systems in low-carbon energy systems with significant cross-sectoral integration. IEEE Trans. Power Syst. 2020. [Google Scholar] [CrossRef]
  22. Wang, J.; Zhong, H.; Tan, C.W.; Chen, X.; Rajagopal, R.; Xia, Q.; Kang, C. Economic Benefits of Integrating Solar-Powered Heat Pumps into a CHP System. IEEE Trans. Sustain. Energy 2018, 9, 1702–1712. [Google Scholar] [CrossRef]
  23. Garofalo, E.; Bevione, M.; Cecchini, L.; Mattiussi, F.; Chiolerio, A. Waste Heat to Power: Technologies, Current Applications, and Future Potential. Energy Technol. 2020, 8, 2000413. [Google Scholar] [CrossRef]
  24. Ebrahimi, K.; Jones, G.F.; Fleischer, A.S. Thermo-economic analysis of steady state waste heat recovery in data centers using absorption refrigeration. Appl. Energy 2015, 139, 384–397. [Google Scholar] [CrossRef]
  25. Antal, M.; Cioara, T.; Anghel, I.; Pop, C.; Salomie, I.; Bertoncini, M.; Arnone, D. DC Thermal Energy Flexibility Model for Waste Heat Reuse in Nearby Neighborhoods. In Proceedings of the Eighth International Conference on Future Energy Systems (ACM e-Energy ’17), New York, NY, USA, 16–19 May 2017; pp. 278–283. [Google Scholar]
  26. Dayarathna, M.; Wen, Y.; Fan, R. Data Center Energy Consumption Modeling: A Survey. IEEE Commun. Surv. Tutor. 2016, 18, 732–794. [Google Scholar] [CrossRef]
  27. Matko, V.; Brezovec, B.; Milanovič, M. Intelligent Monitoring of Data Center Physical Infrastructure. Appl. Sci. 2019, 9, 4998. [Google Scholar] [CrossRef] [Green Version]
  28. Rastegarpour, S.; Caseri, L.; Ferrarini, L.; Gehrke, O. Experimental Validation of the Control-Oriented Model of Heat Pumps for MPC Applications. In Proceedings of the 2019 IEEE 15th International Conference on Automation Science and Engineering (CASE), Vancouver, BC, Canada, 22–26 August 2019; pp. 1249–1255. [Google Scholar] [CrossRef]
  29. Grishina, A.; Chinnici, M.; Kor, A.-L.; Rondeau, E.; Georges, J.-P. A Machine Learning Solution for Data Center Thermal Characteristics Analysis. Energies 2020, 13, 4378. [Google Scholar] [CrossRef]
  30. Durand-Estebe, B.; le Bot, C.; Mancos, J.N.; Arquis, E. Data center optimization using PID regulation in CFD simulations. Energy Build. 2013, 66, 154–164. [Google Scholar] [CrossRef]
  31. Kheirabadi, A.C.; Groulx, D. Cooling of server electronics: A design review of existing technology. Appl. Therm. Eng. 2016, 105, 622–638. [Google Scholar] [CrossRef]
  32. Li, Y.; Wang, X.; Luo, P.; Pan, Q. Thermal-Aware Hybrid Workload Management in a Green Datacenter towards Renewable Energy Utilization. Energies 2019, 12, 1494. [Google Scholar] [CrossRef] [Green Version]
  33. Daniels, S.J.; Rahat, A.A.M.; Everson, R.M.; Tabor, G.R.; Fieldsend, J.E. A Suite of Computationally Expensive Shape Optimisation Problems Using Computational Fluid Dynamics. In Parallel Problem Solving from Nature, Proceedings of the PPSN 2018: Parallel Problem Solving from Nature–PPSN XV, Coimbra, Portugal, 8–12 September 2018; Auger, A., Fonseca, C., Lourenço, N., Machado, P., Paquete, L., Whitley, D., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11102. [Google Scholar]
  34. Lye, K.O.; Mishra, S.; Ray, D. Deep learning observables in computational fluid dynamics. J. Comput. Phys. 2020, 410, 109339. [Google Scholar] [CrossRef] [Green Version]
  35. Jonas, M.; Gilbert, R.R.; Ferguson, J.; Varsamopoulos, G.; Gupta, S. A transient model for data center thermal prediction. In Proceedings of the 2012 International Green Computing Conference, IGCC 2012, San Jose, CA, USA, 4–8 June 2012. [Google Scholar] [CrossRef]
  36. Berezovskaya, Y.; Yang, C.; Mousavi, A.; Vyatkin, V.; Minde, T.B. Modular Model of a Data Centre as a Tool for Improving Its Energy Efficiency. IEEE Access 2020, 8, 46559–46573. [Google Scholar] [CrossRef]
  37. Wang, F.; Huang, Y.; Prasetyo, B. Energy-Efficient Improvement Approaches through Numerical Simulation and Field Measurement for a Data Center. Energies 2019, 12, 2757. [Google Scholar] [CrossRef] [Green Version]
  38. Sánchez, C.; Bloch, L.; Holweger, J.; Ballif, C.; Wyrsch, N. Optimised Heat Pump Management for Increasing Photovoltaic Penetration into the Electricity Grid. Energies 2019, 12, 1571. [Google Scholar] [CrossRef] [Green Version]
  39. Li, L.; Liang, C.; Liu, J.; Nath, S.; Terzis, A.; Faloutsos, C. ThermoCast: A cyber-physical forecasting model for data centers. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Diego, CA, USA, 21–24 August 2011; pp. 1370–1378. [Google Scholar] [CrossRef]
  40. Kumar, V.A. Real Time Temperature Prediction in a Data Center Environment Using an Adaptive Algorithm. Master’s Thesis, The University of Texas at Arlington, Arlington, TX, USA, December 2013. [Google Scholar]
  41. Sharma, M.; Garg, R. An artificial neural network based approach for energy efficient task scheduling in cloud data centers. Sustain. Comput. Inform. Syst. 2020, 26, 100373. [Google Scholar] [CrossRef]
  42. Sasakura, K.; Aoki, T.; Komatsu, M.; Watanabe, T. Rack Temperature Prediction Model Using Machine Learning after Stopping Computer Room Air Conditioner in Server Room. Energies 2020, 13, 4300. [Google Scholar] [CrossRef]
  43. Ilager, S.; Ramamohanarao, K.; Buyya, R. Thermal Prediction for Efficient Energy Management of Clouds Using Machine Learning. IEEE Trans. Parallel Distrib. Syst. 2021, 32, 1044–1056. [Google Scholar] [CrossRef]
  44. Zapater, M.; Risco-Martín, J.L.; Arroba, P.; Ayala, J.; Moya, J.; Hermida, R. Runtime Data Center Temperature Prediction using Grammatical Evolution Techniques. Appl. Soft Comput. 2016, 49, 94–107. [Google Scholar] [CrossRef]
  45. Brenner, P.; Go, D.B.; Buccellato, A.P.C. Data Center Heat Recovery Models and Validation: Insights from Environmentally Opportunistic Computing. In Proceedings of the ASHRAE Winter Conference Technical Program, Dallas, TX, USA, 26–30 January 2013. [Google Scholar]
  46. Wang, S.J.; Jin, C.Z. An extension of earley’s algorithm for extended grammars. In Computational Methods; Liu, G., Tan, V., Han, X., Eds.; Springer: Dordrecht, The Netherlands, 2006. [Google Scholar] [CrossRef]
  47. Akbari, A.; Khonsari, A.; Ghoreyshi, S.M. Thermal-Aware Virtual Machine Allocation for Heterogeneous Cloud Data Centers. Energies 2020, 13, 2880. [Google Scholar] [CrossRef]
  48. Chaudhry, M.T.; Ling, T.C.; Manzoor, A.; Hussain, S.A.; Kim, J. Thermal-Aware Scheduling in Green Data Centers. ACM Comput. Surv. 2015, 47, 1–48. [Google Scholar] [CrossRef]
  49. Cioara, T.; Antal, M.; Antal, C.D.; Anghel, I.; Bertoncini, M.; Arnone, D.; Lazzaro, M.; Mammina, M.; Velivassaki, T.-H.; Voulkidis, A.; et al. Data Centers Optimized Integration with Multi-Energy Grids: Test Cases and Results in Operational Environment. Sustainability 2020, 12, 9893. [Google Scholar] [CrossRef]
  50. Yao, J.; Guan, H.; Luo, J.; Rao, L.; Liu, X. Adaptive Power Management through Thermal Aware Workload Balancing in Internet Data Centers. IEEE Trans. Parallel Distrib. Syst. 2015, 26, 2400–2409. [Google Scholar] [CrossRef]
  51. Moore, J.; Chase, J.; Ranganathan, P.; Sharma, R. Making scheduling “cool”: Temperature-aware workload placement in data centers. In Proceedings of the Annual Conference on USENIX Annual Technical Conference (ATEC ’05), Anaheim, CA, USA, 10–15 April 2005. [Google Scholar]
  52. Singh, S.; Chana, I. A Survey on Resource Scheduling in Cloud Computing: Issues and Challenges. J. Grid Comput. 2016, 14, 217–264. [Google Scholar] [CrossRef]
  53. Yang, J.; Xiao, W.; Jiang, C.; Hossain, M.S.; Muhammad, G.; Amin, S.U. AI-Powered Green Cloud and Data Center. IEEE Access 2019, 7, 4195–4203. [Google Scholar] [CrossRef]
  54. Wang, L.; von Laszewski, G.; Huang, F.; Dayal, J.; Frulani, T.; Fox, G. Task scheduling with ANN-based temperature prediction in a data center: A simulation-based study. Eng. Comput. 2011, 27, 381–391. [Google Scholar] [CrossRef]
  55. Nejad, S.M.M.; Badawy, G.; Down, D.G. Holistic thermal-aware workload management and infrastructure control for heterogeneous data centers using machine learning. Future Gener. Comput. Syst. 2021, 118, 208–218. [Google Scholar]
  56. Van Damme, T.; de Persis, C.; Tesi, P. Optimized Thermal-Aware Job Scheduling and Control of Data Centers. IFAC-Pap. 2017, 50, 8244–8249. [Google Scholar] [CrossRef]
  57. Ni, J.; Jin, B.; Zhang, B.; Wang, X. Simulation of Thermal Distribution and Airflow for Efficient Energy Consumption in a Small Data Centers. Sustainability 2017, 9, 664. [Google Scholar] [CrossRef] [Green Version]
  58. Sun, H.; Stolf, P.; Pierson, J. Spatio-temporal thermal-aware scheduling for homogeneous high-performance computing datacenters. Future Gener. Comput. Syst. 2017, 71, 157–170. [Google Scholar] [CrossRef] [Green Version]
  59. Kumar, M.; Sharma, S.C. Dynamic load balancing algorithm for balancing the workload among virtual machine in cloud computing. Procedia Comput. Sci. 2017, 115, 322–329. [Google Scholar] [CrossRef]
  60. Zhou, J.; Yan, J.; Chen, J.; Wei, T. Peak Temperature Minimization via Task Allocation and Splitting for Heterogeneous MPSoC Real-Time Systems. J. Signal Process. Syst. 2016, 84, 111–121. [Google Scholar] [CrossRef]
  61. Anghel, I.; Cioara, T.; Salomie, I.; Copil, G.; Moldovan, D.; Pop, C. Dynamic frequency scaling algorithms for improving the CPU’s energy efficiency. In Proceedings of the 2011 IEEE 7th International Conference on Intelligent Computer Communication and Processing, Cluj-Napoca, Romania, 25–27 August 2011; pp. 485–491. [Google Scholar] [CrossRef]
  62. Liu, H.; Liu, B.; Yang, L.T.; Lin, M.; Deng, Y.; Bilal, K.; Khan, S.U. Thermal-Aware and DVFS-Enabled Big Data Task Scheduling for Data Centers. IEEE Trans. Big Data 2018, 4, 177–190. [Google Scholar] [CrossRef]
  63. Cioara, T.; Salomie, I.; Anghel, I.; Chira, I.; Cocian, A.; Henis, E.; Kat, R. A Dynamic Power Management Controller for Optimizing Servers’ Energy Consumption in Service Centers. In Service-Oriented Computing, Proceedings of the International Conference on Service-Oriented Computing ICSOC 2010, Stockholm, Sweden, 23–27 November 2010; Maximilien, E.M., Rossi, G., Yuan, S.T., Ludwig, H., Fantinato, M., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6568. [Google Scholar] [CrossRef] [Green Version]
  64. Li, X.; Xie, N.; Tian, X. Dynamic Voltage-Frequency and Workload Joint Scaling Power Management for Energy Harvesting Multi-Core WSN Node SoC. Sensors 2017, 17, 310. [Google Scholar] [CrossRef]
  65. Fang, Q.; Wang, J.; Zhu, H.; Gong, Q. Using Model Predictive Control in Data Centers for Dynamic Server Provisioning. IFAC Proc. Vol. 2014, 47, 9635–9642. [Google Scholar] [CrossRef] [Green Version]
  66. Wang, J.; Shen, T.; Zhao, J.; Ma, S.; Rao, W.; Zhang, Y. Data-driven thermal efficiency modeling and optimization for reheating furnace based on statistics analysis. In Proceedings of the 2015 34th Chinese Control Conference (CCC), Hangzhou, China, 28–30 July 2015; pp. 8271–8275. [Google Scholar] [CrossRef]
  67. Liu, J.; Goraczko, M.; James, S.; Belady, C.; Lu, J.; Whitehouse, K. The Data Furnace: Heating Up with Cloud Computing. In Proceedings of the 3rd USENIX Workshop on Hot Topics in Cloud Computing, Portland, OR, USA, 14–15 June 2011. [Google Scholar]
  68. Tang, Q.; Mukherjee, T.; Gupta, S.K.S.; Cayton, P. Sensor-based Fast Thermal Evaluation Model for Energy Efficient High-Performance Datacenters. In Proceedings of the Fourth International Conference on Intelligent Sensing and Information Processing, Bangalore, India, 15 October–18 December 2006; pp. 203–208. [Google Scholar]
  69. Zhang, S.; Xia, Y. Solving nonlinear optimization problems of real functions in complex variables by complex-valued iterative methods. IEEE Trans. Cybern. 2018, 48, 277–287. [Google Scholar] [CrossRef]
  70. Belotti, P.; Kirches, C.; Leyffer, S.; Linderoth, J.; Luedtke, J.; Mahajan, A. Mixed-integer nonlinear optimization. Acta Numer. 2013, 22, 1–131. [Google Scholar] [CrossRef] [Green Version]
  71. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  72. Laabadi, S.; Naimi, M.; el Amri, H.; Boujemâa, A. The 0/1 Multidimensional Knapsack Problem and Its Variants: A Survey of Practical Models and Heuristic Approaches. Am. J. Oper. Res. 2018, 8, 395–439. [Google Scholar] [CrossRef] [Green Version]
  73. Pardey, Z.; Demetriou, D.; Erden, H.; Vangilder, J.; Khalifa, H.; Schmidt, R. Proposal for standard compact server model for transient data center simulations. ASHRAE Trans. 2015, 121, 413–421. [Google Scholar]
  74. Qarnot Datacenter. Available online: https://qarnot.com/ (accessed on 3 February 2021).
  75. SciPY Optimizer. Available online: https://docs.scipy.org/doc/scipy/reference/optimize.html (accessed on 29 January 2021).
  76. SciKit Learn. Available online: https://scikit-learn.org/stable/ (accessed on 29 January 2021).
  77. Keras. Available online: https://keras.io/ (accessed on 1 February 2021).
  78. Tensorflow. Available online: https://www.tensorflow.org/ (accessed on 1 February 2021).
  79. XGBoost. Available online: https://xgboost.readthedocs.io/en/latest/ (accessed on 1 February 2021).
  80. Heating Blog. Available online: https://www.electricpoint.com/heating/electric-heating/how-to-calculate-kw-required-to-heat-a-room (accessed on 14 April 2021).
  81. Intel® Core™ i7-6950X Study. Available online: https://www.anandtech.com/show/10337/the-intel-broadwell-e-review-core-i7-6950x-6900k-6850k-and-6800k-tested-up-to-10-cores/10 (accessed on 23 March 2021).
Figure 1. Distributed DC and computing power-based heating.
Figure 1. Distributed DC and computing power-based heating.
Sensors 21 02879 g001
Figure 2. Thermal aware workload scheduling optimization subproblems.
Figure 2. Thermal aware workload scheduling optimization subproblems.
Sensors 21 02879 g002
Figure 3. Measurements used in the machine learning processes.
Figure 3. Measurements used in the machine learning processes.
Sensors 21 02879 g003
Figure 4. Test case distributed DC.
Figure 4. Test case distributed DC.
Sensors 21 02879 g004
Figure 5. Data pre-processing pipeline.
Figure 5. Data pre-processing pipeline.
Sensors 21 02879 g005
Figure 6. Example of relevant data samples obtained using the pre-processing pipeline.
Figure 6. Example of relevant data samples obtained using the pre-processing pipeline.
Sensors 21 02879 g006
Figure 7. Power prediction for temperature change.
Figure 7. Power prediction for temperature change.
Sensors 21 02879 g007
Figure 8. Initial conditions: the 4 QRad server heaters’ total power demand and initial temperatures.
Figure 8. Initial conditions: the 4 QRad server heaters’ total power demand and initial temperatures.
Sensors 21 02879 g008
Figure 9. GBR heat demand estimation model power prediction.
Figure 9. GBR heat demand estimation model power prediction.
Sensors 21 02879 g009
Figure 10. CPU active cored due to task scheduling on the left and CPU power demand estimated and predicted on the right.
Figure 10. CPU active cored due to task scheduling on the left and CPU power demand estimated and predicted on the right.
Sensors 21 02879 g010
Figure 11. QRad Simulation Results: temperature evolution (left) and power demand evolution (right).
Figure 11. QRad Simulation Results: temperature evolution (left) and power demand evolution (right).
Sensors 21 02879 g011
Table 1. Server heater model parameters.
Table 1. Server heater model parameters.
ParameterDescriptionParameter Type
c a i r Air specific heat capacity, defined as a physical constant for air property dependent on temperatureConstant
c s e r v e r Server heat capacity, defined as a physical property of the server, intrinsic to the systemExperimentally determined
f a i r Airflow of air pumped by the cooling system over the servers, measured in m 3 s , the intrinsic parameter of the system.Measured during cooling system operation
ε s e r v e r Server thermal efficiency, defined by Equation (22), considered an intrinsic parameter of the systemExperimentally determined
T R O O M Room temperature measured by sensors during system operationMeasured
T s e r v e r Server temperature measured by sensors, directly influencing the room temperatureMeasured
P s e r v e r Server power demand due to workload executionThe output of the model
Table 2. Machine learning models used for determining the heat demand.
Table 2. Machine learning models used for determining the heat demand.
Model TypeModel Description
Linear RegressionThe basic linear regressor was used to determine the baseline for prediction accuracy
Polynomial RegressionA second-degree polynomial regressor. Multiple degrees were considered, but the validation score began to drop after the degree was set to 2.
Gradient Boosted Regression90 estimators with a maximum depth of 4. The samples had a minimum split of 5 and the learning rate was 0.1. The loss was computed using the least-squares method.
Random Forest Regression9 estimators with a maximum depth of 4 are defined.
Support Vector RegressionA support vector regressor with kernel type of radial basis function and parameters: C = 100 , γ = 0.01 , ε = 0.1
K Neighbors RegressionThe K-Nearest Neighbors Regression with 2 neighbors and uniform weights.
Deep Learning RegressionMulti-Layer Perceptron having one input layer, two hidden layers of 128 and 256 neurons, and one output layer. The activation function for the hidden layers is of type Rectified Linear Units (ReLU), and 500 epochs were used for training. The loss function was the mean squared error and the optimizer ADAM. Early stopping was employed with the patience of 50 epochs and a minimum validation loss as the monitor.
Table 3. Evaluation of the server heat transfer model power prediction capability.
Table 3. Evaluation of the server heat transfer model power prediction capability.
ServerMAPERMSPERMSE
QRad Heater11.9814.2820.15
Table 4. Evaluation results of machine learning-based heat models.
Table 4. Evaluation results of machine learning-based heat models.
ModelRMSER2Error MeanError Standard DeviationRMSPEMAPE
Linear Regressor14.40.89−0.6614.048.6213.36
Polynomial Regression33.120.392.0529.1923.3813.65
Random Forest Regressor10.040.92−2.069.557.65.06
Gradient Boosting Regressor10.650.94−1.1210.327.144.74
Support Vector Regression16.280.852.3315.029.716.89
K Neighbors Regressor13.70.841.0912.8510.546.27
Multi-Layer Perceptron Deep Neural Network33.0911.9224.5420.6217.29
Table 5. Residential heating equipment sizing case study.
Table 5. Residential heating equipment sizing case study.
V R O O M [ m 3 ] Number of Exterior WallsThermal Insulation C R O O M c a l   [ kcal / m 3 ] P R o o m R e c c o m e n d e d   [ Watt ] Number of QRad Heaters
301Yes5118384
1No5519914
2Yes5519914
2No6021444
351Yes5121004
1No5522754
2Yes5522754
2No6024505
501Yes5130006
1No5532506
2Yes5532506
2No6035007
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Antal, M.; Cristea, A.-A.; Pădurean, V.-A.; Cioara, T.; Anghel, I.; Antal, C.; Salomie, I.; Saintherant, N. Heating Homes with Servers: Workload Scheduling for Heat Reuse in Distributed Data Centers. Sensors 2021, 21, 2879. https://doi.org/10.3390/s21082879

AMA Style

Antal M, Cristea A-A, Pădurean V-A, Cioara T, Anghel I, Antal C, Salomie I, Saintherant N. Heating Homes with Servers: Workload Scheduling for Heat Reuse in Distributed Data Centers. Sensors. 2021; 21(8):2879. https://doi.org/10.3390/s21082879

Chicago/Turabian Style

Antal, Marcel, Andrei-Alexandru Cristea, Victor-Alexandru Pădurean, Tudor Cioara, Ionut Anghel, Claudia Antal (Pop), Ioan Salomie, and Nicolas Saintherant. 2021. "Heating Homes with Servers: Workload Scheduling for Heat Reuse in Distributed Data Centers" Sensors 21, no. 8: 2879. https://doi.org/10.3390/s21082879

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop