Next Article in Journal
An Assessment of Waveform Processing for a Single-Beam Bathymetric LiDAR System (SBLS-1)
Previous Article in Journal
Opto-Mechanical Eye Models, a Review on Human Vision Applications and Perspectives for Use in Industry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Approach to Predict the Performance of a Stratified Thermal Energy Storage Tank at a District Cooling Plant Using Sensor Data

by
Afzal Ahmed Soomro
1,*,
Ainul Akmar Mokhtar
1,
Waleligne Molla Salilew
1,
Zainal Ambri Abdul Karim
1,
Aijaz Abbasi
2,
Najeebullah Lashari
3 and
Syed Muslim Jameel
4
1
Department of Mechanical Engineering, Universiti Teknologi Petronas, Seri Iskandar 32610, Perak Darul Ridzuan, Malaysia
2
Mechanical Engineering Department, Quaid-e-Awam University of Engineering Science and Technology, Nawabshah 67450, Pakistan
3
Department of Petroleum and Gas Engineering, Dawood University of Engineering & Technology, M.A. Jinnah Road, Karachi 74800, Pakistan
4
School of Engineering, University of Galway, University Road, H91 TK33 Galway, Ireland
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7687; https://doi.org/10.3390/s22197687
Submission received: 1 September 2022 / Revised: 16 September 2022 / Accepted: 5 October 2022 / Published: 10 October 2022
(This article belongs to the Topic Applied Computing and Machine Intelligence (ACMI))

Abstract

:
In the energy management of district cooling plants, the thermal energy storage tank is critical. As a result, it is essential to keep track of TES results. The performance of the TES has been measured using a variety of methodologies, both numerical and analytical. In this study, the performance of the TES tank in terms of thermocline thickness is predicted using an artificial neural network, support vector machine, and k-nearest neighbor, which has remained unexplored. One year of data was collected from a district cooling plant. Fourteen sensors were used to measure the temperature at different points. With engineering judgement, 263 rows of data were selected and used to develop the prediction models. A total of 70% of the data were used for training, whereas 30% were used for testing. K-fold cross-validation were used. Sensor temperature data was used as the model input, whereas thermocline thickness was used as the model output. The data were normalized, and in addition to this, moving average filter and median filter data smoothing techniques were applied while developing KNN and SVM prediction models to carry out a comparison. The hyperparameters for the three machine learning models were chosen at optimal condition, and the trial-and-error method was used to select the best hyperparameter value: based on this, the optimum architecture of ANN was 14-10-1, which gives the maximum R-Squared value, i.e., 0.9, and minimum mean square error. Finally, the prediction accuracy of three different techniques and results were compared, and the accuracy of ANN is 0.92%, SVM is 89%, and KNN is 96.3%, concluding that KNN has better performance than others.

1. Introduction

From the standpoint of heating and air conditioning, stratified thermal energy storage (TES) has become increasingly popular in recent years [1,2]. The advantage of adopting TES for comfort and cooling applications is that energy and cost concerns are no longer an issue. The actual performance of TES is determined by several elements, including mixing at the tank’s inflow, mixing of hot and cold water, heat loss to the environment, and aspect ratio [3].
The above-mentioned factors impact the temperature distribution in the TES tank during the charging period; the transition between hot and cold water in the tank is known as the thermocline thickness ( W T C ). The performance of TES tank is determined through thermocline thickness ( W T C ) [4,5,6]. Multiple methods have been proposed to calculate W T C , such as a small-scale experimental setup, finite element analysis, computational fluid dynamics, and curve-fitting from sensor data.
Yoo et al. computed W T C by extrapolating thermocline edges from the thermocline’s mid-point. The region fringed to the linear slope of the thermocline profile is determined using interpolation. The thermocline edges were not determined at the true upper and lower limits of thermocline profiles, which is one of the methodology’s disadvantages [7]. Steward [8] conducted the steady-state (solving partial differential equations) model for stratified TES. Musser and Behnfleth [9] analyzed the performance of TES tank using the numerical method. Musser and Bahnfleth proposed a more reliable and simple technique of binding the thermocline zone by employing the dimensionless cut-off temperature on each edge of the thermocline region. The quantity of thermocline detected was suggested to be large enough to eliminate the impacts of tiny temperature changes at the thermocline’s extremities, but small enough to capture the majority of the temperature changes. This is how the dimensionless cut-off temperature is defined [9]. Unfortunately, the methods discussed here have two major drawbacks; first, when the temperature readings available are in discrete form, these methods cannot predict W T C accurately [10], and secondly, the computational complexity requires extensive physical-based knowledge. To cope with the above-mentioned problems, data driven approaches, including artificial neural network (ANN), support vector machine (SVM), and k-nearest neighbor (KNN), have been used in this study to predict thermocline thickness. Because of its learning ability and versatile mapping skills without requiring considerable physical based knowledge, [11] machine learning has attracted increased interest in the system modelling and simulation industry [12,13]. Machine learning (ML) have been used to mimic a variety of engineering applications (such as building energy systems, including TES [14,15], thermal performance in battery systems [16], human thermal comfort in passenger vehicles using an organic phase change material [17], agriculture [18], maintenance of gas pipelines subjected to corrosion [19], subsea pipelines annotation [20], predicting water saturation [21], viscosity of methane gas at high temperature conditions [22], integrity of corroded oil and gas pipelines [23]) other than machine learning-based indoor occupancy predictions on PCMs integrated building energy systems [24,25]. A state-of-the art review on the applications of the phase change materials on storage systems can be found in [26]. Machine learning applications in surrogate model and model predictive control are explored in [27]. A computational-efficient model was developed to predict the renewable generation in [28]. A supervised learning surrogate model was developed to improve prediction efficiency in [29].
Some researchers have focused on utilizing FFNN to forecast the TES tank’s temperature distribution. A water thermocline storage tank’s thermal stratification was modelled by G’eczy-Víg and Farkas [30] for both load and load-free conditions. The temperatures at different vertical positions were forecasted every 5 min, utilizing 12 inputs, including the outputs at a previous time-step. Additionally, [31] covered how ANN time-step affected stratification modelling. Soomro and Mokhtar [32] created an ANN model to forecast the variables of the sigmoid dose-response function describing the water TES tank’s temperature profile. Diez et al. [33] developed a model to predict the evolution of temperature distribution of a stratified solar water tank in static mode, in place of the dynamic processes of charging and discharging. The last time-matching step’s temperature values were used as inputs to predict the temperatures at various vertical layers at 10-min intervals. Recently, ANN has been effectively employed for the best control of a district cooling plant with a thermal energy storage tank for predicting the performance of the stratified chilled water TES of a heat pump system [34]. This paper suggests a novel method for controlling thermal energy storage (i.e., ice storage) in a district cooling system and efficiently predicting performance. Jia, Liu et al. 2022 created a system for TES operating strategy optimization by fusing physics-based modelling with deep learning [35]. The ANN has recently seen significant use in applications related to the energy system, as the aforementioned literature on ANN demonstrates. Based on the above discussion, it is clear from the literature that machine learning has recently drawn the attention of various fields and also in energy systems, but has not yet been well utilized for stratified thermal energy storage performance in district cooling plants; this is why, in this study, machine learning has been applied to predict the performance of a stratified thermal energy storage tank with sensor data. An algorithm’s success is determined by the nature of the problem, such as its variables, boundaries, and additional complications, such as patterns in data. The ability of an algorithm to solve a particular problem, however, does not guarantee that it will perform better than a random search [36,37]. Three machine learning techniques, artificial neural networks (ANN), support vector machines (SVM), and K-Nearest Neighbor (KNN), were used in this study.
The structure of the paper is as follows: in Section 2, the methods used to achieve the study objective are discussed. In Section 3, the results generated using machine learning algorithms are presented and discussed. Finally, Section 4 exhibits the conclusion and the future works.

2. Materials and Methods

2.1. Data Collection

The data in this study was collected from the operating TES tank installed at the UTP GDC plant. Figure 1 shows the layout of TES system at GDC UTP and Figure 2 represents the flow during the charging process of TES tank at the GDC plant.
After collecting the data from the TES sensors, the data was analyzed in an Excel worksheet. Figure 3 shows that during charging, the temperature distribution moves from the bottom to the top direction, and the width of the thermocline thickness increases with the passage of time. This behavior has been verified in many previous studies, such as [6,38,39,40]. Evolution of temperature distribution and thermocline thickness during charging is shown in Figure 3.

2.2. Description of TES System

The TES capacity is 10,000 RTh, connected with four vapor compression or electric chillers (EC) which charge the TES tank and have a capacity of 325 RTh each. The cooling demand of the UTP campus and mosque is fulfilled by the GDC plant. The 14 temperature sensors are installed vertically at a distance of 1 m in the tank, and the data acquisition system is used to collect the temperature data, as shown in Figure 1. During off-peak hours, the chillers are used to charge the tank, and during on-peak periods the tank is discharged to supply the chilled water to the campus, as shown in Figure 2. The tank is cylindrical, having an inside diameter of 22.3 m and a height of 15 m. The total storage capacity of the tank is 54,000 m 3 . The lower nozzle is made up of a pipe with a diameter of 0.5 m located at a height of 1.82 m, and the upper nozzle has a diameter of 0.3 and its location is at 12.3 m; diffusers are attached with both nozzles. The insulation material used is polystyrene, with 300 mm thickness, and epoxy paint is used for the internal coating of the tank. The mass flow rate of the electric chiller is designed at 131 m 3 / h , the inlet and outlet temperature are set at 13 °C and 6 °C, respectively.
The input to the machine learning models is the temperature distribution collected from the sensors installed vertically in the tank, and the output data for the machine learning models model is W T C . The outputs were obtained using the non-linear regression technique proposed by [41,42]. Equation (1) has been used to predict W T C .
W T C = 2 × L o g × 1 θ 1 S
where
θ = T T c ( T h T c )

2.3. Data Preprocessing

In order to have an accurate prediction model, dealing with the input data before feeding it to the algorithm is always desirable. Hence, in this study, the input data were normalized and smoothed with two different smoothing techniques, such as moving average filter and median filter. The procedure is shown in Figure 4 below.
A.
Data Normalization
In this study, the data normalization was carried out using the minimax technique, using Equation (3). In this method, the rescaling of the outputs was carried out by transferring one range of values to another range of values [43,44]. Mostly the rescaling is carried out between [0, 1] or [−1, 1]. The linear interpretation formula, such as Equation (3), was used for the rescaling. In this study, the range of normalization was between 0 and 1. The normalization procedure could lessen the learning algorithm’s ambiguity regarding the significance of each parameter with a smaller amplitude. Figure 5 represents that how the data is normalized.
X n o r m = X X m i n X m a x X m i n
where X is the original value, X m i n is the minimum value and X m a x is the maximum value.
B.
Data smoothing/denoising techniques
In this study, the moving average filter and median filter were used to denoise the data. The mean average filter replaces the point with the average values found for the number over a specified frame length, where the moving median filter works with the principle of processing the signal entry by entry, and replacing each input with median values over a certain frame length [45].
The moving average filter is expressed as:
y [ i ] = 1 M j = 0 M 1 x [ i j ]
where the input values are x[i − j], the output value is y [ j ] , and the frame duration is M.
The filter or the smoothing using moving average filter is expressed as:
y = movmean(A,[m 0])
where A is an array of input data, Y is smoothed data, and m is the frame length.
The filter or the smoothing using moving median filter is expressed as:
y = movmedian(A,[m 0])
where A is an array of input data, Y is smoothed data, and m is the frame length.
Figure 6 shows the behavior of the input and output data. The data clearly shows that there are some noises which must be denoised.
Hence the data was smoothed using the moving average filter and median average filter, as shown in Figure 7. The graph shows that T1, T2, T3, and T4 were more smoothed by the mean average filter, T5, T6, T7, T8, T9, and T10 were more smoothed by moving median filter, whereas T11, T12, T13, and T14 show the same smoothing magnitude for both smoothing techniques at the same frame length. The frame length for both filters was 20.
Figure 8 shows smoothed and normalized versus unsmoothed and normalized data. The blue color graph is for unsmoothed normalized data, the red color graph for mean average filter-smoothed and normalized data, and the yellow one is for median average filter-smoothed and normalized data.

2.4. Data Division

Once the dataset was obtained from the plant, it was stored in the CSV file, then the file was uploaded in the MATLAB toolbox where it was divided into three parts. A subset of 185 (70%) from the whole data was used for training purposes. A subset of 39 (15%) from the whole dataset was used for cross validation. A subset of 39 (15%) was used for testing purposes. Table 1 shows the example of division of dataset.

2.5. ANN Modeling

After the data collection and preprocessing, ANN modeling was performed using the feed-forward backpropagation technique [44] as shown in Figure 9. The mathematical form of artificial neural network is presented in below.

Steps in ANN Modeling

Step 1 Determine the output of the input layer.
Step 2 Determine the output of the hidden layer.
The corresponding normalized input value is multiplied with the corresponding initial weight and these products are added together. This summation is again added to with the variable bias. The resulting sum is applied to the activation function. The activation function employed is the log sigmoid function, due to its best prediction capability as proved in some studies, including but not limited to [46], as shown in Equation (7).
O u t p u t = 1 1 + e x
where x is the result of adding the inputs and weights.
Step 3 Determine the output of the output layer.
The same logic, involving multiplication of weights, with inputs followed by summation and its application to the activation function (as in Step 2) is applied, but the activation function used here is linear function, as shown in the Equation (8).
O u t p u t = x
Step 4 Determine the error of the model.
For each training pattern, the error of the output layer is compared to the desired output, and the difference between the two is the error, which is determined using Equation (9).
E = T Y
Y is the forecast value, while T is the target value.
Step 5 Calculate the square of the error.
The square of the error can be calculated using Equation (10).
E 2 = T Y 2
A.
Backward Pass
The purpose of the backward pass is to calculate errors in the hidden and input layers and modify the weights for better predictions.
Step 1 Determine the node’s error in the hidden layer.
The derivative of the activation function is multiplied by error, and the result is used to determine the error. The activation function is derived as follows:
δ 2 = Y × 1 Y × Y T
The activation function’s derivative is given by Y × 1 Y
Step 2 Determine the node’s error in the input layer.
The error is determined by multiplying the activation function’s derivation by the error.
δ 1 = Y × 1 Y × Y T
Step 3 Adjustment of the weights.
There are several methods for finding the minima of a parabola or any function in any dimension. One of the best techniques for training ANNs is the Levenberg–Marquardt optimization algorithm, particularly when the number of weights is high. Although requiring more memory than other techniques, this method is strongly advised as the first choice for supervised tasks. Equation (13) is used to update the weights in a Levenberg–Marquardt training function (LMTF).
w i + 1 = w i J i T J i + μ i I 1 J i T e i
Step 4 Determine whether or not the training should continue.
Repeat the previous steps until the minimum error is reached or the set number of iterations has been exceeded.
Step 5 Denormalization of the output.
This step is crucial to return the values to their original, unnormalized form using Equation (14).
X = X n o r m × X m a x X m i n + X m i n

2.6. Support Vector Machine

The supervised method (the most-isolation approach) is the support vector machines method. By supplying numerical data, it solves classification and regression problems [47]. The precise location for the hyperplane is determined by the support vector machine process through optimization until the hyperplanes are positioned in the maximum margin [48]. The hyperplane with the highest margin is capable of accurately classifying data or other input parameters [49]. The key concern in support vector machine (SVM) is how to determine the hyperplane’s ideal position. The most acceptable position should be chosen after conducting many tries. The hyperplane’s main function is to divide the two classes. Thus, widening the decision boundary between the two classes is suitable. The line crossing across the support vector is referred to as the separate lines to indicate the decision boundary. The concept of SVM is depicted in Figure 10.
All the input data is classified into the necessary classes using a support vector machine. The label, yi, and the set of data, “x1… xn,” are required. The input data is either predicted to be in the class label, yi = 1 or in the class label, yi = −1 [50]. Accordingly, it is written as:
Data = x i , y i x i p , y i 1 , + 1 i = 1 n
where n is the number of data points and p is the feature dimension.
The hyperplane equation is expressed as:
w . x = 0
The equation for 2D space is as follows:
Y = w x + b
From Equations (1)–(3), we can generalize that:
W X i ¯ + b 0 i : y i = + 1 . W X i ¯ + b 0 i : y i = 1 .
where W = w 1 w d is the dimensions row vector corresponding to the normal axis to the classifier, b is the bias, and X i is a provided row vector associated with the i th data point. y i 1 , + 1 is a binary class variable of the i th data point. The bias b controls the distance of the classifier line from the origin as the vector W controls the classifier line’s position.

2.7. K-Nearest Neighbor

The fundamental supervised algorithm is commonly referred to as k-NN. The concept behind KNN is used in many other machine learning methods, therefore to understand other techniques, studying KNN could be a great place to start [51]. The instances are not stored throughout the training phase. In training, some reasonable indexing is required to quickly identify the classes. The k-nearest classes of the one instance must be found via KNN [52]. The effectiveness of classification is dependent on the preference of similarity by calculating the distance between two instances. The k-nearest neighbor algorithm is shown in Figure 11. The Euclidean distance method is the most practical of the various ways to determine the distance between two points. The Euclidean distance between two points in XY dimensions is calculated as follows:
Euclidean   Distance   between   point   A 1   and   B 2 = X 2 X 1 2 + Y 2 Y 1 2

3. Development of Prediction Model

3.1. Classifier Tuning and Validation Techniques

The performance of the classification model prediction is tested using a variety of techniques. It is conventional to use 70% of the data set for training and the remaining 30% for testing the model’s efficacy. This approach is known as leave-p-out. Cross-validation is the second type of model training and testing strategy, and works by dividing the given dataset into training and validation data subsets and then rotating the evaluations among the subsets; it can assess a classifier’s generalized overall accuracy on an unbiased dataset. When the data set is small, cross validation is used. A generalization called k-fold cross-validation divides the data into k−1 training sets and a validation set. The partitions must roughly be equal in size and are formed by randomly choosing data. The average of each performance metric, such as classification accuracy, will be the outcome of cross-validation. K-fold cross validation means using the partition just once while repeating the validation k times. MATLAB programming language was used to develop the classification algorithms. During development of the SVM and KNN prediction model, the frame length for both moving average filter and moving median filter was 20. The frame length was varied and finally a frame length which leads to the best classification accuracy was selected.

3.2. Hyper Parameter Optimization for the Model

The machine learning models are highly dependent on the following hyerparameters as shown in the Table 2.

3.3. Evaluation Performance Indices

After the prediction was achieved from the machine learning models, the output of machine learning models was compared with the actual output, i.e., the training dataset. To compare the performance of the ANN and non-linear regression model results, MSE, as shown in Equation (20), and R2, as shown in Equation (21), were used in this study as follows. However, for the support vector machine and KNN, the confusion matrix [53] was generated, which shows the prediction accuracy.
M S E = 1 n n = 1 n T O 2
R 2 = 1 i = 1 n T O 2 i = 1 n T O ¯ 2

4. Result and Discussion

This section has been divided into two parts. The first part consists of the results of the ANN architecture-building and the second part is related to the validation or comparison of the output values with the real behavior of the thermocline thickness. The algorithms are assessed using the R-square and MSE metrics that were previously discussed. For the train and test datasets, the predicted and actual values are compared. The projected and experimental data in the best-case scenario should be a 45° slope line that may be shown as the best fit curve to visually analyze the deviation.

4.1. Prediction with ANN Model Architecture

The optimum ANN setup, which is dependent on control parameters such as the number of hidden neurons, the number of hidden layers, and the learning rate, was achieved. Back-propagation with the LM algorithm was the learning algorithm used here, as indicated in Table 3.
A total of 263 training data sets were used to train the network. The impact of the number of neurons on model performance can be seen in Figure 12 in terms of R-square and MSE, respectively. The number of neurons has an impact on the R-square of the model, with 10 neurons achieving the best results, and after that, the model began to overtrain, as experienced in [54]. The impact of the learning rate on the R-square and MSE of the model is also shown in Figure 12. The minimum value of MSE and maximum value of R-square is noted at 0.01 learning rate.

4.2. Prediction of Thermocline Thickness

Figure 13 shows the overall regression of the collected and predicted data and Figure 14 shows the comparison between the ANN predicted and actual values during the charging period. The graph shows a very motivated representation, as the data is very close to the central line and gives an R-square of 0.92 for W T C . According to the theory, W T C increases during charging time due to mixing and conduction effects. It can be seen from Figure 14 that during the charging period the W T C increased from 1.94 to 3.5, which supports the theoretical behavior of thermocline during charging. The average error between the ANN predicted and actual values is 2.4%. Figure 14 shows the comparison between the calculated and ANN predicted values of W T C ; it can be seen that the W T C has increased with respect to charging time as expected [9,41,55,56].
Figure 15 represents the changes in the MSE for training, validation, and testing; with respect to the iterations, it can be seen from the graph that the best result was achieved at the 23 iterations, which means that the minimum value of MSE is achieved at 23 iterations or epochs.
The outputs of the model and the actual values are compared for training, validation, and testing and the results are shown in Figure 16. The R value shows the closeness to the targeted values; if the value is greater than 0.9 it means the predictions are satisfactory, and it can be seen from the Figure 16 that the R values for training, validation, and testing are 0.97, 0.94, and 0.95, respectively. A total number of 263 data points were used in the comparison. The closeness of the data to the central line indicates the accuracy of the prediction of the model. It can be seen that the data is very close to the trend line, showing a very good prediction result. The R-square for W T C comparison is 0.92.

4.3. Prediction via KNN and SVM

The Input data magnitude with it’s label shown in Table 4 below.
(A).
Prediction via KNN
Figure 17 shows the confusion matrix for k-nearest neighbor with mean average filter prediction, and it shows that the maximum false positive value is 1, which happened in class 7; the maximum false negative is 1, which happened in class 2, 10, and 11. The result has prevailed that KNN is a promising tool to predict the output. The overall classification accuracy was 96.3%.
Figure 18 shows the confusion matrix for k-nearest neighbor with median average filter prediction and the figure shows that the maximum false positive value is 1, happening in class 3; whereas the maximum false negative is 1, which happened in class 1, 5, 6, 8, 10, and 11. The result prevailed that KNN is also a promising tool to predict the output with median average filter. The overall classification accuracy is 93.5%.
(B).
Prediction via SVM
Figure 19 shows the confusion matrix for the support vector machine with mean average filter prediction, and it shows that the maximum false positive value is 1, happening in class 2, 3, 6, 7, and 11; whereas the maximum false negative is 2, and this happened in class 6.
Figure 20 shows the confusion matrix for the support vector machine with median average filter prediction, and it shows that the maximum false positive value is 5, and this happened in class 2 and 3, whereas there are no false negative values. Even though the SVM accuracy is lower than the prediction accuracy of KNN, it predicts the output at 89.7% using mean average filter, and 82.4% using median average filter. The result shows that k-nearest neighbor predicts more false negative values than false positive values, whereas support vector machine predicts more false positive values than false negative values. Generally, the result prevailed that k-nearest neighbor is the best prediction tool with mean average filter.

5. Conclusions

This study focuses on the evaluation performance of a stratified TES tank installed at the GDC plant of UTP in terms of thermocline thickness, using machine-learning algorithms. The study was conducted using backpropagation with the Lavenberg–Marquadt algorithm, support vector machine, and k-nearest neighbor, trained with the temperature sensor data collected from the GDC plant installed at the university Teknologi PETRONAS, Malaysia. Based on the modeling results, the following conclusions can be drawn:
  • The model was trained by using the data obtained from the temperature sensor data and non-linear regression model based on Musser and Behnfleth’s approach. The model was trained to predict W T C from temperature distribution data obtained from an operating TES tank. The average error during charging between ANN and actual values for W T C was 2.4%. The maximum R-square was 0.92.
  • The ANN model for the TES tank was built based on various parameters i.e., number of hidden layers, numbers of hidden neurons, and learning rate. All the parameters were decided based on a trial-and-error approach. After performing various trials, the optimum parameters were decided as 10 numbers of neurons, 0.01 learning rate, and 1 hidden layer. On the decided parameters, the model results were validated with the results obtained by the non-linear regression model used for data generation. After validation with the model, the R-square achieved was 0.94 between the ANN and the non-linear regression model.
  • The SVM model was also trained with the same data, the R-square obtained using the SVM model was 89% and for KNN it was 96.3%.
  • Finally, based on the results, KNN outperformed the other machine learning models used in this study. The author would like to extend the use of the machine learning model for other parameter predictions, such as figure of merit and for optimization of the operations in future.

6. Research Limitations and Future Work

Although predicting the performance of stratified thermal energy storage tanks using ANN, SVM, and KNN produced excellent results, there are still some limitations in the work, such as:
  • Less number of observations in dataset.
  • The modeling lacks physics understanding.
The above-mentioned limitations can be improved further. The author considers that the work can still be extended as follows:
  • Consider employing more sophisticated machine learning models, such as deep learning models, which have remained unexplored in this area.
  • Using physics-informed neural networks, which can solve the system’s governing equations and lessen the problem of data scarcity in the research.
  • Planning for maintenance can be added to the work.

Author Contributions

Conceptualization, A.A.S. and A.A.M.; methodology, A.A.S. and A.A.M.; software, A.A.S. and W.M.S.; validation, A.A.S. and S.M.J.; formal analysis, A.A.S. and A.A.; investigation, N.L.; resources, A.A.M. and Z.A.A.K.; data curation, A.A.S.; writing—original draft preparation, A.A.S.; writing—review and editing, N.L. and W.M.S.; visualization, A.A.S.; supervision, A.A.M.; project administration, A.A.M.; funding acquisition, A.A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by YUTP-FRG 1/2021 (015LC0-381).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

ANNArtificial neural network
SVMSupport vector machine
KNNK-nearest neighbor
C i n t Integrated capacity
C m a x Ideal capacity
CSVComma separated values
GDCGas district cooling
OOutput
O ¯ Average of output values
RThRefrigeration Ton-hour
LMLavemberg–Marquadt
MSEMean square error
SSlope of thermocline
TESThermal energy storage
T1-T14Temperature in sensor one to fourteen
TTemperature at selected sensor
T c Average cooling temperature
T h Average hot temperature
UTPUniversiti Teknologi PETRONAS
WTCThermocline thickness
XOriginal value
X n o r m Normalized value of X
X m a x Maximum value of X
X m i n Minimum value of X

References

  1. Dincer, I.; Rosen, M. Thermal Energy Storage: Systems and Applications; John Wiley & Sons: Hoboken NJ, USA, 2002. [Google Scholar]
  2. Mahon, H.; O’Connor, D.; Friedrich, D.; Hughes, B.R. A review of thermal energy storage technologies for seasonal loops. Energy 2021, 239, 122207. [Google Scholar] [CrossRef]
  3. Karim, A.; Burnett, A.; Fawzia, S. Investigation of stratified thermal storage tank performance for heating and cooling applications. Energies 2018, 11, 1049. [Google Scholar] [CrossRef] [Green Version]
  4. Majid, A.A.M.; Soomro, A.A.; Akmar, A. Artificial neural network modelling approach for assessment of stratified thermal energy storage tank. ARPN J. Eng. Appl. Sci. 2019, 14, 8. [Google Scholar]
  5. Hosseinnia, S.M.; Akbari, H.; Sorin, M. Numerical analysis of thermocline evolution during charging phase in a stratified thermal energy storage tank. J. Energy Storage 2021, 40, 102682. [Google Scholar] [CrossRef]
  6. Majid, M.A.A.; Muhammad, M.; Hampo, C.C.; Akmar, A.B. Analysis of a Thermal Energy Storage Tank in a Large District Cooling System: A Case Study. Processes 2020, 8, 1158. [Google Scholar] [CrossRef]
  7. Yoo, J.; Wildin, M.; Truman, C. Initial formation of a thermocline in stratified thermal storage tanks. ASHRAE Trans. 1986, 92, 280–292. [Google Scholar]
  8. Stewart, W. Downwardly impinging flows for stratified chilled water storage. In Proceedings of the 1992 National Heat Transfer Conference, San Diego, CA, USA, 9–12 August 1992. [Google Scholar]
  9. Musser, A.; Bahnfleth, W.P. Evolution of temperature distributions in a full-scale stratified chilled-water storage tank with radial diffusers. ASHRAE Trans. 1998, 104, 55. [Google Scholar]
  10. González-Altozano, P.; Gasque, M.; Ibáñez, F.; Gutiérrez-Colomer, R.P. New methodology for the characterisation of thermal performance in a hot water storage tank during charging. Appl. Therm. Eng. 2015, 84, 196–205. [Google Scholar] [CrossRef]
  11. Salilew, W.M.; Karim, Z.A.A.; Lemma, T.A. Investigation of fault detection and isolation accuracy of different Machine learning techniques with different data processing methods for gas turbine. Alex. Eng. J. 2022, 61, 12635–12651. [Google Scholar] [CrossRef]
  12. Manohar, H.J.; Saravanan, R.; Renganarayanan, S. Modelling of steam fired double effect vapour absorption chiller using neural network. Energy Convers. Manag. 2006, 47, 2202–2210. [Google Scholar] [CrossRef]
  13. Lee, D.; Ooka, R.; Ikeda, S.; Choi, W. Artificial neural network prediction models of stratified thermal energy storage system and borehole heat exchanger for model predictive control. Sci. Technol. Built Environ. 2019, 25, 534–548. [Google Scholar] [CrossRef]
  14. Krarti, M. An overview of artificial intelligence-based methods for building energy systems. J. Sol. Energy Eng. 2003, 125, 331–342. [Google Scholar] [CrossRef]
  15. Soomro, A.A.; Mokhtar, A.A.; Akbar, A.; Abbasi, A. Modelling Techniques Used in The Analysis of Stratified Thermal Energy Storage: A Review. MATEC Web Conf. 2018, 225, 01015. [Google Scholar] [CrossRef]
  16. Afzal, A.; Ramis, M. Multi-objective optimization of thermal performance in battery system using genetic and particle swarm algorithm combined with fuzzy logics. J. Energy Storage 2020, 32, 101815. [Google Scholar] [CrossRef]
  17. Afzal, A.; Saleel, C.A.; Badruddin, I.A.; Khan, T.Y.; Kamangar, S.; Mallick, Z.; Samuel, O.D.; Soudagar, M.E.M. Human thermal comfort in passenger vehicles using an organic phase change material–an experimental investigation, neural network modelling, and optimization. Build. Environ. 2020, 180, 107012. [Google Scholar] [CrossRef]
  18. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine learning in agriculture: A review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef] [Green Version]
  19. Mahmoodzadeh, Z.; Wu, K.-Y.; Droguett, E.L.; Mosleh, A. Condition-based maintenance with reinforcement learning for dry gas pipeline subject to internal corrosion. Sensors 2020, 20, 5708. [Google Scholar] [CrossRef] [PubMed]
  20. Stamoulakatos, A.; Cardona, J.; McCaig, C.; Murray, D.; Filius, H.; Atkinson, R.; Bellekens, X.; Michie, C.; Andonovic, I.; Lazaridis, P.; et al. Automatic annotation of subsea pipelines using deep learning. Sensors 2020, 20, 674. [Google Scholar] [CrossRef] [Green Version]
  21. Sambo, C.H.; Hermana, M.; Babasari, A.; Janjuhah, H.T.; Ghosh, D.P. Application of Artificial Intelligence Methods for Predicting Water Saturation from New Seismic Attributes. In Proceedings of the Offshore Technology Conference Asia, Kuala Lumpur, Malaysia, 20–23 March 2018. [Google Scholar]
  22. Sambo, C.; Yin, Y.; Djuraev, U.; Ghosh, D. Application of adaptive neuro-fuzzy inference system and optimization algorithms for predicting methane gas viscosity at high pressures and high temperatures conditions. Arab. J. Sci. Eng. 2018, 43, 6627–6638. [Google Scholar] [CrossRef]
  23. Soomro, A.A.; Mokhtar, A.A.; Kurnia, J.C.; Lashari, N.; Lu, H.; Sambo, C. Integrity assessment of corroded oil and gas pipelines using machine learning: A systematic review. Eng. Fail. Anal. 2022, 131, 105810. [Google Scholar] [CrossRef]
  24. Zhou, Y.; Grids, S. Corrigendum to “Quantification of energy flexibility of residential net-zero-energy buildings involved with dynamic operations of hybrid energy storages and diversified energy conversion strategies” [Sustain. Energy Grids Netw. 21 (2020) 100304]. Sustain. Energy Grids Netw. 2020, 21, 100321. [Google Scholar] [CrossRef]
  25. Zhou, Y.; Zheng, S.; Liu, Z.; Wen, T.; Ding, Z.; Yan, J.; Zhang, G. Passive and active phase change materials integrated building energy systems with advanced machine-learning based climate-adaptive designs, intelligent operations, uncertainty-based analysis and optimisations: A state-of-the-art review. Renew. Sustain. Energy Rev. 2020, 130, 109889. [Google Scholar] [CrossRef]
  26. Zhou, Y.; Zheng, S.; Zhang, G. A state-of-the-art-review on phase change materials integrated cooling systems for deterministic parametrical analysis, stochastic uncertainty-based design, single and multi-objective optimisations with machine learning applications. Energy Build. 2020, 220, 110013. [Google Scholar] [CrossRef]
  27. Zhou, Y.; Zheng, S.; Zhang, G. A review on cooling performance enhancement for phase change materials integrated systems—Flexible design and smart control with machine learning applications. Build. Environ. 2020, 174, 106786. [Google Scholar] [CrossRef]
  28. Zhou, Y.; Zheng, S.; Zhang, G. Machine-learning based study on the on-site renewable electrical performance of an optimal hybrid PCMs integrated renewable system with high-level parameters’ uncertainties. Renew. Energy 2020, 151, 403–418. [Google Scholar] [CrossRef]
  29. Zhou, Y.; Zheng, S. Multi-level uncertainty optimisation on phase change materials integrated renewable systems with hybrid ventilations and active cooling. Energy 2020, 202, 117747. [Google Scholar] [CrossRef]
  30. Géczy-Víg, P.; Farkas, I. Neural network modelling of thermal stratification in a solar DHW storage. Sol. Energy 2010, 84, 801–806. [Google Scholar] [CrossRef]
  31. Géczy-Víg, P.; Farkas, I. Influence of the time step in ANN modelling of thermal stratification of solar storage. IFAC Proc. Vol. 2008, 41, 9575–9578. [Google Scholar] [CrossRef] [Green Version]
  32. Soomro, A.A.; Mokhtar, A.A. Prediction of performance parameters of stratified TES tank using artificial neural network. AIP Conf. Proc. 2018, 2035, 080001. [Google Scholar]
  33. Diez, F.; Navas, L.; Martínez, A. Prediction and identification of the evolution of solar-heated hot water temperature in a stratified tank in static mode using artificial neural networks. Appl. Sol. Energy 2016, 52, 183–192. [Google Scholar] [CrossRef]
  34. Cox, S.J.; Kim, D.; Cho, H.; Mago, P. Real time optimal control of district cooling system with thermal energy storage using neural networks. Appl. Energy 2019, 238, 466–480. [Google Scholar] [CrossRef]
  35. Jia, L.; Liu, J.; Chong, A.; Dai, X. Deep learning and physics-based modeling for the optimization of ice-based thermal energy systems in cooling plants. Appl. Energy 2022, 322, 119443. [Google Scholar] [CrossRef]
  36. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef] [Green Version]
  37. Kumar, A.; Ridha, S.; Ganet, T.; Vasant, P.; Ilyas, S.U. Machine learning methods for herschel–bulkley fluids in annulus: Pressure drop predictions and algorithm performance evaluation. Appl. Sci. 2020, 10, 2588. [Google Scholar] [CrossRef]
  38. Hoffmann, J.-F.; Fasquelle, T.; Goetz, V.; Py, X. Experimental and numerical investigation of a thermocline thermal energy storage tank. Appl. Therm. Eng. 2017, 114, 896–904. [Google Scholar] [CrossRef]
  39. Reddy, K.; Pradeep, N. Stability analysis of the thermocline thermal energy storage system during high flow rates for solar process heating applications. Sol. Energy 2021, 226, 40–53. [Google Scholar] [CrossRef]
  40. Lou, W.; Fan, Y.; Luo, L. Single-tank thermal energy storage systems for concentrated solar power: Flow distribution optimization for thermocline evolution management. J. Energy Storage 2020, 32, 101749. [Google Scholar] [CrossRef]
  41. Waluyo, J.; Majid, M.A.A. Performance Evaluation of Stratified TES using Sigmoid Dose Response Function. J. Appl. Sci. 2011, 11, 1642–1647. [Google Scholar] [CrossRef] [Green Version]
  42. Waluyo, J.; Majid, M.A.A.; Amin, M. Temperature Profile and Thermocline Thickness Evaluation of a Stratified Thermal Energy Storage Tank. Int. J. Mech. Mechatron. Eng. 2010, 1, 7–12. [Google Scholar]
  43. Salehuddin, N.F.; Omar, M.B.; Ibrahim, R.; Bingi, K. A Neural Network-Based Model for Predicting Saybolt Color of Petroleum Products. Sensors 2022, 22, 2796. [Google Scholar] [CrossRef]
  44. Asteris, P.G.; Roussis, P.C.; Douvika, M.G. Feed-forward neural network prediction of the mechanical properties of sandcrete materials. Sensors 2017, 17, 1344. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Kawala-Sterniuk, A.; Podpora, M.; Pelc, M.; Blaszczyszyn, M.; Gorzelanczyk, E.J.; Martinek, R.; Ozana, S. Comparison of smoothing filters in analysis of EEG data for the medical diagnostics purposes. Sensors 2020, 20, 807. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Jo, H.-S.; Park, C.; Lee, E.; Choi, H.K.; Park, J. Path loss prediction based on machine learning techniques: Principal component analysis, artificial neural network, and Gaussian process. Sensors 2020, 20, 1927. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Gunn, S.R. Support vector machines for classification and regression. ISIS Tech. Rep. 1998, 14, 5–16. [Google Scholar]
  48. Lin, C.-F.; Wang, S.-D. Fuzzy support vector machines. IEEE Trans. Neural Netw. 2002, 13, 464–471. [Google Scholar]
  49. Fentaye, A.D.; Ul-Haq Gilani, S.I.; Baheta, A.T.; Li, Y.-G. Performance-based fault diagnosis of a gas turbine engine using an integrated support vector machine and artificial neural network method. Proc. Inst. Mech. Eng. Part A J. Power Energy 2019, 233, 786–802. [Google Scholar] [CrossRef] [Green Version]
  50. Zhou, X.; Zhang, X.; Wang, B. Online support vector machine: A survey. In Harmony Search Algorithm; Springer: Berlin/Heidelberg, Germany, 2016; pp. 269–278. [Google Scholar]
  51. Wu, Y.; Ianakiev, K.; Govindaraju, V. Improved k-nearest neighbor classification. Pattern Recognit. 2002, 35, 2311–2318. [Google Scholar] [CrossRef]
  52. Zouhal, L.M.; Denoeux, T. An evidence-theoretic k-NN rule with parameter optimization. IEEE Trans. Syst. Man Cybern. Part C 1998, 28, 263–271. [Google Scholar] [CrossRef] [Green Version]
  53. Azmi, N.; Kamarudin, L.M.; Zakaria, A.; Ndzi, D.L.; Rahiman, M.H.F.; Zakaria, S.M.M.S.; Mohamed, L. RF-based moisture content determination in rice using machine learning techniques. Sensors 2021, 21, 1875. [Google Scholar] [CrossRef]
  54. Agatonovic-Kustrin, S.; Beresford, R. Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research. J. Pharm. Biomed. Anal. 2000, 22, 717–727. [Google Scholar] [CrossRef]
  55. Waluyo, J. Simulation model of stratified thermal energy storage tank using finite difference method. AIP Conf. Proc. 2016, 1737, 030002. [Google Scholar]
  56. Majid, M.; Hui, P.; Soomro, A. Performance assessment of stratified chilled water thermal energy storage tank at district cooling plant. IOP Conf. Ser. Mater. Sci. Eng. 2020, 863, 012032. [Google Scholar] [CrossRef]
Figure 1. Photograph taken at GDC plant data acquisition.
Figure 1. Photograph taken at GDC plant data acquisition.
Sensors 22 07687 g001
Figure 2. Flow diagram of charging cycle of TES at UTP GDC plant.
Figure 2. Flow diagram of charging cycle of TES at UTP GDC plant.
Sensors 22 07687 g002
Figure 3. Evolution of temperature distribution and thermocline thickness during charging.
Figure 3. Evolution of temperature distribution and thermocline thickness during charging.
Sensors 22 07687 g003
Figure 4. Prediction model development flowchart.
Figure 4. Prediction model development flowchart.
Sensors 22 07687 g004
Figure 5. Example of data normalization.
Figure 5. Example of data normalization.
Sensors 22 07687 g005
Figure 6. Input and output data without preprocessing.
Figure 6. Input and output data without preprocessing.
Sensors 22 07687 g006
Figure 7. Smoothed versus unsmoothed data: blue color for unsmoothed data, red color smoothed data using mean average filter, and yellow for median filter.
Figure 7. Smoothed versus unsmoothed data: blue color for unsmoothed data, red color smoothed data using mean average filter, and yellow for median filter.
Sensors 22 07687 g007
Figure 8. Smoothed and normalized versus unsmoothed and normalized data: blue for unsmoothed normalized data, red for mean average filter-smoothed and normalized data, and yellow for median average filter-smoothed and normalized data.
Figure 8. Smoothed and normalized versus unsmoothed and normalized data: blue for unsmoothed normalized data, red for mean average filter-smoothed and normalized data, and yellow for median average filter-smoothed and normalized data.
Sensors 22 07687 g008
Figure 9. Basic architecture of artificial neural network and neuron.
Figure 9. Basic architecture of artificial neural network and neuron.
Sensors 22 07687 g009
Figure 10. Concept of support vector machines [4].
Figure 10. Concept of support vector machines [4].
Sensors 22 07687 g010
Figure 11. K-Nearest neighbor algorithm.
Figure 11. K-Nearest neighbor algorithm.
Sensors 22 07687 g011
Figure 12. Impact of the number of neurons and learning rates.
Figure 12. Impact of the number of neurons and learning rates.
Sensors 22 07687 g012
Figure 13. Regression of predicted and actual output values.
Figure 13. Regression of predicted and actual output values.
Sensors 22 07687 g013
Figure 14. Comparison of output values.
Figure 14. Comparison of output values.
Sensors 22 07687 g014
Figure 15. Performance of training, testing, and validation.
Figure 15. Performance of training, testing, and validation.
Sensors 22 07687 g015
Figure 16. Regression analysis of training, validation, and testing.
Figure 16. Regression analysis of training, validation, and testing.
Sensors 22 07687 g016
Figure 17. KNN with mean average filter confusion matrix.
Figure 17. KNN with mean average filter confusion matrix.
Sensors 22 07687 g017
Figure 18. KNN with median filter confusion matrix.
Figure 18. KNN with median filter confusion matrix.
Sensors 22 07687 g018
Figure 19. SVM with mean average filter confusion matrix.
Figure 19. SVM with mean average filter confusion matrix.
Sensors 22 07687 g019
Figure 20. SVM with median filter confusion matrix.
Figure 20. SVM with median filter confusion matrix.
Sensors 22 07687 g020
Table 1. Example of division of dataset for one day data.
Table 1. Example of division of dataset for one day data.
Division of DatasetHourT1T2T3T4T5T6T7T8T9T10T11T12T13T14 W T C
70%18:0014.314.314.214.214.114.113.813.613.19.78.07.77.87.53.20
19:0014.314.214.214.114.113.913.513.09.48.08.07.77.97.63.18
20:0014.214.214.114.113.913.612.89.28.07.87.97.77.97.71.97
21:0014.214.114.113.913.512.88.87.97.97.87.97.77.97.82.26
22:0014.114.113.913.612.68.77.77.87.97.87.97.88.07.82.02
23:0014.113.913.512.58.37.97.77.87.97.87.97.88.07.92.03
00:0013.813.512.08.37.97.87.77.87.97.87.97.88.07.92.61
00:0113.411.58.07.97.87.87.77.87.97.87.97.88.07.92.26
00:0210.58.17.97.97.87.87.77.87.97.87.97.88.07.83.13
15%00:037.97.97.97.97.87.87.77.87.97.87.87.57.77.33.04
00:047.97.97.97.87.87.87.77.87.87.67.47.07.17.04.0
15%00:057.97.97.87.87.87.87.77.87.57.06.96.76.86.733.1
00:067.97.97.87.87.87.87.67.36.96.76.76.56.76.620.0
Table 2. Hyper parameters.
Table 2. Hyper parameters.
S.NoModelHyper Parameters
01ANN
  • Hidden layers
  • Hidden Neurons
  • Activation function
02SVM
  • Value of regularized terms C
  • Kernel type
  • Degree of kernel function
  • Hyperplane
03KNN
  • K—the number of neighbors
Table 3. Final Parameters for the model.
Table 3. Final Parameters for the model.
Sr NoParameterRemarks
01Learning algorithmLM
02Activation functionLog-sigmoid
03Number of neurons10
04Learning rate0.01
Table 4. W T C magnitude and labels.
Table 4. W T C magnitude and labels.
Input Data W T C MagnitudeClass/Label
T1–T140.01–0.991
T1–T141.0–1.992
T1–T142.0–2.993
T1–T143.0–3.993
T1–T144.0–4.994
T1–T145.0–5.995
T1–T146.0–5.996
T1–T147.0–7.997
T1–T148.0–8.998
T1–T149.0–9.999
T1–T1410.0–10.9910
T1–T1411.0–11.9911
T1–T1412.0–12.9912
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Soomro, A.A.; Mokhtar, A.A.; Salilew, W.M.; Abdul Karim, Z.A.; Abbasi, A.; Lashari, N.; Jameel, S.M. Machine Learning Approach to Predict the Performance of a Stratified Thermal Energy Storage Tank at a District Cooling Plant Using Sensor Data. Sensors 2022, 22, 7687. https://doi.org/10.3390/s22197687

AMA Style

Soomro AA, Mokhtar AA, Salilew WM, Abdul Karim ZA, Abbasi A, Lashari N, Jameel SM. Machine Learning Approach to Predict the Performance of a Stratified Thermal Energy Storage Tank at a District Cooling Plant Using Sensor Data. Sensors. 2022; 22(19):7687. https://doi.org/10.3390/s22197687

Chicago/Turabian Style

Soomro, Afzal Ahmed, Ainul Akmar Mokhtar, Waleligne Molla Salilew, Zainal Ambri Abdul Karim, Aijaz Abbasi, Najeebullah Lashari, and Syed Muslim Jameel. 2022. "Machine Learning Approach to Predict the Performance of a Stratified Thermal Energy Storage Tank at a District Cooling Plant Using Sensor Data" Sensors 22, no. 19: 7687. https://doi.org/10.3390/s22197687

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop