Next Article in Journal
Action Graphs for Performing Goal Recognition Design on Human-Inhabited Environments
Next Article in Special Issue
Prediction of Computer Vision Syndrome in Health Personnel by Means of Genetic Algorithms and Binary Regression Trees
Previous Article in Journal
Proposition and Real-Time Implementation of an Energy-Aware Routing Protocol for a Software Defined Wireless Sensor Network
Previous Article in Special Issue
Short-Term Energy Demand Forecast in Hotels Using Hybrid Intelligent Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fault Detection System for a Geothermal Heat Exchanger Sensor Based on Intelligent Techniques

by
Héctor Aláiz-Moretón
1,*,†,
Manuel Castejón-Limas
2,†,
José-Luis Casteleiro-Roca
3,†,
Esteban Jove
3,†,
Laura Fernández Robles
2,† and
José Luis Calvo-Rolle
3,†
1
Departamento de Ingeniería de Sistemas y Automática, Universidad de León, 24071 León, Spain
2
Departamento de Ingenierías Mecánica, Informática y Aeroespacial, Universidad de León, 24071 León, Spain
3
Departamento de Ingeniería Industrial, Universidade da Coruña, 15405 Ferrol, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2019, 19(12), 2740; https://doi.org/10.3390/s19122740
Submission received: 10 May 2019 / Revised: 11 June 2019 / Accepted: 16 June 2019 / Published: 18 June 2019
(This article belongs to the Special Issue Selected Papers from HAIS2018)

Abstract

:
This paper proposes a methodology for dealing with an issue of crucial practical importance in real engineering systems such as fault detection and recovery of a sensor. The main goal is to define a strategy to identify a malfunctioning sensor and to establish the correct measurement value in those cases. As study case, we use the data collected from a geothermal heat exchanger installed as part of the heat pump installation in a bioclimatic house. The sensor behaviour is modeled by using six different machine learning techniques: Random decision forests, gradient boosting, extremely randomized trees, adaptive boosting, k-nearest neighbors, and shallow neural networks. The achieved results suggest that this methodology is a very satisfactory solution for this kind of systems.

1. Introduction

In recent years, most countries faced an important challenge in terms of global warming, economic instability and fossil fuels price dependency. In this context, the use of alternative energies has been promoted by the administrations. The most common alternative energy sources are the wind and solar energies, whose technologies have been subjected to significant advances. However, in addition to these two energies, the promotion of other renewable energies, such as oceanic or geothermal energy, have presented important developments in terms of efficiency [1].
Geothermal energy is defined as the heat energy stored under the ground. Dickson and Fanelli, in [2], presented an estimation of the amount of heat inside the earth rounds the 42 × 10 12 W. In spite of this high amount of energy, geothermal installations must be placed in specific areas with suitable geological conditions [3]. Around the world, its use represents 15 MW of non electrical applications, such as industrial processes, bathing or heat pumps, and 9 MW of the electrical ones.
The heat exchanger is a crucial component of a geothermal facility, and its main function is to absorb heat from the ground or transfer it. A geothermal heat exchanger can be placed under the ground in vertical or horizontal configurations [4,5]. On the one hand, vertical configurations are more efficient because at high depths, the ground temperature remains almost constant along the year. This means that, compared to the ambient temperature, the ground temperature would be higher in winter and lower in summer. On the other hand, horizontal configurations are less expensive, since the setup is simpler.
In this kind of facility, where the energy efficiency plays a significant role, the appearance of any kind of anomaly may lead to inefficient performance. Hence, in renewable energy systems, or any industrial plant in general terms, the anomaly detection is a crucial task [6,7,8,9,10,11]. These anomalies can be produced by wrong sensor readings, actuator malfunctions or changes in plant parameters, in general terms [12,13,14,15,16,17,18]. Focusing on the sensors performance, the occasional reading errors can be removed and recovered, making the systems more fault tolerant and robust [19,20,21,22,23,24,25].
This work deals with a geothermal heat pump facility used to provide thermal energy from the ground [26]. To achieve a geothermal system optimization, the good behaviour of the system equipment must be ensured. Then, the prediction of the correct sensor values is a key step to perform a proper fault identification and recovery. An anomaly would lead to a high deviation from the real and predicted value. In this case, the real measurement would be discarded and the value considered by the control system would be the predicted one.
With the aim of improving system performance, machine learning techniques are commonly considered. These techniques rely on actual observations registered from the system that are used to train the model. In this work, the sensor reading prediction is performed using intelligent models, trained with data from a geothermal heat pump installation. Four different intelligent techniques, commonly used for these kind of applications, were applied to the dataset: Shallow neural networks, extremely randomized trees, random decision forest and gradient boosting. Two possible approaches can be considered to obtain the best model. The first approach uses the whole dataset to train a global model. The second approach is based on the division of the dataset to apply the intelligent regression techniques over each group. In all cases, the models were tested using artificially generated outliers, obtaining successful results.
The document is organized following this structure: The next section describes the case of study. Then, Section 3 details the proposed fault detection and recovery system. The experiments performed and the obtained results are presented in Section 4. The results are discussed in Section 5 and, finally, the conclusions are explained in Section 6.

2. Case of Study

This section describes the geothermal heat exchanger facility under study located in a bioclimatic house.

2.1. Sotavento Bioclimatic House

The Sotavento bioclimatic house is a building dedicated to promote the use of alternative energies and the energy savings. These facilities, founded by the Sotavento Galicia Foundation, are located between the councils of Xermade and Monfero (Lugo), in the autonomous community of Galicia (Spain). Its geographical coordinates are 43 21 North, 7 52 West, with an elevation of 640 m above the sea and at a distance of 30 km from the sea.
Two different energy needs must be satisfied in the Sotavento bioclimatic house: The thermal and the electric energy. The thermal system has three different renewable energy sources: Geothermal, solar and biomass. These three sources ensure the thermal demand coverage. The thermal installation can be divided into three parts [27]:
  • Generation group: Three different renewable sources are exploited:
    -
    Geothermal system: A horizontal collector consisting of 5 loops of 100 m is placed under the ground at a depth of 2 m. The heat pump is a MAMY Genius—10.3 kW, and it has a nominal electrical power consumption of 1.9 kW and a nominal thermal power of 8.4 kW. The energy is absorbed from the ground and it is used to heat a mixture of water and glycol.
    -
    Solar thermal: Eight solar panels absorb the solar radiation to heat the ethyleneglycol flowing inside them.
    -
    Biomass boiler system: A biomass boiler type Ökofen, model Pallematic 20, with a configurable power of 20 kW, with a yield of pellets of 90%.
  • Energy accumulation group: The thermal energy storage is ensured using different accumulators. A solar accumulator of 1000 L receives the thermal energy from the solar system. In series, an inertial accumulator of 800 L stores the heat from the boiler and geothermal systems.
  • Consumption group: The thermal system must cover the demand of underfloor heating systems and Domestic Hot Water (DHW). The underfloor heating system is designed to keep the house temperature between 18 C and 22 C. The fluid temperature should remain between 35 C and 40 C. According to the Spanish Technical Building Code, the DHW demands 240 L/day.
In addition to the thermal systems, the Sotavento bioclimatic house has also an electrical installation with two renewable sources: Wind and photovoltaic. The electricity supplies the power systems and the lighting. To avoid power cuts, the house is connected to the power grid when it is demanded.

2.2. The Geothermal System

A more detailed explanation about the geothermal energy system is presented in this subsection. It is divided into two main parts described below: The heat pump and the heat exchanger (Figure 1).
Heat Pump. The Heat Pump has two different circuits; the primary one provides the heat from the ground (the geothermal exchanger) to the heat pump unit, and the other one is connected between the unit and the inertial accumulator. The energy absorbed from the ground is measured by two sensors.
Geothermal exchanger. The horizontal exchanger consists of five different circuits. The ground temperature along the exchanger is monitored using sensors distributed in four different loops. A scheme of the sensors located along the exchanger can be seen in Figure 2. Sensors S28 and S29 measure the energy absorbed from the ground and S401 measures the ground temperature. The rest of the sensors monitor the exchanger temperature in different points.

2.3. The Dataset

The initial dataset corresponds to the temperatures measured by the sensors during one year, registered with a sample time of 10 min.
Sensors S28 and S29 (the input and the output of the heat exchanger) are located inside the house. Hence, when the heat pump is off, these sensors measure the temperature inside the house. For this reason, the temperature is filtered to take into account only the data when the heat pump is on.
As this work proposes a model capable of predicting the sensor measurements to detect anomalies, only data from correct operation is considered. Then, to avoid wrong samples (bad sample time, bad range, open wires, etc.), the dataset was filtered to discard the erroneous data. After this conditioning step, the samples were reduced from 52,705 to 52,699.
However, as the appearance of any kind of anomaly in a sensor must be detected in a short time, the models were implemented with an amount of data corresponding to two days. These measurements are randomly selected from the 52,699 samples.

3. Fault Detection and Recovery (FDR) Approach—Used Techniques

The scheme defined for fault detection and recovery approach is shown in Figure 3. It is possible to divide the figure into two parts: The model and the fault detection and recovery block. The first one gives the prediction of each sensor based on the measurements made by the rest of the sensors. The second one compares the prediction with the real measurement, and analyzes the deviation based on a defined range. If there is a significant deviation, the valid signal is the prediction. Otherwise, the real measurement is set at the output.

3.1. FDR Steps

In this subsection, the necessary steps to accomplish the FDR developed approach are explained.
Sensor fault detection. Initially, a simple methodology for accomplishing sensor fault detection technique is used. The method allows a specific configuration of the range deviation. If the measured sample is out of this range, then a fault is labeled. The deviation percentage is referred to the operating temperature range.
Recovery. If a fault is detected, then it is necessary to recover the wrong sample with a value prediction. This prediction could be based on the other sensors readings, their previous values, and so on. To accomplish the recovery, a model must be implemented with the aim to predict an accurate value.

3.2. Used Techniques

The present subsection shows the different techniques used for accomplishing the objectives of the present research.

3.2.1. Analysis and Preprocessing

From the considered initial data, two different subclasses were created:
  • Day data cases.
  • Night data cases.
Knowing each date of the data recollection and the precise location of the installation under study, the sunrise and sunset times can be obtained. This is the criteria used to split the data in the two subclasses.
To obtain a representative model, some variables of the raw dataset have been selected. In addition, the previous state of some signals is included as an artificial input, for developing each experiment shown in Section 4.
The use of this extra information can be more beneficial than obtaining the model with original data features only. The election of these artificial features is always based on expert knowledge about the system behavior [28].
Based on a data description of the new dataset generated from the raw data, a common pre-processing procedure has been developed, including those experiments with previous values of different sensor like artificial variables.
The criterion for data normalizing is shown in Equation (1):
X i m e a n ( x ) s t d e v ( x )
The Standard Scaler data input pre-processing has been implemented with Python sklearn.preprocessing. StandardScaler [29] library. The main goal of the normalization step is to avoid the very soon convergence in the first iterations, when the training process of a particular regression method begins [30].

3.2.2. Regression Techniques

The recovery methodology purpose is to mimic the actual behaviour of the sensor. Thus, a predictive model trained with data acquired from the sensor is a sensible approach for achieving a computational representation of the sensor. Six different types of predictive models have been tested: Shallow neural networks, extremely randomized trees (ExtraTrees), random decision forests, adaptive boosting, k-nearest neighbors, and gradient boosting.
This choice of regressors pursues to represent the complexity of the sensor’s behaviour by two subtly different approaches: The shallow neural network solution features a single model capable of increasing its complexity by means of enlarging the number of neurons in its hidden layer; on the other hand, the extremely randomized trees, adaptive boosting, random decision forests, and gradient boosting regressors, belong to the ensemble methods family. Ensemble models provide their results by combining those obtained from multiple elementary models. In this case, complexity is approached by enlarging the number of simple models comprised in the ensemble.
The Multilayer Perceptron (MLP) is one of the most frequently used shallow neural network architectures. The good performance of this kind of artificial neural network has been proven in similar works such as [31,32,33]. Previous research [34] proved how this technique is capable of providing satisfactory results in the context of much larger amounts of data than those used in the case of study.
Ensemble methods, on the other hand, are among the most frequently used techniques for the excellent results they usually display. Examples of successful stories can be discovered following Kaggle’s machine learning competitions (https://www.kaggle.com/), where, along with Deep Neural Networks, ensemble methods such as those reported in this research are most frequently the winning techniques.
Each technique and the set of their associated parameters used in this work are explained bellow:
  • Shallow Neural Networks. Artificial Neural Networks can be used as universal approximators [35]. For this paper, a three layer Multi Layer Perceptron architecture was chosen: An input layer for capturing the sensor information, a hidden layer with non linear activation functions, and an output layer with one single neuron and a linear activation function to provide the prediction. The most important hyperparameters governing the regressor performance are the hidden layer size, the maximum number of iterations, the early stopping, the activation function, the nesterov momentum and the solver.
  • K-Nearest Neighbors. This is a representative of instance based techniques or non generalizing learning. Instead of representing the data via a model, this technique stores instance and uses a voting scheme on the nearest neighbors for obtaining the prediction on new data. This technique is a popular choice for setting a baseline for the prediction error. The most important hyperparameter is the number of neighbors.
  • Adaptive Boosting. This technique belongs to the stagewise additive models family. The prediction is based on a weighted sum of the simpler weak estimators it comprises. Each weak estimator is designed to concentrate on those samples that previous estimators found still to be difficult to fit. In this technique, the number of estimators is the most important hyperparameter to tune.
  • Random Decision Forests. Being one of the most popular ensemble methods, Random decision forests (RF) comprise a collection of simple decision trees whose results are considered to emit a final collective result. RF basic components can be built by considering a random limited number of features and/or a random limited number of observations. Thus, each component only has access to a fraction of the information and pays attention to specific details in the portion of information assigned to them. The combination of a number of these simple basic trees most frequently outperforms the results from a larger and more complex single tree. The number of estimators is the most important hyperparameter to tune.
  • Extremely Randomized Trees. They are similar to Random Forests, as they combine an ensemble of decision trees. Nevertheless, a few important differences are worth noting: Firstly, Extra Trees can provide piece-wise multilinear approximations to the training dataset instead of the piece-wise constants one provides by random forests. Secondly, Extra Trees are based on using random values for the optimal cut point choice, instead of bootstrapping to find the optimal cut point [36]. Similarly to RF, one of the most important hyperparameters to tune is the number of basic estimators.
  • Gradient Boosting. This technique builds the model following a stage-wise approach, by adding subsequent basic estimators in order to capture the unexplained information present in the residuals of former weak estimators [37]. The estimators frequently are decision trees and, similarly, the number of basic estimators is among the most important hyperparameters.

4. Experiments and Results

This section describes the different experiments carried out and the results obtained.

4.1. Experiment Definition

Depending on the predictors used in the predictive model, four different experiments are defined:
  • Experiment A: Prediction of sensor S-315 based on S-309 to S-316 signals
  • Experiment B: Prediction of sensor S-315 based on S-309 to S-316 signals and their previous states
  • Experiment C: Prediction of sensor S-315 based on S-309 to S-316 signals and S-315 previous state
  • Experiment D: Prediction of sensor S-315 based on S-309 to S-316 signals, their previous states, and S-315 previous state
In each experiment, the four regression techniques mentioned above —shallow neural networks, extremely randomized trees, random decision forests, and gradient boosting— are used to build two types of models, according to the data used for each one:
  • Global models: In this case the whole data set is used for training a single regressor.
  • Hybrid models: In this case, the data set is split into two groups in accordance to day and night criteria. Two different models are fit, one for day usage and another one for the night hours.

4.2. Error Metrics

In order to compare the different regression models obtained, the following error metrics have been implemented:
  • MAE: Mean Absolute Error. The goal of this metric is to measure the difference between predicted and real values. This metric has some advantages compared to other error measures [38].
    M A E = 1 m k = 1 m | Y k Y ^ k |
    where Y k is the observed value and Y ^ k is the foretold value.
  • LMLS: Least Mean Log Squares. This metric is used as regression loss function in the training process as well as in the validation error measure [39], Equation (3).
    L . M . L . S = 1 m k = 1 m log 1 + 1 2 Y k Y ^ k 2
    where Y k is the observed value and Y ^ k is the foretold value.
  • SMAPE: Symmetric Mean Absolute Percentage Error. The main goal of this metric is to explain relative errors thanks to the use of percentages [40], Equation (4).
    S . M . A . P . E = 2 m k = 1 m | Y k Y ^ k | Y k + Y ^ k
    where Y k is the observed value and Y ^ k is the foretold value.
  • MSE: Mean Squared Error. This metric can include the variance of error, it can be applied in several forecasting problems [41] Equation (5).
    M . S . E . = 1 m k = 1 m ( Y k Y ^ k ) 2
    where Y k is the observed value and Y ^ k is the foretold value.
  • MAPE: Mean Absolute Percentage Error. This error metric is one of the most common measures of the accuracy in regression problems [42], Equation (6).
    M . A . P . E = 100 % m k = 1 m | Y k Y ^ k | Y k
    where Y k is the observed value and Y ^ k is the foretold value.
  • NMSE: Normalised Mean Square Error. This a measure oriented to estimate the overall deviations between observed and predicted values [43], Equation (7).
    N . M . S . E = 1 m k = 1 m ( Y ^ k Y k ) 2 m e a n ( Y ^ k ) m e a n ( Y k )
    where Y k is the observed value and Y ^ k is the foretold value.

4.3. Experiments Setup

For each experiment the dataset was split into two subsets—training and test sets—as customary in data science projects in order to provide the error value on a held out dataset. Such an error represents the capability of the method to generalize the observed behavior to new unseen data. Thus, a fraction comprising 70% of the samples is used for training purposes to adjust the parameters of the models, while a fraction with 30% of the samples is used for final testing. In order to find the best combination of hyperparameters for each model, a grid search with ten fold cross validation has been carried out. The chosen scoring criteria was the negative mean square error. As a preprocessing step, the data is normalized before entering the regression model. In order to avoid leaking information from the validation test during cross validation, both the scaler and the regressor are embedded in a pipeline.
The four families of regressors, the scaler, the pipeline tool, and the grid search with cross validation, are implemented in Scikit-Learn’s machine learning library [44] which provides easy access to these techniques using Python as programming language for computational purposes.
The search space for the best values of the hyperparameters is reported below. Those hyperparameters not mentioned adopt Scikit-Learn default values.

4.3.1. Shallow Neural Network

  • hidden_layer_sizes=[(n,) for n in ( 5, 6, 7, 8)]
  • max_iter=[ 500_000]
  • learning_rate_init=[1e-1, 1e-2, 1e-3]
  • early_stopping=[True]
  • activation=[’relu’]
  • nesterovs_momentum=[True]
  • warm_start=[False]
  • solver=[’lbfgs’]

4.3.2. Extremely Randomized Tree

  • n_estimators=range(10, 100, 5)

4.3.3. Random Decision Forests

  • n_estimators=range(10, 100, 5)

4.3.4. Gradient Boosting

  • n_estimators=range(10, 100, 5)
  • learning_rate=np.linspace(1e-3, 1e-1, 5)
  • n_iter_no_change=[2]

4.3.5. AdaBoost

  • n_estimators=range(10, 100, 5)

4.3.6. K-Neareat Neighbors

  • n_neighbors=range(5, 20, 5)
Table 1, Table 2 and Table 3 show the results obtained in the experiments by the global and hybrid approaches (best ones in bold). According to most error metrics, the ExtraTrees regressor achieves the best results in both global and hybrid approaches. Among these two, the hybrid approach displays better results, particularly according to the mean absolute error criteria, the easiest to interpret by human beings. Figure 4 and Figure 5 display the results obtained by the six types of regressors considered, in this case using the data from experiment A. It is clear that the Extra Trees regressor achieves great resemblance with the actual data recorded from the sensor in what are considered very satisfactory results.

5. Discussion

The number of experiments, regressors and error metrics reported in this paper builds a complex scenario when attempting to establish a single winner solution. As it usually happens in real engineering problems, the solution to a problem is not unique and the context determines the preferred one.
From a strictly numerical point of view, it could be argued that Experiment A frequently displays best error values. In those cases where it fails to outperform other experiments, the results are not significantly different from the minimum.
Considering the experiments from the point of view of their complexity, Experiment A also represents the simplest configuration as it requires the lowest number of input variables; a fact that, in absence of significant differences in performance with respect to the rest of the experiments, also advocates for its designation as the preferred configuration.
In economic terms, the context around this study case does not justify a more complex configuration. In some scenarios, e.g., optimizing a quality feature in manufacturing processes, a marginal improvement in the prediction model leads to significant economic benefits; but that is hardly the case of the study case reported in this paper: A predictive model that is used as a backup for the real sensor and whose reads are only considered during malfunctioning.
According to these former criteria, Experiment A could be considered the best choice, but considering the following practicalities, the final decision might differ. Firstly, an important issue to consider is the intrinsic precision of the actual sensor being modeled. If the performance difference between alternatives is orders of magnitude lower than the sensor precision, then those alternatives are in fact equally optimal. Secondly, the results must be considered from the point of view of the subsequent data consumption. If the sensor data is to be further processed by an algorithm sensitive to a specific precision, it makes little sense considering differences in considerable smaller differences, e.g., comparing two temperature readings with values 82.15 F and 82.17 F when the on–off controller driving a pump already made a decision at a 60 F threshold. This paper deliberately does not specify the particular subsequent model that the sensor signal feeds, as many such systems can be considered. Essentially, it is the engineer’s call to weigh the context factors and choose the optimal solution for the problem at hand, the numerical error scored by each alternative being an important but not unique criterion in the decision making process. For the study case reported in this paper, Experiment A using Extra Trees was adopted as the preferred solution. Nevertheless, the approach proposed relies on training data from a short period of time, two days, which makes it possible to periodically retrain the models and perform the comparison to select the new best choice and adapt for future changes.

6. Conclusions and Future Works

A methodology for recovering data missing in malfunctioning state sensor and the sensor fault detection have been addressed in this research successfully. Sensor fault detection procedure is relaying on tagging data as fault, when a measured sample is out of the range derivation. Moreover, the procedure for recovering data missing is based on the implementation of several experiments with the aim to get the best way to define a model when it is trying to get measurements of a sensor with problems. Input data features election is relevant when a robust regression model wants to be created to predict missing data in a process where the temperature is involved—more concretely, the election of new features and how these are estimated or calculated. In this research, new artificial features based on the sensor values on the previous state are added to achieve and compare a global model and hybrid model for recovering missing data of a sensor. Results prove that a hybrid model implemented with an Extremely Ramdomized Trees regressor, composed by day and night submodels not including previous state values as artificial features, is the best way for recovering data missing. Future works will explore the improvement of the sensor fault detection procedure via anomaly detection techniques such as Isolation Forest, One Class SVM (Support Vector Machines), Local Outlier Factor, and Elliptic Envelope. From the point of view of recovering missing data, new experiments based on time series oriented to prevent the use of previous state information will be implemented. Some new, complex and data fusion models will be used also in the next research phase.

Author Contributions

Data curation, L.F.R.; Investigation, M.C.-L and E.J.; Methodology, L.F.R.; Project administration, H.A.-M. and J.L.C.-R. (José Luis Calvo-Rolle); Software, H.A.-M.; Supervision, J.L.C.-R. (José Luis Calvo-Rolle); Validation, M.C.-L. and J.-L.C.-R. (José Luis Casteleiro-Roca); Writing, original draft, J.-L.C.-R. (José Luis Calvo-Rolle) and E.J.

Funding

Junta de Castilla y León—Consejería de Educación; Project: LE078G18. UXXI2018/000149. U-220. Ministerio de Economía, Industria y Competitividad: Project grant DPI2016-79960-C3-2-P; and NVIDIA GPU Grant Program.

Acknowledgments

We would like to thank “Instituto Enerxético de Galicia” (INEGA) and “Parque Eólico Experimental de Sotavento” (Sotavento Foundation) for their technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kaltschmitt, M.; Streicher, W.; Wiese, A. Renewable Energy; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  2. Dickson, M.H.; Fanelli, M. Geothermal Energy: Utilization and Technology; Routledge: Abingdon, UK, 2013. [Google Scholar]
  3. Ozgener, L.; Ozgener, O. Monitoring of energy exergy efficiencies and exergoeconomic parameters of geothermal district heating systems (GDHSs). Appl. Energy 2009, 86, 1704–1711. [Google Scholar] [CrossRef]
  4. Kakaç, S.; Liu, H.; Pramuanjaroenkij, A. Heat Exchangers: Selection, Rating, and Thermal Design, 2nd ed.; Designing for Heat Transfer, Taylor & Francis: Abingdon, UK, 2002. [Google Scholar]
  5. Sauer, H.; Howell, R. Heat Pump Systems; Krieger Publishing Company: Malabar, FL, USA, 1991. [Google Scholar]
  6. Quintian Pardo, H.; Calvo Rolle, J.L.; Fontenla Romero, O. Application of a low cost commercial robot in tasks of tracking of objects. Dyna 2012, 79, 24–33. [Google Scholar]
  7. Rolle, J.; Gonzalez, I.; Garcia, H. Neuro-robust controller for non-linear systems. Dyna 2011, 86, 308–317. [Google Scholar] [CrossRef]
  8. Alaiz Moretón, H.; Calvo Rolle, J.; García, I.; Alonso Alvarez, A. Formalization and practical implementation of a conceptual model for PID controller tuning. Asian J. Control 2011, 13, 773–784. [Google Scholar] [CrossRef]
  9. Garcia, R.F.; Rolle, J.L.C.; Castelo, J.P.; Gomez, M.R. On the monitoring task of solar thermal fluid transfer systems using NN based models and rule based techniques. Eng. Appl. Artif. Intell. 2014, 27, 129–136. [Google Scholar] [CrossRef]
  10. González Gutiérrez, C.; Sánchez Rodríguez, M.L.; Fernández Díaz, R.Á.; Calvo Rolle, J.L.; Roqueñí Gutiérrez, N.; Javier de Cos Juez, F. Rapid tomographic reconstruction through GPU-based adaptive optics. Log. J. IGPL 2018, 27, 214–226. [Google Scholar] [CrossRef]
  11. Baruque, B.; Porras, S.; Jove, E.; Calvo-Rolle, J.L. Geothermal heat exchanger energy prediction based on time series and monitoring sensors optimization. Energy 2019, 171, 49–60. [Google Scholar] [CrossRef]
  12. Chiang, L.H.; Russell, E.L.; Braatz, R.D. Fault Detection and Diagnosis in Industrial Systems; Springer Science & Business Media: Berlin, Germany, 2000. [Google Scholar]
  13. Casteleiro-Roca, J.L.; Pérez, J.A.M.; Piñón-Pazos, A.J.; Calvo-Rolle, J.L.; Corchado, E. Modeling the electromyogram (EMG) of patients undergoing anesthesia during surgery. In Proceedings of the 10th International Conference on Soft Computing Models in Industrial and Environmental Applications, Burgos, Spain, 15–17 June 2015; pp. 273–283. [Google Scholar]
  14. Vega Vega, R.; Quintián, H.; Calvo-Rolle, J.L.; Herrero, Á.; Corchado, E. Gaining deep knowledge of Android malware families through dimensionality reduction techniques. Log. J. IGPL 2018, 27, 160–176. [Google Scholar] [CrossRef]
  15. Quintián, H.; Casteleiro-Roca, J.L.; Perez-Castelo, F.J.; Calvo-Rolle, J.L.; Corchado, E. Hybrid intelligent model for fault detection of a lithium iron phosphate power cell used in electric vehicles. In Proceedings of the International Conference on Hybrid Artificial Intelligence Systems, Seville, Spain, 18–20 April 2016; pp. 751–762. [Google Scholar]
  16. Jove, E.; Gonzalez-Cava, J.M.; Casteleiro-Roca, J.L.; Méndez-Pérez, J.A.; Antonio Reboso-Morales, J.; Javier Pérez-Castelo, F.; Javier de Cos Juez, F.; Luis Calvo-Rolle, J. Modelling the hypnotic patient response in general anaesthesia using intelligent models. Log. J. IGPL 2018, 27, 189–201. [Google Scholar] [CrossRef]
  17. Gonzalez-Cava, J.M.; Reboso, J.A.; Casteleiro-Roca, J.L.; Calvo-Rolle, J.L.; Méndez Pérez, J.A. A novel fuzzy algorithm to introduce new variables in the drug supply decision-making process in medicine. Complexity 2018, 2018, 9012720. [Google Scholar] [CrossRef]
  18. Casteleiro-Roca, J.L.; Jove, E.; Gonzalez-Cava, J.M.; Méndez Pérez, J.A.; Calvo-Rolle, J.L.; Blanco Alvarez, F. Hybrid model for the ANI index prediction using Remifentanil drug and EMG signal. Neural Comput. Appl. 2018. [Google Scholar] [CrossRef]
  19. Casteleiro-Roca, J.L.; Quintián, H.; Calvo-Rolle, J.L.; Corchado, E.; del Carmen Meizoso-López, M.; Piñón-Pazos, A. An intelligent fault detection system for a heat pump installation based on a geothermal heat exchanger. J. Appl. Log. 2016, 17, 36–47. [Google Scholar] [CrossRef] [Green Version]
  20. Vilar-Martinez, X.M.; Montero-Sousa, J.A.; Calvo-Rolle, J.L.; Casteleiro-Roca, J.L. Expert system development to assist on the verification of “TACAN” system performance. Dyna 2014, 89, 112–121. [Google Scholar]
  21. Casteleiro-Roca, J.L.; Calvo-Rolle, J.L.; Méndez Pérez, J.A.; Roqueñí Gutiérrez, N.; de Cos Juez, F.J. Hybrid Intelligent System to Perform Fault Detection on BIS Sensor During Surgeries. Sensors 2017, 17, 179. [Google Scholar] [CrossRef] [PubMed]
  22. Marrero, A.; Méndez, J.; Reboso, J.; Martín, I.; Calvo, J. Adaptive fuzzy modeling of the hypnotic process in anesthesia. J. Clin. Monit. Comput. 2017, 31, 319–330. [Google Scholar] [CrossRef] [PubMed]
  23. Quintián, H.; Corchado, E. Beta scale invariant map. Eng. Appl. Artif. Intell. 2017, 59, 218–235. [Google Scholar] [CrossRef]
  24. Jove, E.; López, J.A.V.; Fernández-Ibáñez, I.; Casteleiro-Roca, J.L.; Calvo-Rolle, J.L. Hybrid intelligent system topredict the individual academic performance of engineering students. Int. J. Eng. Educ. 2018, 34, 895–904. [Google Scholar]
  25. Jove, E.; Casteleiro-Roca, J.L.; Quintián, H.; Méndez-Pérez, J.A.; Calvo-Rolle, J.L. A fault detection system based on unsupervised techniques for industrial control loops. Expert Syst. 2019, e12395. [Google Scholar] [CrossRef]
  26. Ozgener, L. A review on the experimental and analytical analysis of earth to air heat exchanger (EAHE) systems in Turkey. Renew. Sustain. Energy Rev. 2011, 15, 4483–4490. [Google Scholar] [CrossRef]
  27. Cabrerizo, J.A.R.; Santos, M. ParaTrough: Modelica-based Simulation Library for Solar Thermal Plants. Revista Iberoamericana de Automática e Informática Industrial RIAI 2017, 14, 412–423. [Google Scholar] [CrossRef]
  28. Tuv, E. Feature Selection with Ensembles, Artificial Variables, and Redundancy Elimination. J. Mach. Learn. Res. 2009, 10, 1341–1366. [Google Scholar] [CrossRef]
  29. Developers, S.L. scikit-learn v0.19.1. Available online: https://sklearn.org/modules/classes.html (accessed on 15 January 2019).
  30. Géron, A. Hands-On Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques for Building Intelligent Systems; O’Reilly Media: Sebastopol, CA, USA, 2017. [Google Scholar]
  31. Jove, E.; Gonzalez-Cava, J.M.; Casteleiro-Roca, J.L.; Pérez, J.A.M.; Calvo-Rolle, J.L.; de Cos Juez, F.J. An Intelligent Model to Predict ANI in Patients Undergoing General Anesthesia. In Proceedings of the International Joint Conference SOCO’17-CISIS’17-ICEUTE’17, León, Spain, 6–8 September 2017; Pérez García, H., Alfonso-Cendón, J., Sánchez González, L., Quintián, H., Corchado, E., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 492–501. [Google Scholar]
  32. Casteleiro-Roca, J.L.; Jove, E.; Sánchez-Lasheras, F.; Méndez-Pérez, J.A.; Calvo-Rolle, J.L.; de Cos Juez, F.J. Power Cell SOC Modelling for Intelligent Virtual Sensor Implementation. J. Sens. 2017, 2017, 9640546. [Google Scholar] [CrossRef]
  33. Casteleiro-Roca, J.; Calvo-Rolle, J.; Meizoso-López, M.; Piñón-Pazos, A.; Rodríguez-Gómez, B. Bio-inspired model of ground temperature behavior on the horizontal geothermal exchanger of an installation based on a heat pump. Neurocomputing 2015, 150 Pt A, 90–98. [Google Scholar] [CrossRef] [Green Version]
  34. Alaiz-Moretón, H.; Casteleiro-Roca, J.L.; Robles, L.F.; Jove, E.; Castejón-Limas, M.; Calvo-Rolle, J.L. Sensor Fault Detection and Recovery Methodology for a Geothermal Heat Exchanger. In Hybrid Artificial Intelligent Systems; de Cos Juez, F.J., Villar, J.R., de la Cal, E.A., Herrero, Á., Quintián, H., Sáez, J.A., Corchado, E., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 171–184. [Google Scholar]
  35. Hornik, K. Approximation Capabilities of Multilayer Feedforward Network. Neural Netw. 1991, 4, 251–257. [Google Scholar] [CrossRef]
  36. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  37. Friedman, J. Stochastic Gradient Boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  38. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  39. Campoy, A.M.; Rodríguez-Ballester, F.; Carot, R.O. Using dynamic, full cache locking and genetic algorithms for cache size minimization in multitasking, preemptive, real-time systems. In Proceedings of the International Conference on Theory and Practice of Natural Computing, Caceres, Spain, 3–5 December 2013; pp. 157–168. [Google Scholar]
  40. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef] [Green Version]
  41. Wang, Z.; Bovik, A.C. Mean squared error: Love it or leave it? A new look at signal fidelity measures. IEEE Signal Process. Mag. 2009, 26, 98–117. [Google Scholar] [CrossRef]
  42. Kim, S.; Kim, H. A new metric of absolute percentage error for intermittent demand forecasts. Int. J. Forecast. 2016, 32, 669–679. [Google Scholar] [CrossRef]
  43. Poli, A.A.; Cirillo, M.C. On the use of the normalized mean square error in evaluating dispersion model performance. Atmos. Environ. Part A Gen. Top. 1993, 27, 2427–2434. [Google Scholar] [CrossRef]
  44. Buitinck, L.; Louppe, G.; Blondel, M.; Pedregosa, F.; Mueller, A.; Grisel, O.; Niculae, V.; Prettenhofer, P.; Gramfort, A.; Grobler, J.; et al. API design for machine learning software: Experiences from the scikit-learn project. In Proceedings of the ECML PKDD Workshop: Languages for Data Mining and Machine Learning, Prague, Czech Republic, 23–27 September 2013; pp. 108–122. [Google Scholar]
Figure 1. Heat pump and horizontal exchanger layout.
Figure 1. Heat pump and horizontal exchanger layout.
Sensors 19 02740 g001
Figure 2. Geothermal exchanger sensors layout.
Figure 2. Geothermal exchanger sensors layout.
Sensors 19 02740 g002
Figure 3. Fault detection and recovery approach.
Figure 3. Fault detection and recovery approach.
Sensors 19 02740 g003
Figure 4. Actual data vs. predictions for Experiment A. ExtraTrees, random forest, gradient boosting, and MLP are considered for each model (day, night, and global model).
Figure 4. Actual data vs. predictions for Experiment A. ExtraTrees, random forest, gradient boosting, and MLP are considered for each model (day, night, and global model).
Sensors 19 02740 g004aSensors 19 02740 g004b
Figure 5. Actual data vs. predictions for Experiment A. AdaBoost and k-nearest neighbors are considered for each model (day, night, and global model).
Figure 5. Actual data vs. predictions for Experiment A. AdaBoost and k-nearest neighbors are considered for each model (day, night, and global model).
Sensors 19 02740 g005
Table 1. Global model errors (multiplied by 10 5 ) for extremely randomized trees (ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
Table 1. Global model errors (multiplied by 10 5 ) for extremely randomized trees (ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
ErrorExperimentETGBMLPRFABK-NN
LMLSA2.424.46.03.72.664.76
B2.421.26.73.53.164.76
C2.621.24.53.02.874.76
D2.316.618.73.23.264.76
MAEA243.1880.6495.3280.3263.76353.77
B240.1868.9620.1300.8317.70353.77
C249.5821.5414.9280.7277.65353.77
D243.5768.2855.9283.5336.57353.77
MAPEA29.2106.059.933.731.7242.62
B28.9104.674.936.238.2442.62
C30.098.950.133.833.4042.62
D29.392.5103.234.140.5242.62
MSEA4.848.812.07.45.329.53
B4.942.513.47.06.339.53
C5.242.59.06.15.759.53
D4.633.237.56.36.519.53
NMSEA508.4535.7979.6370.6572.03108.90
B527.9570.0250.6320.7540.69108.90
C561.2569.8867.9407.3537.47108.90
D658.0257.41883.6428.9502.79108.90
SMAPEA29.3106.359.933.731.7542.67
B28.9104.975.036.238.2842.67
C30.099.150.133.833.4442.67
D29.392.7103.434.140.5642.67
Table 2. Day model errors (multiplied by 10 5 ) for extremely randomized trees(ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
Table 2. Day model errors (multiplied by 10 5 ) for extremely randomized trees(ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
ErrorExperimentETGBMLPRFABK-NN
LMLSA2.915.71153.23.33.565.81
B3.219.577.73.93.965.81
C3.229.516.33.94.155.81
D3.134.1304.74.03.735.81
MAEA232.0689.13079.4269.9355.44363.79
B255.1832.21515.3318.7332.95363.79
C280.81102.7727.6321.4483.44363.79
D278.41136.01675.1320.7407.91363.79
MAPEA27.882.8372.632.442.7543.75
B30.6100.2183.238.340.0043.75
C33.8132.887.838.658.2743.75
D33.5136.8202.638.549.1043.75
MSEA5.931.53457.96.77.1211.62
B6.339.0157.57.87.9311.62
C6.459.132.67.98.3011.62
D6.268.3672.68.07.4511.62
NMSEA799.6783.718418.0544.7724.6898.50
B425.2347.07103.5366.3731.7698.50
C459.0193.21652.2361.8111.5198.50
D434.8147.313330.6408.733.0298.50
SMAPEA27.983.0349.532.442.8043.82
B30.7100.4184.338.340.0543.82
C33.8133.188.038.758.3243.82
D33.5137.1198.038.649.1543.82
Table 3. Night model errors (multiplied by 10 5 ) for extremely randomized trees (ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
Table 3. Night model errors (multiplied by 10 5 ) for extremely randomized trees (ET), gradient boosting (GB), multi-layer perceptron (MLP), random forest (RF), adaptive boosting (AB), and k-nearest neighbors (K-NN).
ErrorExperimentETGBMLPRFABK-NN
LMLSA0.050.100.30.100.090.07
B0.040.103254.60.080.100.06
C0.050.100.100.070.090.06
D0.040.10633.30.060.100.06
MAEA38.667.8141.355.642.0039.90
B33.657.614477.549.552.5035.70
C35.365.1110.648.142.0037.80
D33.367.55595.945.952.5037.80
MAPEA4.78.217.16.75.094.83
B4.17.01754.56.06.364.32
C4.37.913.45.85.094.58
D4.08.2678.25.66.364.58
MSEA0.110.210.550.200.180.13
B0.080.216996.170.160.200.11
C0.110.200.200.140.180.12
D0.080.231291.100.130.200.12
NMSEA4735.48055.618847.77114.26805.56822.22
B2407.61427.016796.85333.36388.891355.56
C5349.26283.422194.25732.96805.562355.56
D3829.97114.864166.13324.28055.562356.56
SMAPEA4.78.217.16.75.094.83
B4.17.01711.66.06.364.83
C4.37.913.45.85.094.83
D4.08.2687.55.66.364.83

Share and Cite

MDPI and ACS Style

Aláiz-Moretón, H.; Castejón-Limas, M.; Casteleiro-Roca, J.-L.; Jove, E.; Fernández Robles, L.; Calvo-Rolle, J.L. A Fault Detection System for a Geothermal Heat Exchanger Sensor Based on Intelligent Techniques. Sensors 2019, 19, 2740. https://doi.org/10.3390/s19122740

AMA Style

Aláiz-Moretón H, Castejón-Limas M, Casteleiro-Roca J-L, Jove E, Fernández Robles L, Calvo-Rolle JL. A Fault Detection System for a Geothermal Heat Exchanger Sensor Based on Intelligent Techniques. Sensors. 2019; 19(12):2740. https://doi.org/10.3390/s19122740

Chicago/Turabian Style

Aláiz-Moretón, Héctor, Manuel Castejón-Limas, José-Luis Casteleiro-Roca, Esteban Jove, Laura Fernández Robles, and José Luis Calvo-Rolle. 2019. "A Fault Detection System for a Geothermal Heat Exchanger Sensor Based on Intelligent Techniques" Sensors 19, no. 12: 2740. https://doi.org/10.3390/s19122740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop