Next Article in Journal
Structural Health Monitoring of Dams Based on Acoustic Monitoring, Deep Neural Networks, Fuzzy Logic and a CUSUM Control Algorithm
Next Article in Special Issue
Community- and Data-Driven Services for Multi-Policy Pedestrian Routing
Previous Article in Journal
A Distributed Fault Diagnosis and Cooperative Fault-Tolerant Control Design Framework for Distributed Interconnected Systems
Previous Article in Special Issue
Implementing Replication of Objects in DOORS—The Object-Oriented Runtime System for Edge Computing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Anomaly Detection with Feature Extraction Based on Machine Learning Using Hydraulic System IoT Sensor Data

Department of Information & Statistics, Chungbuk National University, Cheongju 28644, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2479; https://doi.org/10.3390/s22072479
Submission received: 7 February 2022 / Revised: 11 March 2022 / Accepted: 13 March 2022 / Published: 23 March 2022
(This article belongs to the Special Issue Cyber-Physical Systems - from Perception to Action)

Abstract

:
Hydraulic systems are advanced in function and level as they are used in various industrial fields. Furthermore, condition monitoring using internet of things (IoT) sensors is applied for system maintenance and management. In this study, meaningful features were identified through extraction and selection of various features, and classification evaluation metrics were presented through machine learning and deep learning to expand the diagnosis of abnormalities and defects in each component of the hydraulic system. Data collected from IoT sensor data in the time domain were divided into clusters in predefined sections. The shape and density characteristics were extracted by cluster. Among 2335 newly extracted features, related features were selected using correlation coefficients and the Boruta algorithm for each hydraulic component and used for model learning. Linear discriminant analysis (LDA), logistic regression, support vector classifier (SVC), decision tree, random forest, XGBoost, LightGBM, and multi-layer perceptron were used to calculate the true positive rate (TPR) and true negative rate (TNR) for each hydraulic component to detect normal and abnormal conditions. Valve condition, internal pump leakage, and hydraulic accumulator data showed TPR performance of 0.94 or more and a TNR performance of 0.84 or more. This study’s findings can help to determine the stable and unstable states of each component of the hydraulic system and form the basis for engineers’ judgment.

1. Introduction

Hydraulic systems are applied in various industries regardless of scale utilizing the properties of liquid that can be delivered to all parts without the loss of applied pressure. As hydraulic systems are applied in various fields, problems that may arise can manifest in different ways in industrial applications. Fluids within hydraulic systems have generally low flammability, but if they leak in the form of a spray, they can become highly flammable and cause severe problems [1].
With an aircraft that is controlled by a hydraulic system, a problem is contained by controlling the contamination of the fluid inside the system [2]. Inefficiency may arise because of a lack of control of the performance of the hydraulic system, which may waste energy [3]. Therefore, it is necessary to manage the performance, problems, and risk factors of a hydraulic system.
Research and development are in progress to monitor the status of hydraulic systems continuously and to detect and control possible issues in advance to minimize the occurrence of delays and financial problems caused by investigating how to solve hydraulic system problems and providing solutions.
The hydraulic system is largely composed of an accumulator, cooler, pump, and valve, and the characteristics of each are as follows. The hydraulic accumulator of the hydraulic system operates through fluid and gas, and since the fluid is relatively incompressible compared to gas, it is ideal for power transmission. On the other hand, gas has the characteristic of compressing a small volume at high pressure compared to a fluid; thus, energy can be stored in the gas, and thereafter transferred to the fluid as the high-pressure stage is released.
The hydraulic cooler generates heat while converting and transferring energy in the hydraulic system. As the fluid’s temperature increases, internal wear of the system occurs, and the worn particles become unnecessary elements in the fluid and may damage other components. The hydraulic valve can control the fluid’s flow and its pressure by opening and closing. The hydraulic pump is a component that can transfer energy to the fluid and is the core of the hydraulic system.
If at least one of the hydraulic system’s components does not work properly, a problem may occur in the entire system. For example, if the generated heat is not cooled, heat is applied to the fluid, which can cause further damage due to aging and loss of lubricity. In addition, leakage inside the hydraulic system can degrade the performance and efficiency of the system, worsening the problem over time. Various measures and control methods have been proposed to reduce vibration or noise generated in the hydraulic system as a signal of abnormality or the occurrence of a problem [4]. Among them, condition monitoring can reduce human inspection requirements, detect errors before they escalate, reduce maintenance costs, and improve safety and reliability [5]. In addition, it can maximize productivity by minimizing system downtime [6].
As the structure of hydraulic system-related equipment becomes increasingly complex, functions are strengthened and the level of automation is increased; hence, subjective error diagnosis based on visual observation and disassembly reaches its limit. Error diagnosis technology has been continuously developed in line with the level of technological growth. Research on fault diagnosis using machine and deep learning is being actively conducted after diagnosing hydraulic system faults based on signal processing technology and mathematical models [7]. Especially in condition monitoring, high-dimensional datasets have overlapping information to characterize properties of faults. Recently, to construct high-dimensional datasets, an intrinsic mode function (IMF) has been extracted from a raw signal using empirical mode decomposition (EMD) among signal decomposition methods, and numerical features are estimated for each function. However, dimensional reduction is required to prevent low diagnostic performance and overfitting of machine learning models due to high-dimensional datasets [8]. For example, collect real-time high-dimensional measurements from multiple sensors to diagnose faults in internal combustion engines. In constructing a diagnostic model, dimension reduction is required by feature extraction of a high-dimensional dataset in order to separate sensitive information that can positively affect diagnostic results and computational efficiency [9].
Similar to a previous study, this work proposes a supervised learning-based linear discriminant analysis suitable for detecting various types of errors in a wide time range [10]. Robustness is added to the condition monitoring system by implementing the common errors of the raw signals (constant offset, drift, noise, and signal peak) of the sensors used in hydraulic systems [11]. In addition, if the model optimized through random forest is used, even with limited sensor data, it can be sufficient to classify the hydraulic system’s state and can reduce storage capacity and training time [12].
Statistical-based features were extracted from sensor data, which varied depending on whether the vibration data generated in the hydraulic system were in stable or unstable states, to identify differences based on these states [13]. When the classification algorithm was performed by extracting and using new features from vibration data with fast Fourier transform, classification accuracy close to 1 was achieved [14].
We reduced the dimension of observations through principal component analysis (PCA), diagnosed hydraulic valve failures using XGBoost, realized higher accuracy than random forest [15], and built a sensor sensitivity prediction model based on a convolution neural network (CNN), so that the prediction quality matched that of the previously built model [16]. Data related to anomaly and problem detection often have imbalanced target variables; therefore, some studies classify them after applying oversampling, undersampling, and ensemble methods [17].
To detect internal leakage faults that affect the hydraulic system’s dynamic performance and reduce energy efficiency, periodic data were analyzed through trained support vector classifier (SVC) to detect faults early [18]. In addition, a more efficient electricity production process was proposed by integrating the production process control system and IT system by designing a method to detect faults in the water pump early based on data measured by the control device through internet of things (IoT)-enabled predictive maintenance [19].
In previous studies, duplicate data were removed and new information was extracted by mapping data to a new dimension. However, in this study, feature extraction was performed based on the shape and distribution density characteristic of the raw signal to create a new combination of features and by selecting the upper features related to the components of the hydraulic system, which contributed to the improvement of classification performance and diagnosed the abnormalities and faults of each component of the hydraulic system further.
In addition, the performance evaluation of the anomaly detection model was performed by measuring accuracy, precision, true positive rate (TPR), and true negative rate (TNR) from the confusion matrix of classification. In particular, the stability of the hydraulic system was classified through TPR to prevent incorrect system suspension due to misclassification, and accidents were prevented in advance by determining defects in the hydraulic system early through TNR. It can also be proposed as a basis for judgment of engineers and used as a clear measure of judgment by decision-makers.
The structure of this paper is as follows. In the Data Section, sensor data for each hydraulic system component are introduced. Data preparation through data preprocessing, feature extraction, feature selection, and introducing methodologies and model evaluation metrics for anomaly and defect classification through model learning is described in the Study Design Section. The Results Section reveals the classification of abnormalities’ results, defects, and the selection of excellent models through the model evaluation index, and the Discussion Section identifies strengths and weaknesses through comparison with previous studies and suggests future research tasks. Finally, the study is summarized in the Conclusion.

2. Hydraulic System Cyclical Sensor Data

To minimize any additional damage that can occur due to a possible defect in the hydraulic system, we determine whether an anomaly has occurred based on vibration data. Vibration data in a fluidized mechanical system is greatly influenced by lubrication temperature and operating speed, so attention should be paid to test temperature and working speed [20]. Since the vibrations that occur when an abnormality or problem occurs differ from what is normal, abnormality detection is conducted based on this information. Vibration can occur when a series of mechanisms do not operate within the proper range of motion.
The data used in this study can be obtained from the UCI Machine Learning Repository [10]. Data were obtained experimentally using hydraulic test equipment.
A hydraulic system circuit represents a set consisting of several hydraulic components. It controls the flow, position, and pressure of fluid based on circuit components such as hydraulic pumps, pipes, and hydraulic motors. Figure 1 describes a promised representation of a complex hydraulic system circuit for a more convenient view. It is easy to comprehend the hydraulic system circuit of the hydraulic test equipment that generates data using the promised representation in Figure 1. The hydraulic test equipment that generates the data proposed in this paper consisted of a primary circuit connected via an oil reservoir and a secondary cooling filtration circuit. In a working circuit with a main pump (electrical motor power of 3.3 kW), different load levels are cyclically repeated using a proportional pressure relief valve (V11) [15]. Experiments with predefined load levels are possible. The sensor data were collected from a PLC (Beckhoff CX5020) which transmitted the data to a PC via EtherCAT [16,21].
As the system repeated a cycle, it quantitatively changed the status of the four hydraulic components (accumulator, cooler, pump, and valve) presented in Table 1. Data was repeated 2205 times with a cycle of 60 s to change the status of the hydraulic system for each cycle. Furthermore, a binary value about stable flag indicated the presence or absence of an abnormal state according to the hydraulic system’s state corresponding to each cycle.
The pressure (PS1 to PS6, 100 Hz), motor power (EPS1, 100 Hz), volume flows (FS1 to FS2, 10 Hz), temperature (TS1 to TS4, 1 Hz), and vibration (VS1, 1 Hz) presented in Table 2 were obtained from the actual sensor. Cooling efficiency (CE, 1 Hz), cooling power (CP, 1 Hz), and system efficiency (SE, 1 Hz) were values measured by a virtual sensor. Therefore, 17 sensors’ data were measured in total.

3. Study Design

In this study, various analyses were conducted to detect problems and abnormalities in the hydraulic system. The data were pre-processed for correct data shape and feature extraction; the train/test set was separated to verify whether a correct model that was not overfitted to the training data was learned after model training. Appropriate information was extracted from the shape and data distribution of the sensor data, and the extracted features were selected using an appropriate algorithm and used as predictive variables for model training. The feature most related to the components of the hydraulic system was selected and used as a feature when learning the model.
Figure 2 shows the process of extracting 187 or 171 new features per cluster from the data by feature extraction and selecting 20 features that explained the state of hydraulic components well by feature selection to select variables to use for the machine learning model.

3.1. Data Pre-Processing

The sensor data were measured 2205 times with a cycle of 60 s, and the number of vibrations measured for each repetition differed; hence, the data needed to be pre-processed. Therefore, to extract and select features that could be appropriately used for model learning through Figure 3, we divided the sensor data based on 13 clusters defined in advance based on PS1, which best describes a certain cycle. Cluster 1 corresponded to the entire cycle of FS1, and Clusters 2 to 13 were allocated in the divided order.

3.2. Feature Extraction

Feature extraction is a method to generate more suitable and useful features for model training through new combinations based on all features. It is important to use the full information we have, but if the appropriate information for model training exists locally, it can be used for model training, to result in a better model.
We extracted the slope of the linear fit, and the positions of maximum and minimum value for each cluster from the signal shape. Furthermore, the mean, median, Q1, Q3, variance, skewness, and kurtosis were extracted for each cluster from the distribution density characteristic. Additionally, since skewness and kurtosis were not calculated in Clusters 2, 4, 6, 8, 10, and 12 in which there was no gradient in the 1 Hz feature, in total 96 features were excluded, and 2335 new combinations of features were extracted.

3.2.1. Signal Shape

Based on the cluster defined previously in the data-preprocessing step, the slope parameter of the linear regression and the positions of the maximum and minimum values were used as features for each cluster. The slope parameter of the linear regression with the locations of the maximum and minimum values is one of the methods to extract only the necessary information from our entire information and is particularly useful when the relevant information is included only within a specific period. Especially with linear regression, it is computationally very reasonable and provides a first-order approximation. If we used an even higher-order approximation, we could have information close to the original information, but unnecessary information such as noise may have been included, so we used the gradient as a feature [22].

3.2.2. Distribution Density Characteristics

In past studies, time-domain features used statistics such as mean and variance to identify the difference between a general vibration signal and other vibration signals. Furthermore, skewness and kurtosis can be extracted through the probability density function of the vibration signal. Therefore, the sensor data of the hydraulic system are also composed of a time domain, and since it is data that collects vibration signals and repeats a certain period, distribution density characteristics (mean, median, variance, Q1, Q3, skewness, kurtosis) based on pre-classified clusters are extracted [13].

3.3. Feature Selection

Feature selection is a series of processes to improve model performance by reducing model computation cost and by reducing the number of features in developing a predictive model. The model is composed of the most relevant features among those recombined by feature extraction. Its advantage is that the number of dimensions can be managed and the model’s performance degradation can be prevented. The features most related to the state of the hydraulic system components described above were selected and used for model learning.

3.3.1. Spearman’s Rank Correlation Coefficient

Since the Spearman rank correlation coefficient describes the relationship between two variables through a monotonic function between them, it evaluates a monotonic relationship regardless of its linearity. It can be expressed the same as the Pearson correlation coefficient of the rank values of two variables. The range of coefficients can have a value between −1 and 1, and it has a value close to 1 when observations between two variables have similar ranks, and has a value close to −1 when the ranks differ. An absolute value is applied to the calculated correlation coefficient to select a high correlation coefficient regardless of positive or negative correlation.
Table 3 presents the Spearman rank correlation coefficients for the type of feature extraction for each sensor for each hydraulic component and the cluster location, and selects the top 20 in absolute value order regardless of negative or positive values. For example, in the case of “FS1_maxloc_4”, which has the highest correlation coefficient of valve condition, the absolute value of the Spearman rank correlation coefficient with the valve condition in the 5th cluster of the FS1 sensor is 0.946.

3.3.2. Pearson Correlation Coefficient

The Pearson correlation coefficient measures linear correlation between two variables, unlike the Spearman correlation coefficient. It can be expressed as the ratio between the covariance and the product of two variables. The range of the coefficient can have a value between −1 and 1, and the sign of the coefficient is determined according to the slope of the straight line. An absolute value is applied to the calculated correlation coefficient to select a high correlation coefficient regardless of positive or negative correlation.
Table 4 shows the Pearson correlation coefficient calculations for the type of feature extraction for each sensor for each hydraulic component, and the cluster location, and selects the top 20 in absolute value order regardless of negative or positive values. For example, in the case of “CE_Q1_13”, which has the highest correlation coefficient in the cooler condition, the absolute value of the Pearson correlation coefficient with the cooler condition in the 13th cluster of the CE sensor is 0.993.

3.3.3. Boruta Algorithm

The Boruta algorithm is a random forest-based variable selection method. Existing variables that do not affect the model creation are shadow variables created by restoring and extracting existing variables. They are judged to be of insufficient value and are removed. The algorithm can be used without hyper-parameter tuning, and variable importance can be converted into numerical values.
Duplicate features are created for all features, and shadow features are mixed well and combined with existing features to eliminate correlation with response variables. Based on this, a random forest is executed to calculate the z-score. Among the shadow features, the feature with the highest z-score is marked as an important variable with +1 hit. By repeating the above process as many times as possible, random forest analysis is performed, and important and insignificant variables can be tagged.
Figure 4 shows the features selected by calculating the importance of features extracted for each hydraulic element using the Boruta algorithm. Twenty variables were selected for each component, and 16 variables were selected during feature selection of valve condition among hydraulic components through the Boruta algorithm.

3.4. Model Training

After completing the variable selection, a model for anomaly and defect classification was trained based on 20 features related to hydraulic components. The valve condition with feature selection with the Boruta algorithm trained a model based on 15 selected features. The response variable was a stable flag indicating the state of the hydraulic system for a total of 2205 repeated cycles. Linear discriminant analysis (LDA), logistic regression, support vector classifier (SVC), decision tree, random forest, XGBoost, LightGBM, and Multilayer perceptron were used as the methodologies to learn the anomaly and defect classification model. Furthermore, the optimal hyper-parameter was obtained using grid search as cross validation, and the optimization model was determined by preventing overfitting. After determining the model, its performance was evaluated by measuring accuracy, precision, recall, and the true negative rate.

3.4.1. Linear Discriminant Analysis

LDA means that the discriminant function that categorizes an observation’s class is a linear function. In the existing dimension of the dataset, data are separated by projecting it to a smaller dimension that maximizes between-class variance and minimizes within-class variance.
Using each group mean μ c and the total mean μ of our feature matrix, we find the between-group variance B   and within-group variance W   , calculate the eigenvalues and eigenvectors of the transformation matrix W = W 1 B   , and organize them according to the largest eigenvalues. The k   th eigenvector from the first eigenvector means a new space V k of a smaller dimension. Finally, we project X i j into the new space V k .

3.4.2. Logistic Regression

In general, linear regression analysis is suitable for prediction but not for classification, because the response variable of linear regression analysis has no limit on the range of values. Therefore, the z value of the linear regression analysis z = w T x + b is substituted into the sigmoid function y = 1 1 + e z = 1 1 + e w T x + b to have a value between 0 and 1. The right side can be summarized by w T x + b and expressed in a linear state as shown in l o g y 1 y = w T x + b . The left side was made into a logit that took a log to the odds ratio. Logistic regression analysis can be completed by replacing y in logit with π ( x ) . The π ( x ) is P ( y = 1 | X = x ) . The resulting value of logistic regression analysis is the classification probability, and the class is determined based on the classification threshold.

3.4.3. Support Vector Classifier

The support vector classifier (SVC) determines the class based on the support vector. The SVC creates a straight line that separates each class. The SVC classifies the various divisible straight lines using a centerline and a support vector.
To draw a centerline, when the vector w is perpendicular to the centerline and the observation   x is the dot product, the point at which the dot product value c becomes the centerline is w T x = c . The data space can be divided into cases where the dot product is greater than c and less than c .
By transposing c to the left side and replacing c with b , the centerline can be expressed as in Equation w T x + b = 0 , which becomes a hyperplane that separates data, and Equation G ( x ) = s i g n [ w T x + b ] becomes an expression representing each region.
In the SVC, the margin, which is the width between support vectors, is maximized. The margin is maximized because the generalization error tends to be low. If the margin is small, the model may overfit.
m a x w , b M s u b j e c t   t o   { y i ( w T x i + b ) M ,         i = 1 , , n | | w | | = 1
A soft margin through slack variables relaxes the existing SVC’s criteria to allow some misclassified observations. The kernel changes the feature space and applies it to the SVC. When the curved space is straightened after applying the SVC to it, the observations are classified well and the data are classified as a nonlinear decision boundary through the kernel. Kernel types include the polynomial, radial bias function, and sigmoid.

3.4.4. Decision Tree

The decision tree can predict the class or value of the response variable by learning the decision rule from the training dataset. It starts at the root node to predict the class label using the decision tree. The internal node splits according to the properties of the root node. The leaf node is the node located at the bottom of the internal node. It contains the result of predicting the value of the class label or response variable and does not split any more. It has high enough interpretability to be interpreted without statistical knowledge, and there is no assumption about the type of data. It is not affected by multicollinearity and outliers.

3.4.5. Random Forest

Random forest is an ensemble technique composed of multiple decision trees. Individual trees in the random forest predict a class, and the model predicts the specific class chosen by most classifiers. After randomly extracting n observations from observations so that overlap is possible, the process of creating a decision tree is repeated by selecting p features so that overlap is not possible. The final model prediction is made by voting or averaging the results from each decision tree.
Although it is relatively slow due to the characteristics of the random forest, individual decision trees differ, and since not all features are considered, the performance degradation of the model can be prevented in high-dimensional data.

3.4.6. XGBoost

Boosting reduces the weight of correctly classified observations while learning, and assigns weights to difficult-to-classify observations to focus ultimately on the difficult-to-classify observations. The learner created in the previous step is used to change the weights of the dataset to be used in the next step.
Gradient boosting is a method to boost the use of boosting gradients, and it learns in the direction of reducing the residual obtained by comparing the predicted result obtained by including the feature into the learning model and the actual result. If the residual is large, a new model is added to improve the performance continuously by adding a new model that can reduce the difference between the existing predicted result and the actual result. XGBoost is a type of ensemble boosting technique, and has the advantage of fast model learning by parallel processing gradient boosting.

3.4.7. LightGBM

The existing tree-based algorithm uses a level-wise method that splits while maintaining a balanced tree as much as possible. Although the depth of the tree can be minimized, it has a characteristic that requires additional time to balance. LightGBM has a leaf-wise tree partitioning method that extends vertically. Without considering the balance of the tree, it continuously divides the leaf node with the maximum loss to create a deep and asymmetric tree. LightGBM has maximum data loss, so it can minimize the prediction error loss compared to the level-wise tree split method and it takes relatively less learning time and uses less memory compared to XGBoost.

3.4.8. Multilayer Perceptron

A multilayer perceptron consists of three or more node layers of input, hidden, and output layers. Except for the input layer, each node uses a nonlinear activation function. To train the MLP, the weights are changed using backpropagation, which reduces the error by reversing the error between the actual and predicted values. Multilayer and nonlinear functions can distinguish between linearly inseparable data. When all nodes are connected, there is a disadvantage that can lead to inefficiency by including too many parameters.

3.4.9. Cross Validation

The ratio of the steady state to the unstable state of the target variable of the data analyzed is (7449, 756), and the ratio of the unstable state is about 10% compared to that of the steady state. To prevent errors, the cross validation used in our study is stratified k-fold to prevent a specific target variable from being concentrated on a specific fold.
Furthermore, the grid search method is used to select suitable hyper-parameters of the model with the best performance by sequentially inputting predetermined values to find the optimized model hyper-parameter.

3.5. Classification Evaluation Metrics

The evaluation indicators used for model evaluation are accuracy, precision, recall, and true negative rate. Precision is an evaluation index to measure positive prediction performance accurately as the ratio of actual positive observations among observations with positive prediction results. The recall is the ratio of positively predicted actual positive observations that is also called sensitivity or true positive rate (TPR). It is useful to identify issues that may be sensitive when a negative is actually positive.
True negative rate (TNR) is a ratio that is correctly predicted to be negative among observations that are actually negative. Like the TPR, it is useful to identify problems that may be sensitive when a negative is actually positive. Since our study aims to gauge whether the hydraulic system is stable to prevent further damage, it is judged that the hydraulic system is in a stable state but unstable through recall (TPR) rather than accuracy, so we reduce the financial damage and reduce the actual unstable state through TNR. The goal is to reduce a state that is judged stable but could increase the damage.
Accuracy   = TP   +   TN TP + TN + FP + FN Precision   = TP TP + FP Recall   =   True   Positive   Rate   = TP TP + FN True   Negative   Rate   = TN TN + FP

4. Results

We selected the top 20 features that are highly related to each hydraulic system component and trained the model for the experiment. Each table and figure below present optimal hyper-parameters and learning curves per machine learning technique proposed for each hydraulic component being presented. In addition, the confusion matrix for feature selection for each hydraulic system component, the accuracy determined by the grid search evaluation criteria for the training set, and accuracy, precision, recall (TPR), and TNR confirmed by the testing set are presented. Table 5 shows optimal hyper-parameters obtained by executing feature selection-based machine learning using a grid search for each hydraulic component. The optimal combination was found in the search space of the given hyper-parameters. And Table 6 shows confusion matrix and classification evaluation metrics from each machine learning method configured based on the optimal hyper-parameters proposed in Table 5. Similarly, Table 7 and Table 8, Table 9 and Table 10, and Table 11 and Table 12 show results related to valve condition, internal pump leakage, and hydraulic accumulator, respectively.
Figure 5 presents a learning curve to determine whether overfitting or underfitting occurs while learning hyper-parameters for each hydraulic component. Overall, as the number of samples increases, a model with high variance becomes a model with a good bias-variance trade-off. Random forest or XGBoost ML methods tend to have a smaller variance as the number of samples increases. Therefore, if the number of samples can be increased, the variance can be made smaller, making it a model with a good bias-variance trade-off.
Table 5 shows the optimal hyper-parameters obtained by running machine learning based on feature selection using a grid search for cooler condition. All machine learning methodologies were constructed by feature selection through the Boruta algorithm in cooler condition. Table 6 shows the results of the confusion matrix and classification evaluation index for the cooler condition among hydraulic system components. The cooler condition showed the worst classification performance result compared to other components. Furthermore, the classification accuracy was as high as 0.885, indicating the worst classification accuracy among hydraulic components. In particular, the linear discriminant analysis showed the worst performance at 0.655.
The recall performance index also showed the worst classification result in the linear discriminant analysis. The feature selection with the best performance was Spearman’s rank correlation coefficient, which properly classified most of the actual stable states of 0.997, but the TNR was 0.099 in the Boruta algorithm. Additionally, in logistic regression, TPR showed the highest performance for all feature selections, but TNR was zero or 0.006, so it could not detect anomalies. Compared to the recall performance, it showed incomparably poor performance.
In the model with the best recall performance, the SVC using Spearman’s rank correlation coefficient classified the stable state perfectly, but the TNR was 0.093, showing a performance not significantly different from that of the linear discriminant analysis. The models with the best TNR performance were random forest, XGBoost, and LightGBM using the Boruta algorithm. Considering recall and the TNR performance (0.948, 0.762), (0.938, 0.768), (0.935, 0.788), the model that could best judge the cooler condition was selected.
Table 5. Hyper-parameter configuration space and optimal parameter under cooler condition.
Table 5. Hyper-parameter configuration space and optimal parameter under cooler condition.
Hydraulic ComponentML Model (with Feature Selection)Hyper-ParameterTypeSearch SpaceOptimal Parameter
Cooler ConditionLDA (Boruta)----
Logistic Reg.
(Boruta)
CDiscrete0.001, 0.01, 0.1, 1, 10, 100100
penaltyCategorical‘L1’, ‘L2’, ‘elasticnet’L2
SVC (Boruta)CDiscrete0.01, 0.1, 1, 10, 5050
kernelCategorical‘poly’, ‘rbf’, ‘sigmoid’rbf
Decision Tree
(Boruta)
max_featuresCategorical‘auto’, ‘sqrt’, ’log2’auto
ccp_alphaDiscrete0.1, 0.01, 0.0010.001
max_depthDiscrete5, 7, 95
criterionCategorical‘gini’, ‘entropy’entropy
Random Forest
(Boruta)
n_estimatorsDiscrete10, 100, 1000100
max_featuresCategorical‘sqrt’, ‘log2’sqrt
XGBoost
(Boruta)
gammaDiscrete0.5, 1, 21
learning_rateDiscrete0.01, 0.10.01
max_depthDiscrete3, 4, 5, 66
n_estimatorsDiscrete12, 24, 3612
LightGBM
(Boruta)
num_iterationsDiscrete100, 500, 1000500
max_depthDiscrete15, 20, 2515
learning_rateDiscrete0.01, 0.1, 0.30.1
Multilayer Perceptron
(Boruta)
hidden_layer_sizesTuple(50,50,50), (50,100,50), (100,)(50,100,50)
activationCategorical‘tanh’, ‘relu’relu
solverCategorical‘sgd’, ‘adam’adam
alphaDiscrete0.0001, 0.050.0001
learning_rateCategorical‘constant’, ‘adaptive’constant
Table 6. Classification evaluation metrics for cooler condition.
Table 6. Classification evaluation metrics for cooler condition.
Cooler ConditionSpearmanPearsonBoruta
PredictedPredictedPredicted
StableUnstableStableUnstableStableUnstable
LDATrueStable28912791127416
Unstable1501149213615
Train Accuracy0.6520.6550.642
Test Accuracy0.6580.6370.655
Precision0.6580.6520.668
Recall (TPR)0.9970.9620.945
True Negative Rate0.0070.0130.099
Logistic RegressionTrueStable290029002900
Unstable151015101447
Train Accuracy0.5670.6570.664
Test Accuracy0.6580.6580.673
Precision0.6580.6580.668
Recall (TPR)1.0001.0001.000
True Negative Rate0.0000.0000.046
SVCTrueStable2900287325337
Unstable137141104138113
Train Accuracy0.6840.7410.862
Test Accuracy0.6890.7440.830
Precision0.6790.7230.869
Recall (TPR)1.0000.9900.872
True Negative Rate0.0930.2720.748
Decision TreeTrueStable260302771327119
Unstable7972826934117
Train Accuracy0.7220.7660.889
Test Accuracy0.7530.7850.880
Precision0.7670.7720.889
Recall (TPR)0.8970.9550.934
True Negative Rate0.4770.4570.775
Random ForestTrueStable263272642627515
Unstable7576658636115
Train Accuracy0.7540.7730.909
Test Accuracy0.7690.7940.884
Precision0.7780.8020.884
Recall (TPR)0.9070.9100.948
True Negative Rate0.5030.5700.762
XGBoostTrueStable264262741627218
Unstable8467727935116
Train Accuracy0.7480.7820.899
Test Accuracy0.7510.8000.880
Precision0.7590.7920.886
Recall (TPR)0.9100.9450.938
True Negative Rate0.4440.5230.768
LightGBMTrueStable252382543627219
Unstable7279678432119
Train Accuracy0.7620.7740.896
Test Accuracy0.7510.7660.885
Precision0.7780.7910.895
Recall (TPR)0.8690.8760.935
True Negative Rate0.5230.5560.788
Multi-layer
Perceptron
TrueStable29002652527218
Unstable13714747742109
Train Accuracy0.6850.760.853
Test Accuracy0.6890.7760.864
Precision0.6790.7820.866
Recall (TPR)1.0000.9140.938
True Negative Rate0.0930.5100.722
Table 7 shows the optimal hyper-parameters obtained by running machine learning based on feature selection using a grid search for valve condition. Most of the machine learning methodologies consisted of feature selection through Pearson correlation coefficient in valve condition, and in particular, the LDA model was trained using Spearman’s rank correlation coefficient. Table 8 shows the results of the confusion matrix and classification evaluation index of valve condition among hydraulic system components. Among the models that can determine the state of the valve condition, the linear discriminant analysis had significantly improved performance compared to the result in the cooler condition, but its performance was not as good as that of random forest and XGBoost.
XGBoost showed the best classification accuracy with LDA 0.889, SVM 0.949, random forest 0.966, and XGBoost 0.967. In particular, the feature selection model with Boruta algorithm showed good performance in RF and XGBoost.
In the case of recall performance, LDA 0.86 and SVC 0.94 had the maximum recall performance, and random forest and XGBoost showed performance above 0.95. In particular, the feature selection model with the Boruta algorithm showed a performance of 0.97 or higher.
In the case of TNR performance, the feature selection model with the Boruta algorithm showed overall excellent performance. In particular, 0.98 in LDA and 0.92 in random forest and XGBoost showed excellent performance in LDA.
TPR and TNR of the random forest using Spearman’s rank correlation coefficient and Boruta algorithm for feature selection within the valve condition showed the best performance with (0.972, 0.927) and (0.976, 0.927), respectively. XGBoost using the Boruta algorithm (0.969, 0.921) showed similar performance to the random forest.
Table 7. Hyper-parameter configuration space and optimal parameter under valve condition.
Table 7. Hyper-parameter configuration space and optimal parameter under valve condition.
Hydraulic ComponentML Model (with Feature Selection)Hyper-ParameterTypeSearch SpaceOptimal Parameter
Valve ConditionLDA (Spearman)----
Logistic Reg.
(Pearson)
CDiscrete0.001, 0.01, 0.1, 1, 10, 100100
penaltyCategorical‘L1’, ‘L2’, ‘elasticnet’L2
SVC (Pearson)CDiscrete0.01, 0.1, 1, 10, 5050
kernelCategorical‘poly’, ‘rbf’, ‘sigmoid’rbf
Decision Tree
(Pearson)
max_featuresCategorical‘auto’, ‘sqrt’, ’log2’auto
ccp_alphaDiscrete0.1, 0.01, 0.0010.001
max_depthDiscrete5, 7, 97
criterionCategorical‘gini’, ‘entropy’entropy
Random Forest
(Boruta)
n_estimatorsDiscrete10, 100, 1000100
max_featuresCategorical‘sqrt’, ‘log2’sqrt
XGBoost
(Pearson)
gammaDiscrete0.5, 1, 20.5
learning_rateDiscrete0.01, 0.10.1
max_depthDiscrete3, 4, 5, 66
n_estimatorsDiscrete12, 24, 3636
LightGBM
(Boruta)
num_iterationsDiscrete100, 500, 10001000
max_depthDiscrete15, 20, 2515
learning_rateDiscrete0.01, 0.1, 0.30.01
Multilayer Perceptron
(Pearson)
hidden_layer_sizesTuple(50,50,50), (50,100,50), (100,)(50,100,50)
activationCategorical‘tanh’, ‘relu’tanh
solverCategorical‘sgd’, ‘adam’adam
alphaDiscrete0.0001, 0.050.0001
learning_rateCategorical‘constant’, ‘adaptive’constant
Table 8. Classification evaluation metrics for valve condition.
Table 8. Classification evaluation metrics for valve condition.
Valve ConditionSpearmanPearsonBoruta
PredictedPredictedPredicted
StableUnstableStableUnstableStableUnstable
LDATrueStable247432504023159
Unstable6145111403148
Train Accuracy0.8990.9130.858
Test Accuracy0.8890.8840.859
Precision0.9760.9580.987
Recall (TPR)0.8520.8620.797
True Negative Rate0.9600.9270.980
Logistic RegressionTrueStable260302603025238
Unstable10141101416145
Train Accuracy0.9160.9260.899
Test Accuracy0.9090.9090.900
Precision0.9630.9630.977
Recall (TPR)0.8970.8970.869
True Negative Rate0.9340.9340.960
SVCTrueStable267232721826723
Unstable1313881439142
Train Accuracy0.9350.9490.935
Test Accuracy0.9180.9410.927
Precision0.9540.9710.967
Recall (TPR)0.9210.9380.921
True Negative Rate0.9140.9470.940
Decision TreeTrueStable274162791127119
Unstable181331313834117
Train Accuracy0.9460.9550.951
Test Accuracy0.9230.9460.880
Precision0.9380.9550.889
Recall (TPR)0.9450.9620.934
True Negative Rate0.8810.9140.775
Random ForestTrueStable282828492837
Unstable111401213911140
Train Accuracy0.9600.9660.966
Test Accuracy0.9570.9590.959
Precision0.9620.9590.963
Recall (TPR)0.9720.9790.976
True Negative Rate0.9270.9210.927
XGBoostTrueStable2781228282819
Unstable14137814312139
Train Accuracy0.9630.9610.967
Test Accuracy0.9410.9640.952
Precision0.9520.9720.959
Recall (TPR)0.9590.9720.969
True Negative Rate0.9070.9470.921
LightGBMTrueStable2801028282828
Unstable111401114011140
Train Accuracy0.9630.9670.968
Test Accuracy0.9520.9570.957
Precision0.9620.9620.962
Recall (TPR)0.9660.9720.972
True Negative Rate0.9270.9270.927
Multi-layer PerceptronTrueStable263272791125238
Unstable11140131388143
Train Accuracy0.9290.9480.931
Test Accuracy0.9140.9460.896
Precision0.9600.9550.969
Recall (TPR)0.9070.9620.869
True Negative Rate0.9270.9140.947
Table 9 shows the optimal hyper-parameters obtained by running machine learning based on feature selection using a grid search for internal pump leakage. Unlike logistic regression using the Boruta algorithm in the case of internal pump leakage, all other machine learning methodologies were learned based on the Pearson correlation coefficient. Table 10 shows the confusion matrix and classification evaluation index of internal pump leakage. In the case of classification accuracy, the random forest with feature selection with Pearson correlation coefficient showed the best performance with 0.961. Regarding LDA feature selection with Spearman’s rank correlation coefficient, 0.832 showed the worst performance.
In the case of recall, the model with the highest classification accuracy also showed the highest recall performance of 0.986. In particular, random forest and XGBoost showed high recall performance, but random forest slightly outperformed them. Moreover, with TNR, as with recall, the random forest showed 0.901 performance overall. Logistic regression showed the highest TNR performance with 0.960, but TPR was 0.869, which was inferior to random forest. Therefore, in the third hydraulic system component, the random forest using the Pearson correlation coefficient for internal pump leakage showed the best performance of (0.986, 0.901) for each TPR and TNR. The XGBoost and logistic regression feature selected with the Boruta algorithm showed the best performance with (0.983, 0.894) and (0.869, 0.960). The performance fell short of that of the random forest.
Table 9. Hyper-parameter configuration space and optimal parameter under internal pump leakage.
Table 9. Hyper-parameter configuration space and optimal parameter under internal pump leakage.
Hydraulic ComponentML Model (with Feature Selection)Hyper-ParameterTypeSearch SpaceOptimal Parameter
Internal Pump LeakageLDA (Pearson)----
Logistic Reg.
(Boruta)
CDiscrete0.001, 0.01, 0.1, 1, 10, 100100
penaltyCategorical‘L1’, ‘L2’, ‘elasticnet’L2
SVC (Pearson)CDiscrete0.01, 0.1, 1, 10, 5050
kernelCategorical‘poly’, ‘rbf’, ‘sigmoid’rbf
Decision Tree
(Pearson)
max_featuresCategorical‘auto’, ‘sqrt’, ’log2’auto
ccp_alphaDiscrete0.1, 0.01, 0.0010.001
max_depthDiscrete5, 7, 99
criterionCategorical‘gini’, ‘entropy’gini
Random Forest
(Pearson)
n_estimatorsDiscrete10, 100, 1000100
max_featuresCategorical‘sqrt’, ‘log2’sqrt
XGBoost
(Pearson)
gammaDiscrete0.5, 1, 22
learning_rateDiscrete0.01, 0.10.1
max_depthDiscrete3, 4, 5, 66
n_estimatorsDiscrete12, 24, 3636
LightGBM
(Pearson)
num_iterationsDiscrete100, 500, 1000500
max_depthDiscrete15, 20, 252520
learning_rateDiscrete0.01, 0.1, 0.30.3
Multilayer Perceptron
(Pearson)
hidden_layer_sizesTuple(50,50,50), (50,100,50), (100,)(50,100,50)
activationCategorical‘tanh’, ‘relu’relu
solverCategorical‘sgd’, ‘adam’adam
alphaDiscrete0.0001, 0.050.0001
learning_rateCategorical‘constant’, ‘adaptive’constant
Table 10. Classification evaluation metrics for internal pump leakage.
Table 10. Classification evaluation metrics for internal pump leakage.
Internal Pump LeakageSpearmanPearsonBoruta
PredictedPredictedPredicted
StableUnstableStableUnstableStableUnstable
LDATrueStable238522504023753
Unstable221291513640111
Train Accuracy0.8080.8520.785
Test Accuracy0.8320.8750.789
Precision0.9150.9430.856
Recall (TPR)0.8210.8620.817
True Negative Rate0.8540.9010.735
Logistic RegressionTrueStable241492504025238
Unstable31120241276145
Train Accuracy0.8130.8550.825
Test Accuracy0.8190.8550.900
Precision0.8860.9120.977
Recall (TPR)0.8310.8620.869
True Negative Rate0.7950.8410.960
SVCTrueStable246442751526228
Unstable111401513625126
Train Accuracy0.8490.9220.870
Test Accuracy0.8750.9320.880
Precision0.9570.9480.913
Recall (TPR)0.8480.9480.903
True Negative Rate0.9270.9010.834
Decision TreeTrueStable283728282837
Unstable251262412736115
Train Accuracy0.9190.9310.91
Test Accuracy0.9270.9270.902
Precision0.9190.9220.887
Recall (TPR)0.9760.9720.976
True Negative Rate0.8340.8410.762
Random ForestTrueStable2828282427515
Unstable191321313820131
Train Accuracy0.9440.9550.914
Test Accuracy0.9390.9610.921
Precision0.9370.9570.932
Recall (TPR)0.9720.9860.948
True Negative Rate0.8740.9010.868
XGBoostTrueStable28010285527416
Unstable201311613523128
Train Accuracy0.9380.9520.914
Test Accuracy0.9320.9520.912
Precision0.9330.9470.923
Recall (TPR)0.9660.9830.945
True Negative Rate0.8680.8940.848
LightGBMTrueStable281928372828
Unstable121391313819132
Train Accuracy0.9470.9640.943
Test Accuracy0.9520.9550.939
Precision0.9590.9560.937
Recall (TPR)0.9690.9760.972
True Negative Rate0.9210.9140.874
Multi-layer PerceptronTrueStable265252652527218
Unstable261251313825126
Train Accuracy0.8250.9070.9
Test Accuracy0.8840.9140.902
Precision0.9110.9530.916
Recall (TPR)0.9140.9140.938
True Negative Rate0.8280.9140.834
Table 11 shows the optimal hyper-parameters obtained by running machine learning based on feature selection using a grid search for hydraulic accumulator. In the hydraulic accumulator, various feature selections were all used to train the machine learning model. Finally, Table 12 shows the confusion matrix and classification evaluation index of the hydraulic accumulator. For classification accuracy, random forest showed the best performance with LDA 0.878, logistic regression 0.880, SVC 0.930, decision tree 0.909, RF 0.959, XGBoost 0.948, LightGBM 0.980, and MLP 0.923. When all feature selection methods were compared, Spearman’s rank, Pearson correlation coefficient, and Boruta algorithm were 0.980, 0.964, and 0.968, respectively, showing the best performance among classification models by the feature selection method. Particularly, for all of them test accuracy was observed in LightGBM.
Recall showed the best performance of 0.99 or higher in all feature selection methods of random forest. Moreover, for TNR, the SVC obtained by feature selection with the Pearson correlation coefficient was 0.894, showing the best performance. In the random forest, the feature selection model with Boruta algorithm showed slightly lower performance with 0.887. Therefore, random forest and XGBoost using Boruta algorithm for feature selection indicated performances of (0.997, 0.887) and (0.990, 0.868), respectively. In particular, the SVC and LightGBM of the Pearson correlation coefficient, which had the highest TNR performance, was (0.931, 0.894) and (0.986, 0.921), indicating satisfactory performance.
Table 11. Hyper-parameter configuration space and optimal parameter under hydraulic accumulator.
Table 11. Hyper-parameter configuration space and optimal parameter under hydraulic accumulator.
Hydraulic ComponentML Model (with Feature Selection)Hyper-ParameterTypeSearch SpaceOptimal Parameter
Hydraulic AccumulatorLDA (Pearson)----
Logistic Reg.
(Spearman)
CDiscrete0.001, 0.01, 0.1, 1, 10, 100100
penaltyCategorical‘L1’, ‘L2’, ‘elasticnet’L2
SVC (Pearson)CDiscrete0.01, 0.1, 1, 10, 5050
kernelCategorical‘poly’, ‘rbf’, ‘sigmoid’poly
Decision Tree
(Pearson)
max_featuresCategorical‘auto’, ‘sqrt’, ’log2’auto
ccp_alphaDiscrete0.1, 0.01, 0.0010.001
max_depthDiscrete5, 7, 99
criterionCategorical‘gini’, ‘entropy’gini
Random Forest
(Boruta)
n_estimatorsDiscrete10, 100, 10001000
max_featuresCategorical‘sqrt’, ‘log2’sqrt
XGBoost
(Boruta)
gammaDiscrete0.5, 1, 22
learning_rateDiscrete0.01, 0.10.1
max_depthDiscrete3, 4, 5, 66
n_estimatorsDiscrete12, 24, 3636
LightGBM(Spearman)num_iterationsDiscrete100, 500, 1000500
max_depthDiscrete15, 20, 2520
learning_rateDiscrete0.01, 0.1, 0.30.3
Multilayer Perceptron
(Pearson)
hidden_layer_sizesTuple(50,50,50), (50,100,50), (100,)(50,100,50)
activationCategorical‘tanh’, ‘relu’relu
solverCategorical‘sgd’, ‘adam’adam
alphaDiscrete0.0001, 0.050.05
learning_rateCategorical‘constant’, ‘adaptive’constant
Table 12. Classification evaluation metrics for hydraulic accumulator.
Table 12. Classification evaluation metrics for hydraulic accumulator.
Hydraulic AccumulatorSpearmanPearsonBoruta
PredictedPredictedPredicted
StableUnstableStableUnstableStableUnstable
LDATrueStable264262632724941
Unstable291222712439112
Train Accuracy0.8790.8700.816
Test Accuracy0.8750.8780.819
Precision0.9010.9070.865
Recall (TPR)0.9100.9070.859
True Negative Rate0.8080.8210.742
Logistic RegressionTrueStable266242662425634
Unstable291223112034117
Train Accuracy0.8970.880.826
Test Accuracy0.8800.8750.846
Precision0.9020.8960.883
Recall (TPR)0.9170.9170.883
True Negative Rate0.8080.7950.775
SVCTrueStable280102702028010
Unstable211301613526125
Train Accuracy0.9290.9300.933
Test Accuracy0.9300.9180.918
Precision0.9300.9440.915
Recall (TPR)0.9660.9310.966
True Negative Rate0.8610.8940.828
Decision TreeTrueStable27812275152837
Unstable321192512636115
Train Accuracy0.9080.9090.94
Test Accuracy0.9000.9090.902
Precision0.8970.9170.887
Recall (TPR)0.9590.9480.976
True Negative Rate0.7880.8340.762
Random ForestTrueStable289128732891
Unstable251261913217134
Train Accuracy0.9560.9510.956
Test Accuracy0.9410.9500.959
Precision0.9200.9380.944
Recall (TPR)0.9970.9900.997
True Negative Rate0.8340.8740.887
XGBoostTrueStable289128552873
Unstable261252412720131
Train Accuracy0.9520.9370.956
Test Accuracy0.9390.9340.948
Precision0.9170.9220.935
Recall (TPR)0.9970.9830.990
True Negative Rate0.8280.8410.868
LightGBMTrueStable290028642900
Unstable91421213914137
Train Accuracy0.9690.9680.972
Test Accuracy0.9800.9640.968
Precision0.9700.9600.954
Recall (TPR)1.0000.9861.000
True Negative Rate0.9400.9210.907
Multi-layer PerceptronTrueStable28282791126525
Unstable311202312823128
Train Accuracy0.9120.9210.922
Test Accuracy0.9120.9230.891
Precision0.9010.9240.920
Recall (TPR)0.9720.9620.914
True Negative Rate0.7950.8480.848

5. Discussion

In this study, the model’s performance was judged through classification accuracy by hydraulic component and the TPR and TNR. Regarding classification accuracy, the classification performance of cooler, valve, pump, and accumulator in related studies had performances of 1, 1, 0.80, and 0.65, respectively [18]; however, the results of this study showed that the maximum classification performances of 0.88, 0.96, 0.96, and 0.96 were obtained. It had high classification accuracy in the case of internal pump leakage and hydraulic accumulator. So, in this paper, it had accurate performance, but in the cooler condition, classification accuracy was lower than that of related studies, so it needs to be supplemented in future studies.
In the case of TPR and TNR, the TPR in the related study was 0.97 [15], and the results of this study showed similar performance, with more than 0.97. In particular, this study shows the identification performance of stable and unstable states by additionally calculating a TNR of 0.78 in the cooler condition and a TNR of 0.87 or more in the remaining components.
Furthermore, after feature extraction, classification was performed based on the top 20 features selected by feature selection, but it was not known which features contributed more and played a decisive role in prediction. In future studies, it is expected that explainable artificial intelligence (XAI) will be used to measure the importance of a feature, and it will be used to determine anomalies and problems by combining them with industry domain information based on the importance of features used when anomalies or problems occur in actual industrial sites. It will also be helpful in organizing target variables and metadata of the industrial internet of things (IoT) data collected for the first time.
However, the proposed approach in this paper can be a guide to the occurrence of abnormal conditions for each hydraulic element by using machine learning techniques. Even so, a disadvantage is that the analysis to help the decision-maker according to the engineer’s report is insufficient. Therefore, future research using explainable AI is expected so that engineers can understand the results created by machine learning techniques. Additionally, although the time domain exists, since it is experimental data repeatedly measured according to a certain time, a time series analysis approach cannot be attempted, which is a limitation. Therefore, it is expected that anomaly detection will be the subject of future research by approaching time-series analysis based on repeated measurement experimental data.

6. Conclusions

This study deals with the identification of stable and unstable conditions that may occur in hydraulic system components. After classifying data based on a predetermined cluster, features recombined with the shape and distribution density characteristics of observations were extracted. In total 2335 features were used for model learning by selecting variables with high probability due to a high correlation coefficient or high hit with hydraulic system components using a correlation coefficient and Boruta algorithm. Based on the selected features using each feature selection method, the cooler condition, valve condition, internal pump leakage, and stable/unstable condition of the hydraulic accumulator were classified.
The recall rate of all hydraulic system components was 0.94 or higher, and satisfactory results were obtained, but the TNR showed 0.77 in cooler condition. Except for the cooler condition, all components showed a TNR of 0.85 or more, and in particular, the TNR of the valve condition was 0.92, which is very high.
The objective and contributions made through this study are as follows.
  • When learning machine learning or deep learning models, it aims to create a new combination of features with feature extraction and to propose better performance models and various classification evaluation scales by reducing the calculation and feature dimensions required for learning the model through feature selection.
  • Suggesting TNR to reduce the issue of unexpected system shutdown except for in case of the unstable state of each hydraulic component.
  • Justification for the engineer’s judgment by proposing the basis for determining the unstable condition of each hydraulic component.
In addition, in future studies, more specific and interpretive grounds for judgment of engineers and decision makers will be specified using XAI.
TPR and TNR, which were proposed as a result of this paper, can be utilized as a basis for engineers to determine anomalous occurrences and as a factor for decision makers to make situational judgments. Adding this study’s results to the domain used to judge abnormalities and problems in the industrial field can provide a more reliable basis for analysis, and it is expected that the results of XAI in future studies will be added to make a more precise value judgment.

Author Contributions

Formal analysis, D.K.; funding acquisition, T.-Y.H.; supervision, T.-Y.H.; writing—original draft, D.K.; writing—review and editing, T.-Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2017R1D1A3B03028084, 2019R1I1A3A01057696).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This data can be found here: [https://archive-beta.ics.uci.edu/ml/datasets/condition+monitoring+of+hydraulic+systems] (accessed on 5 February 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hoover, J.B.; Bailey, J.L.; Willauer, H.D.; Williams, F.W. Evaluation of Submarine Hydraulic System Explosion and Fire Hazards; NRL Memorandum Report 6180-05-8908; U.S. Naval Research Laboratory: Washington, DC, USA, 2005.
  2. Zhang, R.; Xiang, Y.; Hu, Y.; Zang, H.; Shu, W. Active control of hydraulic oil contamination to extend the service life of aviation hydraulic system. Int. J. Adv. Manuf. Technol. 2018, 96, 1693–1704. [Google Scholar] [CrossRef]
  3. Shen, W.; Huang, H.; Pang, Y.; Su, X. Review of the Energy Saving Hydraulic System Based on Common Pressure Rail. IEEE Access 2017, 5, 655–669. [Google Scholar] [CrossRef]
  4. Ning, C.; Zhang, X. Study on Vibration and Noise for the Hydraulic System of Hydraulic hoist. In Proceedings of the 1st International Conference on Mechanical Engineering and Material Science, Shanghai, China, 28–30 December 2012; pp. 358–360. [Google Scholar] [CrossRef] [Green Version]
  5. Hodge, V.J.; O’Keefe, S.; Weeks, M.; Moulds, A. Wireless Sensor Networks for Condition Monitoring in the Railway Industry: A Survey. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1088–1106. [Google Scholar] [CrossRef]
  6. Márquez, F.; Tobias, A.; Pérez, J.; Papaelias, M. Condition monitoring of wind turbines: Techniques and methods. Renew. Energy 2012, 46, 169–178. [Google Scholar] [CrossRef]
  7. Dai, J.; Tang, J.; Huang, S.; Wang, Y. Signal-Based Intelligent Hydraulic Fault Diagnosis Methods: Review and Prospects. Chin. J. Mech. Eng. 2019, 32, 75. [Google Scholar] [CrossRef] [Green Version]
  8. Juan, J.S.-D.; Miguel, D.-P.; Roque, A.O.-R.; Rene, d.J.R.-T. Multifault Diagnosis Method Applied to an Electric Machine Based on High-Dimensional Feature Reduction. IEEE Trans. Ind. Appl. 2017, 53, 3086–3097. [Google Scholar] [CrossRef] [Green Version]
  9. Pacella, M.; Papadia, G. Fault Diagnosis by Multisensor Data: A Data-Driven Approach Based on Spectral Clustering and Pairwise Constraints. Sensors 2020, 20, 7065. [Google Scholar] [CrossRef] [PubMed]
  10. Helwig, N.; Pignanelli, E.; Schütze, A. Condition monitoring of a complex hydraulic system using multivariate statistics. In Proceedings of the IEEE International Instrumentation and Measurement Technology Conference (I2MTC) Proceedings, Pisa, Italy, 11–14 May 2015; pp. 210–215. [Google Scholar] [CrossRef]
  11. Helwig, N.; Pingnalli, E.; Schütze, A. Detecting and Compensating Sensor Faults in a Hydraulic Condition Monitoring System. In Proceedings of the AMA Conferences, Nürnberg, Germany, 19–21 May 2015. [Google Scholar] [CrossRef]
  12. Alenany, A.; Helmi, A.M.; Nasef, B.M. Comprehensive Analysis for Sensor-Based Hydraulic System Condition Monitoring. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 133–140. [Google Scholar] [CrossRef]
  13. Caesarendra, W.; Tjahjowidodo, T. A Review of Feature Extraction Methods in Vibration-Based Condition Monitoring and Its Application for Degradation Trend Estimation of Low-Speed Slew Bearing. Machines 2017, 5, 21. [Google Scholar] [CrossRef]
  14. Casoli, P.; Pastori, M.; Scolari, F.; Rundo, M. A Vibration Signal-Based Method for Fault Identification and Classification in Hydraulic Axial Piston Pumps. Energies 2019, 12, 953. [Google Scholar] [CrossRef] [Green Version]
  15. Lei, Y.; Jiang, W.; Jiang, A.; Zhu, Y.; Niu, H.; Zhang, S. Fault Diagnosis Method for Hydraulic Directional Valves Integrating PCA and XGBoost. Processes 2019, 7, 589. [Google Scholar] [CrossRef] [Green Version]
  16. König, C.; Helmi, A.M. Sensitivity Analysis of Sensors in a Hydraulic Condition Monitoring System Using CNN Models. Sensors 2020, 20, 3307. [Google Scholar] [CrossRef] [PubMed]
  17. Joo, Y.; Kim, K.; Jeong, J. Performance Comparison of Machine Learning Algorithms for Imbalanced Class Classification in Hydraulic System. In Proceedings of the 14th International Conference on Ubiquitous Information Management and Communication (IMCOM), Taichung, Taiwan, 3–5 January 2020; pp. 1–8. [Google Scholar] [CrossRef]
  18. Jose, J.T.; Das, J.; Mishra, S.K.; Wrat, G. Early detection and classification of internal leakage in boom actuator of mobile hydraulic machines using SVM. Eng. Appl. Artif. Intell. 2021, 106, 104492. [Google Scholar] [CrossRef]
  19. Moleda, M.; Momot, A.; Mrozek, D. Predictive Maintenance of Boiler Feed Water Pumps Using SCADA Data. Sensors 2020, 20, 571. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Jing, L.; Zidan, X. A simulation investigation of lubricating characteristics for a cylindrical roller bearing of a high-power gearbox. Tribol. Int. 2022, 167, 107373. [Google Scholar] [CrossRef]
  21. Chawathe, S.S. Condition Monitoring of Hydraulic Systems by Classifying Sensor Data Streams. In Proceedings of the IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; pp. 898–904. [Google Scholar] [CrossRef]
  22. Schneider, T.; Helwig, N.; Schütze, A. Automatic feature extraction and selection for classification of cyclical time series data. Tech. Mess. 2017, 84, 198–206. [Google Scholar] [CrossRef]
Figure 1. Concise hydraulic system circuit and installed sensors.
Figure 1. Concise hydraulic system circuit and installed sensors.
Sensors 22 02479 g001
Figure 2. Data pre-processing, feature extraction, and feature selection for machine learning.
Figure 2. Data pre-processing, feature extraction, and feature selection for machine learning.
Sensors 22 02479 g002
Figure 3. Pre-defined Clusters 1 to 13 for feature extraction.
Figure 3. Pre-defined Clusters 1 to 13 for feature extraction.
Sensors 22 02479 g003
Figure 4. Feature selection through the Boruta algorithm.
Figure 4. Feature selection through the Boruta algorithm.
Sensors 22 02479 g004
Figure 5. Learning curve of hydraulic component in machine learning.
Figure 5. Learning curve of hydraulic component in machine learning.
Sensors 22 02479 g005
Table 1. Condition hydraulic components and stable flag expression in the hydraulic system.
Table 1. Condition hydraulic components and stable flag expression in the hydraulic system.
Hydraulic ComponentsAnnotationsCondition ValuesExamplesStable Flag
(with Examples)
Cooler Condition Close to Total Failure3%7320: Conditions were stable (7449).
1: Static conditions might not have been reached yet (756).
Reduced Efficiency20%732
Full Efficiency100%741
Valve ConditionOptimal Switching Behavior100%1125
Small Lag90%360
Severe Lag80%360
Close to Total Failure73%360
Internal Pump LeakageNo Leakage01221
Weak Leakage1492
Severe Leakage2492
Hydraulic AccumulatorOptimal Pressure130 bar599
Slightly Reduced Pressure115 bar399
Severely Reduced Pressure100 bar399
Close to Total Failure90 bar808
Table 2. Sensor data measured each minute in the hydraulic system.
Table 2. Sensor data measured each minute in the hydraulic system.
CategorySensorUnitSampling Rate (Hz)
PressurePS1Bar100 Hz
PS2
PS3
PS4
PS5
PS6
Motor PowerEPS1W
Volume FlowFS1L/min10 Hz
FS2
TemperatureTS1°C1 Hz
TS2
TS3
TS4
VibrationVS1mm/s
Cooling EfficiencyCE%
Cooling PowerCPkW
Table 3. Feature selection through Spearman’s rank correlation coefficient.
Table 3. Feature selection through Spearman’s rank correlation coefficient.
Cooler ConditionValve ConditionInternal Pump LeakageHydraulic Accumulator
Sensor|r|Sensor|r|Sensor|r|Sensor|r|
CP_min_10.942FS1_maxloc_40.946SE_min_50.893EPS1_var_20.788
CP_min_100.942PS2_ma_loc_40.944SE_Q1_50.893EPS1_slope_20.774
CP_min_110.942PS3_maxloc_40.930SE_mean_50.893FS1_slope_80.759
CP_Q1_100.942PS3_maxloc_10.930SE_median_50.892PS3_Q3_40.755
CP_median_100.942PS2_slope_40.926SE_min_130.892FS1_var_100.724
CP_mean_100.942PS2_skew_40.926SE_Q1_130.892PS1_var_20.697
CP_Q1_10.942PS2_kurt_40.926SE_Q3_10.892PS3_var_40.678
CP_Q3_100.942PS1_kurt_40.926SE_mean_130.892EPS1_var_30.661
CP_max_110.942PS2_var_40.926SE_median_130.891PS1_var_50.656
CP_max_100.942PS1_skew_40.926SE_Q3_50.891FS1_skew_60.649
CP_mean_110.942EPS1_skew_40.926SE_Q3_130.891FS1_kurt_60.640
CP_Q1_110.942PS2_mean_40.906FS1_Q1_130.888PS1_Q3_20.630
CP_min_130.942FS1_kurt_40.896SE_max_130.888PS3_maxloc_20.615
CP_median_110.942PS1_Q3_40.882SE_max_50.887EPS1_slope_80.607
CP_max_120.942PS2_maxloc_10.870FS1_min_130.880FS1_slope_100.605
CP_Q3_110.942PS3_slope_40.865FS1_mean_130.876PS2_var_50.602
CP_Q3_120.942FS1_var_40.860FS1_median_130.875FS1_var_80.601
CP_min_90.942PS3_Q1_40.849SE_var_20.873EPS1_var_80.600
CP_median_120.942FS1_maxloc_10.839SE_mean_20.873PS1_skew_20.586
CP_mean_120.942PS1_slope_40.828SE_Q3_20.873EPS1_skew_30.580
Table 4. Feature selection through Pearson correlation coefficient.
Table 4. Feature selection through Pearson correlation coefficient.
Cooler ConditionValve ConditionInternal Pump LeakageHydraulic Accumulator
Sensor|ρ|Sensor|ρ|Sensor|ρ|Sensor|ρ|
CE_Q1_130.993PS2_kurt_40.999SE_min_50.951EPS1_var_20.689
CE_min_130.993PS2_skew_40.993SE_Q1_50.950PS3_var_40.655
CE_mean_130.993PS2_slope_40.991SE_median_50.947PS3_maxloc_20.626
CE_median_130.993PS2_maxloc_40.982SE_mean_50.946PS3_max_30.583
CE_Q3_130.993PS3_maxloc_40.977SE_Q3_50.940PS3_Q3_40.582
CE_min_120.993PS1_skew_40.976SE_max_50.920EPS1_skew_30.568
CE_Q1_120.993PS1_kurt_40.976FS1_Q1_50.910PS1_Q3_20.556
CE_mean_120.993EPS1_skew_40.975FS1_min_50.896EPS1_slope_20.536
CE_median_120.993PS2_var_40.974FS1_Q3_60.895PS1_median_20.517
CE_Q3_120.993FS1_maxloc_40.957FS1_mean_50.863FS1_min_60.515
CE_max_120.993PS2_mean_40.948FS1_median_50.863PS3_slope_30.513
CE_max_130.993PS3_Q1_40.940FS1_Q3_50.799SE_maxloc_40.505
CE_min_110.993PS2_Q1_40.929FS1_max_60.763FS1_slope_20.503
CE_Q1_110.993PS3_slope_40.916SE_max_40.678FS1_max_20.502
CE_mean_110.993PS1_Q3_40.915FS1_max_50.665FS1_kurt_120.499
CE_median_110.993FS1_slope_40.903SE_var_20.570FS1_mean_20.493
CE_Q3_110.993FS1_kurt_40.899SE_Q3_40.520FS1_min_80.491
CE_max_110.993FS1_var_40.869SE_mean_10.465FS1_var_20.482
CE_Q1_100.992FS1_Q1_40.863FS1_mean_60.438PS3_var_30.472
CE_median_100.992SE_Q1_40.848SE_Q3_10.436FS1_max_40.470
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, D.; Heo, T.-Y. Anomaly Detection with Feature Extraction Based on Machine Learning Using Hydraulic System IoT Sensor Data. Sensors 2022, 22, 2479. https://doi.org/10.3390/s22072479

AMA Style

Kim D, Heo T-Y. Anomaly Detection with Feature Extraction Based on Machine Learning Using Hydraulic System IoT Sensor Data. Sensors. 2022; 22(7):2479. https://doi.org/10.3390/s22072479

Chicago/Turabian Style

Kim, Doyun, and Tae-Young Heo. 2022. "Anomaly Detection with Feature Extraction Based on Machine Learning Using Hydraulic System IoT Sensor Data" Sensors 22, no. 7: 2479. https://doi.org/10.3390/s22072479

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop