Next Article in Journal
Rapid Estimation of Soil Pb Concentration Based on Spectral Feature Screening and Multi-Strategy Spectral Fusion
Next Article in Special Issue
Progress in Active Infrared Imaging for Defect Detection in the Renewable and Electronic Industries
Previous Article in Journal
Advanced Vibration-Based Fault Diagnosis and Vibration Control Methods
Previous Article in Special Issue
Research on Diesel Engine Fault Status Identification Method Based on Synchro Squeezing S-Transform and Vision Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Preventing Forklift Front-End Failures: Predicting the Weight Centers of Heavy Objects, Remaining Useful Life Prediction under Abnormal Conditions, and Failure Diagnosis Based on Alarm Rules

1
Department of Smart Digital Engineering, INHA University, Incheon 22212, Republic of Korea
2
Doosan Industrial Vehicle, Incheon 22503, Republic of Korea
3
Department of Naval Architecture and Ocean Engineering, INHA University, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(18), 7706; https://doi.org/10.3390/s23187706
Submission received: 15 August 2023 / Revised: 1 September 2023 / Accepted: 3 September 2023 / Published: 6 September 2023

Abstract

:
This paper addresses the critical challenge of preventing front-end failures in forklifts by addressing the center of gravity, accurate prediction of the remaining useful life (RUL), and efficient fault diagnosis through alarm rules. The study’s significance lies in offering a comprehensive approach to enhancing forklift operational reliability. To achieve this goal, acceleration signals from the forklift’s front-end were collected and processed. Time-domain statistical features were extracted from one-second windows, subsequently refined through an exponentially weighted moving average to mitigate noise. Data augmentation techniques, including AWGN and LSTM autoencoders, were employed. Based on the augmented data, random forest and lightGBM models were used to develop classification models for the weight centers of heavy objects carried by a forklift. Additionally, contextual diagnosis was performed by applying exponentially weighted moving averages to the classification probabilities of the machine learning models. The results indicated that the random forest achieved an accuracy of 0.9563, while lightGBM achieved an accuracy of 0.9566. The acceleration data were collected through experiments to predict forklift failure and RUL, particularly due to repeated forklift use when the centers of heavy objects carried by the forklift were skewed to the right. Time-domain statistical features of the acceleration signals were extracted and used as variables by applying a 20 s window. Subsequently, logistic regression and random forest models were employed to classify the failure stages of the forklifts. The F1 scores (macro) obtained were 0.9790 and 0.9220 for logistic regression and random forest, respectively. Moreover, random forest probabilities for each stage were combined and averaged to generate a degradation curve and determine the failure threshold. The coefficient of the exponential function was calculated using the least squares method on the degradation curve, and an RUL prediction model was developed to predict the failure point. Furthermore, the SHAP algorithm was utilized to identify significant features for classifying the stages. Fault diagnosis using alarm rules was conducted by establishing a threshold derived from the significant features within the normal stage.

1. Introduction

This paper focuses on the challenge of mitigating durability degradation and accurately predicting failures in forklifts operating under demanding conditions. As classified by the American Society of Mechanical Engineers (ASME), forklifts, powered vehicles used for various material handling tasks, often encounter durability and performance issues while maneuvering heavy loads across logistics warehouses, construction sites, factories, and similar environments. These issues escalate service and maintenance costs, introduce safety hazards, and potentially endanger human lives. To address these challenges, the integration of prognostics and health management (PHM) technologies become essential. PHM, a comprehensive framework leveraging sensors, aims to assess system health, diagnose anomalies, and predict remaining useful life [1]. By encompassing condition monitoring, assessment, fault diagnosis, and prediction, PHM optimizes decision making for condition-based maintenance [2]. It necessitates the application of a systematic methodology to select appropriate feature engineering techniques and algorithms tailored to the specific context [3]. Notably, the study by Meng et al. [4] presented a comprehensive overview of research trends in PHM, particularly focusing on lithium-ion batteries. Categorizing prediction approaches into physics-based, data-driven, and hybrid categories, their work provides valuable insights into the diverse avenues within the field.
Recognizing the limitations of traditional reliability analysis, which relies on mean time to failure data and failure probability distributions, contemporary efforts are directed towards harnessing sensor-based PHM technologies to overcome these challenges. Previous studies in reliability and fault diagnosis have predominantly focused on statistical analyses under average load conditions or known failure scenarios [5,6,7,8,9]. However, the advent of the internet of things (IoT) is driving the transition to sensor-based PHM in a variety of domains as it evolves into a new phase of asset management [10]. Physics-based PHMs formalize the complex behavior of equipment and systems to help diagnose and predict failures when mechanical systems exceed predefined physical thresholds. In contrast, sensor-based methodologies have attracted considerable attention due to their ability to proficiently cope with complex troubleshooting scenarios due to their powerful representation and automated feature learning capabilities [11]. However, sensor-based PHMs face several challenges. Data collected from multiple sensors are prone to noise during acquisition and transmission. Since each type of failure produces different failure signals, signals and failures may not correspond exactly one-to-one in general [12]. The limitations of directly monitoring raw signals in deep-learning-based PHM are recognized, necessitating a transformation process termed feature engineering [12,13]. This process encompasses noise filtration, statistical feature extraction, frequency conversion, and context-dependent conditional reduction, thereby enhancing the data’s meaningfulness and applicability.
Within the data-driven and deep-learning-based PHM framework, considerable research is dedicated to fault diagnosis and the prediction of remaining useful life. While PHM technologies aim for high fault detection accuracy and predictive capabilities, the practicality of models is also contingent on achieving a lightweight design. Ding et al. [14] introduced a lightweight multiscale convolutional network tailored for bearing fault diagnosis in edge computing scenarios, specifically targeting train bogie bearings. To complement model-based lightweighting efforts, feature-based lightweighting is pursued through dimension reduction. Lee et al. [3] emphasized the importance of effective feature engineering and data dimensionality reduction in PHM.
Addressing the challenge of imbalanced data situations, Zhang et al. [15] proposed an integrated multi-task intelligent bearing fault diagnosis method, leveraging representation learning under unbalanced sample conditions. This approach is particularly relevant given the common scenario of skewed sample ratios between fault and normal data. Furthermore, the scarcity of fault data arising in distinct environments necessitates innovative solutions. To overcome this data scarcity challenge, techniques such as data augmentation via generative adversarial networks (GANs) [16] and sensor signal transformation [17] have been advanced. Numerous studies have contributed comprehensive insights into remaining useful life (RUL) prediction within the PHM process [18,19,20,21,22,23,24,25,26,27,28]. RUL prediction encompasses four pivotal technical stages: data collection, construction of health indices, health state segmentation, and RUL prediction [18]. These stages synergistically form a comprehensive framework for accurate prediction.
The Industrial Truck Association (ITA) classifies forklifts into eight distinct product families, ranging from Class I to Class VIII. These classifications are based on specific usage characteristics and structural differentiations. For the purpose of this investigation, a PHM study was undertaken, focusing on the front-end structure responsible for lifting heavy loads. This structural component serves as the core element within ITA Class I-type electric counterbalance forklifts. Durability evaluation of forklifts typically involves testing under average loads and standardized conditions, conforming to established standards such as ISO 3691-1, ANSI/ITSDF B56.1, EN 1726-1, and ASME B56.1. However, the applicability of these conventional statistical approaches can be limited in non-standard operational settings where anomalies can arise. In real-world work environments, forklifts often encounter abnormal loading conditions that can result in rollovers or structural failures. Moreover, when latent strength deficiencies accumulate during assembly, the risk of structural failure becomes significantly elevated.
Among the various safety concerns linked with forklifts, the most significant hazard is the occurrence of vehicle rollovers. These incidents commonly occur when maneuvering heavy objects that exceed the forklift’s designated capacity or when handling unbalanced or improperly centered loads. Furthermore, the repetitive use of forklifts while carrying unbalanced heavy loads can lead to performance degradation over time. This cumulative deterioration can eventually culminate in safety accidents. This particular failure mode operates outside the established standard and stands as an outlier load condition. It resides beyond the boundaries of average reliability ranges and is, therefore, not explicitly accounted for in typical design considerations. Paradoxically, these abnormal failures frequently occur within real-world operational environments. This inherent discrepancy necessitates the development of abnormal failure prevention technologies. These technologies should extend beyond the confines of existing reliability-based lifetime management approaches. Specifically, there is a need for a PHM technique that leverages machine learning and deep learning methodologies, utilizing sensor data to detect changes in forklift operational states. This enables the accurate classification and prediction of impending failures. Consequently, this study introduces an effective PHM approach designed to mitigate safety accidents stemming from abnormal forklift usage conditions. By addressing these anomalies, the aim is to enhance forklift durability, thereby curbing maintenance costs and fostering a value chain that prioritizes accident prevention.
This paper provides a systematic presentation of forklift failure diagnosis and prediction, aiming to ensure the reliability of the forklift front-end, as illustrated in Figure 1. The following overview outlines the content of each section.
  • In Section 2, vibration data were collected under weight-imbalance conditions, and a subsequent feature engineering process was executed to facilitate a comprehensive classification of the centers of heavy objects. The primary objective of this classification study was to diagnose weight imbalance, a critical factor influencing forklift durability. By addressing this imbalance, the study aimed to proactively prevent factors leading to durability degradation in forklifts, thereby minimizing the risk of vehicle rollover.
  • Continuing in Section 3, vibration and sound data were collected from forklifts operating under abnormal weight-imbalance conditions. Employing another feature engineering process alongside health stage classification, this section further delves into the realm of anomaly detection and classification.
  • Section 4 is devoted to conducting remaining useful life (RUL) analysis and an alarm-rule-based fault diagnosis, serving the overarching goal of enhancing operational maintenance. The RUL prediction was based on the probability model established in Section 3, enabling accurate predictions. Additionally, significant features were extracted from both classifiers and features previously generated in Section 3. This information was then employed for an early front-end failure diagnosis through an alarm-rule-based approach.
  • Lastly, Section 5 encapsulates the findings and conclusions, summarizing the contributions derived from the investigation.
Figure 1. Outline of procedure used to predict the failure of forklift front-end.
Figure 1. Outline of procedure used to predict the failure of forklift front-end.
Sensors 23 07706 g001

2. Diagnosing and Classifying the Weight Center of Heavy Objects Carried by Forklifts

This section addresses the diagnosis of weight imbalances, which have a direct impact on the overall durability of forklifts. To achieve this, a structured approach was followed that included several essential steps: data preprocessing, feature engineering, data augmentation, and careful selection and evaluation of appropriate machine learning models. The overall goal was to develop a comprehensive method to detect the center of gravity, as shown in Figure 2.

2.1. Experimental Data Acquisition and Feature Engineering

Vibration (acceleration) data were acquired from the forklift’s front-end structure and presented a process to diagnose and classify the weight center of heavy objects carried by forklifts. The front-end structure of the forklift consists of a mast, backrest, carriage, and forks, as shown in Figure 3, and the acceleration signals were measured from the outer beam of the mast.
In the measurement experiment for data acquisition, the weight center of heavy objects carried by the forklift was measured in three configurations: center, left, and right. The condition segments of the dataset were classified and organized into center, left, and right according to each center of gravity condition. Two embedded devices (one on the left and one on the right) were attached to the front-end structure of the forklift truck to measure the vibration acceleration in three axes (x, y, z) (sampling rate 500 Hz), as shown in Figure 4. Considering the load conditions under which the forklift operates, the operating environments of the two datasets (datasets 1 and 2) were simulated in the experiments while maintaining a state that included ground noise. In dataset 1 (only driving mode), the vehicle was loaded with 3200 kg of weight and traveled 80 m at maximum speed, as shown on the left in Figure 5. All measurements were taken for approximately 20 min for center, left, and right condition segments to eliminate data imbalances within condition segments. In dataset 2 (complex mode), the forklift made a round trip of 80 m, as shown on the right in Figure 5, and added lifting, lowering, and back-and-forth tilting tasks at the end of the trip. In dataset 2, approximately 32 min of data were acquired, which is 12 min longer than in dataset 1. Similarly, approximately 32 min of data were acquired for the center, left, and right condition segments to eliminate the data imbalance.
Six acceleration signals (2 sensors × (x, y, z accelerations)) were included in Data sets 1 and 2. In Dataset 2, for one acceleration signal, 960,000 feature vectors were collected at 500 (Hz) × 32 (min) × 60 (s/min). Given the large data size, which could lead to inefficient analysis, this study referred to previous studies [27,28] to handle the data. Eight features were extracted from each window at one-second intervals: min, max, peak to peak, mean (abs), rms (root mean square), variance, kurtosis, and skewness. In addition, four features were added by combining the max, rms, and mean (abs) features: crest factor, shape factor, impulse factor, and margin factor, as listed in Table 1.
Through the above process, 12 features were extracted, and the measured data were compressed for efficient analysis. The data compression process transformed the dimensionality of the data from 960,000 × 6 × 1 to 1920 × 6 × 12. In this way, the number of feature vectors in the data was reduced, but the number of features was increased twelvefold, resulting in 72 features. The aim was to enable effective data processing and facilitate the diagnosis of the weight center of heavy objects at a one-second interval. Furthermore, the data range of the feature vectors was scaled from 0 to 1 using min–max normalization. Furthermore, an exponentially weighted moving average (EWMA) with a window size of 2 to 3 s was used to smooth the noise signals of the generated features. This approach helped to minimize the noise and outliers in the feature data. Moving averages average out the effects of past data, and exponentially weighted moving averages have the advantage of exponentially attenuating these effects. Therefore, they are used for tasks such as time series forecasting and noise reduction [29,30]. The exponentially weighted moving average is used in a way that adjusts the value of alpha (α) based on the window size, as shown in Equation (1). As described in Equation (3), this method smooths out noise in the feature vector to minimize outliers while also reducing the influence of past vectors.
α = 2 w i n d o w   s i z e + 1 ,         w i n d o w   s i z e 1
y 0 = x 0
y t = x t + ( 1 α ) x t 1 + ( 1 α ) 2 x t 2 + + ( 1 α ) t x 0 1 + ( 1 α ) + ( 1 α ) 2 + + ( 1 α ) t
The parameters employed within Equations (1) through (3) are listed below:
  • α : the weight coefficient of exponentially moving average.
  • w i n d o w   s i z e : the number of data points used by EWMA.
  • y t : the exponentially weighted moving average value at the current time t.
  • x t : the input data value at the current time t.
  • x t 1 ,   x t 2 , , x 0 : the input data values at past time points.
  • ( 1 α ) t ,   ( 1 α ) t 1 , ,   1 : the weights of past time points.
  • 1 + ( 1 α ) + ( 1 α ) 2 + + ( 1 α ) t : the sum of weights.
The exponentially weighted moving average was applied to take advantage of its ability to minimize outliers in the feature vector. Figure 6 shows the ‘min’ feature extracted from the x-acceleration signal using the EWMA technique. Table 2 and Table 3 contrast features extracted from the dataset before and after the application of EWMA. They showcase the initial data state alongside the effects of EWMA, including smoothing and value adjustments.
Through the previous feature engineering process, as shown in Table 4, 3699 feature vectors were generated from dataset 1 (only driving) in approximately 20 min, and 5755 feature vectors were derived from dataset 2 (complex mode), measured for approximately 32 min. In total, 9454 datasets were obtained, which were further divided into training and test datasets at a 7:3 ratio. The training and test datasets comprised 6617 and 2837 samples, respectively, with each feature vector containing 72 features. An attempt was made to use the training dataset to develop machine learning classifier models and check the performance of the machine learning models on the test dataset.
The number of feature vectors in the dataset was also augmented to minimize overfitting during the training process. Data augmentation was performed using additive white Gaussian noise (AWGN) and long short-term memory autoencoder (LSTM AE), which expanded the training dataset to a maximum of 19,851 samples (Table 5).
AWGN was applied by referring to prior studies [31], and the target signal-to-noise ratio (SNR) was set to 20 dB. An additional dataset could be generated by mixing noise with the original data, as shown in Figure 7. AWGN is a method of adding noise with a Gaussian distribution to the input or output signal of a system. SNR serves as a scale that quantifies the ratio between the signal’s strength and the noise level. A higher SNR value corresponds to a more robust signal, reducing the relative impact of noise. AWGN based on SNR can be expressed as follows [31]:
N t = A 10 S N R / 10 2 · w t
  • N t : the noise of the signal.
  • A : the magnitude of the input or output signal.
  • S N R / 10 : the standard deviation σ of the noise, calculated from the SNR value.
  • w t : the white noise, which follows a Gaussian distribution with a mean of 0 and a variance of 1.
Figure 7. Results of noise mixed augmentation using AWGN.
Figure 7. Results of noise mixed augmentation using AWGN.
Sensors 23 07706 g007
The autoencoder is a neural network that can use unlabeled training data to learn a code that efficiently represents the input data. This type of coding is useful for dimensionality reduction because it typically has much lower dimensionality than the input. In particular, it works as a powerful feature extractor that can be used for the unsupervised pre-training of deep neural networks. An autoencoder consists of an encoder that converts the input to an internal representation and a decoder that converts the internal representation back to output [32]. The output result is called reconstruction because the autoencoder reconstructs the input. This study used the mean square error (MSE) as the reconstruction loss in training. LSTM, an artificial recurrent neural network, was designed to address the vanishing gradients in traditional recurrent neural networks (RNNs) [33]. As the number of hidden layers in a neural network and the number of nodes in each layer increase, the last layer is trained while the initial layer is not trained.
This long-term dependency problem arises from the vanishing gradient problem, where the gradients tend to converge to zero during the gradient propagation process, particularly when the data length increases during the training phase of RNNs [34,35]. On the other hand, unlike traditional RNNs, LSTM can effectively overcome the vanishing gradient problem by incorporating long- and short-term state values in the learning process, enabling successful learning even with long training durations [36,37]. In this study, an LSTM autoencoder consisting of two LSTM layers was implemented because time series data were used, as shown in Figure 8. Each layer is used as an encoder and decoder [38]. Furthermore, the repeat vector was used in the decoder part to restore the compressed representation to the original input sequence. The repeat vector function repeats the compressed latent space representation to produce a representation that matches the sequence length. This allows the decoder to use the compressed representation multiple times to reconstruct the original input sequence. Using the LSTM autoencoder, the original feature vectors were trained as input data, and the output vectors were used as the augmentation dataset. The output dataset was generated and replicated to minimize the MSE, resulting in a dataset with similar characteristics and patterns to the input dataset, as shown in Figure 9. The equations and parameter descriptions for the LSTM autoencoder are as follows [32,33].
h t = L S T M e n c o d e r x t , h t 1
z t = f W z h t + b z
h t = L S T M d e c o d e r z t , h t 1
x t ^ = f W x h t + b x
M S E = 1 n i = 1 n x i x i ^ 2
  • x t : the input time series data.
  • z t : the output (latent variable) of the encoder.
  • x t ^ : the output (reconstructed time series data) of the decoder.
  • h t and h t : the hidden states of the LSTM.
  • W z , b z , W x , and b x : the learnable parameters (weights and biases) of the model.
  • f : the activation function, typically sigmoid or tanh function.
  • MSE: the mean squared error, loss function.
  • x i : the ith element of the input data.
  • x i ^ : the ith element of the model’s prediction (reconstructed data).
  • n : the number of elements in the input data.
Figure 8. Structure of the LSTM autoencoder layers.
Figure 8. Structure of the LSTM autoencoder layers.
Sensors 23 07706 g008
Figure 9. Result of feature augmentation using the LSTM autoencoder.
Figure 9. Result of feature augmentation using the LSTM autoencoder.
Sensors 23 07706 g009

2.2. Result of Classification

To compare the accuracy of failure prediction, the selected classification algorithms were random forest [39] and LightGBM [40]. To enhance their performance, the ‘Bayesian optimization’ method was employed for hyperparameter tuning. Random forest, an ensemble technique, is rooted in decision trees and serves as a classifier. Decision trees build tree-like models based on input variables, efficiently growing the tree by identifying optimal splitting rules at each branch. However, the vulnerability of a single decision tree to overfitting can hinder its ability to generalize well. To address this concern, the random forest algorithm is applied to alleviate overfitting concerns. A decision tree, by itself, operates as a tree algorithm for data classification or prediction. It navigates the classification or prediction process by creating a tree structure grounded in the data, partitioning it into multiple child nodes through evaluations of specific conditions at each node. The criteria of these conditions are typically determined by metrics such as information gain ( I G ) or the Gini index ( G i n i ) . These metrics measure the impurity of class distribution at each node, selecting a splitting criterion that minimizes the difference in impurity before and after the split [39,40].
Information   gain :   I G D p , f = I D p j = 1 m N j N p I D j
Gini   index :   G i n i D p = 1 k = 1 K p k 2
  • D p : the data of the parent node.
  • D j : the data of the j th child node.
  • f : the splitting criterion variable.
  • m : the number of child nodes generated after splitting.
  • N p : the number of data points in the parent node.
  • N j : the number of data points in the j th child node.
  • K : the number of classes.
  • p k : the ratio of the k th class.
Random forest stands out as a notable ensemble learning technique, leveraging the power of decision trees. The methodology of random forest is structured around the collaborative efforts of multiple decision trees, which are subsequently aggregated to yield prediction results. The workflow of random forest unfolds as follows:
  • Bootstrap sample creation: The process starts by randomly selecting a subset of the input data to create what is known as a bootstrap sample. This sample comprises a distinct dataset consisting of data instances randomly extracted from the original input dataset.
  • Multiple decision tree generation: Next, numerous decision trees are generated, each stemming from a bootstrap sample. These decision trees come into existence with a random element, ensuring their diversity and independence.
  • Data prediction by decision trees: Each of the generated decision trees is then utilized to predict the input data. This prediction process is carried out individually for all the decision trees in the ensemble.
  • Aggregation of predictions: The prediction results obtained from the individual decision trees are aggregated. This aggregation can take the form of averaging the predictions or adopting a majority voting approach, depending on the task. The aggregated outcome serves as the foundation for the final predictions made by the random forest model.
By following these steps, random forest harnesses the collective insights of multiple decision trees, effectively enhancing prediction accuracy and generalization capabilities. Random forest addresses overfitting by generating multiple models from different data subsets. This diversification improves robustness against noise and uncertainties. The adjustment of the optimal number of decision trees and the splitting criteria can be performed through hyperparameter tuning. Typically, hyperparameters such as the splitting criteria of decision trees and the tree depth are set. The prediction function of random forest is as follow, where T is the number of generated decision trees and f t x represents the prediction function of the t th decision tree.
f x = 1 T t = 1 T f t x
LightGBM is a machine learning model based on the gradient boosting decision tree (GBDT) algorithm. Gradient boosting works by improving the prediction model as new models compensate for the errors of the previous model. Therefore, multiple decision trees can be combined to develop a more robust model that minimizes overfitting. The working mechanism of lightGBM is similar to the conventional GBDT algorithm, but it utilizes the leaf-wise approach during the splitting process (Figure 10). This approach allows lightGBM to produce more unbalanced trees than the traditional level-wise approach, resulting in improved predictive performance. Furthermore, lightGBM includes the feature to perform splitting using only a subset of the data, ensuring faster processing speed for large-scale datasets. However, lightGBM may result in deeper trees, depending on their leaf-wise characteristics and hyperparameter settings, which may lead to deeper trees [41]. While this can improve the prediction accuracy of the training data, it may result in lower accuracy when predicting new data because of the overfitting problem. In the case of lightGBM, the aim was to minimize the overfitting problem through the feature augmentation conducted previously. The objective function and parameter descriptions for LightGBM are as follows [42]:
O b j Θ = i l y i , y i y i t 1 ^ + f t x i + t Ω f t
Ω f t = γ T + 1 2 λ w 2
  • l y i , y i t 1 ^ + f t x i : the loss functions used in the objective function.
  • y i : the actual value for the i th data.
  • y i t 1 ^ : the prediction from the previous time step (t − 1).
  • f t x i : the prediction of the t th tree for the i th data point x i .
  • Ω f t : a term used to regulate the complexity of the tree.
  • T : the number of leaf nodes in the tree.
  • γ : the cost parameter associated with the number of leaf nodes.
  • λ : a coefficient that regulates the weight of leaf nodes.
  • w : the weights assigned to the tree nodes.
Figure 10. Two kinds of tree growth: (a) level-wise growth, (b) leaf-wise growth.
Figure 10. Two kinds of tree growth: (a) level-wise growth, (b) leaf-wise growth.
Sensors 23 07706 g010
Bayesian optimization is a method for finding the optimal solution that maximizes an arbitrary objective function. This optimization technique can be applied to any function for which observations can be obtained and is particularly useful for optimizing black-box functions with high cost and unknown shapes [43]. Therefore, Bayesian optimization is used mainly as a hyperparameter optimization method for machine learning models, taking advantage of the characteristics of such optimization techniques [44]. The optimized hyperparameters were derived through Bayesian optimization, as shown in Table 6. In Bayesian optimization, the aim is to identify the hyperparameter combination x * that minimizes or maximizes the objective function f x . The objective function typically takes a form similar to Equation (15). η x represents the actual value of the objective function for the hyperparameter combination x . f x   is defined as the sum of the actual objective function value η x and the noise ϵ x . Bayesian optimization involves experimentation with various hyperparameter combinations while modeling both the genuine objective function value η x and the accompanying noise ϵ x . Through iterative processes, the next hyperparameter combination to explore is forecast, advancing the optimization process.
f x = η x + ϵ x
Because forklifts move continuously, vibration data has the characteristics of time series data. The time series data and the state changes in the condition segment (center of heavy objects carried by forklift) do not depend on the state at a single point in time but on the past values. Therefore, the probability of classifying the conditioning segment was the EWMA to diagnose the conditioning segment contextually using the moving average instead of diagnosing the conditioning segment only by the probability at that time, as shown in Table 7. Contextual diagnosis in machine learning is a technique to diagnose by considering the context of the given data [45]. This provides a more profound understanding than simply analyzing and predicting data patterns. It simply considers the context of the data before and after, rather than individual data points, to help make an accurate diagnosis. In addition, it is used effectively for outlier detection in time series [46] and partial data [47]. This study attempted to minimize the effect of noise, such as outliers, using the exponentially weighted moving average for contextual diagnosis. In applying contextual diagnosis, this study examined the effects of the window size on the moving average. Figure 11 presents the learning and prediction process flow.
As a result of contextual diagnosis through the exponentially weighted moving average, the classification probability of each condition segment predicted by machine learning changes, as listed in Table 7. Forklifts carry and transport unbalanced heavy objects during continuous movement or operation, and the centers of heavy objects do not fluctuate on a one-second basis. Therefore, a two to three-second window was used to calculate the moving average probability. As a result, when the condition segment was diagnosed as “center”, applying a moving average to the probabilities in certain outlier segments diagnosed as “left” or “right” would result in lower values influenced by past data. Figure 12 and Figure 13 present these probabilities as graphs as a function of time. From the observed results, the centers of heavy objects carried by the forklift were diagnosed more accurately by the generated classifier.
Because the data imbalance was minimized, the performance was compared using the accuracy score of 24 classifiers, including random forest and lightGBM (Table 8). First, the accuracy increased in all cases as the window size of the exponentially weighted moving average for smoothing the feature vector time series data increased and the alpha value decreased. The classification probabilities of the condition segment predicted by machine learning were subjected to the moving average to achieve a contextual diagnosis. In all cases, the classification accuracy of the condition segment increased gradually as the size of the moving average window increased, and the alpha decreased. On the other hand, although the data augmentation minimized the overfitting of the model, it resulted in the same or slightly lower accuracy.
Applying moving average to the probabilities of the lightGBM model resulted in an overall increase in the scores (cases 13–24). This trend can be seen in the probability graphs diagnosing each condition segment in Figure 12 and Figure 13. However, lightGBM, with its leaf-wise growth strategy, tended to overfit with increasing tree depth and often has probabilities highly skewed towards 0 or 1. Therefore, it was difficult to observe performance improvement in the combination of the augmented dataset and contextual diagnosis for the model (cases 13, 16, 19, and 22).
Compared to lightGBM, random forest exhibited relatively less overfitting and showed gradual fluctuations in probabilities corresponding to changes in the time series. This trend can be observed in Figure 13. The diagnosis of the condition segments was not sigmoidal but rather smooth and gradual because the probability was calculated by averaging the voting values of multiple randomly generated decision trees. These characteristics were expressed in classifiers utilizing the augmented dataset with AWGN and LSTM autoencoders. When applying AWGN and LSTM autoencoders to the dataset and contextual diagnosis in cases 10–12, the application of the probability moving average resulted in a higher score of 0.0331 (3.31%) compared to cases 1–3. On the other hand, the random forest score was lower than lightGBM in all cases when the moving average was not applied (without a contextual diagnosis).
By applying machine learning and contextual diagnosis, the diagnosis of the centers of heavy objects carried by forklifts was performed by the condition segment during the process of lifting, tilting, and moving heavy objects on uneven ground using a forklift. As a result, the random forest model (case 12) achieved a maximum accuracy of 0.9563, while the lightGBM model (case 24) achieved a maximum accuracy of 0.9566.

3. Abnormal Lifting Weight Stage Classification

In Section 3, vibration and sound data were acquired from forklifts operating under repeated abnormal weight lifting conditions. In addition, feature engineering and health stage classification of the data were performed following a systematic PHM procedure as depicted in Figure 14.

3.1. Experimental Data Acquisition and Feature Engineering

Safety accidents and failure situations were induced by continuous lifting of unbalanced heavy objects in a laboratory environment. Data measurements and condition diagnoses were conducted for these situations. During the data measurement process, the forklift repeatedly lifted and lowered an unbalanced load of 1500 kg to the right at a consistent speed every 20 s for five hours. The forklift remained stationary throughout this process and did not perform any driving activity. In addition, four three-axis (x, y, z) acceleration sensors were mounted on the left and right sides of the front-end structure to collect vibration data (Figure 15). Repeated acceleration tests were performed until a failure condition occurred; the forklift swayed, and a loud noise was generated in the structure. In the actual operating environment, it is difficult to diagnose the condition and predict the lifespan by sound because of ambient noise. In this study, a microphone (sampling rate 51.2 kHz) was installed at the driver’s position to eliminate ambient noise in an anechoic chamber environment, and labeling was performed using sound data. At the same time, the noise generated was measured by the forklift. These sound data were used to classify and label the failure stages into three stages: normal, failure entry, and failure.
The dataset obtained in the experiment contains 12 acceleration signals (four sensors × (x, y, z accelerations)). For each signal of one sensor, 512 (Hz) × 300 (min) × 60 (s/min) acceleration data were measured, resulting in a total of 9,216,000 data points. In addition, 921,600,000 (51,200 Hz × 300 min × 60 s/min) sound data were collected. As shown in Figure 16, the plotted sound signal obtained in the time domain made it difficult to track the state changes. By observing the frequency changes over time using the short-time Fourier transform (STFT), it was possible to determine if the forklift was operating over time. On the other hand, state changes and detailed differences were difficult to compare.
The data size was large, and it was difficult to observe the distinctive state changes when analyzing the raw signals. Twelve features were extracted from the time-domain signal, including min, max, peak to peak, mean (abs), RMS, variance, kurtosis, skewness, crest factor, shape factor, impulse factor, and margin factor for every 20 s window to compress the data and solve these difficulties. Based on this, the dimensions of the acceleration signal and sound signal data were reduced to 900 × 12 × 12 and 900 × 1 × 12, respectively. As reported in Section 2, an exponentially weighted moving average (alpha 0.5) was applied to the reduced dataset to minimize the noise signal in the data. After feature engineering, the CART (classification and regression tree) algorithm [40] was used to classify the failure stages over time using sound features. The CART algorithm is a decision tree algorithm for classification and regression analysis that evaluates the importance of each variable based on the input data and prioritizes the important variables to produce a decision tree. The CART algorithm was used to derive 12 decision trees for each feature by pruning to prevent overfitting and classify the status into three levels (Figure 17). Significant branching points could be derived from eight of the 12 decision trees, and the failure stage was classified based on the average value of the branching points derived from the eight decision trees (Table 9). Based on the classification results, breakpoint 1 was determined to be at 1.18 h and breakpoint 2 at 1.77 h, which were used to distinguish the labeling of the data (Table 10). Furthermore, a comparison of the recorded forklift experimental video showed that beyond breakpoint 1, the forklift exhibited early failure symptoms (beginning to shake), demonstrating a complete failure state at breakpoint 2.
The 900 data generated were divided into 630 data for training and 270 data for testing. Given that the dataset is imbalanced, the performance was evaluated using the F1 score during the validation process. The F1 score is one of the metrics that evaluate the accuracy of a model and is calculated as the harmonic mean of precision and recall. When data are unbalanced, the prediction accuracy for a small number of data classifications is often high, but the prediction accuracy for a large number of data classifications is often low. In such situations, relying solely on accuracy may give the impression of high prediction accuracy for a small number of data classifications. On the other hand, the accuracy for most data classifications may be relatively lower, leading to inadequate overall model performance evaluation. The F1 score considers both precision and recall, which makes it more suitable for dealing with imbalanced classification problems [48,49,50]. Therefore, when the data are unbalanced, using the F1 score is a more accurate way to evaluate the model performance. Consequently, the F1 score was used to evaluate the performance in this unbalanced dataset.

3.2. Stage Classification Result

In the previous steps, labels were assigned to instances using sound features. Leveraging these labels, logistic regression and random forest classifiers were trained. The dataset was divided into two distinct cases: case 1 utilized solely vibration features, encompassing 144 dimensions; in contrast, case 2 integrated both vibration and sound features, totaling 156 dimensions. While sound features were primarily introduced for labeling purposes, their inclusion in the training prompted additional cases to assess their impact on classifier performance. In both cases, logistic regression and random forest were selected as the machine learning models, with the details of random forest elaborated upon in Section 2.2. For multiclass classification, logistic regression employs the SoftMax function alongside the cross-entropy loss. In this context, the linear discriminant function h k x is defined as follows:
h k x = w k T x i + b k
where w k and b k denotes the weights and bias for each class k , respectively, while x i represents the i th input variable. p y = j x i , the probability of class j , predicted by the SoftMax function is as follows [32]:
p y = j x i = e h j x i k = 1 K e h k x i
The cross-entropy loss function, J θ , is calculated as follows [32]:
J θ = 1 N i = 1 N j = 1 K y i j log y i j ^
  • y i j : the binary value indicating whether class j is the correct target for the i th data point.
  • y i j ^ : the predicted probability of class j for the i th input data point.
  • N : the total number of data points.
  • K : the total number of classes.
LightGBM, used in the previous section, is prone to overfitting when the dataset size is small. LightGBM was deemed inappropriate because this dataset consists of 900 feature vectors and was not used in this study. No additional hyperparameter tuning was conducted, and only the default parameters provided by the scikit-learn module in Python were utilized. Based on the validation results of the test data shown in Table 11, the logistic regression classifier achieved an F1 score (macro) of 0.9599 for case 1 and 0.9790 for case 2. In the case of random forest, the F1 scores (macro) for cases 1 and 2 were 0.9116 and 0.9220, respectively. Additionally, Figure 18 depicts a confusion matrix, where the x-axis shows the actual (ground truth) class labels, and the y-axis shows the predicted class labels generated by a model. This confirms that it is possible to classify the forklifts as healthy or aging under repeated unbalanced load weight conditions based on acceleration signals.

4. RUL Prediction and Fault Diagnosis with Alarm Rule for Abnormal Lifting

Numerous studies [18,19,20,21,22,23,24,25,26,27,28] have comprehensively explored RUL prediction aligned with fault diagnosis. Zhang et al. [19] introduced an inventive approach by parallelly integrating spatial and temporal features using a hybrid neural network. This model combined a 1D convolutional neural network (CNN) with a bidirectional gated recurrent unit (BiGRU). Furthermore, they assessed the limitations of RUL prediction using CNN, LSTM, and the transformer algorithm on aircraft turbofan engine data. To overcome these limitations, they introduced the integrated multi-head dual sparse self-attention network (IMDSSN), an architecture incorporating the ProbSparse self-attention network (MPSN) and LogSparse self-attention network (MLSN) components [20]. Other strategies in RUL prediction span a broad spectrum. Pham et al. [21] proposed RUL prediction for a methane compressor, employing flow system identification, proportional hazard modeling, and support vector machines. Loutas et al. [22] introduced an ε-support vector machine-based approach for estimating rolling bearing RUL. Gugulothu et al. [23] developed Embed-RUL, addressing differing patterns in embeddings between normal and degraded machines. This technique employs a sequence-to-sequence model based on RNNs to generate embeddings for multivariate time series subsequences. Hinchi et al. [24] introduced a method for bearing RUL prediction centered on a convolutional LSTM network. Niu et al. [25] proposed an RUL prediction technique utilizing a 1D-CNN LSTM network. Jiang et al. [26] presented a time series multiple channel convolutional neural network (TSMC-CNN) integrating CNN with LSTM and attention mechanisms. They further integrated TSMC-CNN with an attention-based long short-term memory (ALSTM) network for bearing RUL prediction. Saidi et al. [28] presented RUL prediction using support vector regression (SVR). These regression-based RUL predictions require well-structured feature vectors and substantial RUL data collection. To address this challenge, several studies have focused on performance enhancement, employing ensemble techniques and a range of model-based approaches, including CNN, LSTM, transformers, attention mechanisms, and SVR. While many works have explored convolutional LSTM networks and sequence-to-sequence models, Ley et al. [18] systematically outlined four technical processes including data collection, health index construction, health stage segmentation, and RUL prediction, with a dedicated focus on RUL.
Considering the algorithms proposed in the literature and the model of Ley et al. in [18], the present study combines appropriate algorithms in an integrated manner, adapted to the unique demands of abnormal lifting scenarios. This customized integration facilitates the model in capturing latent patterns and interactions arising from unbalanced load conditions, thus leading to enhanced RUL predictions. This section demonstrates real-time active diagnosis using the life model equation, even with limited data, thereby enhancing the practicality of the proposed approaches. As Section 3 discussed the classification of health stages, subsequently, in Section 4, the dataset and classifiers are applied to perform RUL prediction and fault diagnosis using alarm rules, thereby establishing a foundation for condition-based management [51]. This study also introduces real-time active forklift condition diagnosis using the life model equation, even with limited data. Moreover, alarm rules are devised utilizing key features extracted from the health stage classifier and designated as health indices. Figure 19 summarizes the RUL calculation process and the methods used.

4.1. Life Prediction Model and RUL Verification

In the case of logistic regression, as shown in Figure 20, the classification probabilities of the classifier were derived as either 0 or 1, making it difficult to track the progressive changes in the state. Therefore, it was not feasible to adopt degradation curves for logistic regression. On the other hand, as shown in Section 3, when the classification probabilities of the entire dataset belonging to each stage were visualized using a random forest classifier, a gradual change in state was observed in the graph (Figure 21). This represents the degradation state of the forklift using classification probabilities, and the probabilities of Stages 1 and 3 were combined and averaged to generate the degradation curve (Figure 22). After analyzing the degradation curve, Stage 3 was reached after approximately 106 min, and the threshold was a probability of 0.7.
The model for RUL was constructed using an exponential function equation, as follows.
  y = a + b × e x p ( c × x t )
Exponential coefficients were determined using the method of least squares, focusing on data collected within the 30 min prior to the diagnostic time. The least squares method is a statistical tool used in regression analysis to determine model parameters that reduce the discrepancy (error) between observed data and the values expected by the model. The following equations and parameters illustrate the implementation of the least squares method with exponential functions, specifically in Equation (20), for a given dataset.
D a t a s e t :   ( x 1 , y 1 , x 2 , y 2 , , x n , y n )
e i = y i a + b × exp c × x t i
E = i = 1 n e i 2 = i = 1 n y i a + b × exp c × x t i 2
  • y i : the observed value of the dependent variable at a point
  • x t i : the value of independent variable x i at a time t .
  • a , b , c: the parameters of the exponential function.
  • e i : the residuals associated with the exponential function’s prediction.
  • E : the objective function to compute the sum of squared residuals, which is minimized to determine the values of a , b , and c .
Based on these characteristics, the exponential function continued to change with time, and the time when the y-value of the exponential function reached 0.7 was calculated to analyze the RUL. During the analysis of the RUL, the confidence interval was included by considering Y ( 0.9 x t ) to Y(1.1 x t ) in the pre- and post-prediction time points. The analysis showed that the early data predicted the failure rather early, as shown in Figure 23. The gap between the actual and the predicted RUL decreased gradually as the failure point approached (Figure 24).

4.2. Alarm-Rule-Based Fault Diagnosis

The SHAP (Shapley additive explanations) algorithm was used to diagnose faults based on the key features and alarm rules. The goal was to determine which features were critical for state classification using the SHAP algorithm with the previously trained random forest classifier. This analysis allowed the identification of the significant features used for state classification. SHAP is an algorithm used to analyze the contribution of features to the predictions made by machine learning models. It helps to determine the importance of each feature in the model’s predictions. The SHAP algorithm considers all features necessary to explain the model predictions and calculates the impact of each feature on the prediction outcome [52]. Calculating the Shapley value for all feature combinations is computationally expensive. Therefore, the SHAP algorithm uses approximations that can be calculated relatively quickly for tree-based models, such as the random forests or XGBoost [29]. The SHAP algorithm provides various interpretation results, such as feature importance, feature contribution, and feature effect, which help to interpret the model and explain the prediction results reliably [52]. The SHAP algorithm improves the interpretability of machine learning models and plays a vital role in model development and helping users to understand the model prediction results [52].
The features that contribute significantly to health diagnostics were identified using the SHAP algorithm. As shown on the left in Figure 25, the ‘abs.mean ch In LH z’ feature contributes the most to the health diagnosis. A progressive state change, similar to the degradation model graph, could be observed by observing the dispersion of the time series of the corresponding feature (Figure 25, right image). The 2-sigma and 3-sigma values of the ‘abs.mean ch In LH z’ feature were extracted from the distribution of the normal data range in stage 1. These values were then used as the threshold for fault diagnosis. Next, an alarm rule was set to diagnose a failure when the exponentially weighted moving average value of the feature (alpha 0.17) exceeds a threshold. As shown in Figure 25 (right), the 3-sigma threshold can diagnose failure just before failure, and the 2-sigma threshold can predict failure at the point of the precursor symptoms.

5. Discussion

This study addresses the critical issue of preventing front-end failures in forklifts by predicting the center of gravity. The study begins by emphasizing the fault diagnosis of front-end failures in forklifts and the significance of accurately predicting the gravity center. In this pursuit, the study acquired acceleration signals during lifting, tilting, and moving heavy objects by the forklift on uneven ground and in various operating environments. These acceleration signals were captured from the outer beams on both the left and right sides of the forklift’s front-end. To ensure effective processing and feature determination, time-domain statistical features were extracted and established as variables by applying a window at one-second intervals. To mitigate potential overfitting, the dataset was augmented with AWGN and LSTM autoencoders. Following these data enhancements, classifiers were used to accurately categorize the center of objects being transported by forklifts during their driving and working phases. This classification task was accomplished using the robust capabilities of the random forest and lightGBM models. The random forest model and lightGBM model were able to predict the center of gravity with an accuracy of 0.9563 and 0.9566, respectively. During the prediction of the gravity center, an exponentially weighted moving average was conducted to smooth out the noise of the features. As a result of applying this moving average, the dataset’s outliers were reduced without necessitating complex noise filtering methods. Moreover, an observation was made: as the window size for the exponentially weighted moving average was increased, the accuracy of the machine learning models also improved. This suggests that a larger window captured more accurate data trends, enhancing predictive performance. By using data augmentation, overfitting was minimized, while maintaining similar or slightly lower accuracy scores. In the lightGBM model, the implementation of a moving average on the classification probabilities showed a tendency to improve the scores. In contrast, lightGBM tended to overfit with increased tree depth, resulting in biased classification probabilities. This made it difficult to improve performance when combining augmented datasets with contextual diagnosis. However, the random forest model showed less overfitting compared to lightGBM and showed gradual changes in classification probabilities over time. These trends were observed in the augmented dataset using the AWGN and LSTM autoencoders. Applying the augmented dataset and contextual diagnosis, along with the moving average to the classification probabilities, resulted in an average score improvement of 3.31%. Notably, the random forest score was lower than the lightGBM score without the moving average.
After predicting the center of gravity, the paper presented a procedure for forecasting the RUL during abnormal operational scenarios and diagnosing failures through the application of alarm rules. To forecast RUL during repetitive unbalanced load conditions, statistical features were extracted from acceleration data using a 20 s window. This data collection took place in an anechoic chamber, with simultaneous recording of sound signals via a microphone. The CART algorithm classified and labeled statistical features derived from the sound signals. Logistic regression and random forest models were then used for failure stage classification, achieving F1 scores of 0.9790 and 0.9220, respectively. Notably, logistic regression achieved the highest score among the classifiers. Conversely, when examining the classifier’s probability change, the probabilities were often skewed toward 0 or 1. This skewed distribution made it difficult to accurately track the forklift’s state changes. Consequently, monitoring the forklift’s degradation status was accomplished through the random forest’s generated classification probabilities. The results showed a gradual change in the condition of the forklift over time. Based on these results, the probabilities generated by the random forest for each stage were combined and averaged to create a degradation curve. The failure point was predicted by the failure threshold that was determined through degradation curve analysis.
In conclusion, predicting the gravity center of objects carried by the forklift yields insights into operations that impact forklift durability. This approach enhances equipment longevity and operational safety. Additionally, providing RUL data facilitates the development of operational plans and efficient maintenance by identifying critical features for setting failure thresholds.

Author Contributions

Conceptualization and methodology, J.-G.L. and Y.-S.K.; validation, formal analysis, investigation, and data curation, J.-G.L. and J.H.L.; research administration, J.H.L.; funding acquisition, J.H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant from the National R&D Project “Development of fixed offshore green hydrogen production technology connected to marine renewable energy” funded by the Ministry of Oceans and Fisheries (1525013967).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author. The data derived from the present study are only partially available for research purposes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tsui, K.L.; Chen, N.; Zhou, Q.; Hai, Y.; Wang, W. Prognostics and health management: A review on data driven approaches. Math. Probl. Eng. 2015, 2015, 793161. [Google Scholar] [CrossRef]
  2. Sheppard, J.W.; Kaufman, M.A.; Wilmering, T.J. IEEE standards for prognostics and health management. In Proceedings of the 2008 IEEE AUTOTESTCON, Salt Lake City, UT, USA, 8–11 September 2008; pp. 97–103. [Google Scholar]
  3. Lee, J.; Wu, F.; Zhao, W.; Ghaffari, M.; Liao, L.; Siegel, D. Prognostics and health management design for rotary machinery systems—Reviews, methodology and applications. Mech. Syst. Signal Process. 2014, 42, 314–334. [Google Scholar] [CrossRef]
  4. Meng, H.; Li, Y.F. A review on prognostics and health management (PHM) methods of lithium-ion batteries. Renew. Sustain. Energy Rev. 2019, 116, 109405. [Google Scholar] [CrossRef]
  5. Klyatis, L.M. Accelerated Reliability and Durability Testing Technology; John Wiley & Sons: Hoboken, NJ, USA, 2012; Volume 70. [Google Scholar]
  6. Elsayed, E.A. Overview of reliability testing. IEEE Trans. Reliab. 2012, 61, 282–291. [Google Scholar] [CrossRef]
  7. Kapur, K.C.; Pecht, M. Reliability Engineering; John Wiley & Sons: Hoboken, NJ, USA, 2014; Volume 86. [Google Scholar]
  8. O’Connor, P.; Kleyner, A. Practical Reliability Engineering; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  9. Zio, E. Reliability engineering: Old problems and new challenges. Reliab. Eng. Syst. Saf. 2009, 94, 125–141. [Google Scholar] [CrossRef]
  10. Kwon, D.; Hodkiewicz, M.R.; Fan, J.; Shibutani, T.; Pecht, M.G. IoT-based prognostics and systems health management for industrial applications. IEEE Access 2016, 4, 3659–3670. [Google Scholar] [CrossRef]
  11. Zhang, L.; Lin, J.; Liu, B.; Zhang, Z.; Yan, X.; Wei, M. A review on deep learning applications in prognostics and health management. IEEE Access 2016, 7, 162415–162438. [Google Scholar] [CrossRef]
  12. Fink, O.; Wang, Q.; Svensen, M.; Dersin, P.; Lee, W.J.; Ducoffe, M. Potential, challenges and future directions for deep learning in prognostics and health management applications. Eng. Appl. Artif. Intell. 2020, 92, 103678. [Google Scholar] [CrossRef]
  13. Forman, G. An extensive empirical study of feature selection metrics for text classification. J. Mach. Learn. Res. 2003, 3, 1289–1305. [Google Scholar]
  14. Ding, A.; Qin, Y.; Wang, B.; Jia, L.; Cheng, X. Lightweight multiscale convolutional networks with adaptive pruning for intelligent fault diagnosis of train bogie bearings in edge computing scenarios. IEEE Trans. Instrum. Meas. 2022, 72, 1–13. [Google Scholar] [CrossRef]
  15. Zhang, J.; Zhang, K.; An, Y.; Luo, H.; Yin, S. An integrated multitasking intelligent bearing fault diagnosis scheme based on representation learning under imbalanced sample condition. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–12. [Google Scholar] [CrossRef]
  16. Bowles, C.; Chen, L.; Guerrero, R.; Bentley, P.; Gunn, R.; Hammers, A.; Dickie, D.A.; Hernández, M.V.; Wardlaw, J.; Rueckert, D. Gan augmentation: Augmenting training data using generative adversarial networks. arXiv 2018, arXiv:1810.10863. [Google Scholar]
  17. Ko, T.; Peddinti, V.; Povey, D.; Khudanpur, S. Audio augmentation for speech recognition. In Proceedings of the Sixteenth Annual Conference of the International Speech Communication Association, Dresden, Germany, 6–10 September 2015. [Google Scholar]
  18. Lei, Y.; Li, N.; Guo, L.; Li, N.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to RUL prediction. Mech. Syst. Signal Process. 2018, 104, 799–834. [Google Scholar] [CrossRef]
  19. Zhang, J.; Tian, J.; Li, M.; Leon, J.I.; Franquelo, L.G.; Luo, H.; Yin, S. A parallel hybrid neural network with integration of spatial and temporal features for remaining useful life prediction in prognostics. IEEE Trans. Instrum. Meas. 2022, 72, 1–12. [Google Scholar] [CrossRef]
  20. Zhang, J.; Li, X.; Tian, J.; Luo, H.; Yin, S. An integrated multi-head dual sparse self-attention network for remaining useful life prediction. Reliab. Eng. Syst. Saf. 2023, 233, 109096. [Google Scholar] [CrossRef]
  21. Pham, H.T.; Yang, B.S.; Nguyen, T.T. Machine performance degradation assessment and remaining useful life prediction using proportional hazard model and support vector machine. Mech. Syst. Signal Process. 2012, 32, 320–330. [Google Scholar]
  22. Loutas, T.H.; Roulias, D.; Georgoulas, G. Remaining useful life estimation in rolling bearings utilizing data-driven probabilistic E-support vectors regression. IEEE Trans. Reliab. 2013, 62, 821–832. [Google Scholar] [CrossRef]
  23. Gugulothu, N.; Tv, V.; Malhotra, P.; Vig, L.; Agarwal, P.; Shroff, G. Predicting remaining useful life using time series embeddings based on recurrent neural networks. arXiv 2017, arXiv:1709.01073. [Google Scholar] [CrossRef]
  24. Hinchi, A.Z.; Tkiouat, M. Rolling element bearing remaining useful life estimation based on a convolutional long-short-term memory network. Procedia Comput. Sci. 2018, 127, 123–132. [Google Scholar] [CrossRef]
  25. Niu, J.; Liu, C.; Zhang, L.; Liao, Y. Remaining useful life prediction of machining tools by 1D-CNN LSTM network. In Proceedings of the 2019 IEEE Symposium Series on Computational Intelligence (SSCI), Xiamen, China, 6–9 December 2019; pp. 1056–1063. [Google Scholar]
  26. Jiang, J.R.; Lee, J.E.; Zeng, Y.M. Time series multiple channel convolutional neural network with attention-based long short-term memory for predicting bearing remaining useful life. Sensors 2019, 20, 166. [Google Scholar] [CrossRef] [PubMed]
  27. Ali, J.B.; Saidi, L.; Harrath, S.; Bechhoefer, E.; Benbouzid, M. Online automatic diagnosis of wind turbine bearings progressive degradations under real experimental conditions based on unsupervised machine learning. Appl. Acoust. 2018, 132, 167–181. [Google Scholar]
  28. Saidi, L.; Ali, J.B.; Bechhoefer, E.; Benbouzid, M. Wind turbine high-speed shaft bearings health prognosis through a spectral Kurtosis-derived indices and SVR. Appl. Acoust. 2017, 120, 1–8. [Google Scholar] [CrossRef]
  29. Holt, C.C. Forecasting seasonals and trends by exponentially weighted moving averages. Int. J. Forecast. 2004, 20, 5–10. [Google Scholar] [CrossRef]
  30. Hunter, J.S. The exponentially weighted moving average. J. Qual. Technol. 1986, 18, 203–210. [Google Scholar]
  31. Vaseghi, S.V. Advanced Digital Signal Processing and Noise Reduction; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  32. Aurélien, G. Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2022. [Google Scholar]
  33. Graves, A.; Graves, A. Long Short-Term Memory. Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar]
  34. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
  35. Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training recurrent neural networks. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16 June–21 June 2013; Volume 28, pp. 1310–1318. [Google Scholar]
  36. Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef]
  37. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  38. Bao, W.; Yue, J.; Rao, Y. A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PLoS ONE 2017, 12, e0180944. [Google Scholar] [CrossRef] [PubMed]
  39. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  40. Breiman, L. Classification and Regression Trees; Routledge: London, UK, 2017. [Google Scholar]
  41. Liang, W.; Luo, S.; Zhao, G.; Wu, H. Predicting hard rock pillar stability using GBDT, XGBoost, and LightGBM algorithms. Mathematics 2020, 8, 765. [Google Scholar] [CrossRef]
  42. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  43. Greenhill, S.; Rana, S.; Gupta, S.; Vellanki, P.; Venkatesh, S. Bayesian optimization for adaptive experimental design: A review. IEEE Access 2020, 8, 13937–13948. [Google Scholar] [CrossRef]
  44. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for hyper-parameter optimization. In Proceedings of the Advances in Neural Information Processing Systems 24 (NIPS 2011), Granada, Spain, 12–15 December 2011; Volume 24. [Google Scholar]
  45. Singh, K.; Upadhyaya, S. Outlier detection: Applications and techniques. Int. J. Comput. Sci. Issues IJCSI 2012, 9, 307. [Google Scholar]
  46. Weigend, A.S.; Mangeas, M.; Srivastava, A.N. Nonlinear gated experts for time series: Discovering regimes and avoiding overfitting. Int. J. Neural Syst. 1995, 6, 373–399. [Google Scholar] [CrossRef] [PubMed]
  47. Kou, Y.; Lu, C.T.; Chen, D. Spatial weighted outlier detection. In Proceedings of the 2006 SIAM International Conference on Data Mining, Bethesda, MD, USA, 20–22 April 2006; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2006; pp. 614–618. [Google Scholar]
  48. Sokolova, M.; Japkowicz, N.; Szpakowicz, S. Beyond accuracy, F-score and ROC: A family of discriminant measures for performance evaluation. In Proceedings of the AI 2006: Advances in Artificial Intelligence: 19th Australian Joint Conference on Artificial Intelligence, Proceedings 19, Hobart, Australia, 4–8 December 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1015–1021. [Google Scholar]
  49. Jeni, L.A.; Cohn, J.F.; De La Torre, F. Facing imbalanced data—Recommendations for the use of performance metrics. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; pp. 245–251. [Google Scholar]
  50. Sun, Y.; Wong, A.K.; Kamel, M.S. Classification of imbalanced data: A review. Int. J. Pattern Recognit. Artif. Intell. 2009, 23, 687–719. [Google Scholar] [CrossRef]
  51. Kumar, S.; Mukherjee, D.; Guchhait, P.K.; Banerjee, R.; Srivastava, A.K.; Vishwakarma, D.N.; Saket, R.K. A comprehensive review of condition based prognostic maintenance (CBPM) for induction motor. IEEE Access 2019, 7, 90690–90704. [Google Scholar] [CrossRef]
  52. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
Figure 2. Classification procedure of imbalanced operations based on the center of gravity.
Figure 2. Classification procedure of imbalanced operations based on the center of gravity.
Sensors 23 07706 g002
Figure 3. Electric-powered counterbalance forklift of ITA Class I type.
Figure 3. Electric-powered counterbalance forklift of ITA Class I type.
Sensors 23 07706 g003
Figure 4. The location of DAQ and the apparatus of a forklift carrying object.
Figure 4. The location of DAQ and the apparatus of a forklift carrying object.
Sensors 23 07706 g004
Figure 5. Two operational scenarios for test datasets: driving only (left), complex mode (right).
Figure 5. Two operational scenarios for test datasets: driving only (left), complex mode (right).
Sensors 23 07706 g005
Figure 6. EWMA-processed feature (min) extracted from the x-acceleration signal.
Figure 6. EWMA-processed feature (min) extracted from the x-acceleration signal.
Sensors 23 07706 g006
Figure 11. Schematic procedure of training and prediction.
Figure 11. Schematic procedure of training and prediction.
Sensors 23 07706 g011
Figure 12. Contextual diagnosis graph of lightGBM (left: original, right: after moving average).
Figure 12. Contextual diagnosis graph of lightGBM (left: original, right: after moving average).
Sensors 23 07706 g012
Figure 13. Contextual diagnosis graph of random forest (left: original, right: after moving average).
Figure 13. Contextual diagnosis graph of random forest (left: original, right: after moving average).
Sensors 23 07706 g013
Figure 14. Classification of failure (health) stages in imbalanced abnormal operations.
Figure 14. Classification of failure (health) stages in imbalanced abnormal operations.
Sensors 23 07706 g014
Figure 15. Images of the repeated abnormal lifting experiment in an anechoic chamber with microphone and vibration sensors.
Figure 15. Images of the repeated abnormal lifting experiment in an anechoic chamber with microphone and vibration sensors.
Sensors 23 07706 g015
Figure 16. Waveform and short-time Fourier transform (STFT) of measured sound data.
Figure 16. Waveform and short-time Fourier transform (STFT) of measured sound data.
Sensors 23 07706 g016
Figure 17. Visual examples of CART algorithm results.
Figure 17. Visual examples of CART algorithm results.
Sensors 23 07706 g017
Figure 18. Confusion matrix result: (left) logistic regression, (right) random forest.
Figure 18. Confusion matrix result: (left) logistic regression, (right) random forest.
Sensors 23 07706 g018
Figure 19. Prediction of RUL and fault diagnosis.
Figure 19. Prediction of RUL and fault diagnosis.
Sensors 23 07706 g019
Figure 20. Probability of logistic regression classifier.
Figure 20. Probability of logistic regression classifier.
Sensors 23 07706 g020
Figure 21. Probability of the random forest classifier.
Figure 21. Probability of the random forest classifier.
Sensors 23 07706 g021
Figure 22. Degradation curve of forklift using the average probability of a random forest classifier.
Figure 22. Degradation curve of forklift using the average probability of a random forest classifier.
Sensors 23 07706 g022
Figure 23. Degradation curve of forklift using the average probability of a random forest classifier.
Figure 23. Degradation curve of forklift using the average probability of a random forest classifier.
Sensors 23 07706 g023
Figure 24. Remaining useful lifetime (RUL) result.
Figure 24. Remaining useful lifetime (RUL) result.
Sensors 23 07706 g024
Figure 25. Results of SHAP and fault diagnosis based on the alarm rules.
Figure 25. Results of SHAP and fault diagnosis based on the alarm rules.
Sensors 23 07706 g025
Table 1. Description of features (crest factor, shape factor, impulse factor, and margin factor).
Table 1. Description of features (crest factor, shape factor, impulse factor, and margin factor).
FeatureDimension
Crest factormax/RMS
Shape factorRMS/mean (abs)
Impulse factormax/mean (abs)
Margin factormax/mean ((abs)2)
Table 2. Table of features before applying the exponentially weighted moving average.
Table 2. Table of features before applying the exponentially weighted moving average.
FeaturesFeature Vectors without Exponentially Weighted Window
1101102103104105m
Min.0.8854330.8430960.7976500.8111610.827289.
Max.0.0782640.0807680.1063960.1086210.116061.
Peak to Peak.0.0939020.1123240.1463130.1423080.140491.
Mean (abs).0.3516970.4014780.4076000.3682020.429075.
RMS.0.2671100.3094120.3180850.2910220.342365.
Variance.0.0735000.0980430.1034760.0869300.119690.
Kurtosis.0.0048850.0080430.0113200.0164850.014369.
Skewness.0.4777520.4697980.4912200.4867680.492277.
Crest Factor.0.0644160.0393630.0968500.1271090.100406.
Shape Factor.0.0496050.0608540.0706670.0807900.088413.
Impulse Factor.0.0338820.0231900.0530790.0700030.057815.
Margin Factor.0.0035830.0020520.0036710.0055340.003556.
Table 3. Table of features after applying the exponentially weighted moving average.
Table 3. Table of features after applying the exponentially weighted moving average.
FeaturesFeature Vectors with Exponentially Weighted Window (3 s)
1101102103104105m
Min.0.8753940.8592450.8284480.8198040.823547.
Max.0.0766630.0787160.0925560.1005880.108325.
Peak to Peak.0.0969100.1046170.1254650.1338870.137189.
Mean (abs).0.3650300.3832540.3954270.3818140.405444.
RMS.0.2767650.2930890.3055870.2983040.320335.
Variance.0.0790950.0885690.0960220.0914760.105583.
Kurtosis.0.0056470.0068450.0090820.0127840.013576.
Skewness.0.4790470.4744230.4828210.4847950.488536.
Crest Factor.0.0523900.0458760.0713630.0992360.099821.
Shape Factor.0.0484460.0546500.0626580.0717240.080069.
Impulse Factor.0.0279110.0255510.0393150.0546590.056237.
Margin Factor.0.0029780.0025150.0030930.0043140.003935.
Table 4. Configuration of feature vectors in the dataset.
Table 4. Configuration of feature vectors in the dataset.
Label: CenterLabel: LeftLabel: RightSum
Dataset 11244120912463699
Dataset 21910192819175755
Sum3154313731639454
Table 5. Dataset size based on feature augmentation.
Table 5. Dataset size based on feature augmentation.
Feature CombinationNo. of Training DataNo. of Test Data
1. Original66172837
2. Original + LSTM AE13,2342837
3. Original + AWGN13,2342837
4. Original + LSTM AE + AWGN19,8512837
Table 6. Hyperparameter tuning results of the random forest and lightGBM.
Table 6. Hyperparameter tuning results of the random forest and lightGBM.
Random ForestLightGBM
bootstrapTrue boosting_type‘gbdt’
ccp_alpha0 class_weightNone
class_weightNone colsample_bytree1
criterion‘gini’ importance_type‘split’
max_depthNone learning_rate0.02
max_features‘auto’ max_depth12
max_leaf_nodesNone min_child_samples20
max_samplesNone min_child_weight0.001
min_impurity_decrease0 min_split_gain0
min_impurity_splitNone n_estimators1000
min_samples_leaf1 num_leaves58
min_samples_split2 random_stateNone
min_weight_fraction_leaf0 reg_alpha0
n_estimators130 reg_lambda0
oob_scoreFalse silent−1
random_state42 subsample0.8
warm_startFalse subsample_for_bin200,000
Table 7. Result of contextual diagnosis.
Table 7. Result of contextual diagnosis.
Before Contextual DiagnosisAfter Contextual Diagnosis
TimeLeft (LLH)Right (RRH)CenterTimeLeft (LLH)Right (RRH)Center
10.001120.000140.9987410.001120.000140.99874
20.003050.000440.9965120.002410.000340.99725
30.002390.000700.9969130.002400.000550.99706
40.001030.000120.9988540.001670.000320.99801
50.001170.000220.9986250.001410.000270.99832
60.024920.001980.9731160.013350.001130.98552
70.067150.014670.9181870.040460.007950.95158
80.065000.003030.9319780.052780.005480.94174
90.137490.017390.8451290.095220.011450.89333
100.070090.015560.91436100.082640.013510.90385
..
..
..
mm
Table 8. Result of case studies.
Table 8. Result of case studies.
Case No.DatasetMachine Learning ModelFeature Moving Average Window SizeSmoothing Factor (α)Contextual Diagnosis Accuracy Score
(Probability Moving Average Window Size)
1 s2 s3 s
Case 1Raw featuresRandom forest1 s1.000.75220.81350.8950
Case 2Raw featuresRandom forest2 s0.670.80190.86220.9186
Case 3Raw featuresRandom forest3 s0.500.83470.87520.9274
Case 4With LSTM AE featuresRandom forest1 s1.000.73920.80470.8904
Case 5With LSTM AE featuresRandom forest2 s0.670.78460.84700.9094
Case 6With LSTM AE featuresRandom forest3 s0.500.82160.87130.9263
Case 7With AWGN featuresRandom forest1 s1.000.74870.82380.9020
Case 8With AWGN featuresRandom forest2 s0.670.79940.86010.9203
Case 9With AWGN featuresRandom forest3 s0.500.82940.88190.9366
Case 10With all featuresRandom forest1 s1.000.74870.85870.9362
Case 11With all featuresRandom forest2 s0.670.80330.88970.9450
Case 12With all featuresRandom forest3 s0.500.83050.90480.9563
Case 13Raw featureslightGBM1 s1.000.76590.84530.9295
Case 14Raw featureslightGBM2 s0.670.82230.88580.9496
Case 15Raw featureslightGBM3 s0.500.86460.91290.9637
Case 16With LSTM AE featureslightGBM1 s1.000.76210.83470.9098
Case 17With LSTM AE featureslightGBM2 s0.670.81600.87730.9369
Case 18With LSTM AE featureslightGBM3 s0.500.84880.90410.9521
Case 19With AWGN featureslightGBM1 s1.000.76420.84210.9161
Case 20With AWGN featureslightGBM2 s0.670.83790.89250.9485
Case 21With AWGN featureslightGBM3 s0.500.86430.91220.9591
Case 22With all featureslightGBM1 s1.000.76380.83890.9221
Case 23With all featureslightGBM2 s0.670.82060.88830.9454
Case 24With all featureslightGBM3 s0.500.85690.91150.9566
Table 9. Stage-labeling results of the sound data features using the CART (classification and regression tree) algorithm.
Table 9. Stage-labeling results of the sound data features using the CART (classification and regression tree) algorithm.
FeatureBreakpoint 1Breakpoint 2
Max0.8921.764
Min0.9251.775
Peak20.9251.775
Skewness1.3081.831
Crest Factor1.3531.708
Shape Factor1.3081.775
Impulse Factor1.3531.764
Margin Factor1.3531.775
Average1.1771.771
Table 10. Labeling and structure of a dataset.
Table 10. Labeling and structure of a dataset.
Stage 1Stage 2Stage 3Total
PhaseNormalSemi-failureFailure-
Dataset Range (h)0–1.181.18–1.771.77–5.00–5.0
Number of data212106582900
Table 11. Validation result of stage classification.
Table 11. Validation result of stage classification.
ModelDatasetAccuracyF1 ScoreF1 Score
(Weighted)(Macro)
Logistic
Regression
Case 1 (vibration feature)0.98150.98140.9599
Case 2 (vibration + sound feature)0.98890.98910.9790
Random
Forest
Case 1 (vibration feature)0.95190.95230.9116
Case 2 (vibration + sound feature)0.95560.95560.9220
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, J.-G.; Kim, Y.-S.; Lee, J.H. Preventing Forklift Front-End Failures: Predicting the Weight Centers of Heavy Objects, Remaining Useful Life Prediction under Abnormal Conditions, and Failure Diagnosis Based on Alarm Rules. Sensors 2023, 23, 7706. https://doi.org/10.3390/s23187706

AMA Style

Lee J-G, Kim Y-S, Lee JH. Preventing Forklift Front-End Failures: Predicting the Weight Centers of Heavy Objects, Remaining Useful Life Prediction under Abnormal Conditions, and Failure Diagnosis Based on Alarm Rules. Sensors. 2023; 23(18):7706. https://doi.org/10.3390/s23187706

Chicago/Turabian Style

Lee, Jeong-Geun, Yun-Sang Kim, and Jang Hyun Lee. 2023. "Preventing Forklift Front-End Failures: Predicting the Weight Centers of Heavy Objects, Remaining Useful Life Prediction under Abnormal Conditions, and Failure Diagnosis Based on Alarm Rules" Sensors 23, no. 18: 7706. https://doi.org/10.3390/s23187706

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop