Next Article in Journal
Efficient Preconditioning Based on Scaled Tridiagonal and Toeplitz-like Splitting Iteration Method for Conservative Space Fractional Diffusion Equations
Previous Article in Journal
Sentiment Analysis: Predicting Product Reviews for E-Commerce Recommendations Using Deep Learning and Transformers
Previous Article in Special Issue
Analysis of Production and Failure Data in Automotive: From Raw Data to Predictive Modeling and Spare Parts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method for Predicting Tool Remaining Useful Life: Utilizing BiLSTM Optimized by an Enhanced NGO Algorithm

1
School of Intelligent Manufacturing, Lishui Vocational & Technical College, Lishui 323000, China
2
School of Mechanical Engineering, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(15), 2404; https://doi.org/10.3390/math12152404
Submission received: 2 June 2024 / Revised: 29 July 2024 / Accepted: 30 July 2024 / Published: 2 August 2024

Abstract

:
Predicting remaining useful life (RUL) is crucial for tool condition monitoring (TCM) systems. Inaccurate predictions can lead to premature tool replacements or excessive usage, resulting in resource wastage and potential equipment failures. This study introduces a novel tool RUL prediction method that integrates the enhanced northern goshawk optimization (MSANGO) algorithm with a bidirectional long short-term memory (BiLSTM) network. Initially, key statistical features are extracted from collected signal data using multivariate variational mode decomposition. This is followed by effective feature reduction, facilitated by the uniform information coefficient and Mann–Kendall trend tests. The RUL predictions are subsequently refined through a BiLSTM network, with the MSANGO algorithm optimizing the network parameters. Comparative evaluations with BiLSTM, BiGRU, and NGO-BiLSTM models, as well as tests on real-world datasets, demonstrate this method’s superior accuracy and generalizability in RUL prediction, enhancing the efficacy of tool management systems.

1. Introduction

A tool is an integral part of machine tools and one of the components most susceptible to wear. It has a utility threshold, defined as the level of wear at which the quality of machined components is negatively impacted. Once this wear threshold is reached, a tool can no longer be used effectively. The time it takes for a tool to reach this failure threshold from its current cutting state is known as the remaining useful life (RUL). Understanding the tool wear threshold and RUL is crucial for maintaining production quality and efficiency [1]. Prematurely replacing tools before their failure leads to escalated costs and elongated manufacturing durations [2]. On the other hand, procrastinating their replacement can drastically undermine product quality and production efficiency [3,4]. Hence, the creation of a dependable and precise tool RUL prediction system is critically imperative.
Existing methods for predicting the RUL of cutting tools mainly fall into two categories: model-based and data-driven methods [5]. Model-based methods, grounded in a deep understanding of tool wear and degradation mechanisms, utilize physical or mathematical models to depict a tool’s life degradation process [6]; however, these methods face limitations due to the influence of various factors creating uncertainty in parameter estimation and their inability to update measurement data in real time, which restrict their practical applicability [7].
As artificial intelligence advances, machine learning approaches that leverage signal features are gaining prominence in predicting the RUL of tools. Leveraging the formidable feature learning prowess of machine learning models, a multitude of researchers have deployed algorithms like the BP neural network, support vector machine, and random forest for tool RUL prediction [8,9,10]. While these methods have demonstrated effectiveness, they frequently fail to adequately capture the temporal dependencies critical in extensive and complex industrial datasets. It is thus pivotal to delve into sophisticated algorithms that capitalize on the nuances of time series data, such as recurrent neural networks (RNNs) and their derivatives, which endow an essential edge in enhancing predictive accuracy. Illustratively, Yu et al. [11] crafted an attention–LSTM model predicated on cutting force signals for tool life prognostication. Rabah et al. [12] conceived a CNN-BiLSTM model that integrates signals of cutting force, vibration, and temperature for the same purpose. Nevertheless, the effective tuning of key parameters within a model still merits additional investigation.
Moreover, it is essential to highlight the importance of selecting a suitable optimization algorithm for determining a model’s ideal hyperparameters, a step that is vital for augmenting prediction accuracy [13,14]. The northern goshawk optimization (NGO) algorithm, noted for its straightforward structure and high precision in convergence, is extensively employed in optimizing parameters within machine learning models. For example, Yang et al. [15] employed the NGO algorithm to adaptively search for LSTM’s hyperparameters, yielding adept predictions of short-term runoff trends. Similarly, Zhong et al. [16] utilized the NGO algorithm to fine-tune the hyperparameters of the BiLSTM network effectively, which enhanced the accuracy of facial expression recognition. Nonetheless, the NGO algorithm suffers from suboptimal convergence precision and a propensity for local optima, attributed to its stochastic initialization and greedy paradigm for updating mechanisms, which can impede the optimization process.
Addressing the discussed challenges, this study introduces a predictive method that utilizes BiLSTM optimized by an enhanced NGO algorithm. Initially, the method employs multivariate variational mode decomposition (MVMD) on the acquired signal data to extract multiple intrinsic mode functions (IMFs), from which temporal and spectral features are extracted. Addressing the shortcomings of extant feature selection techniques and the distinct aspects of tool wear, a combined uniform information coefficient (UIC) and Mann–Kendall trend test (MKT) are applied to select features intimately linked with tool wear. The selected features are then fed into a BiLSTM network for feature learning, culminating in the RUL prediction. Concurrently, we incorporated two innovative enhancement strategies to address the previously identified limitations of the NGO algorithm. An enhanced northern goshawk optimization algorithm is utilized to calibrate the BiLSTM parameters optimally. The primary contributions of this study can be encapsulated in the following points:
An enhanced NGO algorithm has been introduced, incorporating a reverse learning strategy and a modified sine algorithm.
(1)
The convergence performance of the MSANGO algorithm has been evaluated through benchmark tests, with comparisons to the standard NGO algorithm and others.
(2)
The application of the MSANGO algorithm in selecting key parameters for the BiLSTM model significantly enhanced its predictive accuracy.
The remainder of this paper is structured as follows: Section 2 discusses data preprocessing, including signal feature extraction and selection. Section 3 elaborates on the theoretical underpinnings of the MSANGO-BiLSTM model. Section 4 elucidates the experimental design pertinent to our study and delves into a comprehensive analysis of the results. Section 5 encapsulates the concluding remarks.

2. Data Preprocessing

2.1. Extraction of Signal Features

In the machining process, signals typically display nonlinear and non-stationary traits. MVMD is a signal processing technique that decomposes complex multidimensional signals into several intrinsic mode functions (IMFs). Each IMF represents different frequency components of the original signal, helping to reveal the intrinsic characteristics and patterns [17]. This method is particularly suitable for processing non-stationary and nonlinear signals. In this study, a six-layer decomposition was conducted on signal data across seven channels to isolate various frequency components. Following signal decomposition, key features reflecting the state of the tool were extracted from both the time and frequency domains, revealing subtle changes in tool wear and providing a data foundation for subsequent monitoring models, thereby enhancing monitoring accuracy.
Time domain features offer direct information about changes in tool condition, as wear-induced alterations affect the fundamental statistical properties of the signals. Frequency domain features are derived from the power spectral density, uncovering shifts in energy distribution caused by tool wear or damage. These changes typically concentrate in specific frequency ranges, offering a solid physical basis for analysis. From each IMF, 17 time domain features and 5 frequency domain features were extracted, totaling 924 features (7 channels × 6 layers × (17 + 5)). Table 1 and Table 2 provide detailed descriptions of these feature expressions and their physical meanings.

2.2. Feature Selection

2.2.1. Normalization

To eliminate the discrepancies in scale and numerical range between features, and to lay a solid foundation for the subsequent correlation and monotonicity analysis, this study employs the max-min normalization method to normalize the extracted features. Assuming that x ¯ and xi represent the normalized data and the original data, respectively, the formula for max-min normalization is as follows:
x ¯ = x i min { x i } max { x i } min { x i } ,

2.2.2. Uniform Information Coefficient (UIC)

During the tool wear process, the trend in signal characteristic changes is often not a simple linear relationship with the tool’s RUL curve. Introduced in 2022, the UIC is a correlation analysis method based on information theory [18]. The UIC is used to measure the correlation between two variables. Compared to the widely used Pearson correlation coefficient method and Spearman correlation coefficient method, the UIC can not only extract linear correlations between variables but also identify nonlinear correlations. Moreover, compared to the maximum information coefficient, the UIC effectively mitigates the influence of noise on correlation analysis, offering a lower computational cost, making it suitable for rapid dimensionality reduction in signal features with abundant data. The specific calculation method of the UIC is as follows:
For two sets of feature vectors, X = [x1, …, xn] and Y = [y1, …, yn], with a sequence length of n, the mutual information coefficient calculation model can be expressed as follows:
I M I ( X ; Y ) = x X y Y p ( x , y ) log 2 p ( x , y ) p ( x ) p ( y ) ,
where IMI(X; Y) is the mutual information coefficient between X and Y; p(x, y) is the joint probability density function of X and Y; and p(x) and p(y) are the marginal probability density functions of X and Y, respectively.
Using a uniform partitioning method, X and Y can be evenly divided into several segments, expressed as follows:
{ l x = x max x min a , 2 a 1 + n 0.6 2 l y = y max y min b , 2 b 1 + n 0.6 2 ,
where lx and ly represent the partition unit lengths of X and Y, respectively; xmax and xmin are the maximum and minimum values, respectively, of feature vector X; ymax and ymin are the maximum and minimum values, respectively, of feature vector Y; a and b denote the number of segments for X and Y, respectively; and n0.6 signifies the grid size of the partition, typically taken as the 0.6th power of the data volume [18].

2.2.3. Mann–Kendall Trend Test (MKT)

Cutting tools’ life degradation often shows time monotonicity, suggesting that features with strong monotonicity better capture a tool’s degradation path. The MKT is a non-parametric method that excels in analyzing time series with consistent monotonic trends. The MKT is used to identify trends in time series data without assuming a specific distribution, making it particularly useful for detecting monotonic trends. Given its wide applications in areas like hydrology and mechanical diagnostics [19,20], we use the MKT to identify key features with monotonic trends, enhancing the prediction of a tool’s RUL.
When analyzing with the MKT, the time series samples (x1, x2, …, xn) consist of n independent, identically distributed random data. Suppose H0 indicates that there is no trend in these sample data. The alternative hypothesis, H1, is a two-sided test, stating that for all i, jn, and ij, the distributions of xi and xj are different. The test statistic, S, is defined as follows:
S = i = 1 n 1 j = i + 1 n sgn ( x j x i ) ,
sgn ( x j x i ) = { 1 x j > x i 0 x j = x i 1 x j < x i ,
where sgn is the sign function. When n ≥ 8, S approximately follows a normal distribution with a mean of 0 and a variance defined as follows:
V a r ( S ) = n ( n 1 ) ( 2 n + 5 ) i = 1 g t i ( i 1 ) ( 2 i + 5 ) 18 ,
where g is the number of groups, with identical elements in the sequence grouped together; ti is the range of the i-th node. By transforming S into Zmk, Zmk can approximate a standard normal distribution:
Z m k = { S 1 V a r ( S ) S > 0 0 S = 0 S + 1 V a r ( S ) S < 0 ,
Under the conditions of choosing a two-sided test and a significance level, α, if | Z m k | Z 1 α / 2 , the null hypothesis, H0, is rejected. This implies that, at the significance level, α, the sample data exhibit a significant increasing or decreasing trend over time. For the two-sided test, when the absolute value of Zmk is greater than or equal to 1.64, 1.96, or 2.57, it indicates that the test has passed at 90%, 95%, and 99% significance levels, respectively. The value of α is determined based on the sample size.

3. Enhanced-NGO-Algorithm-Optimized BiLSTM Model

3.1. Enhanced Northern Goshawk Optimization (MSANGO)

3.1.1. Northern Goshawk Optimization (NGO)

The NGO algorithm represents an innovative optimization algorithm, inspired by the predatory behavior exhibited by the northern goshawk. This method endeavors to pinpoint the optimal solution by perpetually refining the goshawk’s position [21]. The goshawk’s hunting tactics encompass two distinct stages: the prey-identification phase and the subsequent chase-and-evasion phase. The mathematical formulations underpinning the NGO algorithm, tailored to these specific hunting stages, are delineated below:
(1)
Prey-Identification Phase
During the initial hunting stage of the northern goshawk, it selects a target at random and launches a rapid assault. The mathematical representation capturing the goshawk’s behavior in this phase is delineated below:
P i = X k , i = 1 , 2 , , N ; k = 1 , 2 , , i 1 , , N ,
x i , j n e w , p 1 = { x i , j + r ( p i , j I x i , j ) , F P i < F i x i , j + r ( x i , j p i , j ) , F P i F i ,
X i = { X i n e w , p 1 , F i n e w , p 1 < F i X i , F i n e w , p 1 F i ,
where Pi is the position of the prey targeted by the i-th northern goshawk; F P i is the objective function’s value, equating to the fitness score; k is a random integer within the [1, N] range; and X i n e w , p 1 portrays the updated state of the i-th goshawk, while x i , j n e w , p 1 illustrates its renewed state in the j-th dimension, with F i n e w , p 1 being the corresponding fitness metric. The variable r is a stochastic number confined to the [0, 1] interval. Furthermore, I is a random variable that can take on the value 1 or 2. Both r and I serve as random determinants, facilitating the generation of sporadic NGO algorithm behaviors throughout the search and update processes.
(2)
Chase-and-Evasion Phase
Upon the northern goshawk’s assault on its target, the prey instinctively seeks to flee. In this ensuing pursuit, the goshawk exhibits remarkable agility and speed, ensuring its capability to seize the prey in virtually any scenario. Given that this predatory act transpires within an attack radius denoted as R, the mathematical formulation for this secondary phase is articulated below:
x i , j n e w , p 2 = x i , j + R ( 2 r 1 ) x i , j ,
R = 0.02 ( 1 t T ) ,
X i = { X i n e w , p 2 , F i n e w , p 2 < F i X i , F i n e w , p 2 F i ,
where t is the present iteration count, while T is the maximum allowable iterations. X i n e w , p 2 is the updated state of the i-th northern goshawk within the secondary hunting phase. Concurrently, x i , j n e w , p 2 is the renewed state of the same goshawk in the j-th dimension during this phase. Lastly, F i n e w , p 2 corresponds to the fitness metric associated with the new state.

3.1.2. Motivation for Improvement

In its iterative optimization journey, the NGO algorithm can exhibit challenges like suboptimal convergence precision and susceptibility to local optima. NGO’s approach to initializing the initial population relies on random initialization. While this tactic is simple, it risks compromising the diversity of the population’s solutions, thereby constraining the exploration scope. Furthermore, as deduced from Equation (10), NGO adopts a greedy paradigm for updating the population’s position during the prey-detection phase, which heightens the risk of the algorithm settling into local optima as iterations progress.
Given the foregoing analysis, an augmented NGO algorithm, named MSANGO, has been developed to enhance the original NGO’s effectiveness. Initially, a reverse learning strategy is employed for population initialization, enhancing diversity and aiding in detecting potential optimal solutions across an expanded solution space, thereby improving global search capabilities. Further modifications include integrating a modified sine algorithm (MSA) within NGO’s prey-identification phase to refine search strategies, where the MSA incorporates adaptive weighting coefficients during position updates. This adjustment fosters a robust equilibrium between global exploration and localized exploitation. The MSANGO algorithm’s procedural framework is depicted in Figure 1, with the mathematical rationale for the enhanced strategies detailed subsequently.

3.1.3. Reverse Learning Strategy

The reverse learning strategy is a heuristic for population initialization, enhancing diversity and exploration. It uses existing solutions to create ‘reverse’ counterparts, broadening search domain coverage. Recently, many researchers have adopted this method, achieving improved performance and faster convergence in various tasks [22,23].
Let the solution for individuals, i, of the northern goshawk be xi = (xi,1, xi,2, …, xi,D), where i = 1, 2, …, N, and D is the dimension of the search. The reverse solution for an individual’s position is defined as follows:
x i , j = δ ( l b j + u b j ) x i , j ,
where δ is a stochastic variable uniformly distributed between 0 and 1. The term xi,j is the component of an elite individual’s, i’s, solution in the j-th dimension. x i , j is the reverse solution corresponding to xi,j. The index, j, varies from 1 to D, which represents the dimensions of the solution space. The lower and upper bounds within the jth dimension are indicated by lbj and ubj, respectively. The population initialization process based on reverse learning is as follows:
Step1: Generate an initial population of northern goshawks, RP, with N individuals through random initialization.
Step 2: Apply reverse learning to create reverse solutions for each RP member, forming the reverse population, OP.
Step 3: Combine the RP and OP to form the new initial population, NP, with a total of 2N individuals.

3.1.4. Modified Sine Algorithm

The MSA leverages the sine function from mathematics for iterative optimization, demonstrating robust global search capabilities [24]. It incorporates an adaptive variable inertia weight coefficient during the position-updating process, enhancing local area searches and facilitating an effective balance between global exploration and local exploitation. The position-updating formula of the MSA is presented below:
x i ( t + 1 ) = ω t x i ( t ) + r 1 × sin r 2 × [ r 3 p i ( t ) x i ( t ) ] ,
where t is the current iteration number, and xi(t) is the i-th positional component of an individual, X, at the t-th iteration. Additionally, pi(t) refers to the i-th component of the optimal position variable during the same iteration. The function r1 acts as a nonlinear decreasing function, while r2 and r3 are random numbers within the intervals [0, 2π] and [−2, 2], respectively.
The value of r1 is configured using a nonlinear decreasing pattern, with changes in r1 determined by a cosine function spanning the interval from 0 to π:
r 1 = ω max ω min 2 cos π t T max + ω max + ω min 2 ,
where ωmax and ωmin denote the upper and lower bounds of ωt, respectively. The variable t indicates the current iteration, while Tmax specifies the total allowable iterations.
ωt, an adaptive variable inertia weight, diminishes linearly with an increase in the iteration count:
ω t = ω max ( ω max ω min ) × t T max ,

3.2. MSANGO Algorithm Performance Evaluation

3.2.1. Settings for Algorithm Parameters

The MSANGO algorithm has undergone a comparative analysis against the sine cosine algorithm (SCA), whale optimization algorithm (WOA), and NGO algorithm. It is important to note that the parameter settings for these algorithms are based on the recommendations provided in the relevant scholarly literature. Specifically, in the SCA, r1 = 2 − 2t/tmax. For the WOA, the convergence factor linearly decreases from 2 to 0. In pursuit of robust result validation, each algorithm underwent 1000 iterations and 30 separate trials. The mean and standard deviation values from these 30 trials serve as the benchmarks for assessing convergence performance.
Experiments were performed on a system powered by an AMD R7-5800H CPU at 3.2 GHz, supported by 32 GB of RAM, and using the Windows 11 operating system. The simulations were executed using MATLAB version 2022b.

3.2.2. Benchmark Test Functions and Results

The MSANGO algorithm’s efficacy is validated through simulations with six benchmark test functions. Unimodal functions F1 and F2 assess the algorithm’s convergence speed and accuracy. Multimodal functions F3 and F4 evaluate its capability to circumvent local optima and achieve global optima. Fixed-dimension multimodal functions F5 and F6 investigate the algorithm’s performance in addressing specific dimensional challenges. Table 3 details these functions, with ‘n’ representing the dimensionality of the search space.
The convergence performance of the algorithm is evaluated through the mean and standard deviation of the fitness values. A lower mean indicates greater precision in algorithmic convergence, while a reduced standard deviation signifies enhanced stability. The calculations for mean and standard deviation are defined as follows:
Mean = 1 M i = 1 M fitness ( i ) ,
Std = 1 M i = 1 M ( fitness ( i ) Mean ) 2 ,
where fitness(i) denotes the fitness value of the i-th test, while M refers to the total number of experimental runs.
Table 4 illustrates the convergence curves of various algorithms on benchmark test functions. The MSANGO algorithm demonstrates superior performance on unimodal functions, attaining the lowest average values and enhanced stability. In the realm of multimodal functions, although all algorithms exhibit commendable performance, MSANGO stands out in terms of function F4, with exceptional adaptability and precision. For fixed-dimension multimodal functions, the algorithms’ performances are closely matched, which is particularly evident for function F5, where the results are almost indistinguishable. Notably, for function F6, the NGO algorithm maintains the greatest stability, whereas MSANGO exhibits slightly reduced performance.

3.2.3. Analysis of the Convergence Process

Figure 2 illustrates the optimization performance of the MSANGO algorithm alongside other algorithms for various test functions, providing a visual comparison. For unimodal functions, the SCA, WOA, and NGO algorithm demonstrate lower optimization accuracy and slower convergence. In contrast, the MSANGO algorithm achieves rapid convergence to the theoretical optimum within 500 generations for function F1. With multimodal functions, the MSANGO algorithm reaches the theoretical optimum within just 30 generations for function F3. Although the MSANGO algorithm does not fully converge to the theoretical optimum for function F4, it exhibits superior optimization precision and lower variability compared to the other three algorithms. For fixed-dimension multimodal functions, while all algorithms achieve convergence to the theoretical optimum for function F5, the MSANGO algorithm does so more swiftly. The performance for function F6 shows the MSANGO algorithm closely aligning with the other algorithms.
Consequently, the MSANGO algorithm consistently demonstrates outstanding performance across both unimodal and multimodal functions, characterized by its rapid convergence and capacity for in-depth exploitation. This highlights the algorithm’s effective balance between global exploration and local exploitation capabilities.

3.2.4. Analysis of Time Complexity

Time complexity serves as a crucial metric for evaluating the convergence speed of optimization algorithms. This study employs Big O notation for the computation of time complexity [25]. The Big O notation is derived via the following steps: initially, replace all additional constants in runtime with the constant 1; subsequently, retain only the highest-order term in the modified runtime expression; and finally, if a non-unit highest-order term exists, its multiplicative constant is removed, resulting in a Big O expression. Suppose that t denotes the time required to evaluate the objective function. The time complexity analysis for the NGO algorithm starts with the initialization phase, where generating initial population positions incurs O(N × dim). Each iteration involves exploration and exploitation phases, each having a time complexity of O(N × dim + N × t). Consequently, the total time complexity for the NGO algorithm is O(N × dim + T × (N × dim + N × t)). As the MSANGO algorithm introduces no new loops and retains the original sequence in its computations, its time complexity also adheres to O(N × dim + T × (N × dim + N × t)).
The algorithms were executed on all functions for 30 trials, with each trial consisting of 1000 iterations, to gather running times, which were then averaged across the trials, as detailed in Table 5. For every function, from F1 to F6, the MSANGO algorithm consistently demonstrated shorter running times compared to the NGO algorithm, underscoring its superior efficiency on all tested functions.

3.3. MSANGO-BiLSTM Model

LSTM, an especially proficient variant of RNNs, has proven adept at mitigating the gradient explosion or vanishing challenges inherent to basic RNNs [26]. Within the LSTM architecture, three distinct gates are incorporated: input, forget, and output. The input gate assesses the retention level of the current candidate state’s information; the forget gate modulates the amount of information discarded from the prior internal state; and the output gate governs the volume of information relayed from the internal to the external state. As depicted in Figure 3, a standard LSTM cell is presented, with it symbolizing the input gate; gt is the current candidate state; ft is the forget gate; Ot is the output gate; and Xt is the present-time input. Here, σ is the sigmoid function, bounded between (0, 1), and tanh is the hyperbolic tangent, a prevalent activation function.
BiLSTM incorporates two distinct LSTM hidden layers, enabling it to process sequence data bidirectionally—capturing both historical and forthcoming information. This bidirectional paradigm, offering an edge over the traditional unidirectional LSTM, inherently possesses enhanced predictive potential. As depicted in Figure 4, a detailed schematic of BiLSTM is presented: xt is the BiLSTM input, h t is the forward propagation hidden layer’s output, h t corresponds to the backward propagation hidden layer’s output, and yt represents the BiLSTM output.
In the BiLSTM framework, pivotal parameters such as the initial learning rate, neuron count in the hidden layer, and regularization coefficient play a substantial role in dictating the efficacy of time series predictions; however, the field lacks a comprehensive theoretical foundation with which to guide the calibration of these parameters, often relegating their determination to heuristic methods [27,28]. Considering these factors, this study utilizes the MSANGO algorithm to optimize three specific parameters. The average root mean square error from two tools within the training dataset serves as the fitness function for the MSANGO algorithm, guiding the identification of optimal parameter settings.

4. Application of Tool RUL Prediction

4.1. Description of the Experiment

For the empirical validation of the methodology delineated in this manuscript, the PHM 2010 dataset, originating from the 2010 high-speed CNC machine tool health prediction competition organized by the Prognostics and Health Management Society in New York, was selected [29]. This dataset comprehensively documents the lifecycle of six cutting tools from their inception to their wear-out stage, providing real-time data from multiple sensors. It is highly relevant to tool condition monitoring (TCM) scenarios, as it records tool wear status, facilitating the development and validation of tool wear prediction models [30,31]. The data emanate from experiments conducted on the Röders Tech RFM760 CNC machine tool. The experimental parameters were set as follows: a spindle speed of 10,400 revolutions per minute, a feed rate of 1555 mm per minute, a Y-axis cutting depth of 0.125 mm, and a Z-axis cutting depth of 0.2 mm. Within this experimental framework, the study utilized a ball nose carbide milling cutter for face milling operations on a workpiece with a hardness rating of HRC52.
In the experimental phase, the study incorporated a Kistler 9265B triaxial dynamometer positioned between the workpiece and the machining platform for the quantification of cutting force signals. Concurrently, vibrations across three axes were monitored using three Kistler 8636C accelerometers. Additionally, a Kistler 8152 acoustic emission sensor was affixed to the workpiece to detect acoustic signals. The aggregation of these signals was facilitated by an NI PCI1200 data acquisition card, encompassing seven signal channels and operating at a sampling frequency of 50 kHz. A schematic representation of this experimental setup is depicted in Figure 5.
This work focused on three specific milling cutters (C1, C4, and C6) from the dataset. Each of these milling cutters underwent 315 identical cutting cycles under standardized machining conditions. After each cutting session, the flank wear of the tools was measured offline using a LEICA MZ12 microscope. For the purposes of this study, two of these tools were allocated for training the model, while the third served as the test subject, thereby facilitating an assessment of the model’s regression and generalization capabilities. The configuration of the experimental groups is detailed in Table 6. Furthermore, to filter out the non-representative data typically generated during the initiation and cessation phases of the milling process, a subset of one thousand data points, centered around the midst of each cutting cycle, was earmarked for detailed analysis.

4.2. Results of Feature Selection

Only the top 10% of features, ranked by their UIC values, were retained to reduce model complexity, enhance computational efficiency, and ensure the inclusion of the most informative features, thereby improving the model’s generalization and stability. The final selection step involved employing the MKT to isolate features exhibiting pronounced monotonic trends. For this study, given a sample size exceeding 100, the significance level, α, was established at 0.05. This means that a feature is considered statistically significant at a 95% confidence level if the |ZMK| is 1.96 or higher.
For a lucid illustration of the robust correlation between the optimized feature set and the tool’s RUL, Figure 6 depicts the frequency at which various features appear across the experimental groups, alongside the average UIC value and the mean |ZMK| value associated with each feature.
It is observable that, in all three experimental groups, the frequency of features such as Absm, Rms, and RMSA significantly exceeds that of others, indicating their greater importance and stability in predicting the tool’s RUL. In contrast, the features KF, SF, FV, and FSD have a frequency of occurrence of zero, while Ske and RMSF do not appear in two experimental groups, suggesting their lesser efficacy in representing the tool life degradation process. Furthermore, it is observed that for the majority of features, a strong monotonicity often coincides with high correlation, and vice versa.

4.3. Configuration of BiLSTM Network Structure and Parameters

We developed a deep learning framework incorporating BiLSTM layers, dropout layers, a fully connected layer, and a regression layer, as depicted in Figure 7. The input layer receives sequence data reflecting the number of signal channels, enabling the handling of variable-length time series. Following the input layer, two BiLSTM layers extract long-term dependencies and intricate patterns from the data. To mitigate overfitting, each BiLSTM layer is followed by a dropout layer with a dropout rate of 0.3, enhancing the model’s generalizability. The fully connected layer projects the double-layer BiLSTM’s feature representation into a one-dimensional output space. The regression output layer, employing a sigmoid activation function for predictions, utilizes root mean squared error as its loss metric to prioritize minimizing discrepancies between predicted and actual values.
Training parameters include the use of the Adam optimizer, a maximum of 300 epochs, and a batch size of eight. A gradient threshold of one prevents gradient explosion. The initial learning rate, the number of neurons in the hidden layers, and the L2 regularization coefficient are optimized using the MSANGO algorithm proposed in this paper. The learning rate, starting from an initial value determined by the MSANGO algorithm, is adjusted by a piecewise constant strategy, decreasing by 20% every 10 epochs to further enhance training efficiency.

4.4. Assessment of Prediction Performance

In this work, we employ a trio of evaluation criteria, mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE), to holistically gauge the predictive prowess of the model across diverse facets. The RMSE serves as an indicator of the model’s adeptness at mitigating substantial errors, shedding light on its inherent robustness. Conversely, both MAE and MAPE offer a lucid depiction of the precision of the predictions. A diminutive value across these metrics signifies enhanced model efficacy. The computational expressions for these criteria are delineated below:
MAE = 1 n i = 1 n | y i y ^ i | ,
RMSE = 1 n i = 1 n ( y i y ^ i ) 2 ,
MAPE = 1 n i = 1 n | y i y ^ i y i | × 100 % ,
where yi is the true value, y ^ i is the forecasted value, and n is the total sample count.
Figure 8 illustrates the convergence curves for the NGO and MSANGO algorithms on the training dataset. Observations from Figure 8 indicate that the MSANGO algorithm consistently outperforms the NGO algorithm in terms of convergence across all experimental groups. Notably, in the D2 and D3 groups, the initial fitness values of the MSANGO algorithm are substantially superior to those of the NGO algorithm. This enhanced performance is largely attributable to the MSANGO algorithm’s refined approach, which includes initializing populations through a reverse learning strategy. Furthermore, the incorporation of the modified sine algorithm significantly enhances the MSANGO algorithm’s search efficacy. This improvement enables the algorithm to adeptly avoid local optima, thereby facilitating more effective navigation of the search space and accelerating convergence to optimal solutions.

4.5. Comparative Analysis with Leading-Edge Prediction Model

This study compares the predictive performance of the proposed MSANGO-BiLSTM regression prediction model with other models employed in previous research, including BiLSTM [32] and BiGRU [33]. Both the BiLSTM and BiGRU models are variants of recurrent neural networks, whereas the NGO-BiLSTM model serves to illustrate the benefits of parameter optimization via an optimization algorithm and underscores the necessity for enhancements to the NGO algorithm.
The configurations for the BiLSTM and BiGRU models are elaborately outlined in Table 7. For the NGO-BiLSTM and MSANGO-BiLSTM models, the optimization algorithm’s population size is established at 10, with a maximum iteration limit of 30. The range for optimizing the initial learning rate is set between 0.001 and 0.1, the neuron count within the hidden layers ranges from 50 to 200, and the regularization coefficient spans from 0.001 to 0.1. Detailed descriptions of the BiLSTM network structure and parameters are provided in Section 4.3.
Figure 9, Figure 10 and Figure 11 compare the predictive performance of various models against actual values across three experimental groups. The BiGRU model demonstrates suboptimal fitting performance, attributable to its limited number of neurons and reduced iteration count. The BiLSTM and NGO-BiLSTM models, performing comparably, effectively track the RUL’s decline but sometimes tend to pessimistically estimate a tool’s RUL towards the end, predicting fewer cuts than is actually the case. In contrast, the MSANGO-BiLSTM model excels across all three groups, demonstrating consistent stability and accuracy, particularly in the crucial late stage with minimal deviation from actual values. This underscores the MSANGO-BiLSTM model’s superior fitting ability in tool RUL prediction.
Table 8 and Figure 12 display the performance metrics for each model within the three experimental groups. The results demonstrate that, in the context of predicting tool RUL, the three BiLSTM models generally exhibit superior predictive performance compared to the BiGRU model. Additionally, models that have been subjected to specialized optimization processes, such as MSANGO-BiLSTM and NGO-BiLSTM, achieve superior predictive accuracy compared to the standard BiGRU and BiLSTM frameworks. This highlights the significant role of custom-tailored model optimization in augmenting the effectiveness of predictions in specific applications.
The BiGRU model demonstrates underwhelming performance across the three experimental groups. In comparison, the BiLSTM model significantly improves in terms of MAE, RMSE, and MAPE, showing average reductions of 30.11%, 29.61%, and 3.54%, respectively. The NGO-BiLSTM model matches the performance of the BiLSTM model in groups D1 and D2; however, it achieves reductions in MAE, RMSE, and MAPE of 12.05%, 10.53%, and 57.95%, respectively, in group D3.
Notably, the MSANGO-BiLSTM model excels in the MAE and RMSE metrics, which penalize larger errors more severely. Relative to the BiLSTM, BiGRU, and NGO-BiLSTM models, reductions in MAE and RMSE for the MSANGO-BiLSTM are at least 7.49% and 12.43%, indicating consistent stability and precision in its predictions. Although its MAPE is slightly elevated in group D2, it reduces significantly by at least 36.11% and 4.08% in groups D1 and D3, respectively. Overall, the MSANGO-BiLSTM model’s superior performance underscores its strong generalization capabilities and stability, affirming its efficacy in precise data approximation.
Table 9 demonstrates that models optimized by optimization algorithms exhibit a marked increase in training time, entailing extra computational resources and time expenditures. Future strategies could leverage advanced computational hardware and implement parallel as well as distributed training methods to enhance training efficiency substantially. Moreover, the MSANGO algorithm introduced in this study decreases training costs by at least eight percent relative to the NGO algorithm, providing a more efficient alternative. Despite the increase in training duration, the considerable gains in model performance justify the additional investment in training time.

5. Conclusions

This study introduced a predictive method for a tool’s RUL, utilizing joint feature selection and BiLSTM optimized by enhanced northern goshawk optimization. The efficacy of this novel methodology has been substantiated through meticulous milling experiments, culminating in several pertinent conclusions.
(1)
On an array of six benchmark test functions, the MSANGO algorithm exhibited exemplary convergence capabilities and robustness. This performance testifies to the MSANGO algorithm’s adeptness at striking an effective balance between global and local search mechanisms, courtesy of its two innovative improvement strategies. This equilibrium enables the algorithm to exhibit superior convergence attributes in addressing optimization challenges.
(2)
The joint feature selection method introduced in this article concurrently accounts for both monotonicity and relevance in the selection of features; this method manifests its strengths in identifying features intimately linked with a tool’s RUL. Such an approach offers invaluable insights and directives for the formulation of predictive models for tool RUL.
(3)
The integration of the MSANGO algorithm with BiLSTM notably surpasses the other three predictive models, exhibiting at least a 7.49% and 12.43% improvement in the MAE and RMSE metrics, respectively, across the three experimental groups, thereby refining the accuracy of predictions. Additionally, the MSANGO algorithm compared to the NGO algorithm reduces training time by at least 8%, further demonstrating its efficiency and effectiveness in real-world applications.
The proposed tool RUL prediction method provides an effective approach for predicting tool life in machining operations, demonstrating its practical application value; however, to enhance the method’s applicability and accuracy, future studies might explore its use in more diverse operational settings. While the MSANGO algorithm has shown some potential in handling multimodal test functions, there is room for improvement in its stability. Future efforts could include the development of more consistent strategies or optimization techniques to reduce the algorithm’s variability, thus improving its stability and predictive precision.

Author Contributions

Conceptualization, J.W. (Jianwei Wu); Methodology, J.W. (Jianwei Wu); Software, J.W. (Jiaqi Wang); Validation, J.W. (Jianwei Wu) and J.W. (Jiaqi Wang); Formal analysis, J.W. (Jianwei Wu); Investigation, J.W. (Jianwei Wu); Resources, J.W. (Jianwei Wu); Data curation, J.W. (Jianwei Wu); Writing—original draft, J.W. (Jianwei Wu); Writing—review & editing, H.C.; Visualization, J.W. (Jianwei Wu); Supervision, H.C.; Project administration, H.C.; Funding acquisition, J.W. (Jianwei Wu) and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of the People’s Republic of China (grant number: 51975535) and a project supported by the Scientific Research Fund of Zhejiang Provincial Education Department (no. Y202353171).

Data Availability Statement

The data will be made available by the authors on request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Abbreviations

RULRemaining useful life
MVMDMultivariate variational mode decomposition
IMFIntrinsic mode function
UICUniform information coefficient
MKTMann–Kendall trend test
LSTMLong short-term memory
BiLSTMBidirectional long short-term memory
NGONorthern goshawk optimization
MSAModified sine algorithm
MSANGOEnhanced northern goshawk optimization
SCASine cosine algorithm
WOAWhale optimization algorithm
MAEMean absolute error
RMSERoot mean square error
MAPEMean absolute percentage error

References

  1. Cai, W.L.; Zhang, W.J.; Hu, X.F.; Liu, Y.C. A hybrid information model based on long short-term memory network for tool condition monitoring. J. Intell. Manuf. 2020, 31, 1497–1510. [Google Scholar] [CrossRef]
  2. Warke, V.; Kumar, S.; Bongale, A.; Kamat, P.; Kotecha, K.; Selvachandran, G.; Abraham, A. Improving the useful life of tools using active vibration control through data-driven approaches: A systematic literature review. Eng. Appl. Artif. Intell. 2024, 128, 107367. [Google Scholar] [CrossRef]
  3. Karabacak, Y.E. Deep learning-based CNC milling tool wear stage estimation with multi-signal analysis. Eksploat. Niezawodn. 2023, 25, 168082. [Google Scholar] [CrossRef]
  4. Zhang, X.Y.; Lu, X.; Li, W.D.; Wang, S. Prediction of the remaining useful life of cutting tool using the Hurst exponent and CNN-LSTM. Int. J. Adv. Manuf. Technol. 2021, 112, 2277–2299. [Google Scholar] [CrossRef]
  5. Sayyad, S.; Kumar, S.; Bongale, A.; Kotecha, K.; Abraham, A. Remaining Useful-Life Prediction of the Milling Cutting Tool Using Time-Frequency-Based Features and Deep Learning Models. Sensors 2023, 23, 5659. [Google Scholar] [CrossRef] [PubMed]
  6. Zhu, Q.S.; Sun, W.F.; Zhou, Y.Q.; Gao, C. A tool wear condition monitoring approach for end milling based on numerical simulation. Eksploat. Niezawodn. 2021, 23, 371–380. [Google Scholar] [CrossRef]
  7. An, Q.L.; Tao, Z.R.; Xu, X.W.; El Mansori, M.; Chen, M. A data-driven model for milling tool remaining useful life prediction with convolutional and stacked LSTM network. Measurement 2020, 154, 107461. [Google Scholar] [CrossRef]
  8. Chacón, J.L.F.; de Barrena, T.F.; García, A.; de Buruaga, M.S.; Badiola, X.; Vicente, J. A Novel Machine Learning-Based Methodology for Tool Wear Prediction Using Acoustic Emission Signals. Sensors 2021, 21, 5984. [Google Scholar] [CrossRef]
  9. Cheng, Y.N.; Jin, Y.B.; Gai, X.Y.; Guan, R.; Lu, M.D. Prediction of tool wear in milling process based on BP neural network optimized by firefly algorithm. Proc. Inst. Mech. Eng. Part E J. Process. Mech. Eng. 2023. [Google Scholar] [CrossRef]
  10. Chang, W.Y.; Hsu, B.Y. Tool life prediction via SMB-enabled monitor based on BPNN coupling algorithms for sustainable manufacturing. Ai Edam 2023, 37, e20. [Google Scholar] [CrossRef]
  11. Yu, W.W.; Huang, H.; Guo, R.L.; Yang, P.Q. Tool Wear Prediction Based on Attention Long Short-term Memory Network with Small Samples. Sens. Mater. 2023, 35, 2321–2335. [Google Scholar] [CrossRef]
  12. Bazi, R.; Benkedjouh, T.; Habbouche, H.; Rechak, S.; Zerhouni, N. A hybrid CNN-BiLSTM approach-based variational mode decomposition for tool wear monitoring. Int. J. Adv. Manuf. Technol. 2022, 119, 3803–3817. [Google Scholar] [CrossRef]
  13. Singh, P.; Chaudhury, S.; Panigrahi, B.K. Hybrid MPSO-CNN: Multi-level Particle Swarm optimized hyperparameters of Convolutional Neural Network. Swarm Evol. Comput. 2021, 63, 100863. [Google Scholar] [CrossRef]
  14. Li, X.W.; Qin, X.J.; Wu, J.X.; Yang, J.F.; Huang, Z.X. Tool wear prediction based on convolutional bidirectional LSTM model with improved particle swarm optimization. Int. J. Adv. Manuf. Technol. 2022, 123, 4025–4039. [Google Scholar] [CrossRef]
  15. Yang, C.; Jiang, Y.T.; Liu, Y.; Liu, S.L.; Liu, F.P. A novel model for runoff prediction based on the ICEEMDAN-NGO-LSTM coupling. Environ. Sci. Pollut. Res. 2023, 30, 82179–82188. [Google Scholar] [CrossRef] [PubMed]
  16. Zhong, J.R.; Chen, T.X.; Yi, L.H. Face expression recognition based on NGO-BILSTM model. Front. Neurorob. 2023, 17, 1155038. [Google Scholar] [CrossRef] [PubMed]
  17. Rehman, N.u.; Aftab, H. Multivariate variational mode decomposition. IEEE Trans. Signal Process. 2019, 67, 6039–6052. [Google Scholar] [CrossRef]
  18. Mousavi, A.; Baraniuk, R.G. Uniform Partitioning of Data Grid for Association Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1098–1107. [Google Scholar] [CrossRef] [PubMed]
  19. Wei, X.D.; Yang, J.; Luo, P.P.; Lin, L.G.; Lin, K.L.; Guan, J.M. Assessment of the variation and influencing factors of vegetation NPP and carbon sink capacity under different natural conditions. Ecol. Indic. 2022, 138, 108834. [Google Scholar] [CrossRef]
  20. Guo, F.Z.; Rasmussen, B. Predictive maintenance for residential air conditioning systems with smart thermostat data using modified Mann-Kendall tests. Appl. Therm. Eng. 2023, 222, 119955. [Google Scholar] [CrossRef]
  21. Dehghani, M.; Hubálovsky, S.; Trojovsky, P. Northern Goshawk Optimization: A New Swarm-Based Algorithm for Solving Optimization Problems. IEEE Access 2021, 9, 162059–162080. [Google Scholar] [CrossRef]
  22. Ma, G.Y.; Yue, X.F. An improved whale optimization algorithm based on multilevel threshold image segmentation using the Otsu method. Eng. Appl. Artif. Intell. 2022, 113, 104960. [Google Scholar] [CrossRef]
  23. Wang, Y.J.; Wang, G.G.; Tian, F.M.; Gong, D.W.; Pedrycz, W. Solving energy-efficient fuzzy hybrid flow-shop scheduling problem at a variable machine speed using an extended NSGA-II. Eng. Appl. Artif. Intell. 2023, 121, 105977. [Google Scholar] [CrossRef]
  24. Luo, Y.B.; Dai, W.M.; Ti, Y.W. Improved sine algorithm for global optimization. Expert Syst. Appl. 2023, 213, 118831. [Google Scholar] [CrossRef]
  25. Pelusi, D.; Mascella, R.; Tallini, L.; Nayak, J.; Naik, B.; Abraham, A. Neural network and fuzzy system for the tuning of Gravitational Search Algorithm parameters. Expert Syst. Appl. 2018, 102, 234–244. [Google Scholar] [CrossRef]
  26. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, C.; Ma, H.X.; Hua, L.; Sun, W.; Nazir, M.S.; Peng, T. An evolutionary deep learning model based on TVFEMD, improved sine cosine algorithm, CNN and BiLSTM for wind speed prediction. Energy 2022, 254, 124250. [Google Scholar] [CrossRef]
  28. Yang, B.; Wang, Y.S.; Zhan, Y.D. Lithium Battery State-of-Charge Estimation Based on a Bayesian Optimization Bidirectional Long Short-Term Memory Neural Network. Energies 2022, 15, 4670. [Google Scholar] [CrossRef]
  29. 2010 PHM Society Conference Data Challenge. 2010. Available online: https://www.phmsociety.org/competition/phm/10 (accessed on 29 July 2024).
  30. Karabacak, Y.E. Intelligent milling tool wear estimation based on machine learning algorithms. J. Mech. Sci. Technol. 2024, 38, 835–850. [Google Scholar] [CrossRef]
  31. Mishra, D.; Awasthi, U.; Pattipati, K.R.; Bollas, G.M. Tool wear classification in precision machining using distance metrics and unsupervised machine learning. J. Intell. Manuf. 2023, 1–25. [Google Scholar] [CrossRef]
  32. She, C.X.; Li, K.X.; Ren, Y.H.; Li, W.; Shao, K. Tool wear prediction method based on bidirectional long short-term memory neural network of single crystal silicon micro-grinding. Int. J. Adv. Manuf. Technol. 2024, 131, 2641–2651. [Google Scholar] [CrossRef]
  33. De Barrena, T.F.; Ferrando, J.L.; García, A.; Badiola, X.; de Buruaga, M.S.; Vicente, J. Tool remaining useful life prediction using bidirectional recurrent neural networks (BRNN). Int. J. Adv. Manuf. Technol. 2023, 125, 4027–4045. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the MSANGO algorithm.
Figure 1. Flowchart of the MSANGO algorithm.
Mathematics 12 02404 g001
Figure 2. Flowchart of the MSANGO algorithm. (a) F1 (b) F2 (c) F3 (d) F4 (e) F5 (f) F6.
Figure 2. Flowchart of the MSANGO algorithm. (a) F1 (b) F2 (c) F3 (d) F4 (e) F5 (f) F6.
Mathematics 12 02404 g002
Figure 3. Cell structure diagram of LSTM.
Figure 3. Cell structure diagram of LSTM.
Mathematics 12 02404 g003
Figure 4. Schematic diagram of BiLSTM.
Figure 4. Schematic diagram of BiLSTM.
Mathematics 12 02404 g004
Figure 5. Schematic diagram of the experimental setup.
Figure 5. Schematic diagram of the experimental setup.
Mathematics 12 02404 g005
Figure 6. The frequency at which various features appear across the experimental groups, alongside the average UIC values and the mean |ZMK| values associated with each feature: (a) D1; (b) D2; and (c) D3.
Figure 6. The frequency at which various features appear across the experimental groups, alongside the average UIC values and the mean |ZMK| values associated with each feature: (a) D1; (b) D2; and (c) D3.
Mathematics 12 02404 g006aMathematics 12 02404 g006b
Figure 7. Schematic of the BiLSTM network structure.
Figure 7. Schematic of the BiLSTM network structure.
Mathematics 12 02404 g007
Figure 8. Convergence trajectories of the NGO and MSANGO algorithms. (a) D1 (b) D2 (c) D3.
Figure 8. Convergence trajectories of the NGO and MSANGO algorithms. (a) D1 (b) D2 (c) D3.
Mathematics 12 02404 g008
Figure 9. The predictive results of the four models on experimental group D1. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Figure 9. The predictive results of the four models on experimental group D1. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Mathematics 12 02404 g009
Figure 10. The predictive results of the four models on experimental group D2. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Figure 10. The predictive results of the four models on experimental group D2. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Mathematics 12 02404 g010
Figure 11. The predictive results of the four models on experimental group D3. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Figure 11. The predictive results of the four models on experimental group D3. (a) BiLSTM (b) BiGRU (c) NGO-BiLSTM (d) MSANGO-BiLSTM.
Mathematics 12 02404 g011aMathematics 12 02404 g011b
Figure 12. Comparative of evaluation metrics among each model. (a) D1 (b) D2 (c) D3.
Figure 12. Comparative of evaluation metrics among each model. (a) D1 (b) D2 (c) D3.
Mathematics 12 02404 g012
Table 1. Time domain feature expressions and their physical meanings.
Table 1. Time domain feature expressions and their physical meanings.
FeatureExpressionPhysical Meaning
Maximum value (Max) X max = max { x i } Maximum amplitude of the signal
Minimum value (Min) X min = min { x i } Minimum amplitude of the signal
Mean value (Mean) X ¯ = 1 n i = 1 n x i Central tendency of the signal
Peak-to-peak value (PP) X p = X max X min Range of the signal’s amplitude
Absolute mean (Absm) X a r v = 1 n i = 1 n | x i | Average energy of the signal
Variance (Var) X var = 1 n i = 1 n ( x i x ¯ ) 2 Degree of dispersion of the signal
Standard deviation (Std) X σ = 1 n i = 1 n ( x i x ¯ ) 2 Degree of fluctuation of the signal
Kurtosis (Kur) X k u r = 1 n i = 1 n ( x i x ¯ X σ ) 4 Tendency of the signal to produce extreme values
Skewness (Ske) X s k e = 1 n i = 1 n ( x i x ¯ X σ ) 3 Asymmetry of the signal’s probability distribution
Root mean square (Rms) X r m s = 1 n i = 1 n x i 2 Overall energy or strength of the signal
Form factor (FF) X S F = X r m s X a r v Waveform shape characteristics of the signal
Crest factor (CF) X C F = X max X r m s Relative magnitude of the signal’s peak value
Impulse factor (IF) X I F = X max X a r v Impact characteristics of the signal
Margin factor (MF) X m a r = X max X r Relative height of the signal’s peak value
Root mean square amplitude (RMSA) X R M S A = 1 n i = 1 n x i 2 Distribution of the signal’s amplitude
Kurtosis Factor (KF) X K F = X k u r X var 2 Peakedness of the signal’s distribution
Skewness factor (SF) X S F = X s k e X var 2 Symmetry of the signal’s distribution
Table 2. Frequency domain feature expressions and their physical meanings.
Table 2. Frequency domain feature expressions and their physical meanings.
FeatureExpressionPhysical Meaning
Frequency centroid (FC) F C = ( i = 1 n f i p i ) / ( i = 1 n p i ) Central frequency of the signal’s spectrum
Mean square frequency (MSF) M S F = ( i = 1 n f i 2 p i ) / ( i = 1 n p i ) Average energy level of the frequency components
Root mean square frequency (RMSF) R M S F = M S F Overall energy level of the frequency distribution
Frequency variance (FV) V F = ( i = 1 n ( f i F c ) 2 p i ) / ( i = 1 n p i ) Spread of the frequency distribution
Frequency standard deviation (FSD) R V F = V F Variability of the frequency distribution
Table 3. Benchmark test functions.
Table 3. Benchmark test functions.
ExpressionDimensionSearch SpaceOptimal Value
F 1 ( x ) = i = 1 n x i 2 30[−100, 100]0
F 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | 30[−10, 10]0
F 3 ( x ) = i = 1 n [ x i 2 10 cos ( 2 π x i ) + 10 ] 30[−5.12, 5.12]0
F 4 ( x ) = 20 exp ( 0.2 1 n i = 1 n x i 2 ) exp ( 1 n i = 1 n cos ( 2 π x i ) ) + 20 + e 30[−32, 32]0
F 5 ( x ) = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) cos x 1 + 10 2[−5, 5]0.398
F 6 ( x ) = i = 1 4 c i exp ( j = 1 3 a i j ( x j p i j ) 2 ) 3[1, 3]−3.86
Table 4. Comparative analysis of optimization outcomes for benchmark functions.
Table 4. Comparative analysis of optimization outcomes for benchmark functions.
FunctionStatisticAlgorithm
SCAWOANGOMSANGO
F1Mean0.05454.99 × 10−1522.03 × 10−1780
Std0.16962.49 × 10−15100
F2Mean1.98 × 10−51.83 × 10−1052.14 × 10−920
Std3.17 × 10−54.75 × 10−1057.73 × 10−1050
F3Mean18.1164000
Std22.1622000
F4Mean15.16923.29 × 10−155.77 × 10−154.44 × 10−16
Std7.98762.13 × 10−151.78 × 10−150
F5Mean0.39900.39790.39790.3979
Std0.00144.26 × 10−701.42 × 10−9
F6Mean−3.8547−3.8597−3.8628−3.8594
Std0.00240.00362.66 × 10−150.0028
Table 5. Comparative analysis of CPU execution times for the NGO and MSANGO algorithms.
Table 5. Comparative analysis of CPU execution times for the NGO and MSANGO algorithms.
AlgorithmF1F2F3F4F5F6
NGO0.05500.06020.06560.06660.03430.0579
MSANGO0.03420.03990.04120.04390.01780.0418
Table 6. Configuration of the experimental groups.
Table 6. Configuration of the experimental groups.
Experimental GroupTraining SetTesting Set
D1C4, C6C1
D2C1, C6C4
D3C1, C4C6
Table 7. Parameter settings of comparison models.
Table 7. Parameter settings of comparison models.
ModelHidden
Layers
EpochBatch
Size
NeuronsInitial
Learning Rate
Optimizer
BiLSTM12750161100.01Adam
BiGRU11001630.001Adam
Table 8. Evaluation metrics for the prediction results of each model.
Table 8. Evaluation metrics for the prediction results of each model.
ModelD1D2D3
MAERMSEMAPEMAERMSEMAPEMAERMSEMAPE
BiLSTM20.157426.79550.422010.409814.14620.150926.651932.77770.8976
BiGRU52.230365.28010.955615.266919.22260.403625.904933.95970.4319
NGO-BiLSTM22.529327.70530.624010.616512.55020.174723.439529.32560.3774
MSANGO-BiLSTM18.647323.46380.26967.925210.40440.285418.526721.46230.3620
Table 9. Training time of each model.
Table 9. Training time of each model.
ModelTraining Time/s
D1D2D3
BiLSTM223.6069224.2265219.1590
BiGRU17.872013.772514.4773
NGO-BiLSTM4396.46004295.80604404.4936
MSANGO-BiLSTM4036.36453952.06094028.0939
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, J.; Wang, J.; Chen, H. A Method for Predicting Tool Remaining Useful Life: Utilizing BiLSTM Optimized by an Enhanced NGO Algorithm. Mathematics 2024, 12, 2404. https://doi.org/10.3390/math12152404

AMA Style

Wu J, Wang J, Chen H. A Method for Predicting Tool Remaining Useful Life: Utilizing BiLSTM Optimized by an Enhanced NGO Algorithm. Mathematics. 2024; 12(15):2404. https://doi.org/10.3390/math12152404

Chicago/Turabian Style

Wu, Jianwei, Jiaqi Wang, and Huanguo Chen. 2024. "A Method for Predicting Tool Remaining Useful Life: Utilizing BiLSTM Optimized by an Enhanced NGO Algorithm" Mathematics 12, no. 15: 2404. https://doi.org/10.3390/math12152404

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop