1. Introduction
As an essential component in life, valves mainly play a significant role in regulating flow direction, controlling the opening and closing of pipelines, and ensuring effective sealing performance, which directly impacts their functionality. During valve operation, internal leakage poses a greater challenge than external leakage due to its concealed nature, making accurate diagnosis of internal leakage particularly crucial. Currently, methods for detecting internal leakage in valves can be categorized into two main types: offline detection methods, which require disassembly of the valve, and online detection methods, which do not necessitate disassembly. Online methods include acoustic emission detection [
1,
2], negative-pressure wave detection [
3], and ultrasonic detection [
4]. For instance, Au-Yang, M. K. [
5] employed ultrasonic waves to assess the leakage in check valves; however, ultrasonic signals are susceptible to interference from external noise and require active excitation signals to generate detectable responses, resulting in limited real-time performance. Liu et al. [
6,
7] utilized negative-pressure waves for pipeline leakage localization; their detection capability diminishes at low leakage rates, leading to reduced accuracy. In contrast, the acoustic emission detection method offers greater convenience, adaptability, real-time responsiveness, and sensitivity.
In recent years, numerous scholars have engaged in comprehensive research regarding the extraction of acoustic emission signal features associated with valve leakage. This body of work encompasses extensive investigations across the time domain, frequency domain, and time–frequency domain. In the time domain, Ye et al. [
8] introduced an acoustic emission signal analysis method predicated on standard deviation to identify internal valve leakage. They established a mathematical model by fitting the relationship between the standard deviation and leakage rate, utilizing the least squares method. Regarding time–frequency analysis, Sim H.Y [
9] employed wavelet packet transformation to decompose the signal into various frequency ranges, subsequently calculating the root mean square (RMS) value and assessing valve issues based on fluctuations in the RMS value. Additionally, Liang et al. [
10] utilized wavelet scattering transform (WST) to extract the first three wavelet scattering coefficients from leakage signals, which were then employed as feature vectors. The leakage acoustic emission signal from the valve exhibits nonlinear and nonstationary characteristics, which may diminish the efficacy of the previously mentioned statistical features in differentiating fault states. Adaptive time–frequency analysis methods, such as Ensemble Empirical Mode Decomposition (EEMD) [
11], and Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) [
12], have been increasingly employed by researchers to address certain limitations of earlier signal processing techniques. In light of the fact that the emission signal from the valve can be easily influenced by environmental noise, we propose using CEEMDAN to decompose the signal. This approach aims to extract useful signals while minimizing the impact of noise. By selecting the top five IMFs based on their correlation coefficients, we can focus on the most relevant frequency bands associated with valve internal leakage. To further enhance feature extraction, fuzzy entropy is employed to quantify the complexity and irregularity of these selected IMFs, providing robust features for classification.
In the realm of fault classification, the support vector machine (SVM) [
13,
14,
15] is extensively utilized as an effective classification tool. SVM demonstrates a high level of accuracy in fault identification and classification, even when faced with limited data samples owing to its robust learning capabilities. In light of its strong adaptability, low classification error, straightforward feature vector, and ability to work with small datasets, SVM was chosen as the classifier. The kernel functions of SVM can further augment classification performance in nonlinear scenarios. Among the most frequently employed kernel functions is the Radial Basis Function (RBF) [
16,
17], which typically achieves favorable classification outcomes.
The Hippopotamus Optimization Algorithm (HO) [
18] simulates the defense and evasion strategies of the hippopotamus while optimizing these strategies through location updates. This algorithm demonstrates exceptional performance in enhancing accuracy, improving local search capabilities, and exhibiting strong practicality. Nevertheless, the HO algorithm still holds significant research potential for enhancing global search capabilities, strengthening local development abilities, and preventing convergence to local optimal solutions. Notably, IHO is particularly adept at addressing issues related to local optima and slow convergence rates. The contributions of this paper are outlined as follows:
(1) To address the challenges posed by the non-stationary and nonlinear characteristics of valve leakage signals, the CEEMDAN method was proposed for signal decomposition. Subsequently, the fuzzy entropy of the decomposed Intrinsic Mode Function (IMF) signals was computed, allowing for the extraction of an initial feature set from the signals;
(2) A novel intelligent search algorithm IHO was introduced to enhance the kernel parameters of SVM in order to achieve improved performance. Consequently, the IHO-SVM model was developed;
(3) This paper presented the development of a valve fault diagnosis model referred to as CEEMDAN-IHO-SVM, which is capable of accurately extracting and diagnosing fault characteristics.
The subsequent sections of this article are organized as follows.
Section 2 delineates the pertinent methods that have been utilized and enhanced within this study.
Section 3 presents the experimental results along with a discussion of their implications. Finally,
Section 4 provides a summary of the findings and contributions of this work.
2. Materials and Methods
2.1. The CEEMDAN Algorithm
Empirical Mode Decomposition (EMD) [
19] and its optimization methods are adaptive processing techniques well-suited for analyzing nonlinear, unsteady signals. These methods rely on the characteristic time scale of the signal itself, allowing for decomposition into multiple IMFs and a residual component that reflects the overall trend of the signal. EMD optimization methods, such as EEMD, Complete Ensemble Empirical Mode De-composition (CEEMD), and CEEMDAN, are all derived from EMD, inheriting its advantages, including intuitiveness and adaptability.
CEEMDAN is an advanced ensemble empirical mode decomposition algorithm that incorporates adaptive noise. It builds upon the EMD and EEMD algorithms to address limitations inherent in these earlier methods. During EMD decomposition, waveform aliasing often occurs, leading to modal aliasing—a phenomenon where signal interactions are difficult to distinguish. EEMD was introduced as a noise-assisted data analysis method to mitigate modal aliasing in EMD by adding white noise to the original signal in each iteration before performing EMD decomposition. CEEMDAN further refines this approach by decomposing both the original signal plus white noise and the original signal minus white noise using EMD, followed by averaging the resulting IMFs to eliminate added noise, thereby enhancing decomposition accuracy and stability. However, residual white noise can still affect subsequent processing and analysis in both EEMD and CEEMD. CEEMDAN addresses this issue by more effectively suppressing noise interference and reducing the number of iterations and computational complexity.
The subsequent section illustrates the decomposition process of CEEMDAN applied to the signals.
First, Gaussian white noise is added to the original signals.
where
is the Gaussian white noise weighting factor and
is the Gaussian white noise.
Next, EMD decomposition of the sequence
is performed, and the average value of the first modal components obtained from the decomposition is taken as the first-order component of CEEMDAN.
where
(Intrinsic Mode Function) denotes the first modal component obtained from the CEEMDAN decomposition and
denotes the residual signal after the first decomposition.
After introducing specific Gaussian white noise to the
th order residual signal obtained from the decomposition process, the EMD decomposition is further applied to derive a new component and residual signal.
where
denotes the
th order modal component;
denotes the
th
component after EMD decomposition;
denotes the weight coefficient of the noise added by CEEMDAN for the (
th iteration; and
denotes the
th stage’s residual signal.
The decomposition process of the CEEMDAN algorithm concludes when the residual signals obtained from EMD become monotonic or meet other predetermined criteria.
The flowchart of CEEMDAN decomposition is shown in
Figure 1.
2.2. Fuzzy Entropy
As nonlinear technologies evolve, many nonlinear dynamic methods based on estimations of statistical parameters have been applied to extract fault characteristics. The entropy features derived from information entropy serve as effective tools for characterizing nonlinear properties, which primarily include sample entropy [
20], approximate entropy [
21], and fuzzy entropy [
22], among others. Approximate entropy offers a framework for analyzing the complexity of finite time series; however, it is a statistical measure that quantifies the regularity of a time series, which is characterized by poor statistical stability [
23]. In instances where the time series is excessively brief, sample entropy utilizes a jumping self-similarity function to measure the complexity of the time series, resulting in inaccurate estimated values in real applications [
24]. Sample entropy has lower calculation efficiency, especially for long time series [
25], and may yield inaccurate entropy estimates or result in undefined entropy values [
26]. Fuzzy entropy is a standardized metric for evaluating the uncertainty and complexity of fuzzy sets. It has been demonstrated that fuzzy entropy surpasses sample entropy in several respects, particularly due to its enhanced robustness to noise and its greater suitability for the analysis of short and noisy time series [
27].
Fuzzy entropy is derived from sample entropy, utilizing the degree of membership of elements within a fuzzy set as the probability density function to compute the entropy value. Subsequently, the fuzzy entropy value is determined in accordance with the principles of information entropy. Like approximate entropy and sample entropy, fuzzy entropy aims to quantify the likelihood of new patterns emerging within a sequence. Specifically, a higher fuzzy entropy value signifies an increased probability of the emergence of new patterns, thereby indicating a greater complexity within the sequence. This attribute renders fuzzy entropy a vital instrument for the analysis of uncertainty and complexity in the dynamic evolution of complex systems. The fundamental principles underlying fuzzy entropy are outlined below:
- (1)
Perform phase space reconstruction on a set of time series
of length
.
- (2)
Define the distance between
and
.
- (3)
Introduce fuzzy affiliation degree to measure the similarity between
and
.
- (4)
- (5)
Similarly, define the
function.
- (6)
If the length
of the dataset is finite, the
function is shown below.
2.3. Hippopotamus Optimization Algorithm (HO)
HO is a swarm intelligence optimization algorithm inspired by the social behavior of hippopotamuses, as proposed by Mohammad Hussein Amiri et al. [
18] in 2024. This algorithm seeks to identify optimal solutions to optimization problems by simulating the positional update, defensive strategies against predators, and evasive maneuvers of hippopotamuses in aquatic environments such as rivers or ponds. The optimization process is delineated as follows.
- (1)
Population initialization
The population of hippopotamuses can be mathematically represented using a matrix. Each hippopotamus’s position corresponds to a potential solution, while the updates to its position reflect the values of the decision variables.
where
and
denote the lower and upper bounds and
is a random number in the range of 0 to 1.
- (2)
The hippopotamus’s position is updated in the river or pond (Exploration).
Hippo herds consist of female hippos, immature hippos, male hippos, and dominant male hippos. Adult males may be expelled from the herd upon reaching maturity. Equation (15) describes the position of male hippos.
where
is an integer between 1 and 2,
is a random number between 0 and 1,
denotes the position of the dominant hippopotamus.
Equations (16) and (17) describe the position of the female or immature hippos in the herb.
where
and
are numbers or vectors randomly selected from the five scenarios in the
Equation (18);
is an integer between 1 and 2;
refers to the mean values of some randomly selected hippopotamus with an equal probability of including the current considered hippopotamus;
is a random vector between 0 and 1,
,
and
are random numbers between 0 and 1; and
and
are integer random numbers that can be one or zero.
- (3)
Hippopotamus defense against predators (Exploration).
The defensive mechanism employed by hippos when confronted with a threat involves turning aggressively to confront the predator and producing a loud vocalization.
where
represents a random vector ranging from zero to one.
Equation (20) represents the position of the predator and Equation (21) represents the distance from the hippopotamus to the predator. The mathematical model of this behavior is shown in Equation (22).
where
is a random vector with levy distribution,
is a uniform random number between 1 and 1.5 and
is a uniform random number between 2 and 3,
is a uniform random number between 2 and 4,
represents a uniform random number between −1 and 1, and
is a random vector with dimensions
.
- (4)
Hippopotamus escaping from the predator (Exploitation).
When hippos encounter a group of predators or are unable to effectively repel the threat through defensive measures, they adopt an escape strategy and opt to vacate the current area. The positional update is shown as follows:
choose randomly from three scenarios:
where
and
are random numbers between 0 and 1,
denotes a random vector between 0 and 1, and
is a normally distributed random number.
2.4. Improved Hippopotamus Optimization Algorithm (IHO)
2.4.1. Tent Chaotic Mapping
The chaotic properties, characterized by randomness, ergodicity, and extreme sensitivity to initial conditions, present opportunities for enhancing convergence in algorithm design. Specifically, the ergodicity of sequences produced by chaotic systems guarantees that all states can be explored without repetition within a specified range, which constitutes a significant advantage in the optimization search process. However, the chaotic sequences generated by traditional logistic mapping exhibit a high probability of assuming values at both extremes of the sequence. This results in an uneven distribution of values, which may constrain the efficiency of the search process.
In contrast, Tent mapping [
28] is capable of generating chaotic sequences that exhibit a more uniform distribution within the interval [0, 1]. This characteristic is particularly significant for the initial population construction in optimization algorithms. Utilizing tent mapping for population initialization ensures that the initial solutions are distributed as evenly as possible throughout the solution space. This approach enhances both the comprehensiveness and efficiency of the search process, thereby establishing a solid foundation for subsequent optimization efforts.
The initialization of the hippo population involves the use of chaotic tent sequences. The specific steps can be summarized as follows.
- (1)
Use the mathematical expression of Tent chaotic mapping to generate chaotic sequences. Tent mapping is a simple nonlinear dynamical system whose mathematical expression is shown in Equation (26). The value of
influences the distribution of chaos, which subsequently impacts the generation of the initial population. The value of
is 0.5. When
a = 0.5, it is evenly distributed, and the chaotic sequence also exhibits an even distribution. As different parameters change, it maintains a stable distribution density, allowing for the generation of a robust initial population.
- (2)
Transform the generated chaotic sequence into the range of the hippo population search space, where
is the generated Tent chaotic sequence.
2.4.2. Adaptive Weighting Factor
Inspired by the particle swarm optimization algorithm [
29], the concept of weight is incorporated into the process of updating positions. Traditional methods that utilize a linear weight factor frequently result in suboptimal search outcomes and constrained optimization capabilities, primarily due to their inherent limitations and the constraints imposed during the search process. To address this deficiency, we propose a novel adaptive weight factor designed to dynamically modulate the balance between exploration and exploitation within the algorithm. The formulas for updating the position of the hippo and the corresponding weight factor are articulated as follows:
where
is the maximum number of iterations and
is the current number of iterations. Parameters 5 and 1.7 in the equation are the optimal values determined through extensive experimental simulations. The exponential decay model offers a smooth and gradual method for change. In contrast to linear decay or other abrupt transitions, exponential decay aligns more closely with the natural progression required in the optimization process.
In the initial stage, the large weight factor enables the algorithm to conduct extensive explorations within the search space. In the later stage, the smaller weight factor allows the algorithm to concentrate on promising regions, thereby accelerating convergence. A large initial weight can enhance the diversity of solutions and prevent the algorithm from becoming trapped in local optima during the early stages of optimization. Additionally, the exponential decay method ensures that the weight factor does not abruptly decrease to a minimal value, thereby preserving a degree of exploratory capability throughout the entire optimization process.
Figure 2 illustrates the value of the weight factor
w.
2.4.3. Adaptive Mutation Perturbation
In the third stage of the hippo algorithm, when hippos evade predators, the position update may encounter the challenge of local optima due to the constraints imposed by the random factor
. To address this issue, this paper introduces a perturbation mechanism that combines Cauchy variation and Gaussian variation. Each time the population is updated, the individuals are mutated and perturbed based on the iterative stage. The adaptive mutation strategy can dynamically adjust the ratio of Cauchy to Gaussian variations based on the different stages of the optimization process or the quality of the current solution. In the early stages, the proportion of Cauchy variants is increased to enhance exploration capabilities. In the later stages, Gaussian mutation is emphasized to improve development efficiency.
In the initial phase of the algorithm’s iteration, that is, when
is relatively small, we employ Cauchy mutation [
30,
31] as the primary mode of mutation and assign it a significant weight. The Cauchy distribution is recognized for its long-tail characteristic, which enables the generation of extreme values that are distant from the mean. This property allows for larger mutation steps within the population.
As the number of iterations increases, the weight of the Cauchy mutation is gradually diminished while the weight of the Gaussian mutation [
32,
33,
34] is correspondingly increased. Gaussian mutation is recognized for its capacity to fine-tune searches around the mean value, thereby facilitating a more in-depth and precise exploration within the range of candidate solutions that have been examined.
The flowchart for the improved hippo optimization algorithm are presented in
Figure 3.
2.5. Support Vector Machine (SVM)
As a powerful classification and regression model, SVM has been widely used across various domains. Ali S M et al. [
35] introduced an automated method for detecting valve leakage based on acoustic emission parameters and SVM, achieving an accuracy rate exceeding 98%. Li Z et al. [
36] employed kernel principal component analysis (KPCA) in conjunction with SVM classifiers to ascertain leakage levels, attaining an accuracy rate of 96.75%. Ni L et al. [
37] utilized particle swarm optimization combined with SVM (PSO-SVM) for the intelligent detection of water supply pipeline leaks, demonstrating superior performance compared to backpropagation neural networks (BPNN). Guo et al. [
38] proposed an SV-WTBSVM method for intelligent water supply pipeline leakage detection. Their findings indicated that this algorithm not only preserved the rapid training speed characteristic of TBSVM but also enhanced both classification accuracy and generalization capability. Wang et al. [
39] suggested the application of a convolutional neural network (CNN) model for fault feature extraction and dimensionality reduction, subsequently inputting these features into an SVM model for diagnostic purposes. The CNN-SVM model exhibited superior accuracy relative to other classification models. Originally, SVM was developed for binary classification challenges. However, when addressing multi-class classification issues, it becomes essential to devise an appropriate multi-class classifier. Several strategies for multi-class classification can be employed, including “one-to-one” or “one-to-many”, both of which aim to convert the multi-class classification problem into a sequence of binary classification tasks.
When addressing challenges associated with SVM, it is important to note that if the data are not linearly separable, different kernel functions can be employed in combination to classify nonlinear data. These kernel functions facilitate the mapping of data into a high-dimensional feature space, thereby rendering the data linearly separable A selection of commonly utilized kernel functions in SVM is presented in
Table 1.
Among the various kernel functions, the RBF kernel is one of the most frequently employed. In comparison to other kernel functions, the RBF kernel function exhibits a superior capacity for nonlinear feature mapping, increased model complexity, and the ability to accommodate more intricate models across a broader spectrum of applications. It is characterized by two significant input parameters:
: The regularization parameter, which serves as an indicator of the model’s tolerance to error. A higher value of this parameter indicates a lower tolerance for errors, which may result in overfitting. Conversely, a value that is too low can lead to underfitting. Both excessively large and small values of the regularization parameter can adversely affect the model’s generalization capability.
: The kernel function parameter, which plays a crucial role in determining the distribution of data once they have been mapped into a new feature space. A low value for this parameter tends to result in a greater clustering of data points, whereas a high value necessitates that points should be in close proximity to one another to be classified as belonging to the same group. This latter scenario frequently contributes to the phenomenon of overfitting.
Consequently, to acquire suitable parameters and enhance the classification capability of the model, IHO is employed to optimize the two input parameters and of SVM.
2.6. SVM Parameter Optimization Based on IHO
In this paper, IHO was employed to optimize the parameters and of SVM. The specific steps involved in this process are outlined as follows:
- (1)
Initialize the parameters for IHO. Specify the population number , the maximum number of iterations , and the upper and lower boundary parameters;
- (2)
Select the fitness function. Use the results of five-fold cross-validation on the training set to assess the model’s performance and utilize it as the fitness function;
- (3)
Initialize the population using Tent chaotic mapping;
- (4)
Calculate the initial fitness value of the population and sort to identify the optimal fitness value;
- (5)
According to the formulas pertinent to the corresponding exploration stage, the fitness value of the new position of the hippo is calculated. This value is then compared with the optimal fitness value from the previous iteration to update the position of the optimal individual within the hippopotamus population;
- (6)
Determine whether the maximum number of iterations has been reached. If this condition is met, the optimal value should be outputted and the algorithm should be terminated. If not, then revert to step (5) to continue the iterative loop;
- (7)
The optimal regularization parameter and kernel parameter are determined and subsequently input into the SVM model for the diagnostic classification of the test set.
4. Conclusions
In order to accurately classify valve internal leakage faults, a diagnostic model based on the Improved Hippo algorithm-optimized Support Vector Machine (IHO-SVM) is proposed. The HO was improved by using three strategies, and the search efficiency and performance of the algorithm were improved. The superior performance of the IHO algorithm was demonstrated by testing twelve standard test functions. Then, the IHO algorithm was used to optimize the parameters of SVM in the diagnosis of valve internal leakage. Compared with the SVM optimized by HO, GWO, PSO, WOA, SSA, and the traditional SVM model, in terms of accuracy, the IHO-SVM model had a higher classification accuracy and classification effect. Comparing the accuracy of HO-SVM, IHO-SVM, PSO-SVM, GWO-PSO, WOA-SVM, and SSA-SVM running independently 10 times, the IHO-SVM model was, again, superior in terms of its high classification accuracy and good stability.