Next Article in Journal
Incorporating Cellular Stochasticity in Solid–Fluid Mixture Biofilm Models
Previous Article in Journal
Mean Shift Cluster Recognition Method Implementation in the Nested Sampling Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Coded Permutation Entropy: A Measure for Dynamical Changes Based on the Secondary Partitioning of Amplitude Information

School of Physics and Information Technology, Shaanxi Normal University, Xi’an 710119, China
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(2), 187; https://doi.org/10.3390/e22020187
Submission received: 21 December 2019 / Revised: 31 January 2020 / Accepted: 3 February 2020 / Published: 6 February 2020
(This article belongs to the Section Entropy and Biology)

Abstract

:
An improved permutation entropy (PE) algorithm named coded permutation entropy (CPE) is proposed in this paper to optimize the problems existing in PE based on the secondary partitioning. The principle of CPE algorithm is given, and the performance of it for dynamical change detection is analyzed using synthetic signal, logistic map and Lorenz map. The detection ability of CPE algorithm in different signal-to-noise ratios (SNR) is studied and the algorithm complexity is discussed. The results show that CPE can accurately capture minor feature information and amplify the detection results of dynamical changes compared with PE, weighted permutation entropy (WPE) and amplitude-aware permutation entropy (AAPE), but it has less robustness to noise and requires a higher computation cost than the others. Finally, we use the new algorithm to analyze the rolling bearing fault signals. The application of actual signals illustrates that CPE performs better in detecting abnormal pulse of the rolling bearing when the embedded dimension is small. From all the analyses in this paper, we find that CPE has a better performance for dynamical change detection compared with the other three algorithms when there is a larger repetition rate of permutation pattern in the position sequences.

1. Introduction

Detecting the dynamical changes of complex systems and distinguishing the complexity of output time series offer great practical significance in physics, biomedicine, engineering and economics. Various information-theoretic methods have been developed to measure complexity, such as entropy analysis [1,2,3,4,5,6,7], Lyapunov exponents [8,9], statistical complexity measure [10,11], correlation dimension [12], and symbolic dynamics [13]. Among these methods, entropy analysis is widely used in many fields because its principle is simple, and it can reflect the characteristics of the system effectively.
In 2002, Bandt and Pompe proposed permutation entropy (PE) to measure the natural complexity of time series [4]. After that, this algorithm has attracted the attention of many researchers because it has superior robustness and requires less computation [14,15,16]. Many derivative algorithms, such as multiscale permutation entropy [17,18], weighted permutation entropy (WPE) [19,20,21], Rényi permutation entropy [22,23], generalized permutation entropy [24], multivariate permutation entropy [25], and amplitude-aware permutation entropy (AAPE) [26] have been proposed to extend the application of PE to a variety of fields.
PE algorithm is based on the probability distribution of the permutation pattern obtained by calculating the frequency of each permutation pattern in all position sequence. However, it does not mean that the reconstructed component is completely consistent before ascending order, although they have the same positions sequence [19,25,26,27,28]. First, if the position sequence is ‘1,2,3’, the original data that is indicated by ‘2’ may be located in three different areas [19,25]. Secondly, it may have the same position sequence when a constant value is added to the original data as shown in Reference [26]. Thirdly, it also has the same position sequence when equivalent elements appear in the original data [27,28]. In PE algorithm, all the above situations are marked as the same permutation pattern. Therefore, PE is not accurate when it is used to divide those cases. This is because only counting the frequency of occurrence of each permutation pattern can result in the loss of much detailed information. To solve the first, Fadlallah et al. in 2013 proposed WPE which introduced the amplitude information into PE algorithm using variance to represent the probability [19,20,21]. However, WPE does not solve the problem when a constant value is added to the original data because its variance does not change any more. In 2016, Azami et al. proposed amplitude-aware permutation entropy (AAPE) to settle the second and the third by adding a variable contribution parameter A, which is an adjusting coefficient related to the mean value and the difference between consecutive samples to make the algorithm more flexible [26]. The simulation results show that WPE and AAPE greatly improve the detection ability of amplitude changed signal compared to PE. However, dynamical changes do not only exist in terms of amplitude, and system complexity change detection is also very common. Based on this point, it is not necessarily advantageous to use WPE or AAPE in detecting dynamical changes of system complexity compared to PE.
In addition to the above problem, another question for PE in capturing the characteristic information of the data is that the selection of embedding dimension m is excruciating. If m is small, the reconstructed component will contain few distinct patterns, and it is detrimental to the extraction of abnormal feature [29,30,31]. For example, the potential permutation pattern of m = 3 is 3 ! = 6 , and the detection result of dynamical changes when m > 3 is better than m = 3 [4,24]. Considering in this way, using a large value of m is fine. However, the larger m is, the larger the data length N in PE algorithm will be needed. The relationship between m and N is that N should be considerably larger than m ! [4,22]. In some applications, the length of data is not always sufficient.
The main reason for the above two problems is that it is too rough to calculate the probability only using the permutation pattern as the partitioning criterion in PE algorithm. In Reference [25], He et al. proposed a concept which has the intention of secondary partitioning, and they applied their idea in the multivariate system and presented multivariate permutation entropy (MvPE). In their concept, each permutation pattern in PE is divided into three sub-patterns according to where the intermediate element is. However, MvPE is used to measure the complexity of multivariate systems for embedding dimension m = 3 , and some difficulties prevent its application when m > 3 . In this paper, we make a secondary partitioning, which improving the accuracy of partitioning based on permutation pattern and overcoming the incompatibility of MvPE algorithm for m, and propose coded permutation entropy (CPE).
The paper is organized as follows. In Section 2, CPE algorithm is explained. Dynamical change detection and algorithm complexity are presented in Section 3. We analyze the dynamical detection of three rolling bearing fault signals using CPE algorithm in Section 4. The concluding remarks are recorded in Section 5.

2. Methodologies

2.1. Permutation Entropy

For a scalar sequence { x ( i ) , i = 1 , 2 , , N } , it is embedded into a m-dimension state space as:
X ( j ) = [ x ( j ) , x ( j + τ ) , , x ( j + ( m 1 ) τ ) ] ,
where j = 1 , 2 , , N ( m 1 ) τ , N is the data length of scalar sequence, m and τ denote the embedding dimension and the time delay and m = 3 , , 7 and τ = 1 , 2 , 3 [4].
Then, the reconstructed component X ( j ) is arranged in an increasing order as x ( j + ( k 1 1 ) τ ) x ( j + ( k 2 1 ) τ ) x ( j + ( k m 1 ) τ ) . When an equality occurs, such as x ( j + ( k l 1 1 ) τ ) = x ( j + ( k l 2 1 ) τ ) , x will be permuted according to the size of k l 1 and k l 2 ; specifically, if k l 1 < k l 2 , let x ( j + ( k l 1 1 ) τ ) x ( j + ( k l 2 1 ) τ ) . Therefore, the position sequence which corresponds to the permutation type in Reference [4] can be obtained as follow:
S ( j ) = { k 1 , k 2 , , k m } ,
where k l is the position index of x ( j + ( k l 1 ) τ ) in X ( j ) .
For a certain embedding dimension m, there are m ! potential permutations patterns, and each permutation pattern can be known as F ( v ) , 1 v m ! [4].
Next, we calculate the frequency of each F ( v ) in all S ( j ) , and N u m ( v ) is the count of each F ( v ) . Then, the probability can be marked as
P ( v ) = N u m ( v ) N ( m 1 ) τ , 1 v m ! .
Finally, PE is computed by
H P E ( m ) = v = 1 m ! P ( v ) ln P ( v ) .
It is clear that 0 H P E ( m ) ln ( m ! ) . H P E ( m ) has the maximum value ln ( m ! ) when P ( v ) = 1 m ! . Generally, H P E ( m ) is normalized by ln ( m ! ) as h P E ( m ) = H P E ( m ) ln ( m ! ) .

2.2. Coded Permutation Entropy

The reason WPE and AAPE have a sensitive detection ability regarding the changes in amplitude is that they use specific amplitude information directly in probability calculation. Therefore, they cannot show good performance if the change of amplitude in data is very small whereas prominent changes in system complexity are expected. To solve this problem, we use the initial data as a hidden assistant rather than take it directly, and propose an improved algorithm named CPE. The processes of CPE algorithm are as follows:
Step 1
The steps before getting the position sequence S ( j ) for CPE are the same as those for PE.
Step 2
Obtain the serial number which is the position index of each S ( j ) in all permutation pattern F ( v ) , and record them in column 1 of row j of matrix Q, namely Q ( j , 1 ) .
Step 3
Average the reconstructed components X ( j ) which have the same position sequence, and call the result as the mean sequence M e ( v ) , 1 v m ! .
Step 4
Compare the size of each element in reconstructed component X ( j ) with its corresponding mean sequence M e ( v ) using Equation (5), and record the comparison results in column 1 + k of row j of matrix Q.
Q ( j , 1 + k ) = 2 , X ( j , k ) > M e ( v , k ) 1 , X ( j , k ) = M e ( v , k ) 0 , X ( j , k ) < M e ( v , k ) , j = 1 , 2 , , N ( m 1 ) τ , k = 1 , 2 , , m .
Step 5
A coded sequence matrix Q ( N ( m 1 ) τ , 1 + m ) is constructed and each row of it can be named the coded sequence. Obviously, the number of potential patterns existing in CPE is m ! · 3 m , and so the probability of unique coded sequence is calculated as:
P ( r ) = N u m ( r ) N ( m 1 ) τ , 1 r m ! · 3 m .
Step 6
CPE can be obtained by substituting the probability P ( r ) into the Shannon entropy formula, namely
H C P E ( m ) = r = 1 m ! · 3 m P ( r ) ln P ( r ) .
Similarly, the normalized expression of H C P E ( m ) is h C P E ( m ) = H C P E ( m ) ln ( m ! · 3 m ) . Since PE and CPE are normalized on different base, we use Equation (4) and Equation (7) to calculate the values of PE and CPE in the next analyses.
Based on the above statements about CPE, the flow chart of CPE algorithm is shown in Figure 1. A simple example is presented to illustrate the calculation processes of PE and CPE as shown in Figure 2. The scalar sequence { x ( i ) } of this example is {0.8147, 0.9058, 0.1270, 0.9134, 0.6324, 0.0975, 0.2785, 0.5469, 0.9575, 0.9649} which is the random number drawn from the standard normal distribution. In Figure 2, the position sequence ‘1, 2, 3’ is the fifth in the permutation pattern of m = 3 , and so we set Q ( j , 1 ) = 5 when j = 6 , 7 , 8 which is shown in the bold red line of Figure 2. Due to the reconstructed component X ( 6 ) , X ( 7 ) and X ( 8 ) have the same position sequence ‘1, 2, 3’, the corresponding mean sequence can be recorded by calculating the mean of X ( 6 ) , X ( 7 ) , X ( 8 ) , namely M e ( 5 ) = [ 0.3076 , 0.5943 , 0.8231 ] . Then we compare the size between X ( 6 ) , X ( 7 ) , X ( 8 ) and M e ( 5 ) respectively, and get the coded sequences Q ( 6 , 2 : 4 ) , Q ( 7 , 2 : 4 ) and Q ( 8 , 2 : 4 ) (we only present the specific process of obtaining Q ( 8 , 2 : 4 ) as shown in the bold blue line of Figure 2, and Q ( 8 , 2 : 4 ) = [ 2 , 2 , 2 ] means the value of column 2 through 4 of row 8 in matrix Q(8,4) is 2, 2 and 2, respectively). It is clear that CPE successfully divides the reconstructed component X ( 6 ) , X ( 7 ) , X ( 8 ) into three classes by the process of the secondary partitioning which is circled by a black box in Figure 2. In particular, CPE will degenerate to PE when the probability P ( r ) in Equation (6) is presented by the frequency of the same serial number in Q ( j , 1 ) according to Figure 1 and Figure 2. The same procedures can be used to achieve the secondary partitioning for m > 3 .

3. Simulation Analyses

The number of potential permutation pattern in PE, WPE and AAPE is m ! , and the pattern in CPE is m ! · 3 m due to the introduction of the secondary partitioning. Because of that, we can expect CPE to be more significant than PE, WPE and AAPE in the measure of dynamical changes. Three models, which are generally used in dynamical change detection, are built to verify the performance of CPE algorithm, namely synthetic signal, discrete standard model and continuous standard model. In addition to presenting the corresponding entropy graph, the standard deviation (Std) which is the most common formula to reflect the discreteness of a set of data is also calculated to show the effect of dynamical change detection from the perspective of quantitative analysis. Moreover, the experiments in the case of noise and the algorithm complexity are also analyzed.

3.1. Synthetic Signal

A synthetic signal y ( t ) is produced using the following equation:
y ( t ) = sin ( 2 π · t ) + sin ( 2 π · 10 · t ) 0 < t 10 sin ( 2 π · t ) + sin ( 2 π · 10 · t ) + sin ( 2 π · 20 · t ) 10 < t 20 2 · sin ( 2 π · t ) + 1 . 5 · sin ( 2 π · 10 · t ) + 1 . 25 · sin ( 2 π · 20 · t ) 20 < t 30
where the sampling frequency is 1000 Hz and the sampling time is 30 s. The curve of synthetic signal y ( t ) is shown in Figure 3.
Figure 4a–d show the detection results of PE, WPE, AAPE and CPE for the above time series y ( t ) when m = 3 , τ = 1 , m = 4 , τ = 1 , m = 5 , τ = 1 and m = 6 , τ = 1 , respectively (The value of parameter A existing in AAPE algorithm is 0.5 in all experiments in this paper). A 2000-sample sliding window with the increment of 1 is used in this computation. From Figure 4, we see that the entropy values of PE, WPE, AAPE and CPE increase suddenly at t = 10 s (Jump1), and then it decrease at t = 20 s (Jump2). The reason is that y ( t ) is more complex between 10 s and 20 s, so the corresponding entropy value is larger than those of 0 s–10 s. The only difference for y ( t ) between 10 s–20 s and 20 s–30 s is the amplitude which cause the decrease of entropy value at t = 20 s. Whatever m are, the jumps of CPE are the most obvious among than PE, WPE, and AAPE not only for t = 10 s but also for t = 20 s. Moreover, the Std and the jumps of entropy value of PE, WPE, AAPE and CPE for synthetic signal y ( t ) when m takes on different values are recorded in Table 1. It is clear to find that the Std and the jumps of CPE are the largest in these four algorithms whatever m are. These two aspects indicate that CPE algorithm has some advantages not only for the complexity dynamical change detection but also for the amplitude.

3.2. Discrete Standard Model

A transient logistic map can be presented as [29]:
x i + 1 = r · x i · ( 1 x i ) .
In this example, x 0 = 0.65 , r ( 0 ) = 3.5 and r is consistently increasing to r = 4 by a step of 0.001. For each value of r, a time series x with a length of 3000 is generated. To discard transients, we only use the last 2000 samples of this time series, as shown in Figure 5 [24].
Figure 6 shows the results of PE, WPE, AAPE, and CPE for the dynamical change detection of the logistic map when m are set as different values. From Figure 6, we find that PE, WPE and AAPE have a similar detection result for the chaotic window in the logistic map whatever the value of m are, and CPE decreases more than those of other three algorithms in the chaotic window. Although WPE was proposed to consider the amplitude difference of the same position sequences, it does not significantly improve the detection ability of dynamical changes of chaotic windows in the logistic map compared to PE. We also observe that PE, WPE and AAPE have no prominent decrease when r = 3.628 whether it is m = 3 or m = 4 . However, there is a significant entropy decrease in CPE, which is marked by a red arrow in Figure 6a,b. Table 2 presents the corresponding Std and the jumps of entropy value for the logistic map when m = 3 , 4 , 5 , 6 . Consistent with the result of entropy graph, the Std and the jumps of CPE are still the largest in these four algorithms. It can be concluded that CPE amplifies the detection results of dynamical changes in the logistic map, and has a weak dependence on parameter m when we hope to observe the dynamical changes accurately.

3.3. Continuous Standard Model

In the above simulation, we analyzed the ability of CPE to detect the dynamical changes for a discrete time system. Next, we examine a continuous time system described by the following transient Lorenz equations [29]:
d x d t = 10 ( x y ) d y d t = x z + r ( t ) x y d z d t = x y 8 3 z .
The system is solved using a fourth-order Runge–Kutta method with a time step Δ t = 0.01 . In our simulation, x ( 0 ) = 1 , y ( 0 ) = 0 , z ( 0 ) = 1 , r ( 0 ) = 28 and r ( t ) increases to r ( t ) = 268 by a step of 0.2. For each r ( t ) , we obtain a series with a length of 3000. It is consistent with the logistic map that we only take the last 2000 data to calculate the entropy, and the transient Lorenz map is shown in Figure 7. Theoretically, there are three periodic windows, which existing in 99.524 < r ( t ) < 100.795 , 145 < r ( t ) < 166 and r ( t ) > 214.4 , as indicated by dashed vertical lines in Figure 7 [32].
The choices of both m and τ are related to the time step Δ t . Based on this point, we set m = 3 , τ = 10 , m = 4 , τ = 10 , m = 5 , τ = 10 and m = 6 , τ = 10 [32], and the corresponding entropy value curves are presented in Figure 8a–d, respectively. It is consistent with the logistic map that PE, WPE and AAPE have similar detection results when a periodic window appears in the transient Lorenz map. According to Figure 8, we observe that all these curves of CPE capture the periodic windows at r ( n ) = 166 (Jump1) very well whatever m = 3 or m = 4 . The change existing in r ( n ) = 230.4 (Jump2) also can be detected using CPE when m = 4 rather than m = 6 for PE, WPE and AAPE. It is very beneficial for dynamical change detection if we can produce good results when m is small. Furthermore, the corresponding Std and the jumps of CPE also reflect these performances in complexity dynamical change detection, which is recorded in Table 3.
For PE, Bandt and Pompe recommend m = 3 , , 7 [4]. However, the number of F ( v ) in m = 3 and m = 4 is 6 and 24, hence, it is not very large, which is the reason that m = 3 , 4 is not the best choice for PE algorithm to detect the dynamical changes [29]. WPE takes into account the amplitude differences between the same position sequences, and it improve the performance of PE for the detection of abnormal amplitude change [19]. However, for the above dynamical change detection in system complexity, WPE and AAPE do not demonstrate better detection ability than PE. Both for chaotic windows detection of the logistic map and periodic windows detection of the Lorenz map, CPE shows a superior detection result compared to PE, WPE and AAPE. This result illustrates that the secondary partitioning is very useful in extracting the feature information. In particular, it can greatly improve the detection performance of PE when m = 3 , 4 .

3.4. Robustness to Noise

To study the dynamical change detection performance of these four algorithms at different signal-to-noise ratios (SNR), we add white Gaussian noise to the two series, and the power of them is expressed in dB. The data length of these two series are 20000, and they produced by Equation (9) when set r ( n ) = 3.961 and r ( n ) = 3.962 , respectively. In the logistic map, x ( n ) is a periodic series when r ( n ) = 3.961 and it is a chaotic series when r ( n ) = 3.962 .
The 30 statistical values of PE, WPE, AAPE, and CPE for different SNR when m = 3 , τ = 1 are presented in Figure 9 using errorbar [33]. As we all know, it is certain that the entropy value will decrease with the increase of SNR at step of 1 from SNR = −5. From Figure 9, we find that the entropy value of the periodic series decrease more than the chaotic series with the increasing of SNR from −5 to 25, and it is consistent with the detection result of logistic map without noise as show in Figure 6. Furthermore, either for the periodic series or for the chaotic series, the entropy value of CPE decreases the most from the perspective of absolute values in these four algorithms. As shown in Figure 9, with the increasing of SNR, CPE starts distinguishing these two series when the SNR close to 13, and it is indistinguishable before that. The corresponding SNR for PE, WPE, and AAPE is 10, 9 and 11, respectively. These results show that CPE has less robustness to noise than PE, WPE, and AAPE.

3.5. The Complexity of Four Algorithms

The complexity of four algorithms is discussed in this subsection. All the experiments are run in MATLAB R2014a on the same computer, with a CPU model number of i5-7500 and RAM size of 16 GB. The running time (per second) of four algorithms for the above three models when m has different values are presented in Table 4. From Table 4, we find that the running time is increasing as the increase of m for all these four algorithms. When m is constant, the running time of PE algorithm is the shortest for these three signals. For m = 3 , the running time of WPE is slightly longer than CPE. When m = 4 , 5 , 6 , the running time of the four algorithms is sorted as PE, AAPE, WPE, and CPE from small to large basically. The running time of four algorithms for logistic map when m = 6 is an exception. The reason for WPE and CPE require more running time is that WPE adds the step of calculating the weighted value [19], and CPE adds the steps of computing the mean sequence and comparing the size of the corresponding element. Furthermore, we calculate the growth rate (Gr) of running time of CPE referring to PE as shown in Table 5, and ‘Average’ in Table 5 means the average of Gr for three signals under the same m. Here, Gr can be got by G r = ( T C P E T P E ) · 100 % T P E , where T C P E means the running time of CPE and T P E represents the running time of PE. The Gr of running time of CPE to PE is 121% if we take the mean of ‘Average’ at different m. A similar operation can be used to calculate the Gr of CPE referring to WPE and AAPE, and the corresponding mean value of ‘Average’ is 25% and 72% respectively.
The running time of algorithm is closely related to the other factors such as computer configuration. Therefore, we also list the time complexity of four algorithms, which is shown in Table 6. It is evident to see that the time complexity can be sorted as PE, AAPE, WPE and CPE from small to large, which is basically consistent with the conclusion obtained by running time. Therefore, both from the perspective of running time and time complexity of algorithm, the computation cost of CPE is the largest in these four algorithms.

4. Apply CPE to Rolling Bearing Fault Detection

Rolling element bearings are one of the most prevalent components in rotating machines, and the failure of the bearing is the most frequent reasons for machine breakdown. To verify the effectiveness of CPE algorithm in actual application, the fault data of rolling bearing from Case Western Reserve University are used [34,35,36], and they are inner race fault (IRF), ball fault (BF) and outer race fault (ORF). The sample frequency of them is 12,000 Hz, and the pulse in these signals are considered to be the dynamical changes of actual data [37]. In this experiment, the samples of 0–0.5 s, 0–10 s and 0–0.5 s in these signals are employed as the test data, the window length w is set to 200 samples when analyze IRF and ORF signals and it is equal to 500 for BF signal, and the move step is 1 for these three bearing signals.
The IRF signal is plotted in Figure 10a. Figure 10b–d present the corresponding entropy curves for different m. From Figure 10, it is obvious that the entropy value has a sudden drop when a pulse appeared and the low value lasts for a window length, and the decrease of CPE is the most significant compared with PE, WPE and AAPE when m = 3 and m = 4 . However, when m = 5 , CPE has less pulse detection ability than WPE. The reason is that there are few reconstruct components with the same permutation pattern for w = 200 when m = 5 , and so the advantage of CPE are not prominent in this case. From the aspect of quantitative analysis, when m = 3 and m = 4 , the detection result of CPE are more evident than that by PE, WPE and AAPE because the Std of CPE is larger than the others as shown in Table 7.
For BF signal which is shown in Figure 11a, the entropy curves are presented in Figure 11b–d when m = 3 , 4 , 5 and τ = 1 . The detection performance of CPE algorithm for BF signal are basically consistent with those for IRF signal when m = 3 and m = 4 . The detection difference of CPE for IRF and BF signals is that the result of the latter is obvious when m = 5 according to Figure 10d and Figure 11d. As shown in Table 8, the Std at different m can also illustrate the above two aspects.
Figure 12 presents the amplitude diagram of ORF signal and corresponding entropy curves of four kinds of algorithms at different m. When a fault pulse occurs, the entropy curve of CPE has a slight drop and then a low entropy of one window length appears. The reason is that the pulse intensity of ORF signal is higher than that of IRF signal. So, this different performance of CPE algorithm for IRF signal and ORF signal can help scholars distinguish these two faults better than WPE. Moreover, the four algorithms have basically the same performance for the pulse detection of IRF signal and ORF signal under the same parameters. Whether from Figure 12 or Table 9, the detection performance of CPE is optimal when m = 3 , 4 , and WPE seems to perform better when m = 5 .
Compared with the detection results of simulation data in Section 3, CPE does not seem to improve the detection performance much more than the other three algorithms for the actual data when m = 5 . In theory, for the same embedding dimension m, the larger the value of window length w, the larger the repetition rate of each permutation pattern in all position sequence, and so the advantage of CPE in dynamical change detection will be prominent. To illustrate this point from the perspective of simulation analysis, we employ the BF signal when t = 5 10 s . In this experiment, w is equal to 500, 800 and 1000 respectively under the condition of m = 5 , and the corresponding Std are presented in Table 10. From Table 10, the Std of CPE is increasing with the increase of w when m = 5 , and so the result supports the above theory.
The value of w in three models of Section 3 is equal to 2000, but it is equal to 200 or 500 for the actual signal. Therefore, the probability of appearing the same permutation pattern in all position sequence for the three models is larger than those of the actual signals when m = 5 . Therefore, the detection performance of CPE for the dynamical changes in the three models is more prominent than those for the actual signal when m = 5 . It is obvious that CPE has a better performance for dynamical change detection compared with the other three algorithms when there is a larger repetition rate of permutation pattern in position sequences.

5. Conclusions

In this paper, an improved algorithm named CPE is proposed to solve the problems existing in PE by employing the secondary partitioning. A synthetic signal, a discrete standard model and a continuous standard model are applied to verify the limitations of PE, WPE and AAPE, and illustrate the advantage of CPE. The dynamical change detection performance of these four algorithms are also studied when the signal contains the noise, and then the algorithm complexity is analyzed from the running time and time complexity. Finally, these four algorithms are used to analyze the detection performance of the actual signals. Simulation results show that CPE can amplify the results of dynamical change detection compared to PE, WPE and AAPE, whether for the synthetic signal, logistic map or Lorenz map. In particularly, CPE shows extremely accurate feature fetching for the logistic map and the Lorenz map when m = 3 or m = 4 . However, CPE has less robustness to noise and requires a higher computation cost than PE, WPE and AAPE. The application of CPE in actual signals illustrate that CPE performs better than the other three algorithms in detecting abnormal pulse of rolling bearing when m is small. From all the analyses in this paper, we find that CPE has a better performance for dynamical change detection compared with the other three algorithms when there is a larger repetition rate of permutation pattern in position sequences. CPE is an extremely useful abnormal detection method, and may have more practical applications.

Author Contributions

Conceptualization, H.K. and X.Z.; methodology, H.K. and X.Z.; software, H.K.; validation, H.K. and X.Z.; formal analysis, H.K.; investigation, H.K.; resources, G.Z.; data curation, G.Z.; writing—original draft preparation, H.K.; writing—review and editing, H.K. and X.Z.; visualization, H.K.; supervision, G.Z.; project administration, G.Z.; funding acquisition, X.Z. All authors have read and agree to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 11674208 and 11874252).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PEPermutation Entropy
WPEWeighted Permutation Entropy
AAPEAmplitude-aware Permutation Entropy
MvPEMultivariate Permutation Entropy
CPECoded Permutation Entropy
StdStandard Deviation
SNRSignal-to-Noise Ratio
GrGrowth Rate
IRFInner Race Fault
BFBall Fault
ORFOuter Race Fault

References

  1. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Richman, J.S.; Randall, M.J. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Lobier, M.; Siebenhühner, F.; Palva, S.; Palva, J.M. Phase transfer entropy: A novel phase-based measure for directed connectivity in networks coupled by oscillatory interactions. NeuroImage 2014, 85, 853–872. [Google Scholar] [CrossRef] [PubMed]
  4. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef]
  5. Li, J.; Ning, X. Dynamical complexity detection in short-term physiological series using base-scale entropy. Phys. Rev. E 2006, 73, 052902. [Google Scholar] [CrossRef] [PubMed]
  6. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 2002, 89, 705–708. [Google Scholar] [CrossRef] [Green Version]
  7. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef] [Green Version]
  8. Wolf, A.; Swift, J.B.; Swinney, H.L.; Vastano, J.A. Determining Lyapunov exponents from a time series. Physica D 1985, 16, 285–317. [Google Scholar] [CrossRef] [Green Version]
  9. Ding, R.; Li, J. Nonlinear finite-time Lyapunov exponent and predictability. Phys. Lett. A 2007, 364, 396–400. [Google Scholar] [CrossRef]
  10. López-Ruiz, R.; Mancini, H.; Calbet, X. A statistical measure of complexity. Phys. Lett. A 1995, 209, 321–326. [Google Scholar] [CrossRef] [Green Version]
  11. Martin, M.T.; Plastino, A.; Rosso, O.A. Generalized statistical complexity measures: Geometrical and analytical properties. Physica A 2006, 369, 439–462. [Google Scholar] [CrossRef]
  12. Rolo-Naranjo, A.; Montesino-Otero, M. A method for the correlation dimension estimation for online condition monitoring of large rotating machinery. Mech. Syst. Signal Process. 2005, 19, 939–954. [Google Scholar] [CrossRef]
  13. Hao, B.; Zheng, W. Applied symbolic dynamics and chaos. Dir. Chaos 1998, 7. [Google Scholar] [CrossRef]
  14. Bandt, C.; Keller, G.; Pompe, B. Entropy of interval maps via permutations. Nonlinearity 2002, 15, 1595–1602. [Google Scholar] [CrossRef]
  15. Bandt, C. Ordinal time series analysis. Ecol. Model. 2005, 182, 229–238. [Google Scholar] [CrossRef]
  16. Rosso, O.A.; Larrondo, H.A.; Martin, M.T.; Plastino, A.; Fuentes, M.A. Distinguishing noise from chaos. Phys. Rev. Lett. 2007, 99, 154102. [Google Scholar] [CrossRef] [Green Version]
  17. Aziz, W.; Arif, M. Multiscale permutation entropy of physiological time series. In Proceedings of the 2005 Pakistan Section Multitopic Conference, Karachi, Pakiston, 24–25 December 2005; pp. 1–6. [Google Scholar] [CrossRef]
  18. Liu, T.; Yao, W.; Wu, M.; Shi, Z.; Wang, J.; Ning, X. Multiscale permutation entropy analysis of electrocardiogram. Physica A 2017, 471, 492–498. [Google Scholar] [CrossRef]
  19. Fadlallah, B.; Chen, B.; Keil, A.; Príncipe, J. Weighted-permutation entropy: A complexity measure for time series incorporating amplitude information. Phys. Rev. E 2013, 87, 022911. [Google Scholar] [CrossRef] [Green Version]
  20. Gao, J.; Chen, K.; Wu, P.; Chen, L.; Lin, P. Leakage detection of pipeline based on weighted-permutation entropy. In Proceedings of the 2017 29th Chinese Control And Decision Conference, Chongqing, China, 28–30 May 2017; pp. 2820–2823. [Google Scholar] [CrossRef]
  21. Xia, J.; Shang, P.; Wang, J.; Shi, W. Permutation and weighted-permutation entropy analysis for the complexity of nonlinear time series. Commun. Nonlinear Sci. 2016, 31, 60–68. [Google Scholar] [CrossRef]
  22. Zhao, X.; Shang, P.; Huang, J. Permutation complexity and dependence measures of time series. EPL 2013, 102, 40005. [Google Scholar] [CrossRef]
  23. Mammone, N.; Duun-Henriksen, J.; Kjaer, T.; Morabito, F. Differentiating interictal and ictal states in childhood absence epilepsy through permutation Rényi entropy. Entropy 2015, 17, 4627–4643. [Google Scholar] [CrossRef] [Green Version]
  24. Xu, M.; Shang, P. Generalized permutation entropy analysis based on the two-index entropic form Sq,δ. Chaos 2015, 25, 053114. [Google Scholar] [CrossRef] [PubMed]
  25. He, S.; Sun, K.; Wang, H. Multivariate permutation entropy and its application for complexity analysis of chaotic systems. Physica A 2016, 461, 812–823. [Google Scholar] [CrossRef]
  26. Azami, H.; Escudero, J. Amplitude-aware permutation entropy: Illustration in spike detection and signal segmentation. Comput. Meth. Prog. Biol. 2016, 128, 40–51. [Google Scholar] [CrossRef] [PubMed]
  27. Bian, C.; Qin, C.; Ma, Q.; Shen, Q. Modified permutation-entropy analysis of heartbeat dynamics. Phys. Rev. E 2012, 85, 021906. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, X.; Wang, Y. Fine-grained permutation entropy as a measure of natural complexity for time series. Chin. Phys. B 2009, 18, 2690–2695. [Google Scholar] [CrossRef]
  29. Cao, Y.; Tung, W.W.; Gao, J.B.; Protopopescu, V.A.; Hively, L.M. Detecting dynamical changes in time series using the permutation entropy. Phys. Rev. E 2004, 70, 046217. [Google Scholar] [CrossRef] [Green Version]
  30. Zhang, J.; Hou, G.; Cao, K.; Ma, B. Operation conditions monitoring of flood discharge structure based on variance dedication rate and permutation entropy. Nonlinear Dyn. 2018, 93, 2517–2531. [Google Scholar] [CrossRef]
  31. Kang, H.; Zhang, X.F.; Zhang, G.B. Generalized phase permutation entropy algorithm based on two-index entropy. Sci. Sin. Inform. 2019, 49, 1205–1216. [Google Scholar] [CrossRef] [Green Version]
  32. Gao, J.B.; Cai, H. On the structures and quantification of recurrence plots. Phys. Lett. A 2000, 270, 75–87. [Google Scholar] [CrossRef]
  33. Bai, Y.; Liang, Z.; Li, X. A permutation Lempel-Ziv complexity measure for EEG analysis. Biomed. Signal Process. 2015, 19, 102–114. [Google Scholar] [CrossRef]
  34. Bearing Data Center of Case Western Reserve University. Available online: http://csegroups.case.edu/bearingdatacenter/pages/12k-drive-end-bearing-fault-data (accessed on 30 May 2019).
  35. Wu, S.D.; Wu, P.H.; Wu, C.W.; Ding, J.J.; Wang, C.C. Bearing Fault Diagnosis Based on Multiscale Permutation Entropy and Support Vector Machine. Entropy 2012, 14, 1343–1356. [Google Scholar] [CrossRef] [Green Version]
  36. Wu, S.D.; Wu, C.W.; Humeau-Heurtier, A. Refined scale-dependent permutation entropy to analyze systems complexity. Physica A 2016, 450, 454–461. [Google Scholar] [CrossRef]
  37. Smith, W.A.; Randall, R. Rolling Element Bearing Diagnostics Using the Case Western Reserve University Data: A Benchmark Study. Mech. Syst. Signal Process. 2015, 64–65, 100–131. [Google Scholar] [CrossRef]
Figure 1. The flow chart of CPE algorithm. The left part of this flow chart is the basic steps, and the right is the key steps of CPE algorithm called the secondary partitioning which aims to obtain the coded sequence matrix Q ( N ( m 1 ) τ , 1 + m ) .
Figure 1. The flow chart of CPE algorithm. The left part of this flow chart is the basic steps, and the right is the key steps of CPE algorithm called the secondary partitioning which aims to obtain the coded sequence matrix Q ( N ( m 1 ) τ , 1 + m ) .
Entropy 22 00187 g001
Figure 2. The processes of calculating PE and CPE for a simple example. For the secondary partitioning in CPE, using the serial number constructs Q ( j , 1 ) as shown in the bold red line, and Q ( j , 2 : 4 ) records the size of the relationship between X ( j ) and M e ( v ) as the bold blue line shown.
Figure 2. The processes of calculating PE and CPE for a simple example. For the secondary partitioning in CPE, using the serial number constructs Q ( j , 1 ) as shown in the bold red line, and Q ( j , 2 : 4 ) records the size of the relationship between X ( j ) and M e ( v ) as the bold blue line shown.
Entropy 22 00187 g002
Figure 3. Synthetic signal y ( t ) .
Figure 3. Synthetic signal y ( t ) .
Entropy 22 00187 g003
Figure 4. Detecting dynamical changes existing in the synthetic signal y ( t ) with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 1 ; (b) m = 4 , τ = 1 ; (c) m = 5 , τ = 1 ; (d) m = 6 , τ = 1 .
Figure 4. Detecting dynamical changes existing in the synthetic signal y ( t ) with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 1 ; (b) m = 4 , τ = 1 ; (c) m = 5 , τ = 1 ; (d) m = 6 , τ = 1 .
Entropy 22 00187 g004
Figure 5. The transient logistic map.
Figure 5. The transient logistic map.
Entropy 22 00187 g005
Figure 6. Detecting dynamical changes existing in the transient logistic map with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 1 ; (b) m = 4 , τ = 1 ; (c) m = 5 , τ = 1 ; (d) m = 6 , τ = 1 .
Figure 6. Detecting dynamical changes existing in the transient logistic map with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 1 ; (b) m = 4 , τ = 1 ; (c) m = 5 , τ = 1 ; (d) m = 6 , τ = 1 .
Entropy 22 00187 g006
Figure 7. The transient Lorenz map.
Figure 7. The transient Lorenz map.
Entropy 22 00187 g007
Figure 8. Detecting dynamical changes existing in the transient Lorenz map with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 10 ; (b) m = 4 , τ = 10 ; (c) m = 5 , τ = 10 ; (d) m = 6 , τ = 10 .
Figure 8. Detecting dynamical changes existing in the transient Lorenz map with different m using PE, WPE, AAPE, and CPE: (a) m = 3 , τ = 10 ; (b) m = 4 , τ = 10 ; (c) m = 5 , τ = 10 ; (d) m = 6 , τ = 10 .
Entropy 22 00187 g008
Figure 9. The performance of four algorithms in distinguishing periodic series and chaotic series under different SNR when m = 3 , τ = 1 : (a) PE; (b) WPE; (c) AAPE; (d) CPE.
Figure 9. The performance of four algorithms in distinguishing periodic series and chaotic series under different SNR when m = 3 , τ = 1 : (a) PE; (b) WPE; (c) AAPE; (d) CPE.
Entropy 22 00187 g009
Figure 10. Detecting abnormal pulse existing in IRF signal with different m using PE, WPE, AAPE, and CPE: (a) The IRF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Figure 10. Detecting abnormal pulse existing in IRF signal with different m using PE, WPE, AAPE, and CPE: (a) The IRF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Entropy 22 00187 g010
Figure 11. Detecting abnormal pulse existing in BF signal with different m using PE, WPE, AAPE, and CPE: (a) The BF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Figure 11. Detecting abnormal pulse existing in BF signal with different m using PE, WPE, AAPE, and CPE: (a) The BF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Entropy 22 00187 g011
Figure 12. Detecting abnormal pulse in ORF signal with different m using PE, WPE, AAPE, and CPE: (a) The ORF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Figure 12. Detecting abnormal pulse in ORF signal with different m using PE, WPE, AAPE, and CPE: (a) The ORF signal; (b) m = 3 , τ = 1 ; (c) m = 4 , τ = 1 ; (d) m = 5 , τ = 1 .
Entropy 22 00187 g012
Table 1. The jumps (Jump1, Jump2), and the Std of PE, WPE, AAPE, and CPE for the synthetic signal in different m. Jump1 is equal to M v 2 minus M v 1 and Jump2 is equal to M v 2 minus M v 3 , where M v 1 , M v 2 and M v 3 represent the mean of entropy value when t = 2 s–10 s, t = 12 s–20 s and t = 22 s–30 s respectively.
Table 1. The jumps (Jump1, Jump2), and the Std of PE, WPE, AAPE, and CPE for the synthetic signal in different m. Jump1 is equal to M v 2 minus M v 1 and Jump2 is equal to M v 2 minus M v 3 , where M v 1 , M v 2 and M v 3 represent the mean of entropy value when t = 2 s–10 s, t = 12 s–20 s and t = 22 s–30 s respectively.
Synthetic Signal m = 3 m = 4 m = 5 m = 6
Jump1Jump2StdJump1Jump2StdJump1Jump2StdJump1Jump2Std
PE0.0810.0010.0370.1690.0080.0750.2560.0150.1130.3540.0290.155
WPE−0.0060.0060.0050.0000.0070.0030.0140.0090.0060.0370.0120.015
AAPE0.0730.0030.0320.1350.0060.0600.1900.0070.0850.2510.0070.112
CPE0.1300.0920.0530.2680.1790.1080.3340.1440.1330.4130.1790.164
Table 2. The jumps when r = 3.702 (Jump1) and r = 3.961 (Jump2), and the Std of PE, WPE, AAPE, and CPE for the logistic map in different m. Jump1 is equal to the entropy value when r = 3.701 minus those when r = 3.702 , and Jump2 is equal to the entropy value when r = 3.96 minus those when r = 3.961 .
Table 2. The jumps when r = 3.702 (Jump1) and r = 3.961 (Jump2), and the Std of PE, WPE, AAPE, and CPE for the logistic map in different m. Jump1 is equal to the entropy value when r = 3.701 minus those when r = 3.702 , and Jump2 is equal to the entropy value when r = 3.96 minus those when r = 3.961 .
Logistic Map m = 3 m = 4 m = 5 m = 6
Jump1Jump2StdJump1Jump2StdJump1Jump2StdJump1Jump2Std
PE0.1370.4450.1070.2180.9520.3510.6110.8770.4780.7641.5240.737
WPE0.1440.2400.0890.2450.8040.3270.6890.7100.4470.8461.3300.699
AAPE0.1320.3750.0980.2190.9370.3560.6550.8500.5510.8211.5030.801
CPE0.3970.8570.3850.5891.4670.6571.0032.0910.8441.2232.7501.106
Table 3. The jumps when r ( t ) = 166 (Jump1) and r ( t ) = 230.4 (Jump2), and the Std of PE, WPE, AAPE, and CPE for the Lorenz map in different m. Jump1 is equal to the entropy value when r ( t ) = 166.4 minus those when r ( t ) = 166 , and Jump2 is equal to the entropy value when r ( t ) = 227 minus those when r ( t ) = 230.4 .
Table 3. The jumps when r ( t ) = 166 (Jump1) and r ( t ) = 230.4 (Jump2), and the Std of PE, WPE, AAPE, and CPE for the Lorenz map in different m. Jump1 is equal to the entropy value when r ( t ) = 166.4 minus those when r ( t ) = 166 , and Jump2 is equal to the entropy value when r ( t ) = 227 minus those when r ( t ) = 230.4 .
Lorenz Map m = 3 m = 4 m = 5 m = 6
Jump1Jump2StdJump1Jump2StdJump1Jump2StdJump1Jump2Std
PE−0.073−0.0080.0780.055−0.0030.1540.3550.0380.3750.9530.1810.587
WPE0.0130.0160.1200.1430.0060.2260.4420.0410.3740.9920.2020.570
AAPE−0.0330.0080.0970.1170.0120.1520.3900.0430.3760.9590.1950.579
CPE0.2990.0330.2460.8090.4140.4791.2860.6600.6931.6050.8520.880
Table 4. The running time of PE, WPE, AAPE, and CPE for the above three models in different m.
Table 4. The running time of PE, WPE, AAPE, and CPE for the above three models in different m.
Running Time (Second)
m = 3 m = 4 m = 5 m = 6
Synthetic signalPE10371668367717534
WPE22932800531519823
AAPE13901945423917950
CPE19253021553420811
Logistic mapPE81653290
WPE293673309
AAPE152471386
CPE213789365
Lorenz mapPE2641134711
WPE81100183765
AAPE4057151712
CPE741624411367
Table 5. The Gr of running time of CPE referring to PE in different m.
Table 5. The Gr of running time of CPE referring to PE in different m.
Gr
m = 3 m = 4 m = 5 m = 6
Synthetic signal86%81%51%19%
Logistic map163%131%68%26%
Lorenz map185%295%229%92%
Average145%169%116%55.5%
Table 6. The time complexity of PE, WPE, AAPE, and CPE algorithms.
Table 6. The time complexity of PE, WPE, AAPE, and CPE algorithms.
AlgorithmTime Complexity
PE O ( N ( m 1 ) τ ) · m !
WPE O ( N ( m 1 ) τ ) · 4 m !
AAPE O ( N ( m 1 ) τ ) · m ! ( m 1 )
CPE O ( N ( m 1 ) τ ) · m ! ( m + 1 )
Table 7. The Std of PE, WPE, AAPE, and CPE for IRF signal in different m.
Table 7. The Std of PE, WPE, AAPE, and CPE for IRF signal in different m.
IRFStd
m = 3 m = 4 m = 5
PE0.0100.0610.096
WPE0.0430.1220.261
AAPE0.0110.0680.127
CPE0.1650.1760.081
Table 8. The Std of PE, WPE, AAPE, and CPE for BF signal in different m.
Table 8. The Std of PE, WPE, AAPE, and CPE for BF signal in different m.
BFStd
m = 3 m = 4 m = 5
PE0.0040.0480.094
WPE0.0150.0530.113
AAPE0.0040.0480.100
CPE0.0610.1090.097
Table 9. The Std of PE, WPE, AAPE, and CPE for ORF signal in different m.
Table 9. The Std of PE, WPE, AAPE, and CPE for ORF signal in different m.
ORFStd
m = 3 m = 4 m = 5
PE0.0120.0590.093
WPE0.0620.1200.251
AAPE0.0200.0830.167
CPE0.2140.2270.088
Table 10. The Std of the entropy value for the BF signal when changing the value of N under the condition of m = 5 .
Table 10. The Std of the entropy value for the BF signal when changing the value of N under the condition of m = 5 .
Std
PEWPEAAPECPE
w = 5000.09040.11790.10100.1047
w = 8000.07010.09950.08210.1186
w = 10000.06290.09400.07580.1218

Share and Cite

MDPI and ACS Style

Kang, H.; Zhang, X.; Zhang, G. Coded Permutation Entropy: A Measure for Dynamical Changes Based on the Secondary Partitioning of Amplitude Information. Entropy 2020, 22, 187. https://doi.org/10.3390/e22020187

AMA Style

Kang H, Zhang X, Zhang G. Coded Permutation Entropy: A Measure for Dynamical Changes Based on the Secondary Partitioning of Amplitude Information. Entropy. 2020; 22(2):187. https://doi.org/10.3390/e22020187

Chicago/Turabian Style

Kang, Huan, Xiaofeng Zhang, and Guangbin Zhang. 2020. "Coded Permutation Entropy: A Measure for Dynamical Changes Based on the Secondary Partitioning of Amplitude Information" Entropy 22, no. 2: 187. https://doi.org/10.3390/e22020187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop