Next Article in Journal
A Bayesian Model for Bivariate Causal Inference
Next Article in Special Issue
Application of Time-Scale Decomposition of Entropy for Eye Movement Analysis
Previous Article in Journal
Unexpected Information Demand and Volatility Clustering of Chinese Stock Returns: Evidence from Baidu Index
Previous Article in Special Issue
Complexity-Based Measures of Postural Sway during Walking at Different Speeds and Durations Using Multiscale Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

(Multiscale) Cross-Entropy Methods: A Review

by
Antoine Jamin
1,2,* and
Anne Humeau-Heurtier
2
1
COTTOS Médical, Allée du 9 novembre 1989, 49240 Avrillé, France
2
LARIS – Laboratoire Angevin de Recherche en Ingénierie des Systèmes, Univesity Angers, 62 avenue Notre-Dame du Lac, 49000 Angers, France
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(1), 45; https://doi.org/10.3390/e22010045
Submission received: 9 December 2019 / Revised: 24 December 2019 / Accepted: 26 December 2019 / Published: 29 December 2019
(This article belongs to the Special Issue Multiscale Entropy Approaches and Their Applications)

Abstract

:
Cross-entropy was introduced in 1996 to quantify the degree of asynchronism between two time series. In 2009, a multiscale cross-entropy measure was proposed to analyze the dynamical characteristics of the coupling behavior between two sequences on multiple scales. Since their introductions, many improvements and other methods have been developed. In this review we offer a state-of-the-art on cross-entropy measures and their multiscale approaches.

1. Introduction

To quantify the asynchronism between two time series, Pincus and Singer have adapted the approximate entropy algorithm to a cross-approximate entropy (cross-ApEn) method [1]. Then, other cross-entropy methods—that improve the cross-ApEn—have been developed [2,3,4,5,6,7]. Furthermore, additional cross-entropy methods have been introduced to quantify the degree of coupling between two signals, or the complexity between two cross-sequences [8,9,10]. Cross-entropy methods have recently been used in different research fields, including medicine [5,11,12], mechanics [13], and finance [7,10].
The multiscale approach of entropy measures was proposed by Costa et al. in 2002 to analyze the complexity of a time series [14]. In 2009, Yan et al. proposed a multiscale approach for cross-entropy methods to quantify the dynamical characteristics of coupling behavior between two sequences on multiple scale factors [15]. Then, other multiscale procedures have been published with different cross-entropy methods [16,17]. Multiscale cross-entropy methods have recently been used in different research fields, including medicine [18,19,20,21], finance [6,9], civil engineering [22], and the environment [23].
Cross-entropy methods and their multiscale approaches are used to obtain information on the possible relationship between two time series. For example, Wei et al. applied percussion entropy to the amplitude of digital volume pulse signals and changes in R-R intervals of successive cardiac cycles for assessing baroreflex sensitivity [18]. Results showed that the method is able to identify the markers of diabetes by the nonlinear coupling behavior of the two cardiovascular time series. Moreover, Zhu and Song computed cross-fuzzy entropy on a vibration time series to assess the bearing performance degradation process of motor [13]. Results showed that the method detects trend for bearing degradation process over the whole lifetime. In addition, Wang et al. applied multiscale cross-trend sample entropy to analyze the asynchrony between air quality impact factors (fine particulate matters, nitrogen dioxide, …), and air quality index (AQI) in different regions of China [23]. Results showed that the degree of synchrony between fine particulate matter and AQI is higher than the other air quality impact factor which reveals that fine particulate matter has become the main source of air pollution in China.
Our paper presents a state-of-the-art in three sections: First, the cross-entropy methods are introduced. We detail, in the second section, different multiscale procedures. A multiscale cross-entropy generalization is presented and other specific multiscale cross-entropy algorithms are proposed in the third section.

2. Cross-Entropy Methods

In this section, we classify cross-entropy methods according to their entropy measures: Cross-approximate entropy, cross-sample entropy, and cross-distribution entropy. Other methods that use different cross-entropy-based measures are also detailed. Table 1 shows the twelve measures that are detailed in this section.

2.1. Cross-Approximate Entropy-Based Measures

2.1.1. Cross-Approximate Entropy

Cross-approximate entropy (cross-ApEn), introduced by Pincus and Singer [1], allows to quantify asynchrony between two time series. For two vectors u and v of length N, cross-ApEn is computed as:
cross - ApEn ( m , r , N ) ( v | | u ) = Φ m ( r ) ( v | | u ) Φ m + 1 ( r ) ( v | | u ) ,
where Φ m ( r ) ( v | | u ) = 1 N m + 1 i = 1 N m + 1 log C i m ( r ) ( v | | u ) and C i m ( r ) ( v | | u ) is the number of sequences, of m consecutive points, of u that are approximately (within a resolution r) the same as sequences, of the same length, of v . One major dawback of this approach is that C i m ( r ) ( v | | u ) should not be equal to zero. This is why cross-ApEn is not really adapted for a short time series. Furthermore, it is direction-dependent because often Φ m ( r ) ( v | | u ) is generally not equal to its direction conjugate Φ m ( r ) ( u | | v ) [2]. The value of cross-ApEn computed from two signals can be interpreted as a degree of synchrony or mutual relationship.

2.1.2. Binarized Cross-Approximate Entropy

Binarized cross-approximate entropy (XBinEn), introduced by Škorić et al. [5] in 2017, is an evolution of cross-ApEn to quantify the similarity between two time series. It has the advantage of being faster than cross-ApEn. XBinEn encodes a time series divided into vectors of length m. For two vectors u and v of length N, the XBinEn algorithm follows these six steps:
  • Binary encoding series are obtained as:
    x i = 0 if u i + 1 u i 0 1 if u i + 1 u i > 0 , y i = 0 if v i + 1 v i 0 1 if v i + 1 v i > 0 ,
    where i = 1 , 2 , , N 1 , x i X m ( i ) = [ x i , x i + t , , x i + ( m 1 ) t ] , and y i Y m ( i ) = [ y i , y i + t , , y i + ( m 1 ) t ] .
    The time lag t allows a vector decorrelation to be performed;
  • Vector histograms N X ( m ) ( k ) and N Y ( m ) ( n ) are computed as:
    N X ( m ) ( k ) = i = 1 N ( m 1 ) t I { l = 0 m 1 x i + l · t × 2 l = k } , N Y ( m ) ( n ) = j = 1 N ( m 1 ) t I { l = 0 m 1 y j + l · t × 2 l = n } ,
    where k , n = 0 , 1 , , 2 m 1 , and I { · } is a function that is equal to 1 if the indicated condition is fulfilled;
  • The probability mass functions are obtained as:
    P X ( m ) ( k ) = N X ( m ) ( k ) N ( m 1 ) t , P Y ( m ) ( n ) = N Y ( m ) ( n ) N ( m 1 ) t ,
    where k , n = 0 , 1 , , 2 m 1 ;
  • A distance measure is applied:
    d ( X m ( i ) , Y m ( j ) ) = k = 0 m 1 I { x i + k · t y j + k · t } ,
    where i , j = 1 , , N ( m 1 ) t ;
  • The probability p k m ( r ) that a vector is within the distance r from a particular vector is estimated:
    p k m ( r ) = Pr { d ( X m ( k ) , Y m ) r } ;
  • XBinEn is finally obtained as:
    XBinEn ( m , r , N , t ) = Φ ( m ) ( r , N , t ) Φ ( m + 1 ) ( r , N , t ) ,
    where Φ ( m ) ( r , N , t ) = k = 0 2 m 1 P X ( m ) ( k ) · ln ( p k m ( r ) ) .
This method gives almost the same results as cross-ApEn for a non-short time series. However, it is computationally more efficient than cross-ApEn. Its main disadvantage is that it cannot identify small signal changes. XBinEn is adapted to environments where processor resources and energy are limited but it is not a substitute to cross-ApEn [5]. It is proposed when the cross-ApEn procedure cannot be applied. The value of XBinEn computed from two signals can be interpreted as a degree of relationship between a related pair of time series.

2.2. Cross-Sample Entropy-Based Measures

2.2.1. Cross-Sample Entropy

Cross-sample entropy (cross-SampEn) quantifies the degree of asynchronism of two time series. This method was introduced by Richman and Moorman in 2000 to improve the cross-ApEn limitations (see Section 2.1.1) [2]. Cross-SampEn is a conditional probability measure that quantifies the probability that a sequence of m consecutive points (called sample) of a time series u —that matches another sequence of the same length of another time series v —will still match the other sequence when their length is increased by one sample ( m + 1 ). For two vectors u and v , cross-SampEn is computed as:
cross - SampEn ( m , r , N ) ( v | | u ) = ln A m ( r ) ( v | | u ) B m ( r ) ( v | | u ) ,
where m is the sample length, N is the vectors ( u and v ) length, A m ( r ) ( v | | u ) and B m ( r ) ( v | | u ) are, respectively, the probability that a sequence of u and a sequence of v will match for m + 1 and m points (within a tolerance r).
For two time series u and v of length N, cross-SampEn can also be described as:
cross - SampEn ( u , v , m , r , N ) = ln n ( m + 1 ) n ( m ) ,
where n ( m ) represents the total number of sequences of m consecutive points of u that match with other sequences of m consecutive points of v .
The main difference between cross-ApEn and cross-SampEn is that cross-SampEn shows relative consistency whereas cross-ApEn does not. Unlike cross-ApEn, cross-SampEn is not direction-dependent. However, cross-SampEn generates, sometimes, undefined values for short time series. The value of cross-SampEn computed from two time series can be interpreted as a measure of similarity of the two time series.

2.2.2. Modified Cross-Sample Entropy

Modified cross-sample entropy (mCSE), introduced by Yin and Shang in 2015, has been developed to detect the asynchrony of a financial time series [4]. Inspired by the generalized sample entropy, proposed by Silva and Murta, Jr. [25], the authors proposed to adapt this method to cross-SampEn. The method combines cross-SampEn and nonadditive statistics. For two vectors u and v of length N, mCSE is computed as:
mCSE ( m , r , N ) = log q i = 1 N m n i ( m + 1 ) i = 1 N m n i ( m ) ,
where m is the sample length, q is the entropic index, and n i ( m ) is the number of times that the distance between vectors y m = { v ( i ) , v ( i + 1 ) , , v ( i + m 1 ) : 1 i N m + 1 } and x m = { u ( i ) , u ( i + 1 ) , , u ( i + m 1 ) : 1 i N m + 1 } is less than or equal to the tolerance r. The distance is calculated with d ( x m ( i ) , y m ( i ) ) = max { | u ( i + k ) v ( j + k ) | : 0 k m 1 } .
The value of mCSE computed from two time series can be interpreted as a degree of synchrony between the two time series and it can illustrate some intrinsic relations between the two time series.

2.2.3. Modified Cross-Sample Entropy Based on Symbolic Representation and Similarity

Modified cross-sample entropy based on symbolic representation and similarity (MCSEBSS), introduced by Wu et al. in 2018, has been developed to quantify the degree of asynchrony of two financial time series with various trends (stock markets from different areas [6]). In comparison with cross-SampEn, this method reduces the probability of including undefined entropies and it is more robust to noise. For two vectors u and v of length N, MCSEBSS is computed as:
MCSEBSS ( u , v , m , r , N ) = ln n ( m + 1 ) n ( m ) ,
where m is the sample length and n ( m ) is the number of template matches by comparing s ( u m ( i ) , v m ( j ) ) and r. For u m = { u ( i + k ) } and v m = { v ( i + k ) } ( 0 k m 1 and 1 i N m ), the similarity function s ( u m ( i ) , v m ( j ) ) is calculated as:
s ( u m ( i ) , v m ( j ) ) = #   of   in   count ( i , j ) m , 1 i , j N m ,
where c o u n t ( i , j ) is obtained by the function f defined as:
f = 1 if u m ( i + k ) = v m ( j + k ) 0 if u m ( i + k ) v m ( j + k ) , 0 k m 1 .
The parameter r must be fixed between m n m + 1 and m n m , where n is the maximum number of zeros obtained with c o u n t ( i , j ) to consider u and v similar.
The value of MCSEBSS computed from two time series can be interpreted as a degree of asynchrony of the two time series. A low cross-entropy value indicates a strong synchrony between two signals.

2.2.4. Kronecker-Delta-Based Cross-Sample Entropy

The Kronecker-delta-based cross-sample entropy (KCSE), introduced by He et al. in 2018, has been developed to define the dissimilarity between two time series [7]. KCSE is based on the Kronecker-delta function δ x , y that returns 1 if two variables are equal and 0 otherwise. For two vectors u and v of length N, KCSE is calculated as:
KCSE ( m ) = ln B m + 1 B m ,
where B m = i = 1 N m + 1 KrD u m ( i ) , v m ( i ) N m + 1 and B m + 1 = i = 1 N m KrD u m + 1 ( i ) , v m + 1 ( i ) N m . The dissimilarity, between u m ( i ) = [ u ( i ) , u ( i + 1 ) , , u ( i + m 1 ) ] and v m ( i ) = [ v ( i ) , v ( i + 1 ) , , v ( i + m 1 ) ] , is calculated as:
KrD u m ( i ) , v m ( i ) = δ u m ( i ) , v m ( i ) + δ u m ( i + 1 ) , v m ( i + 1 ) + + δ u m ( i + m 1 ) , v m ( i + m 1 ) n .
Authors show that KCSE is better to classify financial data than multidimensional scaling based on the Chebyshev distance method [7]. The value of KSCE computed from two time series can be interpreted as a degree of irregularity between the two time series.

2.2.5. Permutation-Based Cross-Sample Entropy

The permutation-based cross-sample entropy (PCSE), introduced by He et al. in 2018, is quite similar to KCSE (see Section 2.2.4) [7]. A permutation step has only been added. For two vectors u and v of length N, PCSE is calculated as:
PCSE ( m ) = ln B m + 1 B m ,
where B m = i = 1 N m + 1 KrD permuX m ( i ) , permuY m ( i ) N m + 1 and B m + 1 = i = 1 N m KrD permuX m + 1 ( i ) , permuY m + 1 ( i ) N m . The KrD function is defined in Section 2.2.4. The two vectors permuX m ( i ) and permuY m ( i ) are obtained by a permutation algorithm defined with the permutation entropy [26]. The Video S1 shows an example of a permutation algorithm.
PCSE shows better results than KCSE for synthetic data (ARFIMA model) [7]. However, the two approaches give the same results for financial data [7]. Authors show that KCSE is better to classify financial data than multidimensional scaling based on the Chebyshev distance method [7]. The value of PCSE computed from two time series can be interpreted as degree of irregularity between the two time series.

2.2.6. Cross-Trend Sample Entropy

Inspired by MCSEBSS (see Section 2.2.3), Wng et al. developed the cross-trend sample entropy (CTSE) to quantify the synchronism between two time series with strong trends [23]. For two time series u and v of length N, CTSE is calculated with the following four steps algorithm:
  • The two time series are symbolized as:
    U ( j ) = 1 if u ˜ ( j ) > u ( j ) 0 otherwise , V ( j ) = 1 if v ˜ ( j ) > v ( j ) 0 otherwise , 1 j N ,
    where u ˜ and v ˜ are, respectively, the trend of u and v obtained by polynomial fitting (linear, quadratic or higher order).
  • The template vectors u m and v m are constructed as:
    u m ( i ) = { U ( i + k ) } , v m ( i ) = { V ( i + k ) } ,
    where 0 k m 1 and 1 i N m .
  • The similarity between x m ( i ) and y m ( i ) is calculated as:
    d ( x m ( i ) , y m ( i ) ) = #   of   1   in   C m ( i ) m , 1 i N m ,
    where the i-th symbol vector C m is determined with f, a symbolic function between two template vectors u m and v m , as:
    f = 1 if u m ( i + k ) = v m ( i + k ) 0 otherwise , 0 k m 1 .
  • CTSE is finally computed as:
    CTSE ( u , v , r , N ) = ln n ( m + 1 ) n ( m ) ,
    where n ( m ) is obtained by comparing d ( x m ( i ) , y m ( j ) ) within a tolerance r for 1 i N m .
CTSE has two advantages over MCSEBSS: It is more sensitive to the difference of dynamical characteristic between two signals, and it works well with signals with trends (linear, quadratic, cubic, and sinusoidal) [23]. The value of CTSE computed from two time series can be interpreted as an indicator of dynamical structure regarding the two time series with potential trends.

2.3. Cross-Distribution Entropy-Based Measures

2.3.1. Cross-Distribution Entropy

In 2018, Wang and Shang introduced the cross-distribution entropy (cross-DistEn) to quantify the complexity between two cross-sequences [9]. To generalize the standard statistical mechanics, the authors replaced the standard distribution entropy (DistEn) based on Shannon entropy by DistEn based on Tsallis entropy [9]. The authors showed that cross-DistEn better illustrates the relationships between two vectors than cross-SampEn does [9]. For two times series u and v of length N cross-DistEn follow these four steps:
  • The state-space is reconstructed by building ( N m + 1 ) vectors X ( i ) and Y ( i ) with X ( i ) = { u ( i ) , u ( i + 1 ) , , u ( i + m 1 ) } , 1 i N m , and Y ( j ) = { v ( j ) , v ( j + 1 ) , , v ( j + m 1 ) } , 1 j N m . m is the intended size of the vectors X ( i ) and Y ( i ) ;
  • The distance matrix is built by defining the distance matrix D = { d i , j } with d i , j being the Chebyshev distance between the vectors X ( i ) and Y ( j ) defined as:
    d i j = max { | u i + k ( τ ) v j + k τ | , 0 k m 1 } ;
  • The probability density is estimated by computing the empirical probability density function of the matrix D by applying the histogram approach. If the histogram has M bins, the probability of each bin will be P t with 1 t M ;
  • The cross-distribution entropy based on the Tsallis entropy is computed as:
    c r o s s D i s t E n ( u , v ) = 1 ln ( a ) 1 q 1 ( 1 t = 1 M P t q ) ,
    where q is the order of the Tsallis entropy and a the logarithm base of the entropy computation.
The main advantage of cross-DistEn is that it is adapted for short time series. With financial data, cross-DistEn illustrates better the relationship between signals than cross-SampEn [9]. The value of cross-DistEn computed from two time series can be interpreted as a degree of linkage of the two time series.

2.3.2. Permutation Cross-Distribution Entropy

The permutation cross-distribution entropy (PCDE), introduced by He et al. in 2019, is a variant of cross-DistEn (see Section 2.3.1) [10]. The permutation allows to characterize fluctuations and prevents the impact of spatial distances on results. The PCDE algorithm is the same as the one of cross-DistEn, detailed in Section 2.3.1. However, an additional step is added before step 2 to permute X ( i ) and Y ( j ) with the permutation algorithm mentioned in Section 2.2.5. The distance matrix is therefore constructed with the permuted vectors. The value of PCDE computed from two time series can be interpreted as a degree of dissimilarity between the two time series.

2.4. Other Cross-Entropy-Based Measures

2.4.1. Cross-Conditional Entropy

Cross-conditional entropy (CCE), introduced by Porta et al. in 1919, quantifies the degree of coupling between two signals [8]. A corrected conditional entropy has been introduced to improve the approximate entropy that suffers from limitations when a finite number of sample is considered [27]. CCE is an adaptation of the corrected conditional entropy. For two signals u = { u ( i ) , i = 1 , , N } and v = { v ( i ) , i = 1 , , N } , CCE is computed as:
CCE v / u ( L ) = L 1 p ( u L 1 ) i / L 1 p ( v ( i ) / u L 1 ) × log p ( v ( i ) / u L 1 ) ,
where L is the length of the pattern extracted to be compared, p ( u L 1 ) is the joint probability of the pattern u L 1 ( i ) = ( u ( i ) , u L 1 ( i 1 ) ) , and p ( v ( i ) / u L 1 ) is the probability of the sample v ( i ) given the pattern u L 1 ( i ) . If a mixed pattern, composed by L 1 samples, of u and v : ( v ( i ) , u ( i ) , , u ( i L + 2 ) ) = ( v ( i ) , u L 1 ) , is defined and with the Shannon entropy E ( u L ) = L p ( u L ) log p ( u L ) , CCE can also be described as:
CCE v / u ( L ) = E ( v ( i ) , u L 1 ) E ( u L 1 ) .
For a limited amount of samples, the approximation of CCE always decreases to zero while increasing L. To solve this problem, a modification has been introduced as:
CCE v / u ( L ) = CCE v / u ^ ( L ) + perc v / u ( L ) × E ^ ( v ) ,
where perc v / u is the ratio of mixed patterns found only once over the total number of mixed patterns, CCE v / u ^ ( L ) and E ^ ( v ) are, respectively, the estimates of the CCE v / u ( L ) and E ( v ) based on the considered limited dataset.
CCE can be defined as a measure of unpredictability of one signal when the second is observed because it quantifies the amount of information carried by one signal which cannot be derived from the other. It is not fully a measure of synchronization. The main disadvantage of CCE is that it is not totally adapted for short time series.

2.4.2. Cross-Fuzzy Entropy

Cross-fuzzy entropy (C-FuzzyEn), introduced by by Xie et al. in 2010 [3], is an adaptation of fuzzy entropy, introduced by Chen et al. [28], that quantifies the synchrony or similarity of patterns between two signals [3]. C-FuzzyEn is an improvement of cross-SampEn that is more adapted for short time series and more robust to noise. For two times series u and v of length N, C-FuzzyEn is obtained with the following three steps algorithm:
  • The distance d i j m between X i m and Y j m is computed as:
    d i j m = d [ X i m , Y j m ] = max k ( 0 , m 1 ) | u ( i + k ) u ¯ ( i ) v ( j + k ) v ¯ ( i ) | ,
    where m is the number of consecutive data to compare, X i m = { u ( i ) , u ( i + 1 ) , , u ( i + m 1 ) } u ¯ ( i ) , and Y j m = { v ( i ) , v ( i + 1 ) , , u ( v + m 1 ) } v ¯ ( i ) . u ¯ ( i ) and v ¯ ( i ) are calculated as: u ¯ ( i ) = 1 m l = 0 m 1 u ( i + l ) , and v ¯ ( i ) = 1 m l = 0 m 1 v ( i + l ) ;
  • The synchrony or similarity degree D i j m is computed as: D i j m = μ ( d i j m , n , r ) , where μ ( d i j m , n , r ) is the fuzzy function obtained as:
    μ ( d i j m , n , r ) = exp ( d i j m ) n r ,
    where r and n determine the width and the gradient of the boundary of the exponential function, respectively;
  • Finally, C-FuzzyEn is computed as:
    C - FuzzyEn ( m , n , r ) = ln Φ m ln Φ m + 1 ,
    where Φ m = 1 N m i = 1 N m ( 1 N m j = 1 N m D i j m ) , and Φ m + 1 = 1 N m i = 1 N m ( 1 N m j = 1 N m D i j m + 1 ) .
The value of C-FuzzyEn computed from two time series can be interpreted as the synchronicity of patterns.

2.4.3. Joint-Permutation Entropy

Joint permutation entropy (JPE), introduced by Yin et al. in 2019, quantifies the synchronism between two time series. It is based on permutation entropy that consists of comparing neighboring values of each point and mapping them to ordinal patterns to quantify the complexity of a signal [26]. For two signals u and v , JPE is computed as the Shannon entropy of the d ! × d ! distinct motif combinations { ( π i d , t , π j d , t ) } :
JPE ( d , t ) = i , j : ( π i d , t , π j d , t ) p ( π i d , t , π j d , t ) · ln p ( π i d , t , π j d , t ) ,
where d is the embedded dimension and p ( π i d , π j d ) is the joint probability of { ( π i d , t , π j d , t ) } appearing in the X l d , t = { u l , u l + t , , u l + ( d 1 ) t } and Y l d , t = { v l , v l + t , , v l + ( d 1 ) t } and it is defined as:
p ( π i d , t , π j d , t ) = | | l : l T , type ( X l d , t , Y l d , t ) = ( π i d , t , π j d , t ) | | T ,
where T = N ( d 1 ) t , type ( · ) corresponds to the map from pattern space to symbol space, and | | · | | corresponds to the cardinality of a set.
The main advantages of JPE are the simplicity, the robustness, and the low computational cost. The value of JPE computed from two time series can be interpreted as a degree of correlation between the two time series [29].

3. Multiscale Procedures

To study entropy or cross-entropy measures of time series across scales, a multiscale procedure can be used. In this part we detail, in chronological order, three multiscale methods: The coarse-grained, the time-shift, and the composite coarse-grained approaches.

3.1. Coarse-Graining Procedure

In 2002 Costa et al. introduced the coarse-graining procedure to analyze the complexity, defined by the analysis of the irregularity through scale factors [14]. This method is an improvement, more adapted for a biological time series, of the coarse-graining procedure introduced by Zhang [30]. This procedure has been used in multiscale entropy and cross-entropy methods [6,9,15,20,31,32,33]. For each scale factor, this procedure derives a set of vectors illustrating the system dynamics. For a monovariate discrete signal x of length N, the coarse-grained time series y ( τ ) is calculated as:
y j ( τ ) = 1 τ i = ( j 1 ) τ + 1 j τ x i ,
where τ is the scale factor and 1 j N τ . The length of the coarse-grained vector is N τ . An example of coarse-graining procedure is presented in Figure 1A.

3.2. Time-Shift Procedure

As for the coarse-grained procedure, the time-shift procedure is used to decompose a signal through different scale factors and to perform a multiscale analysis. While coarse-graining procedure uses the averaging of time series on several interval scales, the time-shift procedure applies time shifting in time series. The main disadvantage of a coarse-graining procedure is the loss of pattern information hidden in the time series. To overcome this limitation, Pham used the Higuchi’s fractal dimension (HFD) [36] and proposed a new multiscale analysis [37]. The time-shift procedure illustrates the fractal dimension of a signal. This method has been recently used with entropy and cross-entropy measures [17,37,38,39]. HFD shows stable numerical results for stationary, non-stationary, deterministic, and stochastic time series [40]. For a monovariate discrete signal x of length N, the β time-shift signal y β ( τ ) is calculated as:
y β ( τ ) = ( x β , x β + τ , , x β + N β τ τ ) .
For each time scale τ , β time-shift time series are computed ( β = 1 , 2 , , τ ). An illustration of the time-shift procedure is presented in Figure 1B.

3.3. Composite Coarse-Graining Procedure

The coarse-graining procedure, introduced by Costa et al. [14], increases the variance of estimated entropy values at large scale. To overcome this limitation, by Wu et al. introduced in 2013 a composite coarse-graining procedure [35]. This method has been used with entropy and cross-entropy measures [16,32]. For a monovariate discrete signal x of length N, the k-th composite coarse-grained time series y k ( τ ) is computed as:
y k , j = 1 τ i = ( j 1 ) τ + k j τ + k 1 x i ,
where 1 j N τ . For each time scale τ , k composite coarse-grained time series are computed ( 1 k τ ). An illustration of the composite coarse-graining procedure is presented in Figure 1C.

4. Multiscale Cross-Entropy Methods

4.1. Generalization

Multiscale cross-entropy (MCE) methods consist of applying a cross-entropy measure for each scale factor obtained by a specific procedure. For each scale factor τ , MCE is computed as:
M C E ( X ( τ ) , Y ( τ ) ) = 1 k β = 1 k c r o s s E n ( X β ( τ ) , Y β ( τ ) ) ,
where X ( τ ) and Y ( τ ) are computed with a multiscale procedure (see Section 3), k is the number of time series that are generated by the multiscale procedure ( k = 1 for the coarse-graining procedure and k = τ for the time-shift and the composite coarse-graining procedures), and crossEn is the cross-entropy method used (see Section 2). Table 2 shows the multiscale cross-entropy methods that can be generalized with Equation (35). Before the computation of MCE, a pre-treatment can be performed. For example, the asymetric multiscale cross-SampEn (AMCSE) method [33] decomposes each signal into two, one for the positive trends and the other for the negative trends, before applying a coarse-graining procedure and cross-SampEn.

4.2. Particular Cases

Some multiscale cross-entropy methods cannot follow the generalization previously introduced. In this part we detail three particular methods, in chronological order: The adaptive multiscale cross-SampEn, the refined composite multiscale cross-SampEn, and the percussion entropy.

4.2.1. Adaptive Multiscale Cross-Sample Entropy

The adaptive multiscale cross-sample entropy (AMCSE), introduced by Hu and Liang in 2011, assesses the nonlinear interdependency between different visual cortical areas [41]. The method uses the multivariate empirical mode decomposition (MEMD), introduced by Rehman and Mandic [42], to decompose two time series into intrinsic mode functions (IMFs) that represent the oscillation mode embedded in the data. For two time series u and v , AMCSE is calculated with the following three steps algorithm:
  • The MEMD on u and v is performed to obtain N IMFs;
  • The scales of data are computed in two directions, fine-to-coarse S f 2 c τ and coarse-to-fine S c 2 f τ , with the following two equations:
    S f 2 c τ = i = τ N IMF i , ( τ N ) ,
    S c 2 f τ = i = 1 N + 1 τ IMF i , ( τ N ) .
    The two directions can be used separately or used in tandem to reveal the underlying dynamics of complex time series;
  • For each scale factor τ , the cross-SampEn (see Section 2.2.1) is applied between the two scales of data ( S f 2 c τ and S c 2 f τ ) extracted from vectors u and v .

4.2.2. Refined Composite Multiscale Cross-Sample Entropy

Yin et al. introduced in 2016 the composite multiscale cross-sample entropy (CMCSE) that follows the generalization (see Section 4.1), where the composite coarse-graining procedure and cross-SampEn are used [16]. The main disadvantage of this method is that cross-SampEn generates some undefined values when the number of matched sample is zero. To overcome this limitation, Yin et al. introduced the refined CMCSE (RCMCSE). This method leads to better results with short time series. For two times series u and v of length N, RCMSE is computed with the following three steps algorithm:
  • Coarse-grained time series are obtained with the composite coarse-graining procedure detailed in Section 3.3;
  • For a scale factor τ , the number of matched vector pairs, n k , τ m and n k , τ m + 1 , are calculated for all coarse-grained vectorsl
  • For each scale factor τ , RCMCSE is computed as:
    RCMCSE ( u , v , τ , m , r ) = ln k = 1 τ n k , τ m + 1 k = 1 τ n k , τ m ,
    where m is the dimension and of the matched vector pairs and r is the distance tolerance for the matched vector pairs.

4.2.3. Percussion Entropy

Wu et al. introduced, in 2013, the multiscale small-scale entropy index ( MEI S S ) that is obtained by summing the values of entropy for the first five scale factors [43]. Percussion entropy, introduced by Wei et al. in 2019, allows one to quantify a percussion entropy index (PEI) [18]. The method has been introduced to assess baroreflex sensitivity. PEI compares the similarity in tendency of change between two time-series. This index has been compared to MEI S S . For two time series u and v of length N, PEI is computed with the following three steps algorithm:
  • A binary transformation of u and v is used to obtain x = { x 1 , x 2 , , x N 1 } and y = { y 1 , y 2 , , y N 1 } :
    x i = 0 u ( i + 1 ) u ( i ) 1 u ( i + 1 ) > u ( i ) y i = 0 v ( i + 1 ) v ( i ) 1 v ( i + 1 ) > v ( i ) ;
  • The percussion rate for each scale factor τ is computed as:
    P τ m = 1 n m τ + 1 i = 1 n m τ + 1 count ( i ) ,
    where m is the embedded dimension vectors and count ( i ) represents the match number between A ( i ) = { x i , x i + 1 , , x i + m 1 } and B ( i + τ ) = { y i + τ , y i + τ + 1 , , y i + τ + m 1 } ;
  • PEI is calculated as:
    PEI ( m , n τ ) = ϕ m ϕ m + 1 ,
    where ϕ m = ln τ = 1 n τ P τ m and n τ is the number of scales to consider. Wei et al. [18] have chosen n τ = 5 in accordance with MEI S S .
This algorithm is a generalization of the method developed by Wei et al. [18] for a specific time series, amplitudes of successive digital volume pulse signals and changes in R-R intervals of successive cardiac cycles. At the moment, it has not been used to process other kinds of signals.

5. Conclusions

In this review we proposed a state-of-the-art of cross-entropy measures, multiscale procedures, and multiscale cross-entropy methods. Multiscale cross-entropy methods offer other interesting perspectives for time series analysis. Furthermore, all the cross-entropy methods, detailed in this review, can be translated into multiscale cross-entropy methods with the multiscale procedures presented in this review.

Supplementary Materials

The following are available online at https://www.mdpi.com/1099-4300/22/1/45/s1, Video S1: Permutation entropy–An exemple to obtain permutation vectors.

Author Contributions

Investigation, A.J. and A.H.-H.; supervision, A.H.-H.; writing–original draft, A.J.; writing–review and editing, A.H.-H. All authors have read and agreed to the published version of the manuscript.

Funding

A CIFRE grant N°2017/1165 was awarded by ANRT to the company COTTOS Médical to support the work of graduate student A.J.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pincus, S.; Singer, B.H. Randomness and degrees of irregularity. Proc. Natl. Acad. Sci. USA 1996, 93, 2083–2088. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Xie, H.B.; Zheng, Y.P.; Guo, J.Y.; Chen, X. Cross-fuzzy entropy: A new method to test pattern synchrony of bivariate time series. Inf. Sci. 2010, 180, 1715–1724. [Google Scholar] [CrossRef]
  4. Yin, Y.; Shang, P. Modified cross sample entropy and surrogate data analysis method for financial time series. Phys. A Stat. Mech. Appl. 2015, 433, 17–25. [Google Scholar] [CrossRef]
  5. Škorić, T.; Mohamoud, O.; Milovanovic, B.; Japundzic-Zigon, N.; Bajic, D. Binarized cross-approximate entropy in crowdsensing environment. Comput. Biol. Med. 2017, 80, 137–147. [Google Scholar] [CrossRef]
  6. Wu, Y.; Shang, P.; Li, Y. Multiscale sample entropy and cross-sample entropy based on symbolic representation and similarity of stock markets. Commun. Nonlinear Sci. Numer. Simul. 2018, 56, 49–61. [Google Scholar] [CrossRef]
  7. He, J.; Shang, P.; Xiong, H. Multidimensional scaling analysis of financial time series based on modified cross-sample entropy methods. Phys. A Stat. Mech. Appl. 2018, 500, 210–221. [Google Scholar] [CrossRef]
  8. Porta, A.; Baselli, G.; Lombardi, F.; Montano, N.; Malliani, A.; Cerutti, S. Conditional entropy approach for the evaluation of the coupling strength. Biol. Cybern. 1999, 81, 119–129. [Google Scholar] [CrossRef]
  9. Wang, Y.; Shang, P. Analysis of financial stock markets through the multiscale cross-distribution entropy based on the Tsallis entropy. Nonlinear Dyn. 2018, 94, 1361–1376. [Google Scholar] [CrossRef]
  10. He, J.; Shang, P.; Zhang, Y. PID: A PDF-induced distance based on permutation cross-distribution entropy. Nonlinear Dyn. 2019, 97, 1329–1342. [Google Scholar] [CrossRef]
  11. Škorić, T. Automatic Determination of Cross-Approximate Entropy Parameters for Cardiovascular Time Series. In Proceedings of the 2018 26th Telecommunications Forum (TELFOR), Belgrade, Serbia, 20–21 November 2018; pp. 1–8. [Google Scholar] [CrossRef]
  12. Wang, R.; Li, D.; Wang, J.; Cai, L.; Shi, L. Synchrony analysis using different cross-entropy measures of the electroencephalograph activity in Alzheimer’s disease. In Proceedings of the 2016 9th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Datong, China, 15–17 October 2016; pp. 1541–1545. [Google Scholar] [CrossRef]
  13. Zhu, K.; Song, X. Cross-fuzzy entropy-based approach for performance degradation assessment of rolling element bearings. Proc. Inst. Mech. Eng. E J. Process Mech. Eng. 2018, 232, 173–185. [Google Scholar] [CrossRef]
  14. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale Entropy Analysis of Complex Physiologic Time Series. Phys. Rev. Lett. 2002, 89, 068102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Yan, R.; Yang, Z.; Zhang, T. Multiscale Cross Entropy: A Novel Algorithm for Analyzing Two Time Series. In Proceedings of the 2009 Fifth International Conference on Natural Computation, Tianjin, China, 14–16 August 2009; Volume 1, pp. 411–413. [Google Scholar] [CrossRef]
  16. Yin, Y.; Shang, P.; Feng, G. Modified multiscale cross-sample entropy for complex time series. Appl. Math. Comput. 2016, 289, 98–110. [Google Scholar] [CrossRef]
  17. Jamin, A.; Duval, G.; Annweiler, C.; Abraham, P.; Humeau-Heurtier, A. A Novel Multiscale Cross-Entropy Method Applied to Navigation Data Acquired with a Bike Simulator. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 733–736. [Google Scholar] [CrossRef]
  18. Wei, H.C.; Xiao, M.X.; Ta, N.; Wu, H.T.; Sun, C.K. Assessment of Diabetic Autonomic Nervous Dysfunction with a Novel Percussion Entropy Approach. Complexity 2019, 2019, 11. [Google Scholar] [CrossRef] [Green Version]
  19. Xiao, M.X.; Lu, C.H.; Ta, N.; Jiang, W.W.; Tang, X.J.; Wu, H.T. Application of a Speedy Modified Entropy Method in Assessing the Complexity of Baroreflex Sensitivity for Age-Controlled Healthy and Diabetic Subjects. Entropy 2019, 21. [Google Scholar] [CrossRef] [Green Version]
  20. Castiglioni, P.; Parati, G.; Faini, A. Information-Domain Analysis of Cardiovascular Complexity: Night and Day Modulations of Entropy and the Effects of Hypertension. Entropy 2019, 21, 550. [Google Scholar] [CrossRef] [Green Version]
  21. Xiao, M.X.; Wei, H.C.; Xu, Y.J.; Wu, H.T.; Sun, C.K. Combination of R-R Interval and Crest Time in Assessing Complexity Using Multiscale Cross-Approximate Entropy in Normal and Diabetic Subjects. Entropy 2018, 20. [Google Scholar] [CrossRef] [Green Version]
  22. Lin, T.K.; Chien, Y.H. Performance Evaluation of an Entropy-Based Structural Health Monitoring System Utilizing Composite Multiscale Cross-Sample Entropy. Entropy 2019, 21. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, F.; Zhao, W.; Jiang, S. Detecting asynchrony of two series using multiscale cross-trend sample entropy. Nonlinear Dyn. 2019. [Google Scholar] [CrossRef]
  24. Yin, Y.; Shang, P.; Ahn, A.C.; Peng, C.K. Multiscale joint permutation entropy for complex time series. Phys. A Stat. Mech. Appl. 2019, 515, 388–402. [Google Scholar] [CrossRef]
  25. Silva, L.E.V.; Murta, L.O., Jr. Evaluation of physiologic complexity in time series using generalized sample entropy and surrogate data analysis. Chaos 2012, 22, 043105. [Google Scholar] [CrossRef] [PubMed]
  26. Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  27. Porta, A.; Baselli, G.; Liberati, D.; Montano, N.; Cogliati, C.; Gnecchi-Ruscone, T.; Malliani, A.; Cerutti, S. Measuring regularity by means of a corrected conditional entropy in sympathetic outflow. Biol. Cybern. 1998, 78, 71–78. [Google Scholar] [CrossRef] [PubMed]
  28. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of Surface EMG Signal Based on Fuzzy Entropy. IEEE Trans. Neural Syst. Reh. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
  29. Yin, Y.; Peng, C.K.; Hou, F.; Gao, H.; Shang, P.; Li, Q.; Ma, Y. The application of multiscale joint permutation entropy on multichannel sleep electroencephalography. AIP Adv. 2019, 9, 125214. [Google Scholar] [CrossRef]
  30. Zhang, Y.C. Complexity and 1/$f$ noise. A phase space approach. J. Phys. I France 1991, 1, 971–977. [Google Scholar] [CrossRef]
  31. Wu, H.T.; Lee, C.Y.; Liu, C.C.; Liu, A.B. Multiscale Cross-Approximate Entropy Analysis as a Measurement of Complexity between ECG R-R Interval and PPG Pulse Amplitude Series among the Normal and Diabetic Subjects. Comput. Math. Methods Med. 2013, 2013, 7. [Google Scholar] [CrossRef]
  32. Humeau-Heurtier, A. The multiscale entropy algorithm and its variants: a review. Entropy 2015, 17. [Google Scholar] [CrossRef] [Green Version]
  33. Yin, Y.; Shang, P. Asymmetric asynchrony of financial time series based on asymmetric multiscale cross-sample entropy. Chaos 2015, 25, 032101. [Google Scholar] [CrossRef]
  34. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy to distinguish physiologic and synthetic RR time series. In Proceedings of the Computers in Cardiology, Memphis, TN, USA, 22–25 September 2002; pp. 137–140. [Google Scholar] [CrossRef]
  35. Wu, S.D.; Wu, C.W.; Lin, S.G.; Wang, C.C.; Lee, K.Y. Time Series Analysis Using Composite Multiscale Entropy. Entropy 2013, 15, 1069–1084. [Google Scholar] [CrossRef] [Green Version]
  36. Higuchi, T. Approach to an irregular time series on the basis of the fractal theory. Phys. D Nonlinear Phenom. 1988, 31, 277–283. [Google Scholar] [CrossRef]
  37. Pham, T.D. Time-Shift Multiscale Entropy Analysis of Physiological Signals. Entropy 2017, 19, 257. [Google Scholar] [CrossRef]
  38. Dong, Z.; Zheng, J.; Huang, S.; Pan, H.; Liu, Q. Time-Shift Multi-scale Weighted Permutation Entropy and GWO-SVM Based Fault Diagnosis Approach for Rolling Bearing. Entropy 2019, 21. [Google Scholar] [CrossRef] [Green Version]
  39. Zhu, X.; Zheng, J.; Pan, H.; Bao, J.; Zhang, Y. Time-Shift Multiscale Fuzzy Entropy and Laplacian Support Vector Machine Based Rolling Bearing Fault Diagnosis. Entropy 2018, 20. [Google Scholar] [CrossRef] [Green Version]
  40. Kesić, S.; Spasić, S.Z. Application of Higuchi’s fractal dimension from basic to clinical neurophysiology: A review. Comput. Methods Prog. Biomed. 2016, 133, 55–70. [Google Scholar] [CrossRef]
  41. Hu, M.; Liang, H. Uncovering perceptual awareness of visual stimulus with adaptive multiscale entropy. In Proceedings of the 2011 3rd International Conference on Awareness Science and Technology (iCAST), Dalian, China, 27–30 September 2011; pp. 485–490. [Google Scholar] [CrossRef]
  42. Rehman, N.; Mandic, D.P. Multivariate empirical mode decomposition. Proc. Royal Soc. A Math. Phys. Eng. Sci. 2010, 466, 1291–1302. [Google Scholar] [CrossRef]
  43. Wu, H.T.; Lo, M.T.; Chen, G.H.; Sun, C.K.; Chen, J.J. Novel Application of a Multiscale Entropy Index as a Sensitive Tool for Detecting Subtle Vascular Abnormalities in the Aged and Diabetic. Comput. Math. Methods Med. 2013, 2013, 8. [Google Scholar] [CrossRef]
Figure 1. Examples of multiscale procedures for the ten first points of a time series x . (A) represents the coarse-graining procedure (modified from [34]), (B) shows the time-shift procedure, and (C) illustrates the composite coarse-graining procedure (modified from [35]).
Figure 1. Examples of multiscale procedures for the ten first points of a time series x . (A) represents the coarse-graining procedure (modified from [34]), (B) shows the time-shift procedure, and (C) illustrates the composite coarse-graining procedure (modified from [35]).
Entropy 22 00045 g001
Table 1. Cross-entropy measures, in chronological order, that are presented in this review. Authors, year, reference, and section location are indicated for each item.
Table 1. Cross-entropy measures, in chronological order, that are presented in this review. Authors, year, reference, and section location are indicated for each item.
MethodAuthorsYearRef.Section
Cross-approximate entropyPincus and Singer1996[1]Section 2.1.1
Cross-conditional entropyPorta et al.1999[8]Section 2.4.1
Cross-sample entropyRichman and Moorman2000[2]Section 2.2.1
Cross-fuzzy entropyXie et al.2010[3]Section 2.4.2
Modified cross-sample entropyYin and Shang2015[4]Section 2.2.2
Binarized cross-approximate entropyŠkorić et al.2017[5]Section 2.1.2
Modified cross-sample entropy
based on symbolicWu et al.2018[6]Section 2.2.3
representation and similarity
Kronecker-delta based cross-sample entropyHe et al.2018[7]Section 2.2.4
Permutation based cross-sample entropyHe et al.2018[7]Section 2.2.5
Cross-distribution entropyWang and Shang2018[9]Section 2.3.1
Permutation cross-distribution entropyHe et al.2019[10]Section 2.3.2
Cross-trend sample entropyWang et al.2019[23]Section 2.2.6
Joint permutation entropyYin et al.2019[24]Section 2.4.3
Table 2. Multiscale cross-entropy methods, in chronological order, that can be generalized with Equation (35). For each method, the multiscale procedure and the cross-entropy measure used and the reference are mentioned.
Table 2. Multiscale cross-entropy methods, in chronological order, that can be generalized with Equation (35). For each method, the multiscale procedure and the cross-entropy measure used and the reference are mentioned.
MethodMultiscale ProcedureCross-entropy MeasureReference
Multiscale cross-SampEnCoarse-grainedcross-SampEnYan et al., 2009 [15]
Multiscale cross-ApEnCoarse-grainedcross-ApEnWu et al., 2013 [31]
Asymetric multiscale cross-SampEnCoarse-grainedcross-SampEnYin and Shang, 2015 [33]
Composite multiscale cross-SampEnComposite coarse-grainedcross-SampEnYin et al., 2016 [16]
Multiscale cross-DistEnCoarse-grainedcross-DistEnWang and Shang, 2018 [9]
Modified multiscale cross-SampEn
based on symbolicCoarse-grainedMCSEBSSWu et al., 2018 [6]
representation and similarity
Modified multiscale cross-SampEnCoarse-grainedmCSECastiglioni et al., 2019 [20]
Time-shift multiscale cross-SampEnTime-shiftcross-SampEnJamin et al., 2019 [17]
Time-shift multiscale cross-DistEnTime-shiftcross-DistEnJamin et al., 2019 [17]
Multiscale cross-trend SampEnCoarse-grainedCTSEWang et al., 2019 [23]
Multiscale joint permutation entropyCoarse-grainedJPEYin et al., 2019 [24]

Share and Cite

MDPI and ACS Style

Jamin, A.; Humeau-Heurtier, A. (Multiscale) Cross-Entropy Methods: A Review. Entropy 2020, 22, 45. https://doi.org/10.3390/e22010045

AMA Style

Jamin A, Humeau-Heurtier A. (Multiscale) Cross-Entropy Methods: A Review. Entropy. 2020; 22(1):45. https://doi.org/10.3390/e22010045

Chicago/Turabian Style

Jamin, Antoine, and Anne Humeau-Heurtier. 2020. "(Multiscale) Cross-Entropy Methods: A Review" Entropy 22, no. 1: 45. https://doi.org/10.3390/e22010045

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop