Next Article in Journal
On the Holographic Bound in Newtonian Cosmology
Next Article in Special Issue
Application of Multiscale Entropy in Assessing Plantar Skin Blood Flow Dynamics in Diabetics with Peripheral Neuropathy
Previous Article in Journal
Entropy of Entanglement between Quantum Phases of a Three-Level Matter-Radiation Interaction Model
Previous Article in Special Issue
Changes in the Complexity of Heart Rate Variability with Exercise Training Measured by Multiscale Entropy-Based Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Multivariate Sample Entropy Algorithm for Modeling Time Series Synchronization

Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(2), 82; https://doi.org/10.3390/e20020082
Submission received: 29 November 2017 / Revised: 15 January 2018 / Accepted: 19 January 2018 / Published: 24 January 2018

Abstract

:
Approximate and sample entropy (AE and SE) provide robust measures of the deterministic or stochastic content of a time series (regularity), as well as the degree of structural richness (complexity), through operations at multiple data scales. Despite the success of the univariate algorithms, multivariate sample entropy (mSE) algorithms are still in their infancy and have considerable shortcomings. Not only are existing mSE algorithms unable to analyse within- and cross-channel dynamics, they can counter-intuitively interpret increased correlation between variates as decreased regularity. To this end, we first revisit the embedding of multivariate delay vectors (DVs), critical to ensuring physically meaningful and accurate analysis. We next propose a novel mSE algorithm and demonstrate its improved performance over existing work, for synthetic data and for classifying wake and sleep states from real-world physiological data. It is furthermore revealed that, unlike other tools, such as the correlation of phase synchrony, synchronized regularity dynamics are uniquely identified via mSE analysis. In addition, a model for the operation of this novel algorithm in the presence of white Gaussian noise is presented, which, in contrast to the existing algorithms, reveals for the first time that increasing correlation between different variates reduces entropy.

1. Introduction

Multivariate data analysis tools are essential for characterizing the interaction between the variates of complex systems; applications are wide-ranging and include those in biology [1], climatology [2] and finance [3]. Standard methods for estimating interdependencies between multiple data channels are almost invariably linear; typical examples are cross-correlation and coherence (correlation in the frequency domain). More advanced methods, such as Granger causality [4], offer insight into the temporal ordering of interactions and are widely used in, for example, neuroscience applications, for which directionality information is of value. Partial directed coherence extends the concept of Granger causality to the frequency domain [5].
While progress has been made on nonlinear extensions of the above second-order algorithms, information-theoretic measures, such as mutual information and transfer entropy [6], are intrinsically suited to cater for higher-order interactions. Traditional entropy methods are limited by their requirements for large numbers of samples and sensitivity to noise [7]. To this end, the approximate entropy (AE) algorithm [7] was developed to provide a statistically valid measure of entropy for real-world univariate data that may be noisy and short. It represents the probability that similar patterns (delay vectors—DVs) in a time series will remain similar once the pattern lengths are increased (extended DVs), thereby providing a natural measure of the regularity of a time series. An extension, the sample entropy (SE) algorithm [8], improves the bias issues experienced by AE by omitting self-matches from the similarity calculation stage. In general, a stochastic time series is characterized by high AE/SE, while a regular signal exhibits low AE/SE. Additionally, multiscale extensions of the algorithm (multiscale sample entropy—MSE) [9,10] have expanded its capabilities beyond regularity estimation to evaluating the structural richness of a signal-generating system, its complexity. Both the SE and MSE algorithms have been widely used in physiological applications [11,12,13].
Recently, two multivariate extensions of the SE and MSE algorithms have been developed to evaluate the regularity and complexity for any number of data channels in a rigorous and unified way [14,15]. The key difference between the two multivariate algorithms is the manner in which multivariate DVs are extended. In the first extension, termed the naïve method [14], extended DV subspaces are generated for each variate, and self-matches are computed for each subspace separately. A shortcoming of the naïve method is its failure to account for inter-channel couplings, such as correlation between the variates. This was reflected in the “full method” [15]. Critically, a multivariate MSE (mMSE) method that caters fully for cross-channel dynamics enables the modeling of any complexity that exists between the variates, offering greater insight for physical systems that are typically multivariate and correlated. In [15], the benefits of a multivariate approach over univariate algorithms were demonstrated for applications spanning the categorization of wind regimes and the analysis of human gait recordings.
Despite the clear improvements of the full multivariate method over existing work, there are several concerns regarding its operation. Firstly, the way in which the DVs are extended is such that distances are directly calculated between elements of the different variates, potentially obscuring the physical meaning of the analysis for heterogeneous data. Secondly, such inconsistencies in the alignment of extended DVs can lead to inaccuracies. Finally, empirical results obtained using the full method for white Gaussian noise (WGN) and 1/f noise imply that regularity decreases and complexity increases with increasing correlation. However, viewing increased correlation between the variates as a decrease in cross-channel regularity is not consistent with physical intuition, motivating this work. It must also be noted that it has been suggested that the size of the time delay employed in the SE algorithm can have an effect on the computed entropy [16]; however, for the purposes of consistency across all analyses reported in this study, a unity time delay is employed.
To address these concerns, we propose a novel multivariate sample entropy (mSE) and its multiscale extension for complexity analysis (mMSE). At the core of the algorithm is our novel treatment of multivariate DVs, which ensures (i) element-by-element distances are not computed directly between different variates, and (ii) alignments between multivariate DVs remain consistent and independent of the number of data channels. Simulation results for WGN illustrate how the proposed algorithm interprets increased correlation between the variates as an increase in regularity—a missing result that is in agreement with physical intuition. To support this, we have derived a model for the performance of the algorithm in the presence of bivariate WGN; its numerical outcomes are in agreement with simulations. It is furthermore shown via a random alignment operation applied to the variates that this makes it possible to comprehensively distinguish between within- and cross-channel dynamics, thus providing additional insight into inter- versus intra-modal properties.
The algorithm is validated on synthetic and real data for biological applications. Through simulations, we reveal a previously unstudied feature of multivariate SE algorithms—their ability to detect synchronized regularity dynamics. Unlike other measures of synchronization, which assume temporal locking of phase information (phase synchrony [17,18]) or the existence of a functional mapping between the variates (generalized synchrony [19]), synchronized regularity can exist between time series that are generated by independent processes.

2. Sample Entropy

Motivated by the shortcomings of standard entropy measures for short and noisy time series, the approximate entropy technique was introduced in [7]. It characterizes the likelihood that similar patterns within a time series, the signal DVs, will remain similar when the pattern lengths are increased. A robust extension, which neglects self-matches, called the SE, has been developed [8] and is described below:
  • For lag τ and embedding dimension m, generate DVs:
    X m ( i ) = [ x i , x i + 1 , , x i + τ ( m 1 ) ]
    where i = 1 , 2 , , N τ ( m 1 ) .
  • For a given DV, X m ( i ) , and a threshold, r, count the number of instances, Φ m ( i , r ) , for which d { X m ( i ) , X m ( j } r , i j , where d { · } denotes the maximum norm.
  • Define the frequency of occurrence as
    Φ m ( r ) = 1 N τ ( m 1 ) + 1 i = 1 N τ ( m 1 ) + 1 Φ m ( i , r )
  • Extend the embedding dimension ( m m + 1 ) of the DVs in step (1), and repeat steps (2) and (3) to obtain Φ m + 1 ( r ) .
  • The SE is defined as the negative logarithm of the values for different embedding dimensions, that is,
    SE ( m , r , τ ) = l n Φ m + 1 ( r ) Φ m ( r )
In general, the less predictable or the more irregular a time series, the higher its SE. A block diagram of the algorithm is shown in Figure 1.

2.1. Multiscale Sample Entropy

The MSE calculates the SE over multiple time scales in the data [9,10]. A key insight is given by the fact that SE for WGN (which has no structure) decreases for increasing scale factors, while the SE for 1/f noise, which has a self-similar infinitely repeating behaviour (contains structure), remains constant with scale. In this way, the multiscale extension reveals long-range signal correlations—dynamics closely linked to the signal complexity. If the SE of a time series remains high over multiple scale factors, it is said to exhibit high complexity.

3. Multivariate Sample Entropy

3.1. Existing Algorithms

Two multivariate extensions of the SE algorithm have been proposed in [14,15]. In both cases, the un-extended multivariate DVs are generated on the basis of the approach outlined by Cao et al. [20]. Given a P-variate time series, x k , i where k = 1 , , P is the channel index and i = 1 , , N is the sample index, the multivariate DVs are given by
X M ( i ) = [ x 1 , i , , x 1 , i + ( m 1 1 ) τ 1 , x 2 , i , , x 2 , i + ( m 2 1 ) τ 2 , , x P , i , , x P , i + ( m P 1 ) τ P ]
where M = [ m 1 , m 2 , , m P ] is the multivariate embedding dimension vector and ø = [ τ 1 , τ 2 , , τ P ] is the multivariate time-lag vector. Figure 2 shows a bivariate time series and its multivariate DVs with embedding dimension M = [ 1 , 1 ] .
To extend the multivariate DVs, the naïve approach in [14] generates P DV subspaces, in which the length of a specific variate is extended but other variates are left unchanged. The extended DV subspace for a variate k is given by
X M e k ( i ) = [ x 1 , i , , x 1 , i + ( m 1 1 ) τ 1 , , x k , i , , x k , i + ( m k 1 ) τ k , x k , i + ( m k ) τ k , , x P , i , , x P , i + ( m P 1 ) τ P ]
for i = 1 , , N and e k = e 1 , , e P . For the bivariate case, this approach to DV extension is illustrated in Figure 3. Distances are then calculated between DV pairs in each of the P subspaces to obtain P estimates for the frequency of occurrence, and the final value is obtained by averaging the subspace estimates. The next steps of the algorithm are the same as for the univariate case. A significant shortcoming of the naïve method is that, by measuring distances between the DVs within separate subspaces, it does not cater fully for inter-channel couplings. This was illustrated in [15], where the naïve method could not distinguish between different degrees of correlated variates.
To address these issues, the full method was proposed in [15], in which pair-wise distances are calculated between all DVs across all the subspaces, as described in Equation (5) and illustrated for the bivariate case in Figure 3. This enables the enhanced modeling of inter-channel couplings, as exemplified by its ability to distinguish between different degrees of correlation in the variates. The two approaches are compared in Figure 4 for bivariate WGN with a length of 5000 samples; the mSE parameters were M = [ 1 , 1 ] , τ = [ 1 , 1 ] and r = 0.15 , each of the variates was standardized to unit variance and zero mean, and the correlation between the variates was ρ = 0.95 . We observe that while the naïve algorithm [14] cannot distinguish between correlated and uncorrelated WGN, the full algorithm [15] exhibits different SE values for each scale factor. As the full algorithm described in [15] represents the most recent multivariate extension to date; in the sequel we refer to it as the existing method.
Despite providing clear improvements over the previous mMSE algorithms, there are still issues regarding the existing method that need to be addressed:
  • Inability to cater adequately for heterogeneous data sources.
  • Inconsistencies in the alignment of DVs.
  • Counter-intuitive representation of multivariate dynamics.
Concerning item 1, in the standard method, the DV distances are directly calculated between elements of different variates. This is illustrated in Figure 3 for the bivariate case. If the variates are heterogeneous, that is, if they reflect different components of a biological system (e.g., cardiac and neural activity), then operations involving element-by-element comparisons will hinder the physical meaning of the analysis.
Concerning item 2 above, Figure 3 shows that one of the distances calculated for the extended DVs of a bivariate signal is between [ x 1 , 1 , x 1 , 2 , x 2 , 1 ] T and [ x 1 , 2 , x 1 , 3 , x 2 , 2 ] T . In this instance, as is also the case in the operation of the univariate algorithm, the delays between corresponding elements are the same (+1 in this instance). However, when the distances are calculated between [ x 1 , 1 , x 1 , 2 , x 2 , 1 ] T and [ x 1 , 1 , x 2 , 1 , x 2 , 2 ] T , this results in different delays between the DV elements: 0 ( { x 1 , 1 , x 1 , 1 } ) , 1 ( { x 1 , 2 , x 2 , 1 } ) and +1 ( { x 2 , 1 , x 2 , 2 } ) . This inconsistency in the alignment of the DV elements may cause contradictory results and affect the accuracy of the analysis.
Concerning item 3 above, Figure 4 illustrates that the multivariate SE increases across all scale factors for increasing correlation between the variates. Considering the algorithm output at the first scale factor only, the result indicates that regularity decreases with increasing correlation—a result that is not consistent with physical intuition. Instead, it is expected that as variates become more independent, it will be characterized by a decrease in regularity. Likewise, considering the algorithm output across all scale factors, the result indicates that complexity increases with increasing correlation. No precise definition of complexity exists; however an intuitive expectation is that the number of required mathematical descriptors for a system should increase with complexity. For this reason, it can be assumed that as the variates become more correlated/dependent, the system will be characterized by a decrease in complexity, the “complexity loss theory”.

3.2. Synchronized Regularity

The ability of multivariate SE algorithms to distinguish between coupled regularity dynamics has so far been confined to correlated variates [15] (see also Figure 4). In such studies, the couplings between the variates are static—they remain unchanged across time. Additionally, such relationships are adequately modeled by correlation estimation, not requiring the unique capabilities of a multivariate SE algorithm.
We instead propose a multivariate benchmark test in which the variates are generated from independent random processes but their regularity is synchronized. In other words, the within-channel regularity changes at the same time within each variate. We consider the bivariate signal
x 1 , i = λ 1 , i 1 2 v 1 , i + ( 1 λ 1 , i ) 1 2 u 1 , i
x 2 , i = λ 2 , i 1 2 v 2 , i + ( 1 λ 2 , i ) 1 2 u 2 , i
where i = 1 , , N , N is the number of samples, λ 1 , i and λ 2 , i [ 0 , 1 ] are mixing parameters, v 1 , i and v 2 , i are independent realizations of WGN (zero mean and unit variance), and u 1 , i and u 2 , i are independent realizations of 1/f noise (zero mean and unit variance). We note that no correlation, phase or generalized synchronization exists between the variates, as they are all independent. Instead, the parameter λ i enforces synchronized regularity; when λ 1 , i = λ 2 , i = 0 , the SE in both variates is low as each contains 1/f noise, and when λ 1 , i = λ 2 , i = 1 , the SE in both variates is high as each contains WGN. A dynamic mixing parameter can be generated by
λ k , i = 1 if z i ( β ) > 0 0 if z i ( β ) < 0
where k { 1 , 2 } , i = 1 , , N and z i ( β ) is a random noise time series with a spectral distribution that decays at rate β . When β = 0 , the time series is WGN; when β = 1 , it is 1/f noise; and when β = 2 , it is Brownian noise. Figure 5 shows a bivariate time series generated using the dynamic mixing model described by Equations (6)–(8) with β = 2 ; we observe that the within-channel regularities are synchronized.
The existing mMSE algorithm was applied to distinguish between the scenarios in which (i) the variates are independent, that is, λ 1 , i λ 2 , i ; and (ii) the regularity of the variates is synchronized, that is, λ 1 , i = λ 2 , i . In both scenarios, β = 1.6 (see Equation (8)) was used to generate the mixing parameters. The signal length was N = 15,000 samples; r = 0.4 and M = [ 1 , 1 ] . Figure 6 shows the mMSE for each scenario; a two-tailed two-sample t-test was applied to determine the degree of separation at each scale factor (38 degrees of freedom). We observe that a statistically significant separation was determined at scale factor 4 ( p < 0.05 ) and at all subsequent scale factors. It is also noted that the synchronized regularity caused a decrease in the mSE; this result is in agreement with physical intuition but contradicts results obtained with the same algorithm for correlated variates (see Figure 4b), where coupled regularity dynamics was found to increase the multivariate SE. As real-world processes can be expected to exhibit different forms of coupled regularity dynamics simultaneously, this compromises the accuracy of the algorithm in applications.

3.3. The Proposed Algorithm

At the core of the proposed algorithm, which addresses the above issues, is the manner in which multivariate DVs are extended to calculate the SE. To avoid the direct calculation of distances between the elements of different variates and DV misalignment, and to induce a desired mode-of-operation for correlated data, generating an extended set of multivariate DVs from those in Equation (4) is proposed in the following manner:
X M e ( i ) = [ x 1 , i , , x 1 , i + ( m 1 1 ) τ 1 , x 1 , i + ( m 1 ) τ 1 , x k , i , , x k , i + ( m k 1 ) τ k , x k , i + ( m k ) τ k , , x P , i , , x P , i + ( m P 1 ) τ P , x P , i + ( m P ) τ P ]
for i = 1 , , N . For the bivariate case, the proposed DV extension is illustrated in Figure 7. We observe that distances are only calculated between elements of corresponding variates, making the approach perfectly suited for heterogeneous data.
Although the distances between DVs were not directly calculated between the elements of different variates, the estimation of multivariate SE on the basis of the proposed approach can still detect coupled regularity dynamics as well as existing methods, and in some instances, its performance enables better separation (see the following section). To support the claim of a desired mode-of-operation for correlated data, in the Appendix A, we present a model for the operation of the proposed multivariate algorithm in the presence of bivariate WGN; the model reveals how an increasing correlation between the variates reduces the multivariate SE. This is demonstrated in Figure 8 for the multiscale operation of the algorithm for bivariate WGN. The algorithm was also applied to time series with synchronized regularity between the variates generated using the same mixing model as before (see Equation (8) with β = 1.6 ) ; the results are shown in Figure 9. We note that, unlike the existing algorithm, a significant separation between the two time series also exists at the first scale factor (cf. Figure 4).
It is important to note an increase in the degree of coupling between the variates, either via correlation or synchronized regularity; in both cases this causes a decrease in the SE. It is natural to expect that real-world systems will exhibit different forms of coupling simultaneously and that an algorithm that behaves in a consistent way for each form of coupling is better equipped to model changes in multivariate regularity and complexity.

3.4. Multivariate Surrogates

For a given measure or index, it is common to employ signal surrogates to provide a baseline or reference value. Surrogates are a set of time series that share certain characteristics of the original signal but lack the property whose effect on the measure we wish to study [1]. For instance, in univariate sample analysis, it is common to generate surrogates that retain the spectrum shape of the original signal but that have randomized phase values, in this way creating similar signals with high irregularity and no structure (low complexity). In the same way, for greater insight, the SE of the original signal can be compared with the values obtained for the surrogates.
For the multivariate case, it is desirable to remove any interdependencies between the variates in order to distinguish between within- and cross-channel regularity and complexity. We propose to utilize the multivariate surrogate approach used in previous synchronization measures, for example, in the study of asymmetry [21] and mutual information [22], whereby one of the variates is temporally shifted with respect to the other (random alignment). Figure 10 shows the mSE, obtained with parameters M = [ 1 , 1 ] , τ = [ 1 , 1 ] and r = 0.15 , for correlated bivariate WGN ( ρ = 0.6 ; 5000 samples), as well as the average values obtained for 30 surrogates created by randomly shifting the second variate. We observe that the mSE of the original signal is significantly lower than that obtained for the surrogates, indicating the presence of coupled dynamics between the variates.

4. Simulations

The performance of the proposed algorithm was illustrated over the following case studies: (i) detecting joint regularity dynamics between recordings of cardiac function, and (ii) classifying multichannel recordings of brain function during wake and sleep states. In all cases, unless stated otherwise, the time series were scaled to zero mean and unit variance prior to mSE analysis.

4.1. Detecting Synchronized Cardiac Behaviour

We firstly considered a real-world synchronized regularity scenario. A healthy male subject (aged 32) performed a series of breathing exercises while his cardiac function was monitored via the electrocardiogram (ECG); electrodes were placed on the chest and ECG data were recorded using the gtec g.USBamp, a certified and U.S. Food and Drug Administration (FDA)-listed medical device, with a sampling frequency of 1200 Hz. The time difference between R peaks, sharp dominant peaks in the ECG waveform, was calculated and the R–R interval time series was generated using cubic spline interpolation at regular time intervals of 0.25 s.
For a given 30 s period, the instruction was either to (i) breathe normally (unconstrained), or (ii) breathe at a fixed rate of 15 breathing cycles (inhale/exhale) per minute aided by a metronome (constrained). The instruction was alternated from period to period. The periods of constrained breathing had the effect of inducing different regularity into the R–R interval through the phenomenon of respiratory sinus arrhythmia (RSA), the modulation of the cardiac output by the respiration effort [23]. Two 300 s trials were recorded; in the first, the subject started breathing in a constrained fashion, and in the second, the subject started breathing in an unconstrained fashion. Figure 11 shows a segment of the trial in which the subject started breathing in a constrained fashion. Prior to analysis, a high-pass filter with a cutoff frequency of 0.1 Hz was applied to the R–R interval.
Figure 12 compares the results of the proposed algorithm ( M = [ 1 , 1 ] , τ = [ 1 , 1 ] , and r = 0.4 ) with those obtained by cross-correlation, with the aim to determine the degree of synchronization between the two trials at different time-lags. Spectral analysis of the proposed mSE results revealed that synchronized regularity occurred at 60 s intervals (approximately 1/60 = 0.0167 Hz)—this clearly indicates joint regularity between the recording trials whenever the subject performed the same breathing instruction every 2 × 30 s. Furthermore, the approach revealed the correct time-lag between the trials, with the minimum mSE value occurring at lag −32 s. We observe that the cross-correlation approach was unable to reveal the synchronization in regularity; spectral analysis of the results showed no peak at 0.017 Hz.

4.2. Classifying Sleep States

The proposed algorithm was next applied to recordings of brain function during wake and sleep states. The data, electroencephalogram (EEG) recordings extracted from whole-night polysomnography recordings, were obtained from (http://www.tcts.fpms.ac.be/~devuyst/Databases/DatabaseSpindles/). Each 5 s segment of the EEG was independently annotated by an expert according to the Rechtschaffen and Kales criteria (wake; Rapid Eye Movement (REM); sleep stages S1, S2, S3 and S4). The task was to distinguish between the wake state and the state of slow-wave sleep (stages S3 and S4). Eight 30 min excerpts were available; however only four of these were considered. For two excerpts (2 and 7), there were insufficient wake periods (no sustained wake periods exceeding 1 min); for another excerpt (3), the sampling frequency was too low for meaningful analysis (50 Hz); and for another (8) there were no periods of slow-wave sleep. The excerpts analyzed were 1, 4, 5 and 6; excerpt 1 had a sampling frequency of 100 Hz, and the two channels of the EEG were obtained using the bipolar electrode configurations C3-A1 and FP1-A1; excerpts 4, 5 and 6 had a sampling frequency of 200 Hz, and the two channels of the EEG were obtained using the bipolar electrode configurations CZ-A1 and FP1-A1.
The multivariate SE was estimated ( M = [ 1 , 1 ] , τ = [ 1 , 1 ] , and r = 0.4 ) between corresponding EEG segments of length 2.5 s with a 50% overlap, with segments containing voltage values exceeding ± 100 µV rejected to discard artefacts. The difference between the multivariate SE values for the states of wake and slow-wave sleep were determined using the Bhattacharyya distance, given by
D = 1 4 ln 1 4 σ w 2 σ s 2 + σ s 2 σ w 2 + 2 + 1 4 ( μ s μ w ) 2 σ s 2 + σ w 2
where μ w and σ w denote the mean and standard deviation of the feature values for the wake state, and μ s and σ s are the same values for the slow-wave sleep state. The degrees of separation using the existing and the proposed algorithm are shown in Table 1; we observe that for all excerpts, the degrees of separation between the states of wake and slow-wave sleep were greater using the proposed algorithm (the value of D is larger).

5. Conclusions

We have introduced an mSE algorithm that, unlike existing approaches, caters for heterogeneous data and yields improved insight into coupled regularity dynamics. This is exemplified, in part, by its consistent treatment of correlation and synchronized regularity, both of which exhibit low mSE values; this is in agreement with intuition concerning regularity and the complexity loss theory. Simulation results for the proposed algorithm reveal greater separation between neural data during different sleep stages. Unlike standard tools, which are invariably linear, it is also shown how the approach is sensitive to higher-order synchronization. Multivariate surrogate-generation techniques have been shown to enhance the significance of the results. The concept of synchronized regularity has been illuminated and a benchmark test proposed, which opens a new avenue of research in a number of applications.

Acknowledgments

The authors wish to thank Professor Mary Morrell of the Sleep Unit, Royal Brompton Hospital, Imperial College London, United Kingdom, for all her assistance and guidance concerning the application of our work on sleep. Our collaboration has been supported by a UK Biomedical Research Unit Pump Priming grant.

Author Contributions

David Looney performed the data analysis reported in this paper, under the supervision of Danilo P. Mandic. Tricia Adjei revised this paper. All authors read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AEApproximate entropy
DVDelay vector
ECGElectrocardiogram
EEGElectroencephalogram
mMSEMultivariate multiscale sample entropy
MSEMultiscale sample entropy
mSEMultivariate sample entropy
RSARespiratory sinus arrhythmia
SESample entropy
WGNWhite Gaussian noise

Appendix A. Multivariate SE for WGN

One of the key SE results is that for WGN. We here revisit the model for the operation of the univariate algorithm, presented in [10], and then derive a model for the case of bivariate WGN.
In the case of the univariate algorithm when m = 1 , the denominator in Equation (3) indicates the probability that the distance between two data points is less than or equal to r, that is, | x i x j | r , which we denote by
P { m = 1 , r } = P ( | x i x j | r )
The physical meaning of the numerator in Equation (3) is the probability that the pair-wise distance for DVs with embedding dimension m = 2 , the extended DVs, is less than or equal to r. As the elements in the DVs are independent random variables, the probability can be rewritten as follows (for simplicity, we here assume τ = 1 ):
P { m = 2 , r } = P ( | x i x j | r ) × P ( | x i + 1 x j + 1 | r ) = P { m = 1 , r } P { m = 1 , r }
Generalizing this result for any embedding dimension, for univariate WGN, the SE is given by
SE ( m , r ) = log e P { m , r } P { m + 1 , r } = log e P { m = 1 , r } P { m = 1 , r } P { m = 1 , r } P { m = 1 , r } P { m = 1 , r } P { m = 1 , r } P { m = 1 , r } = log e P { m = 1 , r }
Thus, the SE for WGN is defined by P { m = 1 , r } , which, for zero mean and unit variance, is given by
P { m = 1 , r } = P ( | x i x j | r ) = x i r x i + r p ( x j ) d x j p ( x i ) d x i = 1 2 π x i r x i + r e x j 2 / 2 d x j e x i 2 / 2 d x i = 1 2 π erf x i + r 2 erf x i r 2 e x i 2 / 2 d x i
The term within curly brackets { · } denotes the probability that a point at x i is within distance r of another randomly selected point. This is represented by the area under the normal curve in Figure A1. Numerically evaluating the model for r = 0.15 gives 2.47, which matches the simulation result for the same parameters (see first scale factor in Figure 4b).
Figure A1. The area under the Gaussian curve, centred at x ^ i , denotes the probability that the distance between this point and any other randomly selected point is less than or equal to r.
Figure A1. The area under the Gaussian curve, centred at x ^ i , denotes the probability that the distance between this point and any other randomly selected point is less than or equal to r.
Entropy 20 00082 g0a1
For the case of bivariate WGN, the proposed multivariate algorithm calculates the SE as
mSE ( M , r ) = log e P { M , r } P { M e , r }
where P { M , r } is the probability that the pair-wise distance for DVs with embedding dimension M is less than or equal to r, and P { M e , r } is the probability for the extended case as described by Equation (9). Similarly to the univariate case, independence between the multivariate DVs can be assumed, and the above can be rewritten as
mSE ( M , r ) = log e P { M = [ 1 , 1 ] , r }
where the probability is given by
P { M = [ 1 , 1 ] , r } = P ( | x 1 , i x 1 , k | r , | x 1 , j x 1 , l | r ) = x 1 , k = x 1 , i r x 1 , i + r x 1 , l = x 1 , j r x 1 , j + r p ( x 1 , k , x 2 , l ) δ x 1 , k δ x 2 , l × p ( x 1 , i , x 2 , j ) δ x 1 , i δ x 2 , j
for which, assuming variates with zero mean and unit variance, the joint probability is given by
p ( x 1 , i , x 2 , j ) = 1 2 π 1 ρ 2 e 1 2 ( 1 ρ 2 ) x 1 , i 2 + x 2 , j 2 2 ρ x 1 , i x 2 , j
and ρ is the correlation coefficient. Similarly to the univariate case, the term within curly brackets { · } denotes the probability that a point at x 1 , i , x 2 , j is within distance r of another randomly selected bivariate point; this is represented by the volume under the bivariate surface in Figure A2.
Figure A2. The volume under the surface of a normal bivariate distribution, centred at ( x ^ 1 , i , x ^ 2 , j ), denotes the probability that the distance between this point and any other randomly selected point is less than or equal to r.
Figure A2. The volume under the surface of a normal bivariate distribution, centred at ( x ^ 1 , i , x ^ 2 , j ), denotes the probability that the distance between this point and any other randomly selected point is less than or equal to r.
Entropy 20 00082 g0a2
Unlike the univariate case, however, a closed-form solution for the term enclosed in brackets does not exist for ρ 0 . However, it is possible to approximate the cumulative distribution function:
F ( x ^ 1 , i , x ^ 2 , j ; ρ ) = P ( x 1 , k > x ^ 1 , i , x 2 , l > x ^ 2 , j )
as follows [24]:
F ( x ^ 1 , i , x ^ 2 , j ) L ( x ^ 1 , i , x ^ 2 , j ; ρ ) = F ( x ^ 1 , i ) [ F { ξ ( x ^ 1 , i , x ^ 2 , j ; ρ ) } 1 2 ρ 2 ( 1 ρ 2 ) 1 ξ ( x ^ 1 , i , x ^ 2 , j ; ρ ) p { ξ ( x ^ 1 , i , x ^ 2 , j ; ρ ) } [ 1 + μ ( x ^ 1 , i ) μ 2 ( x ^ 1 , i ) ] ]
where
ξ ( x ^ 1 , i , x ^ 2 , j ; ρ ) = ρ μ ( x ^ 1 , i ) x ^ 2 , j ( 1 ρ 2 )
and
μ ( x ^ 1 , i ) = x ^ 1 , i F ( x ^ 1 , i )
Using this approximation, for which the accuracy decreases for increasing ρ , we can derive an approximation to the probability in Equation (A8) as
P { M = [ 1 , 1 ] , r } R a ( R b + R c )
where
R a = L ( x ^ 1 , i r , x ^ 2 , j r , ρ ) L ( x ^ 1 , i + r , x ^ 2 , j + r , ρ )
R b = L ( x ^ 1 , i r , x ^ 2 , j r , ρ ) L ( x ^ 1 , i + r , x ^ 2 , j r , ρ )
R c = L ( x ^ 1 , i r , x ^ 2 , j r , ρ ) L ( x ^ 1 , i r , x ^ 2 , j + r , ρ )
The above expressions are illustrated in Figure A3. Combining the probability approximation in Equation (A13) with Equation (A6), we obtain a model that approximates the behaviour of the proposed algorithm for bivariate WGN for any correlation between the variates. A comparison of the numerical evaluation of the model and simulation results (averaged over 100 realizations of bivariate WGN; M = [ 1 , 1 ] and r = 0.15 ; variate lengths of 3000 samples) for different correlation coefficient values ( ρ ) is shown in Figure A4. We observe that the model results match the simulation results, particularly for small values of ρ . However, the model is less accurate for larger values of ρ , as the accuracy of the approximation of the cumulative distribution function from [24], on which the model is based, decreases.
Figure A3. Approximating the bivariate cumulative distribution function at the locations shown can be used to approximate P { M = [ 1 , 1 ] , r } (see Equation (A13)).
Figure A3. Approximating the bivariate cumulative distribution function at the locations shown can be used to approximate P { M = [ 1 , 1 ] , r } (see Equation (A13)).
Entropy 20 00082 g0a3
Figure A4. Numerical evaluation of the derived model (Equation (A13)) and simulation results for bivariate white Gaussian noise (WGN) with varying degrees of correlation ( ρ ). As expected, the accuracy of the model decreases for larger correlation values.
Figure A4. Numerical evaluation of the derived model (Equation (A13)) and simulation results for bivariate white Gaussian noise (WGN) with varying degrees of correlation ( ρ ). As expected, the accuracy of the model decreases for larger correlation values.
Entropy 20 00082 g0a4

References

  1. Pereda, E.; Quiroga, R.Q.; Bhattacharya, J. Nonlinear multivariate analysis of neurophysiological signals. Prog. Neurobiol. 2005, 77, 1–37. [Google Scholar] [CrossRef] [PubMed]
  2. Burke, D.J.; O’Malley, M.J. A study of principal component analysis applied to spatially distributed wind power. IEEE Trans. Power Syst. 2011, 26, 2084–2092. [Google Scholar] [CrossRef]
  3. Filis, G.; Degiannakis, S.; Floros, C. Dynamic correlation between stock market and oil prices: The case of oil-importing and oil-exporting countries. Int. Rev. Financ. Anal. 2011, 20, 152–164. [Google Scholar] [CrossRef]
  4. Granger, C.W.J. Investigating Causal Relations by Econometric Models and Cross-spectral Methods. Econometrica 1969, 37, 424–438. [Google Scholar] [CrossRef]
  5. Sameshima, K.; Baccalá, L.A. Using partial directed coherence to describe neuronal ensemble interactions. J. Neurosci. Methods 1999, 94, 93–103. [Google Scholar] [CrossRef]
  6. Schreiber, T. Measuring Information Transfer. Phys. Rev. Lett. 2000, 85, 461–464. [Google Scholar] [CrossRef] [PubMed]
  7. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef] [PubMed]
  8. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
  9. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 2002, 89, 068102. [Google Scholar] [CrossRef] [PubMed]
  10. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef] [PubMed]
  11. Bruce, E.N.; Bruce, M.C.; Vennelaganti, S. Sample entropy tracks changes in electroencephalogram power spectrum with sleep state and aging. J. Clin. Neurophysiol. 2009, 26, 257–266. [Google Scholar] [CrossRef] [PubMed]
  12. Costa, M.; Ghiran, I.; Peng, C.K.; Nicholson-Weller, A.; Goldberger, A.L. Complex dynamics of human red blood cell flickering: Alterations with in vivo aging. Phys. Rev. E 2008, 78, 020901. [Google Scholar] [CrossRef] [PubMed]
  13. Hornero, R.; Abásolo, D.; Escudero, J.; Gómez, C. Nonlinear analysis of electroencephalogram and magnetoencephalogram recordings in patients with Alzheimer’s disease. Philos. Trans. R. Soc. Lond. A Math. Phys. Eng. Sci. 2009, 367, 317–336. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Ahmed, M.U.; Li, L.; Cao, J.; Mandic, D.P. Multivariate multiscale entropy for brain consciousness analysis. In Proceedings of the IEEE Engineering in Medicine and Biology Society (EMBC), Boston, MA, USA, 30 August–3 September 2011; pp. 810–813. [Google Scholar]
  15. Ahmed, M.U.; Mandic, D.P. Multivariate multiscale entropy: A tool for complexity analysis of multichannel data. Phys. Rev. E 2011, 84, 061918. [Google Scholar] [CrossRef] [PubMed]
  16. Kaffashi, F.; Foglyano, R.; Wilson, C.G.; Loparo, K.A. The effect of time delay on Approximate & Sample Entropy calculations. Phys. D Nonlinear Phenom. 2008, 237, 3069–3074. [Google Scholar]
  17. Rosenblum, M.G.; Pikovsky, A.S.; Kurths, J. Phase Synchronization of Chaotic Oscillators. Phys. Rev. Lett. 1996, 76, 1804–1807. [Google Scholar] [CrossRef] [PubMed]
  18. Rosenblum, M.G.; Pikovsky, A.S. Detecting direction of coupling in interacting oscillators. Phys. Rev. E 2001, 64, 045202. [Google Scholar] [CrossRef] [PubMed]
  19. Rulkov, N.F.; Sushchik, M.M.; Tsimring, L.S.; Abarbanel, H.D.I. Generalized synchronization of chaos in directionally coupled chaotic systems. Phys. Rev. E 1995, 51, 980–994. [Google Scholar] [CrossRef]
  20. Cao, L. Dynamics from multivariate time series. Phys. D Nonlinear Phenom. 1998, 121, 75–88. [Google Scholar] [CrossRef]
  21. Bhattacharya, J.; Pereda, E.; Petsche, H. Effective detection of coupling in short and noisy bivariate data. IEEE Trans. Syst. Man Cyberne. Part B Cybern. 2003, 33, 85–95. [Google Scholar] [CrossRef] [PubMed]
  22. Netoff, T.I.; Schiff, S.J. Decreased neuronal synchronization during experimental seizures. J. Neurosci. 2002, 22, 7297–7307. [Google Scholar] [PubMed]
  23. Hayano, J.; Yasuma, F.; Okada, A.; Akiyoshi, M.; Mukai, S.; Fujinami, T. Respiratory Sinus Arrhythmia. Circulation 1996, 94, 842–847. [Google Scholar] [CrossRef] [PubMed]
  24. Cox, D.R.; Wermuth, N. A simple approximation for bivariate and trivariate normal integrals. Int. Stat. Rev. 1991, 59, 262–269. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the multivariate sample entropy (mSE) algorithm.
Figure 1. Block diagram of the multivariate sample entropy (mSE) algorithm.
Entropy 20 00082 g001
Figure 2. A bivariate time series (upper panel) and its delay vectors with embedding dimension M = [ 1 , 1 ] (lower panel).
Figure 2. A bivariate time series (upper panel) and its delay vectors with embedding dimension M = [ 1 , 1 ] (lower panel).
Entropy 20 00082 g002
Figure 3. The extended multivariate delay vectors (DVs) of the multivariate DVs shown in Figure 2 using existing techniques. The naïve method [14] calculates the pair-wise distances within the two subspaces ( P = 1 and P = 2 ) separately, while the full method [15] calculates the pair-wise distances across all the delay vectors.
Figure 3. The extended multivariate delay vectors (DVs) of the multivariate DVs shown in Figure 2 using existing techniques. The naïve method [14] calculates the pair-wise distances within the two subspaces ( P = 1 and P = 2 ) separately, while the full method [15] calculates the pair-wise distances across all the delay vectors.
Entropy 20 00082 g003
Figure 4. Multivariate sample entropy (mSE) estimate for bivariate white Gaussian noise (WGN), correlated and uncorrelated, using the naïve method [14] (a); and the full or existing method [15] (b). The error bars denote ± standard deviation. We observe that only the full/existing method can distinguish between correlated and uncorrelated variates.
Figure 4. Multivariate sample entropy (mSE) estimate for bivariate white Gaussian noise (WGN), correlated and uncorrelated, using the naïve method [14] (a); and the full or existing method [15] (b). The error bars denote ± standard deviation. We observe that only the full/existing method can distinguish between correlated and uncorrelated variates.
Entropy 20 00082 g004
Figure 5. A bivariate time series exhibiting synchronized regularity. (a): The mixing parameter, which in this instance is the same for each variate. (b): The first variate. (c): The second variate. We observe that when λ i = 1 (Equations (6) and (7)), both variates exhibit 1/f-type structures (i.e., for i = 7000 , , 7500 ), and when λ i = 0 (Equations (6) and (7)), both variates exhibit WGN-type structures (i.e., for i = 8000 , , 11,000).
Figure 5. A bivariate time series exhibiting synchronized regularity. (a): The mixing parameter, which in this instance is the same for each variate. (b): The first variate. (c): The second variate. We observe that when λ i = 1 (Equations (6) and (7)), both variates exhibit 1/f-type structures (i.e., for i = 7000 , , 7500 ), and when λ i = 0 (Equations (6) and (7)), both variates exhibit WGN-type structures (i.e., for i = 8000 , , 11,000).
Entropy 20 00082 g005
Figure 6. Multivariate multiscale sample entropy (mMSE) extension using the existing algorithm for signals with (i) independent mixing parameters, and (ii) synchronized mixing parameters; see model described by Equations (6)–(8). The error bars denote ± standard deviation. A two-tailed two-sample t-test was applied to determine the degree of separation between the two scenarios at each scale factor; instances in which the distributions were found to overlap are denoted by “not-significant” (n.s.).
Figure 6. Multivariate multiscale sample entropy (mMSE) extension using the existing algorithm for signals with (i) independent mixing parameters, and (ii) synchronized mixing parameters; see model described by Equations (6)–(8). The error bars denote ± standard deviation. A two-tailed two-sample t-test was applied to determine the degree of separation between the two scenarios at each scale factor; instances in which the distributions were found to overlap are denoted by “not-significant” (n.s.).
Entropy 20 00082 g006
Figure 7. The extended multivariate delay vectors (DVs) of the multivariate DVs shown in Figure 2 using the proposed DV extension method (see Equation (9)).
Figure 7. The extended multivariate delay vectors (DVs) of the multivariate DVs shown in Figure 2 using the proposed DV extension method (see Equation (9)).
Entropy 20 00082 g007
Figure 8. Multivariate sample entropy (mSE) estimate for bivariate white Gaussian noise (WGN), correlated and uncorrelated, using the proposed algorithm. The error bars denote ± standard deviation. We observe that increasing correlation reduces the SE (cf. Figure 4).
Figure 8. Multivariate sample entropy (mSE) estimate for bivariate white Gaussian noise (WGN), correlated and uncorrelated, using the proposed algorithm. The error bars denote ± standard deviation. We observe that increasing correlation reduces the SE (cf. Figure 4).
Entropy 20 00082 g008
Figure 9. The multivariate multiscale sample entropy (mMSE) extension using the proposed algorithm for signals with (i) independent mixing parameters, and (ii) synchronized mixing parameters (see model described by Equations (6)–(8)). The error bars denote ± standard deviation. A two-tailed two-sample t-test was also applied to determine the degree of separation between the two scenarios at each scale factor; instances in which the distributions were found to overlap are denoted by “not-significant” (n.s.). Compare with the existing approaches in Figure 6.
Figure 9. The multivariate multiscale sample entropy (mMSE) extension using the proposed algorithm for signals with (i) independent mixing parameters, and (ii) synchronized mixing parameters (see model described by Equations (6)–(8)). The error bars denote ± standard deviation. A two-tailed two-sample t-test was also applied to determine the degree of separation between the two scenarios at each scale factor; instances in which the distributions were found to overlap are denoted by “not-significant” (n.s.). Compare with the existing approaches in Figure 6.
Entropy 20 00082 g009
Figure 10. The multivariate sample entropy (mSE) for the original signal with correlated variates (black circle) and its multivariate surrogates (red error bars, which denote ± standard deviation). In the case of the surrogates, the interdependency (correlation) between the variates is removed and thus its average mSE value is higher, indicating the presence of coupled dynamics in the original signal.
Figure 10. The multivariate sample entropy (mSE) for the original signal with correlated variates (black circle) and its multivariate surrogates (red error bars, which denote ± standard deviation). In the case of the surrogates, the interdependency (correlation) between the variates is removed and thus its average mSE value is higher, indicating the presence of coupled dynamics in the original signal.
Entropy 20 00082 g010
Figure 11. The R–R interval as the subject performed different breathing instructions (constrained/ unconstrained) every 30 s. As a result of the phenomenon of respiratory sinus arrhythmia, the breathing instructions induced periods of different regularity into the R–R interval.
Figure 11. The R–R interval as the subject performed different breathing instructions (constrained/ unconstrained) every 30 s. As a result of the phenomenon of respiratory sinus arrhythmia, the breathing instructions induced periods of different regularity into the R–R interval.
Entropy 20 00082 g011
Figure 12. Multivariate sample entropy (mSE; (a)) and cross-correlation (b) results when applied to detect synchronized activity between two trials of R–R interval recordings. In the upper corner of each panel in red, the spectral analysis of each approach reveals the dominant synchronization period; only the mSE approach reveals synchronized regularity dynamics at 0.017 Hz (1/60).
Figure 12. Multivariate sample entropy (mSE; (a)) and cross-correlation (b) results when applied to detect synchronized activity between two trials of R–R interval recordings. In the upper corner of each panel in red, the spectral analysis of each approach reveals the dominant synchronization period; only the mSE approach reveals synchronized regularity dynamics at 0.017 Hz (1/60).
Entropy 20 00082 g012
Table 1. The first two columns denote the degree of separation, calculated using the Bhattacharyya distance (D; see Equation (10)), in multivariate sample entropy (mSE) using the existing and proposed methods for bivariate electroencephalogram (EEG) recordings, between the states of wake (W) and slow-wave sleep (SWS), as annotated by an expert. The last columns denote the percentage of the entire recording annotated as wake (% W) and slow-wave sleep (% SWS).
Table 1. The first two columns denote the degree of separation, calculated using the Bhattacharyya distance (D; see Equation (10)), in multivariate sample entropy (mSE) using the existing and proposed methods for bivariate electroencephalogram (EEG) recordings, between the states of wake (W) and slow-wave sleep (SWS), as annotated by an expert. The last columns denote the percentage of the entire recording annotated as wake (% W) and slow-wave sleep (% SWS).
ExistingProposed% W% SWS
Excerpt 1 D = 1.28 D = 1.66 14.519.9
Excerpt 4 D = 1.99 D = 3.54 27.16.5
Excerpt 5 D = 1.10 D = 2.06 10.828.6
Excerpt 6 D = 1.48 D = 1.61 3.332.3

Share and Cite

MDPI and ACS Style

Looney, D.; Adjei, T.; Mandic, D.P. A Novel Multivariate Sample Entropy Algorithm for Modeling Time Series Synchronization. Entropy 2018, 20, 82. https://doi.org/10.3390/e20020082

AMA Style

Looney D, Adjei T, Mandic DP. A Novel Multivariate Sample Entropy Algorithm for Modeling Time Series Synchronization. Entropy. 2018; 20(2):82. https://doi.org/10.3390/e20020082

Chicago/Turabian Style

Looney, David, Tricia Adjei, and Danilo P. Mandic. 2018. "A Novel Multivariate Sample Entropy Algorithm for Modeling Time Series Synchronization" Entropy 20, no. 2: 82. https://doi.org/10.3390/e20020082

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop