Next Article in Journal
On Higher Order Structures in Thermodynamics
Next Article in Special Issue
Conditional Entropy: A Potential Digital Marker for Stress
Previous Article in Journal
Energy Efficiency Optimization in Massive MIMO Secure Multicast Transmission
Previous Article in Special Issue
Clinical Value of Information Entropy Compared with Deep Learning for Ultrasound Grading of Hepatic Steatosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Entropy of Probability Integral Transformed Time Series

by
Dragana Bajić
1,*,
Nataša Mišić
2,
Tamara Škorić
1,
Nina Japundžić-Žigon
3 and
Miloš Milovanović
4
1
Faculty of Technical Sciences, University of Novi Sad, 21000 Novi Sad, Serbia
2
Research and Development Institute Lola Ltd., 11000 Belgrade, Serbia
3
Faculty of Medicine, University of Belgrade, 11000 Belgrade, Serbia
4
Mathematical Institute of the Serbian Academy of Sciences and Arts, 11000 Beograd, Serbia
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(10), 1146; https://doi.org/10.3390/e22101146
Submission received: 1 September 2020 / Revised: 27 September 2020 / Accepted: 9 October 2020 / Published: 12 October 2020
(This article belongs to the Special Issue Entropy in Biomedical Applications)

Abstract

:
The goal of this paper is to investigate the changes of entropy estimates when the amplitude distribution of the time series is equalized using the probability integral transformation. The data we analyzed were with known properties—pseudo-random signals with known distributions, mutually coupled using statistical or deterministic methods that include generators of statistically dependent distributions, linear and non-linear transforms, and deterministic chaos. The signal pairs were coupled using a correlation coefficient ranging from zero to one. The dependence of the signal samples is achieved by moving average filter and non-linear equations. The applied coupling methods are checked using statistical tests for correlation. The changes in signal regularity are checked by a multifractal spectrum. The probability integral transformation is then applied to cardiovascular time series—systolic blood pressure and pulse interval—acquired from the laboratory animals and represented the results of entropy estimations. We derived an expression for the reference value of entropy in the probability integral transformed signals. We also experimentally evaluated the reliability of entropy estimates concerning the matching probabilities.

1. Introduction

The sampling theorem [1] paved a way for pervasive signal processing within the scientific fields where it was once inconceivable. Tools developed for classical thermodynamics or communications engineering found new multidisciplinary implementation.
The function developed to estimate the uncertainty of the communication signals—entropy [2]—attracted the attention of scientists from a range of different fields. Other entropy concepts were accepted as well—Kolmogorov–Sinai [3], Grassberger et al. [4] and Eckmann et al. [5], despite difficult implementation and firm theoretical framework.
To ensure easy implementation, Pincus [6] proposed the approximate entropy (ApEn) that avoids the rigid mathematical requirements of its theoretical predecessors (hence the name—approximate). The researchers readily accepted ApEn and its modification SampEn (sample entropy, [7]), with a commendation of the rapidly growing number of citations [8].
Medical researchers quickly realized the benefits of signal processing [9] and successfully applied ApEn and SampEn, in particular for cardiovascular signals: for heart rate variability (HRV) analysis in patients with type 2 diabetes [10], in patients with heart failure [11], in healthy subjects [12] during exercise and during resting [13], under stressful conditions [7,14] or for age and gender analysis [15,16]. Entropy became important for quantifying the deterministic chaos of HRV [17], cardiac variability [18], and complexity changes in cardiovascular disease [19], but it can also be applied for monitoring the changes in neurocardiovascular dynamics, e.g., in acute brain injury patients [20].
Cross-entropy (XEn) was a straightforward generalization, derived as a measure that estimates the mutual (un)predictability of two simultaneously recorded and interconnected signals [7,21,22]. XEn can be based on ApEn (XApEn) or SampEn (XSampEn). It came into the focus slower than ApEn and SampEn, but it has found its place in cardiovascular analytics [23,24].
Many contributions have introduced the improvement of the original concepts [14,25,26,27,28] or proposed the alternative ways to approximate the entropy [8,29,30,31,32], but ApEn and SampEn have remained on the top, with the advantages summarized in a comprehensive tutorial [33].
The entropy application proposed in this paper is motivated by a mathematical method based on copula theory [34]. The approach decomposes a multivariate joint distribution of D > 1 signals, each with an arbitrary distribution, into D independent uniform marginals and a function that binds them all—the copulas. The copulas reveal the dependency structure of two or more related signals, while independent marginals, with equalized amplitude distributions, preserve the shape of signal fluctuations. The signal transformation that produces the copulas and the corresponding marginals is called the probability integral transform (PIT, or PI-transform) [35].
The goal of this paper is to investigate the changes of entropy when the amplitude distribution of the time series is equalized, while the temporal fluctuations of the signal amplitudes remain intact. In particular, we aim to show the benefits of the probability integral transformation to the reliability of entropy estimates. We used artificial time series—pseudo-random signals with arbitrary distributions, coupled with statistical or deterministic methods that include generators of statistically dependent distributions, linear and non-linear transforms, and unpredictable (but not random) signals of deterministic chaos. We also used real cardiovascular time series—systolic blood pressure and pulse interval—acquired from laboratory animals.
The paper is organized as follows: Section 2 explains Sklar’s theorem, copula distribution, and probability integral transform that motivated this study. It also gives a brief recap of ApEn, SampEn, and XEn procedures. Section 2 then continues by presenting the artificial signals that were used to verify the proposed concept, including pseudo-random signals with Gaussian, beta, and gamma distributions and deterministic chaos (logistic map [36]). The pairs of signals are mutually coupled with correlation coefficient ranging from zero to one. The dependency of the signal samples is achieved by moving average (MA) filter, or by non-linear equations. The applied coupling methods are checked using statistical tests for correlation and autocorrelation. The changes in signal regularity are verified by a multifractal spectrum. The section concludes with a brief description of the experimental protocol and of the signal acquisition from the laboratory animals. Section 3 first empirically defines the threshold, a parameter crucial for reliable entropy estimation. The results of entropy estimation from the artificial signals and the signals of laboratory animals exposed to stress are presented and discussed. An expression for the entropy of random signals with uniform distribution is derived, a reference value to which the entropy of PI-transformed signals can be compared. This expression can be applied for ApEn, SampEn, and XEn. The reliability of entropy estimates concerning the matching probabilities is also experimentally evaluated. Concluding remarks are given in Section 4.

2. Materials and Methods

This section shows the theory underlying the dependency structures of multivariate time series given by a copula density. The copula density was a motivation to estimate the entropy from signals with equalized amplitudes. The ApEn, SampEn, and XEn procedures are also briefly described. The signals to be analyzed are the cardiovascular time series recorded from rats exposed to various stressful situations. However, since the PI-transform influence to entropy estimates are still unknown, first we need to analyze the artificially generated signals with known statistical properties.

2.1. Probability Integral Transform, Sklar’s Theorem and Copula Density

The probability integral transform (PIT, or PI-transform) converts a random variable (RV) x with an arbitrary distribution function Fx(x) into a RV y uniformly distributed on the segment [0, 1] [35]. The function used for transformation is the distribution function of signal x, i.e., y = Fx(x). The resulting distribution function Fy(y) is uniform (Figure 1). The proof can be found in textbooks on probability and random variables (e.g., p. 139, [37]), but it is included for the completeness:
From Figure 1 it is obvious that the probabilities Pr{xx0} and Pr{yy0} are equal. The same applies to the distribution functions: (Fx(x0) = Pr{xx0}) = (Fy(y0) = Pr{yy0}). Additionally, the PIT transformation rule states that y0 = Fx(x0), so the following may be written:
F y ( y 0 ) = Pr { y y 0 } = Pr { x x 0 } = Pr { x F x 1 ( y 0 ) } = F x ( F x 1 ( y 0 ) ) = y 0
The distribution function Fy(y0) is a linear function of y0. Its derivative is a constant, so the distribution of signal y is indeed uniform.
An illustrative example of signals transformed by PIT is presented in Figure 2, showing the systolic blood pressure (SBP) and pulse interval (PI) of a laboratory rat before and after PIT application.
PIT gained its popularity during the early days of the digital era, as its inverse produces a random signal of arbitrary distribution. When software packages started to provide built-in distribution generators, PIT was almost let into oblivion, but not for long. Sklar’s theorem [38], although derived in the early sixties, came into the research focus at the turn of the century and brought PIT to the forefront again.
Sklar’s theorem states that every D-dimensional (multivariate) distribution function H ( x 01 , x 02 , , x 0 D ) = Pr { x 1 x 01 , , x D x 0 D } can be expressed in terms of its uniform marginals F x i ( x 0 i ) = Pr { x i x 0 i } ,   i = 1 , , D , and a joint distribution—a copula C—that binds them, i.e.,
H ( x 01 , x 02 , , x 0 D ) = C ( F x 1 ( x 01 ) , F x 2 ( x 02 ) , , F x D ( x 0 D ) ) .
An alternative interpretation can be formulated if we recollect that each marginal is uniformly distributed, i.e., F x i ( x 0 i ) = u i ,   i = 1 , , D :
C ( u 1 , u 2 , , u D ) = Pr { x 1 F x 1 1 ( u 1 ) , x 2 F x 2 1 ( u 2 ) , x D F x D 1 ( u D ) } .
Despite the abstract theoretical definition, a copula implementation and interpretation are simple. The copulas are distribution functions, and their derivatives are the probability density functions—the copula density.
The copula density depicts the dependency structure (density of dependency) of the composite signals. An ability to visualize the dependency structure, especially for bivariate signals, is a unique advantage of the copulas density. To estimate the empirical copula density—it is sufficient to apply the probability integral transform to the source signals and find their joint probability density function.
An example of empirical copula density is shown in Figure 3. The left panel shows the classical joint probability distribution function of systolic blood pressure (SBP) and pulse intervals (PI) signals of the laboratory rats. The right panel presents the copula density of the same signals (D = 2), in the abstract two-dimensional [0, 1]D copula plane. The dependency structure reveals the linear relationship between SBP and PI that corresponds to baroreflex, a major regulatory feedback that helps to maintain blood pressure at a nearly constant level.
Another property of the copulas is that they quantify the strength of signal coupling. It differs from other similar procedures because it can handle more than two signals, it can capture non-linear dependencies as well, but, above all, it can be adapted to the properties of the observed signals. There are many copula sets (“copula families” [39]), and each one is adapted to the particular signal type. For example, Frank copulas are the most suitable for cardiovascular signals [34]. It is this feature of the copulas that has brought them great popularity in many research domains, from finances [40], telecommunications [41], civil engineering [42], geodesy, [43] climatology [44], to medicine [45] and cardiology [34].

2.2. XEn, ApEn and SampEn

The procedures for estimating ApEn and SampEn, originally introduced in [6,7], are repeatedly described in most papers that implement them. We shall give a brief recap of XEn as a general procedure and outline the differences in respect to ApEn and SampEn.
XEn [25,46] measures the mutual (un)predictability of time series X and Y, each one comprising N samples:
  • Reference series xiX, i = 1, …, N;
  • Follower series yjY, j = 1, …, N.
If ApEn or SampEn are implemented, there is just a single series X and in the remaining explanation Y = X.
Time series must be pre-processed before further analysis. The signal comparability is ensured by z-normalization (standard scaling—mean signal value and standard deviation reduced to 0 and 1, respectively). The estimation of statistical moments requires the stationary time series, ensured by a filter designed specifically for biomedical time series [47].
Time series are then divided into the overlapping vectors of length m (m is usually 2, 3 or 4):
  • Template vector X m ( i ) = [ x i , x i + 1 , , x i + m 1 ] ,   i = 1 , , N m + 1 ;
  • Follower vector Y m ( j ) = [ y j , y j + 1 , , y j + m 1 ] ,   j = 1 , , N m + 1 ;
A distance between each template X m ( i ) and each follower vector Y m ( j ) is defined as a maximal absolute sample difference:
  • d ( X m ( i ) , Y m ( j ) ) = m a x k = 0 , , m 1 | x i + k     y j + k | ,   i , j = 1 , , N m + 1 .
If the distance is less than or equal to the predefined threshold r, the vectors are declared as similar. The template matching probability p ^ i ( m ) ( r ) is proportional to the number of vectors similar to a particular template vector X m ( i ) :
p ^ i ( m ) ( r ) = Pr { d ( X m , Y m ) r | X m = X m ( i ) ,   Y m Y ,   r > 0 } = 1 N m + 1 · j = 1 N m + 1 I   { d ( X m ( i ) , X m ( i ) ) r } .
The sign “^” in (4) denotes an estimate, while I{·} is an indicator function that is equal to 1 if d ( X m ( i ) , Y m ( j ) ) r , otherwise it is equal to zero. It is used as a mathematical description for the counting process, so (4) estimates the relative frequency of vectors similar to the template X m ( i ) . For SampEn and XSampEn, the step (Nm + 1) is excluded from averaging. SampEn also excluded self-matching, when the template vector X m ( i ) is compared to itself.
Averaging the probabilities is different for ApEn and SampEn, and the corresponding cross-entropies. For (X) ApEn, the logarithms of the probabilities (the information contents of each template [2]) are averaged:
Φ ^ ( m ) ( r , N ) = 1 N m + 1 · i = 1 N m + 1 ln ( p ^ i ( m ) ( r ) ) .
For SampEn, the logarithm is taken from the averaged probabilities:
Ψ ^ ( m ) ( r , N ) = ln ( 1 N m · i = 1 N m p ^ i ( m ) ( r ) ) .
The complete procedure is repeated for the vectors of length m + 1, with summands equal to:
Φ ^ ( m + 1 ) ( r , N ) = 1 N m · i = 1 N m ln ( p ^ i ( m + 1 ) ( r ) ) ,   Ψ ^ ( m + 1 ) ( r , N ) = ln ( 1 N m · i = 1 N m p ^ i ( m + 1 ) ( r ) )
Final entropy estimates are:
( X ) A p E n ^ ( m , r , N ) = Φ ^ ( m ) ( r , N ) Φ ^ ( m + 1 ) ( r , N ) , ( X ) S a m p E n ^ ( m , r , N ) = Ψ ^ ( m ) ( r , N ) Ψ ^ ( m + 1 ) ( r , N ) = ln ( i = 1 N m p ^ i ( m ) ( r ) i = 1 N m p ^ i ( m + 1 ) ( r ) )
While the sample entropy is a robust estimator [7], the approximate entropy can suffer from inconsistencies as it is based on the logarithm of probability estimations with accumulating estimation errors (cf. Equation (5)). The time-series length N and threshold r are both pointed out as the primary cause for inconsistencies [7,14,22,25]. One of the main adverse outcomes is the zero matching probability (a template vector with no similar followers). The various corrections proposed in [8,22] converge towards the true entropy values when time series length N converges to infinity. We implemented a simple correction, which turned out to be close to the true entropy value regardless of N [22]:
Φ ^ C ( m ) ( r , N ) = 1 N m + 1 N 0 · i = 1 ,   p ^ i ( m ) ( r ) 0 N m + 1 ln ( p ^ i ( m ) ( r ) ) ,   N 0 = j = 1 N m + 1 I   { p ^ i ( m ) ( r ) = 0 } .
As already mentioned, this paper aims to investigate the ApEn, SampEn, and XEn of PI-transformed cardiovascular time series, and PIT is closely related to the copula density. However, another entropy measure related to the copulas—the copula entropy—already exists [48,49,50]. It is based on the Shannon entropy [2], with corresponding probabilities evaluated as normalized histograms of the empirical copula density.
The Shannon entropy applied to a time series is a static measure. If the order of signal samples is permuted in time, as in isodistributional surrogate signals [51], the Shannon entropy will remain unaltered because the probability density function remains the same, as it is derived from the amplitude values, regardless of their position in time. Hence, Shannon entropy reflects the level of orderliness in spatial, but not in the temporal domain.
The ApEn-based entropies reveal the signal orderliness both in the spatial domain (via threshold r) and in the temporal domain (via the vector length m and its increase to m + 1). However, if the threshold r is set to zero, it was shown that ApEn is equivalent to the Shannon conditional entropy applied to the first-order Markov chain [6]. This is, however, a theoretical abstraction, as ApEn with r = 0 cannot be practically achieved.
It was also shown that ApEn of differentially coded binary time series for r = 0 is equivalent to Shannon’s binary entropy [52]. This result, although feasible, has no practical application.

2.3. Artificial Time Series

To test the entropy of signals before and after the PI-transformation, we generated random signals with Gaussian, gamma and beta distribution. Gamma and beta distributions, with parameters (α, β) equal respectively to (1, 2) and (3, 1), are skewed distributions with amplitude concentrated in different regions. For each example, we have generated 20 signals or signal pairs, comprising N = 3000 samples each, and the distribution of each signal was tested by the Kolmogorov–Smirnov test. Additionally, we generated signals of deterministic chaos, where their unpredictability is governed by deterministic laws of the simple non-linear equation of the logistic map [36]:
x ( i + 1 ) = R P · x ( i ) · ( 1 x ( i ) )
The parameter RP was chosen to be 3.81, a value that guarantees chaotic behavior over the complete signal range without oscillations. Another value, RP = 3.58, generated a chaotic signal, but omitting some amplitude ranges. The probability distribution function (a normalized histogram) of these two signals is presented in Figure 4. The second signal (RP = 3.58) is an example of the signal that cannot be used for the copulas as it is not continuous, so it does not fulfill the theoretical requirements.
The statistical dependency of signal samples is induced using an MA filter:
x ( k ) ( i ) = 1 k · j = 0 k 1 x ( i + j ) ,   i = 1 , , N k + 1 ,   k = 2 , , 10
Statistically dependent time series with distribution functions F x ( x ) and F y ( y ) were created using the copula method as follows: the original signal points (x, y) are PI-transformed into uniform signal points (ux, uy) and then the corresponding joint distribution is created in the unit plane [0, 1]2 using the Frank copula distribution [34,39]:
C ( u x , u y ) = 1 θ · log { 1 + ( e θ · u x 1 ) · ( e θ · u y 1 ) e θ 1 }
Finally, each ( u x , u y ) point is transformed back to the original signal plane using the transform ( x , y ) = ( F x 1 ( u x ) ,   F y 1 ( u y ) ) . This method generates mutually dependent time series X and Y with distributions F x ( x ) and F y ( y ) , where the dependency level is given by the copula parameter θ.
Figure 5 presents an illustrative example of signals X and Y with skewed gamma and beta distributions, generated using copula generator with parameter θ = 3. From their joint probability density function (PDF) (Figure 5c) no conclusions can be drawn about the relationship of X and Y, but strong linear coupling is clearly visible in their dependency structure (empirical copula density, Figure 5d).
Deterministic non-linear dependency is introduced using relationships Y = a·XEXP + b, where a = b = 1 are arbitrary chosen parameters, while parameter EXP ranges from 1.1 to 2.
Figure 6a,b show the linear signal coupling estimated by the Pearson, Kendall and Spearman tests [53]. All three tests use different procedures—the Pearson test uses classical moment theory, while the Kendall and Spearman tests use different ranking procedures. The aim was to examine whether the PI transformation changes the linear coupling of two data series. The corresponding data series are generated using a copula generator and an MA filter.
Figure 6a,b show that the PI transformation does not cause changes in the linear dependence of the two signals. The dependence between adjacent samples of a single signal also remains unchanged after PI-transformation, as shown by the autocorrelation function in Figure 6d. However, the Pearson, Spearman, and Kendall tests, designed to capture correlation, are unable to capture nonlinear dependence, Figure 6c.
The correlation lines of source signals and their PI-transformed counterparts in Figure 6 overlap perfectly, showing that PIT does not alter the coupling of the signal pairs.
To test whether the differentiation would be possible at all, we estimated a multifractal spectrum that describes the fluctuation of the local regularity of the observed signals. The multifractal spectrum is estimated in terms of wavelet leaders [54].
The results in Figure 7 reveal that the PIT induced the changes in signal regularity. The inset in the upper right corner of Figure 7a shows that the spectra of the transformed pseudorandom signals differ from the original spectra; besides, all transformed spectra overlap as the signals get the same distribution (inset in Figure 7a). The spectrum of deterministic chaos reveals monofractal properties. The spectrum remains monofractal after PI-transform (Figure 7b). The spectra of signals filtered by the MA filter differ from the source signals, showing that the local regularity of the signals has changed.

2.4. Time Series Recorded from the Laboratory Rats Exposed to Shaker and Restraint Stress

The real cardiovascular signals we used to check the PIT entropy concept were recorded at Laboratory of Cardiovascular Pharmacology, Medical Faculty, University of Belgrade, from outbred male Wistar rats weighing 330 ± 20 g.
Ten days before the experiments, radio-telemetric probes (TA11PA-C40, DSI, Transoma Medical, St. Paul, MN, USA) were implanted into the abdominal aorta under combined ketamine and xylazine anesthesia, along with gentamicin and followed by metamizole injections for pain relief. The arterial blood pressure (BP) signal was digitized at 1000 Hz and relayed to a computer equipped with Dataquest A.R.T. 4.0 software for analysis of cardiovascular signals.
Rats were randomized into two groups. The first group was exposed to shaker stress, with rats positioned on a platform shaking at 200 cycle/min. The second group was exposed to restraint stress, with rats placed in a Plexiglas restrainer tube (ID 5.5 cm with pores) in the supine position. Arterial blood pressure (BP) waveforms were recorded before (CONTROL) and after the first exposure to stress (SHAKER, RESTRAINT). Other phases of the experimental protocol are not relevant for this study [55].
Systolic blood pressure was derived from the arterial BP as local maxima in the BP waveforms, while the pulse interval (PI) time series was derived as the time distance between successive maximal arterial blood pressure increases. Artifacts were removed semi-automatically, first using the filter designed for cardiovascular time series [56] and then carefully visually examining the BP waveforms and residual artifacts. The signals from rats with traces of unstable health were completely excluded. A very slow varying signal component (mostly the result of rat relocation) was removed using a filter proposed by [47]. De-trended time series should be stationary at least in a wide-sense, i.e., their first and second statistical moment should be time-invariant. Then the mean value and standard deviation estimated from the time series are equal to their statistical counterparts [37], and only then the standard scaling could be reliably implemented. Thus, de-trended time series were checked using a stationarity test [57,58] and those that were not wide sense stationary were eliminated.
The final number of remaining animals per experimental group was n = 6. It was satisfactory according to the variability of the parameters in the control group rats (the statistical software “Power Sample Size Calculation”). All experimental procedures in this study were confirmed by the European Communities Council directive of 24 November 1986 (86/609/ECC), and the School of Medicine, University of Belgrade, Guidelines on Animal Experimentation.

3. Results and Discussion

This section presents the results of our study. Each set of the results is accompanied with the corresponding discussion.

3.1. Threshold Choice

As already pointed out, the threshold value is crucial for the consistency of entropy estimates, and its proper choice should be the first task. However, entropy is also a function of the time series length N—shorter time series require a larger threshold, and the relationship is not linear. A thorough analysis in [22] showed that a reliable estimation of probabilities (4) is a key factor for stable entropy measures. This requires that the threshold values be higher than the generally accepted ones. One of the methods is to plot a threshold profile, i.e., to estimate the entropy for different threshold values and fixed N. Figure 8 presents the XEn profile estimated from the cardiovascular signals, as cross-entropy requires higher threshold values than self-entropy [14,22]. Besides, real signals are a better choice for threshold profiling than stationary artificial data. The vertical lines in Figure 8 show the threshold for which the entropy estimates become consistent. This threshold value is equal to r = 0.3 and is adopted for further entropy estimation. The threshold evaluation method proposed in [22] gives higher threshold values, but in present study, we did not want to differ too much from the classical values.

3.2. Entropy Estimated from Artificial Data

The purpose of artificial data, generated in controlled conditions, is to present the reference regarding the probability integral transformation entropy estimates.
Self-entropy estimates (ApEn and SampEn) are presented in Figure 9. As expected, the entropy of pseudo-random signals depends on their distribution (Figure 9a), but PI-transform eliminates this dependency (Figure 9b). The chosen parameters (N = 3000, r = 0.3) are sufficient to ensure reliable entropy estimates. Obviously, no correction is needed as the original and corrected ApEn estimates perfectly overlaps.
The dependency of signal samples, induced by the MA filter, causes an entropy decrease. The decrease is indeed due to the sample correlation, as the signal distribution remains Gaussian and decrease if entropy is seen both for original signals, and the PI-transformed signals (Figure 9c). On the other hand, the non-linear transform induced by the relation Y = a·XEXP + b also decreases the entropy, but this decrease is due to the distribution change: PI-transform converts the distribution into uniform, and the entropy of all converted signals remains stable, regardless of the exponent EXP (Figure 9d).
Figure 10 shows cross-entropy estimates. The correlation of the signal pairs in Figure 10a has been verified using Spearman, Pearson and Kendall tests (shown in Figure 6a). However, their XEn estimates are constant, revealing that the entropy does not reflect the statistical correlation between two pseudo-random time series. It is in accordance with the entropy procedure, where a template vector is compared with each one of its followers, while the dependency exists only with the followers in its vicinity.
When correlation is induced by the MA filter (Figure 10b), both XApEn and PIT-XApEn decrease with the increase of filter length. SampEn decreases as well, but at a lower rate, while PIT-SampEn remains stable. SampEn is known to be a stable measure, due to the logarithm taken from the average matching probabilities [7]. However, this stability reduces the ability to recognize the subtle changes, and equalizing the distribution further reduces the possibility of recognition, so it might be a disadvantage.
The non-linear relationship between the signals is induced by the relation Y = a·XEXP + b. The corresponding cross-entropies are independent of the level of exponent EXP, if the reference signal is a pseudo-random signal X (source signal). If the reference signal is signal Y, obtained by a non-linear transform of signal X, the cross-entropy decreases for XApEn and XSampEn, but the PIT counterparts remain constant. The reason is the same as for the Figure 9d: the decrease is due to the distribution of reference signal that changes as a consequence of non-linear transform; had the decrease been due to the induced non-linear coupling, the PIT entropy would have changed as well; since it has not changed, the non-linear coupling is not responsible for entropy decrease.
The artificial time series have so far been pseudo-random signals with a given distribution. The deterministic chaos exhibits unpredictability, but not randomness. Figure 11 presents the entropy estimated from the logistic map signals. The entropy of the non-continual chaotic signal (Figure 11a) is very low, revealing a low level of its uncertainty. The second signal (Figure 11b) is genuinely chaotic and it depends on the initial conditions. For this reason, the estimated entropy values are not constant, but their changes are not significant (Figure 11b).
The absolute values of cross-entropy estimates are similar to the self-entropy in Figure 11b. This means that less predictable signal (with higher entropy) is dominant in cross-entropy estimates. On the other hand, the consistency of repeated entropy estimation governed by the reference signal: consistency of X vs. Y cross-entropy is similar to the consistency of X self-entropy (Figure 11a,c); variability of Y vs. X cross-entropy is proportional to the variability of Y self-entropy (Figure 11b,d).

3.3. Entropy Estimated from Cardiovascular Signals of Laboratory Rats Exposed to Stress

The parameters recorded from the laboratory rats are presented in Table 1. It reveals a significant decrease in pulse interval (increase in heart rate) in rats exposed to the restraint stress, other changes are slight and not significant.
The entropy estimates are given in Figure 12.
The horizontal line in Figure 12 shows the theoretical value of ApEn, SampEn, and XEn that can be evaluated for the random signals with uniform distribution:
A p E n ( u n i f , r ) = S a m p E n ( u n i f , r ) = X E n ( u n i f , r ) = ln ( 4 · 3 · r r 2 12 )
A detailed evaluation of the theoretical value is shown in Appendix A. This is a result of perfect randomness and uniformity that can serve as a reference value, without the need to run the tedious simulation studies, e.g., surrogate data tests [51].
Considering the experimental results, the entropy of PI-transformed signals captured slightly more statistically significant differences between the cardiovascular parameters of the animals before and after exposure to stress: while classical entropies found the differences in shaker stress—SBP and PI vs. SBP, and XSampEn found a difference in SBP vs. PI, PIT entropy found additional differences in XApEn of SBP vs. PI, and, in restraint stress, in PI vs. SBP and XSampEn in SBP vs. PI.
Contrary to the artificial signals with the controlled outcome, the reliability of the entropy estimated from real data sources is always a subject of discussion. As already stated, a failure in estimating the matching probabilities, Equation (4), leads to an inconsistent entropy estimation. The reliability of probabilities can be checked using the Jeruchim criterion that defines the minimal signal length required to achieve p ^ i ( m ) ( r ) within a 95% confidence interval [59]. A comprehensive theoretical analysis confirmed the traditional engineering rule that the signal length required for a reliable estimation of a binary event probability should be at least 10 / p ^ i ( m ) ( r ) [59].
The ultimate case of unreliability is the matching probability equal to zero, p ^ i ( m ) ( r ) = 0 . This occurs if the signal length N is too short, or if the threshold r is inadequate, or if the vector length m is too long [22]. However, zero probability can occur in XEn for a completely logical reason: the template vector can comprise amplitudes that can never be found in another signal, so no follower vectors exists. In this case, the zero probability is not a result of an incorrect estimation, but a valid relationship between the two signals.
Figure 13 shows the percentage of reliably estimated matching probabilities, while Figure 14 shows the estimated percentage of zero probabilities.
From both figures, it can be seen that PIT signals have better performances than source signals. The increased number of reliably estimated probabilities in Figure 13 is an outcome of the uniform distribution. The signal amplitudes are equally probable, so the probability that a template finds a matching follower is increased. In distributions with exhibited tails (source signals), some of the templates are less likely to find a matching follower.
The decreased number of zero-matching probability (Figure 14) is another benefit of the probability integral transform. As already said, the distributions of signal pairs for cross-entropy can have non-overlapping segments, so some of the templates will never find the followers. After the PI-transform, the signals would be mapped into the same [0, 1] segment, and non-overlapping segments would not exist.
Figure 13 and Figure 14 also reveal the empirically obtained threshold r = 0.3, although slightly exceeding the traditional values from the literature (0.15–0.25), might not be sufficient for XEn as the values of XApEn and XApEn with correction differ. It is in accordance with the theoretical findings from [22], but we preferred to use the values that are more aligned with the traditional ones.

4. Conclusions

The aim of this paper was to apply the ApEn-based entropies and cross-entropies to the signals submitted to the probability integral transformation. PIT yields the signal with uniform distribution, keeping the signal fluctuations intact. The idea was to eliminate the influence of amplitude distribution, and to estimate the entropy where each amplitude has equal opportunity. Then the true unpredictability of the signal could be estimated without the bias induced by amplitude distribution.
The artificial environment revealed that PIT self-entropy estimates are insensitive to the linear or non-linear signal transformation, if the transformation is induced sample by sample (relationship Y = a·XEXP + b). However, entropy estimates are sensitive to transformations that induce the dependency along the signal itself, e.g., using the MA filter. Considering the cross-entropy, its estimates remain constant when correlation coefficient between the signals X and Y increase from 0 to 1, with a conclusion that statistical correlation cannot be measured by the means of cross-entropy. Cross-entropy, on the other hand, notices if one of the signal is formed from another by inducing the correlation between its successive samples.
The chaotic signals are generated using the formula for deterministic chaos. “Chaos” did not deceive the entropy procedure, so the entropy estimates were quite low, showing the high level of signal predictability. Regardless of apparent chaotic signal appearance, the “deterministic” component could not escape the unbiased entropy measure.
Estimates of the real signals showed that PIT results of signals in stress reveal a slightly increased statistical significance than classical entropy measures. However, the main outcome is the increased estimation reliability, compared to the classical measures. The increased reliability is a consequence of the uniform amplitude distribution over [0, 1] segment and reduced number of zero-matching probabilities.
The entropy estimates of PI-transformed signals are unbiased regarding the amplitude distribution. Their reliability has improved, and a referent value—a ground truth to which entropy estimates can be compared—can be obtained by formula and not by a simulation study.
The future work will be devoted to the evaluation of errors in entropy estimation for ApEn, SampEn, XEn, and their PIT pairs, and to developing the methods for error attenuation. The future work will also include the continuation on thresholds role in inconsistency of entropy estimation.

Author Contributions

Conceptualization, D.B., N.J.-Ž.; methodology, D.B., N.J.-Ž.; software, T.Š.; validation, N.M.; formal analysis, N.M.; investigation, D.B.; resources, M.M.; data curation, N.J.-Ž.; writing—original draft preparation, D.B.; writing—review and editing, N.M.; D.B.; supervision, N.M.; project administration, D.B.; funding acquisition, D.B. All authors have read and agreed to the published version of the manuscript.

Funding

The work of Dragana Bajic was supported in part by grant 451-03-68/2020-14/200156 of the Ministry of Education, Science and Technological Development of the Republic of Serbia.

Acknowledgments

The authors are grateful to junior staff of Laboratory for Cardiovascular Pharmacology, Faculty of Medicine, University of Belgrade for their help with laboratory animals.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

ApEn, SampEn and XEn of PIT transformed signals have an explicit theoretical value ground truth to which they can be compared.
XEn requires the signals with zero mean and standard deviation so the PIT signals have to be z-normalized (standard scale, i.e., zero mean value and unit standard deviation). The uniform probability density function of signals x and y is defined as:
f x ( x ) = f y ( y ) = { 1 2 · a ,   a x , y a 0 ,      elsewhere ;   a = 3 .
The probability that two signal samples are “similar”, i.e., that their absolute difference is below the threshold r is equal to:
p = P r { | x y | r } = a a f x ( x ) · ( y r y + r f y ( y ) · d y ) · d x = = 1 4 · a 2 · ( a a + r ( r + a + x ) · d x + a + r a r 2 · r · d x + a r a ( r + a x ) · d x ) = = 4 · a · r r 2 4 · a 2 = 4 · 3 · r r 2 12 .
The probability that all m pairs of samples are similar (matching probability, Equation (4)) is equal to:
p ^ i m ( r ) = p m = ( 4 · 3 · r r 2 12 ) m .
The summand Φ ( m ) for (X)ApEn then becomes:
Φ ( m ) ( r , N ) = 1 N m + 1 · i = 1 N m + 1 ln ( p ^ i ( m ) ( r ) ) = 1 N m + 1 · i = 1 N m + 1 ln ( p m ) = N m + 1 N m + 1 · m · ln ( p ) = m · ln ( p )
Similarly, Φ ( m + 1 ) ( r , N ) = ( m + 1 ) · ln ( p ) , so, for the case of random time series with uniform distribution, entropy estimates are equal to:
A p E n = X A p E n = Φ ( m ) ( r , N )   Φ ( m + 1 ) ( r , N ) = m · ln ( p ) ( m + 1 ) · ln ( p ) = ln ( p )
where p is given with (A2).
The results for (X) SampEn are the same:
S a m p E n = X S a m p E n = Ψ ( m ) ( r , N ) Ψ ( m + 1 ) ( r , N ) = ln ( i = 1 N m p ^ i ( m ) ( r ) i = 1 N m p ^ i ( m + 1 ) ( r ) ) = ln ( ( N m ) · p m ( N m ) · p m + 1 ) = ln ( p ) .

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 623–656. [Google Scholar] [CrossRef]
  2. Shannon, C.E. Communications in the presence of noise. Proc. IRE 1949, 37, 10–21. [Google Scholar] [CrossRef]
  3. Gustavo, D.; Bernd, S. Information Dynamics: Foundations and Applications, 1st ed.; Springer: New York, NY, USA, 2012; p. 101. ISBN 978-1-4613-0127-1. [Google Scholar]
  4. Grassberger, P.; Procaccia, I. Estimation of the Kolmogorov entropy from a chaotic signal. Phys. Rev. A 1983, 28, 2591–2593. [Google Scholar] [CrossRef] [Green Version]
  5. Eckmann, J.-P.; Ruelle, D. Ergodic theory of chaos and strange attractors. Rev. Mod. Phys. 1985, 57, 617–656. [Google Scholar] [CrossRef]
  6. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef] [Green Version]
  7. Richman, J.S.; Moorman, J.R. Physiological time−series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [Green Version]
  8. Yentes, J.M.; Hunt, N.; Schmid, K.K.; Kaipust, J.P.; McGrath, D.; Stergiou, N. The appropriate use of approximate entropy and sample entropy with short data sets. Ann. Biomed. Eng. 2013, 41, 349–365. [Google Scholar] [CrossRef]
  9. Jelinek, H.F.; Cornforth, D.J.; Khandoker, A.H. ECG Time Series Variability Analysis: Engineering and Medicine, 1st ed.; CRC Press, Taylor & Francis Group: Boca Raton, FL, USA, 2017; pp. 1–450. ISBN 978-1482243475. [Google Scholar]
  10. Li, X.; Yu, S.; Chen, H.; Lu, C.; Zhang, K.; Li, F. Cardiovascular autonomic function analysis using approximate entropy from 24-h heart rate variability and its frequency components in patients with type 2 diabetes. J. Diabetes Investig. 2015, 6, 227–235. [Google Scholar] [CrossRef] [Green Version]
  11. Krstacic, G.; Gamberger, D.; Krstacic, A.; Smuc, T.; Milicic, D. The chaos theory and non-linear dynamics in heart rate variability in patients with heart failure. In Proceedings of the Computers in Cardiology, Bologna, Italy, 14–17 September 2008; pp. 957–959. [Google Scholar] [CrossRef]
  12. Storella, R.J.; Wood, H.W.; Mills, K.M.; Kanters, J.K.; Højgaard, M.V.; Holstein-Rathlou, N.-H. Approximate entropy and point correlation dimension of heart rate variability in healthy subjects. Integr. Psychol. Behav. Sci. 1998, 33, 315–320. [Google Scholar] [CrossRef]
  13. Tulppo, M.P.; Makikallio, T.H.; Takala, T.E.; Seppanen, T.; Huikuri, H.V. Quantitative beat-to-beat analysis of heart rate dynamics during exercise. Am. J. Physiol. Heart Circ. Physiol. 1996, 271, H244–H252. [Google Scholar] [CrossRef]
  14. Boskovic, A.; Loncar-Turukalo, T.; Sarenac, O.; Japundžić-Žigon, N.; Bajić, D. Unbiased entropy estimates in stress: A parameter study. Comput. Biol. Med. 2012, 42, 667–679. [Google Scholar] [CrossRef] [PubMed]
  15. Ryan, S.M.; Goldberger, A.L.; Pincus, S.M.; Mietus, J.; Lipsitz, L.A. Gender- and age-related differences in heart rate dynamics: Are women more complex than men? J. Am. Coll. Cardiol. 1994, 24, 1700–1707. [Google Scholar] [CrossRef] [Green Version]
  16. Wang, S.Y.; Zhang, L.F.; Wang, X.B.; Cheng, J.H. Age dependency and correlation of heart rate variability, blood pressure variability and baroreflex sensitivity. J. Gravit. Physiol. 2000, 7, 145–146. [Google Scholar]
  17. Sunkaria, R.K. The deterministic chaos in heart rate variability signal and analysis techniques. Int. J. Comput. Appl. 2011, 35, 39–46. [Google Scholar] [CrossRef]
  18. Marwaha, P.; Sunkaria, R.K. Cardiac variability time-series analysis by sample entropy and multiscale entropy. Int. J. Med. Eng. Inform. 2015, 7, 1–14. [Google Scholar] [CrossRef]
  19. Chen, C.; Jin, Y.; Lo, I.L.; Zhao, H.; Sun, B.; Zhao, Q.; Zhang, X.D. Complexity change in cardiovascular disease. Int. J. Biol. Sci. 2017, 13, 1320–1328. [Google Scholar] [CrossRef]
  20. Papaioannou, V.; Giannakou, M.; Maglaveras, N.; Sofianos, E.; Giala, M. Investigation of heart rate and blood pressure variability, baroreflex sensitivity, and approximate entropy in acute brain injury patients. J. Crit. Care 2008, 23, 380–386. [Google Scholar] [CrossRef]
  21. Pincus, S.M.; Mulligan, T.; Iranmanesh, A.; Gheorghiu, S.; Godschalk, M.; Veldhuis, J.D. Older males secrete lutenizing hormone and testosterone more irregularly, and jointly more asynchronously than younger males. Proc. Natl. Acad. Sci. USA 1996, 93, 14100–14105. [Google Scholar] [CrossRef] [Green Version]
  22. Skoric, T.; Sarenac, O.; Milovanovic, B.; Japundzic-Zigon, N.; Bajic, D. On consistency of cross-approximate entropy in cardiovascular and artificial environments. Complexity 2017, 1–15. [Google Scholar] [CrossRef] [Green Version]
  23. Lin, T.-K.; Chien, Y.-H. Performance evaluation of an entropy-based structural health monitoring system utilizing composite multiscale cross-sample entropy. Entropy 2019, 21, 41. [Google Scholar] [CrossRef] [Green Version]
  24. Castiglioni, P.; Parati, G.; Faini, A. Information-domain analysis of cardiovascular complexity: Night and day modulations of entropy and the effects of hypertension. Entropy 2019, 21, 550. [Google Scholar] [CrossRef] [Green Version]
  25. Chon, K.H.; Scully, C.G.; Lu, S. Approximate entropy for all signals. IEEE Eng. Med. Biol. Mag. 2009, 28, 18–23. [Google Scholar] [CrossRef] [PubMed]
  26. Lu, S.; Chen, X.; Kanters, J.K.; Solomon, I.C.; Chon, K.H. Automatic selection of the threshold value r for approximate entropy. IEEE Trans. Biomed. Eng. 2008, 55, 1966–1972. [Google Scholar] [CrossRef] [PubMed]
  27. Kaffashi, F.; Foglyano, R.; Wilson, C.G.; Loparo, K.A. The effect of time delay on approximate and sample entropy calculations. Physica D 2008, 237, 3069–3074. [Google Scholar] [CrossRef]
  28. Porta, A.; Baselli, G.; Liberati, D.; Montano, N.; Cogliati, C.; Gnecchi-Ruscone, T.; Malliani, A.; Cerutti, S. Measuring regularity by means of a correctesconditional entropy in sympathetic outflow. Biol. Cybern. 1998, 78, 71–78. [Google Scholar] [CrossRef] [PubMed]
  29. Costa, M.; Goldberger, A.L.; Peng, C.-K. Multiscale entropy analysis of biological signals. Phys. Rev. E 2005, 71, 1–18. [Google Scholar] [CrossRef] [Green Version]
  30. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of surface EMG signal based on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef]
  31. Szczesna, A. Quaternion entropy for analysis of gait data. Entropy 2019, 21, 79. [Google Scholar] [CrossRef] [Green Version]
  32. Behrendt, S.; Dimpfl, T.; Peter, F.J.; Zimmermann, D.J. RTransferEntropy—Quantifying information flow between different time series using effective transfer entropy. SoftwareX 2019, 10, 100265. [Google Scholar] [CrossRef]
  33. Delgado-Bonal, A.; Marshak, A. Approximate entropy and sample entropy: A comprehensive tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef] [Green Version]
  34. Jovanovic, S.; Skoric, T.; Sarenac, O.; Milutinovic-Smiljanic, S.; Japundzic-Zigon, N.; Bajic, D. Copula as a dynamic measure of cardiovascular signal interactions. Biomed. Signal Process Control 2018, 43, 250–264. [Google Scholar] [CrossRef]
  35. Embrechts, P.; Hofert, M. A note on generalized inverses. Math. Methods Oper. Res. 2013, 77, 423–432. [Google Scholar] [CrossRef] [Green Version]
  36. May, R. Simple mathematical models with very complicated dynamics. Nature 1976, 261, 459–467. [Google Scholar] [CrossRef] [PubMed]
  37. Pappoulis, A.; Pillai, U. Probability, Random Variables and Stochastic Processes, 4th ed.; McGraw-Hill: New York, NY, USA, 2002; p. 139. [Google Scholar]
  38. Sklar, A. Fonctions de répartition à n dimensions et leurs marges. Publ. del’Inst. Stat. Univ. Paris 1959, 8, 229–231. [Google Scholar]
  39. Nelsen, R.B. An Introduction to Copulas, 2nd ed.; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  40. Aas, K.; Czado, C.; Bakken, H. Pair-copula constructions of multiple dependence. Insur. Math. Econ. 2009, 44, 182–198. [Google Scholar] [CrossRef] [Green Version]
  41. Zheng, C.; Egan, M.; Clavier, L.; Peters, G.; Gorce, J.-M. Copula-based interference models for IoT wireless networks. In Proceedings of the ICC 2019—53rd IEEE International Conference on Communications, Shanghai, China, 20–24 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  42. Yang, S.C.; Liu, T.J.; Hong, H.P. Reliability of tower and tower-line systems under spatiotemporally varying wind or earthquake loads. J. Struct. Eng. 2017, 143, 04017137. [Google Scholar] [CrossRef]
  43. Modiri, S.; Belda, S.; Hoseini, M.; Heinkelmann, R.; Ferrándiz, J.M.; Schuh, H. A new hybrid method to improve the ultra-short-term prediction of LOD. J. Geod. 2020, 94, 1. [Google Scholar] [CrossRef] [Green Version]
  44. Lazoglou, G.; Anagnostopoulou, C. Joint distribution of temperature and precipitation in the Mediterranean, using the Copula method. Theor. Appl. Climatol. 2019, 135, 1399–1411. [Google Scholar] [CrossRef]
  45. Iyengar, S.G.; Dauwels, J.; Varshney, P.K.; Cichocki, A. Quantifying EEG synchrony using copulas. Proc. IEEE Int. Conf. Acoust. Speech Signal Process. 2010, 505–508. [Google Scholar] [CrossRef]
  46. Wu, H.-T.; Lee, C.-Y.; Liu, C.-C.; Liu, A.-B. Multiscale Cross-Approximate Entropy Analysis as a Measurement of Complexity between ECG R-R Interval and PPG Pulse Amplitude Series among the Normal and Diabetic Subjects. Comput. Math. Methods Med. 2013, 2013, 231762. [Google Scholar] [CrossRef]
  47. Tarvainen, M.P.; Ranta-aho, P.O.; Karjalainen, P.A. An advanced detrending approach with application to HRV analysis. IEEE Trans. Biomed. Eng. 2002, 42, 172–174. [Google Scholar] [CrossRef] [PubMed]
  48. Hao, Z.; Singh, V.P. Entropy-copula method for single-site monthly streamflow simulation. Water Resour. Res. 2012, 48, 1–8. [Google Scholar] [CrossRef]
  49. Ma, J.; Sun, Z. Mutual information is copula entropy. Tsinghua Sci. Technol. 2011, 16, 51–54. [Google Scholar] [CrossRef]
  50. Ince, R.A.A.; Giordano, B.L.; Kayser, C.; Rousselet, G.A.; Gross, J.; Schyns, P.G. A statistical framework for neuroimaging data analysis based on mutual information estimated via a gaussian copula. Hum. Brain Mapp. 2017, 38, 1541–1573. [Google Scholar] [CrossRef] [PubMed]
  51. Blaber, A.P.; Yamamoto, Y.; Hughson, R.L. Methodology of spontaneous baroreflex relationship assessed by surrogate data analysis. Am. J. Physiol. Heart Circ. Physiol. 1995, 268, H1682–H1687. [Google Scholar] [CrossRef] [PubMed]
  52. Skoric, T.; Mohamoud, O.; Milovanovic, B.; Japundzic-Zigon, N.; Bajic, D. Binarized cross-approximate entropy in crowdsensing environment. Comput. Biol. Med. 2017, 80, 137–147. [Google Scholar] [CrossRef]
  53. Agresti, A. Analysis of Ordinal Categorical Data, 2nd ed.; John Wiley & Sons: New York, NY, USA, 2010; ISBN 978-0-470-08289-8. [Google Scholar]
  54. Nguyen, F.; Laval, J.-P.; Kestener, P.; Cheskidov, A.; Shvydkoy, R.; Dubrulle, B. Local estimates of Hölder exponents in turbulent vector fields. Phys. Rev. E 2019, 99, 053114. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Sarenac, O.; Lozic, M.; Drakulic, S.; Bajic, D.; Paton, J.F.; Murphy, D.; Japundzic-Zigon, N. Autonomic mechanisms underpinning the stress response in borderline hypertensive rats. Exp. Physiol. 2011, 96, 574–589. [Google Scholar] [CrossRef] [Green Version]
  56. Wessel, N.; Malberg, H.; Bauernschmitt, R.; Kurths, J. Nonlinear methods of cardiovascular physics and their clinical applicability. Int. J. Bifurc. Chaos 2007, 17, 3325–3371. [Google Scholar] [CrossRef] [Green Version]
  57. Carvajal, R.; Żebrowski, J.J.; Vallverdu, M.; Baranowski, R.; Chojnowska, L.; Poplawska, W.; Caminal, P. Dimensional analysis of HRV in hypertrophic cardiomyopathy patients. IEEE Eng. Med. Biol. Mag. 2002, 21, 71–78. [Google Scholar] [CrossRef]
  58. Bendat, J.S.; Piersol, A.G. Random Data Analysis and Measurement Procedures; Wiley Series in Probability and Statistics; John Wiley & Sons: New York, NY, USA, 1986. [Google Scholar]
  59. Jeruchim, M.C. Techniques for estimating the bit error rate in the simulation of digital communication systems. IEEE J. Sel. Areas Commun. 1984, 2, 153–170. [Google Scholar] [CrossRef]
Figure 1. Probability integral transform. The random variable x with the density function fx(x) is transformed into the uniformly distributed random variable y = Fx(x), with uniform density fy(y) = 1.
Figure 1. Probability integral transform. The random variable x with the density function fx(x) is transformed into the uniformly distributed random variable y = Fx(x), with uniform density fy(y) = 1.
Entropy 22 01146 g001
Figure 2. Waveforms recorded from a laboratory rat before and after the probability integral transform: (a) Systolic blood pressure (SBP) and corresponding probability integral transform (PIT) signal and (b) pulse interval (PI) and corresponding PIT signal. Note that fluctuations of PIT signals follow the fluctuations of the original signals. The insets in the top right corner illustrate the uniformity of the PIT amplitudes along the entire recording time.
Figure 2. Waveforms recorded from a laboratory rat before and after the probability integral transform: (a) Systolic blood pressure (SBP) and corresponding probability integral transform (PIT) signal and (b) pulse interval (PI) and corresponding PIT signal. Note that fluctuations of PIT signals follow the fluctuations of the original signals. The insets in the top right corner illustrate the uniformity of the PIT amplitudes along the entire recording time.
Entropy 22 01146 g002
Figure 3. (a) Joint probability density function of systolic blood pressure (SBP) and pulse interval (PI) of a laboratory rat; (b) copula density of PI-transformed SBP and PI signals. Note that the linear dependency structure along the diagonal in (b) is consistent with the linear cardiovascular relationship of SBP and PI in healthy subjects.
Figure 3. (a) Joint probability density function of systolic blood pressure (SBP) and pulse interval (PI) of a laboratory rat; (b) copula density of PI-transformed SBP and PI signals. Note that the linear dependency structure along the diagonal in (b) is consistent with the linear cardiovascular relationship of SBP and PI in healthy subjects.
Entropy 22 01146 g003
Figure 4. Normalized histogram of logistic map deterministic chaos time series. (a): parameter RP = 3.81. (b): parameter RP = 3.58. The signal in (b) is not continuous, so it does not fulfill the theoretical requirements for copula estimation.
Figure 4. Normalized histogram of logistic map deterministic chaos time series. (a): parameter RP = 3.81. (b): parameter RP = 3.58. The signal in (b) is not continuous, so it does not fulfill the theoretical requirements for copula estimation.
Entropy 22 01146 g004
Figure 5. Statistical properties of skewed distributions. (a) Normalized histogram (empirical PDF) of the gamma (1, 2) distribution; (b) normalized histogram of the beta (3, 1) distribution; (c) joint 2D PDF; (d) copula density. Note that the linear statistical dependency induced by the copula parameter θ = 3 cannot be recognized in the joint probability density function (c), but it is visible in the dependency structure of the copula density (d).
Figure 5. Statistical properties of skewed distributions. (a) Normalized histogram (empirical PDF) of the gamma (1, 2) distribution; (b) normalized histogram of the beta (3, 1) distribution; (c) joint 2D PDF; (d) copula density. Note that the linear statistical dependency induced by the copula parameter θ = 3 cannot be recognized in the joint probability density function (c), but it is visible in the dependency structure of the copula density (d).
Entropy 22 01146 g005aEntropy 22 01146 g005b
Figure 6. Classical properties of source signals and PI-transformed signals. (a) Correlation coefficients of the gamma–beta dependent signal pairs; (b) correlation coefficient of Gaussian signal pairs with dependency induced by the moving average (MA) filter; (c) correlation coefficient of non-linear dependencies; (d) autocorrelation function of a single signal filtered by the MA filter. The results are presented as the mean ± SE (standard error of mean) of 20 signals or signal pairs.
Figure 6. Classical properties of source signals and PI-transformed signals. (a) Correlation coefficients of the gamma–beta dependent signal pairs; (b) correlation coefficient of Gaussian signal pairs with dependency induced by the moving average (MA) filter; (c) correlation coefficient of non-linear dependencies; (d) autocorrelation function of a single signal filtered by the MA filter. The results are presented as the mean ± SE (standard error of mean) of 20 signals or signal pairs.
Entropy 22 01146 g006
Figure 7. Multifractal spectrum of artificial test signals. (a) Spectrum of pseudorandom signals with different distributions; inset in the top right corner shows the spectrum of PI transformed signals; (b) spectrum of logistic map signal (deterministic chaos); (c) spectra of the same signal submitted to the MA filter; (d) spectra of PI-transformed signals from (c).
Figure 7. Multifractal spectrum of artificial test signals. (a) Spectrum of pseudorandom signals with different distributions; inset in the top right corner shows the spectrum of PI transformed signals; (b) spectrum of logistic map signal (deterministic chaos); (c) spectra of the same signal submitted to the MA filter; (d) spectra of PI-transformed signals from (c).
Entropy 22 01146 g007
Figure 8. Threshold profile for signals recorded from laboratory rats: (a) entropy estimates of SBP; (b) cross-entropy estimates for SBP vs. PI signals. The estimates become stable for threshold r = 0.3 (vertical line).
Figure 8. Threshold profile for signals recorded from laboratory rats: (a) entropy estimates of SBP; (b) cross-entropy estimates for SBP vs. PI signals. The estimates become stable for threshold r = 0.3 (vertical line).
Entropy 22 01146 g008
Figure 9. Self-entropy estimates: (a) Gaussian, beta and gamma distributions; (b) PI-transformed Gaussian, beta and gamma distributions; (c) MA filtered Gaussian signal; (d) Gaussian signal with induced non-linear transform, Y = a·XEXP+ b.
Figure 9. Self-entropy estimates: (a) Gaussian, beta and gamma distributions; (b) PI-transformed Gaussian, beta and gamma distributions; (c) MA filtered Gaussian signal; (d) Gaussian signal with induced non-linear transform, Y = a·XEXP+ b.
Entropy 22 01146 g009
Figure 10. Cross-entropy estimates: (a) signal with beta distribution vs. signal with gamma distribution for different correlation coefficients; (b) signal with Gaussian distribution vs. the same signal filtered with the MA filter; (c) signal with Gaussian distribution (X) vs. its non-linearly transformed counterpart Y = a XEXP + b; (d) signals from (c) in different order, Y vs. X.
Figure 10. Cross-entropy estimates: (a) signal with beta distribution vs. signal with gamma distribution for different correlation coefficients; (b) signal with Gaussian distribution vs. the same signal filtered with the MA filter; (c) signal with Gaussian distribution (X) vs. its non-linearly transformed counterpart Y = a XEXP + b; (d) signals from (c) in different order, Y vs. X.
Entropy 22 01146 g010aEntropy 22 01146 g010b
Figure 11. Entropy and cross-entropy estimates of deterministic chaos, generated by logistic maps: (a) entropy of signal X with parameter RP = 3.58, non-continual signal; (b) entropy of signal Y with parameter RP = 3.81; (c) cross-entropy X vs. Y; (d) cross-entropy Y vs. X.
Figure 11. Entropy and cross-entropy estimates of deterministic chaos, generated by logistic maps: (a) entropy of signal X with parameter RP = 3.58, non-continual signal; (b) entropy of signal Y with parameter RP = 3.81; (c) cross-entropy X vs. Y; (d) cross-entropy Y vs. X.
Entropy 22 01146 g011aEntropy 22 01146 g011b
Figure 12. Self-entropy and cross-entropy estimates based on ApEn with corrections and SampEn: (a) self-entropy, rats submitted to the restraint stress; (b) self-entropy, rats submitted to the shaker stress; (c) cross-entropy, rats submitted to the restraint stress; (d) cross-entropy, rats submitted to the shaker stress. Results are presented as mean ± SE (standard error of mean); * denotes the difference between the control and stressed signals at the significance level of p < 0.05. The horizontal line shows the theoretical value for perfect random signals with uniform distribution.
Figure 12. Self-entropy and cross-entropy estimates based on ApEn with corrections and SampEn: (a) self-entropy, rats submitted to the restraint stress; (b) self-entropy, rats submitted to the shaker stress; (c) cross-entropy, rats submitted to the restraint stress; (d) cross-entropy, rats submitted to the shaker stress. Results are presented as mean ± SE (standard error of mean); * denotes the difference between the control and stressed signals at the significance level of p < 0.05. The horizontal line shows the theoretical value for perfect random signals with uniform distribution.
Entropy 22 01146 g012aEntropy 22 01146 g012b
Figure 13. The percentage of probabilities p ^ i m ( r ) that are estimated with high reliability. (a) SBP is the reference signal, PI is the follower; (b) PI is the reference signal, SBP is the follower.
Figure 13. The percentage of probabilities p ^ i m ( r ) that are estimated with high reliability. (a) SBP is the reference signal, PI is the follower; (b) PI is the reference signal, SBP is the follower.
Entropy 22 01146 g013
Figure 14. The percentage of p ^ i m ( r ) = 0 cases that violate the entropy estimation; (a) SBP is the reference signal, PI is the follower; (b) PI is the reference signal, SBP is the follower. r. Horizontal line shows the border for which the percentage of zero probabilities is below 1%.
Figure 14. The percentage of p ^ i m ( r ) = 0 cases that violate the entropy estimation; (a) SBP is the reference signal, PI is the follower; (b) PI is the reference signal, SBP is the follower. r. Horizontal line shows the border for which the percentage of zero probabilities is below 1%.
Entropy 22 01146 g014
Table 1. Systolic blood pressure and pulse interval of rats in baseline and stressed condition.
Table 1. Systolic blood pressure and pulse interval of rats in baseline and stressed condition.
SBP [mmHg]PI [ms]
STRESSBASELINESTRESSBASELINESTRESS
SHAKER116.28 ± 6.82125.55 ± 8.89167.8 ± 13.95157.97 ± 10.98
RESTRAINT107.08 ± 7.83114.37 ± 24.92179.29 ± 13.66131.46 ± 5.71 *
The results are presented as mean ± standard deviation. * denotes the difference between the baseline and stressed signals at the significance level of p < 0.05.

Share and Cite

MDPI and ACS Style

Bajić, D.; Mišić, N.; Škorić, T.; Japundžić-Žigon, N.; Milovanović, M. On Entropy of Probability Integral Transformed Time Series. Entropy 2020, 22, 1146. https://doi.org/10.3390/e22101146

AMA Style

Bajić D, Mišić N, Škorić T, Japundžić-Žigon N, Milovanović M. On Entropy of Probability Integral Transformed Time Series. Entropy. 2020; 22(10):1146. https://doi.org/10.3390/e22101146

Chicago/Turabian Style

Bajić, Dragana, Nataša Mišić, Tamara Škorić, Nina Japundžić-Žigon, and Miloš Milovanović. 2020. "On Entropy of Probability Integral Transformed Time Series" Entropy 22, no. 10: 1146. https://doi.org/10.3390/e22101146

APA Style

Bajić, D., Mišić, N., Škorić, T., Japundžić-Žigon, N., & Milovanović, M. (2020). On Entropy of Probability Integral Transformed Time Series. Entropy, 22(10), 1146. https://doi.org/10.3390/e22101146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop