Next Article in Journal
Passive Continuous Variable Quantum Key Distribution through the Oceanic Turbulence
Previous Article in Journal
A System-Independent Derivation of Preferential Attachment from the Principle of Least Effort
Previous Article in Special Issue
The Impact of Hoffmann Reflex on Standing Postural Control Complexity in the Elderly with Impaired Plantar Sensation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Considerations for Applying Entropy Methods to Temporally Correlated Stochastic Datasets

1
Department of Kinesiology, University of Massachusetts Amherst, Amherst, MA 01003, USA
2
Institute for Applied Life Sciences, University of Massachusetts Amherst, Amherst, MA 01003, USA
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(2), 306; https://doi.org/10.3390/e25020306
Submission received: 29 November 2022 / Revised: 18 January 2023 / Accepted: 29 January 2023 / Published: 7 February 2023
(This article belongs to the Special Issue Entropy-Based Biomechanical Research and Its Applications)

Abstract

:
The goal of this paper is to highlight considerations and provide recommendations for analytical issues that arise when applying entropy methods, specifically Sample Entropy (SampEn), to temporally correlated stochastic datasets, which are representative of a broad range of biomechanical and physiological variables. To simulate a variety of processes encountered in biomechanical applications, autoregressive fractionally integrated moving averaged (ARFIMA) models were used to produce temporally correlated data spanning the fractional Gaussian noise/fractional Brownian motion model. We then applied ARFIMA modeling and SampEn to the datasets to quantify the temporal correlations and regularity of the simulated datasets. We demonstrate the use of ARFIMA modeling for estimating temporal correlation properties and classifying stochastic datasets as stationary or nonstationary. We then leverage ARFIMA modeling to improve the effectiveness of data cleaning procedures and mitigate the influence of outliers on SampEn estimates. We also emphasize the limitations of SampEn to distinguish among stochastic datasets and suggest the use of complementary measures to better characterize the dynamics of biomechanical variables. Finally, we demonstrate that parameter normalization is not an effective procedure for increasing the interoperability of SampEn estimates, at least not for entirely stochastic datasets.

1. Introduction

The biomechanics of human movement are characterized by time-varying behavior that is amenable to analysis using entropy methods, which are grounded in principles of information theory and dynamical systems theory. Recently, entropy methods have received considerable attention from researchers and clinicians interested in human movement [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17]. Despite this growing interest, it is not always clear when and how to leverage these powerful analytical tools while avoiding misapplication and misinterpretation.
The goal of this paper is to highlight considerations and provide recommendations for analytical issues associated with applying entropy methods to temporally correlated stochastic datasets, which are representative of a broad range of biomechanical variables, including spatiotemporal gait parameters and center-of-pressure trajectories. Before addressing specific problems and goals, we provide a brief overview of information theory and the concept of information, entropy as a measure of information, entropy methods for physiological datasets, and uses of entropy methods in biomechanical applications.

1.1. Information Theory and Information

Information theory is a branch of mathematics concerned with probability and statistics that is broadly applicable to probabilistic sequences of measurements [18]. In information theory, information is related to the number of responses a system produces and the associated probabilities of those responses under fixed conditions. Information is closely related to uncertainty; the more uncertain the outcomes, the more information gained from new observations, and vice versa.
The information produced by a system can change from one setting to another when the number or probability of the responses changes, such as when environmental conditions or external constraints are varied. Information refers to how the system could have behaved, not how it did, and reflects the degree of constraint on the responses [19]. Systems characterized by more information can produce more flexible responses (i.e., there is greater use of the response space or more uncertainty about future responses), whereas systems characterized by less information produce more rigid responses (i.e., there is reduced use of the response space or less uncertainty about future responses).

1.2. Entropy and Information

Information can be defined formally from a set of potential responses and their associated probabilities:
H = i p i log 2 p i ,
where p i is the probability associated with the ith response. This definition, which is known as Shannon entropy, is the foundation of modern information theory. Shannon entropy has the property that information is zero when the response is known and maximal when all responses occur with the same likelihood. Intuitively, information increases as the number of possible responses grows and the probabilities remain uniform. Another way to think about information is that it represents the surprise associated with a response—i.e., we are least surprised when the response is known in advance and most surprised when all responses are equally likely.
Shannon entropy measures the average information associated with a static probability distribution, whereas its dynamical counterparts, such as Kolmogorov–Sinai (KS) entropy [20,21], measure the rate of information generation from a deterministic dynamical system. However, KS entropy is not applicable to the finite, noisy datasets commonly measured from physiological systems [22]. To overcome this limitation, several related methods were developed to accommodate datasets obtained from systems with deterministic and stochastic components.

1.3. Entropy Methods for Physiological Data

Modeling the processes that generate physiological datasets is challenging. Entropy methods are model-free analytical tools for quantifying the degree of randomness in a dataset without making assumptions about the sources that generated it. Importantly, entropy methods measure randomness in the temporal organization of a dataset independent of its associated moment statistics [23].
Approximate Entropy (ApEn) was developed to quantify regularity in short, noisy physiological datasets [24]. For stochastic processes, K-S entropy returns infinite values, whereas ApEn returns finite values, allowing stochastic datasets to be compared [25]. ApEn quantifies the likelihood that sequences of a fixed template length (m) that are considered close within a tolerance (r) remain close when the template length is increased (m + 1) for a dataset of N observations. ApEn was subsequently used in a wide range of clinical applications, such as distinguishing age- and disease-related changes to cardiovascular dynamics [26,27].
ApEn has a self-matching bias and lacks relative consistency such that the pattern of results is not always maintained when hyperparameters are changed [25,28,29,30]. Consequently, Sample Entropy (SampEn) was developed to eliminate self-matching and the template-wise approach for computing probabilities [28]. SampEn was shown to be more accurate and reliable than ApEn, and serves as the basis for other entropy techniques, such as Control Entropy [31,32] and multiscale entropy (MSE) [33,34]. A modified definition of SampEn accounts for temporal correlations among nearby template vectors, which can occur due to oversampling, by including a time delay, τ [35].
Because this paper examines temporally correlated stochastic datasets, SampEn is a more suitable method than ApEn. Stochastic processes exhibit weak dependence compared to deterministic processes, creating diffuse dynamics such that nearly all partitions contain near-zero probabilities [29]. The self-matching bias of ApEn is particularly problematic in this case because the conditional probability log(A/B) can be much different than log((A + 1)/(B + 1)). As mentioned before, SampEn also serves as the basis for other entropy algorithms, such as MSE and Control Entropy. Thus, we restrict our focus to SampEn.

1.4. Entropy Methods in Biomechanical Applications

Entropy methods, such as ApEn and SampEn, have been widely used in clinical applications to examine changes to the regularity of physiological data, including cardiovascular [26,27], respiratory [36,37,38], and neural dynamics [39,40,41]. More recently, entropy methods have gained traction in the biomechanics community and provided important insights into human postural control, locomotion, and muscle activity.
Entropy measures have been applied to standing posture to examine how changes to the center of pressure dynamics support precision manual tasks [1], understand how interactions between aging and attentional states impact balance [2], as well as characterize short-term changes from training [3] and long-term changes from neurological disease progression [4]. Entropy has also been applied to gait dynamics to evaluate fall risk [5,6] and identify age-related changes to dual-task function [7,8]. Further insights into age- and disease-related changes to postural control and gait have been gained from MSE approaches [9,10,11,12,13].
SampEn has also been applied to electromyographic records to distinguish voluntary muscle activity from background noise [14], identify the onset of muscle fatigue [15,16], and improve pattern recognition to guide prosthetic limb control [17]. As the use of entropy methods in biomechanical applications grows, it is important to highlight common analytical challenges and identify solutions to facilitate accurate and reliable estimation, improve interpretability, and increase interoperability between datasets.

1.5. Problem Statement

In this paper, we examine problems related to the application of entropy methods to temporally correlated stochastic data, which adequately describe many biomechanical variables, such as spatiotemporal gait parameters. Stochastic datasets are often derived from event-based series—e.g., stride time is commonly measured as the interval between ipsilateral heel strikes. With this type of data, the goal is to understand cycle-to-cycle changes rather than within-cycle changes that can be derived from continuous measurements, such as joint angles [42], which have stronger deterministic characteristics. Below, we consider three problems that arise when working with stochastic datasets.
First, past research has examined the effect of transient spikes or outliers on entropy measures [43,44,45]. More deterministic signals are robust to outliers, and entropy estimates increase only slightly as the number and magnitude of outliers increases [44]. By contrast, stochastic signals are more sensitive to outliers because the standard deviation becomes inflated when the scale of the signal is small relative to the outliers. This makes the radius of similarity, r, large relative to the ‘true’ signal dynamics, which reduces entropy estimates because there are more template matches. Thus, lower entropy estimates can result both from increased regularity or outliers [45]. Despite the importance of these observations, recommendations for handling outliers when working with stochastic datasets are lacking. Should r be adjusted to reduce the bias? Should outliers be removed? More importantly, how do we strike a balance between minimizing bias and preserving the underlying dynamics?
Second, rather than representing uncorrelated noise, stochastic datasets often contain temporal correlations characterized by increases in the statistical likelihood that subsequent variations occur in the same (persistence) or opposite (anti-persistence) direction, which should intuitively be characterized by greater regularity (i.e., lower entropy). The formalization adopted here derives from two related families of correlated stochastic processes: fractional Gaussian noises (fGn) and fractional Brownian motions (fBm) [46]. The fGn are stationary, while the fBm are nonstationary. Both fGn and fBm are described by the Hurst exponent (H), which characterizes temporal correlation properties ranging from anti-persistent to persistent on the interval ] 0 , 1 [ . The cumulative sum of an fGn produces a corresponding fBm with the same H, while the difference of an fBm produces a corresponding fGn. One problem is that distinct temporal correlation properties may be characterized by similar entropy estimates because persistence and anti-persistence create dynamical regularities, making it harder to distinguish potentially meaningful differences between experimental conditions or populations.
Third, another challenge with entropy methods is the limited interoperability of results obtained from independent studies, as others have indicated [47]. Interoperability refers to the ability to access data from multiple sources without loss of meaning and integrate data for analytical purposes, such as making comparisons or generating predictions. Here, we focus on one aspect of interoperability: parameter normalization. Entropy estimates have a lower bound of zero and an upper bound that is determined by multiple hyperparameters. While SampEn is formally independent of dataset length, estimates improve with greater numbers of observations because the conditional probabilities are better approximated [28]. Differences in dataset length may therefore produce differences in SampEn obtained from independent studies, which could be remedied by normalizing parameter estimates.

1.6. Goal Statement and Contributions

The goal of this paper is to provide recommendations to address analytical issues associated with the application of entropy methods, specifically SampEn, to stochastic datasets that are commonly encountered in biomechanical applications. Throughout the paper, we emphasize the importance of leveraging multiple techniques to better describe the dynamics of stochastic datasets. Where applicable, we highlight potential workarounds—some of which have been examined in previous work, others for which challenges remain. Overall, this paper provides a basic framework for applying SampEn to temporally correlated stochastic datasets, which represent a wide range of biomechanical processes, with considerations of issues ranging from process characterization to data cleaning to interpretation of results.
Specifically, we examine the use of autoregressive fractionally integrated moving average (ARFIMA) modeling to estimate the temporal correlation properties of stochastic datasets and classify datasets as stationary or nonstationary. The rationale for using simulated datasets rather than experimental datasets is that the temporal correlation properties are known in advance, making it possible to systematically assess the accuracy and reliability of SampEn across a range of input signals. ARFIMA modeling is then used to aid in data cleaning procedures and reduce biases in SampEn estimates due to the presence of outliers. We then demonstrate the challenge of discriminating the temporal correlation properties of stochastic datasets when using SampEn alone, leading to the suggestion of a multiple method approach. Finally, we demonstrate that parameter normalization is not an effective procedure for increasing interoperability, at least not for fully stochastic datasets.

2. Materials and Methods

2.1. Simulations

To produce simulated data with a range of temporal correlation properties, we used ARFIMA modeling following the methods described in [48], which demonstrated that simulations produced via ARFIMA modeling were more accurate than those generated by the Davies–Harte algorithm [49] and spectral synthesis method [50]. ARFIMA models are an extension of autoregressive integrated moving average (ARIMA) models, which are useful for understanding short-term correlations in stochastic data [51]. ARIMA models have three parameters: the autoregressive order (p), the moving average order (q), and the differencing order (d), which indicates how many times the data need to be differenced to be considered stationary.
ARIMA (p, d, q) models can be used to simulate data with specified parameters. By allowing d to take on fractional values, ARIMA models can produce long-term dependence, leading to the more general ARFIMA models [52]. ARFIMA models are only applicable to stationary data with d ] 0.5 ,   0.5 [ . Conveniently, d is linearly related to the scaling exponent (α) in the fGn/fBm model [48]:
α = 2 d + 1 2
where α   ] 0 , 1 [ for fGn and α   ] 1 , 2 [ for fBm. For fGn α = H and for fBm α = H + 1. This relation was exploited to create stationary and nonstationary stochastic datasets.
ARFIMA simulations have been shown to generate accurate and reliable datasets, independent of the method used for parameter estimation [48]. To create the simulated data, we used the freely available ARFIMA simulations library for MATLAB [53]. Only ARFIMA (0, d, 0) models, which omit the p and q parameters responsible for short-term correlations, were used. These models produce stationary fGn processes with d ] 0.5 ,   0.5 [ , which can be cumulatively summed to obtain nonstationary fBm processes.
We created simulations with expected α values of 0.1 to 1.9 in increments of 0.1 (Table 1). One independent set of simulations was obtained for fGn simulations with 0.1 ≤ α ≤ 0.9. Another independent set of simulations was cumulatively summed to obtain fBm simulations with 1.1 ≤ α ≤ 1.9. Additionally, to obtain α 1, we created simulations with α = 0.99. There were 100 simulations for each expected α value. We also examined three dataset lengths (N = 250, 500, and 1000). Each dataset length was produced independently—i.e., shorter datasets were not cropped from longer ones. The mean and standard deviation of each dataset was 0 and 1, respectively.

2.2. ARFIMA Modeling

To estimate d, ARFIMA (0, d, 0) models were fit to the simulated data using the Whittle approximation in the freely available ARFIMA(p,d,q) estimator library for MATLAB [54], which also requires the MFE toolbox [55]. Because ARFIMA models are restricted to stationary processes and produce bounded estimates of d, the Whittle approximation may not converge. Following recommendations by [56] and procedures adopted by [48], nonstationary processes were differenced, and the model fit to the stationary increments. To classify data as nonstationary, the model was fit before differencing. When the algorithm failed to converge, the data were considered nonstationary, and the model was reapplied to the differenced data. If the data were differenced, the d estimate was adjusted by adding 1 and then converted to α (Equation (2)).
Detrended fluctuation analysis (DFA) has been suggested for classifying stochastic datasets as stationary or nonstationary (e.g., [57]). Temporal correlations were separately estimated using evenly spaced average DFA [58,59] with the following hyperparameters: nmin = 10; nmax = N/4; k = 26, 37, and 47 for the three dataset lengths, respectively; and linear detrending. The decision to use ARFIMA models to measure temporal correlations was supported by observations of comparatively lower biases and standard deviations, which is consistent with previous findings [48]. ARFIMA is both a better estimator (Figure S1) and classifier (Figure S2) of the stochastic processes considered in this paper. A further benefit of the ARFIMA approach is that it does not require hyperparameter selection, which reduces the need to make decisions about selecting appropriate values.

2.3. SampEn(m, r, τ )

SampEn(m, r) has two hyperparameters: the template length, m, and the radius of similarity, r [28]. Both m and r need to be specified. SampEn is estimated as the negative logarithm of the conditional probability (A/B) that sequences of m observations that are close within tolerance r (B) remain close when the sequence is increased to m + 1 observations (A). SampEn eliminates the self-matching bias of ApEn by comparing each template vector to all other template vectors except itself. Furthermore, SampEn computes the conditional probability A/B for the entire sequence of observations instead of template-wise, which avoids the requirement that each template vector has at least one match. Note: SampEn(m, r) is a parameter, while SampEn(m, r, N) is a statistic estimated from a dataset.
To select m, different methods have been proposed, such as fitting autoregressive models [28,43,45]. However, stochastic datasets, such as those considered in this paper, are characterized by a wide range of temporal correlation properties, which would lead to the selection of different m for different ranges of α. We selected m = 1 based on the observation that the precision of SampEn estimates decreased for larger m (Figure S3).
To select r, different methods have also been proposed, such as minimizing the relative error in the conditional probability A/B or SampEn(m, r, N) [43,45]. We selected r = 0.25 based on the observation that the precision of SampEn was worse for r < 0.25 and only marginally improved for r > 0.25 (Figure S4). This decision was made to balance robust estimation against reducing the ability to discriminate distinct temporal sequences when large r values are selected.
A more general definition of SampEn includes a third hyperparameter: the time delay, τ [35]. The time delay, τ , represents the lag between successive elements in the template vectors. Compared to SampEn(m, r), where τ = 1 by default, SampEn(m, r, τ ) produces more accurate estimates when data are oversampled or contain temporal correlations [35,60]. We used τ = 1 for all analyses in this paper. τ can be estimated using the autocorrelation function or average mutual information [61]. For α 0.5, the autocorrelation function decays exponentially, yielding τ = 1. For 0.5   α 1, the autocorrelation function decays more slowly and τ will be greater than 1; τ is even larger for nonstationary processes (α > 1). SampEn estimates obtained from stochastic processes containing long-term correlations (α > 5) were relatively unaffected by different τ values (Figure S5). Thus, τ appears to be less influential when using SampEn to interrogate stochastic datasets. However, τ will likely play a more important role when working experimental datasets that represent a mixture of deterministic and stochastic components.

2.4. Outlier Generation

To examine the impact of extreme observations on SampEn estimates, we obtained baseline measures of SampEn(m = 1, r = 0.25, τ = 1) for each of the simulated datasets. We then adjusted a small proportion of the observations (1%) by adding values drawn from a standard normal Gaussian distribution multiplied by three times the peak-to-peak signal amplitude, the same as previous work [44]. SampEn(m = 1, r = 0.25, τ = 1) was then applied to the contaminated datasets to quantify biases in parameter estimates.

2.5. Parameter Normalization

To create interoperable parameter estimates for SampEn, we examined a normalization procedure based on the maximum entropy attainable for a dataset. SampEn parameter estimates have a lower bound of 0 and an upper bound of ln(Nm1) + ln(Nm) – ln(2) [28]. The lower bound is reached when all template vectors match—i.e., the conditional probability A/B = 1, which is unlikely in experimental data. The upper bound is defined for SampEn(m, r) and explicitly depends on the dataset length, N, and the template length, m. This value is approached when only small numbers of template matches are found. However, a more general expression can be derived for SampEn(m, r, τ ), which includes the time delay, τ :
ln ( N m τ 1 ) + ln ( N m τ ) ln ( 2 ) .
The two expressions are equivalent when τ = 1, which is used throughout this paper. To normalize the entropy estimates, we divided each value by its upper bound—i.e., the maximum entropy value. This may facilitate comparisons of estimates obtained for datasets with different lengths. For example, the upper bounds examined in this study were 10.32, 11.72, and 13.12 for the short (N = 250), medium (N = 500), and long (N = 1000) datasets, respectively. By normalizing the entropy estimates from 0 to 1, it becomes possible to compare datasets of different lengths.
To examine the effect of normalization, we generated simulations with two additional dataset lengths that exceeded those primarily considered in this paper. The longer dataset lengths were N = 5000 and 10,000. Both dataset lengths are quite large compared to the datasets commonly collected in laboratory-based experiments or clinical settings. However, with the growing availability and portability of mobile health monitoring technologies, increasingly large datasets are becoming feasible to obtain.

3. Results

3.1. Estimation of Temporal Correlations Using ARFIMA Modeling

To verify that the simulated data were characterized by the expected temporal correlation properties, ARFIMA (0, d, 0) models were fit and the d estimates were converted to the scaling exponent, α. We then quantified the bias and standard deviation of the estimates, which are presented in Table 2. The short datasets consistently underestimated entropy (Figure 1A), whereas the longer datasets were unbiased except for values close to α = 2 (Figure 1B,C). The standard deviation was the greatest for the short datasets and was elevated for α > 1.5. This trend was reduced in the longer datasets. In summary, the ARFIMA simulated datasets contained reliably different temporal correlation properties ranging from anti-persistent stationary data (e.g., α < 0.5) to persistent nonstationary data (e.g., α > 1.5), as characterized by ARFIMA (0, d, 0) models.

3.2. Classification of Stochastic Data Using ARFIMA Modeling

To examine the use of ARFIMA modeling for data classification, we computed the proportion of the 100 simulations classified as nonstationary for each of the expected α values. An ideal classifier would indicate that data with 0 < α < 1 were stationary and data with 1 < α < 2 were nonstationary. In practice, approximately 50% of the simulations close to the 1/f boundary (i.e., α = 1) were classified as nonstationary (Figure 2A). This was consistent across dataset lengths. Shorter datasets were more prone to misclassification near the boundary, but the error rates were less than 5% for all dataset lengths.
Classification errors decreased the farther the expected α value was from the 1/f boundary. However, classification errors were observed for 0.8 < α < 1.2. To better characterize the classification accuracy of the ARFIMA model, we conducted a follow-up analysis by generating simulated datasets with temporal correlations varying from 0.7 < α < 1.3 in increments 0.01. The same three dataset lengths were examined. We generated 500 simulations for each combination of α and N, leading to a total of 90,000 independent datasets. ARFIMA (0, d, 0) models were again used to obtain d estimates, which were subsequently converted to α and used to assess classification errors.
The follow-up analysis revealed differences in classification accuracy across the dataset lengths (Figure 2B). Short datasets (N = 250) were classified least accurately on both sides of the 1/f boundary with error rates exceeding 5% from 0.90 < α < 1.09. This was not unexpected because the standard deviation of the α estimates was elevated compared to the longer datasets. Medium datasets (N = 500) were classified less accurately than long datasets (N = 1000) below the boundary, with errors exceeding 5% at α = 0.92 and α = 0.95, respectively. By contrast, classification accuracy was comparable between the medium and long datasets above the boundary, with error rates exceeding 5% at α = 1.05. This asymmetry is likely due to the increased standard deviation of the α estimates below compared to above the boundary for the medium datasets.
In summary, ARFIMA modeling is also an accurate and reliable method for classifying whether stochastic datasets are stationary or nonstationary. Classification accuracy is dependent on dataset length near the 1/f boundary (α = 1.0) and improves with more data. In the following section, we demonstrate the utility of this approach for improving entropy estimates following data cleaning procedures.

3.3. Outlier Removal with Data Classification Reduces Biases in Entropy Estimates

The main problem with outliers is that they inflate the standard deviation. Because data are commonly normalized before applying SampEn (e.g., [43]), the amplitude of the ‘true’ data will be reduced, which will increase the likelihood of detecting template matches. The extent to which the standard deviation changes depends on the number and magnitude of the outliers. Datasets were manipulated by adjusting a small proportion of observations (1%) with peak magnitudes up to 3× the range of the dataset.
Entropy estimates were substantially lower in the presence of outliers (Figure 3A–C). Biases were relatively consistent across dataset length and greater than 25% in most cases. However, lower biases were observed for nonstationary datasets with persistent increments (i.e., α > 1.5). The standard deviation of the bias decreased with increasing dataset length, possibly because more observations were manipulated in the longer datasets. Thus, a small number of outliers had a disproportionate impact on entropy estimates, consistent with previous findings [44].
One potential solution is to adjust r because the amplitude of the ‘true’ data is reduced when the data are normalized with outliers. An adjustment can be applied to r by taking the ratio of the standard deviation without outliers to the standard deviation with outliers. For example, if r = 0.2 and the ratio of the standard deviations is 0.5, then the adjusted r is 0.5 r or 0.1. This creates stricter requirements for identifying template matches and improves entropy estimates. To examine the efficacy of this procedure, outliers were identified using a highly conservative threshold of five times the median absolute deviation [62]. The adjusted r values were computed and SampEn was reapplied to the outlier-contaminated datasets.
Adjusting r substantially reduced the bias for all datasets lengths (Figure 3D–F). The bias was less than 5% and the standard deviation less than 3% for α < 1.5. Positive biases were observed for α 1.5. The biases increased as α and N increased. Similar trends were observed for the standard deviation of the bias. These findings indicate that adjusting r reduces the influence of outliers on entropy estimates, particularly for stationary datasets; however, this procedure does not consistently work for nonstationary datasets.
Another potential solution is to remove outliers without adjusting r. In practice, this is a simple procedure but one that requires caution. Here, we assumed that observations flagged as outliers were not representative and should be removed. The maximum number of possible outliers was 3, 5, and 10 for the short, medium, and long datasets, respectively. The conservative threshold of five median absolute deviations detected less than or equal to the maximum possible number of outliers for nearly all datasets. False positives were detected in only 2 of the 5,700 total datasets, demonstrating the conservative nature of this approach.
Removing outliers further reduced the bias for all dataset lengths (Figure 3G–I). The biases were less than 1% and standard deviation less than 2% for α < 1.5. Short datasets showed the greatest variance within this range. Again, positive biases were observed for α 1.5—although the magnitudes were smaller than when adjusting r. This indicates that outlier removal is an effective procedure for minimizing the impact of outliers on entropy estimates. However, similar to adjusting r, outlier removal is not effective for nonstationary datasets with persistent increments (α > 1.5).
To address this limitation, ARFIMA models were applied to the data to classify datasets as stationary or nonstationary. One challenge with identifying outliers from nonstationary datasets is that they do not stand out. However, if the data are differenced, outliers can be easily identified and removed. This procedure was applied to the data and compared to outlier removal without classification. Data classification using ARFIMA modeling reduced biases across the entire α range independent of dataset length (Figure 4). The maximum biases were 1.5%, 3.5%, and 3.1% for the short, medium, and long datasets. However, this method produced some tradeoffs. While the standard deviation of the bias was reduced for the nonstationary datasets with persistent increments (α > 1.5), it increased slightly for nonstationary datasets with anti-persistent increments (α < 1.5).
To summarize, when stochastic datasets contain potential outliers, we recommend using a scaled median absolute deviation to identify and remove those observations. The scale factor should be tuned so that the number of outliers is acceptably small. This procedure was effective for stationary datasets (α < 1), but less so for nonstationary datasets (α > 1). Consequently, ARFIMA modeling was used to classify data as stationary or nonstationary before applying the outlier removal procedure, which substantially reduced biases in entropy estimates with minor tradeoffs in the standard deviation of the bias.

3.4. Problems with Discriminating Temporally Correlated Stochastic Datasets with SampEn

Because stochastic datasets commonly contain temporal correlations, entropy should decrease for both anti-persistent and persistent processes, as well as nonstationary ones, leading to potentially equivalent outcomes despite being characterized by distinctly different temporal correlation properties. We applied SampEn(m = 1, r = 0.25, τ = 1) to the simulated datasets with the following expectations: (1) entropy should be greatest for uncorrelated time series (α = 0.5) and (2) entropy will decrease as temporal correlations increase (α ≠ 0.5). This should produce an inverted parabolic relationship between entropy measured by SampEn and α. Greater numbers of data points were not expected to impact these predictions.
The estimated and expected α values obtained from ARFIMA modeling showed a linear relationship with a slope close to 1 (Figure 5A–C). Biases and standard deviations decreased with increasing dataset length. Thus, ARFIMA modeling accurately characterizes the temporal correlation properties of stationary and nonstationary datasets independent of dataset length, which is consistent with past findings [48].
By contrast, SampEn estimates were characterized by an inverted parabolic relationship across the expected α values independent of dataset length (Figure 5D–F). Stationary datasets containing temporal correlations—i.e., anti-persistent (α < 0.5) and persistent (α > 0.5) processes—were characterized by lower entropy compared to uncorrelated noise (α = 0.5), as expected. The only exception was the short datasets, where α = 0.4 was slightly larger than α = 0.5. Strongly persistent processes, which contain long-term correlations, were characterized by greater regularity than strongly anti-persistent processes, leading to a right-skewed asymmetric relationship between entropy and α. Means and standard deviations are reported in Table 3.
Nonstationary datasets were characterized by decreasing regularity for 1 < α < 2. This makes intuitive sense because these processes become more regular as the increments shift from anti-persistent to uncorrelated to persistent. The precision of SampEn was particularly poor for the nonstationary anti-persistent processes (1 < α < 1.5). Greater precision was observed for the nonstationary persistent processes (1.5 < α < 2), likely because SampEn has a lower bound of zero. Again, means and standard deviations are reported in Table 3.
The nonlinear relationship between SampEn and α suggests the need to use complementary measures when examining stochastic datasets. Specifically, we suggest the use of ARFIMA modeling or other methods capable of estimating temporal correlations. Otherwise, SampEn will have difficulty distinguishing between stationary anti-persistent and persistent processes, as well as identifying subtle differences between nonstationary persistent processes, both of which may indicate relevant differences between experimental conditions or populations.

3.5. Normalization of SampEn Estimates across Different Dataset Lengths

SampEn is bounded from below by ln(1) = 0, which occurs when the conditional probability A/B = 1, and from above by ln(Nm τ 1) + ln(Nm τ ) – ln(2) Equation (3). The upper bound is determined by multiple hyperparameters but is principally influenced by the dataset length, N. This value represents the maximum value possible for specific hyperparameter combinations. Although the radius of similarity, r, is not explicitly included, it defines how close sequences of observations must be to be considered a match. Thus, SampEn estimates will grow with decreasing r.
In practice, the estimate will not approach the maximum value because r is relatively large—commonly close to 0.2 standard deviations—such that the number of matches will also be relatively large. To improve the interoperability of SampEn estimates, the obtained estimate can be normalized by its maximum value, which transforms it to a 0 to 1 scale, where 0 corresponds to complete regularity and 1 corresponds to complete randomness. We now consider the implications and limitations of applying this normalization to temporally correlated stochastic datasets.
SampEn has been shown to be robust to variations in dataset length [28]. This was observed for the stationary datasets (α < 1; Figure 6A). The precision of SampEn estimates increased for the stationary datasets as N increased. By contrast, the distributions of SampEn estimates drifted apart for the nonstationary processes suggesting that greater regularity was detected in processes characterized by the same temporal correlation properties; however, in general, the distribution shapes were similar for increasing dataset lengths.
By contrast, the normalized SampEn estimates produced a distinct pattern of results. The qualitative shape of the SampEn estimates as a function of α were comparable to the unnormalized values. However, normalization produced clear differences in SampEn as a function of dataset length for the stationary datasets, whereas longer datasets were characterized by lower normalized estimates (Figure 6B). This is not surprising, given that longer datasets have larger maximum values; however, it suggests a potential problem because longer datasets with the same temporal correlation properties are characterized by greater regularity. Thus, normalization may not be appropriate for stochastic datasets.

4. Discussion

This paper examined considerations for applying entropy methods, specifically SampEn, to temporally correlated stochastic data, which are representative of variables that are often calculated in biomechanical research, such as spatiotemporal gait parameters and center of pressure trajectories. The main problems considered were (1) biases in SampEn in the presence of outliers, (2) the inability of SampEn to distinguish distinct temporal correlation properties, and (3) the limited interoperability of SampEn estimates obtained for datasets of substantially different lengths.

4.1. ARFIMA Modeling for Estimation and Classification of Temporally Correlated Stochastic Datasets

Solutions to some of the problems mentioned above relied on the ability to classify stochastic data as stationary or nonstationary. In the fGn/fBm model, stationary and nonstationary processes are related by the Hurst exponent [46]. Stochastic processes can also be quantified on a continuous scale by measuring power law relations of changes in the variance over different time windows (i.e., S D ( t ) t α ) or changes to the spectral power over different frequencies (i.e., P ( f ) f β ). Here, we opted for the scaling exponent, α, which ranges from ] 0 ,   1 [ for stationary processes and ] 1 ,   2 [ for nonstationary processes. By generating simulated datasets, we were able to provide insights across the range of temporal correlations that are commonly observed in biomechanical variables.
We estimated the temporal correlation properties and classified the simulated datasets using ARFIMA (0, d, 0) models. There are several notable benefits to this approach. First, the difference order, d, can be directly converted to α, which allows consistent interpretation of the measures obtained from ARFIMA and DFA, which has historically been used in the human movement literature (e.g., [63,64]). Second, ARFIMA models do not require hyperparameter selection—unlike other methods, such as DFA—where multiple decisions must be made to obtain parameter estimates, which reduces the potential for discrepancies across studies. Third, and most importantly, ARFIMA models produce more accurate and precise estimates of temporal correlations compared to more common methods, such as DFA, which contributes to better classification of datasets as stationary or nonstationary.
Accurate and reliable classification of stochastic datasets has been considered previously [57,65]. Here, we demonstrated that data classification is also important for obtaining accurate and reliable SampEn estimates. Data cleaning procedures were more effective, and the interpretability of SampEn estimates were improved when data classification and estimation was conducted using ARFIMA modeling. Moreover, SampEn has trouble distinguishing stationary processes with distinct temporal correlation properties, which suggests the need to leverage multiple analytical techniques to characterize and compare the dynamics of biomechanical datasets, which is discussed in more detail below.

4.2. Outlier Removal Reduces Biases in SampEn Estimates

Past work has shown that outliers artificially reduce entropy estimates by inflating the estimated standard deviation relative to the scale of the ‘true’ dynamics [43,44,45], thus artificially inflating the r hyperparameter. Here, SampEn estimates were biased by approximately 25% when only 1% of the observations were manipulated. The bias persisted when smaller amplitude outliers were generated in separate analyses not reported here. However, the bias magnitude decreased to approximately 20% and 7% for outliers with 2× and 1× the peak-to-peak amplitude, respectively. Because lower SampEn estimates can result both from increased regularity or increased r due to the presence of outliers [45], we recommend that researchers cautiously examine datasets for outliers before applying entropy techniques. Otherwise, the resulting interpretations may be inconsistent and/or incorrect.
We examined several methods for reducing biases, including adjusting the radius of similarity, r, and outlier removal that was agnostic of the classification of the dataset as stationary or nonstationary. Both methods were effective at reducing the bias, particularly for stationary datasets. However, positive biases remained for nonstationary datasets with persistent increments. This occurred because outlier detection was less effective for these types of superdiffusive processes.
To address this limitation, we used ARFIMA modeling to classify datasets as nonstationary. Nonstationary datasets were differenced to improve outlier detection. We did not remove the outliers from the differenced datasets and then re-integrate them. Because each difference results from multiple datapoints, removing them can create sizeable discontinuities in the re-integrated dataset, which can also bias SampEn estimates. Instead, outliers were directly removed from the nonstationary datasets. This procedure substantially reduced biases in SampEn across the entire range of the fGn/fBm model (0 < α < 2).
One concern is that outlier removal is itself a methodological procedure that requires parameter tuning and defining what counts as an outlier. We recommend being conservative when data cleaning by restricting removal only to the most extreme observations. If a substantial proportion of observations are flagged as outliers, the detection threshold may be too strict. Unfortunately, we cannot provide a simple recommendation for determining what percentage of the data would represent an unjustifiable number of outliers. Outliers should be rare unless there is suspicion that the data were generated by a distinct mechanism (e.g., the experimenter noticed that a participant was distracted).
Another concern is that outlier removal can create datasets with different numbers of observations. Outlier replacement could reduce data loss and further improve the precision obtained for short datasets. We experimented with several methods, including mean and median replacement, but did not find additional improvements in entropy estimates. More sophisticated procedures have been developed (e.g., [66]) but were not examined. If outlier removal creates substantial variation in dataset length, it may be helpful to crop all datasets to the same length. Fortunately, the accuracy and precision of SampEn have been shown to be relatively robust to variations in sample size [28].
However, we observed that SampEn estimates were robust to differences in dataset length for stationary stochastic processes but not nonstationary stochastic processes (see Table 3 and Figure 6A). We recommend that researchers ensure that their results are not dependent on dataset length, particularly when working with nonstationary datasets. This can be achieved in several ways. First, if sufficient data are collected, then datasets can be cropped. The pattern of results should be consistent across changes in dataset length, even if the estimates begin to depart. Second, additional data can be collected during pilot work and analyses performed to determine the minimum data needed to obtain reliable estimates under specific experimental conditions or from different populations.
Moving forward, the effectiveness of outlier removal procedures needs to be assessed on ‘real-world’ datasets. The simulated datasets examined in this paper are representative of biomechanical datasets obtained from human participants [67,68]. However, in practice, measurements obtained in laboratory or clinical settings will not be as clean and well-behaved, and outliers may not be as easy to detect. While the scaled median absolute deviation threshold was extremely effective for outlier identification and removal, researchers may need to tailor this approach to meet their specific needs.

4.3. Interpreting SampEn Estimates from Temporally Correlated Stochastic Data

SampEn is a regularity statistic that quantifies the frequency of repeating patterns in a dataset; but it does not indicate the nature of the observed patterns. This is not a problem when researchers are interested in understanding whether data are more regular under one set of conditions compared to another. However, SampEn may be unable to distinguish between distinct temporal correlation properties when used alone (see Table 3).
Our results highlight a significant limitation of SampEn, namely, the many-to-one correspondence between regularity and temporal correlations for stationary stochastic datasets. Entropy methods, such as ApEn and SampEn, were developed to avoid phase space reconstruction, which requires large amounts of noise-free data to estimate KS entropy [22]. Whereas KS entropy is defined in the limits of m  , r  0, and N  , ApEn was developed to conduct a partial characterization using smaller m and larger r that requires less data and is less sensitive to noise [29]. Revisiting a point made by Pincus [29], failures to distinguish differences in regularity do not indicate that the dynamics on the reconstructed attractor are the same. We reiterate a more general point that similar values of SampEn do not indicate that the dynamics being compared are the same.
We were unable to identify straightforward solutions or transformations to remedy this issue. For example, differencing stationary datasets before submitting them to SampEn produced a monotonic increasing function as α increased from 0.1 to 0.99—see also Figure 2 in [35]. This procedure could be useful if all datasets are stationary, but often experimental studies contain datasets on both sides of the 1/f boundary (e.g., [69]). Indiscriminately applying the differencing procedure would re-introduce the many-to-one mapping problem for nonstationary datasets.
There is the additional problem that the dataset length biases the entropy estimates for nonstationary processes such that longer datasets are characterized by greater regularity despite having the same temporal correlation properties as shorter ones (Figure 6). This occurred even though SampEn does not explicitly depend on N. The observation of greater regularity for longer datasets may be attributable to the fact that there are more observations within the radius of similarity, r, such that the conditional probability A/B becomes larger, which would reduce SampEn. The problem is determining which estimates are correct. One solution is to assume that the longer time series produce better estimates. On the other hand, it may be inappropriate to apply SampEn to these types of nonstationary datasets because of the poor precision across estimates. This suggests the need to convert nonstationary processes to stationary processes before applying SampEn. The main limitation of this approach is that it does not solve the many-to-one mapping problem described above. An alternative is to use methods specifically developed to handle nonstationarity, such as Control Entropy (e.g., [31]).
The inability to distinguish stochastic datasets with distinct temporal correlation properties using SampEn creates a problem for researchers and clinicians interested in examining differences in the dynamics of biomechanical variables. There is a greater likelihood of reporting false negatives when experimental conditions or groups are characterized by similar degrees of regularity, even if the underlying dynamics are different. To avoid this, we recommend combining the model-free approach of entropy methods and the model-based approach of ARFIMA modeling to quantify regularity and characterize the temporal correlations producing the observed regularities.

4.4. Normalization of SampEn Estimates

One goal of this paper was to examine the use of a normalization method to increase the interoperability of SampEn estimates obtained from datasets with different lengths. Experimental designs and conditions vary widely depending on the specific research questions, leading to different quantities of data and hyperparameter selections. This makes it hard to compare entropy estimates because the range of possible values is a function of m, τ , and N (Equation (3)).
Because SampEn is bounded, the maximum entropy value can be used to normalize values between 0 and 1. This could, in principle, account for differences in dataset length. Other work has found that SampEn is robust across different dataset lengths [28]. Our results confirm this observation but only for the stationary stochastic processes (α < 1). For the nonstationary processes (α > 1), greater regularity was observed for longer datasets. Coupled with the decreased precision observed for the nonstationary datasets, caution should be taken when applying SampEn to nonstationary data, particularly when datasets have different numbers of observations.
One rationale for applying normalization is to increase interoperability. There are multiple ways to increase interoperability, including data transparency, method sensitivity, and parameter normalization. While parameter normalization was not appropriate for stochastic datasets, other opportunities remain. First, studies applying entropy methods should report hyperparameter values and how those values were selected. Second, unless there are restrictions from funding sources or other substantial barriers, researchers can make datasets and analysis routines publicly available, which was performed for this paper (see Data Availability Statement). Third, normalization may be plausible when working with experimental datasets characterized by a mixture of deterministic and stochastic components where hyperparameter selection can be guided by techniques adopted from phase space reconstruction, such as average mutual information [61] and false nearest neighbors [70]. This should be explored in future work.

5. Conclusions

In summary, this paper has highlighted considerations when applying entropy, specifically SampEn, to stochastic datasets with a range of temporal correlations, which are representative of biomechanical variables. We demonstrated the use of ARFIMA modeling for estimating temporal correlation properties and classifying stochastic datasets as stationary or nonstationary. We then leveraged ARFIMA modeling to improve the effectiveness of data cleaning procedures to mitigate the influence of outliers on SampEn estimates. We also emphasized the limitations of SampEn in distinguishing among stochastic datasets, leading to the suggested use of complementary measures to obtain a more comprehensive understanding of behavioral dynamics. Finally, we demonstrated that the parameter normalization method is not an effective procedure for increasing the interoperability of SampEn estimates for the entirely stochastic datasets examined here.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/e25020306/s1, Figure S1: ARFIMA modeling provides more accurate and reliable estimates than DFA; Figure S2: ARFIMA modeling provides better data classification than DFA; Figure S3: Estimating the template length, m; Figure S4: Estimating the radius of similarity, r; Figure S5: Estimating the time delay, τ .

Author Contributions

Conceptualization, J.L. and M.B.; methodology, J.L. and M.B.; software, J.L.; validation, J.L.; formal analysis, J.L.; investigation, J.L.; resources, M.B.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and M.B.; visualization, J.L.; supervision, M.B.; project administration, M.B.; funding acquisition, M.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC were funded by Army Research Laboratory, Cooperative Agreement # W911NF2120208.

Data Availability Statement

The data, analysis routines, and results presented in this study are openly available at https://doi.org/10.5281/zenodo.7595303.

Acknowledgments

We would like to thank Nader Naghavi for the early discussions that led us to pursue this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Michalska, J.; Zając, R.; Szydło, K.; Gerasimuk, D.; Słomka, K.J.; Juras, G. Biathletes present repeating patterns of postural control to maintain their balance while shooting. PLoS ONE 2022, 17, e0267105. [Google Scholar] [CrossRef]
  2. Rhea, C.K.; Diekfuss, J.A.; Fairbrother, J.T.; Raisbeck, L.D. Postural Control Entropy Is Increased When Adopting an External Focus of Attention. Mot. Control 2019, 23, 230–242. [Google Scholar] [CrossRef]
  3. Diekfuss, J.A.; Rhea, C.K.; Schmitz, R.J.; Grooms, D.R.; Wilkins, R.W.; Slutsky, A.B.; Raisbeck, L.D. The Influence of Attentional Focus on Balance Control over Seven Days of Training. J. Mot. Behav. 2019, 51, 281–292. [Google Scholar] [CrossRef]
  4. Pantall, A.; Del Din, S.; Rochester, L. Longitudinal changes over thirty-six months in postural control dynamics and cognitive function in people with Parkinson’s disease. Gait Posture 2018, 62, 468–474. [Google Scholar] [CrossRef]
  5. Riva, F.; Toebes, M.J.P.; Pijnappels, M.A.G.M.; Stagni, R.; Van Dieën, J.H. Estimating fall risk with inertial sensors using gait stability measures that do not require step detection. Gait Posture 2013, 38, 170–174. [Google Scholar] [CrossRef]
  6. Ihlen, E.A.F.; Weiss, A.; Bourke, A.; Helbostad, J.L.; Hausdorff, J.M. The complexity of daily life walking in older adult community-dwelling fallers and non-fallers. J. Biomech. 2016, 49, 1420–1428. [Google Scholar] [CrossRef]
  7. Ahmadi, S.; Wu, C.; Sepehri, N.; Kantikar, A.; Nankar, M.; Szturm, T. The Effects of Aging and Dual Tasking on Human Gait Complexity During Treadmill Walking: A Comparative Study Using Quantized Dynamical Entropy and Sample Entropy. J. Biomech. Eng. 2018, 140, 011006. [Google Scholar] [CrossRef]
  8. Lamoth, C.J.; van Deudekom, F.J.; van Campen, J.P.; Appels, B.A.; de Vries, O.J.; Pijnappels, M. Gait stability and variability measures show effects of impaired cognition and dual tasking in frail people. J. Neuroeng. Rehabil. 2011, 8, 1–9. [Google Scholar] [CrossRef]
  9. Busa, M.A.; Jones, S.L.; Hamill, J.; van Emmerik, R.E. Multiscale entropy identifies differences in complexity in postural control in women with multiple sclerosis. Gait Posture 2016, 45, 7–11. [Google Scholar] [CrossRef]
  10. Gruber, A.H.; Busa, M.A.; Gorton Iii, G.E.; Van Emmerik, R.E.; Masso, P.D.; Hamill, J. Time-to-contact and multiscale entropy identify differences in postural control in adolescent idiopathic scoliosis. Gait Posture 2011, 34, 13–18. [Google Scholar] [CrossRef]
  11. Busa, M.A.; Van Emmerik, R.E.A. Multiscale entropy: A tool for understanding the complexity of postural control. J. Sport Health Sci. 2016, 5, 44–51. [Google Scholar] [CrossRef]
  12. Manor, B.; Costa, M.D.; Hu, K.; Newton, E.; Starobinets, O.; Kang, H.G.; Peng, C.K.; Novak, V.; Lipsitz, L.A. Physiological complexity and system adaptability: Evidence from postural control dynamics of older adults. J. Appl. Physiol. 2010, 109, 1786–1791. [Google Scholar] [CrossRef]
  13. Zhou, J.; Lipsitz, L.; Habtemariam, D.; Manor, B. Sub-sensory vibratory noise augments the physiologic complexity of postural control in older adults. J. NeuroEngineering Rehabil. 2016, 13, 44. [Google Scholar] [CrossRef]
  14. Zhang, X.; Zhou, P. Sample entropy analysis of surface EMG for improved muscle activity onset detection against spurious background spikes. J. Electromyogr. Kinesiol. 2012, 22, 901–907. [Google Scholar] [CrossRef]
  15. Rampichini, S.; Vieira, T.M.; Castiglioni, P.; Merati, G. Complexity Analysis of Surface Electromyography for Assessing the Myoelectric Manifestation of Muscle Fatigue: A Review. Entropy 2020, 22, 529. [Google Scholar]
  16. Murillo-Escobar, J.; Jaramillo-Munera, Y.E.; Orrego-Metaute, D.A.; Delgado-Trejos, E.; Cuesta-Frau, D. Muscle fatigue analysis during dynamic contractions based on biomechanical features and Permutation Entropy. Math. Biosci. Eng. MBE 2020, 17, 2592–2615. [Google Scholar] [CrossRef]
  17. Zhao, J.; Xie, Z.; Jiang, L.; Cai, H.; Liu, H.; Hirzinger, G. EMG control for a five-fingered underactuated prosthetic hand based on wavelet transform and sample entropy. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 3215–3220. [Google Scholar]
  18. Kullback, S. Information Theory and Statistics; John Wiley & Sons: New York, NY, USA, 1959. [Google Scholar]
  19. Shannon, C.E.; Weaver, W. The Mathematical Theory of Communication; University of Illinois Press: Urbana, IL, USA, 1949. [Google Scholar]
  20. Sinai, Y.G. On the Notion of Entropy of a Dynamical System. Dokl. Russ. Acad. Sci. 1959, 124, 768–771. [Google Scholar]
  21. Kolmogorov, A.N. Entropy per unit time as a metric invariant of automorphism. Dokl. Russ. Acad. Sci. 1959, 124, 754–755. [Google Scholar]
  22. Wolf, A.; Swift, J.B.; Swinney, H.L.; Vastano, J.A. Determining Lyapunov Exponents from a Time-Series. Phys. D 1985, 16, 285–317. [Google Scholar] [CrossRef]
  23. Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef]
  24. Pincus, S.M. Approximate Entropy as a Measure of System-Complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef]
  25. Pincus, S.M.; Gladstone, I.M.; Ehrenkranz, R.A. A regularity statistic for medical data analysis. J. Clin. Monit. Comput. 1991, 7, 335–345. [Google Scholar] [CrossRef]
  26. Fleisher, L.A.; Pincus, S.M.; Rosenbaum, S.H. Approximate Entropy of Heart-Rate as a Correlate of Postoperative Ventricular Dysfunction. Anesthesiology 1993, 78, 683–692. [Google Scholar] [CrossRef]
  27. Pincus, S.M.; Viscarello, R.R. Approximate entropy: A regularity measure for fetal heart rate analysis. Obstet. Gynecol. 1992, 79, 249–255. [Google Scholar]
  28. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiology. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  29. Pincus, S.M. Approximate Entropy (ApEn) as a Complexity Measure. Chaos 1995, 5, 110–117. [Google Scholar]
  30. Pincus, S.M.; Goldberger, A.L. Physiological time-series analysis: What does regularity quantify? Am. J. Physiol. 1994, 266, H1643–H1656. [Google Scholar] [CrossRef]
  31. Bollt, E.M.; Skufca, J.D.; McGregor, S.J. Control entropy: A complexity measure for nonstationary signals. Math. Biosci. Eng. MBE 2009, 6, 1–25. [Google Scholar]
  32. McGregor, S.J.; Bollt, E. Control Entropy: What is it and what does it tell us? Clin. Kinesiol. 2012, 66, 7–12. [Google Scholar]
  33. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 2002, 89, 068102. [Google Scholar] [CrossRef]
  34. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Phys. Rev. 2005, 71, 021906. [Google Scholar] [CrossRef]
  35. Govindan, R.B.; Wilson, J.D.; Eswaran, H.; Lowery, C.L.; Preissl, H. Revisiting sample entropy analysis. Phys. A 2007, 376, 158–164. [Google Scholar] [CrossRef]
  36. Caldirola, D.; Bellodi, L.; Caumo, A.; Migliarese, G.; Perna, G. Approximate entropy of respiratory patterns in panic disorder. Am. J. Psychiatry 2004, 161, 79–87. [Google Scholar] [CrossRef] [PubMed]
  37. Engoren, M. Approximate entropy of respiratory rate and tidal volume during weaning from mechanical ventilation. Crit. Care Med. 1998, 26, 1817–1823. [Google Scholar] [CrossRef]
  38. Sarlabous, L.; Torres, A.; Fiz, J.A.; Gea, J.; Martinez-Llorene, J.M.; Jane, R. Efficiency of mechanical activation of inspiratory muscles in COPD using sample entropy. Eur. Respir. J. 2015, 46, 1808–1811. [Google Scholar] [CrossRef] [PubMed]
  39. Tran, Y.; Thuraisingham, R.A.; Wijesuriya, N.; Nguyen, H.T.; Craig, A. Detecting neural changes during stress and fatigue effectively: A comparison of spectral analysis and sample entropy. In Proceedings of the 2007 3rd International IEEE/EMBS Conference on Neural Engineering, Kohala Coast, HI, USA, 2–5 May 2007; pp. 350–353. [Google Scholar]
  40. Chen, X.; Solomon, I.C.; Chon, K.H. Comparison of the use of approximate entropy and sample entropy: Applications to neural respiratory signal. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 17–18 January 2006; pp. 4212–4215. [Google Scholar]
  41. Song, Y.; Zhang, J. Discriminating preictal and interictal brain states in intracranial EEG by sample entropy and extreme learning machine. J. Neurosci. Methods 2016, 257, 45–54. [Google Scholar] [CrossRef] [PubMed]
  42. McCamley, J.D.; Denton, W.; Arnold, A.; Raffalt, P.C.; Yentes, J.M. On the Calculation of Sample Entropy Using Continuous and Discrete Human Gait Data. Entropy 2018, 20, 764. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Richman, J.S.; Lake, D.E.; Moorman, J.R. Sample Entropy. In Methods in Enzymology; Elsevier: Amsterdam, The Netherlands, 2004; pp. 172–184. [Google Scholar]
  44. Molina-Picó, A.; Cuesta-Frau, D.; Aboy, M.; Crespo, C.; Miró-Martínez, P.; Oltra-Crespo, S. Comparative study of approximate entropy and sample entropy robustness to spikes. Artif. Intell. Med. 2011, 53, 97–106. [Google Scholar] [CrossRef]
  45. Lake, D.E.; Richman, J.S.; Griffin, M.P.; Moorman, J.R. Sample entropy analysis of neonatal heart rate variability. Am. J. Physiology Regul. Integr. Comp. Physiol. 2002, 283, R789–R797. [Google Scholar] [CrossRef]
  46. Mandelbrot, B.B.; Van Ness, J.W. Fractional Brownian motions, fractional noises and applications. SIAM Rev. 1968, 10, 422–437. [Google Scholar] [CrossRef]
  47. Yentes, J.M.; Raffalt, P.C. Entropy Analysis in Gait Research: Methodological Considerations and Recommendations. Ann. Biomed. Eng. 2021, 49, 979–990. [Google Scholar] [CrossRef] [PubMed]
  48. Roume, C.; Ezzina, S.; Blain, H.; Delignieres, D. Biases in the Simulation and Analysis of Fractal Processes. Comput. Math. Methods Med. 2019, 2019, 4025305. [Google Scholar] [CrossRef] [PubMed]
  49. Davies, R.B.; Harte, D.S. Tests for Hurst effect. Biometrika 1987, 74, 95–101. [Google Scholar] [CrossRef]
  50. Barnsley, M.F.; Devaney, R.L.; Mandelbrot, B.B.; Peitgen, H.; Saupe, D.; Voss, R.F. The Science of Fractal Images; Springer: New York, NY, USA, 1988. [Google Scholar]
  51. Box, G.E.P.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  52. Granger, C.W.J.; Joyeux, R. An introduction to long-memory time series models and fractional differencing. J. Time Ser. Anal. 1980, 1, 15–29. [Google Scholar] [CrossRef]
  53. Fatichi, S. ARFIMA simulations (Version 1.0.0.0). Available online: https://www.mathworks.com/matlabcentral/fileexchange/25611-arfima-simulations (accessed on 28 November 2022).
  54. Inzelt, G. ARFIMA(p,d,q) Estimator (Version 1.1.0.0). Available online: https://www.mathworks.com/matlabcentral/fileexchange/30238-arfima-p-d-q-estimator (accessed on 28 November 2022).
  55. Sheppard, K. MFE Toolbox (commit 9622b6c). Available online: https://github.com/bashtage/mfe-toolbox/ (accessed on 28 November 2022).
  56. Diebolt, C.; Guiraud, V. A note on long memory time series. Qual Quant 2005, 39, 827–836. [Google Scholar] [CrossRef]
  57. Delignieres, D.; Ramdani, S.; Lemoine, L.; Torre, K.; Fortes, M.; Ninot, G. Fractal analyses for ‘short’ time series: A re-assessment of classical methods. J. Math. Psychol. 2006, 50, 525–544. [Google Scholar] [CrossRef]
  58. Almurad, Z.M.H.; Delignières, D. Evenly spacing in detrended fluctuation analysis. Phys. A 2016, 451, 63–69. [Google Scholar] [CrossRef]
  59. Liddy, J.J.; Haddad, J.M. Evenly Spaced Detrended Fluctuation Analysis: Selecting the Number of Points for the Diffusion Plot. Phys. A 2018, 491, 233–248. [Google Scholar] [CrossRef]
  60. Kaffashi, F.; Foglyano, R.; Wilson, C.G.; Loparo, K.A. The effect of time delay on Approximate & Sample Entropy calculations. Phys. D Nonlinear Phenom. 2008, 237, 3069–3074. [Google Scholar]
  61. Fraser, A.M.; Swinney, H.L. Independent coordinates for strange attractors from mutual information. Phys. Rev. A Gen. Phys. 1986, 33, 1134–1140. [Google Scholar] [CrossRef]
  62. Leys, C.; Ley, C.; Klein, O.; Bernard, P.; Licata, L. Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median. J. Exp. Soc. Psychol. 2013, 49, 764–766. [Google Scholar] [CrossRef]
  63. Hausdorff, J.M.; Peng, C.K.; Ladin, Z.V.I.; Wei, J.Y.; Goldberger, A.L. Is walking a random walk? Evidence for long-range correlations in stride interval of human gait. J. Appl. Physiol. 1995, 78, 349–358. [Google Scholar] [CrossRef]
  64. Delignieres, D.; Torre, K.; Bernard, P.L. Transition from persistent to anti-persistent correlations in postural sway indicates velocity-based control. PLoS Comput. Biol. 2011, 7, e1001089. [Google Scholar] [CrossRef]
  65. Eke, A.; Herman, P.; Bassingthwaighte, J.B.; Raymond, G.M.; Percival, D.B.; Cannon, M.; Balla, I.; Ikrenyi, C. Physiological time series: Distinguishing fractal noises from motions. Pflügers Arch. Eur. J. Physiol. 2000, 439, 403–415. [Google Scholar]
  66. Liu, H.; Shah, S.; Jiang, W. On-line outlier detection and data cleaning. Comput. Chem. Eng. 2004, 28, 1635–1647. [Google Scholar]
  67. Delignieres, D.; Torre, K. Fractal dynamics of human gait: A reassessment of the 1996 data of Hausdorff et al. J. Appl. Physiol. 2009, 106, 1272–1279. [Google Scholar] [CrossRef]
  68. Hausdorff, J.M.; Purdon, P.L.; Peng, C.K.; Ladin, Z.; Wei, J.Y.; Goldberger, A.L. Fractal dynamics of human gait: Stability of long-range correlations in stride interval fluctuations. J. Appl. Physiol. 1996, 80, 1448–1457. [Google Scholar] [CrossRef]
  69. Ducharme, S.W.; Liddy, J.J.; Haddad, J.M.; Busa, M.A.; Claxton, L.J.; van Emmerik, R.E.A. Association between stride time fractality and gait adaptability during unperturbed and asymmetric walking. Hum. Mov. Sci. 2018, 58, 248–259. [Google Scholar] [CrossRef] [PubMed]
  70. Kennel, M.B.; Brown, R.; Abarbanel, H.D. Determining embedding dimension for phase-space reconstruction using a geometrical construction. Phys. Rev. A 1992, 45, 3403–3411. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Error in the estimated scaling exponent, α, obtained from the ARFIMA (0, d, 0) model fits. (AC). Errors for the short (N = 250), medium (N = 500), and long (N = 1000) datasets, respectively. Each point represents one simulation. The black circles represent the mean error (i.e., bias); error bars represent the standard deviation.
Figure 1. Error in the estimated scaling exponent, α, obtained from the ARFIMA (0, d, 0) model fits. (AC). Errors for the short (N = 250), medium (N = 500), and long (N = 1000) datasets, respectively. Each point represents one simulation. The black circles represent the mean error (i.e., bias); error bars represent the standard deviation.
Entropy 25 00306 g001
Figure 2. Proportion of simulations classified as nonstationary by the ARFIMA (0, d, 0) models. (A). The initial analysis with wider gaps between the expected α values indicated that the classification accuracy was close to 50% near the 1/f boundary but less than 5% (dotted black lines) in nearby regions. There were minimal differences across dataset lengths. (B). The follow-up analysis with finer gaps revealed that classification accuracy was dependent on dataset length. Short datasets (N = 250) were classified least accurately on either side of the 1/f boundary. Medium datasets (N = 500) were classified less accurately than long datasets (N = 1000) but only below the 1/f boundary.
Figure 2. Proportion of simulations classified as nonstationary by the ARFIMA (0, d, 0) models. (A). The initial analysis with wider gaps between the expected α values indicated that the classification accuracy was close to 50% near the 1/f boundary but less than 5% (dotted black lines) in nearby regions. There were minimal differences across dataset lengths. (B). The follow-up analysis with finer gaps revealed that classification accuracy was dependent on dataset length. Short datasets (N = 250) were classified least accurately on either side of the 1/f boundary. Medium datasets (N = 500) were classified less accurately than long datasets (N = 1000) but only below the 1/f boundary.
Entropy 25 00306 g002
Figure 3. Entropy estimates were biased by small numbers of extreme observations. (AC). Biases in entropy estimates, expressed as a percentage, when 1% of the observations were manipulated without adjusting r. Entropy estimates were characterized by negative biases (>25%) except for α > 1.5. (DF). Biases in entropy estimates after adjusting r to account for outliers. Near-zero biases were observed for the stationary datasets (α < 1). Biases were less than 5% for nonstationary datasets with anti-persistent increments (1 < α < 1.5). The bias and standard deviation progressively increased for nonstationary datasets with persistent increments (α < 1.5). (GI). Biases in entropy estimates after outlier removal with similar results as adjusting r. However, removing outliers further reduced the bias and standard deviation for nonstationary datasets with persistent increments (α < 1.5).
Figure 3. Entropy estimates were biased by small numbers of extreme observations. (AC). Biases in entropy estimates, expressed as a percentage, when 1% of the observations were manipulated without adjusting r. Entropy estimates were characterized by negative biases (>25%) except for α > 1.5. (DF). Biases in entropy estimates after adjusting r to account for outliers. Near-zero biases were observed for the stationary datasets (α < 1). Biases were less than 5% for nonstationary datasets with anti-persistent increments (1 < α < 1.5). The bias and standard deviation progressively increased for nonstationary datasets with persistent increments (α < 1.5). (GI). Biases in entropy estimates after outlier removal with similar results as adjusting r. However, removing outliers further reduced the bias and standard deviation for nonstationary datasets with persistent increments (α < 1.5).
Entropy 25 00306 g003
Figure 4. Classification with ARFIMA (0, d, 0) models reduced biases in entropy estimates for nonstationary datasets independent of dataset length. (AC). Biases in entropy estimates obtained following outlier removal when classification was not applied. (DF). Biases in entropy estimates when outlier removal was preceded by data classification. Data classified as nonstationary were differenced and outliers were identified and removed. This procedure reduced the biases in entropy estimates across the entire α range. However, the standard deviation was inflated for nonstationary datasets with anti-persistent increments (1 < α < 1.5), indicating some tradeoffs with this approach.
Figure 4. Classification with ARFIMA (0, d, 0) models reduced biases in entropy estimates for nonstationary datasets independent of dataset length. (AC). Biases in entropy estimates obtained following outlier removal when classification was not applied. (DF). Biases in entropy estimates when outlier removal was preceded by data classification. Data classified as nonstationary were differenced and outliers were identified and removed. This procedure reduced the biases in entropy estimates across the entire α range. However, the standard deviation was inflated for nonstationary datasets with anti-persistent increments (1 < α < 1.5), indicating some tradeoffs with this approach.
Entropy 25 00306 g004
Figure 5. Limitations of SampEn to discriminate temporally correlated stochastic datasets. (AC). The estimated and expected temporal correlation properties, measured by the scaling exponent, α. The estimated α values were characterized by a linear relationship with a slope close to 1. Thus, the temporal correlation properties were accurately characterized by ARFIMA modeling independent of dataset length. The black lines represent the expected values. Each data point represents one of 100 simulated datasets. Means are represented by black circles. Error bars are standard deviations. (DF). By contrast, SampEn estimates showed a many-to-one correspondence with α, characterized by an inverted parabola that peaked for stationary uncorrelated noise (α = 0.5). Thus, SampEn may not reliably distinguish subtle differences in the dynamics of stochastic data. Each data point represents one of 100 simulated datasets. Means are represented by black circles. Error bars are standard deviations.
Figure 5. Limitations of SampEn to discriminate temporally correlated stochastic datasets. (AC). The estimated and expected temporal correlation properties, measured by the scaling exponent, α. The estimated α values were characterized by a linear relationship with a slope close to 1. Thus, the temporal correlation properties were accurately characterized by ARFIMA modeling independent of dataset length. The black lines represent the expected values. Each data point represents one of 100 simulated datasets. Means are represented by black circles. Error bars are standard deviations. (DF). By contrast, SampEn estimates showed a many-to-one correspondence with α, characterized by an inverted parabola that peaked for stationary uncorrelated noise (α = 0.5). Thus, SampEn may not reliably distinguish subtle differences in the dynamics of stochastic data. Each data point represents one of 100 simulated datasets. Means are represented by black circles. Error bars are standard deviations.
Entropy 25 00306 g005
Figure 6. Normalization of SampEn across different dataset lengths. (A). Unnormalized SampEn estimates for different values of the scaling exponent, α, and dataset length, N. Each data point represents one simulated dataset. Estimates obtained for the stationary datasets were robust to differences in dataset length, whereas the nonstationary datasets were more sensitive to dataset length. (B). Normalized SampEn estimates for the different values of the scaling exponent, α, and dataset length, N. Normalization incorrectly suggests that datasets characterized by the same temporal correlation properties contained different degrees of regularity.
Figure 6. Normalization of SampEn across different dataset lengths. (A). Unnormalized SampEn estimates for different values of the scaling exponent, α, and dataset length, N. Each data point represents one simulated dataset. Estimates obtained for the stationary datasets were robust to differences in dataset length, whereas the nonstationary datasets were more sensitive to dataset length. (B). Normalized SampEn estimates for the different values of the scaling exponent, α, and dataset length, N. Normalization incorrectly suggests that datasets characterized by the same temporal correlation properties contained different degrees of regularity.
Entropy 25 00306 g006
Table 1. Parameter values for the simulated data obtained from the ARFIMA (0, d, 0) models.
Table 1. Parameter values for the simulated data obtained from the ARFIMA (0, d, 0) models.
Parameter
Difference Order (d)Scaling Exponent (α)
Stationary (fGn)−0.40.1
−0.30.2
−0.20.3
−0.10.4
0.00.5
0.10.6
0.20.7
0.30.8
0.40.9
0.490.99
Nonstationary (fBm)0.61.1
0.71.2
0.81.3
0.91.4
1.01.5
1.11.6
1.21.7
1.31.8
1.41.9
Table 2. Bias and standard deviation (SD) of the estimated scaling exponents, α.
Table 2. Bias and standard deviation (SD) of the estimated scaling exponents, α.
N = 250N = 500N = 1000
Expected αBiasSDBiasSDBiasSD
0.1−0.0150.058−0.0040.036−0.0050.028
0.2−0.0220.050−0.0090.039−0.0070.025
0.3−0.0190.053−0.0170.038−0.0070.020
0.4−0.0220.053−0.0050.038−0.0060.026
0.5−0.0170.055−0.0120.034−0.0060.024
0.6−0.0240.058−0.0130.044−0.0100.022
0.7−0.0210.059−0.0100.041−0.0040.023
0.8−0.0230.047−0.0160.037−0.0020.025
0.9−0.0220.0510.0010.0380.0010.023
0.99−0.0010.0550.0010.0380.0080.026
1.1−0.0150.057−0.0130.037−0.0020.027
1.2−0.0250.061−0.0090.036−0.0010.024
1.3−0.0190.058−0.0100.036−0.0090.027
1.4−0.0080.045−0.0100.041−0.0080.023
1.5−0.0190.056−0.0080.037−0.0060.029
1.6−0.0190.058−0.0060.050−0.0070.028
1.7−0.0140.0550.0060.070−0.0140.037
1.8−0.0330.0770.0050.067−0.0140.027
1.9−0.0120.075−0.0320.066−0.0320.030
Table 3. Mean and standard deviation (SD) of SampEn(m = 1, r = 0.25, τ = 1) for each expected value of the scaling exponent, α, and dataset length, N.
Table 3. Mean and standard deviation (SD) of SampEn(m = 1, r = 0.25, τ = 1) for each expected value of the scaling exponent, α, and dataset length, N.
N = 250N = 500N = 1000
Expected αMeanSDMeanSDMeanSD
0.11.9270.0431.9270.0311.9260.016
0.21.9470.0531.9390.0301.9370.018
0.31.9590.0471.9530.0271.9510.017
0.41.9710.0501.9620.0291.9640.017
0.51.9690.0431.9650.0241.9680.015
0.61.9570.0451.9620.0291.9600.018
0.71.9470.0461.9400.0321.9340.015
0.81.9060.0521.8930.0391.8810.026
0.91.8300.0691.7810.0641.7780.040
0.991.6970.1091.6460.0831.6130.067
1.11.5120.1611.4420.1411.3150.159
1.21.2730.2051.1890.1560.9630.196
1.31.0330.2340.8170.2180.6800.193
1.40.7440.2160.5390.2040.4310.177
1.50.4810.2100.3420.1880.2280.100
1.60.3100.1590.1950.1090.1310.069
1.70.2110.1370.1260.0680.0740.035
1.80.1440.1040.0790.0430.0510.026
1.90.0910.0800.0580.0430.0350.019
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liddy, J.; Busa, M. Considerations for Applying Entropy Methods to Temporally Correlated Stochastic Datasets. Entropy 2023, 25, 306. https://doi.org/10.3390/e25020306

AMA Style

Liddy J, Busa M. Considerations for Applying Entropy Methods to Temporally Correlated Stochastic Datasets. Entropy. 2023; 25(2):306. https://doi.org/10.3390/e25020306

Chicago/Turabian Style

Liddy, Joshua, and Michael Busa. 2023. "Considerations for Applying Entropy Methods to Temporally Correlated Stochastic Datasets" Entropy 25, no. 2: 306. https://doi.org/10.3390/e25020306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop