Next Article in Journal
Denoising Non-Stationary Signals via Dynamic Multivariate Complex Wavelet Thresholding
Previous Article in Journal
Design of Closed-Loop Control Schemes Based on the GA-PID and GA-RBF-PID Algorithms for Brain Dynamic Modulation
Previous Article in Special Issue
Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Complexity Reduction in Analyzing Independence between Statistical Randomness Tests Using Mutual Information

by
Jorge Augusto Karell-Albo
1,
Carlos Miguel Legón-Pérez
1,
Raisa Socorro-Llanes
2,
Omar Rojas
3 and
Guillermo Sosa-Gómez
3,*
1
Instituto de Criptografía, Facultad de Matemática y Computación, Universidad de la Habana, Habana 10400, Cuba
2
Facultad de Ingeniería Informática, Universidad Tecnológica de la Habana José Antonio Echeverría (CUJAE), Habana 19390, Cuba
3
Facultad de Ciencias Económicas y Empresariales, Universidad Panamericana, Álvaro del Portillo 49, Zapopan 45010, Jalisco, Mexico
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(11), 1545; https://doi.org/10.3390/e25111545
Submission received: 11 October 2023 / Revised: 9 November 2023 / Accepted: 13 November 2023 / Published: 15 November 2023
(This article belongs to the Special Issue Information Theory in Multi-Agent Systems: Methods and Applications)

Abstract

:
The advantages of using mutual information to evaluate the correlation between randomness tests have recently been demonstrated. However, it has been pointed out that the high complexity of this method limits its application in batteries with a greater number of tests. The main objective of this work is to reduce the complexity of the method based on mutual information for analyzing the independence between the statistical tests of randomness. The achieved complexity reduction is estimated theoretically and verified experimentally. A variant of the original method is proposed by modifying the step in which the significant values of the mutual information are determined. The correlation between the NIST battery tests was studied, and it was concluded that the modifications to the method do not significantly affect the ability to detect correlations. Due to the efficiency of the newly proposed method, its use is recommended to analyze other batteries of tests.

1. Introduction

In cryptographic applications, sequences of random numbers are very important. One of the widest-spread applications is the generation of secret keys. These sequences are obtained using random number generators, which, depending on the source of randomness, can be classified as pseudo-random number generators (PRNGs) or truly random number generators (TRNGs). Some authors, such as [1], include a third group of generators called hybrid random number generators (HRNGs). The HRNG combines elements of the previous two generators.
According to [2], the best way to generate “unpredictable” random numbers is to use physical processes, such as atmospheric or thermal noise, cosmic radiation, and some other phenomena; however, number generation from physical processes is relatively inefficient since TRNGs are expensive devices both in terms of execution and applicability. For this reason, ref [3] states that most systems use PRNGs instead of TRNGs.
Weak RNGs have been identified as the cause of many security problems in computer systems. A concrete case is exemplified by MIFARE Classic, an RFID radio frequency identification) card manufactured by NXP Semiconductors that had a weak PRNG, making it possible to attack their channel of supposedly secure communications [4]. Furthermore, ref [5] described pre-play and man-in-the-middle attacks that demonstrated how, due to a weak RNG, a payment system could be abused. An explanation of how it was possible to attack Taiwan’s national database of “Citizen Digital Certificates” was presented in [6], where it was due to poor implementation of the TRNG in the Office-certified Renesas AE45C1 v.01 smart card IC German Federal Information Security (BSI) [7]. These examples highlight the need for further analysis of RNG design and implementation to mitigate the possibility of such attacks in the future.
Because PRNGs are based on deterministic algorithms, it is necessary to examine the output to confirm its suitability for cryptographic applications. This output is statistically analyzed by one or several randomness tests, and the results are evaluated to determine whether the generator is random or not. The statistic associated with each statistical test attempts to identify the presence or absence of a specific pattern. If the pattern is detected, then the lack of randomness in the sequence can be inferred. There are many statistical tests for randomness. The same test can even be applied multiple times with different combinations of parameters [8].
Although some authors [9] recommend using multiple statistical randomness tests to leverage each test’s strengths, applying too many tests may lead to overestimating the properties of a PRNG. Thus, ref [10] suggest that two conceptually different tests may evaluate the same characteristic of randomness and produce correlated results. In addition, not only can the correlation between different randomness tests be studied, but the same test with different parameters can be correlated. Therefore, correlation analysis between randomness tests can also be used for parameter selection. In [11,12], the study of correlations between tests using the Pearson correlation coefficient (CCP) was proposed; however, this coefficient is not capable of detecting nonlinear correlations.
Results reported by [13] demonstrate the need for a greater variety of statistical tests. In their work, they studied eight RNGs, and, in some cases, generators showed difficulties in passing the tests of Alphabits, Rabbit, Small Crush, and Crush batteries, passing the tests of Dieharder and SP800-22. As [13] points out, future batteries of statistical tests should be thoroughly analyzed for possible correlations between their tests before publication. Likewise, a continuous review of the new randomness tests must be carried out to ensure their efficiency and statistical strength and integrate them into the batteries if they are not correlated with the existing tests or are superior to them.
In [12,14], Pearson’s correlation coefficient was used to analyze correlations between tests; however, [2,11] focused on the proportion of regions where p-values are less than 0.01 . Then, ref [15] analyzed the dependencies between nine NIST tests and found the same dependencies as [12]. Another approach proposed by [16] analyzed the difference between p-values corresponding to two different tests and detected correlations according to the distribution of that difference.
In [14], dependencies between some NIST tests were determined using the Pearson correlation coefficient, and patterns were found in the evolution of these dependencies according to specific factors, such as the length of the sequences analyzed by the tests. In [17], a novel method was proposed to detect the correlation between statistical randomness tests using mutual information for the first time. This measure was used to examine the tests present in the NIST SP 800-22 battery, detecting new correlations.
Different articles have verified the effectiveness of mutual information in studying the correlation between the randomness tests of different batteries. In [18], DieHarder, TufTest, and SmallCrush batteries were studied, and in [19], the FIPS 140-2 battery was analyzed using the method proposed in [17]. Mutual information can detect the correlations found by Pearson correlation and new nonlinear dependencies between tests. In [20], the complexity of the method proposed in [17] was estimated, and its high complexity was pointed out as a disadvantage, which prevents testing batteries with more tests or longer sequences. However, mutual information as a new measure of independence between randomness tests represents a significant advance and an important piece for building a definitive test battery [18].
The main objective of this work is to reduce the complexity of the method proposed by [17], which we will call from now on MIRT-1 (Mutual Information to analyze Statistical Randomness Test v1). The step in which the mutual information values corresponding to significant correlations are determined is modified to achieve this. The complexity reduction achieved with this modification is estimated theoretically and then confirmed experimentally. The efficiency of the new proposed method (MIRT-2) will be determined by considering the complexity calculation. On the other hand, the efficacy will be measured according to the number of correlations that it can detect compared to the MIRT-1 method.

2. Preliminaries

2.1. Mutual Information

Mutual information (MI) is an important measure of statistical correlation. It stands out for its ability to detect the correlation between variables and possesses properties that make it an ideal measure of stochastic dependence [21,22,23,24]. Unlike the Pearson correlation coefficient (PCC), which only considers linear dependencies, or other correlation coefficients that only detect monotonic dependencies, MI considers all correlations regardless of their functional form.
The MI between two discrete random variables X and Y is defined as
M I ( X , Y ) = y Y x X p ( x , y ) log p ( x , y ) p ( x ) p ( y ) ,
where p ( x , y ) is the joint probability function of X and Y, and p ( x ) and p ( y ) are the marginal probability distribution functions of X and Y, respectively. MI can also be defined in terms of entropy as
M I ( X , Y ) = H ( X ) + H ( Y ) H ( X , Y ) ,
where H ( X ) and H ( Y ) are the marginal entropy of the variables X and Y, respectively, and H ( X , Y ) is the joint entropy of both variables. For the case of continuous random variables, the sums are replaced by integrals:
M I ( X , Y ) = Y X p ( x , y ) log p ( x , y ) p ( x ) p ( y ) d x d y .
More on the properties of entropy and mutual information for continuous variables and their relationship to the discrete case can be found in [25].

2.2. Estimating Mutual Information

When the probability distributions are unknown, it is not possible to calculate the exact value of M I ( X , Y ) , so it is necessary to calculate a sample estimator M I ^ ( X , Y ) . Estimators of the mutual information M I ^ differ in estimating the probabilities of the marginal and joint densities. Some of the estimators proposed in the literature use discretization [26,27], kernels [28,29] or correlation integrals [30], k-nearest neighbors [31,32], B-splines [33], or Gram–Charlier polynomial expansion [34].
One of the simplest estimators is the maximum-likelihood (ML) estimator (plug-in). With this, entropy is estimated from the observed individual and joint frequencies. Since there are no assumptions about the data distribution, the ML estimator is considered non-parametric. Another of the estimators used is the James–Stein shrinkage estimator. The approach of this estimator is considered semi-parametric because it has characteristics of parametric and non-parametric methods [35]. A comprehensive comparison of estimators can be found in [35,36].
According to [37], the most common mutual information estimator is the naive equidistant binning estimator. This estimator considers each variable’s domain partition in a finite number, n, of discrete intervals (equidistant partition). The number of intervals for each variable is the same, so the parameter to optimize is the number of intervals to discretize or, equivalently, the length of the interval. The binning process begins with selecting the support interval, ( a , b ) . Usually, this interval is constructed from the smallest and largest values in the sample { x 1 , x 2 , , x n } . Given the interval ( a , b ) , a frequency histogram is constructed from the number of points in each interval. The intervals, B k , are half-open, i.e., B k = [ t k , t k + 1 ) for k = , , 1 , 0 , 1 , , , where the points t k satisfy that t k < t k + 1 . The width of each interval is denoted by h k = t k + 1 t k , and the number of elements in the interval is denoted by v k . In this way, given the sample x 1 , x 2 , , x n , we have
v k = i = 1 n 1 x i B k ,
where 1 ( · ) is the indicator function. It is clear that v k 0 and k v k = n . Generally, the intervals are selected so they are of the same width, i.e., h k = h , k . For equal-width intervals, we define
h = b a k .
Many researchers have attempted to determine the optimal value, k, of the intervals, but these methods often rely on strong assumptions about the distribution of the data [38]. Depending on the distribution type and the analysis objective, different values of h can be selected. Experimentation is usually necessary to determine the optimal width. Table 1 presents some of the most used rules for selecting k.
In most cases, the number of bins to discretize is set to k = 10 [50]. This work will discretize it in k = 10 intervals because it is the same approach used in [17] to design the MIRT-1 method.

2.3. Distribution of Estimators

The ML estimator of mutual information M I ^ M L ( X , Y ) = H ^ M L ( X ) + H ^ M L ( Y ) H ^ M L ( X , Y ) has been extensively studied in the literature [51,52]. It is known from [51] that, under certain conditions, such as finite alphabet size and M I > 0 , the following holds:
n ( M I ^ M L M I ) σ ^ N ( 0 , 1 ) .
On the other hand, in [52,53], it was stated that
2 n M I ^ M L χ ( I 1 ) ( J 1 ) 2
where I and J are the sizes of the alphabets of X and Y, respectively.
In [17], the ML, Miller–Madow, James–Stein, and Schurmann–Grassberger estimators were compared as the sample size increased. For the selection of the mutual information estimator, two pairs of variables were used, one of independent variables (see Figure 1) and the other of dependent variables (see Figure 2). It was concluded that, for more than 10,000 observations, the difference between the estimators is very small: 0.0074 for the pair of independent variables and 0.0065 for the dependent ones. However, the James–Stein estimator (shrinkage) was selected since the mutual information took the value of 0 for the independent variables, even for very small sample sizes.
Although for this particular case, only these mutual information estimators are being compared, a more comprehensive study of the entropy estimators can be found in [36]. The entropy estimator James–Stein shrinkage used to calculate mutual information in [17] is defined as
H ^ S H = x K p ^ x S H log 2 p ^ x S H
where
p ^ x S H = λ ^ t x + ( 1 λ ^ ) p ^ x M L
with
λ ^ = 1 x = 1 k ( p ^ x M L ) 2 ( n 1 ) x = 1 k ( t x p ^ x M L ) 2
and
t k = 1 k

2.4. MIRT-1 Method

As the distribution of the shrinkage estimator of the mutual information is unknown, a permutation test was performed to decide which values were significantly greater than 0. A permutation test is a statistical method used to determine the significance of a particular statistic by comparing it to a null distribution generated by randomly permuting the data. In this case, the permutation test determines whether the mutual information between two variables significantly differs from what would be expected by chance.
The null hypothesis ( H 0 ) is that there is no relationship between the two variables, and any observed mutual information is due to chance. The alternative hypothesis ( H 1 ) is that there is a significant relationship between the two variables, and the observed mutual information is not due to chance. If the p-value is less than a predetermined significance level, then the null hypothesis can be rejected in favor of the alternative hypothesis, indicating that there is a significant relationship between the two variables. The permutation test is applied following these steps:
  • Let X = ( x 1 , , x n ) and Y = ( y 1 , , y n ) be continuous random variables.
  • Construct the permuted samples X , π i ( Y ) , i = 1 , k ¯ in such a way that the possible association between X and Y disappears, π i being the permutation i of elements of Y, i.e.,
    • π i S n , i = 1 , k ¯ ;
    • π i π j , for i j ;
    • π 0 is the identity of S n .
  • Estimate the MI of the allowed samples to obtain { Z i } i = 0 k , where Z i = M I ( X , π i ( Y ) ) .
  • The p-value associated with the test is calculated by
    p -value = j = 1 k I Z 0 ( Z i ) k ,
    where I Z 0 ( Z i ) is the indicator function defined by
    1 Z 0 ( x ) = 1 i f x Z 0 , 0 i f x < Z 0 .
  • If p -value α , then the null hypothesis is not-rejected.
In [17], these steps were performed for each pair of random statistical tests, computing q = 10 , 000 permutations in each case.
The MIRT-1 method proposed in [17] is described in the following steps:
  • Select PRNGs.
    The selected generators must generate outputs that satisfy the randomness conditions.
  • Build the data samples using the selected generators.
    Generate n sequences of random numbers of length L to be evaluated using the selected statistical randomness tests.
  • Evaluate each of the n sequences using the k statistical randomness tests to obtain the corresponding p-values for each T i test (with i = 1 , , k ).
  • Compute the MI between sequences of p-values to detect the possible presence of correlations.
    Estimate the MI between pairs ( T i , T j ) of sequences of p-values to detect the presence of correlation.
    In the case of the MIRT-1 method, the estimator used is the shrinkage estimator.
    For a better interpretation in [17], the MI values were normalized, i.e., M I ( T i , T j ) = I M ( T i , T j ) H ( T i ) , where H ( T i ) represents the entropy of the variable T i .
  • Determine the significant correlations to conclude the correlation between the tests using the permutation test. The MI values are grouped in a triangular matrix
    M = H ( T 1 ) M I ( T 1 , T 2 ) M I ( T 1 , T k ) 0 H ( T 2 ) M I ( T 2 , T k ) 0 0 H ( T k ) k × k
    where M I ( T i , T j ) represents the MI between i and j. The diagonal of the matrix contains the values H ( T i ) that represent the entropy of the variable T i .
The complexity of the MIRT-1 method is estimated in [20] using the following:
O ( k 2 · q ( n + d 2 ) ) ,
where k represents the number of statistical tests to be analyzed, q is the number of permutations to determine the significance of the correlations, n is the length of the sequences of p-values analyzed, and d is the number of intervals in the discretization process. In practical applications, it is of interest to increase k, and the parameter d is selected in such a way as to increase the effectiveness of detection. Therefore, the feasible parameters to reduce are q and n, but the reduction in n can affect the effectiveness of the estimator, so, in this work, we will focus on reducing the value q.

3. Reducing the Complexity of the Method Based on Mutual Information to Analyze the Independence between the Statistical Tests of Randomness

This section presents the main contribution of this work: significantly reducing the complexity of the MIRT-1 method.

3.1. Solution Idea

To reduce complexity, it is proposed to eliminate the permutation test, which poses a new problem. In Step 5 of the MIRT-1 method, deciding which values of the mutual information are significant is necessary to conclude the correlation between the tests. Given the continuous random variables X and Y, decide if the value M I ( X , Y ) is significantly greater than 0, and thus conclude if there is some dependency between both variables. The hypothesis test asks the following:
H 0 : M I ( X , Y ) = 0 H 1 : M I ( X , Y ) > 0 ,
where H 0 is the null hypothesis that states the independence between X and Y, and H 1 is the alternative hypothesis, where there would be some association between X and Y.
To determine the significant values without using the permutation test, we propose to replace the James–Stein (shrinkage)-type estimator M I ^ S H , whose distribution is unknown, with another estimator of mutual information whose distribution is known. The mean square error of the new estimator should be reasonably small for the selected sample size. In this paper, we will use a maximum-likelihood estimator transformation M I ^ M L whose distribution is known [51,52]. It is known that the maximum-likelihood estimator of entropy is normally distributed [54] asymptotically with mean
E ( H ^ M L H ) = k 1 2 n + 1 12 n 2 1 i = 1 k 1 p i + O ( n 3 ) ,
and variance,
σ 2 ( H ^ M L ) = 1 n i = 1 k p i ln 2 ( p i ) H 2 + k 1 2 n 2 + O ( n 3 ) .
The mean can be calculated if the order n 2 is neglected. However, for the case of the variance, the terms of order n depend on the unknown probabilities. Therefore, the transformation of Equation (7) will be used since the mean and variance are known, and it is known that the normal distributes asymptotically when the number of degrees of freedom is greater than 30.

3.2. Selection of the Critical Value to Determine the Significant Correlations

To decide whether to reject the null hypothesis of independence between ( T i , T j ) in the new variant of the method, a priori knowledge of the distribution of the transformation 2 n M I ^ M L ( T i , T j ) of estimator M I ^ M L will be used. The transformation 2 n M I ^ M L ( T i , T j ) distributes χ 2 (Equation (7)). For our case, I = J = d is the number of discretization intervals. Then,
2 n M I ^ M L ( T i , T j ) χ ( d 1 ) 2 2 .
To guarantee that the χ 2 distribution of 2 n M I ^ M L ( T i , T j ) approximates the normal distribution, the condition ( d 1 ) 2 > 30 will be imposed on d, which implies that d > 1 + 30 = 6.48 , concluding that, for values of d > 6.48 , we have
2 n M I ^ M L ( T i , T j ) N ( d 1 ) 2 , 2 ( d 1 ) .
For the particular case of study, if we discretize d = 10 > 6.48 intervals, we obtain ( d 1 ) 2 = ( 9 ) 2 = 81 30 , which guarantees that the theoretical distribution of 2 n M I ^ M L ( T i , T j ) closely approximates the normal distribution (Equation (19)). Using this theoretical distribution, the critical value corresponding to the α prefix can be taken, and therefore it is unnecessary to use the permutation test.
For the sampling distribution of 2 n M I ^ M L ( T i , T j ) to be close to its theoretical normal distribution, it is convenient that the length n of the sequence of p-values is big enough. So,
Z M L = 2 n M I ^ M L ( T i , T j ) ( d 1 ) 2 2 ( d 1 ) N ( 0 , 1 )
distributes the normal asymptotically when ( d 1 ) 2 30 . For an α significance level, the critical value for the right tail is Z 1 α . Therefore, if the estimated values of Z M L are greater than the critical value, it can be concluded that there is a correlation between T i and T j with probability α that they are independent. For example, for α = 0.01 , if Z M L > Z 1 0.01 = 2.36 , it is concluded that the value Z M L is significant.

3.3. Method Modification Proposal

From Step 1 to Step 3, the MIRT-1 method remains unchanged. In Step 4, in [17], the shrinkage estimator M I ^ S H was used, while now it is proposed to use the ML estimator I M ^ M L . In Step 5, in [17], the values of M I ^ S H were normalized. A permutation test was applied; on the other hand, in this new method used to determine the significant values of M I ^ M L , the transformation Z M L is computed, and knowledge of the normal distribution of Z M L will be used.
The application of the new MIRT-2 method from this modification is as follows (the steps that changed concerning MIRT-1 are indicated in bold):
  • Select PRNGs.
    The selected generators must generate outputs that satisfy the randomness conditions.
  • Build the data samples using the selected generators.
    Generate n sequences of random numbers of length L to be evaluated using the selected statistical tests of randomness.
  • Evaluate each of the n sequences using the k statistical tests of randomness to obtain the corresponding p-values for each T i test (with i = 1 , , k ).
  • Compute the MI between sequences of p-values to detect the presence of correlations.
    Calculate the MI between pairs ( T i , T j ) of sequences of p-values to detect the presence of correlation.
    The MI estimator used is the ML M I ^ M L
  • Calculate Z M L and compare it with the critical value associated with the default α value. If Z ^ M L > Z 1 α , the null hypothesis of independence between the tests of randomness is rejected.
Table 2 summarizes the parameters chosen for the MIRT-1 and MIRT-2 methods.
The complexity of the MIRT-1 method was calculated in [20]
O ( k 2 · q ( n + d 2 ) ) .
The modification proposed in this paper reduces it to
O ( k 2 ( n + d 2 ) ) ,
since not applying the permutations test is equivalent to taking q = 1 . The reduction achieved is of the order
k 2 · q ( n + d 2 ) k 2 ( n + d 2 ) = q ,
so the modified method is expected to be about q-times faster. For example, in [17,20], q = 10 , 000 was used; for this case, the MIRT-2 method would have a complexity 10,000 times lower than MIRT-1.
The replacement of the shrinkage estimator M I ^ S H of mutual information used in the MIRT-1 method with the maximum-likelihood estimator M I ^ M L allows complexity to be reduced since the permutation test is eliminated and the known distribution of the new estimator is used to set the critical value and select the significant values. This modification could affect the ability to detect correlations if appropriate measures are not taken. To avoid this impact, it is proposed to apply the MIRT-2 method with a sample size equal to 10 000 , and it is recommended not to reduce this value while this M I ^ M L estimator is used. This recommendation is based on the results of Figure 2 of [17] since, for this sample size, the difference between the values of the estimators is very small and the ability to detect correlations is not affected.

4. Experimental Validation

In this section, the reduction in the complexity of MIRT-2 to MIRT-1 will be verified experimentally.

4.1. Experimental Check of Normality of Z M L

4.1.1. Design of Experiment 1

For random sequences, the p-values generated by statistical randomness tests follow a uniform distribution with values between 0 and 1 [9]. For this reason, the experiments generated data that followed the same distribution. To study the normality of the transformation Z M L and the pairs ( X i , Y i ) , i = 1 , , 1000 of independent and identically distributed random variables, where X , Y U ( 0 , 1 ) , X = { x 1 , x 2 , . . . , x 10 , 000 } and Y = { y 1 , y 2 , . . . , y 10 , 000 } . Subsequently, the MI between the pairs was calculated using the ML estimator, and 1000 M I ^ M L observations were obtained. The MI values were transformed using Equation (20) to obtain a sample of 1000 Z M L values. The Anderson–Darling and Kolmogorov–Smirnov tests were applied to verify the normality of Z M L .

4.1.2. Results

Table 3 shows p-values for the Anderson–Darling and Kolmogorov–Smirnov normality tests.
For all cases, the p -value > 0.05 ; therefore, there is insufficient evidence to reject the null hypothesis of normality. In Figure 3, a histogram with the values of Z M L and the estimate of the kernel density (KDE) of the sample data (black) and the curve of normal density (blue) can be observed. The blue line represents the theoretical normal probability density function, while the black curve represents a smoothed density estimate using a kernel density estimation technique.
In the Q-Q graph of Figure 4, it can be seen that the observed values are close to the expected ones. Through normality tests, it was possible to verify experimentally that the transformation Z M L follows a normal distribution under the theoretical argumentation presented in Section 3.2. The red line in the Q-Q plot graph represents the theoretical quantiles of a standard normal distribution, and it serves as a reference line for comparing the distribution of the plotted data to the normal distribution.

4.2. Analysis of the Effectiveness of the Proposed Variant

4.2.1. Design of Experiment 2

Based on the idea of the previous section, the MI between the NIST statistical randomness tests was calculated using the MIRT-2 method. For this, the ML estimator was used, and the symmetric matrix of dimension 17 × 17 was obtained with the values of the MI. It is important to highlight that for selecting the p-values, the approach proposed by [20] was followed. All sequences that did not comply with the requirement were discarded. There was a required number of cycles for the random excursions and random excursions variant tests, thus generating p-values equal to 0. Under the standard normal distribution and the independence assumption, a significance level α = 0.001 corresponds to a critical value C V = 3.090232 . Therefore, if Z M L > 3.09 , the null hypothesis of independence is rejected. Following Step 5 of the MIRT-2 method, those greater than the critical value were selected as significant values of the Z M L .

4.2.2. Results

In Figure 5, the histogram is represented with the k ( k 1 ) / 2 = 136 observations of Z M L for the values of the mutual information between the pairs of statistical tests of randomness. High values of Z M L (greater than 300) are seen on the right, indicating possible correlations between test pairs. In Figure 5 and Figure 6, the dotted line represents the critical value selected for the MI.
In Figure 6, only the observations of Z M L are represented for the values between 0 and 6. It can be noted that the values of Z M L smaller than the critical value behave following a normal distribution, as expected under the independence hypothesis.
Figure 7 shows the mutual information matrix with the significant values for the selected critical value. The values of Z M L greater than C V = 3.09 are indicated in red.
Table 4 compares the correlations detected by the two variants. The paper presented by [17] did not consider the pre-processing of invalid sequences for the random excursions and random excursions variant tests. For this reason, the correlations detected with the MIRT-1 method in [20] and the new MIRT-2 variant proposed in this article will be compared.
The correlation between the CUSUM (b) and non-overlapping template tests was not detected by the new MIRT-2 method, indicating that the new variant for this case loses efficacy concerning MIRT-1. This may be due to substituting the shrinkage estimator for the ML. However, according to what was expressed in [20], this correlation is quite small, so its detection may be due to an error or not easy to detect.

4.2.3. Design of Experiment 3

To verify the reduction in complexity of the new variant of the method concerning MIRT-1, the experiments presented in [17] were repeated. The methods were implemented in R using the entropy library in a computer running Windows 10 (64-bit) operating system, Intel Core i7-3770 3.40 GHz CPU, and 32 GB RAM. The MI with the two versions, MIRT-1 and MIRT-2, was calculated, and the times were compared. This experiment was repeated ten times. Results are shown in Table 5.
As seen in Table 5, the times were reduced on average 19,444-fold approximately, where 19 , 444 > q = 10 , 000 . It can be seen that the reduction in time observed in practice is even greater than the expected theoretical reduction.
The times for detection of correlations between the 17 tests of the NIST battery with sequences of p-values of length 10,000 decreased from approximately 1.7 h to less than 1 s (see Figure 8).

4.3. Analysis of the Consistency and Stability of the MIRT-2 Method

In this section, a detailed analysis of the consistency of the method MIRT-2 and MIRT-1 and the stability of MIRT-2 will be carried out.

4.3.1. Consistency

To analyze the consistency of the estimators M I ^ M L and M I ^ S H of mutual information, a Bland–Altman test was performed. The Bland–Altman test is a statistical method used to assess the agreement between two measurement methods. In this case, the experiment aimed to check the consistency between the estimators M I ^ M L and M I ^ S H .
To perform the Bland–Altman test, the following steps were taken:
  • Generate 10,000 samples of independent and identically distributed random variables U ( 0 , 1 ) .
  • Calculate the mutual information using both M I ^ M L and M I ^ S H estimators for each sample.
  • Calculate the mean and difference between the two methods for each sample. To check the assumptions of normality of the differences, a test for normal distribution, such as the Shapiro–Wilk or Kolmogorov–Smirnov test, can be conducted for the hypothesis that the distribution of the observations in the sample is normal (Figure 9)) (if p < 0.05 , then reject normality).
  • Calculate the mean difference and the LOAs (limits of agreement) (mean difference ± 1.96 times the standard deviation of the differences.)
  • Interpret the results. If the mean difference is close to zero and the limits of agreement are narrow, then the two methods are considered to be in good agreement.
If the mean difference is close to zero and the LOA is narrow, it suggests that the two methods are consistent and can be used interchangeably. On the other hand, if the mean difference is far from zero and/or the LOA is wide, it suggests that the two methods are inconsistent and cannot be used interchangeably.
In this case, the test was performed on 10,000 comparisons of the two estimators. The main measures are summarized in Table 6, resulting in a bias of 0.001768216 and a standard deviation of bias of 0.0002812867. The upper and lower LOAs were calculated to be 0.002319538 and 0.001216894, respectively.
The bias is within the LOA, indicating that there is no significant difference between the two measures. However, the bias is not zero, indicating that there is some systematic difference between the two measures. The upper and lower LOAs are relatively narrow (Figure 10), indicating good agreement between the two measures.
Overall, the results of the Bland–Altman test suggest that the two measures being compared are in good agreement, with a small systematic difference between them.

4.3.2. Stability

To assess the stability of this estimator, a stability test was carried out using a bootstrap test with 10,000 repetitions.
The hypothesis being tested is that the M I ^ M L estimator is stable, producing consistent results when applied to different data samples. The null hypothesis is that the M I ^ M L estimator is unstable, producing inconsistent results when applied to different data samples.
To perform the bootstrap test, multiple data samples were randomly selected with replacements from the original dataset. The M I ^ M L estimator was applied to each sample, and the resulting mutual information values were recorded. This process was repeated 10,000 times to generate a distribution of mutual information values.
This was carried out for each dataset of each of the NIST statistical tests. The results showed that of the seven correlations detected with the original data, after carrying out the bootstrap test, six correlations were maintained (Figure 11). It can be concluded that, in general, the MIRT-2 method presents good stability. It would be necessary to study in depth the reason why the correlation between the overlapping and longest run tests did not remain stable.

5. Conclusions

In this work, we reduced the complexity of the MIRT-1 method proposed by [17] q-fold by modifying the selection criteria for significant values. The complexity reduction was estimated theoretically and confirmed through experimentation. In addition, it was concluded that this modification does not significantly affect the method’s effectiveness in detecting correlations. Therefore, it was proposed for this modification to be implemented in future applications of the MIRT-1 method to enhance its efficiency. As directions for future work, it is recommended to apply the MIRT-2 method to analyze the correlation in other batteries of statistical tests, such as the batteries analyzed in [18,19], with the MIRT-1 method. Some of the batteries that can be studied are ENT, FIPS 140-2, DieHarder, TufTests, and TestU01. On the other hand, it is proposed to continue reducing the complexity of the MIRT-2 method by reducing the value of n. Although it is important to note that reducing n may increase the mean square error of the M I ^ M L estimator used in MIRT-2, this causes a decrease in the method’s effectiveness. The challenge is to reduce n and complexity without losing effectiveness.

Author Contributions

Conceptualization, J.A.K.-A. and C.M.L.-P.; methodology, C.M.L.-P.; formal analysis, R.S.-L. and J.A.K.-A.; investigation, J.A.K.-A., R.S.-L. and G.S.-G.; writing—original draft, J.A.K.-A.; R.S.-L. and G.S.-G.; writing—review and editing, O.R. and G.S.-G.; project administration, O.R. and G.S.-G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Skliar, O.; Monge, R.E.; Medina, V.; Gapper, S.; Oviedo, G. A Hybrid Random Number Generator(HRNG). Rev. De Matemática Teoría Apl. 2011, 18, 265. [Google Scholar] [CrossRef]
  2. Turan, M.S.; Doğanaksoy, A.; Boztaş, S. On Independence and Sensitivity of Statistical Randomness Tests. In Proceedings of the Sequences and Their Applications-SETA 2008, Lexington, KY, USA, 14–18 September 2008; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5203, pp. 18–29. [Google Scholar]
  3. Koçak, O. A Unified Evaluation of Statistical Randomness Tests and Experimental Analysis of Their Relations. Ph.D. Thesis, Middle East Technical University, Üniversiteler Mahallesi, Dumlupınar Bulvarı No:1 06800, Ankara, Turkey, 2016. [Google Scholar]
  4. Garcia, F.D.; de Koning Gans, G.; Muijrers, R.; van Rossum, P.; Verdult, R.; Schreur, R.W.; Jacobs, B. Dismantling MIFARE Classic. In Proceedings of the Computer Security—ESORICS 2008, Torremolinos, Spain, 6–8 October 2008; Lecture Notes in Computer Science. Jajodia, S., Lopez, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 97–114. [Google Scholar] [CrossRef]
  5. Kasper, T.; Silbermann, M.; Paar, C. All You Can Eat or Breaking a Real-World Contactless Payment System. In Proceedings of the International Conference on Financial Cryptography and Data Security, Tenerife, Spain, 25–28 January 2010; Sion, R., Ed.; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6052, pp. 343–350. [Google Scholar] [CrossRef]
  6. Bernstein, D.J.; Chang, Y.A.; Cheng, C.M.; Chou, L.P.; Heninger, N.; Lange, T.; van Someren, N. Factoring RSA Keys from Certified Smart Cards: Coppersmith in the Wild. In Proceedings of the Advances in Cryptology—ASIACRYPT, Bengaluru, India, 1–5 December 2013; Lecture Notes in Computer Science. Sako, K., Sarkar, P., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 341–360. [Google Scholar] [CrossRef]
  7. Bundesamt für Sicherheit in der Informationstechnik. Certification Report BSI-DSZ-CC-0212-2004 for Renesas AE45C1 (HD65145C1) Smartcard Integrated Circuit Version 01; Technical Report; Federal Office for Information Security: Bonn, Germany, 2004. [Google Scholar]
  8. Dunn, W.L.; Shultis, J.K. Pseudorandom Number Generators. In Exploring Monte Carlo Methods; Elsevier: Amsterdam, The Netherlands, 2012; pp. 47–68. [Google Scholar]
  9. Rukhin, A. Statistical Testing of Randomness: New and Old Procedures. In Randomness through Computation; World Scientific: Singapore, 2011. [Google Scholar]
  10. Hernandez-Castro, J.; Barrero, D.F. Evolutionary Generation and Degeneration of Randomness to Assess the Indepedence of the Ent Test Battery. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), Donostia, Spain, 5–8 June 2017; pp. 1420–1427. [Google Scholar] [CrossRef]
  11. Doğanaksoy, A.; Ege, B.; Muş, K. Extended Results for Independence and Sensitivity of NIST Randomness Tests. In Proceedings of the Information Security and Cryptography Conference, Istanbul, Turkey, 25–27 December 2008. [Google Scholar]
  12. Doğanaksoy, A.; Sulak, F.; Uğuz, M.; Şeker, O.; Akcengiz, Z. Mutual Correlation of NIST Statistical Randomness Tests and Comparison of Their Sensitivities on Transformed Sequences. Turk. J. Elec. Eng. Comp. Sci. 2017, 25, 655–665. [Google Scholar] [CrossRef]
  13. Hurley-Smith, D.; Hernandez-Castro, J. Great Expectations: A Critique of Current Approaches to Random Number Generation Testing & Certification. In Proceedings of the Security Standardisation Research, Darmstadt, Germany, 26–27 November 2018; Lecture Notes in Computer Science. Cremers, C., Lehmann, A., Eds.; Springer: Cham, Switzerland, 2018; Volume 11322, pp. 143–163. [Google Scholar] [CrossRef]
  14. Burciu, P.; Simion, E. A systematic approach of NIST statistical tests dependencies. J. Electr. Eng. Electron. Control. Comput. Sci. 2019, 5, 1–6. [Google Scholar]
  15. Sulak, F.; Uğuz, M.; Koçak, O.; Doğanaksoy, A. On the Independence of Statistical Randomness Tests Included in the NIST Test Suite. Turk. J. Elec. Eng. Comp. Sci. 2017, 25, 3673–3683. [Google Scholar] [CrossRef]
  16. Fan, L.; Chen, H.; Gao, S. A General Method to Evaluate the Correlation of Randomness Tests. In Proceedings of the International Workshop on Information Security Applications, Jeju Island, Republic of Korea, 19–21 August 2013; pp. 52–62. [Google Scholar]
  17. Karell-Albo, J.A.; Legón-Pérez, C.M.; Madarro-Capó, E.J.; Rojas, O.; Sosa-Gómez, G. Measuring independence between statistical randomness tests by mutual information. Entropy 2020, 22, 741. [Google Scholar] [CrossRef] [PubMed]
  18. Luengo, E.A.; Cerna, M.B.L.; Villalba, L.J.G.; Hernandez-Castro, J. A New Approach to Analyze the Independence of Statistical Tests of Randomness. Appl. Math. Comput. 2022, 426, 127116. [Google Scholar] [CrossRef]
  19. Luengo, E.A.; Cerna, M.B.L.; Villalba, L.J.G.; Hernandez-Castro, J.; Hurley-Smith, D. Critical Analysis of Hypothesis Tests in Federal Information Processing Standard (140-2). Entropy Int. Interdiscip. J. Entropy Inf. Stud. 2022, 24, 613. [Google Scholar]
  20. Cerna, M.B.L. Nuevas Técnicas Computacionales Para La Estimación de La Independencia de Los Tests de Aleatoriedad. 2021. Available online: https://docta.ucm.es/entities/publication/9c972153-b581-456f-b2a3-3bb7c3c256c9 (accessed on 1 October 2023).
  21. Thomas, J.A.; Cover, T. Elements of Information Theory; John Wiley & Sons, Inc.: New York, NY, USA, 1991; Volume 6, pp. 187–202. [Google Scholar]
  22. Darbellay, G.A. An Estimator of the Mutual Information Based on a Criterion for Conditional Independence. Comput. Stat. Data Anal. 1999, 32, 1–17. [Google Scholar] [CrossRef]
  23. Joe, H. Relative Entropy Measures of Multivariate Dependence. J. Am. Stat. Assoc. 1989, 84, 157–164. [Google Scholar] [CrossRef]
  24. Renyi, A. On the Foundations of Information Theory. Rev. Int. Stat. Inst. 1965, 33, 1–14. [Google Scholar] [CrossRef]
  25. Michalowicz, J.V.; Nichols, J.M.; Bucholtz, F. Handbook of Differential Entropy; CRC: Boca Raton, FL, USA, 2013. [Google Scholar]
  26. Fraser, A.M.; Swinney, H.L. Independent Coordinates for Strange Attractors from Mutual Information. Phys. Rev. A 1986, 33, 1134–1140. [Google Scholar] [CrossRef] [PubMed]
  27. Darbellay, G.; Vajda, I. Estimation of the Information by an Adaptive Partitioning of the Observation Space. IEEE Trans. Inf. Theory 1999, 45, 1315–1321. [Google Scholar] [CrossRef]
  28. Silverman, B.W. Density Estimation for Statistics and Data Analysis; CRC Press: Boca Raton, FL, USA, 1986; Volume 26. [Google Scholar]
  29. Moon, Y.I.; Rajagopalan, B.; Lall, U. Estimation of Mutual Information Using Kernel Density Estimators. Phys. Rev. E 1995, 52, 2318–2321. [Google Scholar] [CrossRef] [PubMed]
  30. Diks, C.; Manzan, S. Tests for Serial Independence and Linearity Based on Correlation Integrals. Stud. Nonlinear Dyn. Econom. 2002, 6. [Google Scholar] [CrossRef]
  31. Paninski, L. Estimation of Entropy and Mutual Information. Neural Comput. 2003, 15, 1191–1253. [Google Scholar] [CrossRef]
  32. Kraskov, A.; Stögbauer, H.; Grassberger, P. Estimating Mutual Information. Phys. Rev. E Stat. Physics Plasmas Fluids Relat. Interdiscip. Top. 2004, 69, 066138. [Google Scholar] [CrossRef]
  33. Daub, C.O.; Steuer, R.; Selbig, J.; Kloska, S. Estimating Mutual Information Using B-spline Functions—An Improved Similarity Measure for Analysing Gene Expression Data. BMC Bioinform. 2004, 5, 118. [Google Scholar] [CrossRef]
  34. Blinnikov, S.; Moessner, R. Expansions for Nearly Gaussian Distributions. Astron. Astrophys. Suppl. Ser. 1998, 130, 193–205. [Google Scholar] [CrossRef]
  35. Yavuz, Z.K.; Aydin, N.; Altay, G. Comprehensive review of association estimators for the inference of gene networks. Turk. J. Electr. Eng. Comput. Sci. 2016, 24, 695–718. [Google Scholar]
  36. Contreras Rodríguez, L.; Madarro-Capó, E.J.; Legón-Pérez, C.M.; Rojas, O.; Sosa-Gómez, G.S.G. Selecting an Effective Entropy Estimator for Short Sequences of Bits and Bytes with Maximum Entropy. Entropy 2021, 23, 561. [Google Scholar] [CrossRef]
  37. Papana, A.; Kugiumtzis, D. Evaluation of Mutual Information Estimators for Time Series. Int. J. Bifurc. Chaos 2009, 19, 4197–4215. [Google Scholar] [CrossRef]
  38. Sulewski, P. Equal-Bin-Width Histogram versus Equal-Bin-Count Histogram. J. Appl. Stat. 2021, 48, 2092–2111. [Google Scholar] [CrossRef] [PubMed]
  39. Sturges, H.A. The Choice of a Class Interval. J. Am. Stat. Assoc. 1926, 21, 65–66. [Google Scholar] [CrossRef]
  40. Cochran, W.G. Some Methods for Strengthening the Common χ2 Tests. Biom. J. Int. Biom. Soc. 1954, 10, 417–451. [Google Scholar]
  41. Lane, D. Online Statistics Education: A Multimedia Course of Study. Available online: https://onlinestatbook.com/ (accessed on 13 May 2023).
  42. Cencov, N.N. Estimation of an Unknown Distribution Density from Observations. Soviet Math. 1962, 3, 1559–1566. [Google Scholar]
  43. Bendat, J.S.; Piersol, A.G. Measurement and Analysis of Random Data, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 1966. [Google Scholar]
  44. Larson, H.J. Statistics: An Introduction; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 1975. [Google Scholar]
  45. Velleman, P. Interactive Computing for Exploratory Data Analysis i: Display Algorithms. Proc. Stat. Comput. Sect. 1976, 142–147. [Google Scholar]
  46. Doane, D.P. Aesthetic Frequency Classifications. Am. Stat. 1976, 30, 181–183. [Google Scholar]
  47. Mosteller, F.; Tukey, J.W. Addison-Wesley Series in Behavioral Science: Quantitative Methods, Reading, Mass; Addison-Wesley: Boston, MA, USA, 1977. [Google Scholar]
  48. Terrell, G.R.; Scott, D.W. Oversmoothed Nonparametric Density Estimates. J. Am. Stat. Assoc. 1985, 80, 209–214. [Google Scholar] [CrossRef]
  49. Ishikawa, K. Guide to Quality Control, UNIPUB/Kraus International; White Plains: New York, NY, USA, 1986. [Google Scholar]
  50. Boulle, M. Optimal Bin Number for Equal Frequency Discretizations in Supervized Learning. Intell. Data Anal. 2005, 9, 175–188. [Google Scholar] [CrossRef]
  51. Zhang, Z. Statistical Implications of Turing’s Formula; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2016. [Google Scholar] [CrossRef]
  52. Wilks, S.S. The Large-Sample Distribution of the Likelihood Ratio for Testing Composite Hypotheses. Ann. Math. Stat. 1938, 9, 60–62. [Google Scholar] [CrossRef]
  53. Menezes, A.J.; van Oorschot, P.C.; Vanstone, S.A. Handbook of Applied Cryptography; CRC Press: Boca Raton, FL, USA, 1997. [Google Scholar]
  54. Madarro-Capó, E.J.; Legón-Pérez, C.M.; Rojas, O.; Sosa-Gómez, G. Information theory based evaluation of the RC4 stream cipher outputs. Entropy 2021, 23, 896. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison of mutual information estimators for a pair of independent tests [17].
Figure 1. Comparison of mutual information estimators for a pair of independent tests [17].
Entropy 25 01545 g001
Figure 2. Comparison of mutual information estimators for a pair of correlated tests [17].
Figure 2. Comparison of mutual information estimators for a pair of correlated tests [17].
Entropy 25 01545 g002
Figure 3. Sampling distribution of the 1000 observations of Z M L obtained from pairs ( X i , Y i ) .
Figure 3. Sampling distribution of the 1000 observations of Z M L obtained from pairs ( X i , Y i ) .
Entropy 25 01545 g003
Figure 4. Quantile plot of the 1000 observations of Z M L obtained from pairs ( X i , Y i ) .
Figure 4. Quantile plot of the 1000 observations of Z M L obtained from pairs ( X i , Y i ) .
Entropy 25 01545 g004
Figure 5. Distribution of the 136 observations of Z M L for values between 0 and 1000.
Figure 5. Distribution of the 136 observations of Z M L for values between 0 and 1000.
Entropy 25 01545 g005
Figure 6. Distribution of Z M L observations for values between 0 and 6.
Figure 6. Distribution of Z M L observations for values between 0 and 6.
Entropy 25 01545 g006
Figure 7. Mutual information matrix with the values of Z M L between the pairs of statistical tests for randomness.
Figure 7. Mutual information matrix with the values of Z M L between the pairs of statistical tests for randomness.
Entropy 25 01545 g007
Figure 8. Representation of the execution time of the methods for ten experiments.
Figure 8. Representation of the execution time of the methods for ten experiments.
Entropy 25 01545 g008
Figure 9. Distribution of the differences in the estimators M I ^ M L and M I ^ S H .
Figure 9. Distribution of the differences in the estimators M I ^ M L and M I ^ S H .
Entropy 25 01545 g009
Figure 10. Bland–Altman plot for comparison of the estimators M I ^ M L and M I ^ S H .
Figure 10. Bland–Altman plot for comparison of the estimators M I ^ M L and M I ^ S H .
Entropy 25 01545 g010
Figure 11. Matrix to illustrate the stability of the MI estimator M I ^ M L .
Figure 11. Matrix to illustrate the stability of the MI estimator M I ^ M L .
Entropy 25 01545 g011
Table 1. Rules for selecting the number k of intervals for the discretization.
Table 1. Rules for selecting the number k of intervals for the discretization.
Rulesk
Sturges [39] 1 + log 2 ( n )
Cochran [40] n / 5
Rice [41] 2 n 3
Cencov [42] n 3
Bendat–Piersol [43] 1.87 ( n 1 ) 0.4
Larson [44] 1 + 2.2 log 2 ( n )
Velleman [45] 2 n if n 100
10 log 10 ( n ) if n > 100
Doane [46] 1 + log 2 ( n ) + log 2 1 + b 1 σ b 1
Mosteller–Tukey [47] n
Terrell–Scott [48] 2 n 3
Ishikawa [49] 6 + n / 50
Table 2. Parameters used for the MIRT-1 and MIRT-2 methods.
Table 2. Parameters used for the MIRT-1 and MIRT-2 methods.
MethodMIRT-1MIRT-2
Distribution of the
p-values
U(0,1)U(0,1)
MI estimatorshrinkage (SH)maximum-likelihood (ML)
Discretizationk = 10k = 10
Selection of
significant values
Comparison with the
critical value obtained
by the permutation test.
Comparison with
the critical value
of the normal distribution
for the α prefix.
Table 3. Tests for normality for the values of Z M L .
Table 3. Tests for normality for the values of Z M L .
Normality Test p-Value
Anderson–Darling A = 0.47518 0.2395
Kolmogorov–Smirnov D = 0.019679 0.8334
Table 4. Comparison of the correlations detected by the MIRT-1 and MIRT-2 methods.
Table 4. Comparison of the correlations detected by the MIRT-1 and MIRT-2 methods.
MIRT-1MIRT-2
App. Ent.≈Serial 1≈Serial 1
CUSUM (f)≈CUSUM (b)
≈Frequency
≈CUSUM (b)
≈Frequency
CUSUM (b)≈Frequency
≈Non-Overlapping
≈Frequency
Long. Run≈Overlapping≈Overlapping
Random Ex.≈Random Ex. Variant≈Random Ex. Variant
Serial 1≈Serial 2≈Serial 2
Table 5. Execution time in seconds of the methods for ten experiments.
Table 5. Execution time in seconds of the methods for ten experiments.
ExecutionMIRT-1MIRT-2Quotient
16249.360.3318,937.45
26230.960.2921,486.07
36250.030.320,833.43
46238.480.3120,124.13
56240.170.4115,219.93
66234.230.3120,110.42
76236.860.2921,506.41
86244.910.3020,816.37
96241.020.3020,803.40
106246.430.4214,872.45
μ 6241.2450.32619,144.92
σ 40.7130.00217,415.26
Table 6. Bland–Altman test measurements.
Table 6. Bland–Altman test measurements.
Bias0.001768216
The standard deviation of bias0.0002812867
Upper LOA0.002319538
Lower LOA0.001216894
Mean of differences/means202.4483
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Karell-Albo, J.A.; Legón-Pérez, C.M.; Socorro-Llanes, R.; Rojas, O.; Sosa-Gómez, G. Complexity Reduction in Analyzing Independence between Statistical Randomness Tests Using Mutual Information. Entropy 2023, 25, 1545. https://doi.org/10.3390/e25111545

AMA Style

Karell-Albo JA, Legón-Pérez CM, Socorro-Llanes R, Rojas O, Sosa-Gómez G. Complexity Reduction in Analyzing Independence between Statistical Randomness Tests Using Mutual Information. Entropy. 2023; 25(11):1545. https://doi.org/10.3390/e25111545

Chicago/Turabian Style

Karell-Albo, Jorge Augusto, Carlos Miguel Legón-Pérez, Raisa Socorro-Llanes, Omar Rojas, and Guillermo Sosa-Gómez. 2023. "Complexity Reduction in Analyzing Independence between Statistical Randomness Tests Using Mutual Information" Entropy 25, no. 11: 1545. https://doi.org/10.3390/e25111545

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop