Next Article in Journal
Demystifying the Two-Armed Futurity Bandit’s Unfairness and Apparent Fairness
Previous Article in Journal
A Classification of Compact Cohomogeneity One Locally Conformal Kähler Manifolds
Previous Article in Special Issue
Estimation in Semi-Varying Coefficient Heteroscedastic Instrumental Variable Models with Missing Responses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Testing Multivariate Normality Based on Beta-Representative Points

1
Department of Statistics and Data Science, BNU-HKBU United International College, Zhuhai 519087, China
2
Guangdong Provincial Key Laboratory of Interdisciplinary Research and Application for Data Science, BNU-HKBU United International College, Zhuhai 519087, China
3
Department of Medical Statistics, University Medical Center Göettingen, 37075 Göettingen, Germany
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(11), 1711; https://doi.org/10.3390/math12111711
Submission received: 30 March 2024 / Revised: 21 May 2024 / Accepted: 27 May 2024 / Published: 30 May 2024
(This article belongs to the Special Issue Computational Statistics and Data Analysis, 2nd Edition)

Abstract

:
Testing multivariate normality in high-dimensional data analysis has been a long-lasting topic in the area of goodness of fit. Numerous methods for this purpose can be found in the literature. Reviews on different methods given by influential researchers show that new methods keep emerging in the literature from different perspectives. The theory of statistical representative points provides a new perspective to construct tests for multivariate normality. To avoid the difficulty and huge computational load in finding the statistical representative points from a high-dimensional probability distribution, we develop an approach to constructing a test for high-dimensional normal distribution based on the representative points of the simple univariate beta distribution. The representative-points-based approach is extended to the the case that the sample size may be smaller than the dimension. A Monte Carlo study shows that the new test is able to control type I error rates fairly well for both large and small sample sizes when faced with a high dimension. The power of the new test against some non-normal distributions is generally or substantially improved for a set of selected alternative distributions. A real-data example is given for a simple application illustration.

1. Introduction

Methodologies for testing multivariate normality (MVN for short) have been studied for more than half a century. Ebner and Henze (2020) [1] gave the most up-to-date review on tests for MVN with the emphasis on L 2 distance. There are rich resources on methodologies for testing MVN. These methodologies were reviewed and commented on in different periods with different emphases; for example, Mardia (1980) focused on MVN tests constructed from the sample Mahalanobis distances (M distance for short, [2]); Horswell and Looney (1992) [3] focused on power comparison among the MVN tests based on measures of multivariate skewness and kurtosis; Remeu and Ozturk (1993) [4] carried out a comprehensive power comparison among various types of tests for MVN and gave a general recommendation; Henze (2002) [2] focused on a critical review on invariant MVN tests; and Mecklin and Mundfrom (2004) [5] gave a review on general MVN tests. These review articles show the fact that tests for MVN can be constructed from various angles, and each angle may provide a unique way to identify a possible source of departure from MVN. Therefore, new methods for testing MVN keep emerging, and statisticians never stop making efforts to develop new MVN tests. The theory on statistical representative points (simply called RP) or principal points ([6,7]) provides a new angle to develop goodness-of-fit tests that includes MVN tests as a special case. Based on our effort to bridge the RP theory and MVN tests ([8,9]), we want to develop another RP-based MVN test using the simple univariate beta distribution. The new RP-based MVN test is easy to carry out by using the publicly accessible website https://www.acsu.buffalo.edu/cxma/UIC/Representative.htm (accessed on 26 May 2024) to obtain the RPs from any beta distribution. It is applicable for both large and small sample sizes after some dimension reduction.
A test for MVN is usually expected to have the property of affine invariance because the multivariate normal distribution family is affine invariant. It is generally known that an affine-invariant test for MVN keeps its null distribution unchanged under the affine transformation of an observed sample ([2]). As a result, the null distribution of an affine invariant test for MVN does not depend on the unknown mean μ and covariance matrix Σ in the multivariate normal distribution N d ( μ , Σ ) . A good example of an affine invariant statistic is the sample Mahalanobis distance. We will focus on the sample M-distance approach to constructing MVN tests in this paper.
Let x 1 , , x n be a set of i.i.d. (independently identically distributed) samples from a d-dimensional continuous distribution. We want to test the hypothesis
H 0 : { x 1 , , x n } c o m e s   f r o m   s o m e   d - d i m e n s i o n a l   n o r m a l   d i s t r i b u t i o n N d ( μ , Σ )
against the general alternative H 1 that implies that the null hypothesis is not true.
A statistic T n x 1 , , x n is said to be affine invariant if it satisfies
T n A x 1 + b , , A x n + b = T n x 1 , , x n
for any b R d (the d-dimensional Euclidean space) and non-singular matrix A R d × d . Denote the sample mean and the sample covariance matrix by
x ¯ = 1 n j = 1 n x i and S n = 1 n j = 1 n x j x ¯ x j x ¯ ,
respectively. The sample M distance between an observation x i and the sample mean x ¯ is defined by
r i 2 = x i x ¯ S n 1 x i x ¯ , i = 1 , , n > d .
It is easy to verify that { r i 2 : i = 1 , , n } are affine invariant under the linear transformation (2).
Many existing tests for MVN are related to the idea of the sample M distance. An early comprehensive Monte Carlo comparison among different tests for MVN was given by Romeu and Ozturk [4]. Their results show that the Mardia’s [10,11] multivariate skewness and kurtosis, which are also based on the sample M distance, are generally recommended because of their competitive power performance against a wide range of alternative distributions. The sample M distance can be considered a dimension reduction approach to characterizing high-dimensional data. The relationship between any two high-dimensional observations is measured by their M distance. For example, Small [12] applied the sample M distance between any high-dimensional observation and the sample mean, which has approximately chi-square distribution under MVN, to the construction of a chi-square plotting method for detecting non-MVN; and Ahn [13] applied the so-called Jack-Knife sample M distance, which has approximately F-distribution under MVN, to the construction of an F-plotting method for detecting non-MVN. In this paper, we will develop a beta-distributed sample M-distance approach to testing MVN. We will employ the statistical representative points of the beta distribution to construct a chi-square-type test for MVN. Section 2 will introduce the beta sample M distance and its distribution under MVN. The test for MVN is transformed into a necessary test for a beta distribution by the traditional chi-square approach. A simple Monte study is given in Section 3 to illustrate the performance of the RP chi-square statistics for both relatively large and small sample sizes. Section 3 also presents a simple application of the RP chi-square approach to a real-data example. Some concluding remarks are given in the last section.

2. The MVN Test Based on Beta-Representative Points

2.1. The Beta M Distance

Suppose that we want to test hypothesis (1) based on an i.i.d. d-dimensional sample x 1 , , x n . The following theorem provides the basis for constructing the RP chi-square test for the hypothesis (1) with a large sample size ( n > d , d = data dimension).
Theorem 1. 
Under hypothesis (1), we have the following two conclusions:
1. 
The adjusted M distance has an exact beta distribution up to a constant:
d i 2 = n ( n 1 ) 2 r i 2 β d 2 , n 1 d 2 ;
2. 
d i 2 : i = 1 , , n are asymptotically independent.
Proof of Theorem 1.
The exact beta distribution β d 2 , n 1 d 2 for each d i 2 in Equation (5) can be derived from Wilks [14] (p. 562). The asymptotic independence of d i 2 : i = 1 , , n holds as a result of the fact that x ¯ μ and S n Σ ( n ) almost surely. Therefore, for a large sample size n , d i 2 is approximately a function of x i , μ and Σ . The independence of the sample x 1 , , x n results in the independence of d i 2 : i = 1 , , n . This completes the proof. □

2.2. The RP Chi-Square Test with a Large Sample Size

Instead of testing hypothesis (1) directly, we can turn to test hypothesis
H 0 : d i 2 : i = 1 , , n   i n   E q u a t i o n   ( 5 )   i s   a   s a m p l e   f r o m   β d 2 , n 1 d 2
against the alternative that H 0 is not true. Equation (6) is a general non-normal goodness-of-fit problem. It is generally carried out by the classical Pearson chi-square test by using the so-called equiprobable classification intervals for computing the Pearson statistic. The idea of equiprobable classification may not be the best option for non-uniform distributions. To improve the performance of the classical chi-square test, the idea of representative points (Fang and He [6]) (or principal points, Flurry [7]) can be employed to determine the classification cells for computing the chi-square statistic.
The beta-representative points are a set of points 0 < B 1 < < B m < 1 (for a selected number of points m) that minimize the quadratic loss function:
ϕ x 1 , , x m = 0 1 min 1 i m x i x 2 f b x ; d 2 , n 1 d 2 d x
where f b x ; d 2 , n 1 d 2 stands for the density function of the beta distribution with parameters d 2 , n 1 d 2 ,
ϕ B 1 , , B m = min 1 i m ϕ x 1 , , x m : 0 < x 1 < < x m < 1 .
The RPs B 1 , , B m for the general beta distribution can be obtained from running the beta distribution in the website: https://www.acsu.buffalo.edu/cxma/UIC/Representative.htm (accessed on 26 May 2024).
Define the following intervals
I 1 = 0 , B 1 + B 2 2 , I 2 = B 1 + B 2 2 , B 2 + B 3 2 , , I m 1 = B m 2 + B m 1 2 , B m 1 + B m 2 , I m = B m 1 + B m 2 , 1
and the probabilities
p i = I i f b x ; d 2 , n 1 d 2 d x , i = 1 , , m .
According to Fang and He [6], p 1 , , p m can be considered a set of “representative probabilities” for the beta distribution β d 2 , n 1 d 2 .
Based on Theorem 1, a test for hypothesis (1) can be approximately (under large sample size n) transferred to a test for hypothesis (6). The χ 2 -statistic for testing hypothesis (6) is computed by:
χ R 2 = i = 1 m n i n p i 2 n p i ,
where n i is the frequency of the transformed approximately i.i.d. sample points d i 2 : i = 1 , , n computed by Equation (5) that are located in the interval I i in Equation (8). It is known that χ R 2 χ 2 ( m 1 ) ( n ) in distribution. The p value for testing hypothesis (6) is computed by
P χ R 2 , v = K χ R 2 z v 2 1 exp z 2 d z , w i t h   v = m 1 , K = 2 v 2 Γ v 2 1 .

2.3. The RP Chi-Square Test with High Dimension and a Small Sample Size

The RP chi-square test for hypothesis Equation (6) in Section 2.2 requires the sample size n to be greater than the dimension d ( n > d ). When facing high dimension with a small sample size ( n d ), the RP chi-square test for hypothesis Equation (6) is no longer applicable. The dimension reduction is based on the idea of principal component analysis (PCA) by Liang et al. [15]. Suppose that we have an i.i.d. sample { x 1 , , x n } and want to test hypothesis (1). Assuming hypothesis (1) is true, we carry out the transformation in [15]:
y i = x 1 + + x i i x i + 1 i ( i + 1 ) , i = 1 , , n 1 ,
{ y 1 , , y n 1 } is an i.i.d. sample from N d ( 0 , Σ ) . Testing hypothesis (1) can be transferred to testing hypothesis
H 0 : { y 1 , , y n 1 }   is   a   sample   from N d ( 0 , Σ )
against general non-normal alternatives. Let
X = ( x 1 , , x n ) : n × d , Y = ( y 1 , , y n 1 ) : ( n 1 ) × d .
Define the eigenvalue-eigenvector problem
1 n 1 Y Y V = V Λ ,
where V = ( v 1 , , v d ) ( d × d ) consists of the eigenvectors, and Λ = diag ( λ 1 , , λ p ) is a diagonal matrix consisting of the eigenvalues. Let
Z q = Y V q , ( n 1 ) × q , V q = ( v 1 , , v q ) , ( p × q )
where q = 1 , , min ( n 1 , d ) 1 is called the projection dimension. Z q consists of the projected data on a lower-dimensional sample space with dimension q < min ( n 1 , d ) . Define
γ j 2 = ( z j z ¯ ) S z 1 ( z j z ¯ ) , e j 2 = n 1 ( n 2 ) 2 γ j 2 , z ¯ = 1 n 1 j = 1 n 1 z j , S z = 1 n 1 j = 1 n 1 ( z j z ¯ ) ( z j z ¯ ) .
Following Theorem 1, we have the following.
Theorem 2. 
Under hypothesis (1) and the Equations (11)–(16), the following two assertions are true:
1. 
The adjusted M distance has an exact beta distribution up to a constant:
e j 2 = n 1 ( n 2 ) 2 γ j 2 β q 2 , n 2 q 2 , q = 1 , , r = min ( n 1 , d ) 1 ;
2. 
e j 2 : j = 1 , , n 1 are asymptotically independent.
Proof of Theorem 2. 
According to Corollaries 2.1–2.2 in [15] and the affine invariance of the γ j 2 ( j = 1 , , n 1 ) in (16), the null distribution of γ j 2 remains unchanged when transforming the original observation matrix X into the observation matrix Y through Equations (11)–(13) and then projecting Y onto Z q through Equation (15) for any q = 1 , , min ( n 1 , d ) 1 . The proof can be completed by applying Theorem 1 with sample size n to the case here with sample size n 1 . □
Similar to the case of n > d in Section 2.2, instead of testing hypothesis (1) directly, we can turn to test hypothesis
H 0 : e j 2 : i = 1 , , n 1   in   Equation   ( 16 )   is   a   sample   from   β q 2 , n 2 q 2
against the alternative that H 0 is not true, where q is any given value from q = 1 , , min ( n 1 , d ) 1 . The χ R 2 statistic defined by Equation (10) (where n is replaced with n 1 in Equations (9) and (10)) can be applied to testing this hypothesis, and its asymptotic null distribution maintains unchanged. Theoretically, each of the projection dimensions q = 1 , , r = min ( n 1 , d ) 1 can be used to construct the χ 2 -test for (18). Based on the Monte Carlo study in [15], the choice of q in the range of [ r / 3 ] q [ r / 2 ] has better empirical power performance, where [ · ] stands for the integer part of a real number (e.g., [2.5] = 2, [3.2] = 2). We will choose q = [ r / 3 ] and q = [ r / 2 ] in the following Monte Carlo study for the case of n d .

3. A Monte Carlo Study and an Illustrative Example

In order to compare the χ R 2 -test (10) under the “representative probabilities” p 1 , , p m in Equation (9) with the traditional chi-squared test, we choose the equiprobable cells for computing the traditional chi-squared test. The chi-square statistic with equiprobable cells was recommended by Voinov et al. [16]. For a selected number of representative points m, define the interval endpoints:
a 1   satisfies   P χ 2 ( m 1 ) < a 1 = 1 m ; a 2   satisfies   P a 1 < χ 2 ( m 1 ) < a 2 = 1 m ; a m 1   satisfies   P a m 2 < χ 2 ( m 1 ) < a m 1 = 1 m ; a m   satisfies   P χ 2 ( m 1 ) > a m = 1 m .
Denote the traditional chi-squared test based on the interval endpoints given by Equation (11) by χ T 2 , which is also an approximate χ 2 ( m 1 ) .

3.1. Comparison between the Empirical Type I Error Rates

Because the chi-square test based on the transformed sample points d i 2 : i = 1 , , n given by Equation (5) are affine invariant under any non-singular linear transformation of the original i.i.d. sample x 1 , , x n , we only need to generate samples from a d-dimensional standard normal N d 0 , I d I d stands for the d × d identity matrix). The simulation results under 10,000 replications for each case are summarized in Table 1 for the significance level α = 0.05 . Simulation results for α = 0.01 and α = 0.10 are also available upon request. The results in Table 1 ( n > d ) and Table 2 ( n d ) demonstrate that the empirical type I error rates are feasibly well controlled near the given significance level α = 0.05 .

3.2. A Simple Power Comparison

In order for the power comparison to be representative, we select three symmetric distributions and three skewed distributions as follows, where the definitions for the multivariate t-distribution and the multivariate Cauchy distribution can be found in the work of Fang, Kotz, and Ng [17].
  • The multivariate t-distribution has a density function of the form
    f t ( x ) = C 1 1 + x 2 m d + m 2 , m > 0 ,
    which is symmetric about the origin f t ( x ) f t ( x ) , where “ · ” stands for the Euclidean norm of a vector. Let m = 5 .
  • The multivariate Cauchy distribution has a density function of the form:
    f c ( x ) = C 1 1 + x 2 m d + 1 2 ,
    which is symmetric about the origin, where C 1 is a normalizing constant depending on the dimension d.
  • The β -generalized normal distribution N d 0 , I d , 1 / 2 with β = 1 / 2 has a density function of the form (by Goodman and Kotz [18]):
    f x 1 , , x d = β d r d / β 2 d Γ d ( 1 / β ) · exp r i = 1 d x i β , x 1 , , x d R d ,
    which is symmetric about the origin, where r > 0 is a parameter. Let r = 1 / 2 and β = 1 in the simulation and denote it by β -g-normal.
  • The shifted i.i.d. χ 2 ( 1 ) has i.i.d. marginals, where each marginal has the same distribution as that of the random variable Y = X E ( X ) , where X χ 2 ( 1 ) , the univariate chi-square distribution with 1 degree of freedom and E ( X ) = 1 . This is a skewed distribution.
  • The distribution N ( 0 , 1 ) + χ 2 ( 2 ) consists of i.i.d. [ d / 2 ] normal N ( 0 , 1 ) marginals and d [ d / 2 ] i.i.d. χ 2 ( 2 ) marginals. This is a skewed distribution.
  • The shifted i.i.d. exp ( 1 ) has i.i.d. marginals, where each marginal has the same distribution as that of the random variable Y = X E ( X ) , where X exp ( 1 ) , the univariate exponential distribution. This is a skewed distribution.
For each of these alternative distributions, we choose the sample size n = 50 , 70 , , 400 . We plot the power values versus the sample size n for both statistics χ R 2 and χ T 2 to obtain a quick visual comparison. Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 demonstrate the comparisons between the two power curves for χ R 2 (the blue one) and χ T 2 (the red one) with dimensions ranging from d = 5 to d = 20 . The simulation is repeated 5000 times. It is observed that the RP chi-square test outperforms the traditional chi-square significantly for all selected symmetric alternative non-normal distributions (like the multivariate t, the Cauchy, and the β -generalized normal distribution) and asymmetric ones (like the chi-square and those with chi-square and exponential as the marginal distributions). Figure 10 and Figure 11 demonstrate the comparisons between the two projected chi-square tests χ R 2 (the blue one) and χ T 2 (the red one) under the multivariate t distribution with df = 5 (Figure 10) and the shifted χ 2 ( 2 ) 2 with i.i.d. marginals. Because both χ R 2 and χ T 2 for testing hypothesis (18) in Section 2 are affine invariant under the transformations (11)–(16) on the original x-sample in (11), we only need to take the zero mean and identity covariance matrix for the simulated samples. Figure 10 and Figure 11 show that the projected RP- χ R 2 substantially outperforms the traditional χ T 2 for both recommended projection dimensions q 1 = [ r / 3 ] and q 2 = [ r / 2 ] with r = min ( n 1 , d ) 1 under the significance level α = 0.05 . The power improvement of the projected RP- χ R 2 over the projected traditional χ T 2 is similar to the case demonstrated by Figure 10 and Figure 11 for other non-normal alternative distributions in Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9. We do not present those similar cases to Figure 10 and Figure 11 to save some space.
It is pointed out that the Pearson chi-square statistic has an asymptotic chi-square distribution under any classification cell intervals. Its approximation speed to the chi-square distribution is 1 / n under a given sample size n ([16]). We choose the sample size n ranging from 50 to 400 in our Monte Carlo study to maintain fairly large sample sizes. While one may choose a larger sample size than 400 or a smaller sample size than 50 in the simulation, we just want to illustrate if our beta-RP-based MVN test is able to control the type I error rate feasibly within some range of the sample size. The choice of equiprobable classification cell intervals is motivated by the empirical study in [16] that shows some good performance of this choice. It is feasible to believe that our beta-RP-based MVN test will also show significant power improvement compared with the traditional chi-square test under a set of arbitrarily chosen cell intervals based on the study in [16].

3.3. An Illustrative Example

Example 1 
(see Example 6.3 of Fang and Wang [19] (pp. 258–262)). The data arose from the problem of standardizing the size for men’s clothes in China in 1976, which involves the data of 12 measurements of the body ( cm ) :
X 1 : height from the waist up X 2 : arm length
X 3 : bust X 4 : neck
X 5 : shoulder length X 6 : width of the front part of chest
X 7 : width of the back part of chest X 8 : height
X 9 : height without head and neck X 10 : height from the waist down
X 11 : waist circumference X 12 : buttocks
A sample of size n = 100 can be found in the work of Fang, Yuan, and Bentler [20]. Fang and Wang [19] implemented the classical skewness and kurtosis statistics and number-theoretic methods to test the multinormality of some subsets of the 12 variables. At α = 0.05 level of significance, they concluded that (1) ( X 1 , X 3 , X 8 , X 10 , X 12 ) has a multivariate normal distribution; (2) ( X 1 , X 3 , X 8 , X 10 ) , ( X 1 , X 8 , X 10 , X 12 and ( X 3 , X 8 , X 10 , X 12 ) have a multivariate normal distribution; and (3) ( X 4 , X 5 , X 6 , X 11 ) and ( X 2 , X 4 , X 6 , X 11 ) have a non-normal distribution.
The p values under different m (the number of representation points) from the two chi-square tests, RP chi-square χ R 2 in Equation (10) and the traditional chi-square χ T 2 using equiprobable intervals defined by the endpoints given by Equation (19), are summarized in Table 3. The results in Table 3 show that both the RP chi-square χ R 2 and the traditional chi-square χ T 2 give results consistent with those given by [19] on the multivariate normality of the four sets of variables: ( X 1 , X 3 , X 8 , X 10 , X 12 ) , ( X 1 , X 3 , X 8 , X 10 ) , ( X 1 , X 8 , X 10 , X 12 ) , and ( X 3 , X 8 , X 10 , X 12 ) , but both χ R 2 and χ T 2 fail to reject the possible non-normality of the two sets of variables ( X 4 , X 5 , X 6 , X 11 ) and ( X 2 , X 4 , X 6 , X 11 ) , while Fang and Wang [19] reject the multivariate normality. This needs to be double checked by some other statistics for testing MVN.

4. Concluding Remarks

The theory of statistical representative points is a natural extension to the mean value of a probability distribution. A comprehensive study on the RP theory can be found in the work of Graf and Luschgy [21]. Application of the RP theory to testing multivariate normality was first proposed by Liang, He, and Yang [8] by using the univariate Student’s t-representative points, and by Wang et al. [9] using the univariate F-representative points. The study in this paper is parallel to those in [8] and of Wang et al. [9]. The results in these papers bridge the gap between the RP theory and testing goodness of fit. The Monte Carlo study in Section 3 shows some impressive power improvement from using the RP chi-square versus the traditional chi-square for both cases of relatively large and small sample sizes. A noticeable point is that the new RP-based test is still applicable in the case that the sample size may be smaller than the dimension, and it still performs very well. This can be regarded as a special credit compared with some existing tests for multivariate normality when faced with high dimension with a small sample size. Although it is not an easy task to prove whether the RP chi-square always improves the traditional chi-square in testing general goodness-of-fit problems, the results in this paper shed some hopeful light on applying the RP theory to the goodness-of-fit area. It is certain that the chi-square statistic is not the sole way to construct the RP-based test. Some other classical tests like those compared by Quesenberry and Miller [22] for general goodness-of-fit purposes can be also employed to test hypothesis (6). A complete comparison among these statistics for testing hypothesis (6) is beyond the scope of this paper. It is also too heavy to compare univariate tests and multivariate tests for high-dimensional normality. The beta RP-based test in this paper provides an additional way to connect the RP theory with goodness-of-fit techniques.

Author Contributions

Conceptualization and methodology, J.L. and Y.C.; simulation and real data analysis, L.X.; simulation double check, reference collection and editing of manuscript, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the UIC Teaching Enhancement Project UICR0400051-21CTL, and the Guangdong Provincial Key Laboratory of Interdisciplinary Research and Application for Data Science, BNU-HKBU United International College (UIC), project code 2022B1212010006. Longhao Xu’s work was carried out while he was studying at UIC.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ebner, B.; Henze, N. Tests for multivariate normality—A critical review with emphasis on weighted L2-statistics. Test 2020, 29, 845–892. [Google Scholar] [CrossRef]
  2. Henze, N. Invariant tests for multivariate normality: A critical review. Stat. Pap. 2002, 43, 467–506. [Google Scholar] [CrossRef]
  3. Horswell, R.L.; Looney, S.W. A comparison of tests for multivariate normality that are based on measures of multivariate skewness and kurtosis. J. Statist. Comput. Simul. 1992, 42, 21–38. [Google Scholar] [CrossRef]
  4. Remeu, J.L.; Ozturk, A. A comparative study of goodness-of-fit tests for multivariate normality. J. Multivar. Anal. 1993, 46, 309–334. [Google Scholar] [CrossRef]
  5. Mecklin, C.J.; Mundfrom, D.J. An appraisal and bibliography of tests for multivariate normality. Int. Stat. Rev. 2004, 72, 123–138. [Google Scholar] [CrossRef]
  6. Fang, K.T.; He, S.D. The Problem of Selecting a Given Number of Representative Points in a Normal Population and a Generalized Mills Ratio; Stanford Technical Report; Stanford Statistics Department: Stanford, CA, USA, 1982; No. 327. [Google Scholar]
  7. Flury, B.A. Principal points. Biometrika 1990, 77, 33–41. [Google Scholar] [CrossRef]
  8. Liang, J.; He, P.; Yang, J. Testing multivariate normality based on t-representative points. Axioms 2022, 11, 587. [Google Scholar] [CrossRef]
  9. Wang, S.; Liang, J.; Zhou, M.; Ye, H. Testing multivariate normality based on F-representative points. Mathematics 2022, 10, 4300. [Google Scholar] [CrossRef]
  10. Mardia, K.V. Measures of multivariate skewness and kurtosis with applications. Biometrika 1970, 57, 519–530. [Google Scholar] [CrossRef]
  11. Mardia, K.V. Tests of univariate and multivariate normality. In Handbook of Statistics; Krishnaiah, P.R., Ed.; North-Holland Publishing Company: Amsterdam, The Netherlands, 1980; pp. 279–320. [Google Scholar]
  12. Small, N.J.H. Plotting squared radii. Biometrika 1978, 65, 657–658. [Google Scholar] [CrossRef]
  13. Ahn, S.K. F-probability plot and its application to multivariate normality. Commun. Stat. Theory Methods 1992, 21, 997–1023. [Google Scholar] [CrossRef]
  14. Wilks, S.S. Mathematical Statistics; Wiley: New York, NY, USA, 1962. [Google Scholar]
  15. Liang, J.; Li, R.; Fang, H.; Fang, K.T. Testing multinormality based on low-dimensional projection. J. Stat. Plan. Inference 2000, 86, 129–141. [Google Scholar] [CrossRef]
  16. Voinov, V.; Pya, N.; Alloyarova, R. A comparative study of some modified chi-squared tests. Commun. Stat. Simul. Comput. 2009, 38, 355–367. [Google Scholar] [CrossRef]
  17. Fang, K.T.; Kotz, S.; Ng, K.W. Symmetric Multivariate and Related Distributions; Chapman and Hall: London, UK; New York, NY, USA, 1990. [Google Scholar]
  18. Goodman, I.R.; Kotz, S. Multivariate q-generalized normal distribution. J. Multivar. Stat. Anal. 1990, 3, 204–219. [Google Scholar] [CrossRef]
  19. Fang, K.T.; Wang, Y. Number-Theoretic Methods in Statistics; Chapman and Hall: London, UK, 1994. [Google Scholar]
  20. Fang, K.T.; Yuan, K.; Bentler, P.M. Applications of sets of points uniformly distributed on a sphere to testing multinormality and robust estimation. In Probability and Statistics; Jiang, Z.P., Yan, S.J., Cheng, P., Wu, R., Eds.; World Scientific: Singapore, 1992; pp. 56–73. [Google Scholar]
  21. Graf, S.; Luschgy, H. Foundations of Quantization for Probability Distributions; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  22. Quesenberry, C.P.; Miller, F.L., Jr. Power studies of some tests for uniformity. J. Stat. Comput. Simul. 1977, 5, 169–191. [Google Scholar] [CrossRef]
Figure 1. Dimension d = 5 .
Figure 1. Dimension d = 5 .
Mathematics 12 01711 g001
Figure 2. Dimension d = 5 .
Figure 2. Dimension d = 5 .
Mathematics 12 01711 g002
Figure 3. Dimension d = 5 .
Figure 3. Dimension d = 5 .
Mathematics 12 01711 g003
Figure 4. Dimension d = 10 .
Figure 4. Dimension d = 10 .
Mathematics 12 01711 g004
Figure 5. Dimension d = 10 .
Figure 5. Dimension d = 10 .
Mathematics 12 01711 g005
Figure 6. Dimension d = 10 .
Figure 6. Dimension d = 10 .
Mathematics 12 01711 g006
Figure 7. Dimension d = 20 .
Figure 7. Dimension d = 20 .
Mathematics 12 01711 g007
Figure 8. Dimension d = 20 .
Figure 8. Dimension d = 20 .
Mathematics 12 01711 g008
Figure 9. Dimension d = 20 .
Figure 9. Dimension d = 20 .
Mathematics 12 01711 g009
Figure 10. Power comparison between two projected χ 2 tests for multivariate normality ( d = 30 , r = min ( n 1 , d ) 1 ).
Figure 10. Power comparison between two projected χ 2 tests for multivariate normality ( d = 30 , r = min ( n 1 , d ) 1 ).
Mathematics 12 01711 g010
Figure 11. Power comparison between two projected χ 2 tests for multivariate normality ( d = 50 , r = min ( n 1 , d ) 1 ).
Figure 11. Power comparison between two projected χ 2 tests for multivariate normality ( d = 50 , r = min ( n 1 , d ) 1 ).
Mathematics 12 01711 g011
Table 1. Empirical type I error rates ( α = 0.05 , n > d ).
Table 1. Empirical type I error rates ( α = 0.05 , n > d ).
Sample Size nRP m χ 2 d = 3 d = 5 d = 10 d = 15 d = 20
n = 50 m = 10 χ P 2 0.03570.03190.03230.03220.0318
χ T 2 0.02980.03190.03090.03100.0307
m = 20 χ P 2 0.07340.05210.04690.04700.0420
χ T 2 0.03710.03350.03510.03240.0339
m = 30 χ R 2 0.06120.06330.06940.06080.0546
χ T 2 0.03220.03440.03380.03750.0336
n = 100 m = 10 χ R 2 0.03300.03290.03250.03360.0366
χ T 2 0.02910.02960.03460.03120.0323
m = 20 χ R 2 0.05060.04480.04260.04160.0404
χ T 2 0.03430.03370.03260.03400.0351
m = 30 χ R 2 0.06230.06900.05420.05150.0525
χ T 2 0.04000.04000.03950.03890.0350
n = 200 m = 10 χ R 2 0.03140.03420.03270.03170.0293
χ T 2 0.03120.02960.03050.03230.0320
m = 20 χ R 2 0.04140.04440.04010.03840.0407
χ T 2 0.03490.03450.03370.03590.0332
m = 30 χ R 2 0.06950.05480.05040.04700.0492
χ T 2 0.03900.03940.03620.03820.0394
n = 400 m = 10 χ R 2 0.03290.02770.03160.03210.0343
χ T 2 0.03250.03140.03330.03140.0302
m = 20 χ R 2 0.04180.04160.03670.03550.0371
χ T 2 0.03600.03550.03570.03520.0364
m = 30 χ R 2 0.05290.04650.04550.04520.0459
χ T 2 0.03720.03540.03890.03750.0381
Table 2. Empirical type I error rates ( α = 0.05 , n d , r = min ( n 1 , d ) 1 ) .
Table 2. Empirical type I error rates ( α = 0.05 , n d , r = min ( n 1 , d ) 1 ) .
( n , d ) = ( 30 , 30 ) ( n , d ) = ( 30 , 50 ) ( n , d ) = ( 50 , 50 )
RP χ 2 q = [ r / 3 ] q = [ r / 2 ] q = [ r / 3 ] q = [ r / 2 ] q = [ r / 3 ] q = [ r / 2 ]
m = 10 χ R 2 0.05440.05380.04720.05940.04120.0438
χ T 2 0.02940.03460.02500.03100.03500.0326
m = 15 χ R 2 0.05500.07060.05080.07020.05360.0566
χ T 2 0.03280.03700.03260.03800.03640.0372
m = 20 χ R 2 0.07420.07840.07060.08860.06100.0666
χ T 2 0.03620.03980.03700.03800.03880.0430
Table 3. p values from the two chi-squared tests for the body data.
Table 3. p values from the two chi-squared tests for the body data.
Subsets χ 2 m = 10 m = 20 m = 30
X 1 , X 3 , X 8 , X 10 , X 12 χ R 2 0.84980.88530.8027
χ T 2 0.86770.84870.5286
X 1 , X 3 , X 8 , X 10 χ R 2 0.91580.90160.8184
χ T 2 0.95580.73520.8518
X 1 , X 8 , X 10 , X 12 χ R 2 0.80340.81330.7007
χ T 2 0.92410.54930.4342
X 3 , X 8 , X 10 , X 12 χ R 2 0.50030.64280.8635
χ T 2 0.40120.54930.8518
X 4 , X 5 , X 6 , X 11 χ R 2 0.43550.14990.3865
χ T 2 0.49440.86780.6886
X 2 , X 4 , X 6 , X 11 χ R 2 0.67140.91410.7850
χ T 2 0.63710.73520.8733
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, Y.; Liang, J.; Xu, L.; Kang, J. Testing Multivariate Normality Based on Beta-Representative Points. Mathematics 2024, 12, 1711. https://doi.org/10.3390/math12111711

AMA Style

Cao Y, Liang J, Xu L, Kang J. Testing Multivariate Normality Based on Beta-Representative Points. Mathematics. 2024; 12(11):1711. https://doi.org/10.3390/math12111711

Chicago/Turabian Style

Cao, Yiwen, Jiajuan Liang, Longhao Xu, and Jiangrui Kang. 2024. "Testing Multivariate Normality Based on Beta-Representative Points" Mathematics 12, no. 11: 1711. https://doi.org/10.3390/math12111711

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop