**1. Introduction**

Let X1 = (*<sup>X</sup>*11, ..., *<sup>X</sup>*1*m*)*<sup>T</sup>*, ... ,X*<sup>k</sup>* = (*Xk*1, ..., *Xkm*)*<sup>T</sup>* be a random sample from *m*-dimensional population. The data set can be regarded as *k* vectors or points in *m*-dimensional space. Recently, there has been significant interest in a high-dimensional datasets when the dimension is large. In a high-dimensional setting, it is assumed that either (i) *m* tends to infinity and *k* is fixed, or (ii) both *m* and *k* tend to infinity. Case (i) is related to high-dimensional low sample size (HDLSS) data. One of the first results for HDLSS data appeared in Hall et al. [1]. It became the basis of research in mathematical statistics for the analysis of high-dimensional data, see, e.g., Fujikoshi et al. [2], which are an important part of the current data analysis fashionable area called *Big data*. Scientific areas where these settings have proven to be very useful include genetics and other types of cancer research, neuroscience, and also image and shape analysis. See a recent survey on HDLSS asymptotics and its applications in Aoshima et al. [3].

For examining the features of the data set, it is necessary to study the asymptotic behavior of three functions: the length X*i* of a *m*-dimensional observation vector, the distance X*i* −X*j* between any two independent observation vectors, and the angle ang(X*i*,X*j*) between these vectors at the

population mean. Assuming that X*i*'s are a sample from *N*(0, *Im*), it was shown in Hall et al. [1] that for HDLSS data the three geometric statistics satisfy the following relations:

$$\|\vec{\chi}\_i\| = \sqrt{m} + \mathcal{O}\_p(1), \quad i = 1, \ldots, k,\tag{1}$$

$$\|\vec{X}\_i - \vec{X}\_j\| = \sqrt{2m} + \mathcal{O}\_{\mathbb{P}}(1), \ i, j = 1, \dots k, \ i \neq j,\tag{2}$$

$$\text{diag}(\vec{\chi}\_i \vec{\chi}\_j) = \frac{1}{2}\pi + \mathcal{O}\_p(m^{-1/2}), \text{ i.j.}\\j = 1, \dots, k, \text{ i.e.} \tag{3}$$

where · is the Euclidean distance and <sup>O</sup>*p* denotes the stochastic order. These interesting results imply that the data converge to the vertices of a deterministic regular simplex. These properties were extended for non-normal sample under some assumptions (see Hall et al. [1] and Aoshima et al. [3]). In Kawaguchi et al. [4], the relations (1)–(3) were refined by constructing second order asymptotic expansions for distributions of all three basic statistics. The refinements of (1) and (2) were achieved by using the idea of Ulyanov et al. [5] who obtained the computable error bounds of order <sup>O</sup>(*m*<sup>−</sup><sup>1</sup>) for the chi-squared approximation of transformed chi-squared random variables with *m* degrees of freedom.

The aim of the present paper is to study approximation for the third statistic ang( X1, X2) under generalized assumption that *m* is a realization of a random variable, say *Nn*, which represents the sample dimension and is independent of X1 and X2. This problem is closely related to approximations of statistics constructed from the random size samples, in particular, to this kind of problem for the sample correlation coefficient *Rm*.

The use of samples with random sample sizes has been steadily growing over the years. For an overview of statistical inferences with a random number of observations and some applications, see Esquível et al. [6] and the references cited therein. Gnedenko [7] considered the asymptotic properties of the distributions of sample quantiles for samples of random size. In Nunes et al. [8] and Nunes et al. [9], unknown sample sizes are assumed in medical research for analysis of one and more than one-way fixed effects ANOVA models to avoid false rejections, obtained when using the classical fixed size *F*-tests. Esquível et al. [6] considered inference for the mean with known and unknown variance and inference for the variance in the normal model. Prediction intervals for the future observations for generalized order statistics and confidence intervals for quantiles based on samples of random sizes are studied in Barakat et al. [10] and Al-Mutairi and Raqab [11], respectively. They illustrated their results with real biometric data set, the duration of remission of leukemia patients treated by one drug. The present paper continues studies of the authors on non-asymptotic analysis of approximations for statistics based on random size samples. In Christoph et al. [12], second order expansions for the normalized random sample sizes are proved, see below Propositions 1 and 2. These results allow for proving second order asymptotic expansions of random sample mean in Christoph et al. [12] and random sample median in Christoph et al. [13]. See also Chapters 1 and 9 in Fujikoshi and Ulyanov [14].

The structure of the paper is the following. In Section 2, we describe the relation between ang( X1, X2) and *Rm*. We recall also previous approximation results proved for distributions of ang( X1, X2) and *Rm*. Section 3 is on general transfer theorems, which allow us to construct asymptotic expansions for distributions of randomly normalized statistics on the base of approximation results for non-randomly normalized statistics and for the random size of the underlying sample, see Theorems 1 and 2. Section 4 contains the auxiliary lemmas. Some of them have independent interest. For example, Lemma 3 on the upper bounds for the negative order moments of a random variable having negative binomial distribution. We formulate and discuss main results in Sections 5 and 6. In Theorems 3–8, we construct the second order Chebyshev–Edgeworth expansions for distributions of ang( X1, X2) and *Rm* in random setting. Depending on the type of normalization, we ge<sup>t</sup> three different limit distributions: Normal, Laplace, or Student's *t*-distributions. All proofs are given in the Appendix A.

#### **2. Sample Correlation Coefficient, Angle between Vectors and Their Normal Approximations**

We slightly simplify notation. Let X*m* = (*<sup>X</sup>*1, ..., *Xm*)*<sup>T</sup>* and Y*m* = (*<sup>Y</sup>*1, ...,*Ym*)*<sup>T</sup>* be two vectors from an *m*-dimensional normal distribution *N*(0, <sup>I</sup>*m*) with zero mean, identity covariance matrix I*m* and the sample correlation coefficient

$$R\_m = R\_m(\vec{\mathbf{X}}\_m, \vec{\mathbf{Y}}\_m) = \frac{\sum\_{k=1}^m X\_k \, \mathbf{Y}\_k}{\sqrt{\sum\_{k=1}^m X\_k^2 \, \sum\_{k=1}^m \, \mathbf{Y}\_k^2}} \cdot \tag{4}$$

Under the null hypothesis *H*0: X*m and* Y*m are uncorrelated*, the so-called null density *pRm* (*y*; *n*) of *Rm* is given in Johnson, Kotz and Balakrishnan [15], Chapter 32, Formula (32.7):

$$p\_{R\_m}(y;m) = \frac{\Gamma((m-1)/2)}{\sqrt{\pi}\,\Gamma((m-2)/2)} \left(1 - y^2\right)^{(m-4)/2} \mathbb{I}\_{(-1,1)}(y)$$

for *m* ≥ 3, where <sup>I</sup>*A*(.) denotes indicator function of a set *A*.


Consider now the standardized correlation coefficient

$$
\overline{R}\_m = \sqrt{m-c} \text{ } R\_m \tag{5}
$$

with some correcting real constant *c* < *m* having density

$$p\_{\overline{\mathbb{R}}\_m}(y;m,c) \quad = \frac{\Gamma((m-1)/2)}{\sqrt{m-c}\sqrt{\pi}\Gamma((m-2)/2)} \left(1 - \frac{y^2}{m-c}\right)^{(m-4)/2} \mathbb{I}\_{\{|r| < \sqrt{m-c}\}}(y),\tag{6}$$

which converges with *c* = O(1) as *m* → ∞ to the standard normal density

$$\varphi(y) = \frac{1}{\sqrt{2\pi}} e^{-y^2/2}, \quad y \in (-\infty, \infty)$$

and by Konishi [16], Section 4, Formula (4.1) as *m* → ∞:

$$F\_{\mathfrak{m}}^{\*}(\mathbf{x}, \mathfrak{c}) := P\left(\sqrt{m - \mathfrak{c}} \, R\_{\mathfrak{m}} \le \mathbf{x}\right) = \Phi(\mathbf{x}) + \frac{\mathbf{x}^{3} + (2(\mathfrak{c} - 1) - 3)\mathbf{x}}{4\left(m - \mathfrak{c}\right)} \varphi(\mathbf{x}) + \mathcal{O}(m^{-3/2}),\tag{7}$$

where <sup>Φ</sup>(*x*) = *x*−∞ *ϕ*(*y*)*dy* is the standard normal distribution function. Note that in Konishi [16] the sample size (in our case the dimension of vectors) is *m* + 1 and *c* = 1 + 2Δ with Konishi's correcting constant Δ. Moreover, (7) follows from the more general Theorem 2.2 in the mentioned paper for independent components in the pairs (*Xk Yk*), *k* = 1, ..., *m*.

In Christoph et al. [17], computable error bounds of approximations in (7) with *c* = 2 and *c* = 2.5 of order <sup>O</sup>(*m*<sup>−</sup><sup>2</sup>) for all *m* ≥ 7 are proved:

$$\sup\_{x} \left| P\left(\sqrt{m-2.5} \ R\_{m} \le x \right) - \Phi(x) - \frac{x^{3} \,\varphi(x)}{4\left(m-2.5\right)} \right| \le \frac{B\_{m}}{(m-2.5)^{2}} \le \frac{B}{m^{2}} \tag{8}$$

and

$$\sup\_{x} \left| P\left(\sqrt{m-2} \; R\_m \le x \right) - \Phi(x) - \frac{\left(x^3 - x\right)\varrho(x)}{4\left(m-2\right)} \right| \le \frac{B\_m^\*}{(m-2)^2} \le \frac{B^\*}{m^2} \tag{9}$$

where for some *m* ≥ 7 constants *Bm* and *B*<sup>∗</sup>*m* are calculated and presented in Table 1 in Christoph et al. [17]: i.e., *B*7 = 1.875, *B*∗7= 2.083 and *B*50 = 0.720, *<sup>B</sup>*<sup>∗</sup>50= 0.982.

Usually, the asymptotic for *Rm* is (9), where *c* = 2 since it is related to the *t*-distributed statistic √*m* − 2 *Rm* / 21 − *<sup>R</sup>*2*m*. With the correcting constant *c* = 2.5, one term in the asymptotic in (8) vanishes. Inordertransfertheoremfromnon-randomrandomdimensionofthe

 to use a to vectors, we prefer (7) with *c* = 0. In a similar manner as proving (8) and (9) in Christoph et al. [17], one can verify the following inequalities for *m* ≥ 3:

$$\sup\_{\mathbf{x}} \left| P\left(\sqrt{m}\mathbf{R}\_{\mathbf{m}} \le \mathbf{x} \right) - \Phi(\mathbf{x}) - \frac{\left(\mathbf{x}^3 - 5\mathbf{x}\right)}{4m} \varphi(\mathbf{x}) \right| \le C\_1 m^{-2}.\tag{10}$$

Let us consider now the connection between the correlation coefficient *Rm* and the angle *θm* of the involved vectors X*<sup>m</sup>*, Y*m*:

$$\theta\_m = \arg(\vec{\chi}\_m, \vec{\Upsilon}\_m). \tag{11}$$

Hall et al. [1] showed that under the given conditions

$$\theta\_m = \frac{1}{2}\pi + \mathcal{O}\_p(m^{-1/2}) \quad \text{as} \quad m \to \infty,$$

where <sup>O</sup>*p* denotes the stochastic order. Since

$$\cos\theta\_m = \frac{||\vec{X}\_m||^2 + ||\vec{Y}\_m||^2 - ||\vec{X}\_m - \vec{Y}\_m||^2}{2||\vec{X}\_m|| \, ||\vec{Y}\_m||} = R\_m(\vec{X}\_m, \vec{Y}\_m) = R\_{m\nu}$$

the computable error bounds for *θm* follows from computable error bounds for *Rm*.

For any fixed constant *c* < *m*, and arbitrary *x* with |*x*| < √*m* − *c π*/2, we obtain for the angle *θm* : 0 < *θm* < *π* :

$$\begin{aligned} P\left(\sqrt{m-c}(\theta\_m - \pi/2) \le x\right) &=& P\left(\theta\_m \le \pi/2 + x/\sqrt{m-c}\right) \\ &=& P\left(\cos\theta\_m \ge \cos(\pi/2 + x/\sqrt{m-c})\right) \\ &=& P\left(R\_m \ge -\sin(x/\sqrt{m-c})\right) \\ &=& P\left(\sqrt{m-c}\ R\_m \le \sqrt{m-c}\ \sin(x/\sqrt{m-c})\right) \end{aligned} \tag{12}$$

because *Rm* is symmetric and *P*(*Rm* ≤ *x*) = *<sup>P</sup>*(− *Rm* ≤ *<sup>x</sup>*).

Equation (12) shows the connection between the correlation coefficient *Rm* and the angle *θm* among the vectors involved. In Christoph et al. [17], computable error bound of approximation in (8) are used to obtain similar bound for the approximation of the angle between two vectors, defined in (11). Here, the approximation (10) and (12) with *c* = 0 lead for any *m* ≥ 3 and for |*x*| ≤ *π* √*m* /2 to

$$\sup\_{x} \left| P\left(\sqrt{m}(\theta\_m - \frac{\pi}{2}) \le x \right) - \Phi(x) - \frac{(1/3)x^3 - 5x}{4m} \varphi(x) \right| \le C\_1 m^{-2}.\tag{13}$$

Many authors investigated limit theorems for the sums of random vectors when their dimension tends to infinity, see, e.g., Prokhorov [18]. In (6) and (7), the dimension *m* of the vectors X*m* and Y*m* tends to infinity.

Now, we consider the correlation coefficient of vectors X*m* and Y*<sup>m</sup>*, where the non-random dimension *m* is replaced by a random dimension *Nn* ∈ N+ = {1, 2, ...} depending on some natural parameter *n* ∈ N+ and *Nn* is independent of *Xm* and *Ym* for any *m*, *n* ∈ N+. Define

$$R\_{N\_w} = \frac{\sum\_{k=1}^{N\_n} X\_k \,\, \mathbf{y}\_k}{\sqrt{\sum\_{k=1}^{N\_n} X\_k^2 \,\, \sum\_{k=1}^{N\_n} \,\, \mathbf{y}\_k^2}} \cdot$$

#### **3. Statistical Models with a Random Number of Observations**

Let *X*1, *X*2, ... ∈ R = (−<sup>∞</sup> ∞) and *N*1, *N*2, ... ∈ N+ = {1, 2, ...} be random variables on the same probability space (<sup>Ω</sup>, A, <sup>P</sup>). Let *Nn* be a random size of the underlying sample, i.e., the random number of observations, which depends on parameter *n* ∈ N+. We suppose for each *n* ∈ N+ that *Nn* ∈ N+ is independent of random variables *X*1, *X*2, ... and *Nn* → ∞ in probability as *n* → ∞. Let *Tm* := *Tm* (*<sup>X</sup>*1,..., *Xm*) be some statistic of a sample with *non-random sample size m* ∈ N+. Define the random variable *TNn*for every *n* ∈ N+:

$$T\_{N\_n}(\omega) := T\_{N\_n(\omega)}\left(X\_1(\omega), \dots, X\_{N\_n(\omega)}(\omega)\right), \ \omega \in \Omega, \ \omega$$

i.e., *TNn*is some statistic obtained from a random sample *X*1, *X*2,..., *XNn*.

The randomness of the sample size may crucially change asymptotic properties of *TNn* , see, e.g., Gnedenko [7] or Gnedenko and Korolev [19].

## *3.1. Random Sums*

Many models lead to random sums and random means

$$S\_{N\_n} = \sum\_{k=1}^{N\_n} X\_k \quad \text{and} \quad M\_{N\_n} = \frac{1}{N\_n} \sum\_{k=1}^{N\_n} X\_{k'}. \tag{14}$$

A fundamental introduction to asymptotic distributions of random sums is given in Döbler [20].

It is worth mentioning that a suitable scaled factor by *SNn* affects the type of limit distribution. In fact, consider random sum *SNn* given in (14). For the sake of convenience, let *X*1, *X*2, ... be independent standard normal random variables and *Nn* ∈ N+ be geometrically distributed with *E*(*Nn*) = *n* and independent of *X*1, *X*2, .... Then, one has

$$\mathbb{P}\left(\frac{1}{\sqrt{N\_n}}S\_{N\_n} \le x\right)\_{\cdot = \cdot} = \int\_{-\infty}^{x} \frac{1}{\sqrt{2\pi}}e^{-u^2/2} du \quad \text{for all } n \in \mathbb{N},\tag{15}$$

$$\mathbb{P}\left(\frac{1}{\sqrt{\mathbb{E}(N\_n)}}S\_{N\_n} \le x\right) \quad \to \quad \int\_{-\infty}^{x} \frac{1}{\sqrt{2}} e^{-\sqrt{2}|u|} du \quad \text{as } n \to \infty,\tag{16}$$

$$\mathbb{P}\left(\frac{\sqrt{\mathbb{E}(N\_n)}}{N\_n}S\_{N\_n} \le x\right) \quad \to \quad \int\_{-\infty}^{x} \left(2 + u^2\right)^{-3/2} du \quad \text{as } n \to \infty. \tag{17}$$

We have three different limit distributions. The suitable scaled geometric sum *SNn* is standard normal distributed or tends to the Laplace distribution with variance 1 depending on whether we take the random scaling factor 1/√*Nn* or the non-random scaling factor 1/√<sup>E</sup>*Nn*, respectively. Moreover, we ge<sup>t</sup> the Student distribution with two degrees of freedom as the limit distribution if we use scaling with the mixed factor 2E(*Nn*)/*Nn*. Similar results also hold for the normalized random mean *MNn* = 1*NnSNn* .

Assertion (15) is obtained by conditioning and the stability of the normal law. Moreover, using Stein's method, quantitative Berry–Esseen bounds in (15) and (16) for arbitrary centered random

variables *X*1 with <sup>E</sup>(|*<sup>X</sup>*1|<sup>3</sup>) < ∞ were proved in (Chen et al. [21], Theorem 10.6), (Döbler [20] Theorems 2.5 and 2.7) and (Pike and Ren [22] Theorem 3), respectively. Statement (17) follows from (Bening and Korolev [23] Theorem 2.1).

First order asymptotic expansions are obtained for the distribution function of random sample mean and random sample median constructed from a sample with two different random sizes in Bening et al. [24] and in the conference paper Bening et al. [25]. The authors make use of the rate of convergence of P(*Nn* ≤ *gnx*) to the limit distribution *<sup>H</sup>*(*x*) with some *gn* ↑ ∞. In Christoph et al. [12], second order expansions for the normalized random sample sizes are proved, see below Propositions 1 and 2. These results allow for proving second order asymptotic expansions of random sample mean in Christoph et al. [12] and random sample median in Christoph et al. [13].

#### *3.2. Transfer Proposition from Non-Random to Random Sample Sizes*

Consider now the statistic *TNn* = *TNn XNn* , *YNn*,, where the dimension of the vectors *XNn* , *YNn* is a random number *Nn* ∈ N+.

In order to avoid too long expressions and at the same time to preserve a necessary accuracy, we limit ourselves to obtaining limit distributions and terms of order *m*<sup>−</sup><sup>1</sup> in the following non-asymptotic approximations with a bounds of order *m*<sup>−</sup>*<sup>a</sup>* for some *a* > 1.

We suppose that the following condition on the statistic *Tm* = *Tm*(*Xm*, *Ym*) with E*Tm* = 0 is met for a non-random sample size *m* ∈ N+:

**Condition 1.** *There exist differentiable bounded function f*2(*x*) *with* sup*x* |*x f* 2(*x*)| < *c*0 *and real numbers a* > 1*, C*1 > 0 *such that for all integer m* ≥ 1

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( m^{\gamma} T\_{\mathbf{m}} \le \mathbf{x} \right) - \Phi(\mathbf{x}) - m^{-1} f\_{\mathbf{2}}(\mathbf{x}) \right| \le C\_1 m^{-a} \tag{18}$$

*where γ* ∈ {−1/2, 0, 1/2}*.*

**Remark 1.** *Relations (10) and (13) give the examples of statistics such that Condition 1 is met. For other examples of multivariate statistics of this kind, see Chapters 14–16 in Fujikoshi et al. [2].*

Suppose that the limiting behavior of distribution functions of the normalized random size *Nn* ∈ N+ is described by the following condition.

**Condition 2.** *There exist a distribution function <sup>H</sup>*(*y*) *with H*(0+) = 0*, a function of bounded variation h*2(*y*)*, a sequence* 0 < *gn* ↑ ∞ *and real numbers b* > 0 *and C*2 > 0 *such that for all integer n* ≥ 1

$$\left| \sup\_{y \ge 0} \left| \mathbb{P} \{ \mathbf{g}\_n^{-1} \mathbf{N}\_n \le y \} - H(y) \right| \le \mathbb{C}\_2 n^{-b}, \quad \quad 0 < b \le 1$$

$$\sup\_{y \ge 0} \left| \mathbb{P} \{ \mathbf{g}\_n^{-1} \mathbf{N}\_n \le y \} - H(y) - n^{-1} h\_2(y) \right| \le \mathbb{C}\_2 n^{-b}, \quad b > 1 \tag{19}$$

**Remark 2.** *In Propositions 1 and 2 below, we get the examples of discrete random variables Nn such that Condition 2 is met.*

Conditions 1 and 2 allow us to construct asymptotic expansions for distributions of randomly normalized statistics on the base of approximation results for normalized fixed-size statistics (see relation (18)) and for the random size of the underlying sample (see relation (19)). As a result, we obtain the following transfer theorem.

**Theorem 1.** *Let* |*γ*| ≤ *K* < ∞ *and both Conditions 1 and 2 be satisfied. Then, the following inequality holds for all n* ∈ N+ *:*

$$\sup\_{\mathbf{x}} \sup\_{\mathbf{x} \in \mathbb{R}} \left| \mathbb{P} \left( \mathbf{g}\_n^{\gamma} T\_{\mathcal{N}\_n} \le \mathbf{x} \right) - G\_n(\mathbf{x}, 1/\mathcal{g}\_n) \right| \le C\_1 \mathbb{E} \left( \mathcal{N}\_n^{-a} \right) + \left( \mathcal{C}\_3 \mathcal{D}\_n + \mathcal{C}\_4 \right) n^{-b},\tag{20}$$

$$G\_{\rm n}(\mathbf{x}, 1/\mathcal{g}\_{\rm n}) = \int\_{1/\mathcal{g}\_{\rm n}}^{\infty} \left( \Phi(\mathbf{x} \, y^{\gamma}) + \frac{f\_{\rm 2}(\mathbf{x} \, y^{\gamma})}{\mathcal{g}\_{\rm n} \mathcal{y}} \right) d\left( H(y) + \frac{h\_{\rm 2}(y)}{n} \right),\tag{21}$$

$$D\_n = \sup\_{\mathbf{x}} \int\_{1/\mathcal{J}\_n}^{\infty} \left| \frac{\partial}{\partial \mathcal{Y}} \left( \Phi(\mathbf{x} \mathbf{y}^\gamma) + \frac{f\_2(\mathbf{x} \mathbf{y}^\gamma)}{\mathcal{Y} \mathcal{g}\_n} \right) \right| d\mathbf{y}\_\prime \tag{22}$$

*where a* > 1, *b* > 0, *f*2(*z*), *h*2(*y*) *are given in (18) and (19). The constants C*1, *C*3, *C*4 *do not depend on n.*

**Remark 3.** *Later, we use only the cases γ* ∈ {−1/2, 0, 1/2}*.*

**Remark 4.** *The domain* [1/*gn*, ∞) *of integration in (21) depends on gn. Thus, it is not clear how Gn*(*<sup>x</sup>*, 1/*gn*) *is represented as a polynomial in g*<sup>−</sup><sup>1</sup> *n and <sup>n</sup>*<sup>−</sup>1*. To overcome this problem (see (26)), we prove the following theorem.*

**Theorem 2.** *Under the conditions of Theorem 1 and the additional conditions on functions <sup>H</sup>*(.) *and h*2(.)*, depending on the convergence rate b in (19):*

$$H(\mathbf{1}/\mathbf{g}\_n) \le c\_1 \mathbf{g}\_n^{-b}, \quad b > 0,\tag{23}$$

$$\begin{array}{ll} i: & \int\_{0}^{1/\mathcal{g}\_{n}} y^{-1} dH(y) \le c\_{2} \mathcal{g}\_{n}^{-b+1}, \\\ ii: & h\_{2}(0) = 0 \quad \text{and} \quad |h\_{2}(1/\mathcal{g}\_{n})| \le c\_{3} n \, \mathcal{g}\_{n}^{-b}, \\\ iii: & \int\_{0}^{1/\mathcal{g}\_{n}} y^{-1} |h\_{2}(y)| dy \le c\_{4} n \, \mathcal{g}\_{n}^{-b}, \end{array} \tag{24}$$

*we obtain for the function Gn*(*<sup>x</sup>*, 1/*gn*) *defined in (21):*

$$\sup\_{\mathbf{x}} \left| G\_{\mathbf{n}}(\mathbf{x}, 1/\mathbf{g}\_{\mathbf{n}}) - G\_{\mathbf{n}, \mathbf{2}}(\mathbf{x}) \right| \leq C \mathbf{g}\_{\mathbf{n}}^{-b} + \sup\_{\mathbf{x}} \left( \left| I\_{1}(\mathbf{x}, \mathbf{n}) \right| \mathbb{I}\_{\{b < 1\}}(b) + \left| I\_{2}(\mathbf{x}, \mathbf{n}) \right| \right) \tag{25}$$

*with*

$$\mathcal{G}\_{n2}(\mathbf{x}) = \begin{cases} \int\_{0}^{\infty} \Phi(\mathbf{x}\,\mathbf{y}') dH(\mathbf{y}), & 0 < b < 1, \\ \int\_{0}^{\infty} \Phi(\mathbf{x}\,\mathbf{y}') dH(\mathbf{y}) + \frac{1}{\mathfrak{J}\_{n}} \int\_{0}^{\infty} \frac{f\_{2}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}} dH(\mathbf{y}), & b = 1 \\ \int\_{\frac{\log\_{b}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}\_{n}}} \int\_{\frac{\log\_{b}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}\_{n}}}^{\infty} \frac{f\_{2}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}} dH(\mathbf{y}) \mathbbm{1}\_{\{\gamma - 0\}}(\gamma) + \frac{1}{\mathfrak{H}} \int\_{0}^{\infty} \Phi(\mathbf{x}\,\mathbf{y}') dh\_{2}(\mathbf{y}), & b > 1. \end{cases} \tag{26}$$
 
$$I\_{1}(\mathbf{x},\mathbf{n}) = \int\_{1/\mathcal{J}\_{\widetilde{\mathcal{B}}}}^{\infty} \frac{f\_{2}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}\_{n}\mathfrak{g}} dH(\mathbf{y}) \text{ for } b \le 1 \quad \text{and} \quad I\_{2}(\mathbf{x},\mathbf{n}) = \int\_{1/\mathcal{J}\_{\widetilde{\mathcal{B}}}}^{\infty} \frac{f\_{2}(\mathbf{x}\,\mathbf{y}')}{\mathfrak{J}\_{n}\mathfrak{n}\mathfrak{g}} dh\_{2}(\mathbf{y}) \text{ for } b > 1.$$

**Remark 5.** *The additional conditions (23) and (24) guarantee to extend the integration range from* [1/*gn*, ∞) *to* (0, ∞) *of the integrals in (26).*

Theorems 1 and 2 are proved in Appendix A.1.

#### **4. Auxiliary Propositions and Lemmas**

Consider the standardized correlation coefficient (5) having density (6) with correcting real constant *c* = 0 and standardized angle √*m*(*<sup>θ</sup>m* − *<sup>π</sup>*/2), see (12). By (10) and (13) for *m* ≥ 3, we have

$$\sup\_{\mathbf{x}} \sup\_{\mathbf{x}} \left| P\left(\sqrt{m} \, \mathbf{R}\_m \le \mathbf{x} \right) - \Phi(\mathbf{x}) - \frac{\left(\mathbf{x}^3 - \mathbf{5} \cdot \mathbf{x}\right)}{4m} \varphi(\mathbf{x}) \right| \le C\_1 m^{-2}, \qquad m \in \mathbb{N}\_+. \tag{27}$$

and for the angle *θm* between the vectors for |*x*| ≤ *π* √*m* /2

$$\sup\_{\mathbf{x}} \sup\_{\mathbf{x}} \left| P\left( \sqrt{m} (\theta\_m - \frac{\pi}{2}) \le \mathbf{x} \right) - \Phi(\mathbf{x}) - \frac{(1/3)\mathbf{x}^3 - 5\mathbf{x}}{4\ m} \,\upvarphi(\mathbf{x}) \right| \le \mathbb{C}\_1 m^{-2}, \qquad m \in \mathbb{N}\_+,\tag{28}$$

where (27) and (28) for *m* = 1 and *m* = 2 are trivial and *C*1 does not depend on *m*.

Suppose *f*2(*<sup>x</sup>*; *a*)=(*a x*3 − 5 *x*) *ϕ*(*x*)/4 with *a* = 1 or *a* = 1/3 when (27) or (28) are considered. Since a product of polynomials in *x* with *ϕ*(*x*) is always bounded, numerical calculus leads to

$$\sup\_{x} |x f\_2'(x; a)| = \sup\_{x} |x \left( a x^4 - (3a + 5) \ge^2 + 5 \right)| \, q(x)/4 \le 0.4.$$

Condition 1 of the transfer Theorem 1 to the statistics *Rm* and *θm* are satisfied with *c*0 = 0.4 and *a* = 2.

Next, we estimate *Dn*(*x*) defined in (22).

**Lemma 1.** *Let gn a sequence with* 0 < *gn* ↑ ∞ *as n* → ∞*. Then, with some* 0 < *<sup>c</sup>*(*<sup>γ</sup>*, *a*) < <sup>∞</sup>*, we obtain with a* = 1 *and a* = 1/3*:*

$$D\_{\mathfrak{n}} = \sup\_{\mathbf{x}} \int\_{1/\mathcal{G}\_{\mathfrak{n}}}^{\infty} \left| \frac{\partial}{\partial \mathbf{y}} \left( \Phi(\mathbf{x} \, y^{\gamma}) + \frac{f\_2(\mathbf{x} y^{\gamma}; a)}{\mathcal{Y} \mathbf{g}\_{\mathfrak{n}}} \right) \right| dy \leq \frac{1}{2} + \frac{\mathfrak{c}(\gamma, a)}{4}. $$

In the next subsection, we consider the cases when the random dimension *Nn* is negative binomial distributed with success probability 1/*<sup>n</sup>*.

#### *4.1. Negative Binomial Distribution as Random Dimension of the Normal Vectors*

Let the random dimension *Nn*(*r*) of the underlying normal vectors be negative binomial distributed (shifted by 1) with parameters 1/*n* and *r* > 0, having probability mass function

$$\mathbb{P}(N\_{\text{fl}}(r) = j) = \frac{\Gamma(j + r - 1)}{\Gamma(j)} \left(\frac{1}{n}\right)^r \left(1 - \frac{1}{n}\right)^{r - 1}, \quad j = 1, 2, \dots \tag{29}$$

with <sup>E</sup>(*Nn*(*r*)) = *r* (*n* − 1) + 1. Then, <sup>P</sup>(*Nn*(*r*)/*gn* ≤ *x*) tends to the Gamma distribution function *Gr*,*<sup>r</sup>*(*x*) with the shape and rate parameters *r* > 0, having density

$$g\_{r^r}(\mathbf{x}) = \frac{r^r}{\Gamma(r)} \mathbf{x}^{r-1} e^{-r\mathbf{x}} \mathbb{I}\_{(0,\infty)}(\mathbf{x}), \quad \mathbf{x} \in \mathbb{R}. \tag{30}$$

If the statistic *Tm* is asymptotically normal, the limit distribution of the standardized statistic *TNn*(*r*) with random size *Nn*(*r*) is Student's *t*-distribution *<sup>S</sup>*2*r*(*x*) having density

$$s\_{\nu}(\mathbf{x}) = \frac{\Gamma((\nu+1)/2)}{\sqrt{\nu \pi} \Gamma(\nu/2)} \left(1 + \frac{\mathbf{x}^2}{\nu}\right)^{-(\nu+1)/2}, \quad \nu > 0, \quad \mathbf{x} \in \mathbb{R}, \tag{31}$$

with *ν* = 2*r*, see Bening and Korolev [23] or Schluter and Trede [26].

**Proposition 1.** *Let r* > 0*, discrete random variable Nn*(*r*) *have probability mass function (29) and gn* := <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1*. For x* > 0 *and all n* ∈ N *there exists a real number <sup>C</sup>*2(*r*) > 0 *such that*

$$\sup\_{r\geq 0} \left| \mathbb{P} \left( \frac{N\_{\mathbb{R}}(r)}{g\_n} \leq x \right) - G\_{\mathbb{r}\mathcal{I}}(x) - \frac{h\_{\mathbb{2};r}(x)}{n} \right| \leq C\_2(r) \, n^{-\min\{r, 2\}},\tag{32}$$

*where*

$$h\_{2,r}(\mathbf{x}) = \begin{cases} 0, & \text{for } r \le 1, \\ \frac{\mathcal{G}\_{r,r}(\mathbf{x}) \left( (\mathbf{x} - 1)(2 - r) + 2Q\_1(\mathbf{g}\_n \mathbf{x}) \right)}{2r}, & \text{for } r > 1. \end{cases}$$

$$Q\_1(y) = 1/2 - (y - \lfloor y \rfloor) \text{ and } \lfloor . \rfloor \text{ denotes the integer part of a number.} \tag{33}$$

Figure 1 shows the approximation of P (*Nn*(*r*) ≤ (*r*(*n* − 1) + <sup>1</sup>)*x*) by *<sup>G</sup>*2,2(*x*) and *<sup>G</sup>*2,2(*x*) + *h*2(*x*)/*<sup>n</sup>*.

**Figure 1.** Distribution function P (*Nn*(*r*) ≤ (*r*(*n* − 1) + <sup>1</sup>)*x*) (black line, almost covered by the red line), the limit law *<sup>G</sup>*2,2(*x*) (blue line) and the second approximation *<sup>G</sup>*2,2(*x*) + *h*2(*x*)/*<sup>n</sup>* (red line) with *n* = 25 and *r* = 2.

**Remark 6.** *The convergence rate for r* ≤ 1 *is given in Bening et al. [24] or Gavrilenko et al. [27]. The Edgeworth expansion for r* > 1 *is proved in Christoph et al. [12], Theorem 1. The jumps of the sample size Nn*(*r*) *have an effect only on the function Q*1(.) *in the term <sup>h</sup>*2;*r*(.)*.*

The negative binomial random variable *Nn* satisfies Condition 2 of the transfer Theorem 1 with *<sup>H</sup>*(*x*) = *Gr*,*<sup>r</sup>*(*x*), *h*2(*x*) = *<sup>h</sup>*2;*r*(*x*), *gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1 and *b* = min{*r* <sup>2</sup>}.

**Lemma 2.** *In Theorem 2 the additional conditions* (23) *and* (24) *are satisfied with <sup>H</sup>*(*x*) = *Gr*,*<sup>r</sup>*(*x*)*, h*2(*x*) = *<sup>h</sup>*2;*r*(*x*)*, gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1 *and b* = min{*r* <sup>2</sup>}*. Moreover, one has for γ* ∈ {−1/2, 0, 1/2} *and f*2(*<sup>z</sup>*; *a*)=(*a z*3 − 5 *z*) *ϕ*(*z*)/4*, with a* = 1 *or a* = 1/3*:*

$$|I\_1(\mathbf{x}, n)| = \begin{cases} \left| \int\_{1/\xi\_n}^{\infty} \frac{f\_2(\mathbf{x} \, \mathbf{y}^\gamma; a)}{\mathfrak{S}\_n \mathbf{y}} d\mathbf{G}\_{r\mathcal{I}}(\mathbf{y}) \right| \le c\varepsilon \, \mathbf{g}\_n^{-r} & r < 1, \\\left| \int\_{1/n}^{\infty} \frac{f\_2(\mathbf{x} \, \mathbf{y}^\gamma; a)}{n \, \mathbf{y}} d\mathbf{G}\_{1,1}(\mathbf{y}) - n^{-1} f\_2(\mathbf{x}; a) \ln n \, \mathbb{I}\_{\{\gamma = 0\}}(\gamma) \right| \le c\_6 n^{-1}, & r = 1, \end{cases} \tag{34}$$

$$|l\_2(\mathbf{x}, n)| = \left| \int\_{1/\xi\_n}^{\infty} \frac{f\_2(\mathbf{x} \, y^\gamma; a)}{g\_n \, n \, \mathbf{y}} dh\_{2\tau}(y) \right| \le \begin{cases} c\gamma g\_n^{-r}, & r > 1, r \ne 2, \\ \left(c\gamma + c\varsigma \ln n \, \mathbb{I}\_{\{\gamma = 0\}}(\gamma)\right) g\_n^{-2}, & r = 2. \end{cases} \tag{35}$$

*Furthermore, we have*

$$0 \le g\_n^{-1} - \left(r n\right)^{-1} \le \left(r - 1\right) \left(r n\right)^{-2} e^{-1/2} \quad \text{for} \quad r > 1, \quad n \ge 2. \tag{36}$$

In addition to the expansion of *Nn*(*r*) a bound of <sup>E</sup>(*Nn*(*r*))−*<sup>a</sup>* is required, where *m*<sup>−</sup>*<sup>a</sup>* is rate of convergence of Edgeworth expansion for *Tm*, see (18).

**Lemma 3.** *Let r* > 0*, α* > 0 *and the random variable Nn*(*r*) *is defined by* (29)*. Then,*

$$\mathbb{E}\left(N\_{\boldsymbol{n}}(\boldsymbol{r})\right)^{-\alpha} \le \mathcal{C}(\boldsymbol{r}) \begin{cases} n^{-\min\{\boldsymbol{r},\boldsymbol{a}\}}, & \boldsymbol{r} \ne \boldsymbol{a} \\ \ln(n) \, n^{-\alpha}, & \boldsymbol{r} = \boldsymbol{a} \end{cases} \tag{37}$$

*and the convergence rate in case r* = *α cannot be improved.* *4.2. Maximum of n Independent Discrete Pareto Random Variables Is the Dimension of the Normal Vectors*

Let *<sup>Y</sup>*(*s*) ∈ N be discrete Pareto II distributed with parameter *s* > 0, having probability mass and distribution functions

$$\mathbb{P}(Y(s) = k) = \frac{s}{s+k-1} - \frac{s}{s+k} \quad \text{and} \quad \mathbb{P}(Y(s) \le k) = \frac{k}{s+k'}, k \in \mathbb{N}, \tag{38}$$

which is a particular class of a general model of discrete Pareto distributions, obtained by discretization continuous Pareto II (Lomax) distributions on integers, see Buddana and Kozubowski [28].

Now, let *<sup>Y</sup>*1(*s*),*Y*2(*s*), ..., be independent random variables with the same distribution (38). Define for *n* ∈ N and *s* > 0 the random variable

$$N\_n(s) = \max\_{1 \le j \le n} \mathcal{Y}\_j(s) \quad \text{with} \quad \mathbb{P}(N\_n(s) \le k) = \left(\frac{k}{s+k}\right)^n, \quad n \in \mathbb{N}.\tag{39}$$

It should be noted that the distribution of *Nn*(*s*) is extremely spread out on the positive integers. In Christoph et al. [12], the following Edgeworth expansion was proved:

**Proposition 2.** *Let the discrete random variable Nn*(*s*) *have distribution function (39). For x* > 0*, fixed s* > 0 *and all n* ∈ N*, then there exists a real number <sup>C</sup>*3(*s*) > 0 *such that*

$$\sup\_{y>0} \left| \mathbb{P} \left( \frac{N\_n(s)}{n} \le y \right) - H\_s(y) - \frac{h\_{2\mathcal{P}}(y)}{n} \right| \le \frac{C\_3(s)}{n^2},$$

$$H\_{\mathbb{S}}(y) = e^{-s/y} \text{ and } h\_{2\mathcal{P}}(y) = s e^{-s/y} \left( s - 1 + 2Q\_1(n \, y) \right) / \left( 2y^2 \right), \, y > 0 \tag{40}$$

*where Q*1(*y*) *is defined in (33).*

**Remark 7.** *The continuous function Hs*(*y*) = *<sup>e</sup>*<sup>−</sup>*<sup>s</sup>*/*y*I(0 ∞)(*y*) *with parameter s* > 0 *is the distribution function of the inverse exponential random variable <sup>W</sup>*(*s*) = 1/*V*(*s*)*, where <sup>V</sup>*(*s*) *is exponentially distributed with rate parameter s* > 0*. Both Hs*(*y*) *and* <sup>P</sup>(*Nn*(*s*) ≤ *y*) *are heavy tailed with shape parameter 1.*

**Remark 8.** *Therefore,* E-*Nn*(*s*) = ∞ *for all n* ∈ N *and* E-*<sup>W</sup>*(*s*) = ∞*. Moreover:*


*On pseudo moments and some of their generalizations, see Chapter 2 in Christoph and Wolf [29].*

**Lemma 4.** *In Transfer Theorem 2, the additional conditions* (23) *and* (24) *are satisfied with <sup>H</sup>*(*y*) = *Hs*(*y*) = *e* −*<sup>s</sup>*/*y , h*2(*y*) = *<sup>h</sup>*2;*s*(*y*) = *s e*<sup>−</sup>*<sup>s</sup>*/*<sup>y</sup> s* − 1 + <sup>2</sup>*Q*1(*<sup>n</sup> y*)/-2 *<sup>y</sup>*<sup>2</sup>, *y* > 0*, gn* = *n and b* = 2*. Moreover, one has for* |*γ*| ≤ *K* < ∞ *and f*2(*<sup>z</sup>*; *a*)=(*a z*3 − 5 *z*) *ϕ*(*z*)/4*, with a* = 1 *or a* = 1/3*:*

$$I\_2(\mathbf{x}, n) = \left| \int\_{1/n}^{\infty} \frac{f\_2(\mathbf{x} \, y^\gamma; a)}{n^2 \, y} dh\_{2; \mathbf{s}}(y) \right| \le C(s) n^{-2}.$$

**Lemma 5.** *For random size Nn*(*s*) *with probabilities* (39) *with reals s* ≥ *s*0 > 0 *and arbitrary small s*0 > 0 *and n* ≥ 1*, we have*

$$\mathbb{E}\left(N\_n(s)\right)^{-\alpha} \le C(s)n^{-\alpha}.\tag{41}$$

The Lemmas are proved in Appendix A.2.

## **5. Main Results**

Consider the sample correlation coefficient *Rm* = *Rm*(X*<sup>m</sup>*,Y*m*), given in (4) and the two statistics *R*<sup>∗</sup>*m* = √*m Rm* and *R*∗∗*m* = *m Rm* which differ from *Rm* by scaling factors. Hence, by (10),

$$\mathbb{P}(\sqrt{m}R\_{\mathfrak{m}} \le \mathbf{x}) = \mathbb{P}(R\_{\mathfrak{m}}^\* \le \mathbf{x}) = \mathbb{P}\left(\frac{1}{\sqrt{m}}R\_{\mathfrak{m}}^{\*\*} \le \mathbf{x}\right) = \Phi(\mathbf{x}) + \frac{(\mathbf{x}^5 - 5\mathbf{x})}{4\ m}\varphi(\mathbf{x}) + r(m) \tag{42}$$

with |*r*(*m*)| ≤ *Cm*−2.

Let *θm* be the angle between the vectors X*m* and Y*<sup>m</sup>*. Contemplate the statistics Θ*m* = *θm* − *π*/2, Θ<sup>∗</sup>*m* = √*m* (*<sup>θ</sup>m* − *π*/2) and Θ∗∗*m* = *m* (*<sup>θ</sup>m* − *π*/2) which differ only in scaling . Then, by (13),

$$\mathbb{P}(\sqrt{m}\Theta\_{m}\le\mathbf{x}) = \mathbb{P}(\Theta\_{m}^{\*}\le\mathbf{x}) = \mathbb{P}\left(\frac{1}{\sqrt{m}}\Theta\_{m}^{\*\*}\le\mathbf{x}\right) = \Phi(\mathbf{x}) + \frac{\left((1/3)\mathbf{x}^{3} - 5\,\mathbf{x}\right)}{4\,m}\varphi(\mathbf{x}) + r^{\*}(m)$$

with |*r*<sup>∗</sup>(*m*)| ≤ *Cm*−2.

Consider now the statistics *RNn* , *<sup>R</sup>*<sup>∗</sup>*Nn* and *<sup>R</sup>*∗∗*Nn* as well as Θ*Nn* , <sup>Θ</sup><sup>∗</sup>*Nn* and <sup>Θ</sup>∗∗*Nn* when the vectors have random dimension *Nn*. The normalized statistics have different limit distributions as *n* → ∞.

#### *5.1. The Random Dimension Nn* = *Nn*(*r*) *Is Negative Binomial Distributed*

Let the random dimension *Nn*(*r*) be negative binomial distributed with probability mass function (29) and *gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1. "The negative binomial distribution is one of the two leading cases for count models, it accommodates the overdispersion typically observed in count data (which the Poisson model cannot)", see Schluter and Trede [26].

It follows from Theorems 1 and 2 and Proposition 1 that if limit distributions for P *gγn Nn*(*r*)1/2−*<sup>γ</sup>RNn*(*r*) ≤ *x* for *γ* ∈ {1/2, 0 − 1/2} exist they are ∞0 <sup>Φ</sup>(*x y<sup>γ</sup>*)*dGr*,*<sup>r</sup>*(*y*) with densities given bellow in the proof of the corresponding theorems:

$$\frac{r^r}{\sqrt{2\pi}\,\Gamma(r)} \int\_0^\infty y^{r-1/2} e^{-(x\cdot y^\gamma + ry)} dy = \begin{cases} s\_{2r}(\mathbf{x}) = \frac{\Gamma(r + 1/2)}{\sqrt{2\pi}\,\Gamma(r)} \left( 1 + \frac{\mathbf{x}^2}{2\pi} \right)^{-(r + 1/2)}, & \gamma = 1/2, \\\ q(\mathbf{x}) = \frac{1}{\sqrt{2}\,\pi} e^{-\mathbf{x}^2/2}, & \gamma = 0, \\\ l\_1(\mathbf{x}) = \frac{1}{\sqrt{2}} e^{-\sqrt{2}|\mathbf{x}|}, & \text{for} \quad r = 1, \\\ \end{cases} \tag{43}$$

where in case *γ* = −1/2 for *r* -= 1 generalized Laplace distributions occur.

## 5.1.1. Student's *t*-Distribution

We start with the case *γ* = 1/2 in Theorems 1 and 2. Consider the statistic *RNn*(*r*) = √*gnRNn*(*r*). The limit distribution is the Student's *t*-distribution *<sup>S</sup>*2*r*(*x*) with 2 *r* degrees of freedom with density (31).

**Theorem 3.** *Let r* > 0 *and (29) be the probability mass function of the random dimension Nn* = *Nn*(*r*) *of the vectors under consideration. If the representation (42) for the statistic Rm and the inequality (32) with gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1 *hold, then there exists a constant Cr such that for all n* ∈ N+

$$\sup\_{x} \left| \mathbb{P} \left( \sqrt{g\_n} R\_{N\_n(r)} \le x \right) - S\_{2r,n}(x; 1) \right| \le C\_r \begin{cases} n^{-\min\{r, 2\}}, & r \ne 2, \\ \ln(n) \, n^{-2}, & r = 2, \end{cases} \tag{44}$$

*where*

$$S\_{2r;n}(\mathbf{x};a) = S\_{2r}(\mathbf{x}) + \frac{s\_{2r}(\mathbf{x})}{rn} \left( a \,\mathbf{x}^3 - \frac{10r \,\mathbf{x} + 5 \,\mathbf{x}^3}{2r - 1} + \frac{(2 - r) \,(\mathbf{x}^3 + \mathbf{x})}{4 \,(2r - 1)} \right). \tag{45}$$

*Moreover, the scaled angle <sup>θ</sup>Nn*(*r*) *between the vectors* <sup>X</sup>*Nn*(*r*) *and* <sup>Y</sup>*Nn*(*r*) *allows the estimate*

$$\sup\_{r} \left| \mathbb{P} \left( \sqrt{\mathcal{g}\_{\mathbb{H}}} \left( \theta\_{\mathrm{N}\_{\mathbb{H}}(r)} - \pi/2 \right) \leq x \right) - S\_{2r,\mathbb{H}}(x; 1/3) \right| \leq C\_r \begin{cases} n^{-\min\{r,2\}}, & r \neq 2, \\ \ln(n) \ n^{-2}, & r = 2, \end{cases}$$

*where <sup>S</sup>*2*r*;*n*(*<sup>x</sup>*; 1/3) *is given in* (45) *with a* = 1/3*.*

Figure 2 shows the advantage of the Chebyshev–Edgeworth expansion versus the limit law in approximating the empirical distribution function.

**Figure 2.** Empirical version of P √*gn RNn*(*r*) ≤ *x* (blue line), limit Student law *<sup>S</sup>*2*r*(*x*) (orange line) and second approximation *<sup>S</sup>*2*r*;*n*(*<sup>x</sup>*; 1) (green line) for the correlation coefficient for pairs of normal vectors with random dimension *<sup>N</sup>*25(2). Here, *x* > 0, *n* = 25 and *r* = 2.

**Remark 9.** *The limit Student's t-distribution <sup>S</sup>*2*r*(*x*) *is symmetric and a generalized hyperbolic distribution which can be written as a regularized incomplete beta function Iz*(*<sup>a</sup>*, *b*)*. For x* > 0*:*

$$S\_{2r}(\mathbf{x}) = \int\_{-\infty}^{\mathbf{x}} s\_{2r}(\mathbf{u}) \, d\mathbf{u} = \frac{1}{2} \left( 1 + I\_{2r/(\mathbf{x}^2 + 2r)}(1/2, r) \right) \quad \text{and} \quad I\_{\overline{z}}(a, b) = \frac{\Gamma(a + b)}{\Gamma(a)} \int\_0^z t^{a - 1} (1 - t)^{b - 1}.$$

**Remark 10.** *For integer values ν* = 2 *r* ∈ {1, 2, ...} *the Student's t-distribution <sup>S</sup>*2*r*(*x*) *is computable in closed form:*

$$\text{the Cauchy law} \quad \mathcal{S}\_1(\mathbf{x}) = \frac{1}{2} + \frac{1}{\pi} \arctan(\mathbf{x}), \qquad \mathcal{S}\_2(\mathbf{x}) = \frac{1}{2} + \frac{\mathbf{x}}{2\sqrt{2 + \mathbf{x}^2}}.$$

$$S\_3(\mathbf{x}) = \frac{1}{2} + \frac{1}{\pi} \left( \frac{\mathbf{x}}{\sqrt{3}(1 + \mathbf{x}^2/3)} + \arctan(\mathbf{x}/\sqrt{3}) \right) \quad \text{and} \quad S\_4(\mathbf{x}) = \frac{1}{2} + \frac{27(\mathbf{x}^2 + 3)\mathbf{x}\left(2\mathbf{x}^2 + 9\right)}{8\left(3\mathbf{x}^2 + 9\right)^{5/2}}.$$

.

**Remark 11.** *If the dimension of the vectors has the geometric distribution Nn*(1)*, then asymptotic distribution of the sample coefficient is the Student law <sup>S</sup>*2(*x*) *with two degrees of freedom.*

**Remark 12.** *The Cauchy limit distribution occurs when the dimension of the vectors has distribution Nn*(1/2)*.*

**Remark 13.** *The Student's t-distributions <sup>S</sup>*2*r*(*x*) *are heavy tailed and their moments of orders α* ≥ 2 *r do not exist.*

#### 5.1.2. Standard Normal Distribution

Let *γ* = 0 in the Theorems 1 and 2 examining the statistics *<sup>R</sup>*<sup>∗</sup>*Nn*(*r*) and <sup>Θ</sup><sup>∗</sup>*Nn*(*r*) = 2*Nn*(*r*)(*<sup>θ</sup>Nn*(*r*) − *<sup>π</sup>*/2).

**Theorem 4.** *Let r* > 0 *and Nn* = *Nn*(*r*) *be the random vector dimension having probability mass function (29). If the representation (42) for the statistic Rm and the inequality (32) with gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1 *hold, then there exists a constant Cr such that for all n* ∈ N+

$$\sup\_{r} \left| \mathbb{P} \left( \sqrt{N\_{\boldsymbol{n}}(r)} \, R\_{N\_{\boldsymbol{n}}(r)} \le x \right) - \Phi\_{\boldsymbol{n}, 2}(x; 1) \right| \le C\_r \begin{cases} n^{-\min\{r, 2\}}, & r \ne 2, \\ \ln(n) n^{-2}, & r = 2, \end{cases} \tag{46}$$

*where*

$$\Phi\_{n;2}(\mathbf{x};a) = \Phi(\mathbf{x}) + \frac{\varrho(\mathbf{x})}{n} \left( \frac{(a\,\mathbf{x}^3 - 5\,\mathbf{x})\,\ln n}{4} \mathbb{I}\_{\{r=1\}}(r) + \frac{\Gamma(r-1)}{4\,\Gamma(r)} \frac{(a\,\mathbf{x}^3 - 5\,\mathbf{x})}{\Gamma(r)} \mathbb{I}\_{\{r > 1\}}(r) \right). \tag{47}$$

*Moreover, the scaled angle <sup>θ</sup>*<sup>∗</sup>*Nn*(*r*) *between the vectors* <sup>X</sup>*Nn*(*r*) *and* <sup>Y</sup>*Nn*(*r*) *allows the estimate*

$$\sup\_{\mathbf{x}} \sup\_{\mathbf{x}} \left| \mathbb{P} \left( \sqrt{N\_n(r)} \left( \theta\_{N\_n(r)} - \pi/2 \right) \le x \right) - \Phi\_{\mathbf{n}, 2}(\mathbf{x}; 1/3) \right| \le C\_r \begin{cases} n^{-\min\{r, 2\}}, & r \ne 2, \\ \ln(n) n^{-2}, & r = 2, \end{cases}$$

*where* <sup>Φ</sup>*n*;2(*<sup>x</sup>*; 1/3) *is given in* (47) *with a* = 1/3*.*

Figure 3 shows that the second order Chebyshev–Edgeworth expansion approximates the empirical distribution function better than the limit normal distribution.

**Figure 3.** Empirical version of P 2*Nn*(*r*) *RNn*(*r*) ≤ *x*(blue line), limit normal law <sup>Φ</sup>(*x*) (orange line) and second approximation <sup>Φ</sup>*n*;2(*<sup>x</sup>*; 1) (green line) for the correlation coefficient for pairs of normal vectors with random dimension *<sup>N</sup>*25(2). Here, *x* > 0, *n* = 25 and *r* = 2.

**Remark 14.** *When the distribution function of a statistic Tm without standardization tends to the standard normal distribution* <sup>Φ</sup>(*x*)*, i.e.,* P(*Tm* ≤ *x*) → <sup>Φ</sup>(*x*)*, then the limit law for* P(*TNn* ≤ *x*) *remains the standard normal distribution* <sup>Φ</sup>(*x*)*.*

5.1.3. Generalized Laplace Distribution

Finally, we use *γ* = −1/2 in Theorems 1 and 2 examining the statistic *g*<sup>−</sup>1/2 *n <sup>R</sup>*∗∗*Nn*(*r*). Theorems 1 and 2 state that if there exists a limit distribution of P *g*<sup>−</sup>1/2 *n <sup>R</sup>*∗∗*Nn* ≤ *x* as *n* → ∞ then it has to be a scale mixture of normal distributions with zero mean and gamma distribution:

$$L\_r(\mathbf{x}) = \int\_0^\infty \Phi(\mathbf{x} y^{-1/2}) dG\_{r,r}(y)$$

having density, see formula (A9) in the proof of Theorem 5:

$$l\_r(\mathbf{x}) = \frac{r^r}{\Gamma(r)} \int\_0^\infty \varphi(\mathbf{x} y^{-1/2}) \, y^{r-3/2} e^{-ry} dy = \frac{2r^r}{\Gamma(r) \sqrt{2\pi}} \left(\frac{|\mathbf{x}|}{\sqrt{2\pi}}\right)^{r-1/2} K\_{r-1/2}(\sqrt{2\pi}|\mathbf{x}|). \tag{48}$$

where *<sup>K</sup>α*(*u*) is the *α-order Macdonald function* or *α*-order modified Bessel function of the third kind. See, e.g., Oldham et al. [30], Chapter 51, or Kotz et al. [31], Appendix, for properties of these functions.

For integer *r* = 1, 2, 3, ... these densities *lr*(*x*), so-called *Sargan densities*, and their distribution functions are computable in closed forms:

$$\begin{aligned} l\_1(\mathbf{x}) &= \frac{1}{\sqrt{2}} \begin{vmatrix} \mathbf{e} - \sqrt{2}|\mathbf{x}| & \text{and} & L\_1(\mathbf{x}) = 1 - \frac{1}{2} \mathbf{e}^{-\sqrt{2}|\mathbf{x}|}, & \mathbf{x} > 0 \\ l\_2(\mathbf{x}) &= \left(\frac{1}{2} + |\mathbf{x}|\right) \mathbf{e}^{-2|\mathbf{x}|} & \text{and} & L\_2(\mathbf{x}) = 1 - \frac{1}{2} (1 + \mathbf{x}) \mathbf{e}^{-2|\mathbf{x}|}, & \mathbf{x} > 0 \\ l\_3(\mathbf{x}) &= \frac{3\sqrt{6}}{16} \left(1 + \sqrt{6}|\mathbf{x}| + 2\mathbf{x}^2\right) \mathbf{e}^{-\sqrt{6}|\mathbf{x}|}) & \text{and} & L\_3(\mathbf{x}) = 1 - \left(\frac{1}{2} + \frac{5\sqrt{6}}{16} \frac{\mathbf{x}}{8} + \frac{3\mathbf{x}^2}{8}\right) \mathbf{e}^{-\sqrt{6}|\mathbf{x}|}, \end{aligned} \tag{49}$$

where *Lr*(−*<sup>x</sup>*) = 1 − *Lr*(*x*) for *x* ≥ 0.

The standard Laplace distribution is *<sup>L</sup>*1(*x*) with variance 1 and density *l*1(*x*) given in (49). Therefore, Sargans distributions are a kind of generalizations of the standard Laplace distribution.

**Theorem 5.** *Let r* = 1, 2, 3 *and (29) be probability mass function of the random dimension Nn* = *Nn*(*r*) *of the vectors under consideration. If the representation (42) for the statistic Rm and the inequality (32) for Nn*(*r*) *with gn* = <sup>E</sup>*Nn*(*r*) = *r*(*n* − 1) + 1 *hold, then there exists a constant Cr such that for all n* ∈ N+

$$\sup\_{r} \sup\_{\mathbf{x}} \left| \mathbb{P} \left( \operatorname{g}\_{n}^{-1/2} N\_{n}(r) \, R\_{N\_{n}(r)} \le \mathbf{x} \right) - L\_{n/2}(\mathbf{x}; \mathbf{1}) \right| \le \mathbb{C}\_{r} \begin{cases} n^{-\min\{r, 2\}}, & r \ne 2, \\ \ln(n) \, n^{-2}, & r = 2, \end{cases} \tag{50}$$

*where*

$$L\_{n2}(\mathbf{x};a) = \begin{cases} L\_1(\mathbf{x}), & r = 1, \\ L\_2(\mathbf{x}) + \frac{a \ x \left| \mathbf{x} \right| - 5 \ge \sqrt{2}}{2(n-1) + 1} e^{-2 \left| \mathbf{x} \right|}, & r = 2, \\ L\_3(\mathbf{x}) + \frac{27}{24(n-1) + 8} \left( \frac{a \ x^3}{\sqrt{2}} - \frac{5 \ge \left| \mathbf{x} \right|}{6} - \frac{5 \ge \mathbf{x}}{6 \sqrt{6}} \right) e^{-\sqrt{6} \left| \mathbf{x} \right|} \\ \qquad + \frac{9 \ge}{2 \cdot n} \left( \frac{1}{12 \sqrt{6}} + \frac{\left| \mathbf{x} \right|}{12} - \frac{\mathbf{x}^2}{6 \sqrt{6}} \right) e^{-\sqrt{6} \left| \mathbf{x} \right|}, & r = 3. \end{cases} \tag{51}$$

*For arbitrary r* > 0*, the approximation rate is given by:*

$$\sup\_{x} \left| \mathbb{P} \left( \operatorname{g}\_{n}^{-1/2} N\_{\mathbb{H}}(r) \, \middle| \, R\_{N\_{\mathbb{H}}(r)} \le x \right) - L\_{\mathbb{T}}(x) \right| \le C\_{r} n^{-\min\left\{r, 1\right\}}.$$

*Moreover, the scaled angle Nn*(*r*) *<sup>θ</sup>Nn*(*r*) *between the vectors* <sup>X</sup>*Nn*(*r*) *and* <sup>Y</sup>*Nn*(*r*) *allows the estimate*

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( \mathcal{g}\_n^{-1/2} N\_n(r) \, \theta\_{\mathcal{N}\_n(r)} \le \mathbf{x} \right) - L\_{n,2}(\mathbf{x}; 1/3) \right| \le C\_r \begin{cases} n^{-\min\{r, 2\}}, & r \ne 2, \\ \ln(n) n^{-2}, & r = 2, \end{cases}$$

*where Ln*;2(*<sup>x</sup>*; 1/3) *is given in* (51) *with a* = 1/3*.*

Figure 4 shows that the Chebyshev–Edgeworth expansion approaches the empirical distribution function better than the limit Laplace law.

**Remark 15.** *One can find the distribution functions Lr*(*x*) *for arbitrary r* > 0 *with formula 1.12.1.3 in Prudnikov et al. [32]:*

$$\begin{split} L\_{r}(\mathbf{x}) &= \ \frac{1}{2} + \frac{2r^{r}}{\sqrt{2\pi}\Gamma(r)} \int\_{0}^{\mathbf{x}} \left(\frac{|\mathbf{x}|}{2r}\right)^{r-1/2} \mathbf{K}\_{r-1/2}(\sqrt{2r}|\mathbf{x}|) d\mathbf{x} \\ &= \ \ \frac{1}{2} + \frac{\mathbf{x}}{(2r)^{(r-1/2)/2}} \Big( \mathbf{K}\_{r-1/2}(\sqrt{2r}\mathbf{x}) \mathbf{L}\_{r-3/2}(\sqrt{2r}\mathbf{x}) + \mathbf{K}\_{r-3/2}(\sqrt{2r}\mathbf{x}) \mathbf{L}\_{r-1/2}(\sqrt{2r}\mathbf{x}) \Big). \end{split}$$

*where* **<sup>L</sup>***α*(*x*) *are the modified Struve functions of order α, for properties of modified Struve functions see, e.g., Oldham et al. [30], Section 57:13.*

**Remark 16.** *The function* (48) *as density of a mixture of normal distributions with zero mean and random variance Wr having gamma distribution* P(*Wr* ≤ *y*) = *Gr*,*<sup>r</sup>*(*y*) *is given also in Kotz et al. [31], Formula (4.1.32) with τ* = *r, σ* = 1/√*r, using Formula (A.0.4) with λ* = −*r* + 3/2 *and the order-reflection formula <sup>K</sup>*−*<sup>α</sup>*(*x*) = *<sup>K</sup>α*(*x*)*. Such a variance gamma model is studied in Madan and Senata [33] for share market returns.*

**Remark 17.** *A systematic exposition about the Laplace distribution and its numerous generalization and diverse applications one finds in the useful and interesting monography by Kotz et al. [31]. Here, these generalized Laplace distributions <sup>L</sup>*1(*x*)*, <sup>L</sup>*2(*x*) *and <sup>L</sup>*3(*x*) *are the leading terms in the approximations of the sample correlation coefficient <sup>R</sup>*∗∗*Nn*(*r*) *of two Gaussian vectors with negative binomial distributed random dimension Nn*(*r*) *and the angle <sup>θ</sup>*∗∗*Nn*(*r*)*between these vectors.*

**Remark 18.** *In Goldfeld and Quandt [34] and Missiakoulis [35] Sargans densities lr*(*x*) *and distribution functions Lr*(*x*) *for arbitrary integer r* = 1, 2, 3, ... *have been studied as an alternative to normal law in econometric models because they are computable in closed form, see also Kotz et al. [31], Section 4.4.3 and the references therein.*

**Figure 4.** Empirical version of P *g*<sup>−</sup>1/2 *n Nn*(*r*) *RNn*(*r*) ≤ *x*(blue line), limit Laplace law *Lr*(*x*) (orange line) and second approximation *Ln*;2(*<sup>x</sup>*; 1) (green line) for the correlation coefficient for pairs of normal vectors with random dimension *<sup>N</sup>*25(2). Here, *x* > 0, *n* = 25 and *r* = 2.

*5.2. The Random Dimension Nn* = *Nn*(*s*) *Is the Maximum of n Independent Discrete Pareto Random Variables*

The random dimension *Nn*(*s*) has probability mass function (39): Since <sup>E</sup>*Nn*(*s*) = ∞ we choose *gn* = *n* and consider again the cases *γ* = 1/2, *γ* = 0 and *γ* = −1/2.

It follows from Theorems 1 and 2 and Proposition 2 that if limit distributions for P *<sup>g</sup>γnRNn*(*s*) ≤ *x*for *γ* ∈ {1/2, 0 − 1/2} exist, they are ∞0 <sup>Φ</sup>(*x y<sup>γ</sup>*)*dHs*(*y*) with densities given below in the proof of the corresponding theorems

$$\frac{s}{\sqrt{2\pi}\pi} \int\_0^\infty y^{-3/2} e^{-(x^2 y^2 / 2 + s/y)} dy = \begin{cases} l\_{1/\sqrt{s}}(\mathbf{x}) = \frac{\sqrt{2\cdot s}}{2} e^{-\sqrt{2\cdot s}|\mathbf{x}|}, & \gamma = 1/2, \\\ q(\mathbf{x}) = \frac{1}{\sqrt{2\cdot\pi}} e^{-x^2/2}, & \gamma = 0, \\\ s\_2^\*(\mathbf{x}; \sqrt{s}) = \frac{1}{2\sqrt{2\cdot s}} \left( 1 + \frac{\mathbf{x}^2}{2\cdot\mathbf{s}} \right)^{-3/2}, & \gamma = -1/2, \end{cases} \tag{52}$$

where *s*∗2 (*x*; √*s*) is the density of the *scaled Student's t-distribution S*∗2 (*x*; √*s*) *with 2 degrees of freedom*, see Definition B37 in Jackman [36], p.507. If *Z* has density *s*∗2 (*x*; √*s*) then *<sup>Z</sup>*/√*s* has a classic Student's *t*-distribution with two degrees of freedom.

## 5.2.1. Laplace Distribution

We start with the case *γ* = 1/2 in Theorems 1 and 2. Consider the statistics √*n RNn*(*s*) and √*n*(*<sup>θ</sup>Nn*(*s*) − *<sup>π</sup>*/2). The limit distribution is now the Laplace distribution

$$L\_{1/\sqrt{s}}(\mathbf{x}) = \frac{1}{2} + \frac{1}{2}\text{sign}(\mathbf{x})\left(1 - e^{-\sqrt{2s}|\mathbf{x}|}\right) \quad \text{with density} \quad l\_{1/\sqrt{s}}(\mathbf{x}) = \frac{\sqrt{2s}}{2} e^{-\sqrt{2s}|\mathbf{x}|}.$$

**Theorem 6.** *Let s* > 0 *and (39) be the probability mass function of the random dimension Nn* = *Nn*(*s*) *of the vectors under consideration. If the representation (42) for the statistic Rm and the inequality (32) with gn* = *n hold, then there exists a constant Cs such that for all n* ∈ N+

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( \sqrt{n} R\_{N\_n(s)} \le \mathbf{x} \right) - L\_{1/\sqrt{s};n}(\mathbf{x}; a) \right| \le C\_s n^{-2} \rho$$

*where*

$$L\_{1/\sqrt{s},\mathrm{pt}}(\mathbf{x};a) = L\_{1/\sqrt{s}}(\mathbf{x}) + \frac{l\_{1/\sqrt{s}}(\mathbf{x})}{8\sin} \left( a \, 2\, \mathrm{s} \, \mathrm{x}^3 - (4-s)\mathbf{x} \left( 1 + \sqrt{2s} \, |\, \mathbf{x}| \right) \right). \tag{53}$$

*Moreover, the scaled angle <sup>θ</sup>Nn*(*s*) *between the vectors* <sup>X</sup>*Nn*(*s*) *and* <sup>Y</sup>*Nn*(*s*) *allows the estimate*

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( \sqrt{n} \ (\theta\_{N\_n(s)} - \pi/2) \le x \right) - L\_{1/\sqrt{s};n}(x; 1/3) \right| \le C\_s n^{-2},$$

*where <sup>L</sup>*1/√*s*;*n*(*<sup>x</sup>*; 1/3) *is given in* (53) *with a* = 1/3*.*

#### 5.2.2. Standard Normal Distribution

Let *γ* = 0 in the Theorems 1 and 2 examine the statistics *<sup>R</sup>*<sup>∗</sup>*Nn*(*s*) and <sup>Θ</sup><sup>∗</sup>*Nn*(*s*) = 2*Nn*(*s*)(*<sup>θ</sup>Nn*(*s*) − *<sup>π</sup>*/2).

**Theorem 7.** *Let s* > 0 *and Nn* = *Nn*(*s*) *be the random vector dimension having probability mass function (39). If the representation (42) for the statistic Rm and the inequality (32) with gn* = *n hold, then there exists a constant Cs such that, for all n* ∈ N+

$$\sup\_{x} \left| \mathbb{P} \left( \sqrt{N\_n(s)} \, \Big| \, \mathcal{R}\_{N\_n(s)} \le x \right) - \Phi(x) - \frac{1}{4n} \varphi(x) \, \mathrm{s}^2 \left( x^3 - 5 \, x \right) \right| \le \mathbb{C}\_s \, n^{-2}, s \ge 0$$

*Moreover, the scaled angle <sup>θ</sup>*<sup>∗</sup>*Nn*(*s*)*between the vectors* <sup>X</sup>*Nn*(*s*) *and* <sup>Y</sup>*Nn*(*s*) *allows the estimate*

$$\sup\_{x} \left| \mathbb{P} \left( \sqrt{N\_n(s)} \left( \theta\_{N\_n(s)} - \pi/2 \right) \le x \right) - \Phi(x) - \frac{1}{4n} \varrho(x) s^2 \left( \frac{1}{3} x^3 - 5x \right) \right| \le C\_\varepsilon n^{-2}. \tag{1}$$

5.2.3. Scaled Student's *t*-Distribution

Finally, we use *γ* = −1/2 in Theorems 1 and 2 examining the statistics *n*<sup>−</sup>1/2 *Nn*(*s*) *RNn*(*s*) and *n* −1/2 *Nn*(*s*) (*<sup>θ</sup>Nn*(*s*) − *<sup>π</sup>*/2). The limit *Scaled Student's t-Distribution S*∗2 (*x*; √*s*) *with two degrees of freedom* is a scale mixture of the normal distribution with zero mean and mixing exponential distribution 1 − *<sup>e</sup>*<sup>−</sup>*sy*, *y* ≥ 0, and it is representable in a closed form, see (A15) below in the proof of Theorem 8:

$$\begin{aligned} \int\_0^\infty \Phi(\mathbf{x}/\sqrt{\eta})d e^{-s/y} &= \int\_0^\infty \Phi(\mathbf{x}/\sqrt{\eta})\mathbf{s}y^{-2}e^{-s/y}d\mathbf{y} = \int\_0^\infty \Phi(\mathbf{x}\sqrt{z})\mathbf{s}e^{-sz}dz \\ &= \int\_0^\infty \Phi(\mathbf{x}\sqrt{z})d(1-e^{-sz})dz = \frac{1}{2} + \frac{\mathbf{x}/\sqrt{s}}{2\sqrt{2}\sqrt{1+\mathbf{x}^2/(2s)}} = S\_2^\*(\mathbf{x}) \end{aligned}$$

**Theorem 8.** *Let s* > 0 *and Nn* = *Nn*(*s*) *be the random vector dimension having probability mass function (39). If the representation (42) for the statistic Rm and the inequality (32) with gn* = *n hold, then there exists a constant Cs such that for all n* ∈ N+

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( n^{-1/2} \, \mathrm{N}\_{\mathrm{n}}(\mathbf{s}) \, \mathrm{R}\_{\mathrm{N}\_{\mathrm{n}}(\mathbf{s})} \le \mathbf{x} \right) - \mathrm{S}\_{\mathrm{n};2}^{\*}(\mathbf{x};\mathbf{1}) \right| \le \mathrm{C}\_{\mathrm{r}} n^{-2} \,\tag{54}$$

*where*

$$S\_{n;2}^\*(\mathbf{x}; \sqrt{s}; a) = S\_2^\*(\mathbf{x}; \sqrt{s}) + \frac{(15a + 3s - 18)\mathbf{x}^3 - 6\mathbf{x}\,\mathbf{s}(6 - s)}{4\,\mathrm{n}\,(\mathbf{x}^2 + 2\,\mathrm{s})^2} \,\mathrm{s}\_2^\*(\mathbf{x}; \sqrt{s})\tag{55}$$

*Moreover, the scaled angle <sup>θ</sup>*<sup>∗</sup>*Nn*(*s*)*between the vectors* <sup>X</sup>*Nn*(*s*) *and* <sup>Y</sup>*Nn*(*s*) *allows the estimate*

$$\sup\_{\mathbf{x}} \left| \mathbb{P} \left( n^{-1/2} N\_n(\mathbf{s}) \, \theta\_{N\_n(\mathbf{s})} \le \mathbf{x} \right) - \mathbf{S}\_{n;2}(\mathbf{x}; \sqrt{s}; 1/3) \right| \le \mathbf{C}\_s n^{-2},$$

*where Sn*;2(*<sup>x</sup>*; √*s*; 1/3) *is given in* (55) *with a* = 1/3*.*

> Theorems 3 to 8 are proved in Appendix A.3.
