Next Article in Journal
The 2-D Cluster Variation Method: Topography Illustrations and Their Enthalpy Parameter Correlations
Next Article in Special Issue
New Estimators of the Bayes Factor for Models with High-Dimensional Parameter and/or Latent Variable Spaces
Previous Article in Journal
An Evolution Based on Various Energy Strategies
Previous Article in Special Issue
Phylogenetic Curved Optimal Regression for Adaptive Trait Evolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Analysis of Finite Populations under Simple Random Sampling

by
Manuel Mendoza
1,*,
Alberto Contreras-Cristán
2 and
Eduardo Gutiérrez-Peña
2
1
Departamento de Estadística, Instituto Tecnológico Autónomo de México, Río Hondo 1, Ciudad de México 01080, Mexico
2
Departamento de Probabilidad y Estadística, Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas, Universidad Nacional Autónoma de México, Apartado Postal 20-126, Ciudad de México 01000, Mexico
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(3), 318; https://doi.org/10.3390/e23030318
Submission received: 12 December 2020 / Revised: 2 March 2021 / Accepted: 3 March 2021 / Published: 8 March 2021
(This article belongs to the Special Issue Bayesian Inference and Computation)

Abstract

:
Statistical methods to produce inferences based on samples from finite populations have been available for at least 70 years. Topics such as Survey Sampling and Sampling Theory have become part of the mainstream of the statistical methodology. A wide variety of sampling schemes as well as estimators are now part of the statistical folklore. On the other hand, while the Bayesian approach is now a well-established paradigm with implications in almost every field of the statistical arena, there does not seem to exist a conventional procedure—able to deal with both continuous and discrete variables—that can be used as a kind of default for Bayesian survey sampling, even in the simple random sampling case. In this paper, the Bayesian analysis of samples from finite populations is discussed, its relationship with the notion of superpopulation is reviewed, and a nonparametric approach is proposed. Our proposal can produce inferences for population quantiles and similar quantities of interest in the same way as for population means and totals. Moreover, it can provide results relatively quickly, which may prove crucial in certain contexts such as the analysis of quick counts in electoral settings.

1. Introduction

Survey sampling is one of the most popular areas of Applied Statistics. Neyman (1934) [1] established the methodological foundations for statistical inference based on random samples obtained from finite populations. This approach, known as design-based inference, has become a standard mostly due to its early adoption by international agencies for gathering and analyzing data from different countries.
The general spirit of these methods is nonparametric as no assumptions are made regarding the distribution of the variable of interest. Given a population of size N and a variable of interest X, inferences focus on some attributes of the set X = { X 1 , , X N } . Special interest is given to the population total, T = i = 1 N X i , and the population mean, X ¯ ( N ) = T / N . For the simpler designs, point-wise estimates based on a sample of size n, { x i 1 , , x i n } , are least squares optimal and are often unbiased. Here, { i 1 , , i n } { 1 , , N } denotes the set of indices corresponding to the n sampled units within the population. Interval estimation relies on the asymptotic normality of the corresponding sampling distribution.
These techniques are widely applicable and extraordinarily useful. On the other hand, however, the results are not necessarily optimal in a general statistical sense (see Godambe (1955) [2] for an early discussion on this topic) and inferences might fail if asymptotic normality does not hold.
The application of Bayesian ideas to survey sampling has been explored for some time now, and falls into the realm of model-based inference (Little (2004) [3]). Pearson (1928) [4] applied inverse probability theory to produce inferences on the proportion of the elements of a finite population satisfying a given condition. His assumptions that the population is finite and that the sampling is without replacement lead to a unique, well-defined model, namely, hypergeometric; thus, his analysis is fully parametric. Pearson produced a posterior distribution for the unknown proportion under the assumption of a uniform prior and showed that, in general, it is not compatible with the frequentist results based on asymptotic normality.
Aggarwal (1959) [5] formulated the estimation of the mean of a finite population as a decision problem and proved that under both simple random sampling and stratified sampling, point-wise classical results can be recovered as a limit of the Bayesian solution when a quadratic loss function is used and the prior variance goes to infinity. In a similar fashion, Godambe (1965) [6] discussed a class of admissible classical estimates and proved in passing that they can also be seen as Bayesian estimates, whereas Aggarwal (1966) [7] generalized his previous results to the case of two-stage sampling.
On the other hand, Godambe (1966) [8], using a quadratic loss function, found a Bayesian estimate for the population total which depends on the prior distribution only through the predictive expected value of X for the nonsampled units. Specifically, if a sample of size n is obtained whose elements have indexes in a set S { 1 , , N } , then T ^ = i S x i + i S E ( X i ) , where E ( X ) stands for the corresponding prior predictive mean.
Ericson (1969) [9] assumed X are obtained from a sequence of exchangeable random variables; therefore, their joint distribution can be represented as p ( x ) = i = 1 N p ( x i | θ ) p ( θ ) d θ . This representation motivated the idea of an infinite superpopulation from which X can be considered a set of i.i.d. observations. The superpopulation is described through the parametric model p ( x | θ ) , and the remaining uncertainty regarding the parameter θ is taken into account via a prior p ( θ ) . If the selection probability for each unit does not depend on the value of X (a noninformative sampling scheme), the sample { x i 1 , , x i n } can also be regarded as a set of i.i.d. observations from the superpopulation.
This framework provides a setting where inferences on the population total T = i S x i + i S X i can be obtained by noting that, upon observing the sample, T o b s = i S x i is a known constant, whereas T n o b s = i S X i can be described through its corresponding posterior predictive distribution. Ericson derived the results for the case of a continuous variable, assuming a normal distribution for the superpopulation. He proved that the asymptotic classical results can be obtained as the prior becomes noninformative. His approach can be replicated with any other model p ( x | θ ) , although he recognized that adoption of a specific model for the superpopulation is a restrictive assumption. In the case where the support of X is finite and completely known, he proposed a multinomial model for the superpopulation and, via a (Dirichlet) conjugate analysis, obtained a robust solution which recovers Pearson’s results for the binary case. This approach, however, is not general as it requires the support of X to be known and finite.
In the case of complex survey designs, both parametric and nonparametric Bayesian approaches have been proposed. Fox example, Si et al. (2015) [10] use a hierarchical approach in which they model the distribution of the weights of the nonsampled units in the population and simultaneously include them as predictors in a nonparametric Gaussian-process regression model. More recently, Rahnamay Kordasiabi and Khazaei (2020) [11] study a Bayesian nonparametric model, based on a Dirichlet process prior, that considers the inverse-probability weights as the only available information. Their model is essentially a Bayesian nonparametric mixture of regression models for the survey outcomes with the weights as predictors. Unfortunately, as in the frequentist approach, there are several ways of incorporating the inclusion probabilities into the analysis, and it is not always clear which of the available methods is the most adequate in a given situation, see, for example, Gelman (2007) [12]. Si, et al. (2020) [13] combine Bayesian prediction and weighting to provide a unified approach to survey inference.
The above notwithstanding, even in the simple random sampling case there does not seem to exist a conventional procedure that can be used as a kind of default for Bayesian survey sampling. In this paper, we discuss a flexible Bayesian nonparametric approach to the analysis of samples from finite populations in the context of simple random sampling, which can be easily generalized to the case of stratified sampling schemes. Our proposal can deal with both continuous and discrete variables, and can produce inferences for population quantiles and similar quantities of interest in the same way as for population means and totals. Moreover, it can provide results relatively quickly, which may prove crucial in certain contexts such as the analysis of quick counts.
The layout of the paper is as follows. In the next section, we make the case for the use of nonparametric, robust models. We then present our proposal in Section 3. Details of the implementation as well as some illustrative examples concerning some basic sampling schemes are given in Section 4. Finally, Section 5 contains some concluding remarks.

2. The Search for Robustness

Several authors have followed the Bayesian approach to analyze complex survey designs using specific superpopulation models (see Royall and Pfeffermann (1982) [14] and Treder and Sedransk (1996) [15], for example). Other proposals have tried to robustify the analysis with respect to the choice of the superpopulation distribution. Binder (1982) [16] replaced the Dirichlet prior distribution of Ericson with a Dirichlet prior process and removed the requirement on the support of X. However, he only offered an asymptotic approximation to the complete posterior distribution of T, which happens to be normal. The results coincide with their classical counterpart.
Lo (1988) [17] proposed a bootstrap mechanism, based on a Pólya urn scheme, to simulate copies of the entire population from which estimates of the finite population distribution and of some parameters of interest can be obtained. The underlying distribution is a Dirichlet-multinomial process, which can be interpreted as an approximation to a posterior distribution when a “noninformative” prior process is adopted. In that paper, the posterior expected value for the population mean was calculated and shown to coincide with the classical estimator. Conditions for asymptotic normality are established and the classical interval estimates are also reproduced.
In Lazar, et al. (2008) [18], the idea of generating “copies” of the entire finite population using the Pólya urn for a fixed sample size was pursued. There, the main interest was no longer on asymptotic results. Instead, the population parameters were computed for each copy, and thus the relevant posterior distribution was approximated via simulation. Point-wise as well as interval estimates can then be computed for any sample size. This procedure involves, however, a discrete predictive distribution whose support is restricted to that of the observed sample. To get around this constraint, Martínez-Ovando, et al. (2014) [19] proposed a nonparametric approach which allows one to incorporate any prior information available. The uncertainty about the population distribution of the (continuous) variable of interest was described through a species sampling model. The authors obtained an expression for the marginal predictive distribution of each unsampled unit and exhibited the posterior mean of the population total. They also argued that the complete posterior distribution for the population total can be obtained via a convolution of the individual predictive distributions. They illustrated their approach with simple random sampling, as well as with strata and unplanned domains. This approach is somewhat limited as the underlying variable is assumed to be continuous. This is a major issue, as in many applications the variable of interest is discrete.
More recently, as pointed out in the previous section, other authors have dealt with complex survey designs from a nonparametric Bayesian perspective. In that setting, one of the main issues is how to take into account the sampling weights of the design. Among other ideas, some authors have used the weights to build a pseudo-likelihood function (see Savitsky and Toth (2016) [20], for example), whereas others have fitted a regression model as part of the inference process (Rahnamay Kordasiabi and Khazaei (2020) [11]).
Our approach is based on a Bayesian nonparametric Dirichlet process mixture (DPM) model. Such models are extremely flexible; a large class of population distributions can in fact be consistently estimated using this kind of model (Ghosal, et al. (1999) [21]; Ghosh and Ramamoorthi (2003) [22]). Specifically, if the true density is in the Kullback–Leibler support of the DPM prior, then the corresponding posterior is consistent.

3. Our Proposal

From a general perspective, inferences on the population total, as well as on any other population parameter Q ( X ) , can be obtained by recalling that Q is a function of the complete set of values X = { X 1 , , X N } , which can then be regarded as a set of unknown parameters. Thus, a prior distribution must then be chosen to describe the knowledge concerning X , and the sample must be used to update this knowledge and get the corresponding posterior distribution. Once this distribution is available, the posterior for Q must be obtained. The exchangeability assumption for X = { X 1 , , X N } leads to the use of a (parametric) model p ( x | θ ) for the population distribution. In this setting, the random variables X 1 , , X N are conditionally i.i.d. given θ according to this model, which would correspond to the superpopulation distribution.
As discussed above, this parametric structure may be simple and the analysis leading to the predictive distribution of Q ( X ) may be straightforward, but it has also been shown to be rather restrictive as a proposal for a general Bayesian survey sampling methodology as it depends on the specific superpopulation model that is used (see Sedransk (2008) [23], for example). Some authors have relaxed the parametric assumption and assumed instead that the population distribution belongs to some nonparametric family, see, for example, Rahnamay Kordasiabi and Khazaei (2020) [11], who assume that the underlying variable of interest is continuous.
Moving to a nonparametric framework, let us assume that X j | θ j p ( x | θ j ) , where { θ j } are conditionally independent and such that θ j | G G . A widely used nonparametric family of prior distributions for G is that corresponding to the Dirichlet Process (Ferguson (1973) [24]). If a random probability measure G follows a Dirichlet Process with precision parameter α and base probability measure G 0 —denoted G D P ( α , G 0 ) —then, according to Sethuraman (1994) [25], it can be represented as
G ( · ) = j = 1 ω j δ θ j ( · ) ,
where θ 1 , θ 2 , are independent and with distribution G 0 , and the weights { ω j } j are obtained from a sequence { υ j } of independent and identically distributed Beta ( 1 , α ) random variables by setting ω 1 = υ 1 and ω s j = υ j l < j ( 1 υ l ) , for j = 2 , 3 , . In the stick breaking representation (1), δ θ denotes the measure with a point mass of 1 at θ .
Adopting a Dirichlet Process as the prior distribution for G leads us to the DPM model for the superpopulation distribution:
f ( x | G ) = p ( x | θ ) d G ( θ ) .
Note that, using (1), we can write
f ( x | G ) = j = 1 ω j p ( x | θ j ) .
Walker (2007) [26] provides a slice sampler MCMC algorithm which uses latent variables to produce posterior samples from (2). In this sampling scheme, the assumption that the random measure G has been integrated out is not required, and therefore the Pólya Urn scheme is not used. Kalli, et al. (2011) [27] suggest some efficiency improvements to the algorithm in Walker (2007) [26].
Here, we propose a strategy that provides flexible approximations to the target predictive distribution. In addition, we look for models for which the required unsampled units can be easily simulated without any restriction on their values. Moreover, our proposal is able to cope with both the continuous and the discrete cases.

3.1. Continuous Variables

In the continuous case, we use a Gaussian Dirichlet process mixture DPM model for the unknown density function of the variable of interest in (2). Based on the experience with this model, a reasonable level of flexibility can be expected when using it to approximate continuous but otherwise arbitrary densities (see Escobar and West (1995) [28]). The analysis of this DPM model produces a posterior distribution over the set of continuous densities on the real line, and the predictive distribution for a single observation corresponds to the expectation of that posterior distribution. For our purposes, however, we need a sample of size M, say, from the joint posterior predictive distribution of the unsampled units. This is a multivariate distribution of dimension N n whose components are not independent. Fortunately, this sample can be easily obtained if we independently simulate M densities from the posterior distribution mentioned above (each of these densities will be a known mixture of distributions) and then simulate (also independently) each unsampled unit from this univariate model. The slice sampler algorithm in Kalli, et al. (2011) [27] is used to follow these steps. See Appendix A for details concerning the MCMC algorithm used in this paper.

3.2. Discrete Variables

The case of discrete variables is, by far, more challenging from a parametric point of view. The use of Poisson and related models, as well as various mixtures thereof, has proved to be insufficient to describe the population distribution in a variety of applications. However, in a recent contribution, Canale and Dunson (2011) [29] developed a procedure to estimate the probability function of a discrete random variable that parallels that of Escobar and West (1995) [28]. Given a sample of a discrete random variable { X j } , they introduce a collection of continuous latent variables { Y j } from which the discrete variables are obtained via a rounding process. The density of these latent variables f ( y | G ) is then modeled using a Gaussian DPM model as in (2), and the probability mass function p ( x | G ) of interest is obtained by integration over the rounding intervals which relate the latent variable and the discrete data, namely, p ( x = j | G ) = a j a j + 1 f ( y | G ) d y , where a j and a j + 1 are thresholds lying on a suitable grid. Thus, this approach takes advantage of the flexibility of the continuous model in order to deal with the discrete case. The algorithm of Canale and Dunson (2011) [29] produces a nonparametric posterior distribution over the set of discrete distributions. As in the continuous case, we use this mechanism as a building block to simulate samples from the predictive distribution which corresponds to the expectation of this posterior distribution.
At any rate, after the simulation process we get M copies of the unsampled units on the population which, together with the units in the sample, provide M copies of the entire population. For each copy, the parameter of interest Q ( X ) can be computed, and thus a sample of size M from the predictive distribution of Q will be available for inference purposes. This procedure applies to any parameter, not just for the usual population mean or total. In the particular case of the total, it suffices to get the predictive distribution using only the unsampled units; the observed part can be later added as a constant. This is not the case for other nonlinear parameters such as the population quantiles. Nevertheless, the algorithm is general and can be used to produce inferences for any parameter. In the following section, several simulated examples as well as a real-data example are analyzed to show the type of results that can be obtained with this procedure when estimating the total, a quantile, and a ratio-type parameter of a finite population.

4. Implementation and Examples

As explained in the previous section, we propose to use a Bayesian nonparametric model as a prior for the distribution of θ in the population distribution p ( x | θ ) . Here, we follow Escobar and West (1995) [28] and use a Dirichlet Process for this purpose. Specifically, we assume that, conditionally on G , the distribution of θ = ( μ , V ) follows a Dirichlet Process, where the base measure G 0 is given by a normal-inverse gamma distribution. The corresponding density is given by
g 0 ( μ , V | m , τ , s , S ) = N ( μ | m , τ V ) × InverseGamma ( V | s / 2 , S / 2 ) ,
with hyperparameters m , τ , s and S , where τ is a positive scale factor and the Inverse-Gamma density has shape s / 2 and scale S .
Let N ( x | μ , V ) denote a Normal distribution with mean μ and variance V. If p ( x | θ ) = N ( x | μ , V ) , then X follows a Gaussian DPM. In order to obtain posterior samples via a Gibbs Sampler, we use a more recent algorithm than that of Escobar and West (1995) [28]. This algorithm is described in Kalli, et al. (2011) [27] and lies within the framework of slice sampler methods (e.g., Walker (2007) [26]. Because the Bayesian analysis for the Gaussian DPM produces a posterior distribution over the set of continuous densities on the real line, we can conceptualize posterior samples as realizations of a random continuous probability density. An important feature of the algorithms proposed by Walker (2007) [26] and Kalli, et al. (2011) [27] is that at each loop of the Gibbs sampler we have a posterior sample of this random density.
In the continuous case, once the convergence of the MCMC algorithm has been attained, at iteration i , the algorithm in Kalli, et al. (2011) [27] provides us with posterior samples K i , ( π 1 , μ 1 , V 1 ) , , ( π i , μ K i , V K i ) , which in turn define the posterior probability density
f ( x ) = π 1 N ( x | μ 1 , V 1 ) + + π K i N ( x | μ K i , V K i ) .
Samples from this density can be used to approximate the predictive distribution of Q ( X ) . With this in mind, we simulate the unobserved part of the population by (independently) sampling X n + 1 , , X N from (3) and then, with these values and the observations D = { x 1 , , x n } , compute Q ( X ) . We then repeat these steps for iterations j = i + 1 , , i + M , to obtain a sample of size M from the predictive distribution of Q ( X ) . In the case of a discrete variable, the simulated continuous variable is treated as latent and the observations of the variable of interest are obtained through the rounding process described in Section 3.2.

4.1. Simulated Examples

In each of the examples discussed in this section, we used the following values for the hyperparameters of the prior distribution: s = 4 , S = 2 , τ = 10,000, and m = x ¯ n , where x ¯ n is the sample mean.

4.1.1. Continuous Case

A collection of N = 6000 independent observations of a random variable X were generated according to the mixture f ( x ) = 0.26 N ( x | 10 , 9 ) + 0.10 N ( x | 7 , 4 ) + 0.30 N ( x | 13 , 16 ) + 0.24 N ( x | 22 , 6.25 ) + 0.10 N ( x | 30 , 16 ) . This data set will be our finite population. Its total is T = 92,716.01. From this population, a random sample of size n = 120 was obtained. From this sample, we get the posterior predictive distribution for Q ( X ) = T shown in Figure 1 (right panel). This corresponds to a simulation of M = 1000 copies of the entire population, and the total T was computed for each copy. A Bayesian estimate for T can be produced once we choose an appropriate loss function. In particular, if we use a quadratic loss function, the estimate is given by E ( T | D ) . In this case, we have E ( T | D ) ^ = 94,202.64. Moreover, a 0.95 posterior predictive probability interval is given by (84,559.05, 103,520.17). These results can be compared with those obtained via the classical procedures: T ^ = 93,914.97 and the (asymptotic) 95 % confidence interval: ( 85836.7 , 101993.2 ) . Figure 1 (right panel) shows the true total value T = 92,716.01 with a red dotted vertical line. Figure 1 (left panel) shows a simple approximation to the corresponding prior predictive distribution. This approximation was obtained using the same algorithm as for the posterior, but based on a minimal sample taken at random from the sample of size n. This shows how the sample contributes to the posterior predictive distribution of the parameter of interest.

4.1.2. Discrete Case

Similar to the continuous case, for this illustration we generated a set of N = 6000 i.i.d. observations from the mixture f ( x ) = 0.10 P o ( x | 6 ) + 0.15 P o ( x | 14 ) + 0.20 P o ( x | 18 ) + 0.10 P o ( x | 22 ) + 0.25 P o ( x | 30 ) + 0.20 P o ( x | 32 ) , where Po ( x | λ ) denotes the Poisson distribution with mean λ . This data set will take the role of our finite population. The corresponding total is given by T = 135,692. From this population, we obtained a random sample of size n = 30 and used our proposal to estimate T. The posterior predictive distribution of T is shown in Figure 2. As before, this corresponds to a simulation of M = 1000 copies of the entire population. Again, a Bayesian estimate for T , corresponding to a quadratic loss function is given by E ( T | D ) ^ = 127,890.5, and a 0.95 posterior predictive probability interval is given by ( 109 , 463 , 146 , 872 ) . The classical analysis leads to T ^ = 128,200 with a 95 % (asymptotic) confidence interval given by (107,047, 149,353). Figure 2 shows the true total value T = 135,692 with a red dotted vertical line. Both methods, Bayesian and classical, yield similar results.

4.2. A Real-Data Example

In many countries, the final results of the elections for president and governors are only released a few days after the election day. In order to mitigate the uncertainty that this delay may bring, both civil organizations and the media produce statistical estimates of the results, which are announced the very night of the election. In Mexico, the authority in charge of the electoral processes—the Instituto Nacional Electoral (INE)—produces its own estimates under the name of “INE Quick Count”.
To illustrate our proposal, we used a database with results for each one of the 7463 poll stations that the INE installed on 1 July 2018, for the Governor election in the State of Guanajuato (https://ieeg.mx/computos-finales/, accessed on 13 November 2020). A total number of 2,281,115 votes were recorded that day, and the election was won by Diego Sinhue Rodríguez Vallejo (DSRV) with 1,140,133 votes, 49.98 % of the votes cast.
A random sample of 500 polling stations was chosen to produce our inferences. This is the same size the INE used for its Quick Count in 2018. There were other candidates in that election, but for the sake of simplicity, here we focus on the estimation of the proportion of votes obtained by DSRV. To this end, we defined two variables for each polling station: X, the number of votes in favor of DSRV, and Y, the total number of votes recorded. We need to estimate the population total of the variable X ( T X ), the population total of the variable Y ( T Y ), and the parameter of interest P = T X / T Y .
Figure 3 shows the posterior predictive distribution of T X . We observe that this distribution is unimodal and asymmetric with a heavy tail on the right. The 0.95 probability interval is given by [1,080,506, 1,150,908], and it captures the true value 1,140,133. In the case of T Y , the posterior predictive distribution is shown in Figure 4. This distribution is also unimodal and looks more symmetric; the 0.95 probability interval results as [2,182,122, 2,296,042]; it also captures the true value, 2,281,115. With the two simulated samples, { T i X : i = 1 , , M } and { T i Y : i = 1 , , M } , we obtained another sample { P i : i = 1 , , M } where P i = T i X / T i Y for i = 1 , , M . The approximation to the posterior distribution of the parameter of interest P based on this sample appears in Figure 5. The distribution is unimodal and fairly symmetric, and the 0 . 95 probability interval is given by [ 0.4777 , 0.5186 ] . This interval captures rather well the true voting proportion, 0.4998 . If these results were used to announce the outcome of the election the very night of the election day, the Quick Count would report that, with probability 0.95, the percentage of votes in favor of DSRV lies between 47.77 % and 51.86 % or, equivalently, that the percentage of votes in favor of DSRV lies in the interval 49.815 % ± 2.045 % .

4.3. Larger Population Sizes

To obtain an approximation which is close to the predictive distribution of Q ( X ) , M may need to be in the range of thousands. On the other hand, in many real-life applications the population size N can be in units of tens or hundreds of thousands; in some cases, it could even reach millions. Thus, a loop implementing computations leading to the estimation of Q ( X ) , as described after Equation (3), could easily require millions of iterations, making the procedure computationally unfeasible if results are required in a short period of time. Reducing the size of M will reduce the accuracy and precision of the approximation to the predictive density for Q ( X ) . However, a more important issue is that the impact of taking smaller values of M may be insignificant if the value of N is sufficiently large. To get around this problem, we shall focus on a procedure to compute the simulated values of Q ( X ) , which uses only a fraction of the unobserved part of the population. Specifically, in this section we will discuss how to take advantage of asymptotic theory in order to use only a rather small fraction of the unobserved population to approximate the value of Q ( X ) .
For the sake of illustration, let us consider the case where the parameter Q ( X ) is the population total Q ( X ) = i = 1 N X i , and let
X ¯ ( N ) = Q ( X ) / N = 1 N i = 1 N X i ,
be the population mean. This mean can be written as
X ¯ ( N ) = 1 N { x 1 + + x n + i = n + 1 N X i } ,
where the set { X n + 1 , , X N } represents the unobserved part of the population. As described before, computing a predictive value of X ¯ ( N ) can be very expensive, depending on the magnitude of M and, more dramatically, on the magnitude of N . Note that by defining x ¯ ( n ) = 1 n i = 1 n x i and X ¯ ( N n ) = 1 N n j = 1 N n X n + j , we can write
X ¯ ( N ) = 1 N n x ¯ ( n ) + ( N n ) X ¯ ( N n ) .
We propose to approximate X ¯ ( N n ) with X ¯ ( R n ) = 1 R n j = 1 R n X n + j , where R is an integer substantially smaller than N . Then, we will be using the approximation
X ˜ N = 1 N { n x ¯ ( n ) + ( N n ) X ¯ ( R n ) } ,
for the predictive value of X ¯ ( N ) . This requires us to simulate R n (instead of N n ) values from (3). As for the value of R , a simple argument based on the Central Limit Theorem can be used to propose a convenient value of R for the approximation.
For a large enough value of R , we have
X ¯ ( R n ) N X ¯ ( N ) , σ 2 / ( R n ) ,
where X ¯ ( N ) is given by (4) and σ 2 = VAR ( X ) . These are the mean and the variance of the variable of interest X in the finite population of size N . Thus, for α ( 0 , 1 ) ,
P | X ¯ ( R n ) X ¯ ( N ) | σ R n Z ( 1 α / 2 ) = 1 α
or, in terms of the relative error,
P | X ¯ ( R n ) X ¯ ( N ) | X ¯ ( N ) σ X ¯ ( N ) Z ( 1 α / 2 ) R n = 1 α ,
where we assume X ¯ ( N ) > 0 . Now assume this relative error is not larger than ϵ . Then, we must use R such that
σ X ¯ ( N ) Z ( 1 α / 2 ) R n = ϵ ,
leading to
R n = σ X ¯ ( N ) 2 Z ( 1 α / 2 ) ϵ 2 .
Finally, using the observed data, we can estimate X ¯ ( N ) and σ 2 by x ¯ ( n ) and s ( n ) 2 = 1 n i = 1 n ( x i x ¯ ( n ) ) 2 , respectively, so that
R n + Z ( 1 α / 2 ) ϵ 2 × s ( n ) x ¯ ( n ) 2 .

A Numerical Example

A population of size N = 100,000 was simulated from the model in Section 4.1.1, resulting in the population total of T = 1,548,577. From these simulated values we take a sample of n = 1000 observations. For these data we have x ¯ ( n ) = 15.724 , s ( n ) = 7.908 . If, for example, we fix α = 0.05 and ϵ = 0.05 , we get
R = n + 1.96 0.05 2 × 7.908 15.24 2 = n + 39.2 2 × 0.5029663 2 1000 + 1537 × 0.252 = 1388 .
Thus, in this case we set R = 1400 . Following the argument of the previous section, once the convergence of the MCMC algorithm has been attained, at each loop we can simulate R n = 1400 1000 = 400 observations from the density (3) in order to approximate X ¯ ( N n ) in Equation (5). The reader may want to compare this with the task of simulating N n = 99,000 observations at each loop, assuming that this procedure is repeated M = 1000 times, say.
With the aim of studying the effect of the magnitude of R in the estimates, we shall perform a brief simulation study using different values of R. To this end, for each fixed value of R we compute 10 different estimates of the population total T. Let us denote these by T ^ 1 ( R ) , , T ^ 10 ( R ) ; Figure 6 reports the root mean squared error (RMSE), 1 10 t = 1 10 ( T ^ t ( R ) T ) 2 , for different values of R as indicated in the horizontal axis. The blue horizontal line corresponds to the value of the RMSE 1 10 t = 1 10 ( T ^ t ( 100 , 000 ) T ) 2 , obtained when R = N = 100,000, Note that, as all these RMSEs are not larger than 2500, then relative to the magnitude of the population total, the error in all cases is bounded by 25 , 000 T = 0.016 , i.e. 1.6 % . Additionally, for each value of R in the horizontal axis, Figure 7 shows the sample standard deviation 1 9 t = 1 10 ( T ^ t ( R ) T ¯ ( 10 ) ) 2 obtained from T ^ 1 ( R ) , , T ^ 10 ( R ) , T ¯ 10 = 1 10 t = 1 10 T ^ t ( R ) , again as a percentage of the population total. These deviations are bounded by 3000 T = 0.0019 × 100 % , i.e., 0.19 % . Figure 8 compares two simulated posterior predictive distributions, obtained using R = 1400 (upper panel) and R = 100,000 (lower panel). As expected, the predictive distribution with the smaller value of R involves a larger amount of uncertainty. However, as the previous calculations show, this increase has a rather small relative effect on the value of the point-wise estimate. A similar conclusion is reached if the length of the probability intervals is also measured in relative terms.
We shall now use these simulated data to illustrate the estimation of Q ( X ) given by the 95 % quantile of the population values. Following the scheme described at the beginning of Section 4, at each loop of the Gibbs sampler we simulate the unobserved part of the population and use these values together with the sample D to compute the 95 % quantile, so we end up with a sample of M = 1000 simulated values of Q ( X ) . The upper left panel in Figure 9 shows the model considered as the superpopulation model (this is described in Section 4.1.1). The upper right panel shows the finite population of N = 100,000 samples from the superpopulation model, there a (dotted) vertical red line shows the true 95 % quantile of the finite population, which corresponds to a value of X ( 0.95 ) = 30.12 . The lower left panel shows the sample D of size n = 1000 . The lower right panel shows our approximation to the predictive distribution of the 95 % quantile, and the blue (dotted) vertical lines show a 95 % predictive interval for X ( 0.95 ) . This interval is given by ( 29.5 , 32.3 ) and contains the true value. Again, if a quadratic loss function is used, the point-wise estimate for X ( 0.95 ) is given by E ( Q ( X ) | D ) ^ = 30.94 ; this is indicated in Figure 9 with a magenta dotted vertical line. In this example, all the unsampled values were simulated within the MCMC algorithm. In practice, however, a shortcut as that proposed for the means or totals can be explored to reduce the cost of the computation.
The approximate procedure to obtain inferences for the population mean and total is simple and feasible. With regard to other parameters, a similar asymptotic argument can be used to provide approximate inferences. In particular, in the case of quantiles, there are many contributions discussing the convergence of a sample quantile to its respective population counterpart. For example, the following result can be found in Section 2.3 of Serfling (1980) [30]. Let X be a random variable with density function f , for which the quantile of order q is given by x ( q ) , and assume f ( x ( q ) ) > 0 . Then, if x ( q : n ) is a sample quantile of order q from a random sample of size n of X , it follows that
n 1 / 2 f ( x ( q ) ) ( x ( q : n ) x ( q ) ) { q ( 1 q ) } 1 / 2 is AN ( 0 , 1 )
where AN stands for “asymptotically normal”.
Thus, for n large enough
P | x ( q : n ) x ( q ) | Z ( 1 α / 2 ) ( q ( 1 q ) ) 1 / 2 n 1 / 2 f ( x ( q ) ) = 1 α .
If we fix the error bound
Z ( 1 α / 2 ) ( q ( 1 q ) ) 1 / 2 n 1 / 2 f ( x ( q ) ) = ϵ
We get
n = Z ( 1 α / 2 ) 2 ( q ( 1 q ) ) ϵ 2 ( f ( x ( q ) ) ) 2 .
Here, all components are known with the exception of f ( x ( q ) ) . However, in our case this quantity can be approximated with f ˜ ( x ( q : n ) ) , where f ˜ is the predictive density provided by the simulation process. Alternatively, we can use f ˜ ( x ˜ ( q ) ) , where x ˜ ( q ) is the quantile of order q corresponding to the predictive density f ˜ . For other quantities of interest defined as functions of parameters for which asymptotic normality holds, the delta method could be used to provide the corresponding approximation rule.
We close this section with an example that illustrates how to produce joint inferences about the first and third population quartiles, X ( 0.25 ) and X ( 0.75 ) , respectively. We repeated the steps described after Equation (3) but, instead of the 95% quantile, at each iteration of the Gibbs sampler we computed Q ( X ) = ( X ( 0.25 ) , X ( 0.75 ) ) . Thus, we end up with a sample of size M = 1000 from the posterior predictive distribution of the vector Q ( X ) . Figure 10 shows a scatter plot of this sample; the red dot represents the true values of the population quartiles, X ( 0.25 ) = 9.35 and X ( 0.75 ) = 21.35 . The sample can also be used to produce an estimate of the joint density of X ( 0.25 ) and X ( 0.75 ) ; this is shown in Figure 11.
Finally, the same simulated sample can also be used to provide inferences on other relevant quantities such as the interquartile range X ( 0.75 ) X ( 0.25 ) . Figure 12 shows the posterior predictive distribution of the interquartile range. The vertical dotted red line shows the true value of the population interquartile range, which is 12 in this case.

5. Discussion

In recent years, a number of methods have been proposed to produce Bayesian inferences in the setting of survey sampling, with emphasis on the case of complex designs. One of the main differences among these contributions has to do with the way in which they incorporate the sampling weights into the analysis. Often, these procedures are nonparametric and rely on simulation techniques that render them uncompetitive, in terms of computing time, when compared with their frequentist counterparts. We would argue that there is no such thing as a standard Bayesian survey sampling method, even for the simpler designs. In this paper, we have focused on the simple random sampling case. The Bayesian survey sampling approach proposed here is nonparametric, and thus it avoids specific parametric assumptions that are often unwarranted. It can deal with both the continuous and the discrete cases in a unified manner, provides inferences for any parameter of the finite population of interest, and can be easily generalized to stratified sampling schemes. We propose that our approach be used as a kind of default procedure in the simple random sampling case. Moreover, we believe it can also be used as a benchmark for other proposals applicable to complex designs.
Admittedly, in its original formulation our approach can be as computationally expensive as the other methods. Fortunately, however, asymptotic results can be used to avoid the simulation of the entire unsampled part of the population. The modified procedure discussed in Section 4.3 is computationally efficient; this can be crucial in certain contexts, such as the analysis of quick counts in electoral settings. A limitation of this method, however, is that its applicability in the case of arbitrary parameters (other than the population mean, total, or quantiles) depends on the availability of suitable asymptotic results.
A proper comparison would involve a more thorough simulation exercise under a wide range of scenarios. Nevertheless, note that even in this limited first study, our proposal produces competitive results and provides the user with a set of samples from the predictive distribution of X that can be used to estimate any characteristic of interest in the population under study. An important instance arises when we are interested in a population quantile instead of the total of an attribute in the population. In such a case, commonly used frequentist procedures no longer provide a simple, automatic way to produce interval estimates except in the asymptotic scenario.

6. Materials and Methods

All of the examples of Section 4 were implemented using the R statistical language and environment (R Core Team, 2020 [31]).

Author Contributions

Conceptualization, M.M.; methodology, M.M. and E.G.-P.; software, A.C.-C.; validation, M.M., A.C.-C., and E.G.-P.; formal analysis, M.M.; investigation, M.M., A.C.-C., and E.G.-P.; resources, A.C.-C.; data curation, A.C.-C.; writing—original draft preparation, M.M.; writing—review and editing, E.G.-P.; visualization, A.C.-C.; supervision, M.M.; project administration, E.G.-P.; funding acquisition, M.M., A.C.-C., and E.G.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Grant IN106114-3 of the Programa de Apoyo a Proyectos de Investigación e Innovación Tecnológica, UNAM, Mexico.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset analyzed in Section 4.2 is publicly available at: https://ieeg.mx/computos-finales/ (accessed on 13 November 2020).

Acknowledgments

The first author wishes to thank Asociación Mexicana de Cultura, A.C. Partial support from the Sistema Nacional de Investigadores is also gratefully acknowledged.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DPMDirichlet Process Mixture
DSRVDiego Sinhue Rodríguez Vallejo
INEInstituto Nacional Electoral
MCMCMarkov Chain Monte Carlo

Appendix A

Consider the Sethuraman [25] representation for the Dirichlet process given in Section 3 which leads to Equation (2), where f ( x | θ ) = N ( x | μ , V ) . For the data { x 1 , , x n } , Walker (2007) [26] and Kalli, et al. (2011) [27] introduce latent variables { u i : i = 1 , , n } and { d i : i = 1 , , n } , such that 0 < u i < 1 , i = 1 , , n and d i is a finite integer, i = 1 , , n . Walker (2007) [26] states that the vector ( x i , u i , d i ) has joint density given by
f ( x i , u i , d i = k | G ) = 1 ( u i < ω k ) N ( x i | μ k , V k ) ,
with corresponding marginal f ( x i | G ) given in (2). For each i , the role of u i is to simplify (2) to a finite sum, with a number of terms given by the cardinality N i of the set A ( u i ) = { j : u i < ω j } , on the other hand, d i is a selector, whose value indicates which term in this finite sum generated the observation x i . A usual problem with slice sampler algorithms, is that posterior samples of the latent (augmented) variables feature high correlation implying slow mixing. Additionally, for the method proposed in Walker (2007) [26], when sampling u 1 , . . . , u n , the cardinality of the sets A ( u i ) may increase, making numerical computations slow. To overcome these problems, Kalli, et al. (2011) [27] suggest using a deterministic and decreasing sequence { ξ j } j = 1 , such that ξ j > 0 , j = 1 , 2 , , in order to define the joint density
f ( x i , u i , d i = k | G ) = 1 ( u i < ξ k ) ω k ξ k N ( x i | μ k , V k ) ,
which also has corresponding marginal for x i given by (2). For our implementation, we used one of the possibilities suggested by Kalli, et al. (2011) [27] for the sequence { ξ j } j = 1 , namely, ξ j = { 1 2 } j , j = 1 , 2 , .
For each i = 1 , , n , let us denote by N i the cardinality of the finite set { j : ξ j > u i } , the selector d i takes values in the set 1 , 2 , , N i . Kalli, et al. (2011) [27] state that the augmented likelihood corresponding to the a sample of size n , { x i , u i , d i } i = 1 n is
l ( { x i , u i , d i } i = 1 n | G ) = i = 1 n 1 ( u i < ξ d i ) w d i ξ d i N ( x i | μ d i , V d i ) .
Let us denote by g 0 ( μ , V | m , τ , s , S ) the Normal-Inverse gamma density with hyperparameters m , τ , s , S , the prior distribution for the sequences { θ j = ( μ j , V j ) } j and { υ j } j in Sethuraman representation for a Dirichlet Process is given by
(a)
{ ( μ j , V j ) } j = 1 is a sequence of independent random variables from g 0 ( μ , V | m τ , s , S ) .
(b)
{ υ j } j = 1 are independent and with distribution beta ( υ | 1 , α ) .
The augmented likelihood (A2) and the prior for the sequences { ( μ j , V j ) } j and { υ j } j given in (a) and (b) can be used with Bayes’ theorem to obtain the full conditional distributions for all unknown quantities. Note that as the introduction of the latent variable u i yields a finite sum and the selector d i takes values in the set { 1 , 2 , . . . , N i } , then at each loop of the Gibbs sampler, samples of the parameters ( μ j , V j ) and υ j are required only for j in the set { 1 , , N } , where N = max { N 1 , , N n } . Let D j denote the set { i : d i = j } and n j be the cardinality of D j , the full conditional distributions for the Gibbs sampler are
(fc1)
π ( μ j | ) = N μ j m τ + i D j x i 1 τ + n j ; τ V j 1 + τ n j .
(fc2)
π ( V j | ) = InverseGamma V j n j + s + 1 2 ; ( μ j m ) 2 2 τ + S 2 + i D j ( x i μ j ) 2 2 .
(fc3)
π ( υ j | ) = Beta ( υ j | 1 + n j ; α + i = 1 n 1 ( d i > j ) ) .
(fc4)
π ( u i | ) 1 ( 0 < u i < ξ d i ) .
(fc5)
π ( d i = k | ) 1 ( k : u i < ξ k ) w k ξ k N ( x i | μ k , V k ) .
We use the next algorithm to implement a Gibbs sampler based on the full conditional distributions above.

Algorithm

  • At the t-th step, assume we have sampled values
    { u i , t 1 } i = 1 n ; { d i , t 1 } i = 1 n ; { ( μ j , t 1 , V j , t 1 ) } j = 1 N ( t 1 ) ; { υ j , t 1 } j = 1 N ( t 1 ) ,
    where the number of components in the mixture is N ( t 1 ) .
  • Begin by obtaining samples u 1 , t , , u n , t , where each u i , t is sampled from a Uniform ( u | 0 , ξ d i , t 1 ) density, i = 1 , , n . Use these to obtain N i = max { l : ξ l > u i , t } , i = 1 , , n , then update the number of components N ( t ) = max { N 1 , , N n } .
  • To obtain { ( μ j , t , V j , t ) } j = 1 N ( t ) ; { υ j , t } j = 1 N ( t ) consider two cases:
    (i)
    If N ( t ) < = N ( t 1 ) , then for each j = 1 , 2 , , N ( t )
    (i.1)
    if n j > 0 , then replace ( μ j , t 1 , V j , t 1 ) by a sample ( μ j , t , V j , t ) from (fc1) and (fc2) above.
    (i.2)
    if n j = 0 , then replace ( μ j , t 1 , V j , t 1 ) by a sample ( μ j , t , V j , t ) from g 0 ( μ , V | m , τ , s , S ) .
    Then, obtain a sample υ j , t from (fc3). These steps define the updated set of parameters { ( μ j , t , V j , t ) } j = 1 N ( t ) ; { υ j , t } j = 1 N ( t ) .
    (ii)
    if N ( t ) > N ( t 1 ) , then for each j = 1 , 2 , , N ( t 1 ) produce a sample ( μ j , t , V j , t ) by repeating steps (i.1) or (i.2) depending on the value of n j . For each j = N ( t 1 ) + 1 , , N ( t ) , produce a sample ( μ j , t , V j , t ) from g 0 ( μ , V | m , τ , s , S ) . Then, for each j = 1 , 2 , , N ( t ) obtain a sample υ j , t from (fc3). These steps define the updated set of parameters { ( μ j , t , V j , t ) } j = 1 N ( t ) ; { υ j , t } j = 1 N ( t ) .
  • Use { υ j , t } j = 1 N ( t ) to compute { w j , t } j = 1 N ( t ) by using the relations w 1 = υ 1 , w j = υ j l < j ( 1 υ l ) , then update the selectors by sampling { d i , t } i = 1 n from the full conditional (fc5).

References

  1. Neyman, J. On the two different aspects of the representative method: The method of stratified sampling and the method of purposive selection. J. R. Stat. Soc. 1934, 97, 558–625. [Google Scholar] [CrossRef]
  2. Godambe, V.P. A unified theory of sampling from finite populations. J. R. Stat. Soc. Ser. B. 1995, 17, 269–278. [Google Scholar] [CrossRef]
  3. Little, R.J. To model or not to model? Competing modes of inference for finite population sampling. J. Am. Stat. Assoc. 2004, 99, 546–556. [Google Scholar] [CrossRef] [Green Version]
  4. Pearson, K. On a method of ascertaining limits to the actual number of marked members in a population of given size from a sample. Biometrika 1928, 20A, 149–174. [Google Scholar] [CrossRef]
  5. Aggarwal, O.P. Bayes and minimax procedures in sampling from finite and infinite populations–I. Ann. Math. Stat. 1959, 30, 206–218. [Google Scholar] [CrossRef]
  6. Godambe, V.P. Admissibility and Bayes estimation in sampling finite populations. I. Ann. Math. Stat. 1965, 36, 1707–1722. [Google Scholar] [CrossRef]
  7. Aggarwal, O.P. Bayes and minimax procedures for estimating the arithmetic mean of a population with two-stage sampling. Ann. Math. Stat. 1966, 37, 1186–1195. [Google Scholar] [CrossRef]
  8. Godambe, V.P. A new approach to sampling from finite populations. II. Distribution-free sufficiency. J. R. Stat. Soc. Ser. B 1996, 28, 320–328. [Google Scholar] [CrossRef]
  9. Ericson, W.A. A note on the posterior mean of a population mean. J. R. Stat. Soc. Ser. B 1969, 31, 332–334. [Google Scholar] [CrossRef]
  10. Si, Y.; Pillai, N.S.; Gelman, A. Bayesian nonparametric weighted sampling inference. Bayesian Anaysis 2015, 10, 605–625. [Google Scholar] [CrossRef] [Green Version]
  11. Rahnamay Kordasiabi, S.; Khazaei, S. Prediction of the nonsampled units in survey design with the finite population using Bayesian nonparametric mixture model. Commun. Stat. Simul. Comput. 2020, 21, 93–105. [Google Scholar] [CrossRef]
  12. Gelman, A. Struggles with survey weighting and regression modeling. Stat. Sci. 2007, 22, 153–164. [Google Scholar] [CrossRef] [Green Version]
  13. Si, Y.; Trangucci, R.; Graby, J.S.; Gelman, A. Bayesian hierarchical weighting adjustment and survey inference. Surv. Methodol. 2020, 46, 181–214. [Google Scholar]
  14. Royall, R.M.; Pfeffermann, D. Balanced samples and robust Bayesian inference in finite population sampling. Biometrika 1982, 69, 401–409. [Google Scholar] [CrossRef]
  15. Treder, R.P.; Sedransk, J. Bayesian sequential two-phase sampling. J. Am. Stat. Assoc. 1996, 91, 782–790. [Google Scholar] [CrossRef]
  16. Binder, D.A. Non-parametric Bayesian models for samples from finite populations. J. R. Stat. Soc. Ser. B 1982, 44, 388–393. [Google Scholar] [CrossRef]
  17. Lo, A.Y. A Bayesian bootstrap for a finite population. Ann. Stat. 1988, 16, 1684–1695. [Google Scholar] [CrossRef]
  18. Lazar, R.; Meeden, G.; Nelson, D. A noninformative Bayesian approach to finite population sampling using auxiliary variables. Surv. Methodol. 2008, 34, 51–64. [Google Scholar]
  19. Martínez-Ovando, J.C.; Olivares-Guzmán, S.I.; Roldán-Rodríguez, A. A Bayesian nonparametric framework to inference on totals of finite populations. In The Contribution of Young Researchers to Bayesian Statistics (Chapter 15); Lanzarone, E., Leva, E., Eds.; Springer: New York, NY, USA, 2014. [Google Scholar]
  20. Savitsky, T.D.; Toth, D. Bayesian estimation under informative sampling. Electron. J. Stat. 2016, 10, 1677–1708. [Google Scholar] [CrossRef]
  21. Ghosal, S.; Ghosh, J.K.; Ramamoorthi, R.V. Posterior consistency of Dirichlet mixtures in density estimation. Ann. Stat. 1999, 27, 143–158. [Google Scholar]
  22. Ghosh, J.K.; Ramamoorthi, R.V. Bayesian Nonparametrics; Springer: New York, NY, USA, 2003. [Google Scholar]
  23. Sedransk, J. Assessing the value of Bayesian methods for inference about finite population quantities. J. Off. Stat. 2008, 24, 495–506. [Google Scholar]
  24. Ferguson, T. Bayesian analysis of some nonparametric problems. Ann. Stat. 1973, 1, 209–230. [Google Scholar] [CrossRef]
  25. Sethuraman, J. A constructive definition of Dirichlet priors. Stat. Sin. 1994, 4, 639–650. [Google Scholar]
  26. Walker, S.G. Sampling the Dirichlet mixture model with slices. Commun. Stat. Simul. Comput. 2007, 36, 45–54. [Google Scholar] [CrossRef]
  27. Kalli, M.; Griffin, J.E.; Walker, S.G. Slice sampling mixture models. Stat. Comput. 2011, 21, 93–105. [Google Scholar] [CrossRef]
  28. Escobar, M.D.; West, M. Bayesian density estimation and inference using mixtures. J. Am. Stat. Assoc. 1955, 90, 577–588. [Google Scholar] [CrossRef]
  29. Canale, A.; Dunson, D.B. Bayesian kernel mixtures for counts. J. Am. Stat. Assoc. 2011, 106, 1528–1539. [Google Scholar] [CrossRef] [Green Version]
  30. Serfling, R. Approximation Theorems of Mathematical Statistics; Wiley: New York, NY, USA, 1980. [Google Scholar]
  31. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2020; Available online: https://www.R-project.org/ (accessed on 13 November 2020).
Figure 1. Prior (left panel) and posterior (right panel) predictive distribution of the population total. Continuous case.
Figure 1. Prior (left panel) and posterior (right panel) predictive distribution of the population total. Continuous case.
Entropy 23 00318 g001
Figure 2. Posterior predictive distribution of the population total. Discrete case.
Figure 2. Posterior predictive distribution of the population total. Discrete case.
Entropy 23 00318 g002
Figure 3. Posterior predictive distribution of the number of votes in favor of Diego Sinhue Rodríguez Vallejo (DSRV).
Figure 3. Posterior predictive distribution of the number of votes in favor of Diego Sinhue Rodríguez Vallejo (DSRV).
Entropy 23 00318 g003
Figure 4. Posterior predictive distribution of the total number of votes recorded.
Figure 4. Posterior predictive distribution of the total number of votes recorded.
Entropy 23 00318 g004
Figure 5. Posterior predictive distribution of P.
Figure 5. Posterior predictive distribution of P.
Entropy 23 00318 g005
Figure 6. Root mean squared errors for different values of R.
Figure 6. Root mean squared errors for different values of R.
Entropy 23 00318 g006
Figure 7. Sample standard deviation for different values of R.
Figure 7. Sample standard deviation for different values of R.
Entropy 23 00318 g007
Figure 8. Simulated posterior predictive distributions.
Figure 8. Simulated posterior predictive distributions.
Entropy 23 00318 g008
Figure 9. Superpopulation model, finite population, sample, and predictive distribution for the 95 % quantile.
Figure 9. Superpopulation model, finite population, sample, and predictive distribution for the 95 % quantile.
Entropy 23 00318 g009
Figure 10. Simulated values from the joint posterior predictive distribution of X ( 0.25 ) and X ( 0.75 ) .
Figure 10. Simulated values from the joint posterior predictive distribution of X ( 0.25 ) and X ( 0.75 ) .
Entropy 23 00318 g010
Figure 11. Joint posterior predictive density of Q ( X ) = ( X ( 0.25 ) , X ( 0.75 ) ) .
Figure 11. Joint posterior predictive density of Q ( X ) = ( X ( 0.25 ) , X ( 0.75 ) ) .
Entropy 23 00318 g011
Figure 12. Posterior predictive distribution of the interquartile range X ( 0.75 ) X ( 0.25 ) .
Figure 12. Posterior predictive distribution of the interquartile range X ( 0.75 ) X ( 0.25 ) .
Entropy 23 00318 g012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mendoza, M.; Contreras-Cristán, A.; Gutiérrez-Peña, E. Bayesian Analysis of Finite Populations under Simple Random Sampling. Entropy 2021, 23, 318. https://doi.org/10.3390/e23030318

AMA Style

Mendoza M, Contreras-Cristán A, Gutiérrez-Peña E. Bayesian Analysis of Finite Populations under Simple Random Sampling. Entropy. 2021; 23(3):318. https://doi.org/10.3390/e23030318

Chicago/Turabian Style

Mendoza, Manuel, Alberto Contreras-Cristán, and Eduardo Gutiérrez-Peña. 2021. "Bayesian Analysis of Finite Populations under Simple Random Sampling" Entropy 23, no. 3: 318. https://doi.org/10.3390/e23030318

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop