Next Article in Journal
On the Qualitative Behavior of Third-Order Differential Equations with a Neutral Term
Next Article in Special Issue
Spatial-Temporal 3D Residual Correlation Network for Urban Traffic Status Prediction
Previous Article in Journal
Some Properties Concerning the JL(X) and YJ(X) Which Related to Some Special Inscribed Triangles of Unit Ball
Previous Article in Special Issue
Bayesian Reference Analysis for the Generalized Normal Linear Regression Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimating the Variance of Estimator of the Latent Factor Linear Mixed Model Using Supplemented Expectation-Maximization Algorithm

1
Department of Statistics, IPB University, Bogor 16680, Indonesia
2
Faculty of Spatial Sciences, University of Groningen, 9747 Groningen, The Netherlands
3
Department of Statistics, University of Padjadjaran, Bandung 16426, Indonesia
*
Authors to whom correspondence should be addressed.
Symmetry 2021, 13(7), 1286; https://doi.org/10.3390/sym13071286
Submission received: 3 June 2021 / Revised: 1 July 2021 / Accepted: 15 July 2021 / Published: 17 July 2021
(This article belongs to the Special Issue Symmetry in Statistics and Data Science)

Abstract

:
This paper deals with symmetrical data that can be modelled based on Gaussian distribution, such as linear mixed models for longitudinal data. The latent factor linear mixed model (LFLMM) is a method generally used for analysing changes in high-dimensional longitudinal data. It is usual that the model estimates are based on the expectation-maximization (EM) algorithm, but unfortunately, the algorithm does not produce the standard errors of the regression coefficients, which then hampers testing procedures. To fill in the gap, the Supplemented EM (SEM) algorithm for the case of fixed variables is proposed in this paper. The computational aspects of the SEM algorithm have been investigated by means of simulation. We also calculate the variance matrix of beta using the second moment as a benchmark to compare with the asymptotic variance matrix of beta of SEM. Both the second moment and SEM produce symmetrical results, the variance estimates of beta are getting smaller when number of subjects in the simulation increases. In addition, the practical usefulness of this work was illustrated using real data on political attitudes and behaviour in Flanders-Belgium.

1. Introduction

The latent factor multivariate linear mixed model (LFLMM) is a combination between the Factor Analysis (FA) and the Linear Mixed Model (LMM), as proposed by [1]. The model aims to analyze longitudinal data sets with large numbers of multivariate responses, i.e., high-dimensional longitudinal data. The authors proposed estimation of the LFLMM by means of the EM algorithm, which is a closed-form solution. They showed by way of simulation that EM estimation of the LFLMM provides accurate parameter estimates and is more efficient in terms of adding variables other than time variables in the model than alternatives like the structural equation model. As shown by [2,3], the combination of fixed and random effects and the interaction of covariates with time can be straightforwardly handled by the LFLMM estimated by EM.
The LFLMM assumes that the responses are continuous and that the number of latent variables is known [1]. Moreover, convergence of the EM algorithm is sometimes slow. The main disadvantage, however, is that the EM algorithm does not produce standard errors of the estimator of the regression coefficients because it does not calculate the derivatives of the likelihood function, which are often complicated and tedious to derive [4,5]. Thus, it is difficult to study the effects of different covariates or fixed variables for different latent factors simultaneously.
The general Supplemented EM algorithm was proposed by [6] to obtain the standard errors by calculating the complete information matrix as the base of the variance-covariance matrix of the estimator. The Supplemented EM algorithm has been applied to various kinds of models, notably item response models [6,7,8,9]. However, its suitability and features in the case of application to the LFLMM have not been investigated yet. In this study, we extend the work of [1] by employing the Supplemented EM algorithm as a by-product of the EM estimator for the case of fixed variables. We used simulation studies to investigate the computational aspects of the Supplemented EM algorithm and used a real data example to illustrate the practical usefulness of this work.
The remainder of this study is organized as follows. In Section 2, we specify the LFLMM and summarize the EM algorithm to estimate it. Section 3 presents the Supplemented EM algorithm. Section 4 and Section 5 present the results of the simulation and real data example. Conclusions follow in Section 6.

2. The LFLMM and the EM Algorithm

Following [1], an LFLMM can be composed of two parts. The first is the factor analysis model, which represents the relationships between the observed and latent variables. This is similar to the structural equation model, which explains the relationship of the latent variables and the measurement indicators carried out through factor analysis [10]. This part can be written as:
Y i t = Λ η i t + ϵ i t
Specifically, for the i-th of N   individuals , we observe j = 1, …, J responses which characterize d latent factors ( η i t = ( η i t 1 , , η i t d ) ,   d < J ) at time t , t = 1 , , T i , where T i is the number of time periods for subject i . Λ is the matrix of factor loadings and ϵ i t = ( ϵ i t 1 , ,   ϵ i t J ) the vector of measurement errors for subject i at time t . It is assumed that ϵ i t J ~ N ( 0 , τ j 2 ) and ϵ i t j ϵ i t h , j h . In matrix notation, Equation (1) reads:
Y i = ( I T i Λ ) η i + ϵ i
where
Y i = ( y i 1 ,   ,   y i T i ) [ J × T i , 1 ] η i = ( η i 1 ,   ,   η i T i ) [ d × T i , 1 ] Λ [ J × d ] = ( λ 1 λ J )
The second part of the LFLMM is a multivariate linear mixed model containing the fixed and random effects for each latent variable ( η i t ) . For individual i , i = 1 , 2 ,     N , at time t , t = 1 ,   2 ,   ,   T i , and latent variable l ,   l   = 1 ,   2 ,   , d , we thus have:
η i t l = x i t l β l + z i t l a i l + ε i t l
where x i t l and z i t l are the elements of design matrices of the p   fixed   variables and q random effects, respectively. β l is an unknown   coefficient , a i l = ( a i 1 l , ,   a i q l ) and ε i l = ( ε i 1 l , ,   ε i T i l ) , l = 1 , 2 ,     d , are the random effects and errors for subject i and factor l ,   respectively . The random effects are assumed to be normally distributed with mean 0 and variance-covariance matrix V ( a ) = Σ a . It is assumed that Σ a captures the changes among the latent variables [1]. For example, a positive covariance between the random effects for the latent variables 1 and 2 means that if for a given individual i the latent variable 1 increases over time, the latent variable 2 also increases for that individual. Note that in this setting, the covariates are included in the multivariate linear mixed model (MLMM) of Equation (3) but not in the factor analysis model of Equation (1).
In matrix notation, Equation (3) reads:
η i = X i β + Z i a i + ε i
where
X i = ( x i 1   x i T i ) [ d × T i , p × d ] ,   x i t = ( x i t 1 0 0 x i t 2 0 0   0   0 x i t d ) [ d ,   p × d ] Z i = ( z i 1   z i T i ) [ d × T i , q × d ] ,   z i t = ( z i t 1 0 0 z i t 2 0 0   0   0 z i t d ) [ d ,   q × d ] β = ( β 1 , ,   β d ) [ p × d , 1 ] a i = ( a i 1 ,   ,   a i d ) [ q × d , 1 ] N ( 0 , Σ a ) ε i = ( ε i 1 ,   ,   ε i T i ) [ d × T i , 1 ] ,   ε i t N ( 0 ,   Σ ε )
The marginal distribution of Y i is assumed multivariate normal with mean:
E   ( Y i ) = ( I T i Λ ) X i β
and variance-covariance matrix
V ( Y i ) = ( I T i Λ ) V ( η i ) ( I T i Λ ) + I T i d i a g ( τ 1 2 ,   ,   τ J 2 )
The first term in V ( Y i ) denotes the variances and covariance of the latent factors and the last term the variances of the error term, ϵ i t . The mean and variance-covariance matrix of η i are E ( η i ) = X i β and V ( η i ) = Z i Σ a Z i + I T i Σ ε , respectively.
To estimate the LFLMM by EM, we summarize it below, as proposed by [1]. Before going into detail, we observe that { η i ,   a i } is treated as missing data. Hence, the complete dataset is { Y i ,   X i ,   Z i ,   η i ,   a i } whereas the observed data is { Y i ,   X i ,   Z i } . It follows that the complete data likelihood is:
L = i = 1 N P ( Y i | η i , Λ ,   τ 2 ) P ( η i | X i , Z i ,   a i , β , Σ ε ) P ( a i | Σ a )
The corresponding complete data loglikelihood is:
log L = i = 1 N [ log P ( Y i | η i , Λ ,   τ 2 ) + log P ( η i | X i , Z i ,   a i , β , Σ ε ) + log P ( a i | Σ a ) ]
where
i = 1 N log P ( Y i | η i , Λ ,   τ 2 ) = i = 1 N j = 1 J log P ( Y i j | η i , Λ j ,   τ j 2 ) = i = 1 N j = 1 J [ n i 2 log τ j 2 1 2 τ j 2 ( Y i j η i Λ j ) ( Y i j η i Λ j ) ]
i = 1 N log P ( η i | X i , Z i ,   β , a i , Σ ε ) = i = 1 N t = 1 T i log P ( η i t | X i , Z i ,   a i , β ,   Σ ε ) = i = 1 N t = 1 T i [ 1 2 log | Σ ε | 1 2 ( η i t X i t β Z i a i ) Σ ε 1 ( η i t X i t β Z i a i )   ]
i = 1 N log P ( a i | Σ a ) = 1 2 i = 1 N [ log | Σ a | 1 2 a i Σ a 1 a i ]
Let the θ denote the parameter vector ( Λ ,   τ 2 ,   β ,   and   Σ ε ) , θ ( w ) be the ML estimate of θ at the w th iteration for w = 0 , 1 , ., and Q ( θ | θ ( w ) ) the expectation of the joint loglikelihood for the complete data   { Y i ,   X i ,   Z i ,   η i ,   a i } conditional on the observed data { Y i ,   X i ,   Z i } :
Q ( θ | θ ( w ) ) = E { log L ( θ | Y i ,   X i ,   Z i ,   η i ,   a i ) | Y i ,   X i ,   Z i , θ ( w ) }
Then the ( w + 1 ) th iteration of the EM algorithm consists of (i) the E–step, which is the expectation of the joint loglikelihood computed according to (10) and (ii) the M-step, which maximizes Q ( θ | θ ( w ) ) to yield θ ( w + 1 ) . Further details on EM estimation of the LFLMM can be found in [1].

3. The Supplemented EM

Below we discuss the Supplemented EM algorithm, denoted as SEM. Before going into detail, we observe that the main purpose of this study is to estimate the standard errors of the fixed effects, β .
Consider the mapping M defined by iteration w of the EM algorithm:
β ( w + 1 ) = M ( β ( w ) ) ,   for   w = 0 , 1 ,
when the parameter vector converges to β ,   we   obtain   β = M ( β ) . For M ( β ) continuous we have by Taylor expansion in the neighbourhood of β
β ( w + 1 ) = M ( β ( w ) ) M ( β ) + D M ( β ( w ) β ) = β + D M ( β ( w ) β )
where
D M = ( M h ( β ) β g ) | β = β
g = 1 , 2 , ,   k and h = 1 , 2 , , k is the k × k Jacobian matrix of M ( β ) = ( M 1 ( β ) , , M k ( β ) ) evaluated at the ML estimate of β with k = p × d .   DM is known as the rate matrix. To obtain the loglikelihood of β , we consider the complete data density of the LFLMM:
f ( { Y i ,   X i ,   Z i ,   η i ,   a i } | θ ) = f ( { Y i ,   X i ,   Z i } | θ ) f ( { η i ,   a i } | { Y i ,   X i ,   Z i } , θ )
where f ( { Y i ,   X i ,   Z i } | θ ) is the density of the observed data and f ( { η i ,   a i } | { Y i ,   X i ,   Z i } , θ ) the density of missing data, given the observed data. Thus, the loglikelihood of β given the complete data is:
log L ( β | { Y i ,   X i ,   Z i ,   η i ,   a i } ) = log L ( β | { Y i ,   X i ,   Z i } ) + log   f ( { η i ,   a i } | { Y i ,   X i ,   Z i } , β )
where log L ( β | { Y i ,   X i ,   Z i } ) is the observed-data loglikelihood and log L ( β | { Y i ,   X i ,   Z i ,   η i ,   a i } ) is the complete data loglikelihood.
The asymptotic variance-covariance matrix of β , V ( β ) , is the inverse of the observed information matrix ( I o ) . In the case of the LFLMM, the observed data is { Y i ,   X i ,   Z i } so that V ( β ) is:
V ( β ) = I o 1 ( β | { Y i ,   X i ,   Z i } )
where I o ( β | { Y i ,   X i ,   Z i } )   is the information matrix of the observed data loglikelihood (which is assumed to exist). That is [6,11]:
I o ( β | { Y i ,   X i ,   Z i } ) = E [ 2 log L ( β | { Y i ,   X i ,   Z i } ) β · β ]
Equation (15) is difficult to evaluate directly using the EM algorithm [6,11]. As a way out, [7] suggested to evaluate the complete data information matrix:
I o ( β | { Y i ,   X i ,   Z i ,   η i ,   a i } ) = E [ 2 log L ( β | { Y i ,   X i ,   Z i ,   η i ,   a i } ) β · β ]
The conditional complete data information, given the observed data evaluated at β = β , is:
I o c = E [ I o ( β | { Y i ,   X i ,   Z i ,   η i ,   a i } | { Y i ,   X i ,   Z i } , β   ]  
After taking second derivatives, averaging over f ( { η i ,   a i } | { Y i ,   X i ,   Z i } , β ) , and evaluating at β = β , Equation (13) implies:
I o ( β | ( { Y i ,   X i ,   Z i } ) = I o c I o m
where the missing information matrix ( I o m ) is
I o m = E [ 2 log f   ( { η i ,   a i } | { Y i , X i , Z i } , β   ) β · β ]
ref [12] interpreted Equation (18) as
o b s e r v e d   i n f o r m a t i o n = c o m p l e t e   i n f o r m a t i o n m i s s i n g   i n f o r m a t i o n
and called it the “missing information principle”. Equation (18) can be written as:
I o ( β | ( { Y i ,   X i ,   Z i } ) = ( I I o m I o c 1 ) I o c ,
where I is the k × k identity matrix and I o m I o c 1 is the matrix of the fraction of missing information [7,11]. According to [13], the rate of convergence of the EM algorithm is determined by the fraction of missing information in the neighborhood of β :
D M = I o m I o c 1
Substituting D M = I o m I o c 1 into Equation (20) and inverting, the asymptotic variance-covariance matrix of β , V ( β ) is:
V ( β ) = I o c 1 ( I D M ) 1
From the equality ( I P ) 1 = ( I P + P ) ( I P ) 1 = I + P ( I P ) 1 it follows that:
V ( β ) = I o c 1 { I + D M ( I D M ) 1 } = I o c 1 + I o c 1 D M ( I D M ) 1
or
V ( β ) = I o c 1 + Δ V ( β )
where Δ V ( β ) is the increase of the diagonal elements of V ( β ) related to missing information.
Calculation of the D M matrix can be done using the code and output of the original EM algorithm as follows [6,7]. The D M matrix represents the differential of the parameter mappings during the EM algorithm. Hence, each element of the D M matrix represents a component-wise increase of the rate of convergence per iteration of the EM algorithm. Let r g h be the ( g , h )   th element of the D M matrix. From Equation (13), we have:
r g h = M h ( β ) β g = lim β g β g M h ( β 1 , , β g 1 , β g , β g + 1 ,   , β k ) M h ( β ) β g β g = lim w M h ( β ( w ) ( g ) ) M h ( β ) β g ( w ) β g l i m w r g h ( w )
g = 1 , 2 , ,   k and h = 1 , 2 , , k
where β ( w ) ( g ) is called the semi-active parameter set
β ( w ) ( g ) = ( β 1 , , β g 1 , β g ( w ) , β g + 1 ,   , β k )   ,   w = 1 ,   2 ,
which converges to β g . Note that only the g th component in β ( w ) ( g ) takes a value different from its maximum likelihood estimate.
To calculate   r g h , the Supplemented EM algorithm requires θ = { Λ ,   τ 2 ,   β and   Σ ε } and θ ( w ) = { Λ ( w ) ,   τ 2 ( w ) ,   β ( w ) and   Σ ε ( w ) } for w = 1 ,   2 , as input .   θ can be obtained by the EM algorithm using a set of arbitrarily chosen initial parameters θ i n i t   including θ ( w ) for w = 1 ,   i . e . ,   θ ( 1 ) . The starting point θ ( 1 ) may, but need not, be close θ . The algorithm below closely follows [14,15].
  • Select input: θ ( w ) and θ
  • Set θ ( w ) for w = 1 ,   2 , . Then take the E step and M step of the LFLMM EM algorithm to produce θ ( w + 1 ) .
  • For rows = 1 , 2 , ,   k :
    (i)
    Set β ˜ ( w ) ( g ) be equal to β , except for the g th element: ( β ˜ ( w ) ( g ) = ( β 1 , , β g 1 , β g ( w ) , β g + 1 ,   , β k ) )
    (ii)
    Run the LFLMM EM algorithm with β ˜ ( w ) ( g ) as the current estimate of β to obtain β ˜ ( w + 1 ) ( g ) .
    (iii)
    Calculate the g th row of r g h ( w ) as
    r g h ( w ) = β ˜ h ( w + 1 ) ( g ) β h β g ( w ) β g   ,   for   h = 1 , 2 , , k
The output after a single run of the Supplemented EM algorithm (Step 1 and 2) are β w + 1 and r g h ( w )   g = 1 , 2 , ,   k and h = 1 , 2 , , k . Based on the final estimates of DM, V ( β ) is calculated using (24). The diagonal elements of V are the variance of β .

4. Simulation

To evaluate the statistical properties and computational aspects of the SEM, we set up a simulation study. The number of subjects ( N ) is set at 500, 1000, and 1500 with six time periods. The number of simulations ( S ) is set at 50 and 250. The other set-up of the simulations is adopted from [1]. Particularly, we use the same initial values of the parameters of the LFLMM model (12 items, 2 latent factors, and a simple structure to model the relationship between the items and the latent factors). It is done to check if the bias resulting from the Supplemented EM algorithm on LFLMM is in line with the results presented in [1]. Table 1 presents the absolute difference for the true parameters, and the averages of the SEM estimates are calculated as a measure of performance.
Table 1 shows that the absolute difference of σ a , 11 has a range from 0 to 0.0444 ( N = 500 and S = 250 ) . The results are in line with [1] the parameters of the measurement model (factor loadings and error variances) are estimated more precisely than those of the latent mixed regression model. Overall, these results indicate that with the increasing number of subjects in the simulation, the absolute difference between the actual parameters and the SEM average is getting smaller. This means the SEM can estimate the model parameter very well.
Although the results in Table 1 indicate that the accuracy of the estimates in the latent mixed regression model (No. 23–43) is not as good as in the measurement model part (No. 1–22), through Figure 1a–c, it can be shown that the median of boxplots (which generally is close to the true parameters) are all at the same level. This means that the parameters of the latent mixed regression models ( β ,   σ a ,   σ ε ) are estimated more precisely, especially for the number of simulations S = 250 . Furthermore, all the boxplots are also shown to have different distributions of views with an increasing number of subjects in the simulation. This is indicated by the smaller size of the boxplot as the number of subjects increases for both numbers of simulations.
The results from the simulations of the Supplemented EM algorithm in estimating the asymptotic variance matrix of beta is summarized as a standard deviation of beta in Table 2. We also calculate the standard deviation of beta using the 2nd moment, V ( β ^ ) = 1 S 1 s = 1 S ( β ^ s β ^ ¯ ) 2 as a benchmark to compare with the standard deviation of beta of SEM. Both the 2nd moment and SEM produce symmetrical results the parameter estimate for the standard deviation of beta is getting smaller with the increasing number of subjects in the simulation. Overall, it can be concluded that by using SEM, changes in the parameter estimate for the standard deviation of beta are not too different for all number of subjects (Figure 2). Therefore, the simulation results suggest that the asymptotic variance of beta from the Supplemented EM Algorithm can be used to estimate the asymptotic variance of beta in real data analysis.

5. Real Data Example

The real data-set that we used to illustrate the development of the Supplemented EM algorithm is the political attitudes and behavior data of Flemish. The data was designed to include a representative sample of the target population under the Belgian electorate. The Flemish data set (Flemish and Dutch speaking respondents from Brussels Capital Region) consists of 1274 respondents, who have been interviewed three times (1991, 1995, and 1999) [16,17,18]. There are four latent factors measured on political attitudes of Flemish used, i.e., Individualism, Nationalism, Ethnocentrism, and Authoritarianism. This data has been analyzed using various methods by several authors, including [19,20,21,22,23]. There are three interesting questions in this real data case, i.e., how Individualism, Nationalism, Ethnocentrism, and Authoritarianism of the Flemish develop over time; whether there is an association between these four developments, and whether the gender of the respondent affects the change patterns of latent developments.
I, N, E, and A in Table 3 correspond to Individualism, Nationalism, Ethnocentrism, and Authoritarianism, respectively. a 11 and a 12 are the random intercept and random slope for Individualism. a 21 and a 22   are the random intercept and random slope for Nationalism. a 31 and a 32   are the random intercept and random slope for Ethnocentrism. a 41 and a 42   are the random intercept and random slope for Authoritarianism. The positive correlation of random intercept between a 11 and a 21 , a 11 and a 31 , a 11 and a 41 suggests that the development of Individualism and other political attitudes is highly related, which highest correlated with Ethnocentrism. The results indicate that those who have a better sense of Individualism tend to have a better sense of Nationalism, Ethnocentrism, and Authoritarianism. The results find a positive correlation of random intercept between a 21 and a 31 , a 21 and a 41 . It suggests that those who have a better sense of Nationalism tend to have a better sense of Ethnocentrism and Authoritarianism, as well as those who have a better sense of Ethnocentrism tend to have a better sense of Authoritarianism. There is also a positive correlation of random slope between a 12 and a 22 . It means that if one subject’s Individualism decreases over time, then it is reasonable to expect that his or her Nationalism will decrease over time and vice versa. This also holds between Individualism and Ethnocentrism and between Individualism and Authoritarianism. The positive correlation of random slope between a 22 and a 32 , meaning that if one subject’s Nationalism decreases over time, then it is reasonable to expect that his or her Ethnocentrism will decrease over time. The correlation matrix of random effects confirms that all latent factors have a positive correlation over time.
The significance of parameter estimate of β is analyzed via the z values. By using the Supplemented EM algorithm, the standard errors of β for all parameters can be calculated. The standard errors of β are listed in Table 4. Using a 95 percent confidence interval of β , almost all confidence intervals do not include the null value, except the slope of Male on Authoritarianism. Hence there are statistically significant differences in the parameter estimate of β . In other words, all latent factors of Flemish people decrease over time, with Ethnocentrism having the highest rate of decline over time (−0.252) and Nationalism the lowest (−0.177). On average, the Individualism and Nationalism of the male respondent are higher than that of the female. However, Ethnocentrism of the male respondent is lower than that of the female.

6. Conclusions

This paper proposed the Supplemented EM algorithm for LFLMM in estimating the asymptotic variance-covariance matrix as a by-product of the EM estimator for the case of fixed variables in the model. Results from simulation studies suggest that the Supplemented EM algorithm can estimate the model very close to the initial parameters.
As a result of the development of EM algorithm of LFLMM, the Supplemented EM algorithm is very slow to converge, as stated by [1], especially when the number of simulations is 250 times with 1500 subjects. For this reason, further research is needed to find techniques that can be used to accelerate the speed of the algorithm. Several approaches to speed the EM algorithm have been proposed and can be found in [24,25,26] (the ECM algorithm), [27] (the ECME algorithm), and [28] (the Parameter-Expanded EM algorithm).

Author Contributions

Conceptualization, Y.A.; methodology, Y.A.; software, Y.A.; validation, K.A.N. and H.F.; formal analysis, Y.A.; data curation, Y.A. and T.T.; writing—original draft preparation, Y.A.; writing—review and editing, K.A.N., H.F., A.S. and T.T.; visualization, T.T.; supervision, K.A.N., H.F. and A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by RUG and IPB University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from ISPO which were used under license, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of ISPO.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. An, X.; Yang, Q.; Bentler, P.M. A latent factor linear mixed model for high-dimensional longitudinal data analysis. Stat. Med. 2013, 32, 4229–4239. [Google Scholar] [CrossRef] [PubMed]
  2. Kondaurova, M.V.; Bergeson, R.R.; Xu, H.; Kitamura, C. Affective Properties of Mothers’ Speech to Infants with Hearing Impairment and Cochlear Implants. J. Speech Lang. Hear. Res. 2015, 58, 590–600. [Google Scholar] [CrossRef] [Green Version]
  3. Wang, J.; Luo, S. Multidimensional latent trait linear mixed model: An application in clinical studies with multivariate longitudinal outcomes. Stat. Med. 2017, 36, 3244–3256. [Google Scholar] [CrossRef]
  4. Ng, S.K.; Krishnan, T.; McLachlan, G. The EM algorithm. In Handbook of Computational Statistics; Springer: Berlin, Germany, 2004; pp. 137–168. ISBN 9783642215513. [Google Scholar]
  5. Mclachlan, G.J.; Krishnan, T. The EM Algorithm and Extensions Second Edition, 2nd ed.; Wiley: New York, NY, USA, 2007; ISBN 9780471201700. [Google Scholar]
  6. Meng, A.X.; Rubin, D.B. Using EM to Obtain Asymptotic Variance-Covariance Matrices: The SEM Algorithm. J. Am. Stat. Assoc. 1991, 86, 899–909. [Google Scholar] [CrossRef]
  7. Cai, L. SEM of another flavour: Two new applications of the supplemented EM algorithm. Br. J. Math. Stat. Psychol. 2008, 61, 309–329. [Google Scholar] [CrossRef] [PubMed]
  8. Cai, L.; Lee, T.; Lee, T. Covariance Structure Model Fit Testing Under Missing Data: An Application of the Supplemented EM Algorithm Covariance Structure Model Fit Testing Under Missing Data: An Application of the Supplemented EM Algorithm. Multivar. Behav. Res. 2009, 44, 281–304. [Google Scholar] [CrossRef] [PubMed]
  9. Tian, W.; Cai, L.; Thissen, D.; Xin, T. Numerical Differentiation Methods for Computing Error Covariance Matrices in Item Response Theory Modeling: An Evaluation and a New Proposal. Educ. Psychol. Meas 2012, 73, 412–439. [Google Scholar] [CrossRef]
  10. Caraka, R.E.; Noh, M.; Chen, R.C.; Lee, Y.; Gio, P.U.; Pardamean, B. Connecting climate and communicable disease to penta helix using hierarchical likelihood structural equation modelling. Symmetry 2021, 13, 657. [Google Scholar] [CrossRef]
  11. Pritikin, J.N. A comparison of parameter covariance estimation methods for item response models in an expectation-maximization framework. Cogent Psychol. 2017, 4, 1–11. [Google Scholar] [CrossRef]
  12. Orchard, T.; Woodbury, M. A Missing Information Principle: Theory and Applications. In Theory of Statistics; University of California Press: Berkeley, CA, USA, 1972; Volume 1, pp. 697–715. Available online: https://projecteuclid.org/download/pdf_1/euclid.bsmsp/1200514117 (accessed on 1 June 2021).
  13. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum Likelihood from Incomplete Data via the EM Algorithm A. J. R. Stat. Soc. Ser. B 1977, 39, 1–38. [Google Scholar]
  14. Little, R.J.A.; Rubin, D.B. Statistical Analysis with Missing; Wiley: New York, NY, USA, 2002; ISBN 3175723993. [Google Scholar]
  15. Abel, G.J. International Migration Flow Table Estimation. Ph.D. Thesis, University of Southampton, Southampton, UK, 2009. [Google Scholar]
  16. Interuniversitair Steunpunt Politieke-Opinieonderzoek. General Election Study: Codebook and Questionnaire; ISPO: Leuven, Belgium, 1991; ISBN 9067841161. [Google Scholar]
  17. Interuniversitair Steunpunt Politieke-Opinieonderzoek. General Election Study: Codebook and Questionnaire; ISPO: Leuven, Belgium, 1995; ISBN 9067841366. [Google Scholar]
  18. Interuniversitair Steunpunt Politieke-Opinieonderzoek. General Election Study: Codebook and Questionnaire; ISPO: Leuven, Belgium, 1999. [Google Scholar]
  19. Billiet, J. Church Involvement, Individualism, and Ethnic Prejudice among Flemish Roman Catholics: New Evidence of a Moderating Effect. J. Sci. Study Relig. 1995, 34, 224–233. [Google Scholar] [CrossRef]
  20. Billiet, J.; Coffe, H.; Maddens, B. Een Vlaams-nationale identiteit en de houding tegenover allochtonen in een longitudinaal perspectief. In Proceedings of the Paper Presented at the Marktdag Sociologie; Universitaire Pers Leuven: Leuven, Belgium, 2005. [Google Scholar]
  21. Toharudin, T.; Oud, J.H.L.; Billiet, J.B. Assessing the relationships between Nationalism, Ethnocentrism, and Individualism in Flanders using Bergstrom’s approximate discrete model. Stat. Neerl. 2008, 62, 83–103. [Google Scholar] [CrossRef]
  22. Toharudin, T.; Oud, J.H.L.; Billiet, J.; Folmer, H. Measuring Authoritarianism with Different Sets of Items in a Longitudinal Study. In Methods, Theories, Andempirical Applications in the Social Sciences; Salzborn, S., Davidov, E., Reinecke, J., Eds.; Springer: Heidelberg, Germany, 2012; pp. 193–200. ISBN 9783531188980. [Google Scholar]
  23. Angraini, Y.; Toharudin, T.; Folmer, H.; Oud, J.H.L. The Relationships between Individualism, Nationalism, Ethnocentrism, and Authoritarianism in Flanders: A Continuous Time-Structural Equation Modeling Approach. Multivar. Behav. Res. 2014, 49, 41–53. [Google Scholar] [CrossRef] [PubMed]
  24. Meng, X.; Rubin, D.B. Maximum likelihood estimation via the ECM algorithm: A general framework. Biometrika 1993, 80, 267–278. [Google Scholar] [CrossRef]
  25. Van Dyk, D.A.; Meng, X.; Rubin, D.B. Maximum Likelihood Estimation via the ECM Algorithm: Computing The Asymptotic Variance. Stat. Sin. 1995, 5, 55–75. [Google Scholar]
  26. Li, H.; Tian, W. Slashed lomax distribution and regression model. Symmetry 2020, 12, 1877. [Google Scholar] [CrossRef]
  27. Liu, B.Y.C.; Rubin, D.B. The ECME algorithm: A simple extension of EM and ECM with faster monotone convergence. Biometrika 1994, 81, 633–648. [Google Scholar] [CrossRef]
  28. Liu, C.; Rubin, D.B.; Wu, Y.N. Parameter Expansion to Accelerate EM: The PX-EM Algorithm. Biometrika 1998, 85, 755–770. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) Boxplot the parameters estimate of the latent mixed regression models ( β ) . (b) Boxplot the parameters estimate of the latent mixed regression models ( σ a ) . (c) Boxplot the parameters estimate of the latent mixed regression models ( σ ε ) .
Figure 1. (a) Boxplot the parameters estimate of the latent mixed regression models ( β ) . (b) Boxplot the parameters estimate of the latent mixed regression models ( σ a ) . (c) Boxplot the parameters estimate of the latent mixed regression models ( σ ε ) .
Symmetry 13 01286 g001aSymmetry 13 01286 g001b
Figure 2. Line plot for the standard deviation of beta.
Figure 2. Line plot for the standard deviation of beta.
Symmetry 13 01286 g002
Table 1. The absolute difference between the true parameters and the SEM averages.
Table 1. The absolute difference between the true parameters and the SEM averages.
NoParaTrueSEMNoParaTrueSEM
N 50010001500 N 50010001500
S 502505025050250 S 502505025050250
1 λ 2 , 1 10.00130.00120.00100.00090.00090.000923 β 1 1 10.00400.00360.00320.00040.00180.0007
2 λ 3 , 1 10.00190.00150.00130.00090.00120.000924 β 1 2 −10.01110.00790.00080.00080.00340.0017
3 λ 4 , 1 10.00150.00090.00050.00090.00120.001025 β 2 1 00.00380.00180.00120.00070.00160.0003
4 λ 5 , 1 10.00140.00160.00070.00090.00070.000926 β 2 2 00.00310.00390.00140.00210.00440.0006
5 λ 6 , 1 10.00160.00140.00110.00090.00130.001027 β 3 1 10.00120.00160.00080.00030.00070.0004
6 λ 7 , 2 10.00210.00180.00130.00140.00160.001328 β 3 2 10.00160.00100.00020.00040.00000.0005
7 λ 8 , 2 10.00150.00160.00170.00140.00100.001329 β 4 1 10.00250.00270.00020.00010.00070.0002
8 λ 9 , 2 10.00200.00150.00150.00140.00190.001330 β 4 2 −10.00180.00250.00040.00010.00030.0001
9 λ 10 , 2 10.00150.00150.00150.00140.00220.001731 σ a , 11 30.03320.04440.01520.01540.00390.0160
10 λ 11 , 2 10.00190.00120.00170.00170.00160.001632 σ a , 12 10.04100.01200.00240.00570.02650.0085
11 τ 1 2 0.50.00070.00010.00060.00040.00090.000433 σ a , 13 1.50.02890.02000.02180.01410.02290.0133
12 τ 2 2 0.50.00030.00060.00000.00110.00190.000634 σ a , 14 10.02860.01310.00770.00830.01690.0094
13 τ 3 2 0.50.00270.00010.00210.00010.00050.000735 σ a , 22 30.02330.01050.02470.00170.02760.0150
14 τ 4 2 0.50.00070.00090.00020.00020.00050.000836 σ a , 23 10.00030.00450.01580.00420.01240.0051
15 τ 5 2 0.50.00020.00060.00100.00130.00100.000637 σ a , 24 20.00730.00200.00570.00040.01260.0089
16 τ 6 2 0.50.00440.00100.00170.00050.00040.000438 σ a , 33 30.00140.01250.03830.02600.03870.0274
17 τ 7 2 0.50.00050.00080.00160.00210.00170.000439 σ a , 34 10.00800.00750.02360.00690.00430.0022
18 τ 8 2 0.50.00270.00120.00060.00060.00110.000940 σ a , 44 30.00210.00150.02150.00200.01480.0094
19 τ 9 2 0.50.00160.00100.00040.00090.00010.000441 σ ε , 11 0.50.00200.00130.00280.00050.00150.0013
20 τ 10 2 0.50.00090.00120.00120.00050.00030.000842 σ ε , 12 0.20.00170.00060.00060.00030.00060.0001
21 τ 11 2 0.50.00010.00010.00130.00000.00120.000143 σ ε , 22 0.50.00210.00330.00270.00290.00280.0023
22 τ 12 2 0.50.00110.00120.00010.00020.00110.0002
Table 2. The parameter estimates for V ( β ^ ) .
Table 2. The parameter estimates for V ( β ^ ) .
Number of SubjectsParameterThe 2nd MomentSEM
5025050250
500 β 1 1 0.05120.05280.02490.0251
β 1 2 0.05870.05300.01870.0195
β 2 1 0.03780.03920.01100.0126
β 2 2 0.03910.03790.01640.0158
β 3 1 0.01840.01900.02450.0268
β 3 2 0.01840.01900.02120.0200
β 4 1 0.02830.02830.01050.0105
β 4 2 0.02880.03080.01450.0145
1000 β 1 1 0.03380.03270.01790.0184
β 1 2 0.03420.03390.01380.0134
β 2 1 0.02320.02320.00710.0077
β 2 2 0.02530.02320.01050.0100
β 3 1 0.00950.01000.01760.0184
β 3 2 0.00890.00950.01300.0130
β 4 1 0.01410.01300.00770.0077
β 4 2 0.01300.01300.01100.0105
1500 β 1 1 0.02760.02530.01520.0152
β 1 2 0.02700.02650.01050.0110
β 2 1 0.01870.01820.00630.0063
β 2 2 0.01640.01920.00840.0084
β 3 1 0.00710.00710.01450.0152
β 3 2 0.00630.00710.01050.0105
β 4 1 0.01100.01050.00630.0063
β 4 2 0.01100.01050.00840.0084
Table 3. Correlation matrix of random effects.
Table 3. Correlation matrix of random effects.
Random EffectsINEA
a 11 a 12 a 21 a 22 a 31 a 32 a 41 a 42
I a 11 10.8920.8320.9030.9660.9300.9560.880
a 12 0.89210.8540.8780.9370.9310.9130.895
N a 21 0.8320.85410.3110.8640.8830.8560.835
a 22 0.9030.8780.31110.9180.8930.8990.861
E a 31 0.9660.9370.8640.91810.9460.9730.919
a 32 0.9300.9310.8830.8930.94610.9460.918
A a 41 0.9560.9130.8560.8990.9730.94610.873
a 42 0.8800.8950.8350.8610.9190.9180.8731
Table 4. Parameter estimates of β .
Table 4. Parameter estimates of β .
ParameterEstimate S E   ( β ^ ) LowerUpper
Individualism −0.1950.004−0.203−0.187
Nationalism −0.1770.028−0.231−0.123
Ethnocentrism −0.2520.011−0.273−0.231
Authoritarianism −0.1860.002−0.190−0.182
Slope of Male on Individualism0.1100.0100.0910.129
Slope of Male on Nationalism0.2190.0170.1850.253
Slope of Male on Ethnocentrism−0.0380.003−0.045−0.031
Slope of Male on Authoritarianism0.0220.017−0.0110.055
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Angraini, Y.; Notodiputro, K.A.; Folmer, H.; Saefuddin, A.; Toharudin, T. Estimating the Variance of Estimator of the Latent Factor Linear Mixed Model Using Supplemented Expectation-Maximization Algorithm. Symmetry 2021, 13, 1286. https://doi.org/10.3390/sym13071286

AMA Style

Angraini Y, Notodiputro KA, Folmer H, Saefuddin A, Toharudin T. Estimating the Variance of Estimator of the Latent Factor Linear Mixed Model Using Supplemented Expectation-Maximization Algorithm. Symmetry. 2021; 13(7):1286. https://doi.org/10.3390/sym13071286

Chicago/Turabian Style

Angraini, Yenni, Khairil Anwar Notodiputro, Henk Folmer, Asep Saefuddin, and Toni Toharudin. 2021. "Estimating the Variance of Estimator of the Latent Factor Linear Mixed Model Using Supplemented Expectation-Maximization Algorithm" Symmetry 13, no. 7: 1286. https://doi.org/10.3390/sym13071286

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop