1. Introduction
Economic and statistical models, especially dynamic stochastic general equilibrium (DSGE) models, are susceptible to misspecification problems since they consist of simplifications (sometimes remarkably strong) of reality. The omission of relevant variables, incorrect functional forms, distributional assumptions, and incompleteness of systems of relations are common and often coincide in estimation and forecasting procedures [
1]. Thus, we desire to employ estimation methods with good performance in the presence of misspecification. Here, robustness means insensitivity to small deviations from assumptions adopted, as defined by [
2].
Misspecification problems can have global (non-local) or local natures. In moment-based estimation, global misspecification occurs when no parameter value is compatible with the moment restrictions, regardless of the sample size. Local misspecification occurs when these conditions are not satisfied in part of the sample, which disappears asymptotically [
3]. See
Appendix A for details on this and see
Appendix B for a more detailed discussion of local and global misspecification in moment-based estimation. In likelihood-based estimation, consistency can be significantly affected by the validity of distributional assumptions; the quasi-maximum likelihood estimator may or may not be consistent for particular parameters of interest [
4,
5]. In this approach, we can consider this case as global misspecification. The authors of [
6] and other papers introduce local misspecification through a nuisance parameter that contaminates the sample and whose effect disappears asymptotically. Depending on the estimation approach, both local and global misspecification problems are usually present in DSGE estimation.
In the current literature on DSGE modeling, we observe two distinct approaches, as highlighted by [
7]. The first, which uses the likelihood principle, emerged in [
8] and is based on approximations of the model’s police functions obtained by linearizing the equilibrium conditions around the steady state. This technique allows for evaluating the likelihood function using the Kalman filter or particle filter. The parameters are estimated either by classical inference maximizing the likelihood function (maximum likelihood—ML) or by Bayesian inference (BI) combining the likelihood function and prior distribution to obtain the parameters’ posterior distribution.
The moment-based approach, in turn, uses a set of moment conditions generated from the model’s first-order conditions (FOCs). The first works that employed GMM in DSGE estimation were [
9,
10,
11]. Among the advantages of GMM estimation about ML estimation (and BI) are (i) to require fewer constraints for the data distribution and (ii) to require less computational capacity and, consequently, less estimation time [
7,
12]. Ref. [
13] concluded that moment-based methods, more specifically GMM and the simulated method of moments (SMM), present better results than the ML estimator in terms of estimation speed and robustness to specification problems. Ref. [
12] showed that GMM and SMM estimators obtained more accurate estimates for model parameters, even in small samples.
This paper aims to analyze the performance of moment-based estimators belonging to generalized empirical likelihood (GEL) and generalized minimum contrast (GMC) families in DSGE estimation, using GMM, ML, and BI as benchmarks. The generalized empirical likelihood (GEL) and generalized minimum contrast (GMC) methods generalize the use of moment methods when performing the nonparametric estimation of the process distribution, obtaining better properties in finite samples in terms of bias and some additional robustness properties. Thus, it is interesting to analyze whether the use of these methods can bring gains in the estimation of DSGE models since these models are commonly estimated from small samples and under relevant simplifications and restrictions.
In minimum contrast estimation, entropy measures, especially those rooted in Kullback–Leibler divergence, are often used to quantify the “distance” between an empirical distribution and a theoretical model. This approach is valuable when likelihood functions are complex or unspecified, allowing for the use of contrast functions like entropy to approximate them [
14,
15]. By minimizing the “contrast” (e.g., the divergence) between observed and model-based distributions, estimators can achieve a form of alignment that captures the underlying statistical structure with minimal assumptions.
Entropy concepts, particularly in generalized empirical likelihood (GEL), serve a dual role. They adjust likelihood estimates by introducing penalties for models that diverge from empirical data, thereby helping to align model-based probabilities more closely with observed data. GEL methods use entropy-based divergences, such as Kullback–Leibler divergence, to balance fit with the need for generalization, reducing overfitting by choosing model parameters that satisfy the given constraints [
16]. This information-theoretic approach to estimation, encompassing methods like GEL and minimum contrast, emphasizes entropy’s role in making robust inferences that respect both model and data uncertainties.
Our paper extends other contributions to the employment of GEL/GMC estimators in the estimation of DSGE and economic models in general. Two moment-based estimators were considered in DSGE estimation by [
7]. The author estimated a DSGE model using GMM and exponentially tilted empirical likelihood (ETEL), a GEL/GMC estimator. ETEL did not obtain results as close to true values as the GMM estimator, and, in addition, the ETEL results presented a high standard deviation. The two estimators presented difficulties in dealing with the identification of the utility function curvature parameter. The author encouraged the continuity of research in this field due to the advantages presented by empirical likelihood, such as (i) the direct use of the equilibrium conditions of the model since it is not necessary to compute the police function; (ii) the flexibility to assume distributions for the stochastic process of the economy; and (iii) the preservation of the nonlinear structure of the equilibrium conditions of the model.
Several applications of moment-based GEL/GMC estimators have been made in finance. Among the covered topics, we highlight the following: portfolio selection based on the empirical likelihood and Hellinger distance [
17]; asset pricing models under misspecification and robustness analysis [
18]; estimation of the discretized stochastic differential equation for interest rates [
19]; dealing with hard assumptions of Black–Scholes option pricing [
20]; obtaining risk measures of the risk of loss on a specific portfolio, such as value at risk and expected shortfall [
21]; improvement in estimation precision for portfolio optimization [
22]; and portfolio efficiency tests with conditioning information in the presence of data contamination [
23]. These papers explored the good robustness and other properties of estimators belonging to the class of GEL/GMC estimators under misspecification and data contamination. They found promising results both in simulations and empirical applications.
Ref. [
18] highlighted the possibilities of robustness analysis in the estimation of misspecified asset pricing models using GEL/GMC methods. Ref. [
19] found that GEL/GMC estimators (mainly the ETEL) outperform the GMM, in terms of bias and mean squared error, in the estimation of stochastic differential equations for interest rates. Ref. [
23] showed that GEL estimators perform better than the GMM in portfolio efficiency tests with conditioning information in the presence of data contamination, such as heavy tails and outliers. Ref. [
24] applied adjusted empirical likelihood to make robust inferences about the Sharpe ratio in asset pricing and [
25] analyzed the properties of the maximum empirical likelihood, maximum empirical exponential likelihood, and maximum log Euclidean likelihood estimators to estimate spatial autocorrelation models.
We deal with a real business cycle (RBC) model that can be considered the core of current DSGE models. We verify by means of Monte Carlo experiments if the studied estimators generate satisfactory results in terms of bias and variance measures in situations where the estimated model is correctly specified. We also emphasize the robustness analysis under both local and global misspecification situations. While in the correctly specified model productivity shocks follow a normal distribution, in misspecified models, we generate productivity shocks following a Student’s t-distribution or a normal distribution with the inclusion of single or several outliers.
The objective of this article is to analyze moment condition-based methods applied to the estimation of DSGE models, with a particular focus on extending the use of the generalized method of moments (GMM) to alternative approaches grounded in moment conditions. Specifically, this study evaluates the strengths and limitations of methods such as empirical likelihood and generalized minimum contrast, considering their performance under both correctly specified and misspecified model settings.
Among the main results of our paper for the estimation of a DSGE model are the following: (i) the empirical likelihood (EL) estimator, as well as its version with smoothed moment conditions (SEL), and Bayesian inference (BI) obtained, in this order, the best performances, even in misspecification cases; (ii) continuous updating empirical likelihood (CUE), minimum Hellinger distance (HD), exponential tilting (ET), and their smoothed versions presented intermediate comparative performance; (iii) ETEL, exponential tilting Hellinger distance (ETHD) estimators, and their smoothed versions were compromised by the occurrence of atypical estimates; (iv) smoothed and non-smoothed versions of the GEL/GMC estimators exhibited very similar performances; and (v) GMM, especially in the over-identified case, and ML estimators performed worse than their competitors.
These experiments show some cases of real problems that may affect the DSGE estimation. Thus, our study contributes to the still limited literature on the robust estimation of DSGE models. As these models have high importance in the analysis and conduct of economic policies, our study contributes some recommendations about using estimation methods in situations of both correct and incorrect specification.
This paper is structured in four additional sections besides this introduction and the Appendix. In
Section 2, the economic model is presented.
Section 3 presents the estimators and Monte Carlo design.
Section 4 discusses the results.
Section 5 concludes this paper. Lastly, the Appendix shows the derivation of the moment conditions used in the moment-based approach and the definitions of local and global misspecification.
4. Results
In this section, we show tables and figures that summarize the main results of this paper.
Table 2 presents the correlations between the moment conditions. Note that some correlations are considerably high even between the original moment conditions—the case of conditions
and
, for example. The correlation is greater than 50% between
and
and between
and
in all the DGPs. Despite this, the moment conditions appear to have been informative enough to identify the parameters for most estimators. It should be noted that difficulties arising from using artificial moment conditions are the possibility of stochastic singularity between moment conditions, which could lead to serious implications regarding computational implementation.
In
Table 3, we show the results of the J, LM, and LR tests for the validity of the moment conditions of the over-identified moment-based estimators. We observe that while the J test rejected the null “correct specification” at the significance level of 5% between 26.40% and 33.45% for the GMM, the same test rejected almost all the replications for the GEL/GMC estimators. For the latter, the LM and LR tests rejected, at the same level of significance, 63.25% to 69.80% and 75.25% to 79.80% of the specifications, respectively. Therefore, under correct specification, the tests indicated rejection of the null hypothesis. In an experiment not reported here, with the increase in the sample from 200 to 2000 observations, the rejection ratio at 5% of the null in the J test for the GMM dropped to 1.55%, considering the
DGP I. That is, there was a noticeable performance problem with the test in a small sample. For the tests based on the GEL/GMC estimators, on the other hand, there was no improvement in performance with increasing the sample size.
In econometric analysis, tests such as the likelihood ratio (LR), Lagrange multiplier (LM), and J tests play a central role in evaluating model specification and the validity of moment conditions, particularly in the context of the generalized empirical likelihood (GEL) and generalized method of moments (GMM) estimators. However, these tests often display notable limitations in finite samples, which can lead to an over-rejection of the null hypothesis even when the model is correctly specified.
The J test, widely used to assess the validity of moment conditions, is prone to over-rejecting the null hypothesis under certain conditions. This tendency is often linked to the choice of instruments and the variability inherent in small samples. Over-rejection can lead to misleading conclusions about model misspecification, even when the specified model aligns well with the data. Additionally, in such cases, parameter estimates derived from the GMM may exhibit downward median bias, further complicating statistical inference. Finite-sample studies of the generalized method of moments (GMM) highlight important tradeoffs and challenges. Ref. [
61] observed that using short lags for instruments tends to yield nearly asymptotically optimal estimates, while longer lags introduce bias and misleading confidence intervals. He also noted that tests for overidentifying restrictions perform well in small samples but are slightly biased toward accepting the null hypothesis. Ref. [
62] found that the J test exhibits minimal size distortion in some cases but is biased toward over-rejection in others, with parameter estimates often showing downward median bias in cases of over-rejection. Ref. [
46] emphasized the importance of how moment conditions are weighted, with continuous updating estimators generally showing less bias but sometimes resulting in fat-tailed sample distributions. This affects confidence intervals and the reliability of overidentifying restriction tests.
Similarly, LR and LM tests are sensitive to sample size, with their performance in finite samples often reflecting size distortions. These distortions increase the likelihood of Type I errors, where valid models are incorrectly rejected. Empirical evidence has demonstrated that the sensitivity of these tests to minor deviations in the data or sampling variability is particularly problematic when dealing with small samples, as is often the case in applied econometrics.
The frequent over-rejection of valid models by these tests raises important concerns, as the results may not necessarily indicate genuine model misspecification but rather reflect the limitations of the test methodologies in finite samples. This issue underscores the need for caution in interpreting the outcomes of LR, LM, and J tests, especially in small-sample contexts. To address these challenges, researchers can employ strategies to mitigate the effects of finite-sample limitations. For example, bootstrap methods are often used to generate more accurate critical values that account for the specific characteristics of the sample, reducing the likelihood of over-rejection. Additionally, the careful selection of instruments and the use of alternative testing procedures designed to perform better in small samples can enhance the robustness of inference.
For the parameters, the results are summarized in
Table 4,
Table 5,
Table 6,
Table 7 and
Table 8 and in
Figure 1,
Figure 2,
Figure 3,
Figure 4 and
Figure 5. To facilitate the description of the results, each parameter appears in one table and one figure, separately distinguishing the DGP considered. In the tables, divided into just- and over-identified cases (for moment-based methods), we have the mean of estimates, median of estimates, bias given by the difference between mean and true parameter, mean squared error (MSE), and mean absolute error (MAE). The figures show the distributions of the parameters generated by each estimator considering 2000 replications. As a matter of space, the distributions generated by both just- and over-identified moment-based methods were superimposed (continuous black and dotted gray lines, respectively). We use the optimization method
nlminb to implement the GMM and GEL/GMC estimators. For Bayesian inference (BI), the RWMH-MCMC was constructed with a chain size of 5500 (burn-in of 1500) and maximization routine
csminwel. Increases in the size of the chain did not lead to an improvement in results.
The smoothed and non-smoothed versions of the GEL/GMC estimators presented very similar performances, making it impossible to differentiate them. Thus, in several moments, we will mention both versions throughout the text by adding (S) before the estimator’s name. Another general highlight is that the occurrence of extreme estimates for ETEL, ETHD, and their smoothed versions (
Figure 1,
Figure 2,
Figure 3,
Figure 4 and
Figure 5) deteriorate the mean, MSE, and MAE of these estimators but not the median since this is more robust to the presence of atypical values (
Table 4,
Table 5,
Table 6,
Table 7 and
Table 8). This behavior was present in the estimation of all the parameters, although it has appeared in different ways: estimates at the limit of the parametric space (parameter
), very discrepant estimates about true values (
and
), and both problems (
b and
). The performance problem of these estimators could not be overcome by using other optimization methods available in the
gmm package.
The parameters
,
, and
b were those for which the estimators analyzed in this paper returned the estimates closer to the true values, except for (i) ETEL, ETHD, and their smoothed versions, for the reason highlighted in the previous paragraph; (ii) GMM that, even considering the median, obtained inferior performance regarding
and
b in the over-identified case; and (iii) ML in the case of the
. In
Figure 2 and
Figure 3, we see that
and
b distributions generated by the over-identified GMM were asymmetric to the right and reasonably spaced, corroborating the poor performance reflected in the statistics contained in
Table 5 and
Table 6. The GMM’s estimation of the parameters
and
was poor in both the just- and over-identified cases.
In general, the parameters and were the most difficult to estimate, and good performance of the (S)EL should be highlighted in obtaining accurate estimates for both. In the case of , a problem with other estimators may derive from the true value (0.9) being close to 1, the upper limit of the parametric space. Much of the estimates were biased down, becoming closer to the true value in poor specifications, something expected due to the inclusion of extreme values that lead to increased data variability. Thus, this result should be treated as a coincidence, not a sign of robustness.
Specifically about parameter
, different experiments generated similar results, and the addition of deviations from a normal distribution (of the productivity shock) led to increases in the MSE and MAE of the estimators—most notably in the case of multiple outliers (
DGP IV) and less in the case of a single outlier (
DGP II) (
Table 4). The lowest MSEs were recorded by BI, GMM, (S)CUE, and (S)EL, while the lowest MAEs were for BI and (S)EL. The problem with (S)ETEL and (S)ETHD was due to the existence of a considerable probability mass close to zero, a lower limit of the parametric space (
Figure 1). Even under correct specification (
DGP I), ML presented poorer performance than the GMM, GEL/GMC, and BI due to the concentration of estimates around 0.9, which may be associated with the existence of a local maximum in the objective function of this estimator.
The estimation of
is usually the most difficult since this parameter controls the curvature of the utility function, the primary source of nonlinearity in the model, as pointed out by [
7,
12]. For this parameter, we have that ML and the over-identified GMM presented the worst performance in terms of bias, MSE, and MAE (
Table 2). On the other hand, the just-identified GMM presented good performance due to a low MSE and MAE. (S)EL, (S)CUE, and (S)ET delivered the best results in terms of bias, MSE, and MAE, while (S)ETEL and (S)ETHD were among the worst. Estimates of the latter presented concentration around the true value (1.8) and 0.75. Due to a similar problem, ML delivered a greater bias, MSE, and MAE. For ML, estimates were concentrated around 1.8 and 30, a number quite far from the true value.
(S)EL and ML obtained good performance in estimating parameter
b in all the considered drawings. A clear improvement in the (S)EL was observed from the just-identified to the over-identified case. This improvement did not occur with the (S)CUE, whose performance eventually worsened but not enough for it to leave the list of estimators with the best results (
Table 6). The just-identified GMM also generated reasonable estimates compared to the others, especially for data generated by
DGP I,
DGP III, and
DGP II, in this order. Even with an informative prior, the BI delivered wrong estimates for the parameter
b compared with ML in terms of bias, MSE, and MAE. It occupied an intermediate position in the performance classification of estimators for parameter
b. The over-identified GMM delivered the best results, while the results of ETEL and (S)ETHD were compromised due to atypical estimates (
Figure 3).
The results for
presented remarkable peculiarities because almost all the estimators provided estimates concentrated around 0.9 (true value) and 1 (
Figure 4). ET, CUE, ETEL, HD, ETHD, and their smoothed versions presented this peculiarity—
of the analyzed estimators. Despite not having this behavior, the over-identified GMM obtained the largest bias. In contrast, the just-identified GMM obtained one of the largest biases, with the mean and median values close to one in both cases. The main positive performances were registered by EL and SEL, which delivered concentrated estimates only around the true value and the lowest MSE and MAE among all the estimators (
Table 7). Despite reporting values close to true, ML generated a reasonably wide distribution around this value. BI, in turn, despite delivering a good MSE, returned more biased estimates than ML, and its MAE was at least twice greater than the MAE of the (S)EL in all the analyzed cases, even counting with an informative prior.
We impose a positivity restriction for
since this parameter appears squared in the moment conditions, allowing for two solutions: the true 0.007 and the opposite −0.007. Such a feature tends to explain why, in the case of the (S)ET, (S)CUE, (S)HD, (S)ETEL, and (S)ETHD and the GMM (the latter to a lesser extent) reported values close to zero (
Figure 5). In a way, the maximization algorithm could have “looked” at the other admitted value (−0.007), and it was prevented from reaching it due to the lower limit imposed. Several GEL/GMC estimators delivered atypical results due to estimates much higher than true in some replications. The ML distribution concentrated around 0.006 and 0.01 in the estimation using data of the
DGP I and around 0.008 and 0.013 in the case of
DGP IV. In the correctly specified case (
DGP I) and the case of contamination by only a sample-centered outlier (
DGP III), BI delivered one of the best results in terms of bias, MSE, and MAE. It was followed closely by (S)EL (
Table 8). In misspecification estimations using
DGP II and
DGP IV, the (S)EL estimator tended to outperform BI in terms of bias, MSE, and MAE. Considering the existence of a fixed parameter, (S)EL can be considered more robust than BI in the estimation of
since it was able to deliver more accurate estimates with a smaller MSE and MAE.
In general, the (S)EL estimator, in both just- and over-identified cases, and BI obtained the best performances in terms of bias, MSE, and MAE, in both correctly specified and misspecified models. Good performance of the EL in a correctly specified model was expected since it has good asymptotic properties in this situation. Because it incorporates prior information, BI generates good results, mainly due to the good configuration of the prior distribution—which was the case since both the prior and initialization of all the estimators were defined from the true values of the parameters. In addition, it is worth noting that, despite the intermediate comparative performance, the performances of the (S)CUE, (S)HD, and (S)ET estimators were shown to be good tools for DSGE estimation.
The GMM, mainly in the over-identified case, and ML presented performances considerably below a good part of their competitors due to inaccurate estimates. The non-zero empirical moment conditions of the over-identified GMM tend to compromise the performance when data are exposed to disturbances. Along this line, the GEL/GMC estimators have the advantage of always satisfying the constraints due to weighting by implicit probability to act directly on observations and not on moment conditions, as in the GMM. The ML estimator delivered concentrated estimates at points other than the true values, which may be associated with the presence of a local maximum in its objective function, showing that despite full information, this method can be reasonably unstable.
We can interpret the inferior results of the GMM by the fact that the generalized empirical likelihood and generalized minimum contrast methods, along with their smoothed variants, offer significant advantages over the generalized method of moments (GMM) estimators. One key advantage is that GEL methods do not require the estimation of a weight matrix, which is essential in the over-identified GMM. Estimating this matrix can be computationally complex and may introduce additional errors, whereas GEL inherently incorporates the moment conditions without relying on such a step, simplifying the process and reducing potential inaccuracies. Smoothed versions of GEL, such as the smoothed empirical likelihood (SEL), further improve estimation by achieving higher-order accuracy in finite samples. Furthermore, GEL methods are nonparametric, making them more flexible and requiring fewer assumptions about the underlying distribution of errors.
The finite-sample properties of the generalized method of moments (GMM) and generalized empirical likelihood (GEL) are further influenced by the methodology used to estimate the GMM parameters, particularly when employing two-step and iterated GMM approaches. These methodologies introduce additional sources of bias and inefficiency, which can significantly impact the reliability of GMM estimators in finite samples.
In the two-step GMM, the estimation process is divided into two stages. The first step involves obtaining an initial estimate of the parameters, often using an identity weighting matrix or some simple approximation. This preliminary estimate is then used to calculate the optimal weighting matrix in the second stage, which minimizes the asymptotic variance of the estimator. While this approach is theoretically efficient asymptotically, in finite samples, the second-stage weighting matrix depends on the first-stage estimates, introducing a form of feedback bias. This bias arises because the variability in the first-stage estimates propagates into the second stage, amplifying the estimation error. The magnitude of this bias grows with the number of moment conditions, as the estimation of the weight matrix becomes less stable when the dimensionality of the moment conditions is highly relative to the sample size.
The iterated GMM aims to mitigate this issue by repeatedly updating the parameter estimates and the weighting matrix until convergence. This iterative procedure improves asymptotic efficiency and reduces the dependence on the initial choice of the weighting matrix. However, in finite samples, the iterated GMM does not fully eliminate the bias introduced by the staged estimation process. The iterated updates can exacerbate finite-sample sensitivity, particularly when the moment conditions are weakly informative or when the sample size is small relative to the number of moments. Furthermore, the iterated GMM can become computationally burdensome, and convergence to a global optimum is not guaranteed in nonlinear settings, adding further practical limitations.
In contrast, generalized empirical likelihood (GEL) methods, such as empirical likelihood (EL), exponential tilting (ET), and continuous updating estimator (CUE), avoid these staged estimation processes entirely. GEL constructs its objective function directly based on the likelihood principle, which ensures that the weighting of the moment conditions is determined endogenously. This approach eliminates the need for pre-estimating a weighting matrix, thereby avoiding the feedback bias inherent in the two-step and iterated GMM. GEL’s continuous updating structure is particularly advantageous in finite samples, as it incorporates all the moment conditions simultaneously without relying on intermediate steps or iterative updates. The CUE, a variant of GEL, further optimizes the empirical likelihood function directly over both the parameters and the moment conditions, offering an efficient and bias-robust alternative to the iterated GMM.
The smoothing techniques applied in GEL also contribute to improved numerical stability, especially in situations where the moment functions are non-smooth or involve discontinuities. This is especially useful in complex models or irregular data structures, where optimization using traditional methods can be challenging. By addressing these issues, GEL and its smoothed variants provide a reliable and efficient framework for estimation, outperforming the GMM in terms of robustness, flexibility, and finite-sample accuracy.
The performance of ET- and HD-based estimators—(S)ETEL and (S)ETHD—was strongly compromised by atypical estimates in some replications. These estimators could deliver good results mainly when comparing the performance of different estimators under specification problems—especially the ETHD estimator that presents good theoretical properties under both correct specification and misspecification. However, it should be emphasized that other studies that consider other sources of misspecification must be made. In addition, other analyses that consider different initialization can be made to analyze the performance of these estimators in other situations. It would be interesting to analyze the choice of optimization methods used in the implementation of estimators, something outside the scope of this paper.
The exponentially tilted empirical likelihood (ETEL) estimator combines the strengths of empirical likelihood (EL) and exponential tilting (ET) to achieve desirable properties under both correct and incorrect model specifications. Ref. [
47] demonstrated that ETEL maintains the low bias of EL in correctly specified models while avoiding EL’s issues under misspecification. However, its finite-sample performance can be significantly compromised by atypical estimates caused by extreme values. This behavior often results in estimates clustering at the boundaries of the parametric space or diverging substantially from true values, highlighting its lack of robustness to outliers. Small sample sizes exacerbate this issue, increasing estimator variance and limiting its reliability, which underscores the need for regularization techniques to mitigate these limitations.
Similarly, the exponentially tilted Hellinger distance (ETHD) estimator was developed to address the lack of robustness in traditional Hellinger distance (HD) estimators under global model misspecification. Ref. [
3] showed that while HD estimators are efficient under correct specifications, they fail to maintain
-consistency when the model is globally misspecified. ETHD integrates ET to enhance robustness against such misspecifications; however, its finite-sample performance is similarly vulnerable to extreme values, leading to parameter estimates that deviate significantly from the true values. Additionally, ETHD’s sensitivity to tuning parameters and data characteristics, such as contamination and the underlying distribution of estimating functions, further challenges its practical application.
One of the difficulties related to using moment-based methods, as highlighted by [
7], is obtaining a sufficient number of moment conditions for estimating DSGE models with many parameters. Resorting to the definition of artificial moment conditions can generate problems in estimation, making it difficult to use those methods. Another limitation of moment-based methods is the impossibility of recovering latent variables as performed in state-space representation used by ML and BI. However, some GEL/GMC estimators can obtain good results even in situations of misspecification, as found in this paper.
Bayesian inference is a powerful tool for estimating dynamic stochastic general equilibrium (DSGE) models, offering a structured approach to incorporate prior knowledge and quantify parameter uncertainty. However, its effectiveness is closely tied to the specification of prior distributions. In situations where the available data are not sufficiently informative, the choice of priors can significantly influence the posterior distributions, potentially leading to results that reflect the imposed priors more than the underlying data. This issue is particularly pronounced in small-sample contexts, where limited data exacerbate the dominance of prior assumptions. Consequently, the reliability of the estimation outcomes may be compromised, as the posterior inferences may be more indicative of the prior configurations than the empirical evidence. Therefore, careful consideration and sensitivity analysis of prior choices are essential to ensure robust and credible parameter estimates in Bayesian estimation of DSGE models.
About BI, despite the advantages of using prior information already mentioned in [
13], it should be emphasized that if the data are not informative enough, the prior configuration can largely dominate the posterior distribution. Thus, sensitivity tests involving prior distribution are fundamental to guarantee the robustness of results. In addition, another disadvantage of BI is the time required to obtain the posterior distribution, which is usually considered high.
5. Conclusions
The estimation of dynamic stochastic general equilibrium (DSGE) models involves a tradeoff between computational efficiency and the thoroughness of parameter inference. Generalized empirical likelihood (GEL) and generalized method of moments (GMM) estimators are often favored for their computational efficiency. These methods rely on moment conditions derived from the model, allowing for parameter estimation without the need to solve the full likelihood function, which simplifies computations and reduces processing time. However, this efficiency may come at the cost of statistical efficiency, particularly in small samples, where GEL and GMM estimators can exhibit higher variance compared to maximum likelihood (ML) estimators.
In contrast, methods like maximum likelihood (ML) and Bayesian inference (BI) provide a more comprehensive framework for parameter estimation by utilizing the full likelihood function. Bayesian estimation, in particular, offers a systematic approach to incorporate prior information and quantify parameter uncertainty through posterior distributions. This process typically involves Markov chain Monte Carlo (MCMC) methods, which, while powerful, are computationally intensive. The need to sample from complex, high-dimensional posterior distributions can lead to substantial computational times, posing challenges for large-scale models or real-time policy analysis.
To evaluate the performance of different moment-based estimators belonging to generalized empirical likelihood (GEL) and generalized minimum contrast (GMC) families in the estimation of DSGE models, we performed a Monte Carlo analysis considering different data-generating processes to verify the results under both correct and incorrect specifications. As a benchmark, we consider the generalized method of moments (GMM), maximum likelihood (ML), and Bayesian (BI) estimators.
The main results found were the following: (i) the just- and over-identified empirical likelihood (EL) estimator, as well as the smoothed version (SEL), and Bayesian inference (BI), obtained, in this order, the best performances in terms of bias, MSE, and MAE, in situations where the estimated model has and does not have specification problems; (ii) continuous updating empirical likelihood (CUE), the minimum Hellinger distance (HD), exponential tilting (ET), and their smoothed versions presented intermediate comparative performance; (iii) the performance of exponentially tilted empirical likelihood (ETEL), the exponential tilting Hellinger distance (ETHD), and their smoothed versions was strongly compromised by atypical (distant of the true values) estimates in some replications; (iv) smoothed and non-smoothed versions of the GEL/GMC estimators showed very similar performances, so it is impossible to distinguish them; and (v) the GMM estimator, especially in the over-identified case, and the ML estimator presented poor performances due to the inaccurate estimates.
In general, the performance of some GEL/GMC estimators, more specifically EL and its version with smoothed moment conditions, was similar (and in some cases even higher) to the Bayesian estimator and superior to the GMM and ML estimators. We emphasize that EL delivering estimates as good as those of the Bayesian method is an outstanding result because the latter has the advantage of incorporating prior information. However, some difficulties associated with defining the sufficient number of informative moment conditions may make the use of GEL/GMC (and GMM) estimators unfeasible. Since GEL/GMC estimators always satisfy their restrictions, they are considerably more advantageous than the over-identified GMM. On the other hand, Bayesian inference has difficulties such as the prior distribution dominating the final result in small samples and the long time spent in estimation. Thus, GEL/GMC estimators can be good tools for DSGE estimation, given their good characteristics in the misspecification context and their easy and fast computational implementation.
The estimation of parameters using methods such as the generalized method of moments (GMM), continuous updating generalized empirical likelihood (GEL), and generalized minimum contrast estimators (GMC) is often challenged by numerical and optimization instabilities. These challenges primarily arise from the complexity of the optimization landscapes, sensitivity to initial conditions, and the potential presence of multiple local minima. Addressing these issues is critical for enhancing the reliability and robustness of these estimators.
In the GMM, numerical instabilities frequently stem from the estimation of the optimal weighting matrix. When the number of moment conditions is large relative to the sample size, the sample covariance matrix of the moments can become ill-conditioned, leading to unreliable parameter estimates. Two-step and iterated GMM procedures exacerbate these issues, as the feedback loop between the initial parameter estimates and the subsequent weighting matrices amplifies estimation errors. Furthermore, convergence to suboptimal solutions is a recurring problem, particularly in high-dimensional settings or when the moment conditions are weak.
GEL methods, such as the continuous updating estimator (CUE), are designed to address some of the shortcomings of the GMM by jointly estimating parameters and the weighting matrix. However, GEL is not immune to optimization difficulties. These methods often involve solving non-convex optimization problems that are highly sensitive to initial conditions, which increases the risk of converging to local optima. The computational complexity of GEL grows with the dimensionality of the parameter space and the number of moment conditions, further complicating its practical application in large-scale problems.
In the case of GMC, which minimizes a contrast function between the empirical and model-implied distributions, the choice of contrast function plays a crucial role. Poorly chosen contrast functions can lead to optimization surfaces that are flat or irregular, making it difficult for numerical algorithms to converge reliably. The estimation of implied probabilities adds another layer of complexity, as these probabilities often require iterative numerical procedures that can be prone to convergence issues, particularly when starting values are poorly selected.
The accuracy of parameter estimation in these methods depends heavily on the specification of moment conditions and the correct computation of implied probabilities. Misspecified moment conditions result in biased or inconsistent estimators, while errors in computing implied probabilities can further distort parameter estimates. In particular, the iterative procedures used in GEL and GMCE for computing implied probabilities may fail to converge in high-dimensional parameter spaces, complicating the estimation process.
To address these challenges, researchers have proposed various strategies. Regularization techniques, for example, can stabilize the estimation of the weighting matrix in the GMM, especially when dealing with a large number of moment conditions [
63]. By incorporating penalty terms into the optimization objective, regularization reduces overfitting and enhances numerical stability. Robust optimization algorithms [
64], including quasi-Newton methods, simulated annealing, and adaptive learning rate techniques, are also beneficial for GEL and GMC methods, as they mitigate sensitivity to initial conditions and improve the likelihood of converging to a global optimum. The careful selection of instruments, particularly in the GMM, enhances efficiency and stability by reducing the dimensionality of the problem and mitigating weak identification issues. Additionally, the adoption of advanced numerical methods and high-performance computing strategies, such as parallel processing, can alleviate the computational burden associated with these estimation procedures.
Moment-based estimators can be valuable tools for estimating parameters in dynamic stochastic general equilibrium (DSGE) models due to their flexibility and minimal reliance on distributional assumptions. However, these methods have notable limitations, particularly when dealing with the latent variables that are central to DSGE models.
These estimators rely on aligning theoretical model-implied moments with empirical data moments, but it does not directly estimate latent variables, such as structural shocks or unobservable state variables. This indirect approach can lead to inefficiencies in parameter estimation, especially in cases where latent variables significantly influence the dynamics of the model. Moreover, the inability to explicitly account for latent components introduces challenges in identifying structural parameters, as the available moment conditions may not fully capture the model’s underlying dynamics. This limitation is exacerbated when moment conditions are misspecified, potentially leading to biased or inconsistent parameter estimates.
To address these challenges, moment-based estimators can be augmented with filtering techniques, such as the Kalman filter or particle filters, which are well suited to handling latent variables. For linear DSGE models with Gaussian assumptions, the Kalman filter can be used to estimate the state-space representation, efficiently handling the unobserved components of the model. By combining moment-based methods with the Kalman filter, researchers can iteratively estimate structural parameters while accounting for the latent states, improving the robustness of the estimation.
In the case of nonlinear DSGE models or those involving non-Gaussian features, particle filters provide a flexible alternative. These filters approximate the posterior distribution of latent variables through sequential Monte Carlo methods using the estimated fixed parameters. Combining the GMM, GEL and GMC with particle filters allows for a more comprehensive treatment of latent variables, although this approach requires careful attention to computational demands and convergence properties. By integrating moment-based estimators with filtering techniques, it is possible to overcome the limitations of traditional moment-based estimators in the context of DSGE models. This combined approach enhances the ability to estimate structural parameters accurately while accounting for the latent variables that are integral to the model’s structure, leading to more reliable and efficient results.