Next Article in Journal
Consistency in Estimation and Model Selection of Dynamic Panel Data Models with Fixed Effects
Previous Article in Journal
Bayesian Approach to Disentangling Technical and Environmental Productivity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Approach to Model Verification, Falsification and Selection

Department of Economics Temple University, Philadelphia, PA 19122, USA
*
Author to whom correspondence should be addressed.
Econometrics 2015, 3(3), 466-493; https://doi.org/10.3390/econometrics3030466
Submission received: 10 March 2015 / Revised: 7 June 2015 / Accepted: 16 June 2015 / Published: 29 June 2015

Abstract

:
This paper shows that a qualitative analysis, i.e., an assessment of the consistency of a hypothesized sign pattern for structural arrays with the sign pattern of the estimated reduced form, can always provide decisive insight into a model’s validity both in general and compared to other models. Qualitative analysis can show that it is impossible for some models to have generated the data used to estimate the reduced form, even though standard specification tests might show the model to be adequate. A partially specified structural hypothesis can be falsified by estimating as few as one reduced form equation. Zero restrictions in the structure can themselves be falsified. It is further shown how the information content of the hypothesized structural sign patterns can be measured using a commonly applied concept of statistical entropy. The lower the hypothesized structural sign pattern’s entropy, the more a priori information it proposes about the sign pattern of the estimated reduced form. As an hypothesized structural sign pattern has a lower entropy, it is more subject to type 1 error and less subject to type 2 error. Three cases illustrate the approach taken here.
JEL Classification:
C15; C18; C51; C52

1. Introduction

In general, a demonstration that economics is a science would be comprised of showing that economics organizes its subject matter in some fashion that limits the configurations that may be taken on by the data. Organizing the subject matter in a fashion that limits the permissible realizations of the data may be termed “theory”. If the limits on the data required by the theory are not found, then the theory has been falsified (Popper [1]). The potential for falsification provided by propositions about the economy is termed by Samuelson [2] as the source of economics’ “meaningful theorems”.
For example, an allocation-of-time model might produce the result that more schooling results in greater lifetime earnings, ceteris paribus. It is the all-other-things-equal clause that produces the wrinkle. Angrist and Krueger [3] suggested that children’s dates of birth sort children into groups with more or less education for a given chronological age, and therefore we can observe the effect of schooling on earnings independent of ability. Butcher and Case [4] argue that it is gender of first-born-sibling that can be used to identify the value of schooling. These two papers might be thought of as (non-)nested alternatives. 1 Barron, Ewing and Waddell [6] consider the effect of high school athletics participation on lifetime earnings differences. Since they do not account for a quarter of birth or gender of first-born-siblings, their theory imposes zero restrictions on the theories of the first two papers. Not surprisingly, all three papers report statistically significant returns to educational activity, be it sports participation or years of schooling. All three of these papers use instrumental variables estimation. None of them have asked to what extent the alternative structural models restrict the outcomes in the first stage of estimation. The implicit first stage reduced form is not fraught with the statistical difficulties inherent in reporting structural results common to all three papers and would therefore be the proper seat for model falsification. 2 Determination of the permissible signs for the reduced form coefficients, given the structural model, is known as a “qualitative analysis”. To state the issue succinctly: It is quite possible to fit alternative “models” or “theories” to the data and find statistically significant results even though the proposed structural models could not possibly have generated the observed data. 3
This paper shows that a qualitative analysis can always be used in evaluating a model’s validity both in general and compared to other hypothesized models. Issues that can be approached via qualitative analysis are identified and manners of assessing them are illustrated. To do this, algorithmic approaches are used to process a structure’s sign pattern information in order to reach conclusions about a model’s scientific content, its potential for falsification (and associated type 1 error) and issues at stake in its acceptance (and associated type 2 error). 4 It is demonstrated that such an analysis can always potentially falsify the hypothesized structural sign patterns or support an analysis of the relative likelihoods of alternative structural hypotheses, if neither are falsified. It is also noted that a partially specified structural hypothesis can be sometimes falsified by estimating as few as one reduced form equation. Additionally, zero restrictions in the structure can themselves be falsified; and, when so, current practice proposes estimated structural arrays that are impossible. It is further shown how the information content of the hypothesized structural sign patterns can be measured using Shannon’s [7] concept of entropy. In general, the lower the hypothesized structural sign pattern’s entropy, the more a priori information it proposes about the sign pattern of the estimated reduced form. As an hypothesized structural sign pattern has a lower entropy, it is more subject to type 1 error and less subject to type 2 error. We want to emphasize that the techniques presented here are proposed as complements to, rather than replacements for, currently accepted econometric practice.
Following Samuelson [2], we propose that economic “theory” about some feature of the economy is expressed by an enumeration of n-many endogenous variables Y to be explained by m-many exogenous variables Z. Further, relationships among these variables are expressed by n-many relationships,
f i (Y, Z) = 0, i = 1, 2, …, n
A propos of the returns to schooling papers cited above, the issue of whether or not the system (1) presents a refutable hypothesis is studied by the method of comparative statics, as specified by a linear system of differentials associated with disturbing a solution (which we assume always exists) to the system (1),
j = 1 n f i y j d y j   +   k = 1 m f i z k d z k   =   0 ,   i   =   1 , 2 , , n
For the econometrician to bring the system (2) to the data, it is often assumed that the relationships in (2) are (at least locally) linear. In particular, that over the range of the data at issue, the signs of the partial derivatives in (2) are invariant. Adding unobserved disturbance terms expressed by the n-vector U, (2) is given by the equation below,
βY = γZ + δU
where β, γ and δ are appropriately dimensioned matrices. 5 Equation (3) is usually called the structural form. The disturbance U is a random vector with zero means and finite, positive definite covariance matrix. 6 Bringing Equation (3) to the data requires manipulation. To avoid a variety of statistical issues the system to be estimated is expressed as,
Y = π   Z + ψ U ,
where π = β−1γ. (4) is usually called the reduced form, from which the unknowns in Equation (3) are subsequently derived. Given this, the issue of falsification becomes that of the limitations on the entries of the estimated π = β−1γ, denoted by π , as derived from the theoretical proposals about the entries of {β, γ}.
When Samuelson [2] considered the issue in exactly these terms, he first noted that the theory might specify the directions of influence among the endogenous and exogenous variables, i.e., the sign patterns of β and γ. Given this, due to a “qualitative analysis”, it might be possible to go through the algebra of calculating π = β−1γ and finding that certain entries of π must have particular signs, based entirely upon the sign patterns of β and γ, independent of the magnitudes of their entries. If so, then when a π is realized, if the required signs do not show up, the theory has been falsified. 7 On reflection, Samuelson proposed that as a practical matter, a qualitative analysis was extremely unlikely to be successful. To work, all of the potentially millions of terms in the expansions of β’s determinant and cofactors necessary in computing β−1 would all have to have the same sign, independent of magnitudes. Samuelson viewed this as too unlikely to be taken seriously. Accordingly, he proposed other sources for limitations on the signs, of π, e.g., that β is a stable matrix or is derived from the second order conditions to a (perhaps constrained) optimization problem.
Nevertheless, a literature on the conditions such that a successful qualitative analysis could be conducted was developed (well summarized in Hale et al. [9]). For example, a liberal sprinkling of zeros among β’s entries can drastically reduce the number of nonzero terms in the expansions of its determinant and cofactors. Still, even so, the conditions for a successful qualitative analysis were found to be severe and are rarely satisfied. 8 Accordingly, there is virtually no tradition of assessing a model’s qualitative properties that might falsify entries in sgn π . Actually, Samuelson’s [2] other criteria are rarely explored as well. It is tempting to worry that the degree to which economic models are “scientific”; that is, falsifiable based on the data, is really not well in-hand.
The purpose of this paper is to demonstrate that the above view of the potential usefulness of qualitative analysis is miscast: starting with Samuelson’s [2] pessimism through the development of the extremely restrictive conditions such that entries of π can be shown to have particular signs, based upon a qualitative analysis. Instead, we show that a qualitative analysis can always be successfully conducted and that it provides important insight into the “scientific” content of an economic model and can always enumerate outcomes for sgn π that would falsify hypothesized sign patterns for the structural arrays.
In particular, it will be shown that:
-
a specification of the sign patterns of the structural arrays always limits the possible sign patterns that can be taken on by the estimated reduced form;
-
the falsifiability of hypothesized sign patterns for the structural arrays is independent of identifiability 9;
-
zero restrictions specified for the structural arrays can be falsified independent of the sign pattern of the nonzero entries. When this happens, multi-stage least squares will always be in error when providing estimates of the structural arrays, i.e., the sign patterns of the estimated structural arrays with zero restrictions imposed are impossible, given the sign pattern of the estimated reduced form;
-
a partially specified structural hypothesis sometimes can be falsified by a small number of estimated reduced form equations, even as few as one;
-
the conditional probability that an hypothesized structural sign pattern is actually true when not falsified can be estimated using Bayes’ formula;
-
the relative likelihood of competing structural hypotheses can be assessed, given the sign pattern of the estimated reduced form, based upon their qualitative characteristics; and,
-
hypothesized structural arrays can be assessed for their information content using Shannon’s [7] measure of entropy.
In the next section, a method of qualitative falsification is presented, including using a measure of entropy to measure the information content of a structural hypothesis. In the section following that, issues in “verifying” a structural hypothesis that is not falsified are discussed. Section 4 works through three case studies. The last section provides a summary. An appendix provides additional detail about our method of analysis.

2. Qualitative Falsification

Lady and Buck [8] (2011) and Buck and Lady [11] (2012) showed that the falsification of a qualitative specification of a model does not require that any individual signs of β−1 (or more generally sgn π) be signable based upon sgn β (and more generally, also sgn γ). Instead, patterns of signs in sgn π may not be possible, even though no individual entry is signable. In this section, this principle is generalized compared to this earlier work. In the next section, the new issue of verification is presented.
An easy example to use for explication is β: 2 × 2 without zero entries. For this and all of the examples below, it is assumed that β is irreducible 10 and not singular. There are 16 possible sign patterns that can be hypothesized for β. Of these, eight have an entry with a sign the opposite of the other three. For these cases, the determinant of β is signable, independent of the magnitudes of its entries. Since the adjoint of β is signable, for these cases there is only one possible sign pattern for β−1. For the other eight cases, there are only two possible sign patterns for β−1; however, no entry of sgn β−1 is the same for these two possibilities, since they are the negative of each other. Accordingly, for any of these eight cases, an hypothesized sgn β would be falsified by sgn π (assuming for the moment that γ = I), if it took on one of the 14 sign patterns that are not possible for the hypothesized sgn β, even though no individual entry of sgn β−1 is signable.
For larger arrays (and/or for γ ≠ I), although the same principles apply, the algorithmic burden of revealing them quickly becomes more difficult. To cope with this, we developed a Monte Carlo algorithm for investigating the possible sign patterns of the reduced form based upon a qualitative specification of the structural form. 11 In brief, sgn β and sgn γ (if need be) are randomly given quantitative values for their entries consistent with the hypothesized sign pattern, π = β−1γ is computed, and the sign pattern outcome of this computation is saved. The iterations/repetitions can be done millions of times. For sufficiently small systems, the set of all possible sign patterns that the reduced form might take on, given the hypothesized sign pattern of the structural arrays, can be tabulated with a vanishingly small potential for error as the number of repetitions increases. 12 A central point of the analysis is that this set of allowed sign patterns always limits the sign patterns the reduced form can take on. Given this, there are always sign patterns that the reduced form array might take on that are impossible, given the hypothesized structural sign patterns, and if so, the given structure is thus falsified.
Consider as the next example the case of hypothesizing that sgn β is 3 × 3 and all positive and γ = I. 13 In principle, there are 512 such 3 × 3 sign patterns that sgn β−1 can take (barring zeros as we always do for the examples we provide, although this is not necessary). Of these, only 102 were found by the Monte Carlo in tens of millions of iterations. Given this, it appears that 410 sign patterns for sgn β−1 are impossible for the hypothesized sgn β. The analytic basis for some sign patterns being impossible can be quickly developed. Let the qualitative inverse of sgn β be defined as:
Definition (qualitative inverse). Sgn π is a qualitative inverse of sgn β if and only if there exist magnitudes for the entries of β, consistent with the given sgn β, such that sgn β−1 = sgn π.
Given this, it is immediate that the proposed sign patterns must be such that magnitudes can be assigned so that:
  • Sgn π is a qualitative inverse of sgn β only if it is possible that i) βπ = I; and, ii) πβ = I.
This possibility can be algorithmically investigated directly. For β 3 × 3 and all positive, of the 512 possible 3 × 3 sign patterns that sgn π might take on, only 216 sign patterns satisfy i) or ii) individually; and, there are only 102 sign patterns that satisfy both. These are what the Monte Carlo found. Since i) and ii) are clearly necessary for the case here, their satisfaction also turned out to be sufficient for sgn π to be a qualitative inverse for the hypothesized sgn β. As we will see below, satisfaction of both of these conditions may not always be sufficient. From the standpoint of falsification for β all positive and of any dimension, it is immediate that a proposed, i.e., estimated, sgn π cannot satisfy both i) and ii) unless it has a positive and negative entry in each row and column. More generally, sgn π cannot be a qualitative inverse of a proposed sgn β unless, for i) above the ith row of β and the jth column of π share a pair of common nonzero entries of the same sign for i = j and both this and another pair of common nonzero entries of the opposite sign for i ≠ j; and, also for the ith row of π and the jth column of β for ii) above.
For the special case that we are considering, why not simply set β ^   =   π 1 and compare sgn β ^ with the hypothesized sgn β? If any signs are different, then the hypothesized sgn β has been falsified and there is no need to worry about the derivations immediately above. This could indeed be done for the exactly identified case, which given our simplifying assumptions requires that β have no zero restrictions. However, even for this very special case, if sgn β ^ ≠ sgn β, this outcome might not be taken as falsification. It might be that sgn π is a qualitative inverse of both the data based sgn β ^ and the hypothesized sgn β. Given this, further analysis, not to speak of empirical work, might be viewed as necessary. This situation is discussed in the next section. Generally, for n ≠ m and/or γ ≠ I setting β ^   =   π 1 is not possible and/or appropriate. For these more general cases, if (3) is exactly identified, then unique arrays β ^   a n d   γ ^ can be found based upon π ; nevertheless, as pointed out above, this may not lead to falsification, at least not without further analysis.
For the special case at issue here, if zero restrictions are imposed upon sgn β, then (3) is over identified and the zero restrictions will be imposed upon the derivation of β ^ . Under these circumstances, there is more than one way to recover β ^ from π . For these, sgn β ^ may not be the same. Even more serious, the zero restrictions themselves can be falsified. If this is the case for the sgn π that was found, whatever method is used to recover β ^ results in a mistake if these zero restrictions are imposed; namely, the estimated array β ^ is impossible, given sgn π ; i.e., sgn π is not a qualitative inverse of sgn β ^ .
As an easy to follow example, let the hypothesized β be given by,
sgn   β   =   [ ? ? ? ? ? ? 0 ? ? ] .
For this array, there is the single zero restriction, β31 = 0. The remaining entries marked “?” are nonzero, but can be either positive or negative. For this case, assume that, based on the data 14
sgn   π ^   =   [ ? ? ? + ? + + ? ] .
This result cannot be the qualitative inverse for any array for which β31 = 0, since the (1,3) cofactor is signable, independent of magnitudes. Of the 512 possible sign patterns that sgn π might have taken on from the data, 32 of them have the above 2 × 2 sign pattern corresponding to the (1,3) cofactor. As discussed above, altogether there are eight such signable sign patterns for this cofactor. As a result, 256 of the possible 512 sign patterns that sgn π might take on falsify the given zero restriction, independent of the signs of the other entries of the proposed sgn β or estimated sgn π . If any of these show up for sgn π , then however β ^ is derived, it is a mistake, i.e., impossible, if the zero restriction is imposed. 15
To dramatize the potential for the zero restrictions themselves to falsify a structural hypothesis, and potentially lead to an error when deriving β ^ from π , consider as an example an inference structure for β (assumed to be irreducible 16) that contains only one cycle of inference. Consider this case for n = 5,
sgn   β = [ ? 0 0 0 ? ? ? 0 0 0 0 ? ? 0 0 0 0 ? ? 0 0 0 0 ? ? ]
As before, the entries marked “?” are nonzero, but may be of any sign. The number of possible qualitative inverses for this array can be derived directly. In particular, the expansion of each cofactor of β includes only one term, since there is only one path of inference between any pair of (endogenous) variables. There are 10 nonzeros which can take on 210 = 1024 distinct sign patterns. Accordingly, adjoint β can only take on 1024 distinct sign patterns. These in turn can be divided into two groups for which the patterns in one group are the negative of the patterns in the other group. As it happens, the determinant is signed for half of the 1024 cases, but this does not make any difference. Given the possible sign patterns of the adjoint, there are only 1024 possible sign patterns for sgn β−1, regardless of how the determinant’s (nonzero) sign plays out. Yet, for a 5 × 5 array, there are 33,554,432 possible sign patterns, barring zeros. For this case, of these, 33,553,408 are impossible, based upon the zero restrictions on β alone (plus γ = I). For this inference structure, the proportion of qualitative inverses among the possible sign patterns decreases significantly as n increases, since there are 2nn possible sign patterns for which only 22n are qualitative inverses. It is worth emphasizing the points just made. If the zero restrictions of the structural hypothesis are falsified by the sign pattern of the estimated reduced form, then the estimated structural arrays with these zero restrictions imposed are impossible, given the sign pattern of the estimated reduced form. Since current practice does not include conducting a qualitative analysis, this leaves the prospect of the outcome of many estimations as unknowingly being utterly impossible.
If two or more entries of the adjoint of β are signable, which could well be the case if the nonzeros in β are sparse, then a criterion in addition to i) and ii) above for the falsification of β based upon sgn π is in-hand; namely, the entries of sgn π that correspond to the signable entries of the adjoint must always have the same or opposite signs as appropriate, or the hypothesized β is falsified. This circumstance can be robustly detected by the Monte Carlo and algorithmically applied directly. To organize this idea formally, let B = [Bij] be the adjoint of β.
Definition (adjoint-consistent). If sgn Bij = sgn Buv (resp. sgn Bij = −sgn Buv) independent of magnitudes, then sgn π is adjoint-consistent with sgn β if and only if sgn πij = sgn πuv (resp. sgn πij = −sgn πuv).
Thus, a third criterion would be,
iii)
An estimated sgn π is a qualitative inverse of sgn β only if the estimated sgn π is adjoint consistent with sgn β.
Once two or more signable entries in sgn B are enumerated, satisfaction of iii) can be algorithmically investigated directly. Taken together, i), ii), and iii) are necessary. As shown in the example below, they may not be sufficient. The ultimate requirement for qualitative inverses of sgn β would be the existence of a solution to a system of inequalities, as given below.
For a given sgn π , consider the system(s) of inequalities as written out symbolically as the expansions of β’s cofactors and determinant,
sgn   B   =   sgn   π ^ ,   and   det   β   >   0 ;   and / or ,   sgn   B   =   sgn   π ^ ,   and   det   β   <   0
It is immediate that a given empirical sgn π ^ is a qualitative inverse of a proposed sgn β if and only if at least one of the systems in (5) has a solution. Even if i), ii), and iii) are satisfied, a given empirical sgn π may not be such that one or both of the systems (5) has a solution. The necessary and sufficient algorithmic method for assessing if (5) is satisfied, if it exists, is beyond the scope of this paper. It was the resolution of this issue (among others) for which the Monte Carlo was developed as a heuristic. For a given empirical sgn π , if the hypothesized sgn β is quantitatively sampled repeatedly and β−1 computed, and it is repeatedly found that, given this, the given sgn π ≠ sgn β−1, then the inference is that the given sgn π is not a qualitative inverse of the hypothesized sgn β. The chances of incorrectly falsifying the hypothesized structure are made ever smaller by increasing the number of quantitative Monte Carlo iterations taken, but the chances are never absolutely zero. For this reason, once falsification is strongly suspected based upon the Monte Carlo, a direct investigation of (5) can be attempted, hoping to analytically confirm what the Monte Carlo suggests. Examples of this approach are given in the appendix of Buck and Lady [11].
As an example of applying all of the criteria presented above, consider the case (with γ = I),
sgn   β   =   [ + + 0 + + 0 0 + + 0 + + ]
This sign pattern, negative main diagonal entries and non-negative off diagonal entries, is an example of a Metzler [12] matrix, a form corresponding to the excess demand functions for multimarket equilibria for which all commodities are (weakly, if zeros are allowed) gross substitutes. The four zero restrictions are imposed here for the purpose of demonstrating all of the falsification criteria discussed above.
Application of the above criteria revealed the following: Of the 65,536 possible 4 × 4 sign patterns for sgn β−1, only 4096 are possible, given the zero restrictions, independent of the signs of the nonzeros. To reiterate this point yet again, if the sign pattern of the estimated reduced form had any of the other 61,440 sign patterns, then any sign pattern proposed for sgn β ^ derived from the observed reduced form with these zero restrictions imposed would be a mistake. Direct analysis revealed that only 625 of the possible sign patterns for sgn β−1 satisfied i) above, 400 satisfied ii), and only 28 satisfied both. Further analysis revealed that entries of β adjoint were signable, as given below,
Adjoint   β   =   [ * * a a b b a a a a c c a a * * ]
In the above, entries marked “a”, “b”, and “c” always have the same sign. Taking this into account for the 28 sign patterns that satisfied both i) and ii), only 20 satisfied iii), i.e., presented the pattern of equal signs given above. Of these, the Monte Carlo only found 18. Since the outcome of i), ii), and iii) can be inspected directly, the findings are clearly necessary. Of 20 sign patterns that satisfied all of these necessary conditions, two were nevertheless not solutions of (5). For these two, further analysis would be appropriate to confirm that the finding falsified the hypothesized structure, i.e., to uncover the contradiction that prevents the satisfaction of (5).
A feature of the above analysis was the complete specification of the qualitative structure of (3) and (4). It should be noted that falsification might be the case if only parts of the structure are qualitatively specified and less than all of the reduced form equations are estimated. For example, if the sign pattern of only one column of β is hypothesized, then if only one row of π is estimated, the (partial) structural hypothesis has been falsified if the signs of the entries of that row are the opposite from the corresponding nonzeros in the hypothesized column of β, since the criterion ii) above would not be satisfied. Other examples of falsifying partially specified structural models can be readily proposed.
The simplifying assumptions for the analysis so far, n = m and γ = I, were made to permit the principles of the analysis to be readily revealed. Relaxing these simplifications increases the analytic burden of the analysis; nevertheless, a qualitative specification of the structural form will inevitably impose restrictions on β−1 that always result in these restrictions being translated to the reduced form in the more general cases.
For the structural form exactly- or over-identified, estimates β ^   a n d   γ ^ can be derived, although they may not be unique and may be otherwise problematical. Accordingly, the sign patterns of such estimates may not be viewed as decisive from the standpoint of falsification via hypothesis testing; unlike qualitative falsification, which is decisive (absent large sampling errors). For the under-identified structural form, such estimates cannot be made; nevertheless, qualitative falsification can proceed as before.
Consider the structural hypothesis:
sgn   β   =   [ + + + + + + + + + ]     a n d   sgn   γ   =   [ + + + + + + + + + ]
For this under-identified structural form, estimates for β ^   a n d   γ ^ cannot be made. Nevertheless, qualitative analysis shows that of the 512 3 × 3 reduced form sign patterns, only 343 are possible for this sign pattern for the structural form. The structural hypothesis would be falsified if any of the other 169 sign patterns were estimated for the reduced form.
Qualitative Falsification and Type 1 Error. As noted above, a structural hypothesis may be rejected because the sign pattern of the estimated reduced form is not found among those admissible reduced forms generated by the Monte Carlo algorithm that we have employed. Given this, further investigation is called for in order to reveal the inconsistency as related to a solution to (5). If this is found, i.e., if the sign pattern of the estimated reduced form is demonstrably impossible, given the hypothesized structural sign pattern, the result may be due to errors in the data, and to reject the structural hypothesis would constitute a type 1 error. The issue of this subsection is to assess the propensity for a structural hypothesis to be incorrectly rejected.
As an example (to be used in the next section as well), let the structural sign patterns given below be termed “system #1”:
sgn   β   =   [ + + + 0 + + + + + + + + 0 0 + + ]   a n d   sgn   γ   =   [ + 0 0 0 0 + 0 0 0 0 + 0 0 0 0 + ]
Of the 65536 4 × 4 sign patterns that the estimated reduced form might take on, simulation results report that only 535 are possible, which constitute less than 1% of the total number. The Monte Carlo simulation draws the absolute values of the nonzeros in the above arrays from a uniform distribution over the open interval [0,10]. This is sufficient to potentially find all possible reduced form sign patterns. However, suppose the distributional rule for populating β and γ is not otherwise supported by data or analysis, which can make a difference as pointed out below. An indicative method to portray the potential for type 1 one error due to sampling error is to draw “error terms” from a normal distribution with mean zero and add these to each nonzero entry used to populate the β and γ arrays, and then compute π ˜ = β−1γ. If the sign pattern found for π ˜ is not one of the (535 for system #1) possible sign patterns, then the associated falsification is a type 1 error due to sampling error 17. For this example, this can only happen if the error term that was drawn puts a negative value in one of the array entries when added to the sampled quantitative value that was supposed to be positive. Such a “wrong” sign is intended to represent errors in the data. As the number of possible sign patterns permitted by the structural sign patterns is smaller, the potential for type 1 error due to sampling errors is larger. For system #1 above, with the standard deviation of the “error term” = 0.2, there were 4.68% of 500,000 quantitative Monte Carlo repetitions which falsified system #1’s sign pattern. This result is dependent in part on the distributional assumption employed in the Monte Carlo repetitions. The influence of this approach can be further investigated by the Monte Carlo simulation, but it is beyond our scope to do so.
For the purpose of comparison, a second structural system was formulated, basically by moving three zero restrictions from γ to β and otherwise changing two signs in β. Given this, system #2 is specified as:
sgn   β   =   [ + 0 0 0 + + 0 + 0 + + 0 + + ]   a n d   sgn   γ   =   [ + 0 0 + 0 0 0 + 0 0 0 0 + ]
The Monte Carlo reports that for system #2, that there are only 65 possible sign patterns for the reduced form. From the above discussion, since there are fewer possible reduced form sign patterns (or better yet, a smaller proportion of the total), the potential for type 1 error would be larger for system #2 compared to system #1. This result is reflected in the results tabulated in Table 1 below.
Table 1. Proportion of Type 1 Error.
Table 1. Proportion of Type 1 Error.
Standard Deviation of the Error DistributionProportion Falsified (500,000 Samples)
System #1System #2
0.20.04680.0672
0.40.09760.1395
0.60.14200.1964
0.80.18470.2604
10.23660.3256
Type 1 Error and Entropy. The result of the last subsection can be generalized somewhat. Consider that for the sign pattern of an n × m reduced form array, barring zeros, that there are nm bits of information (say “1” for “+” and “0” for “−“), one bit for each entry. Given a hypothesized structural sign pattern, the possible reduced form sign patterns are limited to Q < 2nm; and for these, given the algebra of computing π = β−1γ and the distributional rules for assigning values to the nonzeros of {β, γ}, each possible reduced form sign pattern has a particular likelihood of occurrence. Let Fi be the frequency of occurrence of the ith possible reduced form sign pattern (see below for how “i” is computed). Given this, the entropy of the distribution (Shannon [7], see also Cover and Thomas [13]) is given by:
E n t r o p y ( sgn β ,   sgn γ )   =     i     Q F i log ( F i )
where log(Fi) is to the base 2. For example, for one simulation of system #1 above with 3,000,000 samples taken, the entropy of the resulting distribution of (the 535 possible) reduced form sign patterns was 8.18. The unit of this measure is bits. It measures the information achieved by estimating the reduced form. The maximum entropy is 16 for the 4 × 4 sign pattern (barring zeros) and would be the model’s entropy if all of the 65,536 4 × 4 reduced form sign patterns were possible and equally likely. The minimum entropy is zero if only one sign pattern for the reduced form is possible. Our algorithm for drawing the quantitative realizations of the structural sign pattern is designed to find the possible reduced form sign patterns, but has no econometric justification for the distributional (uniform) rule used in the iterations. Nevertheless, we will use the frequency distributions found in the analysis below. If a default entropy is computed which “simply” treats all of the possible reduced form sign patterns as equally likely, then doing this for system #1 causes the entropy measure to increase a small amount to 9.06. The key item driving the measure is the number of possible sign patterns, as opposed to the frequency distribution of the possibilities.
As the entropy measure is high, the structural hypothesis provides less information about the expected outcome of the reduced form’s estimated sign pattern, and conversely. To revise the measure to be larger as the information provided by the structural hypothesis is greater, the following amended measure is proposed,
I N F O % ( sgn β ,   sgn γ ) )   =   100 ( 1     E n t r o p y ( sgn β ,   sgn γ ) n m )
For system #1 this comes out to 48.94% and for system #2 (which is more limiting) it comes out to 66.4%. Accordingly, system #2 provides more a priori information about the expected outcome of the estimated reduced form’s sign pattern (only 65 are possible compared to 535 for system #1) and thus is more susceptible to type 1 error compared to system #1.

3. Qualitative Verification

Although contingent upon the propensity for type 1 error, qualitative falsification is otherwise entirely decisive. Given the hypothesized sign pattern of the structural form, either the estimated reduced form sign pattern is possible, or it is not. If not, subject to error, the structural hypothesis has been falsified.
On the other hand, if the sign pattern of a structural hypothesis is not falsified, this does not establish that the structural hypothesis is “true.” Indeed, as Samuelson (op. cit.) pointed out, if degenerate hypotheses are allowed, there are a virtually unlimited number of hypothesized structural forms that could be consistent with, i.e., not falsified by, a particular outcome for the estimated reduced form. In fact, science does not establish “the” true hypothesized system. Instead, a system may be used that is persistently consistent with the data, perhaps chosen by, say, the principle of Occam’s Razor; but always, such a system awaits its replacement by another system that matches its explanatory power to the present point, but provides further explanation of data that the initial system cannot explain, e.g., such as features of the orbit of Mercury around the Sun that were explained by general relativity, but were not explained by Newtonian physics. 18 As a result, what we will term a model’s “verification” is an open-ended, in some sense never-ending, exercise of continuous testing, evaluation, and potential comparison with competing hypotheses. Open-ended or not, the point of this section is to present a number of procedures due to which a qualitative analysis can help evaluate the validity of an hypothesis and compare it to competitors.
The analysis that we develop here will be conditional upon the lists of endogenous and exogenous variables. It will always be assumed that β is irreducible and that γ is otherwise configured such that the n × m reduced form has no logical zeros; and further, we will assume that there are no zeros otherwise in the estimated reduced form. In this frame of reference, a type 2 error is the acceptance of a hypothesized structural form due to its not being falsified, even though the hypothesized structure is not correct. The propensity for a type 2 error can be assessed in general or in comparison to one or more competing hypotheses.
The latter issue might be addressed by assessing the likelihood that a hypothesized structural sign pattern might fail to be falsified by accident, i.e., if the sign pattern of the estimated reduced form were in some sense random and was found to be a member of the permissible set of reduced forms. From this perspective, it is intuitive that system #2 is less likely to be randomly accepted compared to system #1, since the number of possible outcomes for the sign pattern of the reduced form is substantially less for system #2. There is nothing wrong with this insight, but the analysis can be pushed further, specifically in terms of estimating the conditional probability for each system’s validity, given the outcome of the estimated reduced form sign pattern, p (system #1 or #2 sgn π ). Given this, the likelihood of type 2 error is one minus this conditional probability.
For the two systems proposed for illustration here, of the possible reduced form sign patterns for each, 28 are common to both of them. We will develop our illustration in terms of these 28 sign patterns. To start, the Monte Carlo (3,000,000 repetitions for each system) provides an immediate estimate of the conditional probability of a reduced form sign pattern for each system as provided by the frequency with which that sign pattern appeared. These results are given in Table 2 below for the two hypothesized systems. The reduced form sign pattern for the first row of Table 2 is:
sgn     π ^ ( 8914 )   =   [ + + + + + + ]
The base 2 index for this sign pattern is 0010001011010010 (“0” for “−” and “1” for “+”) which equals 8914 in base 10. The base 10 indices in the other rows are computed in the same way and are found in the first column of the table. In the second column of the table, the sign pattern of the reduced form at issue is written out row by row, laid end-to-end.
Not surprisingly, the frequencies for system #2 are uniformly larger than system #1 since significantly fewer sign patterns are possible for this system, although exceptions are possible.
Table 2. Frequency of Reduced Form Sign Patterns in Sample.
Table 2. Frequency of Reduced Form Sign Patterns in Sample.
π Base 10 IndexSign Pattern of the Reduced Form Row by RowFrequency in Sample of 3,000,000
System #1System #2
CountFrequency: p( π |sys #1)CountFrequency: p( π |sys #2)
8914− − + − − − + − + + − + − − + −31010.00103844550.02815
8915− − + − − − + − + + − + − − + +16610.00055701680.02339
9878− − + − − + + − + − − + − + + −24480.00082170470.00568
9879− − + − − + + − + − − + − + + +36420.00121280490.00935
9938− − + − − + + − + + − + − − + −23290.00078147310.00491
9939− − + − − + + − + + − + − − + +45760.00153848420.02828
21925− + − + − + − + + − + − − + − +19330.000641225560.04850
26262− + + − − + + − + − − + − + + −108090.00360510520.01702
26263− + + − − + + − + − − + − + + +65900.00220427240.01424
38189+ − − + − + − + − − + − + + − +17030.00057188890.0063
38249+ − − + − + − + − + + − + − − +8790.00029367550.01225
39273+ − − + + − − + − + + − + − − +79960.00267773550.02578
40237+ − − + + + − + − − + − + + − +24420.0008187280.00291
40297+ − − + + + − + − + + − + − − +42910.00143165010.0055
41562+ − + − − − + − − + − + + − + −5600.00019443390.01478
41563+ − + − − − + − − + − + + − + +10610.00035993350.03311
41682+ − + − − − + − + + − + − − + −11370.00038430870.01436
41683+ − + − − − + − + + − + − − + +17130.00057288380.00961
42586+ − + − − + + − − + − + + − + −28840.00096202510.00675
42587+ − + − − + + − − + − + + − + +76930.002561654800.05516
42646+ − + − − + + − + − − + − + + −52290.001741820540.06068
42647+ − + − − + + − + − − + − + + +105180.00351985040.03283
42706+ − + − − + + − + + − + − − + −78330.00261482810.01609
42707+ − + − − + + − + + − + − − + +264050.008807532000.02511
43610+ − + − + − + − − + − + + − + −20290.00068707590.02359
43611+ − + − + − + − − + − + + − + +16900.00056766900.02556
54573+ + − + − + − + − − + − + + − +12800.00043566230.01887
56621+ + − + + + − + − − + − + + − +71700.00239554210.01847
The estimates in columns 4 and 6 can be used in computing the desired conditional probability using Bayes’ formula:
p ( s y s t e m # 1   o r # 2   | sgn π ^ ) = p ( sgn π ^   | s y s t e m # 1   o r   # 2 ) p ( s y s t e m # 1   o r   # 2 ) p ( sgn   π ^ )
To use this formula, in addition to the estimates of the conditional probabilities provided in Table 2, estimates of the prior probabilities for each system and the reduced form sign pattern must also be made. Presumably, in a particular empirical context, there would be reasons for assigning these values. Here, for the sake of illustration, we will proceed as follows:
Consider that for the 32 entries in β and γ, the two hypothesized systems agree on the signs for 23 of them. Given this, for all combinations of the nine disputed signs, there are 512 (= 29) possible structural sign patterns. Assume that all of these are equally likely; hence, the prior probability for each such system = 1/512 = 0.00195, including systems #1 and #2. These 512 structural sign patterns might be termed the “universe” from which the specific structural sign patterns were selected, i.e., as what the theory dictates and (evidently) leaves undecided. 19 This universe can be sampled and the frequency found for each of the 28 reduced form patterns enumerated in Table 2. This was done for 300,000,000 Monte Carlo iterations and the results are reported in the Appendix (along with the work up for computing the conditional probabilities for each system for each reduced form sign pattern as given in Equation (8)). The estimates found for these conditional probabilities are given in Table 3 below.
Table 3. Estimated Conditional Probabilities for System #1 and System #2.
Table 3. Estimated Conditional Probabilities for System #1 and System #2.
sgn π Base 10 IndexP(System #1|sgn π )P(System #2|sgn π )
89140.000440.01214
89150.000140.00597
98780.001880.01303
98790.000850.00656
99380.001080.00679
99390.000200.00372
219250.000100.00725
262620.002910.01377
262630.000520.00334
381890.002780.03071
382490.000180.00773
392730.002210.02130
402370.002550.00915
402970.002160.00831
415620.000160.01259
415630.000130.01225
416820.000280.01057
416830.000090.00152
425860.000820.00580
425870.000670.01454
426460.000650.02276
426470.000440.00413
427060.002060.01270
427070.000740.00211
436100.000430.01508
436110.000230.01043
545730.000150.00676
566210.000570.00439
Inspection of Table 3 reveals that for each of the 28 reduced form sign patterns, system #2 is more likely than system #1. Since the prior probabilities of each system and the reduced form sign pattern are the same for each system, this inequality will be the same as the conditional probabilities for each reduced form sign pattern with respect to each system. Accordingly, relative likelihoods for the systems can be determined based upon the Table 2 frequency distributions even if the prior probabilities of the systems and the reduced form sign pattern cannot be estimated. The generality brought forward by these results is that lower entropy systems, which provide more information about the reduced form sign patterns, will be more likely if not falsified and their acceptance will be less prone to type 2 error regardless of the distribution of the structural error.

4. Examples

Three examples of the above analytic point of view are presented in this section. The first is a (famously) under-identified model of the relationships among advertising, market concentration, and price-cost margins. The second is the Oil Market Simulation (OMS), a model of multi-market equilibria that was utilized by the U.S. Department of Energy to determine the world oil price as an input to its long term forecasting system. Remarkably, the sign pattern of the reduced form for this model is entirely determined by the hypothesized sign patterns of the structural arrays. The third example is Klein’s Model I. This is a concise, macroeconomic model that has been widely utilized for pedagogical purposes in both the econometrics literature and the literature on qualitative analysis. We consider estimates of the structural forms for all of these models. Utilizing the Monte Carlo to simulate the corresponding signs and zero restrictions hypothesized for the structural arrays, we found that the sign pattern of the estimated reduced form was not possible given the hypothesized sign patterns of the structural arrays. Further, in all three cases, we found that the sign patterns of the estimated reduced forms were not possible for any structural sign patterns that imposed the zero restrictions proposed for the structural arrays, regardless of the signs of the nonzero entries in these arrays. In some, but not all cases, we can point to the inconsistencies responsible for rejecting the estimated reduced form sign pattern. For the cases in which we are unable to account for the falsification, we note that the point of using the Monte Carlo is to enable such an analysis numerically, when a more explicit, analytical approach is not readily practicable.

4.1. An Under-Identified Model of Industrial Organization

A model of the relationships among advertising, concentration, and price-cost margin was presented by Strickland and Weiss [15]. In the econometrics literature, e.g., Amemiya [16], Greene [17] and Gujarati [18] the paper is cited because the model fails the rank test for identification. The model is also presented in texts in industrial organization, e.g., Greer [19], Scherer and Ross [20], and Carlton and Perloff [21]. The model is also subjected to a qualitative analysis in Buck and Lady [22]. The model is expressed as βY = γZ, where the Y vector of endogenous variables has three entries: advertising divided by sales (Ad/S), concentration (C), and the price cost margin (M); and, the Z vector of exogenous variables has six entries: growth of the industry (Gr), a dummy for durable goods (Dur), minimum efficient scale divided by sales (MES/S), capital stock divided by sales (K/S), geographic dispersion of the industry (GD), and the percentage of industry sales that is represented by consumer demand (CD/S). The sign patterns of the structural arrays proposed by Strickland and Weiss are:
sgn   β   =   [ + + + 0 + + ] ;   a n d ,   sgn   γ   =   [ ? 0 0 0 0 0 0 0 0 0 + 0 ]
The “?” for γ12 indicates that the authors were not sure about this sign; and, for the analysis reported on here, this entry is set both positive and negative in the quantitative Monte Carlo repetitions, each about half the time. In Buck and Lady [22], the reduced form of the model was estimated using data derived from the 1992 Census of Manufacturers and the corresponding National Input-Output Tables. 20 The sign pattern of this result is:
sgn   π ^   =   [ + + + + + + + + + + + + ]
In Buck and Lady [22], 10 of the 18 entries of the reduced form could be signed assuming that β was a stable matrix (which meant that its inverse would have all negative entries). Given this, the analysis proceeded based upon traditional qualitative principles directed at the signs in the reduced form individually. This was not done here. Instead, the Monte Carlo proceeded with quantitative draws for the structural sign patterns above, including the ambiguous (nonzero) sign of γ12. It was found that no entry of the reduced form was signable analytically. Nevertheless, in tens of millions of iterations, the Monte Carlo only found 18 sign patterns for the corresponding reduced form. This result shows that the qualitatively specified structural hypothesis imposes extreme limits upon the possible reduced form sign pattern, since in principle there are 262,144 possible 3 × 6 sign patterns, barring zeros. As a default, assuming that each possible sign pattern is equally likely, the information measure yields a value of INFO% = 76.8, i.e., the hypothesized structural sign patterns (if true) provide 76.8% of the information provided by the fully specified reduced form sign pattern. As discussed, for a hypothesis with low entropy (i.e., high information content), the likelihood of type 1 error is relatively high and the likelihood of type 2 error is relatively low. Significantly, the sign pattern of the estimated reduced form was not one of the 18 found, falsifying the structural hypothesis. 21
We continued the analysis by assessing the limitations on the reduced form sign pattern provided by the zero restrictions in the structural arrays. This was done by setting all of the nonzeros in β and γ above to “?”. The quantitative Monte Carlo draws then proceeded to set these entries both positive and negative, independently, each about half the time. The imposed restrictions on the structure were of course reduced, compared to working with the hypothesized signs. Nevertheless, in tens of millions of samples only 4096 sign patterns for the reduced form were found (about 1.5% of the 262,144 possible 3 × 6 sign patterns for the reduced form). Treating these sign patterns as equally likely gave INFO% = 33.3 for the zero restrictions alone. Significantly, the sign pattern of the estimated reduced form was not found among these 4096 sign patterns, a somewhat decisive falsification of the structural hypothesis. Or, at the very least, an outcome that should prompt more analysis concerning the structural hypothesis.

4.2. The Oil Market Simulation (OMS)

The OMS (System’s Science, Inc. [23], EIA [24]) was used in the 1990s by the Department of Energy’s Energy Information Administration (EIA) in conjunction with a model of domestic energy markets in the preparation of annual long term forecasts and special studies as requested. The model’s output is the world oil price (WOP), given estimates of supply and demand in some number of regions world-wide (seven in the version of the model used by EIA). Four other equations give the demand for OPEC production as needed to balance world-wide supply and demand; and, additionally express the relationship between the WOP in terms of OPEC’s rate of capacity utilization. In general, the relationships used were nonlinear.
This model was subjected to a traditional (i.e., entry by entry of the reduced form) qualitative analysis by Hale and Lady [10]. They found that all the entries in the reduced form were signable, based upon the hypothesized sign patterns for the structural arrays. Further, this result was independent of the level of aggregation specified for the model, i.e., the number of regions considered, so long as the algebraic form and sign patterns of the regional supply and demand relationships remained the same. This analysis was reiterated in Hale et al. [9]. Lady and Buck [8] aggregated the model to only one non-OPEC region and we will work here with that aggregated version of the model. In these terms, the structural form of the model is expressed as,
[ 1 0 0 0 0 β 16 0 1 0 0 0 β 26 β 31 β 32 1 0 0 0 0 0 β 43 1 0 0 0 0 0 β 54 1 0 0 0 0 0 β 65 1 ]   [ d D d S d D O d C A P U T d R d W O P ]   =   [ γ 11 0 0 0 0 γ 22 0 0 0 0 0 0 0 0 γ 43 0 0 0 0 0 0 0 0 γ 64 ]   [ d D 1 d S 1 d M a x C a p d W O P 1 ]  
For this notation, the lower case d is the differential operator, D is world oil demand, S is non-OPEC world oil supply, DO is the demand for OPEC oil, CAPUT is the rate of OPEC capacity utilization, R is the percentage change (in decimal) of the current WOP over last year’s, MaxCap is maximum OPEC capacity and WOP is the world oil price. The hypothesized sign patterns for the structural arrays are given below,
sgn   β   =   [ + 0 0 0 0 + 0 + 0 0 0 + + 0 0 0 0 0 + 0 0 0 0 0 + 0 0 0 0 0 + ] ,   a n d   sgn   γ   =   [ + 0 0 0 0 + 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 + ]
Following Hale and Lady’s [10] qualitative methodology, the reduced form for this aggregated version of the model must have the following sign pattern,
sgn   π   =   [ + + + + + + + + + + + + ]
In brief, this is because an analysis of the expansion of det β reveals that all terms have the same sign; and further, for each cofactor in β’s adjoint, all terms are the same sign, positive or negative depending. All of this is based entirely upon the hypothesized structural sign patterns, independent of the magnitudes of the nonzero entries. For this result, INFO% = 100. If the hypothesized structure is true, nothing about the reduced form sign pattern is learned by estimating it. Lady and Buck [8] then estimated the reduced form based upon recent data with the following result (Table 4),
Table 4. OMS Estimated (1983–2006) Reduced Form Sign Pattern.
Table 4. OMS Estimated (1983–2006) Reduced Form Sign Pattern.
dD-1dS-1dMaxCapdWOP-1
dD+++
dS−*+−*
dDO++*++*
dCaput−*+*+*
dR++*
dWOP+−*
“*” indicates a “wrong” sign compared to the predicted reduced form sign pattern.
There are nine entries with different signs than those required by the structural hypothesis. The basis for the inconsistencies corresponding to these terms for this model is immediate, given the outcome of the qualitative analysis of the structural arrays.
The zero restrictions in the structural hypothesis (as above) can be tested by replacing each nonzero in the structural arrays with “?”. As before, in the quantitative Monte Carlo iterations, these terms are independently chosen positive or negative, each about half of the time. Based on this, the Monte Carlo was repeated tens of millions of times. For this, only 4099 reduced form sign patterns were found, out of the possible 224 = 16,777,216 outcomes for the 6 × 4 sign patterns that the estimated reduced form could take on (barring zeros). The zero restrictions thus provide significant limitations on the outcome of the reduced form estimation and INFO% = 50 for this case. Significantly, the estimated reduced form sign pattern was not among those found. As before, this outcome should at least prompt significant reconsideration of the hypothesized structure.

4.3. Klein’s Model I

Klein’s Model I (Klein [25]) is an over-identified model that has appeared in the econometrics literature, e.g., Berndt [26] and Greene [17] for a variety of pedagogical reasons. It has also appeared in a number of papers on qualitative analysis, e.g., Maybee and Weiner [27], Hale et al. [9], Lady [28], and Buck and Lady [22]. In a qualitative context, the model was chosen to illustrate that for sufficiently sparse structural arrays (perhaps with other convenient features) entries in the reduced form would be signable due to adding a small number of (as called) “side conditions” on the relative values of some of the nonzero entries proposed for the structure. The idea, in brief, is that an inspection of the terms in the expansions of β’s determinant and cofactors might find that “almost” all are of the same sign. Given this, the embellished structural hypothesis, i.e., sign pattern plus side condition(s), would result in (at least some) entries in the corresponding reduced form being signable; and, this in turn could be tested by comparison to the sign pattern of the estimated reduced form.
Suppressing the error vector, Klein’s model is expressed as,
β Y = γ Z
where
[ 1 0 a 1 0 a 2 0 0 0 1 0 0 b 1 0 0 0 0 1 0 0 0 c 1 1 1 0 1 0 0 0 0 0 0 1 1 1 0 0 0 1 0 0 1 0 0 0 0 1 0 0 1 ] [ C I W 1 Y P W E ] = [ a 1 a 3 0 0 0 0 0 0 b 2 b 3 0 0 0 0 0 0 0 c 2 c 3 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 1 0 ] [ W 2 P 1 K 1 E 1 Y e a r T X G ]
In this structure the endogenous variables are private consumption (C), investment (I), the private wage bill (W1), income (Y), profits or nonwage income (P), the sum of private and government wages (W), and private product (E); and the exogenous variables are the government wage bill (W2), lagged profits (P−1), end of last period capital stock (K−1), lagged private product (E−1), years since 1931 (Year), taxes (TX), and government consumption (G).
The sign patterns of the arrays proposed by Klein [25] are as follows,
sgn   β = [ 0 + 0 + 0 0 0 0 0 + 0 0 0 0 0 0 0 + + + 0 0 0 0 0 0 0 + 0 0 0 + 0 0 0 0 0 0 + 0 0 ] and   sgn   γ = [ 0 0 0 0 0 0 + 0 0 0 0 0 0 0 0 0 0 0 0 0 0 + 0 0 0 0 0 0 0 0 0 0 0 0 0 + 0 0 0 0 0 ]
The sign pattern of the estimated reduced form of Klein’s model is reported in Goldberger [29], using data from 1921–1941, as
E s t i m a t e d   sgn   π ^   =   [ + + + + + + + + + + + + + + + + + + + + + + + + + + + ]
Without side conditions, no entry in the reduced form is signable, based upon the hypothesized sign patterns for the structural arrays. The size of this model is too large for our (PC-based) computing platform to enumerate all possible reduced form sign patterns, but repeated quantitative samples of the hypothesized structure can be taken and the outcome for the corresponding reduced form can be compared to a proposed reduced form, such as that above as reported in Goldberger. This was done in Buck and Lady [8]. Tens of millions of samples were taken and the above reduced form sign pattern was not found. Buck and Lady [8] (p. 2826) then found inconsistencies in the hypothesized structural arrays that made certain of the signs in the estimated reduced form impossible. It was patterns of signs that were at issue, since no individual entry was signable, e.g., although not individually signable, π24 and π25 cannot have opposite signs.
We have gone further and assessed the zero restrictions and the estimable entries of the structural arrays. As before, nonzeros were replaced by “?” for which both positive and negative values were chosen, each about half the time. The structural arrays were configured for the analysis as given below,
sgn   β = [ 0 ? 0 ? 0 0 0 0 0 ? 0 0 0 0 0 0 0 ? + + 0 0 0 0 0 0 0 + 0 0 0 + 0 0 0 0 0 0 + 0 0 ] and   sgn   γ = [ ? ? 0 0 0 0 0 0 ? ? 0 0 0 0 0 0 0 ? ? 0 0 0 0 0 0 0 + 0 0 0 0 0 0 0 0 0 0 0 0 0 + 0 0 0 0 0 ] ,
Tens of millions of samples were taken and the sign pattern of the reduced form sign pattern estimated by Goldberger was not found. This not only calls the hypothesized structure into question, but also makes the continuation of the analysis by estimating the values of appropriate entries of the hypothesized structure problematical. In particular, if the zero restrictions (and values for certain of the other entries) are imposed when constructing the structural coefficient estimates, the result could not possibly have generated the data upon which the estimated reduced form is based (absent errors). Specifically, with the zero restrictions imposed, all possible sign patterns for the structure could not have generated the sign pattern of the estimated reduced form.

5. Conclusions

Economics as a science should produce meaningful theorems that when true limit the range of data outcomes. The purpose of this paper is to demonstrate the potential of a qualitative analysis in evaluating models when confronted with data. A qualitative analysis is proposed as a complement to other econometric methods that depend upon more quantitative features of the data to be analyzed. The algorithmic methods used in illustrating our points are indicative and not definitive.
The use of qualitative methods can be compelling to the degree to which the theory does not usually provide a great deal of information beyond a system’s inference structure and the directions of influence among its variables. Unlike the bulk of the literature on such analyses, we show that insightful, and in the case of falsification, decisive, results are always possible and do not depend upon unusually restrictive or uncommon features of the mathematical content of a model. Indeed, even a partially specified structural model can be falsified by as few as one reduced form result.
The process involves using a model’s structural sign pattern information to assess its potential for falsification (and its associated type 1 error) and for its acceptance (and its associated type 2 error). Nevertheless, in the particular case of an estimated reduced form’s sign pattern falsifying the zero restrictions of a hypothesized structural form, the estimation of the structural arrays with the zero restrictions imposed will always lead to a mistake, i.e., the reduced form sign pattern is impossible for estimated structural arrays with the zero restrictions imposed.
The proposed qualitative analysis can enable the information content of a model to be estimated, based upon the entropy of the distribution of possible reduced form sign patterns that corresponds to the hypothesized structural sign patterns. As a model’s information content is greater, i.e., as it imposes greater limits on the possible outcomes of the reduced form sign patterns, the model’s falsification is more subject to type 1 error and the model’s acceptance is less subject to type 2 error.

Acknowledgment

The manuscript has benefitted greatly from the comments of three anonymous referees. Any errors remain our own.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix: A Qualitative Tool Kit

The Monte Carlo approach we utilize in this paper was first presented in Lady and Buck [8]. We have added a number of enhancements to the method first presented as discussed below. Algorithmic approaches and the sufficient conditions for conducting a qualitative analysis presented previously include: Lancaster [30]; Bassett, Maybee and Quirk [31]; Ritschard [32]; Maybee [33]; and Lady [34]. These all were designed to detect and resolve any entries in the reduced form that were signable, i.e., had the same sign, given the sign patterns of the structural arrays, but independent of the magnitudes of the nonzero entries. The algorithms were computer-based, were to one degree or another difficult to formulate based upon their textual descriptions, and (so far as we know) were not, and are not, widely utilized.
Software and instructions for its use that support some of the results presented in this paper are available at: http://astro.temple.edu/~gmlady/RF_Finder/Finder_Page.htm [35]. The Monte Carlo algorithm used here undertakes repeated draws for the magnitudes of the entries of the structural arrays, consistent with the specified sign patterns. Then, π = β−1γ is computed. Given this, it is a simple procedure to count the number of times each entry of π is positive and the number of times negative (if zero the sample is discarded). This is an extremely robust method to determine if any of the entries are signable, since if so they will always turn out to have the same sign, regardless of the size of the sample. Further, these simple counts will also reveal if entries of π always have the same, or different signs, as dictated by entries of β’s adjoint being signable, since if so the counts of such will be the same, independent of the size of the sample. If the sign pattern of an estimated reduced form is in-hand, it is additionally straight-forward to see if the sign patterns of any of the sampled reduced forms are the same as the estimated reduced form. The size of the arrays is not particularly limiting for this “simple search”, except for the time needed to construct samples of a given size. If the sign pattern of the estimated reduced form is not found after repeated Monte Carlo draws, then the counts of positive and negative entries in the sampled reduced form may reveal the reason, i.e., due to signable entries or entries required to always have the same or different signs. Typically (unfortunately), the sampled reduced forms will not provide these regularities and call for further analysis if the sign pattern of the estimated reduced form is not found, suggesting that the structural hypothesis is falsified.
Processing the algebra of the computation π = β−1γ is one of several obvious features of the analysis that calls for substantial development and algorithmic support. One immediate technique is to check to see if subgroupings of the reduced form sign pattern are resulting in the apparent falsification. This is accomplished by only checking such a subgrouping when comparing the sampled reduced forms to the estimated reduced form. When such subgroupings are enumerated, then the algebra of computing the reduced form can be written out with the appropriate focus to determine the problem of satisfying the systems (5). This was done in Lady and Buck [8] and Buck and Lady [11].
The method used here to enumerate all possible reduced form sign patterns, the associated frequency distribution in the sample taken, and the subsequent computation of the entropy of the structural hypothesis is sensitive to the size of the system. For a given system, barring zeros as we assume, there are 2mn possible reduced form sign patterns. The (long) integer used in our computing platform is limited to +/− 231. Accordingly, we cannot tabulate indexed counts of the reduced form sign patterns except for mn < 30. This limitation can be mitigated by using other computing platforms or indexing schemes.
The distributions from which the quantitative values of the structural arrays are chosen can be set for each entry independent of the others, always set at a particular cardinal value as might be the case for accounting equations in the structural system, or otherwise proposed by the theory, e.g., the marginal propensity to consume is not only positive, but also might be limited to be less than one. Further, some entries can be set at different signs as well as different values. In our Monte Carlo procedures, the following notation is used for the sign of an entry in the structural arrays: (1, −1, 0) for respectively (+, −, 0), 2 for equally probable + or −, 3 for equally probable (+, −, 0), 4 for equally probable + or 0, and 5 for equally probable − or 0. When testing the zero restrictions of the structural arrays, all nonzeros are assigned a “2”. In the example using systems #1 and #2 in Section 3, the structural arrays specifying the “universe” of 512 possible structural sign patterns was specified as:
sgn   β   =   [ 1 2 4 0 4 1 1 4 1 4 1 1 5 0 1 1 ]   a n d   γ   =   [ 1 5 0 0 5 1 0 0 0 5 1 0 0 0 0 1 ]
Using this convention, these arrays were sampled and the corresponding reduced forms computed 300,000,000 times. The counts were tabulated for each time one of the twenty-eight reduced form sign patterns at issue was generated. The frequency distributions of these counts were then used in Equation (8) for the prior probability of each of the sign patterns for π . Table A1 and Table A2 below give the workup for computing the conditional probabilities for each system provided in Table 3.
Table A1. Conditional Probability for System #1.
Table A1. Conditional Probability for System #1.
sgn π Base 10 IndexP(sgn π )P(Sys#1)P(sgn π |Sys #1)P(Sys#1/sgn π )
89140.004520.001950.001030.00044
89150.007640.001950.000550.00014
98780.000850.001950.000820.00188
98790.002780.001950.001210.00085
99380.001410.001950.000780.00108
99390.014810.001950.001530.00020
219250.013040.001950.000640.00010
262620.002410.001950.003600.00291
262630.008310.001950.002200.00052
381890.000400.001950.000570.00278
382490.003090.001950.000290.00018
392730.002360.001950.002670.00221
402370.000620.001950.000810.00255
402970.001290.001950.001430.00216
415620.002290.001950.000190.00016
415630.005270.001950.000350.00013
416820.002650.001950.000380.00028
416830.012320.001950.000570.00009
425860.002270.001950.000960.00082
425870.007400.001950.002560.00067
426460.005200.001950.001740.00065
426470.015490.001950.003510.00044
427060.002470.001950.002610.00206
427070.023240.001950.008800.00074
436100.003050.001950.000680.00043
436110.004780.001950.000560.00023
545730.005440.001950.000430.00015
566210.008210.001950.002390.00057
Table A2. Conditional Probability for System #2.
Table A2. Conditional Probability for System #2.
sgn π Base 10 IndexP(sgn π )P(Sys#2)P(sgn π |Sys #2)P(Sys#2/sgn π )
89140.004520.001950.028150.01214
89150.007640.001950.023390.00597
98780.000850.001950.005680.01303
98790.002780.001950.009350.00656
99380.001410.001950.004910.00679
99390.014810.001950.028280.00372
219250.013040.001950.048500.00725
262620.002410.001950.017020.01377
262630.008310.001950.014240.00334
381890.000400.001950.00630.03071
382490.003090.001950.012250.00773
392730.002360.001950.025780.02130
402370.000620.001950.002910.00915
402970.001290.001950.00550.00831
415620.002290.001950.014780.01259
415630.005270.001950.033110.01225
416820.002650.001950.014360.01057
416830.012320.001950.009610.00152
425860.002270.001950.006750.00580
425870.007400.001950.055160.01454
426460.005200.001950.060680.02276
426470.015490.001950.032830.00413
427060.002470.001950.016090.01270
427070.023240.0019500.025110.00211
436100.003050.001950.023590.01508
436110.004780.001950.025560.01043
545730.005440.001950.018870.00676
566210.008210.001950.018470.00439

References

  1. K. Popper. The Logic of Scientific Discovery. New York, NY, USA: Harper and Row, 1959. [Google Scholar]
  2. P. Samuelson. Foundations of Economic Analysis. Cambridge, MA, USA: Harvard University Press, 1947. [Google Scholar]
  3. J.D. Angrist, and A.B. Krueger. “Does compulsory school attendance affect schooling and earnings? ” Q. J. Econ. 106 (1991): 979–1014. [Google Scholar] [CrossRef]
  4. K.F. Butcher, and A. Case. “The effect of sibling composition on women’s education and earnings.” Q. J. Econ. 109 (1994): 531–563. [Google Scholar] [CrossRef]
  5. M.R. Rosenzweig, and K.I. Wolpin. “Natural “natural experiments” in Economics.” J. Econ. Lit. 38 (2000): 827–874. [Google Scholar] [CrossRef]
  6. J.M. Barron, T.E. Ewing, and G.R. Waddell. “The Effect of High School Athletic Participation on Education and Labor Market Outcomes.” Rev. Econ. Statist. 82 (2000): 409–421. [Google Scholar] [CrossRef]
  7. C.E. Shannon. “A Mathematical Theory of Communication.” Bell Syst. Techn. J. 27 (1948): 379–423. [Google Scholar] [CrossRef]
  8. G. Lady, and A. Buck. “Structural Models, Information and Inherited Restrictions.” Econ. Model. 28 (2011): 2820–2831. [Google Scholar] [CrossRef]
  9. D. Hale, G. Lady, J. Maybee, and J. Quirk. Nonparametric Comparative Statics and Stability. Princeton, NJ, USA: Princeton University Press, 1999. [Google Scholar]
  10. D. Hale, and G. Lady. “Qualitative Comparative Statics and Audits of Model Performance.” Linear Algebra Its Appl. 217 (1995): 141–154. [Google Scholar] [CrossRef]
  11. A.J. Buck, and G. Lady. “Structural Sign Patterns and Reduced Form Restrictions.” Econ. Model. 29 (2012): 462–470. [Google Scholar] [CrossRef]
  12. L. Metzler. “Stability of Multiple Markets: The Hicks Conditions.” Econometrica 13 (1945): 277–292. [Google Scholar] [CrossRef]
  13. T.M. Cover, and J.A. Thomas. Elements of Information Theory. New York, NY, USA: Wiley Series in Telecommunications, 1991. [Google Scholar]
  14. W. Isaacson. Einstein, His Life and Universe. New York, NY, USA: Simon and Schuster, 2007. [Google Scholar]
  15. A.D. Strickland, and L.W. Weiss. “Advertising concentration, and price-cost margins.” J. Polit. Econ. 84 (1976): 1109–1121. [Google Scholar] [CrossRef]
  16. T. Amemiya. Advanced Econometrics. Cambridge, MA, USA: Harvard University Press, 1985. [Google Scholar]
  17. W. Greene. Econometric Analysis, 5th ed. Upper Saddle River, NJ, USA: Prentice Hall, 2003. [Google Scholar]
  18. D. Gujarati. Basic Econometrics. New York, NY, USA: McGraw-Hill Higher Education, 2003. [Google Scholar]
  19. D.F. Greer. Industrial Organization and Public Policy, 2nd ed. New York, NY, USA: Macmillan, 1984. [Google Scholar]
  20. F.M. Scherer, and D. Ross. Industrial Market Structure and Economic Performance. Boston, MA, USA: Houghton Mifflin, 1990. [Google Scholar]
  21. D.W. Carlton, and J.M. Perloff. Modern Industrial Organization, 3rd ed. Glenview, IL, USA: Scott, Foresman, 2000. [Google Scholar]
  22. A.J. Buck, and G. Lady. “Falsifying Economics Models.” Econ. Model. 22 (2005): 777–810. [Google Scholar] [CrossRef]
  23. System’s Science Inc. The Oil Market Simulation Model, Documentation Report. Washington, DC, USA: System’s Science Inc., 1985. [Google Scholar]
  24. Energy Information Administration. Oil Market Simulation User’s Manuel; Washington, DC, USA: Energy Information Administration, 1990.
  25. L. Klein. Economic Fluctuations in the U.S., 1921–1941. Cowles Commission for Research in Economics Monograph No. 11; New York, NY, USA: John Wiley and Sons, 1950. [Google Scholar]
  26. E.R. Berndt. The Practice of Econometrics, Classic and Contemporary. New York, NY, USA: Addison-Wesley, 1996. [Google Scholar]
  27. J. Maybee, and G. Wiener. “From qualitative matrices to quantitative restrictions.” Linear Multilinear Algebra 22 (1988): 229–248. [Google Scholar] [CrossRef]
  28. G. Lady. “Topics in nonparametric comparative statics and stability.” Int. Adv. Econ. Res. 6 (2000): 67–83. [Google Scholar] [CrossRef]
  29. A.S. Goldberger. Econometric Theory. New York, NY, USA: John Wiley and Sons, 1964. [Google Scholar]
  30. K. Lancaster. “The Solution of Qualitative Comparative Statics Problems.” Q. J. Econ. 53 (1966): 278–295. [Google Scholar] [CrossRef]
  31. L. Bassett, J. Maybee, and J. Quirk. “Qualitative Economics and the Scope of the Correspondence Principal.” Econometrica 36 (1968): 544–563. [Google Scholar] [CrossRef]
  32. G. Ritschard. “Computable Qualitative Comparative Statics Techniques.” Econometrica 51 (1983): 1145–1168. [Google Scholar] [CrossRef]
  33. S. Maybee. “A Method for Identifying Sign Solvable Systems.” Master Thesis, University of Colorado, Boulder, CO, USA, 1986. [Google Scholar]
  34. G. Lady. SGNSOLVE.EXE Analysis Package; prepared as a job of work for the Energy Information Administration; Washington, DC, USA: U.S. Department of Energy, 1993.
  35. “RF Finder Web Site.” Available online: http://astro.temple.edu/~gmlady/RF_Finder/Finder_Page.htm (accessed on 18 June 2015).
  • 1Rosenzweig and Wolpin [5] show that both the Angrist-Krueger [3] and the Butcher-Case [4] models produce biased results, but in opposite directions. The implication is that both models are wrong, which begs the question of why both studies show statistically significant results.
  • 2All three papers report results for a single structural equation estimated by IV methods without specifying the signs and zero restrictions in the other equations of the implicit structural model. By ignoring this step their models cannot be falsified. As a consequence of the absence of the ability to falsify, these papers cannot claim to have modelled causation. Rather, they have at best provided more precise conditional partial correlations. This is a common shortcoming in the “natural experiment” literature.
  • 3At the risk of belaboring the point, consider a model car. The model car is not an exact replica of a real car, but we could ask what part of the real car is correctly described by the model. What we show in this paper is that empirical implementation of the model car may lead us to believe that we have correctly described some part of the real car, even though the model could not possibly have generated the observed data. See Section 4 below for examples of well-known models that could not possibly have generated the data used to estimate model parameters.
  • 4We show below that even in the absence of sampling error it is possible to commit type 1 and type 2 errors.
  • 5It is worth noting that the system in (3) can encompass a wide variety of models, not the least of which could be a Keynesian model or a structural VAR. These models can be examined using a qualitative analysis (Lady and Buck [8]). Also see the cases considered in Section 4.
  • 6In hypothesis testing there is either an appeal to the Central Limit Theorems or an assumption of normality in order to apply the usual test statistics. For the purposes of the work presented here the actual distribution of the error is of no importance as long as the variances of the random variables in U are not excessively large. In simulations not reported here we have found that the error vector variance has to be orders of magnitude greater than the variation in the independent variables before there is any impact on the falsification procedures presented here. One might even argue that if the structural error is orders of magnitude larger than the variation in the independent variables then the model is incompletely specified, in which case a mistaken falsification is forgivable.
  • 7Sampling error is an issue that may lead one to conclude that an estimate is different from, say, zero when in fact it is not because the test statistic is not constrained to fall in a certain part of the distribution’s support. An exception would be testing for unit roots, in which the parameter of interest is bounded away from one under the null hypothesis and as a consequence the test statistic is also bounded. In the realm of the present paper, if the specified model is TRUE in its assigned zero restrictions and sign patterns then in expectation the reduced form coefficients estimators mathematically cannot take the wrong signs. The sampling error occurs because the reduced form error is a weighted average of the structural disturbance and might admit falsifiable signs. We reiterate the point in footnote 5, that in Mont Carlo simulations the structural error must be orders of magnitude larger than the variation in Z to induce a rejection of a true null. The more important point remains that the admissible outcomes for the reduced form estimates can be small in number and structural forms other than the one specified can produce an admissible reduced form. In short, the world in which theory informs model specification is a good deal more restricted than classical statistics.
  • 8Hale and Lady [10] is an interesting exception and the model presented by them is discussed in the context of this paper in Section 4.
  • 9In this paper we deal only with identification made possible through zero restrictions imposed on β and γ of Equation (3). In an earlier paper we dealt with identifiability and falsification of a structural VAR using the error covariance; see Lady and Buck [8].
  • 10A matrix is reducible if and only if it can be placed into block upper-triangular form by simultaneous row/column permutations. A square matrix which is not reducible is said to be irreducible. The inverse of an irreducible matrix has no entries that must be equal to zero, a characteristic that we make use of when running the Monte Carlo. Limiting the analysis to irreducible systems is convenient, but not necessary for any of our analytical points.
  • 11This algorithmic approach and its use are briefly described in the Appendix. Suffice it to say that the Monte Carlo is run for millions of repetitions.
  • 12Regardless of the actual number of possible sign patterns for an inverse, the outcomes of the inversion have a multinomial distribution. The Monte Carlo is a maximum likelihood estimator of the parameters of that multinomial distribution. For further discussion of this as well as the bias in our estimator for the information content of the model see Lady and Buck [8].
  • 13This is not a trivial example even though the system is exactly identified. In standard practice the estimates of the structural coefficients can be solved from the reduced form using indirect least squares. In spite of being exactly identified the proposed model can still be wrong, produce a set of first stage results that would by our methodology falsify the proposed model, yet produce estimates of some of the proposed model coefficients that were of the proposed sign and be statistically significant.
  • 14Strictly, sgn a = 1, −1, 0 as a > 0, a < 0, a = 0. We will use the signs themselves as a felicitous convention.
  • 15For our example, the condition γ = I is also imposed. Accordingly, it is this condition plus β31 = 0 that is being falsified, not just the zero restriction on β.
  • 16Recall that there can be no logical zeros in the inverse of an irreducible matrix.
  • 17There is fundamentally no difference between what we have done to model sampling error and an alternative approach in which one might first use a Monte Carlo to populate an admissible π and then perturb those πi,j to simulate sampling error.
  • 18See Isaacson [14] (Chapter 9).
  • 19It might be noted that this hypothesized “universe” can itself be falsified. For example, an all positive reduced form is impossible for any of the 512 structural sign patterns.
  • 20We were not able to obtain the dataset used by Strickland and Weiss [15].
  • 21In an actual application, if a reduced form sign pattern is not found by the Monte Carlo, additional analysis would be appropriate to isolate the particular inconsistencies presented by the hypothesized sign patterns for the structural arrays. We will set that as beyond our scope, accepting the results of the numerical method employed. The chances of missing a sign pattern in tens of millions of quantitative samples for the size of the arrays discussed here is presumably quite small. A means of estimating the likelihood of making such a mistake is given in Lady and Buck [8] (p. 2825).

Share and Cite

MDPI and ACS Style

Buck, A.J.; Lady, G.M. A New Approach to Model Verification, Falsification and Selection. Econometrics 2015, 3, 466-493. https://doi.org/10.3390/econometrics3030466

AMA Style

Buck AJ, Lady GM. A New Approach to Model Verification, Falsification and Selection. Econometrics. 2015; 3(3):466-493. https://doi.org/10.3390/econometrics3030466

Chicago/Turabian Style

Buck, Andrew J., and George M. Lady. 2015. "A New Approach to Model Verification, Falsification and Selection" Econometrics 3, no. 3: 466-493. https://doi.org/10.3390/econometrics3030466

Article Metrics

Back to TopTop