Skip to main content
Log in

Model-Implied Instrumental Variable—Generalized Method of Moments (MIIV-GMM) Estimators for Latent Variable Models

  • Published:
Psychometrika Aims and scope Submit manuscript

Abstract

The common maximum likelihood (ML) estimator for structural equation models (SEMs) has optimal asymptotic properties under ideal conditions (e.g., correct structure, no excess kurtosis, etc.) that are rarely met in practice. This paper proposes model-implied instrumental variable – generalized method of moments (MIIV-GMM) estimators for latent variable SEMs that are more robust than ML to violations of both the model structure and distributional assumptions. Under less demanding assumptions, the MIIV-GMM estimators are consistent, asymptotically unbiased, asymptotically normal, and have an asymptotic covariance matrix. They are “distribution-free,” robust to heteroscedasticity, and have overidentification goodness-of-fit J-tests with asymptotic chi-square distributions. In addition, MIIV-GMM estimators are “scalable” in that they can estimate and test the full model or any subset of equations, and hence allow better pinpointing of those parts of the model that fit and do not fit the data. An empirical example illustrates MIIV-GMM estimators. Two simulation studies explore their finite sample properties and find that they perform well across a range of sample sizes.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1.
Figure 2.
Figure 3.
Figure 4.
Figure 5.
Figure 6.
Figure 7.

Similar content being viewed by others

References

  • Anderson, J.C., & Gerbing, D. (1984). The effect of sampling error on convergence, improper solutions, and goodness-of-fit indices for maximum likelihood confirmatory factor analysis. Psychometrika, 49, 155–173.

    Article  Google Scholar 

  • Anderson, T.W., & Amemiya, Y. (1988). The asymptotic normal distribution of estimators in factor analysis under general conditions. The Annals of Statistics, 16, 759–771.

    Article  Google Scholar 

  • Angrist, J.D., & Pischke, J. (2009). Mostly harmless econometrics: an empiricist’s companion. Princeton: Princeton University Press.

    Google Scholar 

  • Bauldry, S. (forthcoming). miivfind: a program for identifying model-implied instrumental variables (MIIVs) for structural equation models in Stata. Stata Journal.

  • Bentler, P.M. (1982). Confirmatory factor analysis via noniterative estimation: a fast, inexpensive method. Journal of Marketing Research, 19, 417–424.

    Article  Google Scholar 

  • Bentler, P.M., & Yuan, K. (1999). Structural equation modeling with small samples: test statistics. Multivariate Behavioral Research, 34, 181–197.

    Article  Google Scholar 

  • Bollen, K.A. (1989). Structural equations with latent variables. New York: Wiley.

    Google Scholar 

  • Bollen, K.A. (1996a). An alternative Two Stage Least Squares (2SLS) estimator for latent variable equations. Psychometrika, 61, 109–121.

    Article  Google Scholar 

  • Bollen, K.A. (1996b). A limited information estimator for LISREL models with and without heteroscedasticity. In G.A. Marcoulides & R.E. Schumacker (Eds.), Advanced structural equation modeling (pp. 227–241). Mahwah: Erlbaum.

    Google Scholar 

  • Bollen, K.A. (2001). Two-stage least squares and latent variable models: simultaneous estimation and robustness to misspecifications. In: R. Cudeck, S.D. Toit, & D. Sörbom (Eds.), Structural equation modeling: present and future, a festschrift in honor of Karl Jöreskog (pp. 119–138). Lincolnwood: Scientific Software International.

    Google Scholar 

  • Bollen, K.A. (2012). Instrumental variables in sociology and the social sciences. Annual Review of Sociology, 38, 37–72.

    Article  Google Scholar 

  • Bollen, K.A., & Bauer, D.J. (2004). Automating the selection of model-implied instrumental variables. Sociological Methods & Research, 32, 425–452.

    Article  Google Scholar 

  • Bollen, K.A., Kirby, J.B., Curran, P.J., Paxton, P.M., & Chen, F. (2007). Latent variable models under misspecification: two-stage least squares (2SLS) and maximum likelihood (ML) estimators. Sociological Methods & Research, 36, 48–86.

    Article  Google Scholar 

  • Bollen, K.A., & Pearl, J. (2013). Eight myths about causality and structural equation models. In S. Morgan (Ed.), Handbook of causal analysis for social research, New York: Springer.

    Google Scholar 

  • Bollen, K.A., & Stine, R. (1990). Direct and indirect effects: classical and bootstrap estimates of variability. Sociological Methodology, 20, 115–140.

    Article  Google Scholar 

  • Bollen, K.A., & Stine, R. (1992). Bootstrapping goodness-of-fit measures in structural equation models. Sociological Methods & Research, 21, 205–229.

    Article  Google Scholar 

  • Boomsma, A., & Hoogland, J.J. (2001). The robustness of LISREL modeling revisited. In R. Cudeck, S.D. Toit, & D. Sörbom (Eds.), Structural equation modeling: present and future, a festschrift in honor of Karl Jöreskog (pp. 139–168). Lincolnwood: Scientific Software International.

    Google Scholar 

  • Browne, M.W. (1984). Asymptotically distribution-free methods for the analysis of the covariance structures. British Journal of Mathematical & Statistical Psychology, 37, 62–83.

    Article  Google Scholar 

  • Browne, M.W., & Cudeck, R. (1993). Alternative ways of assessing model fit. In K.A. Bollen & J.S. Long (Eds.), Testing structural equation models (pp. 136–162). Newbury Park: Sage.

    Google Scholar 

  • Chausse, P. (2012). gmm: generalized method of moments and generalized empirical likelihood (R package). http://cran.r-project.org/web/packages/gmm/index.html.

  • Cragg, J.G. (1968). Some effects of incorrect specification on the small sample properties of several simultaneous equation estimators. International Economic Review, 9, 63–86.

    Article  Google Scholar 

  • Davidson, R., & MacKinnon, J.G. (1993). Estimation and inference in econometrics. New York: Oxford University Press.

    Google Scholar 

  • Foster, E.M. (1997). Instrumental variables for logistic regression: an illustration. Social Science Research, 26, 487–504.

    Article  Google Scholar 

  • Glanville, J.L., & Paxton, P. (2007). How do we learn to trust? A confirmatory tetrad analysis of the sources of generalized trust. Social Psychology Quarterly, 70, 230–242.

    Article  Google Scholar 

  • Godambe, V.P., & Thompson, M. (1978). Some aspects of the theory of estimating equations. Journal of Statistical Planning and Inference, 2, 95–104.

    Article  Google Scholar 

  • Hall, A.R. (2005). Generalized method of moments. Oxford: Oxford University Press.

    Google Scholar 

  • Hägglund, G. (1982). Factor analysis by instrumental variables. Psychometrika, 47, 209–222.

    Article  Google Scholar 

  • Hansen, L.P. (1982). Large sample properties of generalized method of moments estimators. Econometrica, 50, 1029–1054.

    Article  Google Scholar 

  • Hu, L.T., Bentler, P.M., & Kano, Y. (1992). Can test statistics in covariance structure analysis be trusted? Psychological Bulletin, 112, 351–362.

    Article  PubMed  Google Scholar 

  • Ihara, M., & Kano, Y. (1986). A new estimator of the uniqueness in factor analysis. Psychometrika, 51, 563–566.

    Article  Google Scholar 

  • Jöreskog, K.G. (1969). A general approach to confirmatory maximum likelihood factor analysis. Psychometrika, 34, 183–202.

    Article  Google Scholar 

  • Jöreskog, K.G. (1973). A general method for estimating a linear structural equation system. In: A.S. Goldberger & O.D. Duncan (Eds.), Structural equation models in the social sciences (pp. 85–112). New York: Academic Press.

    Google Scholar 

  • Jöreskog, K.G. (1977). Structural equation models in the social sciences: specification, estimation, and testing. in: P.R. Krishnaiah (Ed.), Applications of statistics (pp. 265–287). Amsterdam: North-Holland.

    Google Scholar 

  • Jöreskog, K.G. (1983). Factor analysis as an error-in-variables model. In: Wainer, H. & Messick, S. (Eds.) Principles of Modern Psychological Measurement (pp. 185–196). Hillsdale: Erlbaum.

    Google Scholar 

  • Kirby, J.B., & Bollen, K.A. (2009). Using instrumental variable tests to evaluated model specification in latent variable structural equation models. Sociological Methodology, 39, 327–355.

    Article  PubMed Central  PubMed  Google Scholar 

  • Kolenikov, S. (2011). Biases of parameter estimates in misspecified structural equation models. Sociological Methodology, 41, 119–157.

    Article  Google Scholar 

  • Kolenikov, S., & Bollen, K.A. (2012). Testing negative error variances: is a Heywood case a symptom of misspecification? Sociological Methods & Research, 41, 124–167.

    Article  Google Scholar 

  • Lawley, D.N. (1940). The estimation of factor loadings by the method of maximum likelihood. Proceedings of the Royal Society of Edinburgh, 60, 64–82.

    Google Scholar 

  • Madansky, A. (1964). Instrumental variables in factor analysis. Psychometrika, 29, 105–113.

    Article  Google Scholar 

  • Mardia, K.V. (1970). Measures of multivariate skewness and kurtosis with applications. Biometrika, 57, 519–530.

    Article  Google Scholar 

  • Mátyás, L. (Ed.) (1999). Generalized method of moments estimation. Cambridge: Cambridge University Press.

    Google Scholar 

  • Micceri, T. (1989). The unicorn, the normal curve, and other improbable creatures. Psychological Bulletin, 1, 156–166.

    Article  Google Scholar 

  • Muthén, L.K., & Muthén, B. (1998–2010). Mplus user’s guide. Los Angeles: Muthén & Muthén.

    Google Scholar 

  • Nevitt, J., & Hancock, G.R. (2004). Evaluating small sample approaches for model test statistics in structural equation modeling. Multivariate Behavioral Research, 39, 439–478.

    Article  Google Scholar 

  • Newey, W.K., & McFadden, D. (1986). Large sample estimation and hypothesis testing. In R.F. Engle & D. McFadden (Eds.), Handbook of Econometrics (Vol. 4, 1st ed., pp. 2111–2245). Amsterdam: Elsevier.

    Google Scholar 

  • Paxton, P.M., Curran, P., Bollen, K.A., Kirby, J., & Chen, F. (2001). Monte Carlo simulations in structural equation models. Structural Equation Modeling, 8, 287–312.

    Article  Google Scholar 

  • Pew Research Center (1998). Trust and citizen engagement in metropolitan Philadelphia: a case study. Washington: The Pew Research Center for the People and the Press.

    Google Scholar 

  • Sargan, J.D. (1958). The estimation of economic relationships using instrumental variables. Econometrica, 26, 393–415.

    Article  Google Scholar 

  • Satorra, A. (1990). Robustness issues in structural equation modeling: a review of recent developments. Quality and Quantity, 24, 367–386.

    Article  Google Scholar 

  • Satorra, A., & Bentler, P.M. (1994). Corrections to test statistics and standard errors in covariance structure analysis. In A. von Eye & C.C. Clogg (Eds.), Latent variable analysis (pp. 399–419). Thousand Oaks: Sage.

    Google Scholar 

  • Searle, S.R. (1982). Matrix algebra useful for statistics (1st ed.). New York: Wiley.

    Google Scholar 

  • Skrondal, A., & Hesketh, S.R. (2004). Generalized latent variable modeling. Boca Raton: Chapman & Hall/CRC.

    Book  Google Scholar 

  • Staiger, D., & Stock, J.H. (1997). Instrumental variables regression with weak instruments. Econometrica, 65, 557–586.

    Article  Google Scholar 

  • StataCorp (2011). Stata statistical software: release 12. College Station: StataCorp.

    Google Scholar 

  • Stock, J.H., & Yogo, M. (2005). Testing for weak instruments in linear IV regression. In D.W.K. Andrews (Ed.), Identification and Inference for Econometric Models (pp. 80–108). New York: Cambridge University Press.

    Chapter  Google Scholar 

  • Stock, J.H., Wright, J.H., & Yogo, M. (2002). A survey of weak instruments and weak identification in generalized method of moments. Journal of Business & Economic Statistics, 20, 518–529.

    Article  Google Scholar 

  • van der Vaart, A.W. (1998). Asymptotic statistics. New York: Wiley.

    Book  Google Scholar 

  • Wooldrige, J.M. (2010). Econometric analysis of cross section and panel data. Cambridge: MIT Press.

    Google Scholar 

  • Yuan, K., & Hayashi, K. (2006). Standard errors in covariance structure models: asymptotic versus bootstrap. British Journal of Mathematical & Statistical Psychology, 59, 397–417.

    Article  Google Scholar 

Download references

Acknowledgements

We gratefully acknowledge the support of NSF SES 0617276 and SES-0617193.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kenneth A. Bollen.

Appendices

Appendix A. Notation Example

In the section on “From Latent to Observed Variables,” we introduced the notation of y =ZA+u. To illustrate this notation, suppose that the first latent variable equation for the ith case is

$$ \eta _{1i}=\alpha _{\eta 1}+\beta _{12}\eta _{2i}+\gamma _{11}\xi _{1i}+\gamma _{12} \xi _{2i}+\zeta _{1i} $$
(A.1)

with y 1i , y 2i ,x 1i , and x 2i as the scaling indicators for η 1i ,η 2i ,ξ 1i , and ξ 2i , respectively. By replacing each latent variable by its scaling indicator minus its error [e.g., η 1i =(y 1i ϵ 1i )], the observed variable counterpart to this first latent variable equation is

(A.2)

The Z 1 for this first equation is

$$ \mathbf{Z}_{1}=\left [ \begin{array}{c@{\quad }c@{\quad }c@{\quad }c} 1 & y_{21} & x_{11} & x_{21} \\ 1 & y_{22} & x_{12} & x_{22} \\ \vdots & \vdots & \vdots & \vdots \\ 1 & y_{2N} & x_{1N} & x_{2N}\end{array} \right ] $$
(A.3)

and u 1 is the vector

$$ \mathbf{u}_{1}=\left [ \begin{array}{c} u_{11} \\ u_{12} \\ \vdots \\ u_{1N}\end{array} \right ] =\left [ \begin{array}{c} -\beta _{12}\epsilon _{21}-\gamma _{11}\delta _{11}-\gamma _{12}\delta _{21}+\epsilon _{11}+\zeta _{11} \\ -\beta _{12}\epsilon _{22}-\gamma _{11}\delta _{12}-\gamma _{12}\delta _{22}+\epsilon _{12}+\zeta _{12} \\ \vdots \\ -\beta _{12}\epsilon _{2N}-\gamma _{11}\delta _{1N}-\gamma _{12}\delta _{2N}+\epsilon _{1N}+\zeta _{1N}\end{array} \right ] . $$
(A.4)

The Z 2,…,Z p+qn are constructed in an analogous fashion.

In the full system of equations, the coefficient vector A contains all of the intercepts, factor loadings and other coefficients in the model. It is the partitioned vector

$$ \mathbf{A=}\left [ \begin{array}{c} \mathbf{A}_{1} \\ \mathbf{A}_{2} \\ \vdots \\ \mathbf{A}_{p+q-n}\end{array} \right ], $$
(A.5)

where A j contains the intercept and coefficients for the jth equation in the model.

Continuing with the previous example of the equation for y 1i where y 1i depends on y 2i , x 1i , and x 2i ,

$$ \mathbf{A}_{1}=\left [ \begin{array}{c} \alpha _{\eta _{1}} \\ \beta _{12} \\ \gamma _{11} \\ \gamma _{12}\end{array} \right ]. $$
(A.6)

The other A j vectors are formed in a similar way.

Each of the regression coefficients in the original LISREL-type model appears only once in A. Variance and covariance parameters are not estimated by MIIV-GMM. Also, A has a lot of structure, with zeroes corresponding to the lack of a direct effect of latent or observed variables. Thus, there is a one-to-one correspondence between entries of A, on one hand, and entries of the collection B,Γ,Λ x , and Λ y , on the other.

Appendix B. Selection of MIIVs

The basic process for selecting MIIVs starts with all observed variables in the model and eliminates as potential MIIVs any variables that are directly or indirectly influenced by the errors or unique factors that are part of or that are correlated with the composite disturbance for a given equation. The remaining variables are the MIIVs for the given equation. More specifically, finding the MIIVs involves the following steps:

  1. 1.

    Make a list of all observed variables in the model since these are the potential MIIVs;

  2. 2.

    Make a list of all errors or unique factors (ϵs, δs, or ζs) that are included in the composite disturbance term that is part of the equation of interest;

  3. 3.

    Eliminate any observed variable that is directly or indirectly influenced by the errors or unique factors noted in Step 2;

  4. 4.

    Eliminate any observed variable that is directly or indirectly influenced by an error or unique factor that is correlated with the errors or unique factors noted in Step 2;

  5. 5.

    The remaining observed variables are the MIIVs for the given equation.

This procedure can be implemented in several ways. One is by visual inspection of the path diagram of the model. Another is by looking at the reduced-form model for each observed variable and determining whether the disturbances or errors in question have an effect. Finally, Bollen and Bauer (2004) provide a SAS macro to implement this check, and Bauldry (forthcoming), an implementation of the same algorithm in Stata. In virtually all identified SEMs that we have examined, there are sufficient MIIVs to estimate all equations in the model. There is no need to search for additional observed variables once a researcher starts with an (over)identified model. This is different than the usual IV approach where a researcher searches for auxiliary IVs that were not part of the original structure (Bollen 2012).

Appendix C. GMM Theory and Technical Aspects

This Appendix outlines the general theory of the generalized method of moments estimates. It restates the results given in the original development of Hansen (1982), as well as in comprehensive reviews such as Hall (2005) and Newey and McFadden (1986).

Given the p-variate data vector z and model parameters θ, the generalized method of moments works with q-variate vector of functions g(z,θ) that combine the data and parameters in such a way that in the population,

$$ \mathrm{E}\bigl[ g(\mathbf{z},\boldsymbol{\theta}_0) \bigr] = 0 $$
(C.1)

for the unique “true” value θ 0. Equations (C.1) are typically referred to as “moment conditions” in economics, or “estimating equations” in statistics (van der Vaart 1998; Godambe & Thompson, 1978).

The GMM proceeds as follows. First, the sample analogues to the estimating equations are formed:

$$ g_N(\boldsymbol{\theta}) = \frac{1}{N} \sum _{i=1}^N g(\mathbf{z}_i,\boldsymbol{ \theta}). $$
(C.2)

Second, these estimating equations are collected together into a quadratic form:

$$ Q_N(\boldsymbol{\theta};\mathbf{Z},W) = g_N'( \boldsymbol{\theta}) \mathbf{W}_N g_N(\boldsymbol{ \theta}), $$
(C.3)

where W N is a conforming q×q weight matrix, possibly obtained from the data. Third, this quadratic form is minimized with respect to θ to obtain the parameter estimates:

$$ \widehat{\boldsymbol{\theta}}_N = \arg \min_{\boldsymbol{\theta}} Q_N(\boldsymbol{\theta};\mathbf{Z},{\mathbf{W}}_N). $$
(C.4)

Thus, a GMM estimator \(\widehat{\boldsymbol{\theta}}_{N}\) is defined by a combination of the estimating equations g(z,θ) and the weight matrix W. Conceptually, both of them are at researcher discretion. However, generally the estimating equations are strongly determined by the model of interest (including our case of the model-implied instrumental variables), and some choices of the weight matrix W are obviously better than others, as explained below.

In the MIIV-GMM methodology discussed in this paper, the estimating equations are given by Equation (16):

$$g(\mathbf{z}_i,\boldsymbol{\theta}) = \mathbf{V}_{i}^{\prime }\bigl(\mathbf{y}_{i}^{\ast }-\mathbf{Z}_{i}\mathbf{A}\bigr). $$

After algebraic simplifications, these estimating equations reduce to linear combinations of cross-products ξ i ζ i , ξ i ϵ i , ξ i δ i , ζ i ϵ i , ζ i δ i and ϵ i δ i . Since these pairs of variables are assumed to be uncorrelated, the estimating equations have indeed zero expectations, as required by (C.1).

The desirable properties of the GMM estimates include consistency, asymptotic normality and, with an optimal choice of the weight matrix W N , asymptotic efficiency. Also, specification tests of whether the assumptions (C.1) are supported by the data are available. All of these results are asymptotic, and their justification requires certain regularity conditions.

Consistency of the GMM estimates (Newey & McFadden, 1986, Theorem 2.6, p. 2132; Hall 2005, Theorem 3.1, p. 68) is obtained under the following conditions:

  1. 1.

    z i ∼ i.i.d.;

  2. 2.

    \({\mathbf{W}}_{N} \stackrel{p}{\rightarrow} \mathbf{W}\) (Hall 2005, Assumption 3.7);

  3. 3.

    W is positive semidefinite (Hall 2005, Assumption 3.7);

  4. 4.

    WE[g(z,θ)]=0 iff θ=θ 0 (Hall 2005, Assumptions 3.3 and 3.4);

  5. 5.

    \(\boldsymbol{\theta}_{0} \in \operatorname{int}\boldsymbol{\Theta} \in R^{p}\) (Hall 2005, Assumption 3.5);

  6. 6.

    Θ is compact (Hall 2005, Assumption 3.8);

  7. 7.

    g(z,θ) is continuous at each θ with probability 1 (Hall 2005, Assumption 3.2);

  8. 8.

    E[sup θ g(z,θ)∥]<∞ (Hall 2005, Assumptions 3.2 and 3.10).

Instead of Condition 1, Hall (2005) uses a weaker conditions of strict stationarity and ergodicity of the data (Assumption 3.1), in which case i is the time index. Hall (2005) Assumption 3.1 also allows for heteroscedasticity of the measurement errors and unique variances.

Let us apply these conditions to the MIIV-GMM framework. The conditions on the weight matrix are satisfied for all the matrices we consider in this paper. The fourth condition is satisfied when the model is identified. Continuity of the estimating equations is trivial, as they are linear in the parameters. Finally, the last condition on E[sup θ g(z,θ)∥] is satisfied under the fourth-order cross-moments condition given in the section on the model and assumptions.

Asymptotic normality additionally requires the following conditions (Newey & McFadden, 1986, Theorem 3.4, p. 2148; Hall 2005, Theorem 3.2, p. 71):

  1. 9.

    g(z,θ) is continuously differentiable in the neighborhood of θ 0 with probability approaching 1 (Hall 2005, Assumptions 3.5 and 3.12);

  2. 10.

    E[g(z,θ 0)]=0 and E[∥g(z,θ 0)∥2]<∞ (Hall 2005, Assumption 3.11);

  3. 11.

    E[sup θ ∥∇ θ g(z,θ)∥]<∞ (Hall 2005, Assumption 3.2);

  4. 12.

    \(\operatorname{rank}E[ \nabla_{\boldsymbol{\theta}} g(\mathbf{z},\boldsymbol{\theta}_{0}) ] = p = \operatorname{dim}\boldsymbol{\Theta}\) (Hall 2005, Assumption 3.6);

  5. 13.

    GWG is non-singular for G=E[∇ θ g(z,θ)];

  6. 14.

    \(\sup_{\boldsymbol{\theta}} \| \frac{1}{n} \sum_{i} \nabla_{\boldsymbol{\theta}} g(\mathbf{z}_{i},\boldsymbol{\theta}) - \mathrm{E}[\nabla_{\boldsymbol{\theta}} g(\mathbf{z},\boldsymbol{\theta})] \| \stackrel{p}{\rightarrow} \mathbf{0}\) (Hall 2005, Assumption 2.13).

Let us apply these conditions to the MIIV-GMM framework. Smoothness of g(z,θ) is trivial since g(z,θ) is linear in θ. Finite second moment of g(z,θ 0) is ensured by the fourth-order cross-moments condition given in the section on the model and assumptions. For the estimating equations given by (16), the gradients with respect to the parameters A are given by

$$ \nabla_{\boldsymbol{\theta}} g(\mathbf{z},\boldsymbol{\theta}) = - \mathbf{V}_{i}^{\prime } \mathbf{Z}_{i}. $$
(C.5)

Finiteness of its absolute value follows from the finiteness of the second moment of the data. The condition on the matrix GWG is one of the conditions for the estimator (18) to be properly defined. The condition on the rank of the moment derivative matrix is similar to the condition of nondegenerate Jacobian in the likelihood context, and is satisfied whenever there are no perfectly collinear dependent variables in the model. The last condition is satisfied for i.i.d. data by virtue of the central limit theorem (CLT) for the derivatives of the moment conditions, since the first terms are of order O p (n −1/2).

Under the same set of conditions, the asymptotic variance estimator with GMM estimates plugged in for the population parameters is consistent for the target variance (Newey & McFadden, 1986, Theorem 4.5, p. 2160; Hall 2005, Section 3.5.1) when the data are i.i.d. For dependent or heteroscedastic data, one additionally needs (Hall 2005, Section 3.5.3):

  1. 15.

    sup θ E∥ 2 g(z i ,θ)/ θ j θ k ∥<∞ for all j,k in a neighborhood of θ 0.

As is easily seen, Equations (C.5) do not involve the parameters explicitly or implicitly, so this condition is easily satisfied for MIIV-GMM. Alternatively, if heteroscedasticity is a function of observed or unobserved variables present in the model, as is the case in the simulations, the expected values operators in the population conditions (C.1) and (15) will include integration over the variables that cause heteroscedasticity. After this integration is performed, the data form a skewed and kurtotic distribution. In our first simulation, this is demonstrated by the highly significant results of Mardia’s test.

Asymptotic efficiency is achieved with an optimal choice of the weight matrix W N . Namely, W N needs to converge to the asymptotic variance of the estimating equations g(z,θ 0). The result is given in Theorem 5.2 of Newey and McFadden (1986, p. 2165) and Theorem 3.4 of Hall (2005, p. 88), and does not require any additional assumptions beyond those necessary for asymptotic normality of the estimates.

A distinction should be made of the use of the term “moments” in the three literatures related to the current paper. In the statistics literature, a “moment” is universally understood as the expected value of a power (most typically, a positive integer power) of a random variable X, possibly centered, i.e., E[X k] or E[(Xμ)k] where μ=E[X], or their sample analogues. In the covariance modeling approach to structural equation modeling, “moments” refer to covariances σ jk =E[(X j μ j )(X k μ k )]. A further distinction is made of the sample, population, and implied moments. In the econometrics literature, the term “moment” is used more loosely to indicate any relation between (vector-valued) data X and (vector-valued) parameter θ such that E[g(X,θ)]=0. This generalization covers the standard uses: (i) E[Xμ]=0 for the population mean; (ii) E[X 2μ 2σ 2]=0 for the population variance; (iii) E[(X j μ j )(X k μ k )−σ jk (θ)]=0 for the covariance structure models. It also allows for other uses, such as the normal equations in regression, E[x j (yxβ)]=0, or the instrumental variables orthogonality conditions, E[z k (yxβ)]=0. At this level of generality, the econometric “moments” have the same meaning as “estimating equations” in statistics, the point we made in Sections 2 and 5.1. As the impetus for this paper comes from bringing the econometric ideas into latent variable modeling, we use the term “moment” in the latter, econometric, sense to denote the functions of data and parameters. For the MIIV-GMM application, the moments we use in estimation are given by Equation (15), in which the parameters are implicitly present in the composite error u.

Another way to look at these terminology distinctions is to observe that the covariance structure methods, such as MLE, ADF and other least square methods, make multiple steps from (i) setting up the model, as is done in Section 2, to (ii) deriving the implied second moments to (iii) minimizing the discrepancy between the sample and the implied moments to (iv) forming the variances of the sample moments to (v) utilizing the delta method to derive the standard errors of the parameter estimates. MIIV-GMM is, in fact, more straightforward, as it uses the equations from the latent variable and the measurement model with some minimal transformations, and obtains the standard errors explicitly with analytically available formulae that do not involve any derivatives. Given its greater simplicity, it is not surprising that the method works quite well in small samples even despite severe nonnormality of the data.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Bollen, K.A., Kolenikov, S. & Bauldry, S. Model-Implied Instrumental Variable—Generalized Method of Moments (MIIV-GMM) Estimators for Latent Variable Models. Psychometrika 79, 20–50 (2014). https://doi.org/10.1007/s11336-013-9335-3

Download citation

  • Received:

  • Revised:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11336-013-9335-3

Key words

Navigation