1F— 1 RARY lllllllllllllA‘ll\\lllvlllll\y‘llll MM 1293110540 4598 m“ Michigm Sue University LIBRARY This is to certify that the thesis entitled Parameter Estimation and Model Construction for Recursive Causal Models with Unidimensional Measurement presented by David Wayne Gerbing has been accepted towards fulfillment of the requirements for Ph.D. Psychology degree in . m (J - QA’Q\N\ g \ tilt" VVE (’F M ' of u aJor pr essor Date July 27, 1979 0-7 639 .1511. 32 4 JAN 132526081 9 Wm . ( "fitlxfl1"“ #5 i‘h - 57"! IUN 2 6 2001 351029 0 1 A41. OVERDUE FINES ARE 25¢ PER DAY PER ITEM Return to book drop to remove this checkout from your record. 1“- PARAMETER ESTIMATION AND MODEL CONSTRUCTION FOR RECURSIVE CAUSAL MODELS WITH UNIDIMENSIONAL MEASUREMENT By David Wayne Gerbing A DISSERTATION Submitted to Michigan State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Department of Psychology 1979 ABSTRACT PARAMETER ESTIMATION AND MODEL CONSTRUCTION FOR RECURSIVE CAUSAL MODELS WITH UNIDIMENSIONAL MEASUREMENT BY David Wayne Gerbing Two kinds of linear models often used in the social sciences are measurement models and causal models. Measurement models specify the relation of latent variables to observed variables and causal models relate latent variables to each other so that their ordering in the model specifies the outcome of an underlying causal process. The purpose of this paper is to compare the utility of two distinct estimation procedures which are embodied in two separate computer programs: PACKAGE and LISREL. The comparison is limited to recursive models with unidimensional measurement models. LISREL is based on the more recently developed procedures of "full information maximum likelihood". LISREL simultaneously estimates the parameters of both the measurement and causal models. PACKAGE is based on least squares techniques. The parameters of the measurement model are estimated with centroid factor analysis with communalities in the diagonal, and then, in a separate step, the estimated correlations among the latent variables are subjected to an ordinary least squares (OLS) path analysis. The results of this paper show PACKAGE to be a superior technique. David Wayne Gerbing If the model is correctly specified, then both PACKAGE and LISREL recover the underlying structure, although LISREL does so at a much greater cost in computer time than does PACKAGE. If the model is misspecified, then LISREL spreads the errors related to the misspeci- fied equations throughout the entire system so that all of the parameter estimates tend to be affected by the misspecification. In particular, even if the measurement model is correctly specified, a misspecification of the causal model precludes the correct recovery of the correlations among the latent variables. 0n the other hand, PACKAGE not only separates the estimation of the parameters of the two models, it is based on single equation techniques which localize the errors. Some authors have claimed that the use of the LISREL first deri- vatives for detecting misspecification is superior to the use of residuals. Instead, the use of derivatives was shown to be misleading in certain cases, and under none of the circumstances investigated in this paper did the derivatives provide more information than the residuals. One of the claimed advantages for LISREL was the capability of allowing for correlated disturbance terms and correlated measurement errors. Yet counter-examples were constructed in which the use of correlated errors was misleading. In particular, these examples demonstrate that, contrary to the current literature, omitted variables do not lead to correlated errors. False indication of correlated errors can be produced by (a) ad-hoc composites and (b) missing paths. In certain cases the covariance of "correlated disturbances" is exactly the OLS residual for a misspecified model defined by the deletion of a path. Dedicated to my best friend Monica Gerbing ii ACKNOWLEDGMENTS I would first like to thank my parents, Gordon and Marilyn Gerbing, and my grandparents, Alfred and Edna Bottorf, who have encouraged me from the beginning to be intellectually curious. Not only did they spend so many hours tutoring me from an early age, I was always treated with respect and trusted with responsibility. I would also like to thank those few teachers in high school and college who were able to inspire, teach, and at the same time provide encour- agement and understanding. For the past five years as a graduate student, I would especially like to thank my committee members not only for serving on the committee, but also for the additional support and knowledge they have provided throughout graduate school Charles Wrigley provided some helpful criticisms of the dissertation--especially regarding the historical development of factor analysis. Neal Schmitt provided suggestions for revising particular passages and also contributed to the overall organization of the dissertation. He has also answered many of my questions and provided encouragement and guidance over the last five years. Raymond Frankmann, in addition to contributing to the disserta- tion, has been an invaluable resource person since I arrived at Michigan State. He has been the one person I could turn to when, as was often the case, I needed help in negotiating through the maze of iii paperwork and bureaucratic procedures--or just learning my way around campus. His help was always appreciated, but especially so when it was most needed—-for this final certification. Bill Schmidt has contributed greatly to my graduate education. He has provided instruction over a wide variety of material that would not otherwise have been available. Moreover, he was the source of much of the material in this dissertation. Jack Hunter was my chairman and the guiding force throughout graduate school. He, more than anyone else, has contributed to my career and he is the person who encouraged me to major in quantitative psychology. I am grateful for the opportunity to work with him. Jack is unique in that he is both a mathematician and at the same time a substantive social scientist. He is also a friend who could provide guidance and moral support whenever it was needed. It was his emphasis on meaning and his ability to challenge the prevailing traditions that formed the theme of this work. I look forward to continuing our friendship and our professional relationship well beyond graduate school. My wife Monica is not only my best friend, but she has contri- buted hundreds of hours in directly helping me survive graduate school. She did all of the typing for this dissertation and she did it flawlessly-~including the mathematical terminology and Greek letters. She has also helped me in all other phases of life. I couldn't ask for a better dea1--and I hope she couldn't either. iv LIST OF LIST OF ‘ LIST OF Chapter I. TABLE OF CONTENTS TABLES O O O O O O O O O O O O O O C O FIGURES O O O O O O O O O O O O O O O APPENDICES . . . . . . . INTRODUCTION . . . . . . . . . . . . . . Causal Models . . . . . . . . . . . . . . An example of a causal model . . . . . . . . . Structural equations . . . . . . . . . . . . The covariance structure of a causal model Measurement Models . . . . . . . . . . . . Latent variables . . . . . . . . . . . . . The covariance structure of unidimensional measurement madels O O O O O O O O O O O I O O 0 Evaluation of the Full Causal Model . . . . . . . OLS and centroid factor multiple groups analysis . . LISREL . . . . . . . . . . . . . . . . The LISREL model . . . . . . . . . . . . The LISREL covariance structure . . . . . . . . The metric of the latent variables in a simultaneous LISREL analysis . . . . . . . . . . . . LISREL parameter estimation . . . . . . . . . LISREL two-step analysis . . . . . . . . Detecting Misspecification . . . . . . . . . . The actual vs. the specified model . . . . . . . The residual matrix (OLS and LISREL) . . . . . . Confidence intervals about parameter estimates (OLS and LISREL) . . . . . . . . . . . . . . First derivatives (LISREL) . . . . . . . . . . Likelihood ratio test (LISREL) . . . . . . . . "Reliability" of the model (LISREL) . . . . . . Miscellaneous oddities (LISREL) . . . . . . . Page viii ix U'lboNv—i 00‘ 12 12 16 17 19 21 25 26 27 27 27 28 28 30 32 33 Chapter II. III. IV. OVERVIEW AND PROCEDURE . . . . . . . . . . PACKAGE vs. LISREL . . . . . . . . . . . . The Equivalence of MGRP and OLS with LISREL Given Correct Models . . . . . . . . . . . . Procedure . . . . . . . . . . . . . TWO-STEP vs. ONE-STEP ESTIMATION: PACKAGE vs. LISREL . Single Equation vs. Full Information Methods . . An Example of a Misspecified Causal Model . . . Analysis of the measurement model . . . . . . . OLS path analysis . . . . . . . . . . . . LISREL simultaneous analysis . . . . . . . . . A comparison of OLS and LISREL . . . . . . . . Analysis of a variety of models and specifications . SPECIFICATION ERROR IN BOTH CAUSAL AND MEASUREMENT MODELS : AD-HOC COWOSITES o o o o o o o o The Pattern of Residuals from an Ad-hoc Composite Misspecified as a Construct . . . . . . A Re-interpretation of the Costner and Schoenberg AtlaIYSis O O O O O O O O O O O O O O O A More Complicated Model . . . . . . . . . . The Use of Derivatives in Detecting Misspecification in Ad-hoc Composites . . . . . . . . . . . . A re-analysis of previous studies . . . . . . . An extrapolation of the Sorbom procedure . . . Summary . . . . . . . . . . . . . . . . MISSING VARIABLES, MISSING PATHS, AND CORRELATED ERROR S O O O O C O O I O O O O O I The Effect of a Missing Variable on the Fit of the Model . . . . . . . . . . . . . . . . Previous assumptions . . . . . . . . . . . . A test of the assumptions: Deletion of an endogenous variable . . . . . . . . . . . . . . A test of the assumptions: Deletion of an exogenous variable . . . . . . . . . . . . . . Correlated Errors in Practice: The Use of Derivatives to Respecify a Model . . . . . . . . . . vi Page 34 34 36 38 42 42 45 45 46 47 51 51 54 56 61 64 68 68 7O 72 74 74 74 75 80 82 Chapter Correlated Errors: APPENDICES . . LIST OF REFERENCE NOTES LIST OF REFERENCES Summary vii Page 86 88 101 102 LIST OF TABLES Table Page 1 LISREL Notation . . . . . . . . . . . . . l8 2 Correct and Estimated Factor Correlations from OLS and LISREL and the Residuals . . . . . . . . 48 3 Parameter Values and Estimates . . . . . . . . 49 and n 50 2 3 O O O 5 Comparison of PACKAGE vs. LISREL . . . . . . . 52 4 The Residuals of the Indicators of n 6 Residuals from the Endogenous Indicators in Figure 7 . 58 7 MGRP Residuals and Partial Correlations . . . . . 62 8 Correlations of the Endogenous Indicators with E Partialled Out . . . . . . . . . . . . 63 9 The Second Order Correlations of the Endogenous Indicators . . . . . . . . . . . . . . 64 10 LISREL Residuals and Derivatives . . . . . . . 66 11 Partial Correlations . . . . . . . . . . . 67 12 Second Order Correlations . . . . . . . . . . 67 13 Derivatives of the Measurement Errors of the Endogenous Indicators . . . . . . . . . . . . . . 69 14 Residuals of the Misspecified Model of Figure 10 . . 79 15 Derivatives for the Model in Figure 6 Computed by LISREL O O O O 0 O O O O O O O O O O 83 viii Figure 10 11 12 13 14 LIST OF FIGURES A path diagram . . . . . . . . . . . . . A two-factor multiple groups measurement model A flow chart of the model building process . The LISREL partitioning of Z . . . . . . Four actual causal models used in this study . . . Actual model (a) from Figure 5 and an accompanying misspecified model . . . . . . . . . . . The actual and misspecified models from Costner and Schoenberg (1973) with parameter estimates computed by LISREL O I O C O O O O O O O O O 0 Actual model (a) from Figure 5 and a second misspecified model . . . . . . . . . . The actual and respecified models presented by Duncan (1975) . . . . . . . . . . . . Actual model (d) from Figure 5 and an accompanying misspecified model . . . . . . . . . Deletion of a block of exogenous variables which affects only a single endogenous variable . . . Actual model (d) from Figure 5 and misspecified and respecified versions of this model . . . . . Actual model (a) from Figure 5 and a third misspecified model with correlated disturbance terms 0 O O O O O O O O O O O I 0 0 An example of MGRP iterations . . . . . . . . ix Page 13 19 40 45 57 65 76 78 81 81 85 90 Appendix A B LIST OF APPENDICES COMMUNALITIES IN A MULTIPLE GROUPS ANALYSIS ANALYTIC SOLUTION FOR THE PARAMETER ESTIMATES OF MISSPECIFIED MODELS . . . . . . . . COMPUTATIONAL DETAILS OF A SIMULTANEOUS LISREL ANALYSIS 0 C 0 C O I C O C O O O O RELATIVE SIZES OF THE INDICATOR RESIDUALS WITHIN AND BETWEEN CONSTRUCTS IN AN AD-HOC COMPOSITE IF ALL THE INDICATORS HAVE EQUAL FACTOR LOADINGS . Page 88 91 95 100 CHAPTER I INTRODUCTION Causal Models The primary goal of science is theory construction. Most theories are descriptions of the causal processes among a set of variables. The direct study of the underlying causal processes, i.e., the study of the system dynamics, requires the construction of mathematical models of these processes expressed in terms of differential or difference equa- tions (e.g., Hunter & Cohen, Note 1; Hunter, Nicol & Gerbing, Note 2). A model of dynamics allows one to construct a developmental sequence of the variables in the system--a trajectory--which can be compared to the actual behavior of the system. Within the social sciences, however, the causal processes are usually not studied directly. Instead a causal theory, most often expressed qualitatively, is used to predict the relations among the variables. The model is tested by observing these relations §f£g£_the operation of the causal processes. Thus the distinction is drawn between (a) the underlying causal processes and (b) the effect these processes have on the relations between the system variables. And the outcome of the process, not the process per se, is usually the object of study within the social sciences. The statistical methodology employed to test a causal theory by studying the outcome of the causal processes is, in part, provided by regression analysis. For example, if a model identifies variable X as a causal antecedent of variable Y, then it follows that variables X and Y are related. If this relation is linear, then (a) the variables should be correlated, and (b) the slope parameter of the regression of Y on X, which indicates the linear relation between Y and X with the other predictor variables in the equation held constant, is interpreted as the indicator of the causal impact X has on Y. A "complete" system involves a set of variables such that many of the variables are consequent to some variables and antecedent to other variables. A causal theory which predicts such a network of relations can be tested by the analysis of a set of simultaneous regression equations. But the distinction should always be maintained between this model of relations among the variables following the outcome of the causal processes and the model of dynamics which is tested by the outcome model, even if the process model exists only in the form of verbal relations. An example of a causal model. Consider a theory which predicts that X1 causes X2 and that X2 causes X3 but X1 does not directly influence X3. The implications of this theory for the relations among the variables can be represented in either diagrammatic or equation form. If the relations among the variables are linear, the equations which define the model are: X2 ’ lex1 + U2 X + U X3 7 p32 2 3 The equivalent representation of the model in diagrammatic form appears in Figure 1. X1 is the single exogenous variable of the system (i.e., no antecedents Specified) and X2 and X3 are the two endogenous variables of the system. No constant term appears in the equations since the variables are mean deviates. The error terms, also called residuals or disturbances and denoted by U indicate that the speci- i, fication of the causal influences of the variables X2 and X3 is incomplete. Each disturbance is also specified to be uncorrelated with the predictor variables in the corresponding equation and with the other disturbances. That is, the disturbances are specified as random influences. 1’21 Figure l. A path diagram. p32 The model illustrated above is called a recursive model. For some authors (e.g., Duncan, 1975) a model is recursive if all of the causal linkages or paths form a hierarchy, i.e., the paths flow so that a path never leads back to the same variable from which the path began. Other authors (e.g., Heise, 1975) require that a recursive model be both hierarchical and that the disturbance terms across equations be inde- pendent. In this paper, the latter definition will be used to aid the distinction between hierarchical models with correlated disturbances and recursive models. Structural equations. The regression equations are called structural equations and the regression weights are called structural parameters or path coefficients. The diagram is called a path diagram. The structural equations and/or the path diagram define a causal model which is a specific interpretation of the underlying causal processes described by the theory. Since the relations between the variables are linear, the expected differences (denoted by "A") would be given by: AX AX and AX 2 a p21 1 3 = p32sz For example, if manipulation or natural processes changed the value of X1 by one unit for all of the individuals in the population, then the theory specifies that the mean difference in X2 would be: sz = 921(1) = p21 and predicted mean difference in X3 would be: Ax3 = p32AX2 = p32921 Thus the path coefficients indicated the amount of change that can be expected in the system variables given a change in antecedent variables following the operation of the underlying causal processes. The distinction between structural equations and regression equations which are not structural equations is the distinction between theory and blind prediction. If the variables can be measured, any variable can be regressed on any other variable or set of linearly independent variables without consideration of causality. For example, consider the situation in which two variables, Y2 and Y3, share a common antecedent Y1 but are not causally related in any other way. The variable Y3 can be mathematically expressed as a function of Y2 even though this relationship does not mirror an existing causal relationship. From a purely mathematical perspective, the equation Y3 = YY + U implies AY = YAYZ, but this derived relationship among 2 3 the changes in the two variables has no counterpart in physical processes. That is, a manipulation of a unit change in Y2 independent of Y will not cause a change in Y 1 3' The covariance structure of a causal model: An example. The key to the evaluation of causal models is the calculation of the covari- ances among the variables predicted by the model. The central idea is that the model imposes constraints on the covariances among the variables and it is these constraints on which the test of the model is based. For example, consider the causal model presented at the beginning of this paper in which X1 causes X2 and X2 causes X3. Multiplying both sides of the equation for X by X yields: 3 1 x3X1 = p32X2X1 + x1U3 Taking expected values, E(X3X1) = p32E(X2X1) + E(X1U3) Since the variables are mean deviates, 0(X3X1) = p320(X2X1) + 0(X1U3) But 0(X1U3) = 0. And if the variables are standardized, r31 = p32‘21 Finally, r since there is only a single predictor in the p32 = 32 equation for X3, so r31 = r32’21 Thus the model imposes a structure upon the correlations among the variables. In this simple example it is this single constraint which provides a test of the model by comparing the observed value of r31 with the predicted value E obtained by multiplying the observed 31 values of r32 and r21. This test of the model that r A = r is equivalent to tests based 31 31 on partial correlation coefficients and multiple regression coefficients. Since r = r31 ' r32r21 r31 ' r31 31-2 2 a 2 a = 2 3 2 k (1 ‘ r32) (1 ‘ r21) (1 ' r321) (1 ' r21) the model implies r 31.2 = 0. If X3 is regressed on X1 and X2 and 831.2 is the corresponding multiple regression weight, then: A B = r31 ' r32r21 = r31 ’ r31 31-2 1 _ r 2 1 _ 2 21 r21 Thus, the model implies 831.2 = O. The variables X3 and X1 are related since r31 # 0, but only through the intervening variable X2. The general principle of constructing a model defined by regres- aion equations and then using this model to derive the constraints imposed by the model on the covariances among the variables has been known for over 50 years. In 1921, Wright introduced "path analysis" as a test of recursive causal models, though it was not until Simon's (1957) work, followed by Duncan (1966) and Blalock (1964) that path analysis began to be used by social scientists other than econometri- cians. The computation of the predicted correlations of a causal model is covered extensively in the literature on causal models. Standard references are the texts by Heise (1975), Duncan (1975), Asher (1976), and the article by Lewis-Beck (1974). Measurement Models Latent variables. Before the variables can be related by a set of regression equations, they must be measured. The variables of interest, the variables in the causal model, are called the latent variables since they are not directly observed. The observed variables are the indicators of the latent variables. The relation between the indicators and the latent variables is specified by the measurement or factor model. The measurement model is a set of simultaneous regres- sion equations of the indicators regressed on the latent variables. Each measure is an imperfect indicator of a latent variable because of random response error and invalidity. The problem of measurement error for the analysis of causal models is more serious than simply the lack of precise measures of the latent variables. The analysis of a causal model based on fallible data yields biased parameter estimates (e.g., Wiley, 1973). For a regression equation with a single predictor, measurement error in the independent variable attenuates the slope coefficient in proportion to the reliability of the independent variable. However, for multiple regression equations the effect of measurement error of the independent variables on the estimates of the regression coefficients is usually unpredictable. Wiley (1973) presented an example of a two-predictor multiple regression equation in which the presence of measurement error reversed the magnitude of the sample parameter estimates from the values of the population parameters. The problem of measurement error can be countered by providing multiple indicators of each latent variable. The use of multiple indicators improves the precision of the estimates of the latent variables and allows reliability estimates of the composite scores to be computed. Traditionally, these reliability estimates are used to correct for attenuation the correlations among the latent variables computed from the observed composites. The covariance structure of unidimensional measurement models. The testing of measurement models in which the measures are partitioned into clusters such that the measures in each cluster are postulated to be alternate indicators of only a single common latent variable is called Spearman factor analysis or oblique multiple groups analysis (Tryon, 1939; Holzinger, 1944). The original conceptualizations were provided by Spearman in 1904. The computations of a multiple groups analysis involve any factor analytic method by which a single factor is extracted from each group. The computations provide estimates of the factor loadings of each indicator on the group factors. An extensive discussion of the construction and evaluation of unidimensional measurement models, an example of a procedure called confirmatory factor analysis, is provided by Hunter (Note 3) and Hunter and Gerbing (Note 4). Measurement models are evaluated according to the same principles which underlie the evaluations of causal models. The basic idea is to derive the covariances among the observed variables predicted by the model. The covariance structure of unidimensional measurement models is given below. As presented in Figure 2, let the observed variables X1 and X2 be indicators of latent variable (or true score or factor) F and let Y be an indicator of G. The errors of measurement are specified as random. The curved double-headed arrow in Figure 2 represents a correlation without the specification of causality. 9% O 6% a Figure 2. A two-factor multiple groups measurement model. In equation form, X = 81F + e l 1 X2 = 82F + e2 and Y = BYG + eY where r(F,ei) = r(G,ei) = r(ei,ej) = 0. The covariance structure of the indicators can be described by considering (a) the covariances among indicators of different factors, and (b) the covariances among indicators of the same factor. Consider first the relation between X1 and Y. The model implies: XlY = BIBYFG + BlFe1 + BYGeY + eleY E(X1Y) = BIBYE(FG) + BlE(Fe1) + BYE(GeY) + E(e1eY) Standardizing the observed and latent variables, 8 B r11! 2 1 YrFG where 81 and 82 are standardized regressions weights. Since B=r 1 and BY = r 1F YG by substitution, rlY = rlrrrcrcr Thus one test of the measurement model is the comparison of the observed value of rl with the value predicted by the model rlY’ which Y is the product of the observed values of r rFG’ and r 1F, GY' Hunter and Gerbing (Note 4) call the expression r1Y = rlFrFGrGY the product rule for external consistency of items. By similar logic, they also derive the product rule for external consistency of items and 10 factors, which is, r19 = rlrrrc If X1 and X2 are both indicators of the game factor F, an equivalent test based on these covariance restrictions can be used to test the unidimensional measurement model. The product rules imply that the correlations of X1 and X2 with other items or other factors are proportional to their respective factor loadings on their own group factors. That is, for some other factor C, These proportionality rules are equivalent to the tetrad difference criterion, which is usually written as r1Gr2F = rZGrlF 0r rerZF = rZleF They were presented in the context of factor analysis by Spearman in 1907 and first illustrated in the context of data analysis by Burt (1909) and Spearman (1914). The stability of this proportionality of the correlations of two indicators of the same factor across other indicators and/or factors is indexed by the following formula, originally used as a measure of the similarity of two factors. k§1r(xixk) 11(ij k) (2 n {r(XiXk)} 2)L5 (:12 {r The remaining step in the analysis is the computation of the communality estimates which is accomplished with an iterative procedure outlined in Hunter (Note 3) and Hunter and Gerbing (Note 4). The communality of interest for each observed variable is the communality of the indicator across the remaining indicators in the corresponding group. As Hunter (Note 3) demonstrated, if these communalities are inserted in the diagonal of the correlation matrix, then the observed correlation matrix among the indicators is transformed to the covariance matrix of true scores. Hunter (Note 3) also demonstrated, by example, that factoring 15 a covariance matrix of true scores implies that the computed factor loadings and factor-factor covariances are corrected for attenuation (Spearman, 1907) due to the measurement error in the composite (i.e., factor) scores. At the same time, the use of communalities also eliminates the Spuriously large "part-whole" correlations between indicator and factor, i.e., communalities eliminate the upward bias of correlating an indicator with a composite of which it is a part. However, if the indicators can not be measured without error, communalities must be used in the multiple groups analysis if parameter values are to be correctly estimated. The convergence of this iterative process for these data is discussed in Appendix A. The input to the multiple groups analysis is the correlation matrix of the observed variables. The input to the path analysis is the correlation matrix among the latent variables. Parameter estimates of recursive models have traditionally been accomplished with ordinary least squares (OLS). That is, the estimated parameter values are those values that minimize the sum of the squared errors of the sample data points about the sample regression surface for each equation in the model. Both centroid factor multiple groups analysis and OLS path analysis are called single equation estimation procedures. That is, the computations of these analyses are accomplished "equation by equation"; only the variances and covariances of the variables which appear in a particular equation are used in the computations of the estimated para- meters of that equation. The general equation for a multiple indicator measurement model is Xi = 11F + ei 16 since each indicator or observed variable is a function of only a single factor plus measurement error. In a recursive model, the factors can be ordered so that all variables causally antecedent to a given factor are listed ahead of it. If the variables are so listed, then the general equation for a recursive path analysis is i—l F=£PF+U 1 k=1 ik k 1 The computation of both the centroid factor multiple groups analysis and the OLS path analysis may be accomplished with PACKAGE (Hunter & Cohen, 1969). The use of the multiple groups subprogram is explained by Hunter and Cohen (Note 6) and Hunter and Gerbing (Note 4). The path analysis subprogram is explained by Hunter (Note 7). LISREL. Based on the work of Lawley (1943) and Bock and Bargmann (1966), J6reskog (e.g., 1967, 1978) has developed an alternative analysis of causal and measurement models which differs from the OLS with correction for attenuation strategy along two different dimensions. J6reskog has also developed a computer program, LISREL, which contains the computational algorithms (J6reskog & SSrbom, 1978). Both analytic methods, OLS with correction for attenuation and LISREL were developed to account for the biasing effects of measurement error. However, LISREL simultaneously estimates the parameters of the measurement and causal models. LISREL also is not restricted to testing multiple indicator measurement models in which each indicator is an indicator of only a single factor, i.e., multiple groups measurement models. Nor does LISREL require the measurement errors to be uncorrelated. In terms of the causal model, LISREL is not restricted to the analysis of recursive path models. The causal model may be nonrecursive in that the 17 disturbances may be correlated and/or "feedback loops" may appear in the model. The complexity of the general LISREL model requires the use of matrix algebra in its representation. Since J6reskog's parameteriza- tion of the model includes an explicit distinction between exogenous and endogenous variables, there is also a much greater notational complexity. Many of these symbols are defined in Table 1. Other symbols include ny for the covariance matrix of the exogenous and endogenous observed variables and Z for the covariance matrix of all of the indicators. The LISREL model. Joreskog has parameterized the general measure- ment model in separate equations for the exogenous latent variables and a model for the endogenous latent variables. y Ayn +_e and E(fl_€')‘-"O pxl pxm mxl pxl X Ax; +6 and E(§§') = 0 qx1 qxn nxl qxl The 1th row of the factor pattern matriceslly andllx corresponds to the ith endogenous or exogenous indicator. The jth column of each of the matrices refers to the jth endogenous or exogenous factor. Thus if the scalar equations take the form of y1 = Ky n + Ei or xi = Ax g + 61 i i as is the case for the multiple indicator models with one indicator per factor, then each rowofAyandAx will contain a single A with the remainder of the elements in the row equal to zero. The parameterization of the causal model is best understood by beginning closer to the more traditional path analytic parameterization, 18 Table 1 LISREL Notation Variables Number Variable of such Covariance psymbol variables matrix Observed x q 2 xx Exogenous Unobserved 5 n ¢ Measurement error 6 q 96 Observed 2 y p yy Unobserved n m C Endogenous Measurement error e p 08 Disturbance c m w Parameters Scalar Matrix Function Ax Ax A x 54x1 Measurement A model Y A A i y y E ——v-yi F Y Y Causal g n model Bji 19 N F.=£P F +U, 1 k=1 ik k 1 where there are N latent variables, designated by F. In matrix form, £= PE +9 le NxN le le If a distinction is made between exogenous and endogenous factors, then the model can be rewritten as n=An+I§+£ mxl mxm mxl mxn nxl mxl For recursive models, A is a lower triangular matrix with 0's down the main diagonal and U) E E(5__c_') is a diagonal matrix with the disturbance variances down the main diagonal. However, J6reskog's parameterization of the causal model is based on the following transformation of the above equation. (I - A)fl = P §_+ C B3=F§+£ andEgg')=0 where B E I - A. Thus B has 1's down the main diagonal and the path coefficients are reversed in sign. For recursive models, B is still lower triangular. The LISREL covariance structure. Given the three equations which define the model, it is now possible to derive the covariances among the indicators predicted by the model. The derivation is based on the partitioning of 2 shown in Figure 4. I I Z = - - - -I ----- I I Figure 4. The LISREL partitioning of Z. 20 For example, to compute the implied covariances among the indicators of the exogenous variables, Xxx; E(xx') = E{(l\x_€_+_6_)(Ax§_+_6_)'} MAX; g'AX') + meg 53') + E(§§'Ax') + E(§_6_') =AXE(§§_')AX' +AXE(_§') + E(§§_')Ax' + E(_<_5_§)' __ ' AX¢AX + 96 which is the usual expression for the correlation of the observed variables expressed as a function of the factor pattern matrix, the correlation matrix among the oblique factors, and the communalities. Similar derivations lead to X E Egyxf) = yx E{(Ayn+_€> Hmm .uaofiofimmmoo sung oco haao :a Houum uouuo oz nouns oz mac + mmoz aofiumofimwommmmfia wawuoouma musofiowmmooo numm md—OH UNHGHHOU .HOUUNh mmcwwmoa uouomm coaumufiuu qmmmHg .m> mo r(i,'fi) or A > L n. D 1 i For example, A11 = .45 and Afi. = .30, as shown in Figure 7. 1 In terms of predicting the pattern of residuals for the misspeci- fied model, consider first the correlations among the indicators of the same properly specified construct. The correlation among these indi- cators adheres to the product rule for internal consistency, i.e., for h the it and jth indicators of n, r(i.j) = r(i.n) r(n,j) But for the misspecified model, £ua>=flnmflmp And, from above, I). A |>|A~A~| Di nj Hi nj If the residual is defined as Res(1.j) s r(1.j) - £(1.j) then the expected residuals of an ad-hoc composite misspecified as a construct should be positive for the indicators of the same construct 60 in the correctly specified model, as they are in this example in which Res(y1,y2) = .09 and Res(y3,y4) = .11. Now consider the residuals among the indicators of different constructs in the ad-hoc composite. The observed correlations for the indicators were computed with the product rule for external consistency. For example, r(y1.y3) = r(y1.nl)r(n1.n2) r(nZJZ) (.45) (.48) (.55) = .119 But the predicted correlations among these indicators which have been falsely placed in the same cluster will again be computed using the product rule for internal consistency. For example, (.30) (.53) = .159 So the residual is .119 - .159 = .040. Thus there are two competing influences on the relative sizes of r(i,j) and r(i,j), where i and j are indicators of different constructs in the actual model. As shown before, the actual factor loadings are larger than the computed factor loadings. This discrepancy is large and the residual is positive, to the extent r(nl,n2) is small. What is new for this situation is the presence of r(nl,n2) in the expres- sion for the actual correlation r(i,j). The smaller r(n1,n2), the smaller the product r(i,rh)‘r(nl,n2):r(n2,j). Let the value of r(n1,n2) decrease, but let the remainder of the actual model be unchanged. If the value of this product decreases faster than the discrepancies between the actual and computed factor loadings, the residuals could even become negative, as they are in this example. But 61 these residuals should at least tend to be smaller than the residuals of indicators of the same construct. This result is proved in Appendix D for the special case of equal factor loadings in the actual model. A Re-interpretation of the Costner and Schoenberg Analysis Costner and Schoenberg (1973) investigated the effects of misspec- ification with the strategy utilized in this paper. One of their models, which is presented in Figure 7, was misspecified to form what is here labelled an ad-hoc composite. Costner and Schoenberg (1973) predicted the pattern of residuals resulting from the analysis of this model without any formal justification. They simply indicated that "our intuition is that . . . we might expect large residuals between . . . the indicators of [E] on the one hand and the indicators of [n1] on the other, or between indicators of [E] and [n2]" (p. 175). Following an analysis of the misspecified model with LISREL, the authors noted that "the actual pattern of residuals . . . does not conform to this pattern at all" (p. 176). How should the model be respecified given the residual matrix? Costner and Schoenberg (1973) believed that the respecification of a model when confronted with a matrix of nonzero residuals among the indicators should be to allow the error variables for indicators with the largest residual to be correlated. They respecified the model by relaxing the r(sl,sz) = 0 constraint and noted that (a) the respeci- fied model fits relatively well, and (b) this result was misleading since the respecified model was not the correct model. The authors then concluded on the basis of this example and others that "the respecification suggested by an intuitive appraisal of the pattern of residuals may be grossly misleading" (p. 177). Rejecting the 62 information provided by the residuals, the authors go on to devise a laborious procedure for detecting misspecification which involves separately testing all possible combinations of two-indicator models and then testing specified combinations of three-indicator models. However, the obtained pattern of residuals does conform to the predictions of the previous section. The residuals between indicators of the same construct in the correctly specified model are positive. The residuals between indicators from different constructs are not only smaller but negative in this example. Contrary to the conclusion of Costner and Schoenberg (1973), the residuals do appear to provide useful information for the respecification of this misspecified model. The incorrect assumption is that positive residuals imply correlated errors. Does the use of PACKAGE provide information not provided by LISREL? The application of the multiple groups analysis to the Costner and Schoenberg (1973) example generated the residuals and partial correlations presented in Table 7. Table 7 MGRP Residuals and Partial Correlations Residuals Partial Correlations Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 y1 .00 .09 ' —.06 -.07 y1 1.00 .14 ' -.07 -.06 I I y 09 000 I -006 7.001 y .14 1000 ' -008 -007 2 _______ , _______ 2 ______ L _______ y3 -.06 -.06 ' .00 .11 y3 -.07 -.08 ' 1.00 .17 I I y4 -.07 -.01 ' .11 .00 y4 -.06 -.07 ' .17 1.00 I I 63 The residuals derived from PACKAGE follow exactly the same pattern as the LISREL residuals. However, since the specified model contains only two factors, the LISREL simultaneous analysis is equivalent to a confirmatory factor analysis. Alternate residuals can be obtained by partialling out a single factor at a time-—a straightforward operation with the PACKAGE subprogram PARTIAL following the use of the MGRP subprogram. In the actual model r(n1,n2) can be decomposed entirely into the spurious influence of an exogenous variable E, i.e., r(n1,n2-E) = 0. This same relationship among the factors is mirrored by the indicators of the factors, as illustrated in Table 8. Again, contrary to the con- clusion of Costner and Schoenberg (1973), the residuals provide directly usable information for locating specification errors. Table 8 Correlations of the Endogenous Indicators with E Partialled Out Y1 y2 Y3 Y4 yl 1.00 .21 : .oo .oo y2 .21 1.00 : .oo oo y3 .oo .oo : 1.00 .21 y4 .oo .00 z .21 1.00 Finally, Costner and Schoenberg (1973) could have applied a test suggested by Spearman in 1914 that would also have unambiguously deter- mined that the misspecified model contained an ad-hoc composite. Spearman noted that the proportionality constraint which follows from the product rule for external consistency implies that the correlations 64 of two indicators of the same construct across other variables are "perfectly correlated" (1914, p. 109). These "intercolumnar" or "second order" correlations may be computed by correlating the corre- lations of each pair of indicators with the diagonal value of 1.00 defined as missing data. The resulting matrix of second order corre- lations is presented in Table 9. Actually, similarity coefficients (e.g., Hunter, 1973) should be used instead of the second order corre- lation coefficients since proportionality is a stricter criterion than linearity. Second order correlations were computed because of the availability of the computer program with a missing data provision. Table 9 The Second Order Correlations of the Endogenous Indicators I I yz 1.00 1.00 1 .26 08 __________ '_ _ _ — ._ _ _ _ _ y3 .05 .26 . 1.00 1.00 I y4 —.02 .09 , 1.00 1.00 A More Complicated Model Since the misspecified model analyzed by Costner and Schoenberg (1973) contained only two latent variables, a simultaneous LISREL analysis is equivalent to a LISREL CFA analysis. Consider the actual and misspecified models in Figure 8. The factor loadings of the correctly specified factors were correctly recovered while the factor loadings of the ad-hoc composite were attenuated from their true values. The simultaneous and CFA solutions in terms of factor loadings, 65 residuals, and derivatives were identical. The LISREL residuals among the six indicators and the derivatives for the measurement error corre- lations between the six indicators of the ad-hoc composite are presen- ted in Table 10. The pattern is redundant between the two matrices. Contrary to the assertion of Costner and Schoenberg (1973), the residuals clearly indicate subdivision of the indicators into two sets. Contrary to S6rbom (1975), the derivatives falsely suggest correlated errors . Actual model Misspecified model Figure 8. Actual model (a) from Figure 5 and a second misspecified model. The partial correlations of the indicators of the ad-hoc composite from the PACKAGE analysis are presented in Table 11. This pattern 66 Table 10 LISREL Residuals and Derivatives Residuals Y1 y2 y3 Y4 y5 Y6 .00 .23 .14 : -.11 -.11 .08 23 .00 .09 : -.11 -.11 .08 14 .09 .00 : -.08 —.08 .06 ____________ 1“ - _ _ _ _ _ - _ _ _ _ -.11 —.11 -.08 1 .00 .17 .11 —.11 -.11 -.08 : .17 .00 .06 I -.08 -.O8 -.06 I .11 .06 .00 Derivatives VI 22 Y3 Y4 ys Y6 I .00 -.41 -.23 , .24 .21 .13 I -.41 .00 -.13 I .22 .18 11 I -.23 —.13 .00 , .14 .12 .07 ____________ '- _ _ _ _ _ _ _ _ _ _ _ .24 .22 .14 , .00 -.37 .20 I .21 .18 .12 , —.37 .00 .09 I 13 .11 .07 , -.20 -.09 .00 67 Table 11 Partial Correlations V1 2, Y3 Y4 ys Y6 y1 .00 .28 .16 : -.19 -.15’ -.11 y2 .28 .00 .10 : - 15 -.12 -.08 y3 .16 .10 .00 : -.11 -.08 -.06 y4 -.19 -.15 -.11 : .00 .28 .16 y5 -.15 - 12 —.08 : .28 .00 .10 y6 -.11 -.08 -.06 : .16 .10 .00 Table 12 Second Order Correlations y1 y2 y3 y4 y5 y6 yl 1.00 1.00 1.00 : .04 .11 .08 y2 1.00 1.00 1.00 : .11 .14 .10 y3 1.00 1.00 1.00 : .10 11 .07 y4 .04 .11 .10 : 1.00 1.00 1.00 y5 11 .14 .11 : 1.00 1.00 1.00 .08 .10 .07 : 1.00 1.00 1.00 68 the partial correlations was again similar to the pattern of the residuals and/or derivatives from the LISREL analysis. The matrix of second order correlations is presented in Table 12. In practice, the indicators of the ad-hoc composite might not be listed in subclusters as they are in the examples presented here. Use of the PACKAGE subprogram ORDER on the matrix of partial or second order correlations would reorder the variables so that the indicators which define the component constructs are listed consecutively. The Use of Derivatives in Detecting Misspecification in Ad-hoc Composites A re-analysis of previous studies. SSrbom (1975) was interested in models with correlated errors. He accepted Costner and Schoenberg's (1973) conclusion regarding the problems with the residual matrix for detecting misspecification, but he sought a less troublesome alterna- tive than the procedure outlined by Costner and Schoenberg. SBrbom (1975) advocated the use of the first derivatives as an alternative to the residuals for locating misspecification. "We should relax the zero-restriction for that element which gives the largest decrease in F" (p. 143), i.e., the element with the largest first derivative. He demonstrated that the procedure worked for an actual model with corre- lated errors which was misspecified as a model with uncorrelated errors. However, Sfirbom's (1975) procedure would lead to false conclusions in the case of the collapsed indicator model or ad-hoc composite. For the Costner and Schoenberg (1973) example, the matrix of first deriva- tives for the covariance matrix of measurement errors is given in Table 13. The pattern of first derivatives is redundant with the 69 Table 13 Derivatives of the Measurement Errors of the Endggenous Indicators y1 y2 Y3 Y4 y1 .00 -.23 I .07 .09 l y2 - 23 .00 I .07 10 ________ 4_-______. y3 .07 07 I .00 - 12 I y4 .09 10 I -.12 .00 I information supplied by the matrix of residuals. The only difference between the patterning of the derivatives and the residuals is that the derivatives have opposite algebraic signs. If the model were respeci- fied by falsely assuming correlated measurement errors for Y1 and y2 and for Y3 and y4, then the actual structure would not be recovered, as Costner and Schoenberg (1973) have already noted. Saris, Pijper and Zegwaart (1978) also noted that the Costner and Schoenberg (1973) procedure for detecting misspecification was "quite time-consuming and not completely clear as to how one should proceed in all circumstances" (p. 152). Following Sbrbom (1975), they were concerned only with first derivatives, but they sought to improve SBrbom's procedure by considering the correlations among the deriva- tives. For each of the m fixed parameters, they computed the function In. A W1 7 jflrij 1 where 91 is the estimated first derivative of the ith parameter, and $11 is the estimated correlation between the derivatives of the 1th and th j parameters. They proposed "a stepwise procedure . . . where at 70 each step the restriction is dropped with the highest 3_in absolute value" (Saris et al., 1978, p. 158). Saris et al. (1978) began with two-factor models which contained (a) correlated measurement errors and/or (b) observed variables which were indicators of both factors. The misspecified model was always the corresponding two-factor model with unidimensional measurement, i.e., no correlated error variances and all indicators were indicators of only a single factor. They concluded that "the Saris procedure per- forms better than Sfirbom's" (p. 163). However, like SBrbom, these authors never began with a recursive model with unidimensional measure- ment. Like Sorbom, they never considered residuals. Thus the "improvement" suggested by Saris et al. (1978) leads to the same error of the Sarbom (1975) method when applied to the Costner and Schoenberg (1973) example. Since all derivatives except the derivatives of the measurement errors of the endogenous indicators shown in Table 13 were zero, the application of the Saris et al. method to these data would also lead to a false respecified model which con- tains correlated measurement errors. The only potential difference between the procedures is that the correlated measurement errors might be added in a different order. An extrapolation of the Sarbom_procedure. As constraints of the model are relaxed, the apparent fit of the model continually improves. In the ad-hoc composite example, the suggestion of SBrbom (1975) to assume correlated measurement errors does improve the apparent fit of the model though the respecified model is false. The question addressed in this section is, what is the end result of allowing increasingly more measurement error covariances? 71 The largest residual from the LISREL analysis of the misspecified model in Costner and Schoenberg (1973) was between the indicators of y1 and y2. The largest derivative was between the measurement error covariances of the corresponding error terms. If this error covariance is freed, it assumes the value of .198, the fit of the model improves since F(§) decreases from .588 to .142, and the factor loadings of the two indicators decrease about .04. The largest residual is now between y3 and y4 and the largest derivative is between the corresponding covariance between their error terms. If the model is respecified again by adding this second measurement error covariance, the model fits perfectly, i.e., F(§) = 0.000. The estimated regression para- meters continue to change with each respecification since the factor loadings of y3 and y4 decrease about .09 and, more interestinly, the coefficient relating the latent variables becomes equal to 1.00. The generality of this result can be checked by examining the more complicated model which appears in Figure 8. The misspecified model was successively respecified by freeing a new measurement error covar- iance on each respecification. The chosen covariance on each round corresponded to the largest derivative or, equivalently, the largest residual. The model was respecified until perfect fit was obtained. In general, as more measurement error covariances are added, the factor loadings decrease in magnitude while the estimated regression parameters of the causal model increase in magnitude. The fit of the model continues to improve until six covariances have been included, at which point F(§) - .014. Yet in this "perfect" model, B is equal to the nonsense value of .911. And this "perfect" model is, in actuality, a misspecified ad-hoc composite. 72 The pattern of correlated errors reveals the nature of the misspecification. The misspecified model fit the data perfectly if all the covariances between measurement errors of indicators of the same construct were unconstrained. For the Costner and Schoenberg (1973) model this criterion was achieved after only two measurement error covariances were freed since the ad-hoc composite contained the indi- cators from only two factors and each factor had only two indicators. The present model required six free error covariances since there were three indicators for each factor. The apparent fit of a model with correlated measurement errors means only that the model is misspecified. If the model fits poorly without correlated errors, but fits very well with correlated errors, and if the indicators can be partitioned into clusters with positively correlated errors within clusters and zero or negative covariances between clusters, then at least one of the latent variables is an ad-hoc composite. The indicators should be partitioned accordingly. However as was shown earlier, the same analysis can be performed directly on the original model residuals. There is no need to obtain an intermediate solution with "correlated errors". Summary In contemporary path models, a construct is defined as a latent variable whose indicators form a unidimensional set in the sense of Spearman (1904). To use such models, one must "reduce" composites which are measured by a conglomerate of constructs into component variables. Such an ad-hoc composite can be regarded as a model which has been misspecified by collapsing the indicators of distinct constructs into a single latent variable. Moreover, the recognition of 73 such a misspecification can be accomplished by the examination of the residuals--although previous work has claimed that the residuals were not helpful in detecting misspecification. Contrary to previous claims, the derivatives from a LISREL analysis were shown not to add any more information than the residuals. Indeed the recommendations of Sgrbom (1975) and Saris et al. (1978) were shown to lead to the false assertion of correlated errors. CHAPTER V MISSING VARIABLES, MISSING PATHS, AND CORRELATED ERRORS No existing path model contains all the relevant causal factors. If only complete models were capable of analysis using OLS, then there would be no situations in which OLS could be used. This is precisely the claim of a number of contemporary authors. They argue that missing variables always result in correlated disturbances and hence the use of OLS is never justified in real data sets. The following examples show this claim to be false. On the contrary, these examples suggest that most recommendations for correlated disturbances are misguided responses to data sets which call for the addition of missing paths in the misspecified models. The Effect of 3 Missing Variable on the Fit of the Model Previous assumptions. The traditional estimation procedure used in path analysis is OLS, which assumes that the disturbance terms across equations are uncorrelated. However, several authors have argued that correlated disturbances could be produced by missing variables. These arguments are important since the complete specifi- cation of all of the variables in a causal process is impossible. Presumably, if a common antecedent of some of the endogenous variables is omitted, the omitted variable will "appear" in the corre- sponding disturbance terms since the disturbance represents the influences which operate on the "dependent variable" of the equation 74 75 which are not described in the set of predictor variables in the equation. For example, "[The realism of] the assumption that the disturbance terms are mutually uncorrelated in any given instance will depend upon the completeness of the causal system . . . . Whenever common causes of the disturbance terms for two or more equations can be located or measured, they should be explicitly introduced into the equations as additional variables" (Namboodiri, Carter, & Blalock, 1975, pp. 446 and 448). "If the same explanatory factor is excluded from.more than one equation, the effect of that factor will be present in more than one error term and will cause the error terms to be somewhat correlated . . ." (Hanushek & Jackson, 1977, pp. 230 and 231). "The uncorrelated residuals assumption is basically equivalent to the assertion that there is no confounding variable impinging upon both X1 and X2 where by a confounding variable is meant any unmeasured factor that directly influences two or more of the measured variables" (Asher, 1976, p. 16). "Correlated error terms arise when omitted variables simultaneously influence different observed variables" (Saris, Pijper, & Zegwaart, 1978, p. 161). A test of the assumptions: Deletion of an endogenous variable. The authors cited above believe that a missing variable results in correlated errors. The example in Figure 9, which was presented by Duncan (1975), appears to contradict this principle, although this example did not appear in a discussion of correlated errors. Duncan (1975) noted that "the OLS estimator of Y in [the respecified model] 31 estimates Y without bias. The principle is that insertion of an 21832 "intervening variable" into one path of an initial model does not inValidate that model, but merely elaborates it" (p. 109). The following discussion is an elaboration and extension of these principles. 76 Actual model Respecified model Figure 9. The actual and respecified models presented by Duncan (1975). 77 Consider the more complex model presented in Figure 10 where the misspecification is defined by the deletion of n The parameter 3. estimates presented in the misspecified model were computed with OLS. The residuals of the misspecified model are presented in Table 14. The pattern of residuals is that which would be produced by the deletion of a path. If the misspecified model is respecified by the inclusion of a path from n to n4 with 841 = .09, then (a) the other 1 parameter estimates remain unchanged and (b) the respecified model fits perfectly, i.e., all residuals are zero. The difference between the two models is that in the misspecified model, the elimination of the path from n1 to n3 to n4 was not replaced by a direct path from n1 to n4. Thus the causal impact of n on 114 was not accounted for in the 1 misspecified model. Any model is incomplete in the sense that there are missing variables. Otherwise all multiple correlations in the model would be 1.00. What is crucial is not that variables are missing, but that the causal effects of the missing variables are represented by paths connecting the variables which are observed. If paths corresponding to indirect causation are present, then any subset of a recursive model can also be fit by a recursive model. Deleted endogenous variables will not affect the fit of the model if the direct antecedents of the deleted variable directly influence the direct consequents of the deleted variable. If this conditiOn is met, the total effects of the antecedent variables on the consequent variables are the same in both models. That is, deletion of an endogenous variable does not lead to a misspecification unless paths are incorrectly deleted in the reduced model. Thus the deletion of an endogenous variable does not imply that 78 Actual model .30 0 .30 .30 a .30 Misspecified model Figure 10. Actual model (d) from Figure 5 and an accompanying misspecified model. 79 Table 14 Residuals of the Misspecified Model of Figure 10 E 01 02 04 05 E .00 .00 .00 .03 .01 01 .00 .00 .00 .09 .03 n2 .00 .00 .00 .00 .00 n4 .03 .09 .00 .00 .03 .01 .03 .00 .03 .00 80 the disturbance terms of any of the variables remaining in the model should be correlated. A test of the assumptions: Deletion of an exggenous variable. If a single variable is the only consequent of a block of variables and this variable is antecedent to another block of variables, then the deletion of the first block of variables implies that the specified model will fit the data perfectly. This general situation is illus- trated in Figure 11. The variables in the deleted block of variables do not directly influence the correlations of the variables in the con- sequent block of variables. Thus, the result of deleting these varia- bles is similar to the result from deleting an endogenous variable. The deletion of the variable(s) does not affect the fit of the model and so does not lead to correlated disturbance terms between any of the remaining variables in the model. The deletion of an antecedent variable or block of antecedent variables becomes salient, however, if the deleted variable(s) directly influence variables within the consequent block of variables. For example, the misspecified model presented in Figure 12 which is defined by the deletion of E does not fit the data generated by the correct model because E directly influences n2 and n as well as n1. The 3 misspecified model could be "fixed" by adding a path from n1 to n3, but this respecification amounts to replacing the spurious effect of E on and n by a direct effect. However, even though the respecified n1 3 model is conceptually incorrect, it fits the data perfectly, so there are no correlated disturbance terms in the respecified model. 81 ANTECEDENT BLOCK OF VARIABLES 40+ CONSEQUENT IBIKNSK CM? VARIABLES Figure 11. Actual model @+ CONSEQUENT BLOCK OF VARIABLES Specified model Deletion of a block of exogenous variables which affects only a single endogenous variable. Misspecified model Figure 12. Actual model Respecified model Actual model (d) from Figure 5 and misspecified and respecified versions of this model. 82 Correlated Errors in Practice: The Use of Derivatives to Respecify a Model Given a maximum likelihood estimation procedure such as that used by LISREL, a test for locating a potential misspecification for a solution which has converged is based on the values of the first partial derivatives of the likelihood function with respect to each of the fixed parameters. If the model fits the data poorly, some of the derivatives of the parameters fixed at constant values should differ from zero. S5rbom (1975) recommends that a misspecified model may be respecified by using these derivatives. He advises that "we should relax the . . . restriction for the element which gives the largest decrease in F(E)" (p. 143). That element is the element with the largest partial derivative. Consider the actual and misspecified model in Figure 6. Since 832 was falsely fixed at zero in the misspecified model, the derivative with respect to 832 should not be zero in a LISREL solution of the misspecified model given the data generated by the actual model. However there were three nonzero first derivatives in the LISREL analysis of the path model in Figure 6 and these derivatives appeared in both the B and 0 matrices. The corresponding derivatives were also nonzero in the LISREL simultaneous analysis of the model in Figure 6 in which each of the latent variables was measured with three indicators with factor loadings of .80, .60, and .40. These derivatives appear in Table 15. In both the two-step and simultaneous LISREL solutions, the largest derivative was for o(E2,C3). In the simultaneous analysis the value of this derivative was almost twice the size of the next largest derivative over all the parameters, which is the derivative for 832. 83 Table 15 Derivatives for the Model in Figure 6 Computed by LISREL —— 6.-.... 0(E2,E3) ‘-53 -.34 B32 ‘-46 -.18 8 ‘-40 -.15 23 84 If the model were respecified by relaxing the constraint which yielded the largest derivative, the fit of the respecified model would be improved, but the respecified model would not be the model which generated the data. Thus for this model, SSrbom's (1975) advice is false. Moreover, it was the deletion of a path--not the deletion of a variable--that led to the generation of correlated disturbances. The rules for models with correlated disturbances are just the rules for models with missing paths. The equivalence between correlated disturbances and a misspecifi- cation defined by a missing path is even clearer in the following example. The actual and misspecified models in Figure 13 are the same models presented in Figure 6 except that the three disturbance covar- iances are ppp_constrained at zero in the misspecified model. These disturbance terms are included in Figure 13 for heuristic purposes only. The parameter estimates of the misspecified model are computed in Appendix B. The estimated regression parameters of the two misspecified models are equal. Of particular interest, however, is the comparison between the single nonzero OLS residual and the single nonzero disturbance covariance: they are equal. That is, The correlated disturbance term in the nonrecursive misspecified model is simply the lack of fit in the corresponding recursive misspecified model. The use of correlated disturbances provides no more information than was available from the information provided by the traditional OLS solution. If the disturbance covariances are fixed at zero as they are for the OLS solution, the model does not fit. If the disturbance 85 Actual model 8 Misspecified model Figure 13. Actual model (a) from Figure 5 and a third misspecified model with correlated disturbance terms. 86 covariances are free parameters whose values are to be computed from the data, then the model fits the data perfectly. But to say that this model with free disturbance covariances fits the data is to say nothing at all given that the OLS solution does not fit the data. Correlated Errors: Summary There are situations such as the analysis of longitudinal data (Hunter, Coggin, & Gerbing, Note 9) in which an analysis with corre- lated errors is appropriate. But the results of this paper suggest that there are at least two situations in which the use of correlated errors is inappropriate or at best misleading: (a) correlated measure— ment errors as a substitute for decomposing ad-hoc composites, and (b) correlated disturbances as a substitute for addition of a path. To propose a model with correlated errors in either of these two circumstances is to do nothing more than propose a misspecified non- recursive, unidimensional model. To claim that "OLS is biased if there are correlated disturbances in the presence of a deleted path" is to say nothing more than that the parameter estimates from OLS are wrong because the model is false. If LISREL with correlated disturbances is used instead of OLS to estimate the parameters, the model is still false. The parameter estimates are different but equally misleading. Although the apparent fit of the respecified model with correlated errors may be dramatically improved, the model is still wrong. Most of the simulation studies which examine the properties of the statistical procedures for simultaneous equation models begin with a model or models which contain correlated errors, e.g., Hanushek and Jackson (1977) or Cragg (1968). Hanushek and Jackson (1977) concluded that "there is a noticeable increase in the bias in the ordinary least 87 squares estimator as the correlation between the error terms in the two equations increases" (p. 237), which is only to say that as the model becomes more and more false, the OLS estimators become less and less useful. 0r consider the studies by Costner and Schoenberg (1973), SBrbom (1975), or Saris et al. (1978) in which they attempted to deter- mine useful indices for locating misspecification. These authors, with but the one exception in Figure 7, used a given model with correlated errors to generate the data and then attempted to fit a recursive, unidimensional model to this data. Under these conditions it is not surprising that all of these authors have failed to realize that the residuals may contain some very useful information for detecting misspecification. Contemporary use of correlated errors and correlated disturbances has led to erroneous models in most cases. Perhaps the fact that LISREL permits correlated errors should be regarded as a flaw in the program instead of an advantage. APPENDICES APPENDIX A COMMUNALITIES IN A MULTIPLE GROUPS ANALYSIS Unlike most contemporary authors, Kenny (1979) recognizes that the classical multiple groups analysis is a confirmatory factor analysis. However he completely abandons multiple groups in favor of LISREL. A "factor analytic solution to [a multiple indicator model] . . . is the multiple group solution . . . which is rarely used. Like most factor analytic solutions, it . . . suffers from the communality problem; that is, communalities must be specified in advance" (p. 138). This is a serious misunderstanding of the factor analysis literature since communalities, like any other parameter, must be estimated. This is true of LISREL as well. The LISREL estimate of the communality of 2 the ith observed variable is A1. Nunnally (1978) is one of the strongest contemporary advocates of the classical multiple groups approach, but he argues that "the diagonal unities in a centroid factor multiple groups analysis can make the loadings seem spuriously high . . . . In a sense, there is nothing wrong with this, because it is the correct mathematical solution. The illusory appearance of large loadings could be reduced by . . . the use of SMCs as communalities . . . . However, this is not really necessary" (p. 420). Nunnally (1978) not only fails to recognize the possibility of iterating for the communality values, he 88 89 does not recognize the need. However, Hunter (Note 7) has shown that except in certain unusual cases, the communalities in a multiple groups analysis will converge to their correct values on successive iterations. The convergence for each indicator of the measurement model used in this study is graphed in Figure 14. The 0th iteration for each indicator is the initial computation obtained before the first iteration, i.e., the value Nunnally (1978) recommends. If "the correct mathematical solution" implies that the underlying structure is recovered with perfect accuracy in the absence of sampling or specification error, and if the indicators are measured with error, then communalities must be inserted into the diagonal of the correlation matrix. 90 1.000- I .80 Cb \ 2 H O C a: 60 CD _J c: I CD I-— L) a: .40 U. 0200'" 0 4.000 8.000 121000 ' 16:000 NUMBER OF ITERRTIONS Figure 14. An example of MGRP iterations. APPENDIX B ANALYTIC SOLUTION FOR THE PARAMETER ESTIMATES OF MISSPECIFIED MODELS The parameter estimates and consequently the residuals of the misspecified causal models can be solved analytically by the following algorithm developed for this paper: (1) Generate the predicted correlations of the correctly specified model, 2. (2) Generate the predicted correlations of the misspecified model, 2. (3) Solve for the parameters of the misspecified model in terms of E. (4) Express the solution of the parameters of the misspecified model in terms of the parameters of the correctly specified model. This procedure is illustrated for the misspecified models which appear in Figure 6 and Figure 13 and the corresponding actual model, which is the same in both figures. For the models in Figure 6, r(E.nl) = Y £(a.n1) = I 91 92 r(n1,n2) = 821 r(nl’nz) = 821 r(nl.n3) = 831 + 821832 r(nl,n3) = 831 r("2"”3) = 832 + 831821 r("2’”3) = 31821 Given these correlations, the OLS parameter estimates and the residuals of the misspecified model may be computed. The OLS estimates of the parameters of the misspecified model are A Y = r(nl.E) 821 = r(02.nl) 831 = r(n3.nl) which can be expressed in terms of the parameters of the actual model. A Y = r(n19g) = Y 831 = r("3’“1) = 831 + 821832 The OLS residuals of the misspecified model are defined by S - 8. Since theilmatrix is the observed matrix for these examples, the residuals are defined by Z - 2. I .4 I 4) ll .4 l .4 II C r(€,nl) - £(53n1) — r(E.02) - £(a.n2) = 1321 — 1821 = 1321 - 1321 = o I 4 'm r(E.n3) - £(E.n3) - YB B B 0 + .4 'm u: h) 'm h) [—0 I 31 ’ Y 21 32 = 831 + 821832 " B31 ’ 821832 = 0 r(02.n3) - r(nz.n3) = 832 + B31821 - 831821 2 ’ 832 + 831321 ‘ B31521 ’ 821832 I ‘m 93 Consider the misspecified model in Figure 13 which is identical to the misspecified model in Figure 6 except that the disturbance covar- iances are unconstrained. The covariance structure of the misspecified model is presented below. r(E.nl) = i A r(nl.n3> = 331 + 0(cl.c3) r(n2.n3) = 831821 + 8318(61.62) + 8210(61.c3) + 0(E2:E3) The parameters of the misspecified model appear to be just identified since there are six equations in six unknowns, Y, 821, 831, 6(E1,C2), 0(cl,c3), and 6(c2,c3). (The disturbance variances are functions of these six parameters.) Since there are as many equations as there are unknowns, the parameters of the misspecified model can be expressed in terms of the predicted correlations. Moreover, a just identified model fits any data set perfectly since there are no overidentifying restrictions or degrees of freedom to test the fit of the model. So each predicted correlation r equals the corresponding actual correlation r Since ij ij ' the actual correlations can be expressed in terms of the parameters of the actual model, solving for the parameters of the misspecified model in terms of the correlations implied by the misspecified model is equivalent to expressing the parameters of the misspecified model in terms of the parameters of the actual model for a just identified model. That is, the FIML solution computed by LISREL can be obtained 94 heuristically for the just identified model. These computations for this example are presented below. Y = E(§.n> = r(E.nl) = Y r(nz.E) r(02.E) B A 21 B =———-—— =—————=—-——=B 21 Y Y Y 21 g = r(n3.E) = r(n3.E) = Y831 + Y832821 = B + B B 31 y Y Y 31 32 21 A = r(nlinz) - 821 = B - B 21 21 = 0 = (831832821) (8 8+31 832 821) = 0 = (832+ 831821) (8 31+ 832 821)(821) 2 ’ 1332'l ’ 821) APPENDIX C COMPUTATIONAL DETAILS OF A SIMULTANEOUS LISREL ANALYSIS The Relation Between the Indicator Residuals and Factor Residuals Consider the model in Figure 6 such that each latent variable has three indicators with factor loadings .80, .60, and .40, and Y = .30, 821 = 831 = .35, and 832 = .40. Also consider the residuals from the simultaneous LISREL analysis between the indicators of n2 and 03 which appear in Table 4. Since the residuals are defined as the difference of the corresponding observed and predicted correlations, these residuals may be computed by first computing the observed and predicted correlations. The observed indicator correlations were generated by the product rule for external consistency. For example, since the actual correla- tion between NZ and n3 reported in Table 2 is .523, (.8) (.523) (.8) = .334 LISREL computed the corresponding predicted correlations according to the same product rule. The difference is that the estimated values of the factor correlations were substituted for the actual values. That is, r(3'21’3’31) = A21r("2’”3) A31 95 96 Since the factor loadings were correctly recovered by LISREL, A A21 = A31 = .80 The LISREL estimate of r(n2,n3), reported in Table 2, was .234. £(y21.y31> = (.8) (.234) (.8) = .150 Thus the residual of the indicators y21 and y31 is Res(y21’Y31) E r(Y219Y31) — r(y21!Y3l) .334 - .150 = .184 So, And .184 is the same value computed by LISREL as listed in Table 4. The keys to reconstructing the residuals between indicator corre- lations are the residuals among the factor correlations and the respective factor loadings. This can be more explicitly stated by first recomputing the residual between y21 and y31. RES(y21,y31) E r(Y213y31) - r(y21’y31) = A21"”2’”3)A31 ' A21"“2’”3)A31 [Res(n2,n3)] (A21A31) That is, Res(y21,y3l) = (.523 - .234) (.8) (.8) II (.289) (.8) (.8) = .184 This result can be generalized to any Factors F and G with indicators 1 and j respectively, Res (yFi’yGi) = Res (F,G) AFi AGJ. 97 The entire block of residuals for the indicators of factors F and G can be described as _ I (‘11 rFG) AFAG where;F is the vector of factor loadings for the NF indicators of F and A is the vector of factor loadings for the N factor loadings of -G G G. The resulting "block" of residuals is a matrix of order NF x NG' For example, let n23; F, n3 3 G, A% = Aé E [,8 .6 .4]. Then I- 1 p .18 .14 .09 .64 .43 .32.1 .14 .10 .07 (.289) .48 .36 .24 .04 .07 .05 .32 .24 .16 L J - L a For this example in which each factor in the model had the same number of indicators with an identical pattern of factor loadings, the entire residual matrix of indicator correlations can be expressed as a Kronecker product of (a) the entire residual matrix of the factor correlations and (b) the factor loadings. That is, (BIND " 2311111)) = (ZFAC ' iF.II.(:)® (AA') 12x12 4x4 3x3 The Kronecker product, designated by(3> is defined as the matrix formed by justaposing the scalar products of every element of the first matrix with the entire second matrix. That is, every element of (ZFAC - EFAC) is "replaced" by the multiple of that element with the matrix (AAf). Thus the complete 12 x 12 matrix of residuals defined by Z - E where 2 is computed in a LISREL simultaneous analysis can be expressed as a function of the actual and estimated factor loadings and factor correlations. The actual factor loadings and factor correlations are given and the estimated factor loadings are approximately equal to the 98 actual factor loadings. The problem of reconstructing the residuals of the indicator correlations has been reduced to reconstructing the residuals of the factor correlations. Computation of the Parameter Estimates of the Causal Model LISREL selects those parameter estimates which minimize a function F(E) of the residuals among the indicator correlations. But minimizing the residuals among the indicator correlations is just minimizing the residuals among the factor correlations. Since LISREL is a full infor- mation technique, the parameter estimates are computed to simultaneously minimize the factor correlation residuals across the entire model. Given the present example in which the misspecification is defined by the deletion of the path from n2 to n3, there are two conflicting "pressures" which must be simultaneously resolved. First, the only source of the correlation between and in the misspecified model T12 03 is the common antecedent n1, i.e., but the actual correlation between n2 and n3 is decomposed as r(nz.n3) = 832 + 821831 To the extent 832 is large, B and 832 must be biased upward to 31 minimize Res(n2,n3). That is, LISREL "would like" to increase the and 8 until values of B 32 31 A But, as a second consideration, LISREL cannot adjust both 831 and 832 without affecting the residuals Res(nl,n2) and Res(nl,n3). To the 99 extent that 8>B and 8 31 31 > B 32 32 the values of Res(n1,n2) and Res(n ) will increase. 2’"3 The resulting parameter values computed by LISREL are a compro- mise. The values of 831 and 832 are increased over their OLS counter- parts, which reduces Res(n2,n3) for the LISREL solution compared to the A corresponding OLS residual. However, this increase in B and 832 31 increases Res(nl,n3) and Res(n2,n3) over their OLS counterparts, which are equal to zero. Thus the simultaneous LISREL solution provided incorrect factor correlations and obscured the detection of the misspecification in terms of the residuals. APPENDIX D RELATIVE SIZES OF THE INDICATOR RESIDUALS WITHIN AND BETWEEN CONSTRUCTS IN AN AD-HOC COMPOSITE IF ALL THE INDICATORS HAVE EQUAL FACTOR LOADINGS The examination of the residuals from a misspecified ad-hoc composite should reveal a pattern in which the largest residuals were between the indicators of the same construct. The residuals of the indicators of different constructs may be negative, but they should at least tend to be smaller than the residuals of indicators of the same nconstruct. For example, let yi and yj be indicators of nF and yk be an indicator of nG. Let all of the actual factor loadings be equal. If yi, yj, and yk are pooled together and misspecified as a construct n, then Resij - Res1k = (rij - rij) - (rik - rik) = rij ' r13 ' r1k + rik = riFer ' rifirjfi'- rirrrcrck + rifirkfi If riF = er = rkG’ R s - Res = r2 - r r - r2 r + r r 8 ij ik 12 ffi jfi iF FG ffi kfi' 2 2 r1F ' riFrFG So if rFG < 1 and if the actual factor loadings are equal, then the residual for y1 and yj, indicators of the same construct, will always be larger than the residual for yi and yk, indicators of different constructs. 100 LIST OF REFERENCE NOTES LIST OF REFERENCE NOTES Hunter, J. E., & Cohen, S. H. Mathematical models of attitude change in the passive communication context. Unpublished book, 1973. Hunter, J. E., Nicol, T. S., & Gerbing, D. W. Calculus and mathematical model building in the social sciences. Book submitted for publication, 1979. Hunter, J. E. Cluster analysis: Reliability, construct validity, and the multiple indicators approach to measurement. Paper delivered at a meeting of the U.S. Civil Service Commission Panel on Advanced Statistics, Washington, D.C., 1977. Hunter, J. E., & Gerbing, D. W. Unidimensional measurement and confirmatory factor analysis (Occasional Paper No. 20). East Lansing, MI: The Institute for Research on Teaching, College of Education, Michigan State University, May 1979. Hunter, J. E., Gerbing, D. W., & Boster, F. J. Machiavellian beliefs and personalipy: The construct invalidity of the Machiavellian dimension. Paper submitted for publication, 1979. Hunter, J. E., & Cohen, S. H. PACKAGE: User's manual. East Lansing, MI: Computer Institute for Social Science Research, Michigan State University, 1970. Hunter, J. E. Personal communication, 1978. Gruvaeus, G. T., & JSreskog, K. G. A computer program for minimizing a function of several variables (Research Bulletin 70-14). Princeton, NJ: Educational Testing Service, February 1970. Hunter, J. E., Coggin, T. D., & Gerbing, D. W. Random and specific error in panel data: A multiple indicator measurement model. Paper submitted for publication, 1979. 101 LIST OF REFERENCES LIST OF REFERENCES Asher, H. B. Causal modeling. Sage University Paper series on Quantitative Applications in the Social Sciences, No. 07-003. Beverly Hills & London: Sage Publications, 1976. Blalock, H. M., Jr. Causal inferences in nonexperimental research. Chapel Hills, NC: University of North Carolina Press, 1964. Bock, R. D., & Bargmann, R. E. Analysis of covariance structures. Psychometrika, 1966, 31, 507-534. Burt, C. Experimental tests of general intelligence. British Journal of Psychology, 1909, 3, 94-177. Burt, C. Factors of the mind. London: University of London Press, Ltd., 1940. Burt, R. S. Confirmatory factor-analytic structures and the theory construction process. Sociological Methods and Research, 1973, 131-189. Burt, R. S. Interpretational confounding of unobserved variables in structural equation models. Sociological Methods and Research, 1976, 5, 3-52. Christie, R., & Geis, F. L. Studies in Machiavellianism. New York: Academic Press, 1970. Costner, H. L., & Schoenberg, R. Diagnosing indicator ills in multiple indicator models. In A. S. Goldberger & O. D. Duncan (Eds.), Structural equation models in the social sciences. New York: Seminar Press, Inc., 1973. Cragg, J. G. Some effects of incorrect specification on the small- sample properties of several simultaneous-equation estimators. International Economic Review, 1968, 2, 63-86. Cronbach, L. J. Validation of educational measures. In Proceedings of the 1969 invitational conference on testing problems. Princeton: Educational Testing Service, 1970. Duncan, 0. D. Path analysis: Sociological examples. American Journal of Sociology, 1966, 12, 1-16. 102 103 Duncan, 0. D. Introduction to structural equation models. New York: Academic Press, Inc., 1975. Fletcher, R., & Powell, M. J. D. A rapidly convergent descent method for minimization. The Computer Journal, 1963,_§, 163-168. Green, S. B., Lissitz, R. W., & Mulaik, S. A. Limitations of coefficient alpha as an index of test unidimensionality. Educational and Psychological Measurement, 1977,_3Z, 827-838. Hanushek, E. A., & Jackson, J. E. Statistical methods for social scientists. New York: Academic Press, Inc., 1977. Heise, D. R. Causal analysis. New York: John Wiley & Sons, 1975. Holzinger, K. J. A simple method of factor analysis. Psychometrika, 1944, 3, 257—262. Hunter, J. E. Methods of reordering the correlation matrix to facilitate visual inspection and preliminary cluster analysis. Journal of Educational Measurement, 1973, 39, 51-61. Hunter, J. E., & Cohen, S. H. PACKAGE: A system of computer routines for the analysis of correlational data. Education and Psychological Measurement, 1969,_33, 697-700. Hunter, J. E., Hunter, R. F., & Lopis, J. E. A causal analysis of attitudes toward leadership training in a classroom setting. Human Relations, in press. Jareskog, K. G. Testing a simple structure hypothesis in factor analysis. Psychometrika, 1966, 33, 165-190. Jareskog, K. G. Some contributions to maximum likelihood factor analysis. Psychometrika, 1967,‘33, 443-482. J5reskog, K. G. A general approach to confirmatory maximum likelihood factor analysis. Psychometrika, 1969, 33, 183-202. Jareskog, K. G. A general method for analysis of covariance structures. Biometrika, 1970, 31, 239-251. Jareskog, K. G. Statistical analysis of sets of congeneric tests. Psychometrika, 1971, 3Q, 109-133. J5reskog, K. G. Structural analysis of covariance and correlation matrices. Psychometrika, 1978, 33, 443-477. J5reskog, K. G., & Sarbom, D. LISREL IV: Analysis of linear structural relationships by the method of maximum likelihood. Chicago: National Educational Resources, Inc., 1978. Kenny, D. A. Correlation and causation. New York: John Wiley & Sons, 1979. 104 Kohn, M. L., & Schooler, C. The reciprocal effects of the substantive complexity of work and intellectual flexibility: A longitudinal assessment. American Journal of Sociology, 1978, s3, 25-52. Land, K. C. Identification, parameter estimation, and hypothesis testing in recursive sociological models. In A. S. Goldberger and O. D. Duncan (Eds.), Structural equation models in the social sciences. New York: Seminar Press, Inc., 1973. Lawley, D. N. The application of the maximum likelihood method to factor analysis. British Journal of Psychology, 1943, 33, 172-175. Lewis-Beck, M. S. Determining the importance of an independent variable: A path analytic solution. Social Science Research, 1974,_3, 95-107. Long, J. S. Estimation and hypothesis testing in linear models containing measurement error. Sociological Methods and Research, 1976,_3, 157-206. Magnusson, D. Test theory. Reading, MA: Addison-Wesley Publishing Co., 1967. Mulaik, S. A. The foundations of factor analysis. New York: McGraw- Hill Book Co., 1972. Namboodiri, N. K., Carter, L. F., & Blalock, H. M., Jr. Applied multivariate analysis and experimental designs. New York: McGraw—Hill Book Co., 1975. Nunnally, J. Psychometric methods. New York: McGraw—Hill Book Co., 1978. Saris, W. E., de Pijper, W. M., & Zegwaart, P. Detection of specification errors in linear structural equation models. In K. E. Schuesster (Ed.), Sociological methodology. San Francisco: Jossey-Bass, 1978. Simon, H. A. Models of man. New York: John Wiley & Sons, 1957. SBrbom, D. Detection of correlated errors in longitudinal data. British Journal of Mathematical and Statistical Psychology, 1975, 33, 138-151. Spearman, C. "General intelligence" objectively determined and measured. American Journal of Psychology, 1904, 33, 201-293. Spearman, C. Demonstration of formulae for true measurement of correlation. American Journal of Psychology, 1907, 33, 161-169. Spearman, C. The theory of two factors. Psychological Review, 1914, 33, 101-115. 105 Thurstone, L. L. Multiple factor analysis. Psysholog3pal Review, 1931, 33, 406-427. Tryon, R. C. Cluster ana3ysis. Ann Arbor, MI: Edwards Brothers, Inc., 1939. Tryon, R. C., & Bailey, D. E. Cluster analysis. New York: McGraw— Hill Book Co., 1970. Tucker, L. R., & Lewis, G. A reliability coefficient for maximum likelihood factor analysis. Psychometrika, 1973, 33, 1-10. Wiley, D. E. The identification problem for structural equation models with unmeasured variables. In A. S. Goldberger and O. D. Duncan (Eds.), Structural equation models in the social sciences. New York: Academic Press, Inc., 1973. Wishart, J. The generalized product moment distribution in samples from a normal multivariate population. Biometrika, 1928, 20A, 32-52. ““ Wright, S. Correlation and causation. Journal of Agricultural Research, 1921, 39, 557-585. ..:I:I;:x;“‘;- m .:I~m-' ' ‘4-éuh . punt... . , . ,