Measurement invariance or measurement equivalence is a statistical property of measurement that indicates that the same construct is being measured across some specified groups.[1] For example, measurement invariance can be used to study whether a given measure is interpreted in a conceptually similar manner by respondents representing different genders or cultural backgrounds. Violations of measurement invariance may preclude meaningful interpretation of measurement data. Tests of measurement invariance are increasingly used in fields such as psychology to supplement evaluation of measurement quality rooted in classical test theory.[1] Measurement invariance is often tested in the framework of multiple-group confirmatory factor analysis (CFA).[2] In the context of structural equation models, including CFA, measurement invariance is often termed factorial invariance.[3]
In the common factor model, measurement invariance may be defined as the following equality:
where [math]\displaystyle{ f(\cdot) }[/math] is a distribution function, [math]\displaystyle{ \textit{Y} }[/math] is an observed score, [math]\displaystyle{ \boldsymbol{\eta} }[/math] is a factor score, and s denotes group membership (e.g., Caucasian=0, African American=1). Therefore, measurement invariance entails that given a subject's factor score, his or her observed score is not dependent on his or her group membership.[4]
Several different types of measurement invariance can be distinguished in the common factor model for continuous outcomes:[5]
The same typology can be generalized to the discrete outcomes case:
Each of these conditions corresponds to a multiple-group confirmatory factor model with specific constraints. The tenability of each model can be tested statistically by using a likelihood ratio test or other indices of fit. Meaningful comparisons between groups usually require that all four conditions are met, which is known as strict measurement invariance. However, strict measurement invariance rarely holds in applied context.[6] Usually, this is tested by sequentially introducing additional constraints starting from the equal form condition and eventually proceeding to the equal residuals condition if the fit of the model does not deteriorate in the meantime.
Although further research is necessary on the application of various invariance tests and their respective criteria across diverse testing conditions, two approaches are common among applied researchers. For each model being compared (e.g., Equal form, Equal Intercepts) a χ2 fit statistic is iteratively estimated from the minimization of the difference between the model implied mean and covariance matrices and the observed mean and covariance matrices.[7] As long as the models under comparison are nested, the difference between the χ2 values and their respective degrees of freedom of any two CFA models of varying levels of invariance follows a χ2 distribution (diff χ2) and as such, can be inspected for significance as an indication of whether increasingly restrictive models produce appreciable changes in model-data fit.[7] However, there is some evidence the diff χ2 is sensitive to factors unrelated to changes in invariance targeted constraints (e.g., sample size).[8] Consequently it is recommended that researchers also use the difference between the comparative fit index (ΔCFI) of two models specified to investigate measurement invariance. When the difference between the CFIs of two models of varying levels of measurement invariance (e.g., equal forms versus equal loadings) is below −0.01 (that is, it drops by more than 0.01), then invariance in likely untenable.[8] The CFI values being subtracted are expected to come from nested models as in the case of diff χ2 testing;[9] however, it seems that applied researchers rarely take this into consideration when applying the CFI test.[10]
Equivalence can also be categorized according to three hierarchical levels of measurement equivalence.[11][12]
Tests of measurement invariance are available in the R programming language.[13][14]
The well-known political scientist Christian Welzel and his colleagues criticize the excessive reliance on invariance tests as criteria for the validity of cultural and psychological constructs in cross-cultural statistics. They have demonstrated that the invariance criteria favor constructs with low between-group variance, while constructs with high between-group variance fail these tests. A high between-group variance is indeed necessary for a construct to be useful in cross-cultural comparisons. The between-group variance is highest if some group means are near the extreme ends of the closed-ended scales, where the intra-group variance is necessarily low. Low intra-group variance yields low correlations and low factor loadings which scholars routinely interpret as an indication of inconsistency. Welzel and colleagues recommend instead to rely on nomological criteria of construct validity based on whether the construct correlates in expected ways with other measures of between-group differences. They offer several examples of cultural constructs that have high explanatory power and predictive power in cross-cultural comparisons, yet fail the tests for invariance.[15][16] Proponents of invariance testing counter-argue that the reliance on nomological linkage ignores that such external validation hinges on the assumption of comparability.[17]
Original source: https://en.wikipedia.org/wiki/Measurement invariance.
Read more |