1,899
20
Essay, 14 pages (3500 words)

Editorial: measurement invariance

Multi-item surveys are frequently used to study scores on latent factors, like human values, attitudes, and behavior. Such studies often include a comparison, between specific groups of individuals or residents of different countries, either at one or multiple points in time (i. e., a cross-sectional or a longitudinal comparison or both). If latent factor means are to be meaningfully compared, the measurement structures of the latent factor and their survey items should be stable, that is “ invariant.” As proposed by Mellenbergh (1989), “ measurement invariance” (MI) requires that the association between the items (or test scores) and the latent factors (or latent traits) of individuals should not depend on group membership or measurement occasion (i. e., time). In other words, if item scores are (approximately) multivariate normally distributed, conditional on the latent factor scores, the expected values, the covariances between items, and the unexplained variance unrelated to the latent factors should be equal across groups.

Many studies examining MI of survey scales have shown that the MI assumption is very hard to meet. In particular, strict forms of MI rarely hold. With “ strict” we refer to a situation in which measurement parameters are exactly the same across groups or measurement occasions, that is an enforcement of zero tolerance with respect to deviations between groups or measurement occasions. Often, researchers just ignore MI issues and compare latent factor means across groups or measurement occasions even though the psychometric basis for such a practice does not hold. However, when a strict form of MI is not established and one must conclude that respondents attach different meanings to survey items, this makes it impossible to make valid comparisons between latent factor means. As such, the potential bias caused by measurement non-invariance obstructs the comparison of latent factor means (if strict MI does not hold) or regression coefficients (if less strict forms of MI do not hold).

Traditionally, MI is tested for in a multiple group confirmatory factor analysis (MGCFA) with groups defined by unordered categorical (i. e., nominal) between-subject variables. In MGCFA, MI is tested at each constraint of the latent factor model using a series of nested (latent) factor models. This traditional way of testing for MI originated with Jöreskog (1971), who was the first scholar to thoroughly discuss the invariance of latent factor (or measurement) structures. Additionally, Sörbom (1974 , 1978 ) pioneered the specification and estimation of latent factor means using a multi-group SEM approach in LISREL ( Jöreskog and Sörbom, 1996 ). Following these contributions the multi-group specification of latent factor structures has become widespread in all major SEM software programs (e. g., AMOS Arbuckle, 2006 , EQS Bender and Wu, 1995 , LAVAAN Rosseel, 2012 , Mplus Muthén and Muthén, 2013 , STATA STATA, 2015 , and OpenMx Boker et al., 2011 ). Shortly thereafter, Byrne et al. (1989) introduced the distinction between full and partial MI. Although their introduction was of great value, the first formal treatment of different forms of MI and their consequences for the validity of multi-group/multi-time comparisons is attributable to Meredith (1993). So far, a tremendous amount of papers dealing with MI have been published. The literature on MI published in the 20th century is nicely summarized by Vandenberg and Lance (2000). Noteworthy is also the overview of applications in cross-cultural studies provided by Davidov et al. (2014), as well as a recent book by Millsap (2011) containing a general systematic treatment of the topic of MI. The traditional MGCFA approach to MI-testing is described by, for example, Byrne (2004), Chen et al. (2005), Gregorich (2006), van de Schoot et al. (2012), Vandenberg (2002) and Wicherts and Dolan (2010). Researchers entering the field of MI are recommended to first consult Meredith (1993) and Millsap (2011) before reading other valuable academic works.

Recent developments in statistics have provided new analytical tools for assessing MI. The aim of this special issue is to provide a forum for a discussion of MI, covering some crucial “ themes”: (1) ways to assess and deal with measurement non-invariance; (2) Bayesian and IRT methods employing the concept of approximate MI; and (3) new or adjusted approaches for testing MI to fit increasingly complex statistical models and specific characteristics of survey data.

Dealing with Measurement Non-invariance

If the test for MI indicates that strict MI across groups or time is not established, no sound psychometric basis is provided for the comparison of latent factor means. The absence of such psychometric basic is the first topic dealing with measurement non-invariance. A nice example of a situation in which such psychometric basis is absent is provided in the paper by Lommen et al. (2014). These authors show that comparing posttraumatic stress in soldiers before and after war-zone related traumatic events (the wars in Afghanistan or Iraq) is virtually impossible due to instability in thresholds. For a researcher this conclusion may be hard to digest, especially if the success of the study relies entirely on the possibility to make such meaningful comparisons over time. Within the context of their study the authors recommend considering pre- and post-symptom scores as representing separate constructs.

In the same vein, a failure to establish less strict forms of MI may be worrisome if meaningful comparisons of structural relationships between latent factor means are important to the study (e. g., the comparison of the magnitude of a correlation, regression, or path coefficient across groups/time). Hox et al. (2015), show how the non-establishment of less strict forms of MI can (partly) be explained and corrected for. They show that, in the context of mixed-mode surveys, non-invariance can be the effect of selection or measurement differences due to mode (e. g., web survey, telephone survey, face-to-face interview).

Detecting non-invariant items is the next topic dealing with measurement non-invariance. In the contribution of de Roover et al. (2014) a method is proposed based on cluster-wise simultaneous component analysis (SCA). Their method aims at detecting non-invariant items. Barendse et al. (2014) examined a Bayesian restricted (latent) factor analysis (RFA) method for the same purpose, namely detecting items violating the MI assumption. They concluded that Bayesian RFA methods are especially suited for detecting measurement bias.

Our special issue also contains a discussion on the importance of understanding whether the presence of (in)correctly specified factorial invariance parameters influences the assessment of other factor model parameters (e. g., intercepts, error variances, latent factor variances, and latent factor means). In a simulation study, Guenole and Brown (2014) investigated whether ignoring the non-invariant underlying structure of the latent factor leads to substantial regression parameter bias in categorical item factor analyses (CIFA). The authors urge researchers to avoid ignoring sources of non-invariance in CIFA when non-invariance occurs in both loadings and thresholds even if this occurs in only one item.

Approximate Measurement Invariance

A relatively new research avenue in the MI literature deals with the use of Bayesian structural equation models (BSEM) to relax strict forms of MI (see Muthén and Asparouhov, 2012 ). In particular, exact zero constraints on the cross-group differences between all relevant measurement parameters (e. g., factor loadings and item intercepts) are substituted by “ approximate” zero constraints. Instead of forcing item intercepts to be exactly equal across groups, a substantive prior distribution (around zero) is used to bring the parameters closer to one another, while allowing for some “ wiggle room.” If there are many small differences between the groups in terms of intercepts or factor loadings, approximate MI seeks a balance between adherence to the requirements of MI, making comparisons possible, and obtaining a well-fitting model (i. e., a model that is more realistic given the data at hand). When the classical MI tests do not hold given the data, approximate MI represents a promising (and more realistic) alternative; the cross-group differences between all relevant measurement parameters are “ hopefully” close enough to zero to allow making meaningful latent factor mean comparisons.

A tutorial paper introducing the method of approximate MI is presented by van de Schoot et al. (2013). Further, our special issue contains empirical examples comparing the results of Bayesian approximate MI to the results of the more traditional ways of MI-testing as applied to specific questionnaires: e. g., the Portrait Values Questionnaire, using data from the European Social Survey including data on many countries and many time points ( Cieciuch et al., 2014 ; Zercher et al., 2015 ), the Hedonic and Eudaimonic Motives for Activities scale ( Bujacz et al., 2014 ), and the Golombok-Rust Inventory of Marital State ( Chiorri et al., 2014 ).

Furthermore, our special issue contains two extensions of approximate MI to the field of IRT (see also Fox and Verhagen, 2010 ). Instead of using substantive prior distributions as in the Bayesian approximate MI method, the method described by Fox establishes a measurement scale across countries and conceptualizes country-specific non-invariance in item parameters as random deviations through country-specific random item effects. In such conceptualization cross-group comparisons can still be made even in the presence of non-invariant items. Kelcey et al. (2014) developed a method based on Fox’s approximate MI approach which is applicable whenever measurements are nested within raters and cross-classified among, for instance, countries. Another contribution to our special issue by Muthén and Asparouhov (2014) concerns the use of the alignment method (see also Asparouhov and Muthén, 2014 ) in IRT models, a method which is essential when applying approximate MI. This method minimizes a loss function which makes sure that there are a few large non-invariant measurement parameters instead of many smaller non-invariant measurement parameters, an optimal alignment strategy which resembles the rationale underlying rotation of factor solutions in EFA.

Testing for MI in Increasingly Complex Statistical Models

For some complex statistical models, the traditional multi-group (MGCFA) approach to MI-testing has to be adjusted to meet the specific requirements of the data and/or the model. Examples of such adjustments can be found in our special issue. An assumption embedded within many methods to test for MI is that the grouping (i. e., auxiliary) variable is unordered (i. e., nominal). Wang et al. (2014) present a method to test for MI in cases in which the auxiliary variable is ordered or continuous. Verdam and Oort (2014) illustrate MI-testing for Kronecker restricted SEM models, which constitute parsimonious models that provide an alternative to longitudinal latent factor models. Adolf et al. (2014) examine MI in the context of multiple-occasion and multiple-subject time series models. In such models, MI has to be established (a) over time within subjects, (b) over subjects within occasions, and (c) over time and subjects simultaneously. Boom (2014) investigated MI in the context of children’s development of increasingly advanced strategies over time, in for instance the way they deal with mathematical problems (e. g., strategies on how children learn to multiply numbers below 10). The use of different strategies is scored as a variable and development is seen as the movement from one strategy to a more advanced one and Boom shows how MI plays a crucial role when analyzing such data. Jak (2014) uses a multi-level framework and proposes an extension to the SEM framework, moving from models describing two-level data to models describing three-level data. Within this framework MI invariance can be tested across level 2 as well as across level 3 clustering variables.

Another application of MI finds its origin in multi-trait multi-method models (MTMM; Eid and Diener, 2006 ), in which multiple methods (or scales) and raters are used to quantify the set of latent factors under study. Geiser et al. (2014) demonstrate the advantage of moving from an exclusively covariance- or correlation-based MTMM approach to an approach that includes latent factor means. This approach results in more fine-grained information about convergent validity and method effects when testing for MI. Albeit being analyzed differently, a comparable design to the MTMM is the two-way rating design utilized in situations where subjects have to judge to what extent a particular scale or variable pertains to a particular concept or situation. Kroonenberg (2014) presents an approach applicable to the assessment of MI in two-way rating designs. In his approach, a hierarchy of models is proposed, each one conceptualizing a form of MI, varying in terms of strictness.

Conclusion

Our special issue contains numerous simulation studies aiming at demonstrating the possibilities and limitations of different analytical tools to test for various forms of MI; tutorial papers providing the hands-on support needed when using the recent developed analytical tools to test for MI, as well as illustrations of how the analytical tools may be meaningfully applied in different fields of research when addressing issues related to MI across groups or time.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The first author was supported by a grant from The Netherlands Organization for Scientific Research: NWO-VENI-451-11-008. The second author was supported by the basic research program of the International Laboratory for Socio-Cultural Research at HSE Moscow.

References

Adolf, J., Schuurman, N., Borkenau, P., Borsboom, D., and Dolan, V. (2014). Measurement invariance within and between subjects: a distinct problem in testing the equivalence of intra- and inter-individual model structures. Front. Psychol . 5: 883. doi: 10. 3389/fpsyg. 2014. 00883

||

Arbuckle, J. L. (2006). Amos (Version 7. 0) [Computer Program]. Chicago, IL: SPSS.

Asparouhov, T., and Muthén, B. (2014). Multiple-group factor analysis alignment. Struct. Equ. Model . 21, 1–14. doi: 10. 1080/10705511. 2014. 919210

||

Barendse, M. T., Albers, C. J., Oort, F. J., and Timmerman, M. E. (2014). Measurement bias detection through Bayesian factor analysis. Front. Psychol . 5: 1087. doi: 10. 3389/fpsyg. 2014. 01087

||

Bender, P. M., and Wu, E. J. C. (1995). EQS for Windows User’s Guide . Encino, CA: Multivariate Software.

Boker, S., Neale, M., Maes, H., Wilde, M., Spiegel, M., Brick, T., et al. (2011). OpenMx: an open source extended structural equation modeling framework. Psychometrika 76, 306–317. doi: 10. 1007/s11336-010-9200-6

||

Boom, J. (2014). A new visualization and conceptualization of categorical longitudinal development: measurement invariance and change. Front. Psychol . 6: 289. doi: 10. 3389/fpsyg. 2015. 00289

||

Bujacz, A., Vittersø, J., Huta, V., and Kaozmarek, L. D. (2014). Measuring hedonia and eudaimonia as motives for activities: cross-national investigation through traditional and Bayesian structural equation modeling. Front. Psychol . 5: 984. doi: 10. 3389/fpsyg. 2014. 00984

||

Byrne, B. M. (2004). Testing for multigroup invariance using AMOS graphics: a road less traveled. Struct. Equ. Model . 11, 272–300. doi: 10. 1207/s15328007sem1102_8

|

Byrne, B. M., Shavelson, R. J., and Muthén, B. (1989). Testing for the equivalence of factor covariance and mean structures: the issue of partial measurement invariance. Psychol. Bull . 105, 456–466. doi: 10. 1037/0033-2909. 105. 3. 456

|

Chen, F. F., Sousa, K. H., and West, S. G. (2005). Teacher’s corner: testing measurement invariance of second-order factor models. Struct. Equ. Model . 12, 471–492. doi: 10. 1207/s15328007sem1203_7

|

Chiorri, C., Day, T., and Malmberg, L.-E. (2014). An approximate measurement invariance approach to within-couple relationship quality. Front. Psychol . 5: 983. doi: 10. 3389/fpsyg. 2014. 00983

||

Cieciuch, J., Davidov, E., Schmidt, P., Algesheimer, R., and Schwartz, S. H. (2014). Comparing results of an exact versus an approximate (Bayesian) measurement invariance test: a cross-country illustration with a scale to measure 19 human values. Front. Psychol . 5: 982. doi: 10. 3389/fpsyg. 2014. 00982

||

Davidov, E., Meuleman, B., Cieciuch, J., Schmidt, P., and Billiet, J. (2014). Measurement equivalence in cross-national research. Annu. Rev. Sociol . 40, 55–75. doi: 10. 1146/annurev-soc-071913-043137

|

de Roover, K., Timmerman, M. E., DeLeersnyder, J., Mesquita, B., and Ceulemans, E. (2014). What’s hampering measurement invariance: detecting non-invariant items using clusterwise simultaneous component analysis. Front. Psychol . 5: 604. doi: 10. 3389/fpsyg. 2014. 00604

||

Eid, M., and Diener, E. (2006). Handbook of Multimethod Measurement in Psychology . Washington, DC: American Psychological Association.

Fox, J.-P., and Verhagen, A. J. (2010). “ Random item effects modeling for cross-national survey data,” in Cross-cultural Analysis: Methods and Applications , eds E. Davidov, P. Schmidt, and J. Billiet (London: Routledge Academic), 467–488.

Geiser, C., Burns, G. L., and Servera, M. (2014). Testing for measurement invariance and latent mean differences across methods: interesting incremental information from multitrait-multimethod studies. Front. Psychol . 5: 1216. doi: 10. 3389/fpsyg. 2014. 01216

||

Gregorich, S. E. (2006). Do self-report instruments allow meaningful comparisons across diverse population groups? Testing measurement invariance using the confirmatory factor analysis framework. Med. Care 44(11 Suppl 3), 78–94. doi: 10. 1097/01. mlr. 0000245454. 12228. 8f

||

Guenole, N., and Brown, A. (2014). The consequence of ignoring measurement invariance for path coefficients in structural equation models. Front. Psychol . 5: 980. doi: 10. 3389/fpsyg. 2014. 00980

||

Hox, J. J., de Leeuw, E. D., and Zijlmans, E. A. O. (2015). Measurement equivalence in mixed mode surveys. Front. Psychol . 6: 87. doi: 10. 3389/fpsyg. 2015. 00087

||

Jak, S. (2014). Testing strong factorial invariance using three-level structural equation modeling. Front. Psychol . 5: 745. doi: 10. 3389/fpsyg. 2014. 00745

||

Jöreskog, K. G. (1971). Statistical analysis of sets of congeneric tests. Psychometrika 36, 109–133.

Jöreskog, K. G., and Sörbom, D. (1996). LISREL 8 User’s Reference Guide . Uppsala: Scientific Software International.

Kelcey, B., McGinn, D., and Hill, H. (2014). Approximate measurement invariance in cross-classified rater-mediated assessments. Front. Psychol . 5: 1469. doi: 10. 3389/fpsyg. 2014. 01469

||

Kroonenberg, P. M. (2014). Assessing factorial invariance of two-way rating designs using three-way methods. Front. Psychol . 5: 1495. doi: 10. 3389/fpsyg. 2014. 01495

||

Lommen, M. J. J., van de Schoot, R., and Engelhard, I. M. (2014). The experience of traumatic events disrupts the stability of a posttraumatic stress scale. Front. Psychol . 5: 1304. doi: 10. 3389/fpsyg. 2014. 01304

||

Mellenbergh, G. J. (1989). Item bias and item response theory. Int. J. Educ. Res . 13, 127–143. doi: 10. 1016/0883-0355(89)90002-5

|

Meredith, W. (1993). Measurement invariance, factor analysis and factorial invariance. Psychometrika 58, 525–543. doi: 10. 1007/BF02294825

|

Millsap, R. E. (2011). Statistical Approaches to Measurement Invariance . New York, NY: Routledge.

Muthén, B. O., and Asparouhov, T. (2014). IRT studies of many groups: the alignment method. Front. Psychol . 5: 978. doi: 10. 3389/fpsyg. 2014. 00978

||

Muthén, B. O., and Asparouhov, T. (2012). Bayesian structural equation modeling: a more flexible representation of substantive theory. Psychol. Methods 17, 313–335. doi: 10. 1037/a0026802

||

Muthén, B. O., and Muthén, L. K. (2013). Mplus Version 7. 11 Statistical Analysis with Latent Variables: User’s Guide . Los Angeles, CA: Muthén & Muthén.

Rosseel, Y (2012). lavaan: an R package for structural equation modeling. J. Stat. Software 48, 1–36.

STATA. (2015). Structural Equation Modeling Reference Manual 2015 . College Station, TX: Stata Press.

Sörbom, D. (1974). A general method for studying differences in factor means and factor structure between groups. Br. J. Math. Stat. Psychol . 27, 229–239. doi: 10. 1111/j. 2044-8317. 1974. tb00543. x

||

Sörbom, D. (1978). An alternative to the methodology for analysis of covariance. Psychometrika 43, 381–396. doi: 10. 1007/BF02293647

|

Vandenberg, R. J. (2002). Toward a further understanding of and improvement in measurement invariance methods and procedures. Organ. Res. Methods 5, 139–158. doi: 10. 1177/1094428102005002001

|

Vandenberg, R. J., and Lance, C. E. (2000). A review and synthesis of the measurement invariance literature: suggestions, practices, and recommendations for organizational research. Organ. Res. Methods 3, 4–70. doi: 10. 1177/109442810031002

|

van de Schoot, R., Kluytmans, A., Tummers, L., Lugtig, P., Hox, J., and Muthen, B. (2013). Facing off with Scylla and Charybdis: a comparison of scalar, partial and the novel possibility of approximate measurement invariance. Front. Psychol . 4: 770. doi: 10. 3389/fpsyg. 2013. 00770

||

van de Schoot, R., Lugtig, P., and Hox, J. (2012). A checklist for testing measurement invariance. Eur. J. Dev. Psychol . 9, 486–492. doi: 10. 1080/17405629. 2012. 686740

|

Verdam, M. G. E., and Oort, F. J. (2014). Measurement bias detection with Kronecker product restricted models for multivariate longitudinal data when the number of measurement occasions is large. Front. Psychol . 5: 1022. doi: 10. 3389/fpsyg. 2014. 01022

|

Wang, T., Merkle, E., and Zeileis, A. (2014). Score-based tests of measurement invariance: use in practice. Front. Psychol . 5: 438. doi: 10. 3389/fpsyg. 2014. 00438

|

Wicherts, J. M., and Dolan, C. V. (2010). Measurement invariance in confirmatory factor analysis: an illustration using IQ test performance of minorities. Educ. Measure . 29, 39–47. doi: 10. 1111/j. 1745-3992. 2010. 00182. x

|

Zercher, F., Schmidt, P., Cieciuch, J., and Davidov, E. (2015). The comparability of the universalism value over time and across countries in the European Social Survey: exact versus approximate measurement equivalence. Front. Psychol . 6: 733. doi: 10. 3389/fpsyg. 2015. 00733

||

Thank's for Your Vote!
Editorial: measurement invariance. Page 1
Editorial: measurement invariance. Page 2
Editorial: measurement invariance. Page 3
Editorial: measurement invariance. Page 4
Editorial: measurement invariance. Page 5
Editorial: measurement invariance. Page 6
Editorial: measurement invariance. Page 7
Editorial: measurement invariance. Page 8
Editorial: measurement invariance. Page 9

This work, titled "Editorial: measurement invariance" was written and willingly shared by a fellow student. This sample can be utilized as a research and reference resource to aid in the writing of your own work. Any use of the work that does not include an appropriate citation is banned.

If you are the owner of this work and don’t want it to be published on AssignBuster, request its removal.

Request Removal
Cite this Essay

References

AssignBuster. (2022) 'Editorial: measurement invariance'. 22 October.

Reference

AssignBuster. (2022, October 22). Editorial: measurement invariance. Retrieved from https://assignbuster.com/editorial-measurement-invariance/

References

AssignBuster. 2022. "Editorial: measurement invariance." October 22, 2022. https://assignbuster.com/editorial-measurement-invariance/.

1. AssignBuster. "Editorial: measurement invariance." October 22, 2022. https://assignbuster.com/editorial-measurement-invariance/.


Bibliography


AssignBuster. "Editorial: measurement invariance." October 22, 2022. https://assignbuster.com/editorial-measurement-invariance/.

Work Cited

"Editorial: measurement invariance." AssignBuster, 22 Oct. 2022, assignbuster.com/editorial-measurement-invariance/.

Get in Touch

Please, let us know if you have any ideas on improving Editorial: measurement invariance, or our service. We will be happy to hear what you think: [email protected]