To Check or Not to Check? A Comment on the Contemporary Psychometrics (ConPsy) Checklist for the Analysis of Questionnaire Items
Abstract
:1. Introduction
2. Sample Size Requirements
“In the absence of detailed information about the model and data complexities, a common approach is to consider the subjects-to-variables (STV) ratio, which refers to the ratio of the number of participants (subjects) to the number of measured variables (items) in the study.”
“ConPsy recommends a minimum sample size of 200 individuals as a reasonable guideline for models with up to 25 items.”
“As a general guideline, ConPsy recommends a minimum STV ratio of 8:1. However, [...], smaller STVs (0.7) are acceptable in larger samples exceeding 300 individuals.”
“It is essential to consider that when both confirmatory factor analysis (CFA) and exploratory factor analysis (EFA) are employed in a study, the sample size requirement doubles, as each method necessitates separate datasets.”
3. Recurring Call for Ordinal Factor Analysis
“It is unfortunately common in applied psychometrics to encounter situations where categorical data is mistakenly treated as continuous and normally distributed, particularly in the context of factor analysis. ConPsy emphasises the importance of using the appropriate method for each type of data at every stage of the analysis. It highlights the need to recognise and account for the categorical nature of the data to ensure accurate and valid results.”
“ConPsy advises the use of factor extraction methods for categorical data when the items a) have fewer than five response options, b) have five options but floor and ceiling effects are present (common for instance when screening tools for a certain diagnosis are administered to general population), or if c) the data are ordinal, and the assumption of equidistant categories does not necessarily hold. Biased estimates may emerge when the number of categories is below five and/or the categorical distribution is highly asymmetric (Beauducel & Herzberg, 2006; Rhemtulla et al., 2012).”
4. Estimation Methods in Exploratory and Confirmatory Analysis
5. Reliability
“Cronbach’s alpha assumes that measurement error is random and is influenced by the sample used. It also assumes unidimensionality and equal factor loadings across indicators, which may not always be justified for latent constructs.”
“Reliability measures have also been developed within the framework of latent variable modelling, with McDonald’s omega proposed as a more suitable measure for reliability in studies using structural equation modelling (SEM) compared to Cronbach’s alpha [...]. Omega is based on the congeneric model, which allows for non-equivalent factor loadings across items.”
6. Measurement Invariance
“Measurement invariance is a crucial property of a measurement scale that ensures bias due to exogenous factors does not affect the measurement. If measurement invariance is not established, interpreting score differences is theoretically not advised.”
“[...] establishing measurement invariance in psychometric scales is crucial for meaningful comparisons of factor and total scores (structural invariance) between different groups or conditions”
7. Validity
“In psychometric evaluation, the first step is to assess the dimensionality of a tool, as reliability is to be assessed for each dimension separately. Validity on the other hand is the last to be assessed as a tool cannot be valid unless shown reliable.”
8. Quantifying Model Error in Factor Analysis in Increased Standard Errors
9. Are Factor Models Needed at All for the Evaluation of Measurement Instruments?
10. To Check or Not to Check?
Funding
Conflicts of Interest
Abbreviations
CFA | confirmatory factor analysis |
ConPsy | contemporary psychometrics |
CTT | classical test theory |
EFA | exploratory factor analysis |
FA | factor analysis |
MI | measurement invariance |
ML | maximum likelihood |
SEM | structural equation model |
STV | subjects-to-variables |
ULS | unweighted least squares |
References
- Vitoratou, S.; Uglik-Marucha, E.; Hayes, C.; Pickles, A. A comprehensive guide for assessing measurement tool quality: The contemporary psychometrics (ConPsy) checklist. PsyArXiv 2023. Available online: https://psyarxiv.com/t2pbj/ (accessed on 30 August 2023). [CrossRef]
- DeVellis, R.F.; Thorpe, C.T. Scale Development: Theory and Applications; Sage: Thousand Oaks, CA, USA, 2021. [Google Scholar]
- Kyriazos, T.A. Applied psychometrics: Sample size and sample power considerations in factor analysis (EFA, CFA) and SEM in general. Psychology 2018, 9, 2207. [Google Scholar] [CrossRef]
- Marsh, H.W.; Hau, K.T.; Balla, J.R.; Grayson, D. Is more ever too much? The number of indicators per factor in confirmatory factor analysis. Multivar. Behav. Res. 1998, 33, 181–220. [Google Scholar] [CrossRef]
- Lüdtke, O.; Ulitzsch, E.; Robitzsch, A. A comparison of penalized maximum likelihood estimation and Markov Chain Monte Carlo techniques for estimating confirmatory factor analysis models with small sample sizes. Front. Psychol. 2021, 12, 615162. [Google Scholar] [CrossRef]
- Draxler, C.; Alexandrowicz, R.W. Sample size determination within the scope of conditional maximum likelihood estimation with special focus on testing the Rasch model. Psychometrika 2015, 80, 897–919. [Google Scholar] [CrossRef]
- Zimmer, F.; Draxler, C.; Debelak, R. Power analysis for the Wald, LR, score, and gradient tests in a marginal maximum likelihood framework: Applications in IRT. Psychometrika, 2022; ahead of print. [Google Scholar] [CrossRef]
- Groskurth, K.; Bluemke, M.; Lechner, C.M. Why we need to abandon fixed cutoffs for goodness-of-fit indices: An extensive simulation and possible solutions. Behav. Res. Methods, 2023; ahead of print. [Google Scholar] [CrossRef]
- Heene, M.; Hilbert, S.; Draxler, C.; Ziegler, M.; Bühner, M. Masking misfit in confirmatory factor analysis by increasing unique variances: A cautionary note on the usefulness of cutoff values of fit indices. Psychol. Methods 2011, 16, 319–336. [Google Scholar] [CrossRef]
- McNeish, D.; Wolf, M.G. Dynamic fit index cutoffs for confirmatory factor analysis models. Psychol. Methods 2023, 28, 61–88. [Google Scholar] [CrossRef]
- McNeish, D. Dynamic fit index cutoffs for factor analysis with Likert, ordinal, or binary responses. PsyArXiv 2023. [Google Scholar] [CrossRef]
- McNeish, D.; Wolf, M.G. Direct discrepancy dynamic fit index cutoffs for arbitrary covariance structure models. PsyArXiv 2023. [Google Scholar] [CrossRef]
- Flora, D.B.; Curran, P.J. An empirical evaluation of alternative methods of estimation for confirmatory factor analysis with ordinal data. Psychol. Methods 2004, 9, 466–491. [Google Scholar] [CrossRef] [PubMed]
- Rhemtulla, M.; Brosseau-Liard, P.É.; Savalei, V. When can categorical variables be treated as continuous? A comparison of robust continuous and categorical SEM estimation methods under suboptimal conditions. Psychol. Methods 2012, 17, 354–373. [Google Scholar] [CrossRef]
- Robitzsch, A. On the bias in confirmatory factor analysis when treating discrete variables as ordinal instead of continuous. Axioms 2022, 11, 162. [Google Scholar] [CrossRef]
- Robitzsch, A. Why ordinal variables can (almost) always be treated as continuous variables: Clarifying assumptions of robust continuous and ordinal factor analysis estimation methods. Front. Educ. 2020, 5, 589965. [Google Scholar] [CrossRef]
- Grønneberg, S.; Foldnes, N. Factor analyzing ordinal items requires substantive knowledge of response marginals. Psychol. Methods, 2022; ahead of print. [Google Scholar] [CrossRef]
- Foldnes, N.; Grønneberg, S. Pernicious polychorics: The impact and detection of underlying non-normality. Struct. Equ. Modeling 2020, 27, 525–543. [Google Scholar] [CrossRef]
- Robitzsch, A. Regularized generalized logistic item response model. Information 2023, 14, 306. [Google Scholar] [CrossRef]
- Kyriazos, T.; Poga-Kyriazou, M. Applied psychometrics: Estimator considerations in commonly encountered conditions in CFA, SEM, and EFA practice. Psychology 2023, 14, 799–828. [Google Scholar] [CrossRef]
- Whittaker, T.A.; Schumacker, R.E. A Beginner’s Guide to Structural Equation Modeling; Routledge: New York, NY, USA, 2022. [Google Scholar] [CrossRef]
- Robitzsch, A. Comparing the robustness of the structural after measurement (SAM) approach to structural equation modeling (SEM) against local model misspecifications with alternative estimation approaches. Stats 2022, 5, 631–672. [Google Scholar] [CrossRef]
- Robitzsch, A. Model-robust estimation of multiple-group structural equation models. Algorithms 2023, 16, 210. [Google Scholar] [CrossRef]
- Robitzsch, A. Modeling model misspecification in structural equation models. Stats 2023, 6, 689–705. [Google Scholar] [CrossRef]
- Cronbach, L.J. Coefficient alpha and the internal structure of tests. Psychometrika 1951, 16, 297–334. [Google Scholar] [CrossRef]
- Cronbach, L.J.; Shavelson, R.J. My current thoughts on coefficient alpha and successor procedures. Educ. Psychol. Meas. 2004, 64, 391–418. [Google Scholar] [CrossRef]
- Ellis, J.L. A test can have multiple reliabilities. Psychometrika 2021, 86, 869–876. [Google Scholar] [CrossRef]
- Meyer, P. Understanding Measurement: Reliability; Oxford University Press: Cambridge, UK, 2010. [Google Scholar] [CrossRef]
- Robitzsch, A.; Lüdtke, O. Some thoughts on analytical choices in the scaling model for test scores in international large-scale assessment studies. Meas. Instrum. Soc. Sci. 2022, 4, 9. [Google Scholar] [CrossRef]
- Lucke, J.F. The α and the ω of congeneric test theory: An extension of reliability and internal consistency to heterogeneous tests. Appl. Psychol. Meas. 2005, 29, 65–81. [Google Scholar] [CrossRef]
- McNeish, D. Thanks coefficient alpha, we’ll take it from here. Psychol. Methods 2018, 23, 412–433. [Google Scholar] [CrossRef] [PubMed]
- Davenport, E.C.; Davison, M.L.; Liou, P.Y.; Love, Q.U. Reliability, dimensionality, and internal consistency as defined by Cronbach: Distinct albeit related concepts. Educ. Meas. 2015, 34, 4–9. [Google Scholar] [CrossRef]
- McDonald, R.P. Test Theory: A Unified Treatment; Lawrence Erlbaum: Mahwah, NJ, USA, 1999. [Google Scholar] [CrossRef]
- Brennan, R.L. Generalizability theory and classical test theory. Appl. Meas. Educ. 2010, 24, 1–21. [Google Scholar] [CrossRef]
- Hancock, G.R.; Mueller, R.O. Rethinking construct reliability within latent variable systems. In Structural Equation Modeling: Present and Future—A Festschrift in Honor of Karl Jöreskog; Cudeck, R., du Toit, S., Sörbom, D., Eds.; Scientific Software International: Lincolnwood, IL, USA, 2001; pp. 195–216. [Google Scholar]
- Green, S.B.; Yang, Y. Reliability of summed item scores using structural equation modeling: An alternative to coefficient alpha. Psychometrika 2009, 74, 155–167. [Google Scholar] [CrossRef]
- Ellis, J.L. The Spearman-Brown formula and reliabilities of random test forms. arXiv 2023, arXiv:2308.13811. [Google Scholar] [CrossRef]
- Hunter, J.E. Probabilistic foundations for coefficients of generalizability. Psychometrika 1968, 33, 1–18. [Google Scholar] [CrossRef]
- Markus, K.A.; Borsboom, D. Reflective measurement models, behavior domains, and common causes. New Ideas Psychol. 2013, 31, 54–64. [Google Scholar] [CrossRef]
- McDonald, R.P. Generalizability in factorable domains: “Domain validity and generalizability”. Educ. Psychol. Meas. 1978, 38, 75–79. [Google Scholar] [CrossRef]
- McDonald, R.P. Behavior domains in theory and in practice. Alta. J. Educ. Res. 2003, 49, 212–230. [Google Scholar]
- Mulaik, S.A. Foundations of Factor Analysis; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar] [CrossRef]
- Stout, W. A nonparametric approach for assessing latent trait unidimensionality. Psychometrika 1987, 52, 589–617. [Google Scholar] [CrossRef]
- Krakowczyk, J.B.; Beckord, J.; Planert, J.; Kohl, P.; Schweda, A.; Teufel, M.; Bäuerle, A. Development and psychometric evaluation of the human-made disaster-related distress scale (MMDS). Psychiatry Res. 2023, 324, 115193. [Google Scholar] [CrossRef]
- Robitzsch, A.; Lüdtke, O. Why full, partial, or approximate measurement invariance are not a prerequisite for meaningful and valid group comparisons. Struct. Equ. Model. 2023; ahead of print. [Google Scholar] [CrossRef]
- Funder, D.C.; Gardiner, G. MIsgivings about measurement invariance. PsyArXiv 2023. [Google Scholar] [CrossRef]
- Welzel, C.; Inglehart, R.F. Misconceptions of measurement equivalence: Time for a paradigm shift. Comp. Political Stud. 2016, 49, 1068–1094. [Google Scholar] [CrossRef]
- De Boeck, P.; Gore, L.R. The janus face of psychometrics. In Essays on Contemporary Psychometrics; van der Ark, L.A., Emons, W.H.M., Meijer, R.R., Eds.; Springer: Cham, Switzerland, 2023; pp. 31–46. [Google Scholar] [CrossRef]
- Zwitser, R.J.; Glaser, S.S.F.; Maris, G. Monitoring countries in a changing world: A new look at DIF in international surveys. Psychometrika 2017, 82, 210–232. [Google Scholar] [CrossRef]
- Camilli, G. The case against item bias detection techniques based on internal criteria: Do item bias procedures obscure test fairness issues? In Differential Item Functioning: Theory and Practice; Holland, P.W., Wainer, H., Eds.; Erlbaum: Hillsdale, NJ, USA, 1993; pp. 397–417. [Google Scholar]
- El Masri, Y.H.; Andrich, D. The trade-off between model fit, invariance, and validity: The case of PISA science assessments. Appl. Meas. Educ. 2020, 33, 174–188. [Google Scholar] [CrossRef]
- Shealy, R.; Stout, W. A model-based standardization approach that separates true bias/DIF from group ability differences and detects test bias/DTF as well as item bias/DIF. Psychometrika 1993, 58, 159–194. [Google Scholar] [CrossRef]
- Robitzsch, A. Robust and nonrobust linking of two groups for the Rasch model with balanced and unbalanced random DIF: A comparative simulation study and the simultaneous assessment of standard errors and linking errors with resampling techniques. Symmetry 2021, 13, 2198. [Google Scholar] [CrossRef]
- Nye, C.D.; Drasgow, F. Effect size indices for analyses of measurement equivalence: Understanding the practical importance of differences between groups. J. Appl. Psychol. 2011, 96, 966–980. [Google Scholar] [CrossRef]
- Robitzsch, A. Estimation methods of the multiple-group one-dimensional factor model: Implied identification constraints in the violation of measurement invariance. Axioms 2022, 11, 119. [Google Scholar] [CrossRef]
- Funder, D.C. Misgivings: Some Thoughts about “Measurement Invariance”. Internet Blog Entry. 31 January 2020. Available online: https://bit.ly/3caKdNN (accessed on 30 August 2023).
- Schmitt, N.; Kuljanin, G. Measurement invariance: Review of practice and implications. Hum. Resour. Manag. Rev. 2008, 18, 210–222. [Google Scholar] [CrossRef]
- Schmitt, N.; Golubovich, J.; Leong, F.T.L. Impact of measurement invariance on construct correlations, mean differences, and relations with external correlates: An illustrative example using Big Five and RIASEC measures. Assessment 2011, 18, 412–427. [Google Scholar] [CrossRef] [PubMed]
- Li, H. The resolution of some paradoxes related to reliability and validity. J. Educ. Behav. Stat. 2003, 28, 89–95. [Google Scholar] [CrossRef]
- Feldt, L.S. Can validity rise when reliability declines? Appl. Meas. Educ. 1997, 10, 377–387. [Google Scholar] [CrossRef]
- Kane, M.T. A sampling model for validity. Appl. Psychol. Meas. 1982, 6, 125–160. [Google Scholar] [CrossRef]
- Polikoff, M.S. Instructional sensitivity as a psychometric property of assessments. Educ. Meas. 2010, 29, 3–14. [Google Scholar] [CrossRef]
- Naumann, A.; Hochweber, J.; Klieme, E. A psychometric framework for the evaluation of instructional sensitivity. Educ. Assess. 2016, 21, 89–101. [Google Scholar] [CrossRef]
- Wu, H.; Browne, M.W. Quantifying adventitious error in a covariance structure as a random effect. Psychometrika 2015, 80, 571–600. [Google Scholar] [CrossRef] [PubMed]
- Wu, H. An Empirical Bayesian Approach to Misspecified Covariance Structures. Unpublished Dissertation, Ohio State University, Columbus, OH, USA, 2010. Available online: https://bit.ly/3HGuLFT (accessed on 30 August 2023).
- Robitzsch, A. Linking error in the 2PL model. J 2023, 6, 58–84. [Google Scholar] [CrossRef]
- Uanhoro, J.O. Modeling misspecification as a parameter in Bayesian structural equation models. Educ. Psychol. Meas. 2023; ahead of print. [Google Scholar] [CrossRef]
- Uher, J. Psychometrics is not measurement: Unraveling a fundamental misconception in quantitative psychology and the complex network of its underlying fallacies. J. Theor. Philos. Psychol. 2021, 41, 58–84. [Google Scholar] [CrossRef]
- Nunnally, J.C.; Bernstein, I.R. Psychometric Theory; Oxford University Press: New York, NY, USA, 1994. [Google Scholar]
- Buonaccorsi, J.P. Measurement Error: Models, Methods, and Applications; CRC Press: Boca Raton, FL, USA, 2010. [Google Scholar] [CrossRef]
- Carroll, R.J.; Ruppert, D.; Stefanski, L.A.; Crainiceanu, C.M. Measurement Error in Nonlinear Models; CRC Press: Boca Raton, FL, USA, 2006. [Google Scholar] [CrossRef]
- Fuller, W.A. Measurement Error Models; Wiley: New York, NY, USA, 1987. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Robitzsch, A. To Check or Not to Check? A Comment on the Contemporary Psychometrics (ConPsy) Checklist for the Analysis of Questionnaire Items. Eur. J. Investig. Health Psychol. Educ. 2023, 13, 2150-2159. https://doi.org/10.3390/ejihpe13100151
Robitzsch A. To Check or Not to Check? A Comment on the Contemporary Psychometrics (ConPsy) Checklist for the Analysis of Questionnaire Items. European Journal of Investigation in Health, Psychology and Education. 2023; 13(10):2150-2159. https://doi.org/10.3390/ejihpe13100151
Chicago/Turabian StyleRobitzsch, Alexander. 2023. "To Check or Not to Check? A Comment on the Contemporary Psychometrics (ConPsy) Checklist for the Analysis of Questionnaire Items" European Journal of Investigation in Health, Psychology and Education 13, no. 10: 2150-2159. https://doi.org/10.3390/ejihpe13100151
APA StyleRobitzsch, A. (2023). To Check or Not to Check? A Comment on the Contemporary Psychometrics (ConPsy) Checklist for the Analysis of Questionnaire Items. European Journal of Investigation in Health, Psychology and Education, 13(10), 2150-2159. https://doi.org/10.3390/ejihpe13100151