Next Article in Journal
The Effect of Secondary Education Teachers’ Metacognitive Knowledge and Professional Development on Their Tacit Knowledge Strategies
Next Article in Special Issue
Measurement Invariance of the WISC-V across a Clinical Sample of Children and Adolescents with ADHD and a Matched Control Group
Previous Article in Journal / Special Issue
Gender Differences in the Wechsler Intelligence Scale for Children in a Large Group of Italian Children with Attention Deficit Hyperactivity Disorder
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

WISC-V Measurement Invariance According to Sex and Age: Advancing the Understanding of Intergroup Differences in Cognitive Performance

by
Marcela Rodríguez-Cancino
* and
Andrés Concha-Salgado
Department of Psychology, Universidad de La Frontera, Temuco 4811322, Chile
*
Author to whom correspondence should be addressed.
J. Intell. 2023, 11(9), 180; https://doi.org/10.3390/jintelligence11090180
Submission received: 29 April 2023 / Revised: 15 June 2023 / Accepted: 29 June 2023 / Published: 6 September 2023
(This article belongs to the Special Issue Construct Validity of the WISC)

Abstract

:
This study sought to verify whether the constructs measured on the WISC-V are equivalent according to sex and age group in Chilean students to substantiate intergroup comparisons. For this, the measurement invariance of two variants of the five-factor intelligence model was explored with the ten primary subtests (hierarchical and oblique) using multigroup confirmatory factor analysis. Seven hundred and forty participants between 6 and 16 years of age from the Chilean standardization sample were assessed. The results show complete invariance according to sex, but incomplete according to the age group. The implications of these findings in both the professional area of psychology and future research are discussed.

1. Introduction

In order to use a test ethically and rigorously within a psychological assessment process, the quality of its psychometric properties must be explored and guaranteed (Muñiz et al. 2015). This refers, for example, to the reliability of its results, evidence of the validity of the interpretations that emerge from its scores, updated and representative normative data on the target population, and proven fairness, i.e., the absence of biased elements (in its items, materials, etc.) deleterious to the performance of a particular group (APA 2020; AERA et al. 2018; Muñiz et al. 2015; Muñiz and Fonseca-Pedrero 2019; Vinet et al. 2023).
Given the tremendous professional responsibility involved in test use for diagnosis, intervention, or treatment decision-making, verifying its fairness should be explored before large-scale implementation (Marín 2021; Vinet et al. 2023). At the psychometric level, some strategies that can test for fairness are measurement invariance analysis and differential item functioning (DIF). These techniques can identify if a measurement instrument is free of bias, endorsing the use of its scores in a defined context and population (Elosua 2003; Gómez-Benito et al. 2010; International Journal of Testing 2017; McGill et al. 2020; Muñiz and Fonseca-Pedrero 2019; Pedrero 2019).

1.1. Measurement Invariance

In psychology measurement, a test that is designed to measure a specific attribute should reveal differences between individuals only if those individuals actually differ in the attribute measured; if this is not met, it is possible to suspect the existence of a measurement bias which is equivalent to the systematic inaccuracy of the measurement (Millsap 2011).
In clinical or educational assessment, when the norms of a test have been developed for the general population, it is relevant to ask whether these norms can be applied in the same way in the assessment of individuals belonging to the sub-groups of a population, according to their sex, educational level, age, and ethnicity, among others. Given this, measurement invariance allows us to establish whether a test manages to generate a fair measure of the performance of people belonging to different sub-groups of the population, or in other words, whether it allows us to determine if the test works in an equivalent (invariant) way among the groups compared. When this is not demonstrated, it allows us to think about the need to have specific norms for the sub-groups of a population (Wicherts 2016).
Measurement invariance analysis can verify if the association between the items (or test scores) and the latent factors (or latent traits that are being measured) of the individuals do not depend on belonging to a particular group, i.e., whether it is invariant or stable among the groups being compared (Van de Schoot et al. 2012). Group membership should not alter the relationships between the observed and latent variables (Olivera-Aguilar et al. 2018). The invariant scores support the quality of the inferences made after comparing the sub-groups of a population on a test, guaranteeing that the differences in the scores are not related to characteristics inherent to the group being tested nor biases in the test, but rather to real variations in the measured construct (Chen et al. 2015; Elosua 2005; Leitgöb et al. 2023; Pauls et al. 2020; Pedrero 2019; Van de Schoot et al. 2012).
The factors that generate measurement bias in a test could be how the subtests are administered, cultural aspects of its items, and examinee-related problems such as familiarity with the content, language problems, and educational differences. They must be detected to ensure proper test use, since even Fluid Reasoning tests considered “culture-free” may exhibit measurement biases (Wicherts 2016).
According to Putnick and Bornstein (2016) and Dimitrov (2010), the levels of invariance are usually defined as (1) configural invariance (equivalence of the factorial structure), (2) metric invariance (weak invariance, equivalence of factor loadings), (3) scalar invariance (strong invariance, equivalence of the intercepts), and (4) residual invariance (Strict invariance, equivalence in the indicator residuals). Reaching the level of configural invariance demonstrates that in the compared groups, the latent factors are specified by the same manifested variables (Elosua 2005). Then, the equivalence of the factor loadings that reflect the relevance of these indicators of the factor they define would be shown when the level of metric invariance is reached.
Subsequently, researchers should demonstrate scalar invariance, which means that the mean differences in the latent construct capture all the mean differences in the shared variance of the items (Putnick and Bornstein 2016). Therefore, the equivalence of the intercepts would imply that the compared groups have the same starting point in the measured variable. For example, if the scores on a vocabulary subtest of an intelligence test are compared between boys and girls, the equivalence of the intercepts would mean that both groups have the same mean score on this subtest conditional on the ability, although their means differ across other variables. The equivalence of the intercepts is essential because, if not fulfilled, the differences in the means of the measured variables between the groups cannot be solely attributed to group differences in the measured construct. In other words, if the equivalence of the intercepts is violated, the scores of the two groups cannot be comparable, and it cannot be valid to conclude the differences in the underlying construct (Van de Schoot et al. 2012).
Finally, residual invariance means that the sum of specific variance and error variance is similar across groups; namely, if it is possible to demonstrate the most demanding level of measurement invariance, it means that the residuals between the different groups are equivalent, which suggests that the measurement errors are similar between the groups (Putnick and Bornstein 2016). Although demonstrating residual invariance is essential for determining full factorial invariance, this step is not a prerequisite for comparing mean differences because the residuals are not part of the latent factor (Vandenberg and Lance 2000).
Establishing the measurement invariance of a test is an indispensable requirement within a practical or professional setting (Meredith 1993) and the construction of norms when a test is used within multicultural contexts (Wicherts 2016). Likewise, in the scientific field, comparisons between groups should only be made once it has been proven that the constructs measured are understood in the same way among the groups compared, which can be explored and guaranteed with measurement invariance analyses (Chen et al. 2015).
In the field of intelligence assessment, testing for measurement invariance is an excellent way to empirically explore whether the subtests or items of a test that measures cognitive abilities function in the same way across different sub-groups of a population to support the use of its norms (Wicherts 2016). Several studies on measurement invariance with the most well-known cognitive assessment instruments (WISC-R, WISC-IV, WAIS-III, WAIS-IV, JAT, RAKIT, KAIT, WJ IV COG) have shown varying results depending on the ethnicity, sex or age of the participants, showing that some of them pass tests of invariance (Chen et al. 2015; Dolan 2000; Tommasi et al. 2015), some fail (Dolan et al. 2004; Wicherts et al. 2004; Wicherts and Dolan 2010), and others partially achieve it (Abad et al. 2016; Bulut et al. 2021; Chen and Zhu 2012; Dolan et al. 2006; Immekus and Maller 2010). According to Wicherts (2016), it is disturbing to see that the tests commonly used in clinical or educational practice do not pass invariance tests, since they would be demonstrating that the scores obtained are not only due to differences in the latent cognitive abilities being measured, which should encourage the creation of specific sub-groups of a population, or bias corrections in the test items or subtests.
Another test with which invariance testing studies have been carried out in various parts of the world is the Wechsler Intelligence Scale for Children in its fifth edition (WISC-V), considering variables such as sex, ethnic group or age groups, among others (Chen et al. 2015; Dombrowski et al. 2021; Pauls et al. 2020; Reynolds and Keith 2017; Scheiber 2016). This scale evaluates cognitive functioning in childhood and adolescence by obtaining the total intelligence quotient (FSIQ), five primary indexes (Verbal Comprehension, Visual Spatial, Fluid Reasoning, Working Memory, Processing Speed), and five secondary indexes (Quantitative Reasoning, Auditory Working Memory, Nonverbal, General Ability, and Cognitive Competence), and currently includes of the evaluation of intelligence as the “gold standard,” used worldwide and adapted to various cultural realities (Chen et al. 2015; Coronel et al. 2006; Flanagan and Alfonso 2017; Giofrè et al. 2022; Niileksela and Reynolds 2019; Rosas et al. 2022; Wechsler 2014; Weiss et al. 2019).
WISC-V integrates the most recent contributions of the CHC (Cattell Horn Carroll) Theory of Intelligence, as well as evidence emanating from research in developmental psychology, cognitive neuroscience, and working memory measurement models (Canivez and Watkins 2016; Forns and Amador 2017; Wechsler 2014).
The CHC Hierarchical Model of Intelligence has been recognized as the first consensus, comprehensive, and empirically validated taxonomy of cognitive abilities (Brenlla 2013; Flanagan and Dixon 2014; McGrew et al. 2023). Theoretically underpins the interpretations of the results from the Wechsler scales (WAIS-IV, WISC-IV, and WISC-V in the United States), as well as other recognized test of wich the objectives are to measure the construct of intelligence (Brenlla 2013; Canivez et al. 2016).
Similar to the CHC intelligence model, the conceptualization of intelligence and the factor structure of WISC-V considers a hierarchical structure that includes a general intelligence factor, five broad level primary factors, which in turn, contain more specific cognitive skills (narrow level), and measure skills of fluid intelligence (Gf), crystallized intelligence (Gc), working memory (Gsm), Visual Processing (Gv); Auditory Processing (Ga), Processing Speed (Gs), among others (Brenlla 2013; Kaufman et al. 2016; McGrew et al. 2023; Flanagan and Dixon 2014).
Van de Vijver et al. (2019) and Pauls et al. (2020) refer to the indispensability of investigating the factorial invariance of the WISC to verify the universality of the intelligence construct that supports the interpretation of the scores of this instrument., i.e., to confirm that the same theoretical structure is suitably adjusted for the different sub-groups being compared.
A series of studies (shown in Table 1) have tested invariance of the WISC-V according to sociodemographic variables, finding mixed results. In some cultural realities, strict levels of invariance are reached in terms of participants’ sex, age, or ethnic group. In contrast, in others, its achievement is partial, considering the same grouping variables. This reveals an inconclusive panorama regarding the equivalence of the intelligence construct measured by this scale in different population sub-groups. Table 1 summarizes the main findings reported in scientific literature on invariance studies with the most up-to-date version of the WISC.

1.2. WISC-V Invariance According to Sex

As noted in Table 1, evidence of configural, metric, scalar, and residual invariance has been found according to the sex of those being evaluated in the standardization sample in the United States (Chen et al. 2015) and Taiwan (Chen et al. 2020). Also, evidence of configural, metric, and scalar invariance has been reported according to sex in the Afro-American, Hispanic, and Caucasian populations (Scheiber 2016) and clinical samples (Dombrowski et al. 2021) that endorse the interpretative strategy and performance comparisons among these groups.
Pauls et al. (2020) in Germany found partial scalar invariance according to the participants’ sex, noting that four of the fifteen subtests belonging to the cognitive domains of Verbal Comprehension, Fluid Reasoning, and Processing Speed were not stable between boys and girls. In addition, Smith and Graves (2021) found a level of partial scalar invariance according to sex in a sample of Afro-American children, demonstrating inequality in the subtests of Similarities (Verbal Comprehension) and Coding (Processing Speed). These results suggest that it is impossible to compare these groups and question the quality of the interpretations made from the performance of those evaluated with a possible presence of measurement bias (Meredith 1993).

1.3. WISC-V Invariance According to Age Group

Unlike the variable sex, measurement invariance of the WISC-V based on the age group has been less explored, emphasizing evidence of strict invariance only within the standardization sample in the United States (Reynolds and Keith 2017) and Taiwan (Chen et al. 2020). For their part, Dombrowski et al. (2021) only reached partial scalar invariance with the age group (6–8; 9–11; 12–14; 15–16), identifying imbalances in the subtests that comprise the cognitive domain of Fluid Reasoning. This lack of invariance on the most demanding levels indicates that the meaning of the measured constructs differs between the compared groups and that the scores are possibly affected by some measurement bias that must be examined at a greater depth (Vandenberg and Lance 2000).

1.4. WISC-V Invariance Studies in Chile

In Chile, the WISC-V has recently been standardized. It has psychometric evidence for its reliability, internal structure validity, its relation with other variables, and updated normative data in a broad sample of children and adolescents (Rosas et al. 2022). Regarding exploring fairness, only the study of Rodríguez-Cancino et al. (2021) has been conducted in Chile in a sample of 480 schoolchildren according to their origin (urban n = 320; rural n = 160). This study tested the invariance of the hierarchical model that includes the ten primary subtests grouped within five indexes, reaching the level of configural and partial metric invariance, in which nine of the ten subtests are invariant among the compared groups, except for the Similarities subtest that belongs to the Verbal Comprehension index. No other studies have explored measurement invariance according to the gender or age group in this country, neither with the current nor previous version of WISC.

1.5. Performance Comparisons with WISC-V According to Sex and Age

Concerning the variable sex, a study conducted with the standardization sample in the United States of WISC-V found higher scores for the indexes of Working Memory, Processing Speed, Nonverbal, Cognitive Competence, and FSIQ in girls. By contrast, boys only surpassed them on the Quantitative Reasoning index (Kaufman et al. 2016). Similar results were found in Spain (Hernández et al. 2017) and France (Grégoire 2020), where there was agreement that girls perform significantly superior to boys, particularly for the Processing Speed Index.
A similar trend is observed in Chile, given that in the comparisons of the performance on the WISC-V according to sex, it is the girls who achieve higher scores than the boys, specifically in the domain of Processing Speed (Garolera and Navarro 2020; Rosas et al. 2022).
Among other studies that have explored the differences in the performance on the WISC according to the sex of those evaluated, it is relevant to emphasize the findings of Giofrè et al. (2022). The authors carried out a meta-analysis reviewing 79 studies conducted using different versions of the Wechsler scales for children (WISC, WISC-R, WISC-III, WISC-IV, and WISC-V). They found that, throughout the versions, the boys tend to present better performances in visual spatial and crystallized intelligence tasks. In contrast, girls demonstrate significantly superior performance in processing speed tasks, in addition to showing that for the older versions of WISC, more significant differences were observed attributable to sex than for the most recent versions.
These findings are consistent with those reported by Grégoire (2020) in France, who analyzed the differences by sex in the performance on different versions of the WISC (WISC-R, WISC-III, WISC-IV, and WISC-V), with the data from the standardization samples demonstrating that the differences in performance between girls and boys have gradually been reduced through the versions of the scale, and that on the WISC-V, they disappear for FSIQ and in four of the five primary indexes. On the WISC-V, this author found better performance in the Figure Weights and Arithmetic subtests in favor of boys, and significantly higher performance of girls for Coding, Symbol Search, Cancellation, Comprehension, and Picture Span.
With other standardized instruments, a study by Bulut et al. (2021) found similar results indicating that girls perform better than boys in processing speed; however, boys surpass girls in the cognitive skills of working memory, visual spatial skills, and crystallized intelligence. According to these authors, the differences attributable to sex in intelligence vary between age groups and over time due to the inequality in the maturation rates between boys and girls, noting that younger girls surpass boys’ performance. This trend is inverted as they grow.
Regarding age groups, scientific literature has evidenced changes in cognitive ability development through childhood and adolescence (Ardila 2011; Sattler 2003). During childhood, intelligence tends to develop rapidly, with significant cognitive advancements occurring within various domains, such as language, memory capacity, problem-solving, and reasoning. As children transition into adolescence, their intelligence continues to evolve, although at a slower pace compared to childhood, and their cognitive abilities become more refined (Rosas et al. 2019). Adolescents typically experience enhanced reasoning skills, increased ability for abstract thinking, and improved executive functions such as planning, decision-making, and self-regulation (Center on the Developing Child at Harvard University 2011; Diamond 2013; Rosselli et al. 2010).
Because norms have been created according to age, no empirical studies have compared the performance of age groups on the WISC-V. However, there are studies of measurement invariance according to age, as mentioned previously.
As invariance testing is a mechanism to validate performance comparisons between groups, it is worth asking if the findings of differences in the performance on the WISC-V between boys and girls in Chile, and according to their age group, are possibly explained by real variations in the latent factor or by belonging to one of the compared groups. If explained by the latter, this will reflect the presence of measurement bias that would impede comparison and could reasonably make doubt the quality of the interpretations of the results.
On the other hand, determining the existence of biases in the measurement with WISC-V according to variables like sex or age group is especially critical in Chile, since no studies have produced evidence on these aspects. This instrument must be applied on a mandatory basis in the national school system to define the delivery of specialized support and establish the diagnosis of an intellectual disability (Ministerio Educación Chile 2009). Hence, it must have psychometric properties guaranteed for these purposes.

1.6. The Present Study

Having as a target the study of biases in the tests, the present study endeavors to answer international directives on the ethical use of instruments through the review of possible sources of fairness or bias of the WISC-V, to contribute to the accomplishment of fair psychological assessments and the incorporation of good practices in psychodiagnostics.
From the prepared problem and on a general level, the present study describes evidence of fairness on the WISC-V in measuring cognitive functioning among the Chilean child-youth population according to sex and age group.
Research question: Are the constructs measured using WISC-V equivalent according to the sex and age group of Chilean children and adolescents?
Hypothesis 1. 
The factorial model of the WISC-V is invariant according to sex.
Hypothesis 2. 
The factorial model of the WISC-V is invariant according to the participants’ age group.

2. Materials and Methods

2.1. Participants

The sample comprised 740 children and adolescents between 6 and 16 years of age, stratified according to the school’s funding scheme (Municipal = 33.6%, Subsidized Private = 34.9%, and Private = 31.5%). The data were collected through the WISC-V Standardization Project in Chile (Rosas et al. 2022), conducted by a research team at the Centro de Desarrollo de Tecnologías de Inclusión de la Pontificia Universidad Católica de Chile (CEDETi-UC).
The participants, from seven geographic zones across the country, were selected through non-probability purposive sampling, with the following inclusion criteria: (a) between 6 years and 16 years of age, and (b) not recently assessed with a similar scale. The exclusion criteria were a clinical diagnosis and/or permanent or temporary special education needs. Table 2 contains the complete information on the frequencies by sex and age group. The definition of the age groups for this study followed the previous evidence in studies that performed similar analyses, such as that by Chen et al. (2020) and the one by Dombrowski et al. (2021).

2.2. Instruments

Wechsler Intelligence Scale for Children, Fifth Edition (WISC-V)

This individually administered clinical instrument was developed to assess intellectual abilities in children and adolescents between 6 and 16 years of age (Rosas and Pizarro 2018; Wechsler 2014). The Chilean version of the WISC-V was standardized using data from 754 children and adolescents, stratified by socioeconomic level, and from 7 of the 15 regions of the country (Rosas et al. 2022). Regarding internal structure validity, the correlation matrix between the subtests and indexes in the Chilean sample behaved as theoretically expected (higher correlations between subtests belonging to the same cognitive domain). Also, the confirmatory factor analysis for the hierarchical model that includes ten primary subtests showed an excellent fit, where χ2(30) = 52.245, p < 0.001, χ2/df = 1.7, RMSEA = 0.031, CFI = 0.99, TLI = 0.99. The scale includes 15 subtests, of which 10 are primary and 5 complementary, organized within five cognitive domains, as illustrated in Table 3. Information about the reliability coefficients of the subtests and indexes for the Chilean sample is also included.
The Chilean version of the WISC-V obtains a Full Scale Intelligence Quotient (FSIQ) indicator after the administration of 7 primary subtests, 5 primary indexes, applying ten primary subtests and 5 secondary indexes, all standardized scores expressed on a point scale (M = 10, SD = 03) for the subtests and as composite scores (M = 100, SD = 15) for the indexes and FSIQ (Kaufman et al. 2016; Rosas and Pizarro 2018; Weiss et al. 2019).

2.3. Procedure

The data were collected by the WISC-V Standardization Project in Chile (Rosas et al. 2022), administering the instrument only to the children or adolescents who voluntarily agreed to participate (demonstrated by the signature of an informed assent) and whose parents authorized it by signing an informed consent form.
Each child was evaluated individually at their school and during the school day. The scale administration lasted between 60 and 90 min, the variation of which depended on the characteristics of each child evaluated (Rosas et al. 2022).
In terms of ethical safeguards, all the procedures carried out in the Standardization Project were approved by the Scientific Ethics Committee of the Pontificia Universidad Católica de Chile. The consent documents included explanations about the project’s aims, the instrument’s administration, and the right to withdraw from participating, at any time, without entailing any disadvantage. These documents specified safeguarding confidentiality and data use, indicating that they would be used anonymously and only for scientific and academic purposes.
To use the data in this study, authorization was sought from CEDETi-UC, the institution that owns the rights to the WISC-V in Chile and endorsed this study.

2.4. Data Analyses

For the analyses of invariance, the hierarchical five-factor model proposed by Wechsler (2014) for the American standardization sample was chosen, which also has suitable levels of psychometric fit with the Chilean standardization sample (see Figure 1), made up of 10 primary subtests, 5 indexes and a second-order general intelligence factor (Rosas et al. 2022). Following the findings of several similar investigations (Dombrowski et al. 2021; Reynolds and Keith 2017; Smith and Graves 2021), an alternative first-order model was also tested that includes 5 primary indexes (oblique factors) and 10 primary subtests (indicators), corresponding to the structure of the first-level WISC-V constructs (see Figure 1).
As a preliminary step in the factorial analysis of invariance, the baseline factorial models were tested for the entire sample (n = 740) and separately for each group according to sex and age group through a first and second-order confirmatory factor analysis (CFA). It should be noted that the 11 age groups, from 6 to 16 years old, were clustered following the Chen et al. (2020) and Dombrowski et al. (2021) strategy, namely: 6–8, 9–11, 12–14, and 15–16. According to the guidelines of Putnick and Bornstein (2016) and Dimitrov (2010), different levels of nested models included more restrictions at each level successively, simultaneously for the groups through the multigroup confirmatory factor analysis (MGCFA).
Along the same lines as Dombrowski et al. (2021), a series of progressive restrictions were applied to the models to determine their equivalence based on sex (boy/girl) and age group. First, the level of configural invariance was explored by verifying whether the structure of the WISC-V was stable between groups without applying restrictions to the parameters of the models. After this, the factor loadings in each group were restricted, except for the one in the reference group. Then, the fit indexes were compared to those of the configural level to test the metric invariance (or weak) level. Later, restrictions were applied to the intercepts to determine the level of scalar invariance (or strong), and the fit was compared with the level of metric invariance. Finally, to test the level of residual invariance (strict), i.e., to assess if the residuals of the indicators are equivalent between the groups, they were restricted, and then their fit was compared to the level of scalar invariance (Dimitrov 2010; Putnick and Bornstein 2016). In the case of the hierarchical model, the first order was tested, and then the second order in each step of the progressive levels of invariance (Dimitrov 2010).
The analyses were performed with the maximum likelihood, using a robust standard error (MLR) estimator for both the CFA and the MGCFA. As indicators to establish the fit of the factorial models, the root mean square error of approximation (RMSEA), comparative fit index (CFI), and the Tucker-Lewis index (TLI) were considered. This way, the presence of an optimal fit was considered when the CFI and TLI ≥ 0.95, and RMSEA < 0.05, and a reasonable fit when the CFI and TLI ≥ 0.90, and RMSEA < 0.08 (Browne and Cudeck 1992; Hu and Bentler 1999).
The criteria to establish invariance in the comparison of the models (metric versus configural, scalar versus metric) were: variations ≥ 0.010 in CFI, ≥0.015 in RMSEA, or ≥0.030 in SRMR indicate the absence of metric invariance and a difference ≥ 0.010 in CFI, ≥0.015 in RMSEA or ≥0.010 in SRMR suggest the absence of scalar and residual invariance (Chen 2007; Chen et al. 2020; Cheung and Rensvold 2002). All data analyses were carried out with the standardized scores of the subtests using Mplus 8.5 software.
In order to determine whether there were differences based on sex on the 10 WISC-V sub-test scores and the 5 indexes, two one-way MANOVAs were conducted. Also, a one-way ANOVA was performed to compare the FSIQ between boys and girls. A partial eta squared (η2p) was used to study effect size considering the Cohen (1988) rule (0.01, small; 0.06, medium; 0.14, large).

3. Results

3.1. Descriptions of WISC-V Scores According to Sex

At the subtest level, and considering a mean of 10 and a standard deviation of 3, both boys and girls performed slightly higher than 10 in BD, SI, MR, DS, VC, FW, VP, and PS. In CD and SS, the boys were below 10, and girls above this theoretical average. Regarding skewness and kurtosis, most values were less than 0.5 in absolute value, indicating a general trend toward symmetric and mesokurtic distributions, consistent with the normal transformation of the scores. In the VCI, VSI, FRI, WMI, and FSIQ indexes, boys and girls obtained averages slightly above 100 points, and only males were below that average in PSI. The trend in the descriptive measures of shape also accounts for symmetric and mesokurtic distributions.
Box’s test of equality of covariances matrices and Levene’s test of equality of error variances were not significant; therefore, the assumptions of the MANOVA were demonstrated. As seen in Table 4, there were significant differences in favor of boys for DS: F(1, 707) = 4.495, p = 0.034; η2p = 0.006 (small). On the contrary, significantly higher means were observed in girls for CD: F(1, 707) = 21.994, p < 0.001, η2p = 0.030 (small) and SS, F(1, 707) = 6.929, p = 0.009, η2p = 0.010 (small). Regarding the indexes, only for PSI did girls outperformed boys, F(1, 707) = 19.244, p < 0.001, η2p = 0.026 (small). All other comparisons of the primary sub-tests, indexes, and the FSIQ, showed no significant effects.

3.2. Analysis of Invariance According to Sex and Age Group

3.2.1. Baseline Models

The results of the CFA for the total sample and the groups separately show suitable fit indexes, both for the second-order five-factor model and for the first-order five-factor model, according to the sex and age of the participants (see Table 5).

3.2.2. Configural Invariance

Evidence of equivalence was found for the factor structure of the hierarchical five-factor model regarding the sex (see Table 6) and age groups (see Table 7). Also, for the five-factor oblique model in the two variables (see Table 8 and Table 9).

3.2.3. Metric Invariance

In the hierarchical five-factor model, the equivalence of factor loadings for sex was demonstrated both concerning first-order and second-order indicators as well as in the oblique five-factor model (see Table 6 and Table 8).
For the variable age group, only partial metric invariance was reached for the first-order indicators of the hierarchical five-factor model (see Table 7). The same occurred in the oblique five-factor model (see Table 9). According to the analyses reviewed, the most remarkable differences in the factorial weights are found in the subtests of the cognitive domain of Fluid Reasoning (Matrix Reasoning and Figure Weights) in the two factorial models tested. Figure 2, Figure 3, Figure 4 and Figure 5 show the comparison of the factor loadings for each model according to sex and age group.

3.2.4. Scalar Invariance

Concerning the equivalence of the intercepts of the indicators according to the variable sex, the level of scalar invariance was reached in the second-order five-factor model. It should be emphasized that although the ΔCFI value slightly exceeds what is permitted, it was decided to continue with the analysis since the other two indicators were fulfilled perfectly (see Table 6). In the oblique five-factor model, the equivalence of the intercepts based on sex is verified (see Table 8). For the age group, the scalar invariance level was not reached (see Table 7 and Table 9). Figure 2, Figure 3, Figure 4 and Figure 5 show the comparison of the intercepts for each model according to sex and age group.

3.2.5. Residual Invariance

Finally, the equivalence in the residuals of the indicators was reached only according to sex in both the hierarchical second-order five-factor model and the oblique five-factor model (see Table 6 and Table 8).
The summary of the reached invariance levels for the two models according to the two comparison variables can be seen in Table 10.

4. Discussion

According to international guidelines, the verification of a test’s psychometric properties guarantees that its use is fair and relevant to a specific population (AERA et al. 2018). Professionally, having this information is especially helpful when an evaluator must select the instruments to use as part of a psychological evaluation, taking into account the evidence around the equivalence of the measurement or the information of the normative data. This allows to determine if the test is appropriate according to the sociodemographic characteristics of the person being evaluated (APA 2020). On the other hand, and in the scope of the study, exploring the measurement invariance would make it possible to clarify if the scores that people obtain on a test genuinely reflect the level of the evaluated construct, or if this could depend on belonging to a particular group of a population, a situation in which its use for making intergroup performance comparisons would be questionable (Chen et al. 2015; Elosua 2005). Thus, the evidence offered by the exploration of measurement invariance is valuable and necessary in psychology, both in professional practice and research.
The present study sought to generate evidence about the fairness of the WISC-V, considering sociodemographic variables (sex and age group) in the measurement of the cognitive functioning of the Chilean child youth population to establish whether the constructs measured on the scale are equivalent within the sub-groups evaluated.

4.1. Invariance According to Sex

When taking the variable sex into account, the results show that in both the hierarchical and oblique models, the latent constructs are made up of the same manifested variables for boys and girls (configural invariance), i.e., there is equivalence in the form of measurement models, noting that the constructs of the WISC-V are measured with the same number of factors and the same number of indicators within each factor. Furthermore, in the following level of invariance (metric), equivalence was found for boys and girls in the factor loadings of each subtest, in the factor to which they belonged, establishing that in the two measurement models tested, the manifested variables are also predicted by the latent variables. These results are consistent with studies in other cultural realities, such as the United States (Chen et al. 2015; Dombrowski et al. 2021; Scheiber 2016), Taiwan (Chen et al. 2020), or Germany (Pauls et al. 2020).
Scalar invariance was also reached according to the sex of the participants in the oblique model, reflecting that boys and girls have the same starting point on the measurement scale of the subtests, i.e., they have the same subtests means. This finding demonstrates the absence of bias, and allows for the comparison of the results of these groups, since it is feasible to suppose that with the differences in the performance means in the measured variables, these would be attributable to fundamental differences in the latent construct. Moreover, these results are consistent with Dombrowski et al. (2021), who demonstrated scalar invariance according to sex using the oblique model.
The last level of measurement invariance tested in the present study (residual invariance) was also achieved according to the sex of the participants within the Chilean sample, similar to Chen et al. (2015) in the United States and Chen et al. (2020) in Taiwan, demonstrating equality among the residuals of the indicators in both models tested. If the residuals are invariant, then the variance, distribution, and correlation of the residuals are similar across all groups. This result reflects that sex differences in children’s performance on WISC-V are real differences in the constructs measured and cannot be attributed to measurement errors. Finally, it should be noted that the AIC indicator suggests that the higher-order model does not differ from the correlated model.
Thus, the results endorse the professional use of the interpretative and analytical strategy of the scores proposed by the WISC-V for Chilean boys and girls, i.e., the score calculation derives from grouping the 10 subtests that comprise the 5 primary indexes. In terms of research, the analysis of the differences in performance on the WISC-V must be carried out in light of the findings of the measurement invariance analysis in this study that legitimizes making intergroup comparisons, because when strict levels of invariance are reached it may be inferred that the differences in performance means are attributable to real variations in measured construct, and not due to measurement biases (Chen et al. 2015; Elosua 2005; Leitgöb et al. 2023).
Consistent with the evidence available in Chile, the comparisons made in the standardization sample showed that although in terms of FSIQ there are no differences in the performance on the WISC-V according to the sex of the participants, a better performance, statistically significant and in favor of the girls is noted on the Processing Speed index (Rosas et al. 2022), a result that was also found in this study and that is consistent with the studies by Kaufman et al. (2016), Hernández et al. (2017), and Grégoire (2020) within other cultural realities. Furthermore, concerning the subtests that comprise the WISC-V, according to the report by Garolera and Navarro (2020) in Chile, girls perform better on tasks that require scanning and visual discrimination, short-term visual memory, and quick and accurate decision-making (Coding and Symbol Search). In contrast, boys are more adept at managing auditory verbal information in working memory (Digit Span), as concluded in the present study. These differences, free of measurement error, could demonstrate that girls process visual stimuli faster or have more precise fine motor skills, better visuomotor coordination, or more effective attentional control than boys. Also, boys show a more remarkable ability to identify auditory verbal information, retain it in temporary storage, and resequence it for problem-solving. It should be noted that these results should be interpreted with caution, given that the effect size of these differences is small, which means that the magnitude of these differences may be limited.

4.2. Invariance According to Age Group

As far as the analyses of measurement invariance that considered the age group of the participants (6–8; 9–11; 12–14; 15–16), it should be emphasized that for both the hierarchical five-factor model and the oblique five-factor model the level of configural invariance was satisfied, which endorses the structure of the measurement models tested, i.e., equivalence is noted in the factorial structure (of manifest variables and latent factors) in all the age ranges. This result establishes that the ten primary subtests of the WISC-V are grouped in the same five latent factors in the different age groups of the Chileans evaluated, as in other cultural contexts where this has been explored (Chen et al. 2020; Dombrowski et al. 2021; Reynolds and Keith 2017).
Although the configuration of the models tested is equivalent among the compared groups, on the following level (metric invariance), it was observed that it is only possible to establish a partial equivalence as to the relevance of each of the subtests on the factor they define when the age of those evaluated is considered, noting a variant functioning particularly in the Matrix Reasoning and Figure Weights subtests, that make up the Fluid Reasoning index on the WISC-V and assess the skills of quantitative and inductive reasoning, simultaneous processing, and abstract thought. Although the little evidence on the matter is diverse, it should be highlighted that Dombrowski et al. (2021) partially achieved a higher level of (scalar) invariance than in the present study. The scalar and residual invariance levels were not reached in the Chilean sample, unlike what was reported by Chen et al. (2020) and Reynolds and Keith (2017) who found levels of strict invariance between the compared groups when their age group of reference was included.
As previously mentioned, their analysis detected imbalances in the subtests of the Fluid Reasoning index (Matrix Reasoning and Figure Weights). In these subtests, the task is for children to discover logical patterns underlying problems presented to them through visual stimuli. Although one might indeed think that in this non-verbal reasoning task it is less likely to find biases, as stated by Wicherts (2016), even tests considered “culture-free” may exhibit measurement biases therefore, according to these results, the possible presence of biased elements of the instrument must be explored. Wicherts (2016) points out that the factors that generate measurement bias in a test can be related to the form of administration, cultural aspects, problems of familiarity with the task, or educational differences, so it would be convenient that new research explores these aspects.
On the other hand, it should be noted that according to the contributions of the CHC theory, the subtests in which imbalances were found measure Fluid Reasoning (Gf), i.e., the mental operations used by an individual when faced with a relatively new task that cannot be performed automatically (Flanagan and Dixon 2014). From this definition, resolving new tasks requires the reorganization or transformation of information, perception of relationships between patterns, or inductive and deductive reasoning skills, which develop at a different pace and gradually during childhood and adolescence (Rosas et al. 2019). The results of this study suggest that the items of the Matrix Reasoning and Figure Weight subtests fail to measure these skills in the same way in children and adolescents. The two WISC-V models tested in this study (hierarchical and oblique) place these two subtests in a Fluid Reasoning domain. However, it is worth questioning whether this distribution appropriately accounts for this cognitive domain at all stages of the life cycle. Considering that the results of this study did not demonstrate the equivalence of the measure, in addition to the exploration of bias, it would be advisable to explore other factor models that explore new groupings of the subtests (for example, four-factor models), verify whether there are better factor structures, or different models, within the age ranges studied.
At the professional level, for the Chilean sample, the results of the present study reflect that there are elements affecting the measurement of the skills of quantitative and inductive reasoning through analysis of visual stimuli that must be taken into account in the assessment of cognitive functions with the WISC-V at different stages of development. At the research level, these results do not support the possibility of making performance comparisons on the WISC-V using the age of those evaluated as a comparison variable.

4.3. Implications

As previously stated, the evidence found in the present study endorses the use of the WISC-V professionally and in research when the sex of the examinees is considered, since the demonstration of measurement invariance reflects that differences in performance are attributable to real variations in the measured construct and that it is being detected effectively by this instrument. Nevertheless, when the age of those evaluated is considered, the results of this study suggest the need to explore the possible presence of measurement biases more exhaustively.

4.4. Future Studies

The differences in performance on the WISC-V in Chilean children indicate that girls are capable of processing visual information, making decisions, and executing them more quickly and efficiently than boys and that boys are better at retention tasks and auditory verbal information processing, which invites further studies to identify cultural, educational, or specific elements of the developmental trajectory of girls that could explain these differences. In this vein, a study that can collect evidence based on the response process would make it possible to approach this objective and enrich the evidence of validity that this instrument currently has in the Chilean population. On the other hand, regarding the variable age group, since the possibility of making intergroup performance comparisons based on this condition needs to be legitimized, it is not recommended that explorations of this type be carried out at the research level. Instead, it is encouraged to carry out studies that allow these findings to be deepened to achieve an understanding of what factors explain the variant functioning in the measurements of Fluid Reasoning that the WISC-V provides, for example, through differential item functioning (DIF).

Author Contributions

Both authors contributed substantially and equally to this research article. Conceptualization, M.R.-C. and A.C.-S.; methodology, M.R.-C. and A.C.-S.; software, A.C.-S.; validation, M.R.-C. and A.C.-S.; formal analysis, M.R.-C. and A.C.-S.; investigation, M.R.-C. and A.C.-S.; resources, M.R.-C. and A.C.-S.; data curation, M.R.-C. and A.C.-S.; writing—original draft preparation, M.R.-C. and A.C.-S.; writing—review and editing, M.R.-C. and A.C.-S.; visualization, M.R.-C. and A.C.-S.; supervision, M.R.-C. and A.C.-S.; project administration, M.R.-C.; funding acquisition, M.R.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Dirección de Investigación, Universidad de La Frontera, Project DIUFRO DI22-0030 and Agencia Nacional de Investigación y Desarrollo de Chile (ANID) Project FONDECYT Iniciación N°11230429.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and was approved by the Scientific Ethical Committee, Social Sciences, Arts and Humanities, of the Pontificia Universidad Católica de Chile (protocol code 15091200 and date of approval: 10 December, 2015) for human studies.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are not available due to ethical or privacy restrictions.

Acknowledgments

We want to thank the research team of Centro de Desarrollo de Tecnologías de Inclusión de la Pontificia Universidad Católica de Chile (CEDETi-UC), for providing data from the WISC-V Standardization Project of Chile.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Abad, Francisco J., Miguel A. Sorrel, Francisco J. Román, and Roberto Colom. 2016. The Relationships between WAIS-IV Factor Index Scores and Educational Level: A Bifactor Model Approach. Psychological Assessment 28: 987–1000. [Google Scholar] [CrossRef]
  2. AERA, APA, and CME. 2018. Estándares Para Pruebas Educativas y Psicológicas (M. Lieve, Trad.). Washington, DC: American Educational Research Association. [Google Scholar]
  3. APA. 2020. APA Guidelines for Psychological Assessment and Evaluation. Washington, DC: APA. [Google Scholar] [CrossRef]
  4. Ardila, Rubén. 2011. Inteligencia. ¿Qué sabemos y qué nos falta por investigar? Revista de la Academia Colombiana de Ciencias Exactas Físicas y Naturales 35: 97–103. [Google Scholar]
  5. Brenlla, María Elena. 2013. Interpretación Del WISC-IV. Puntuaciones Compuestas y Modelos CHC. Ciencias Psicológicas 7: 183–97. [Google Scholar] [CrossRef]
  6. Browne, Michael, and Robert Cudeck. 1992. Alternative Ways of Assessing Model Fit. Sociological Methods & Research 21: 230–58. [Google Scholar] [CrossRef]
  7. Bulut, Okan, Damien C. Cormier, Alexandra M. Aquilina, and Hatice C. Bulut. 2021. Age and Sex Invariance of the Woodcock-Johnson IV Tests of Cognitive Abilities: Evidence from Psychometric Network Modeling. Journal of Intelligence 9: 35. [Google Scholar] [CrossRef]
  8. Canivez, Gary L., and Marley W. Watkins. 2016. Review of the Weschler Intelligence Scale for Children—Fifth Edition: Critique, Commentary, and Independent Analyses in Intelligent Testing with the WISC-V. Hoboken: Wiley. [Google Scholar]
  9. Canivez, Gary L., Marley W. Watkins, and Stefan C. Dombrowski. 2016. Factor Structure of the Wechsler Intelligence Scale for Children—Fifth Edition: Exploratory Factor Analyses with the 16 Primary and Secondary Subtests. Psychological Assessment 28: 975–86. [Google Scholar] [CrossRef]
  10. Center on the Developing Child at Harvard University. 2011. Building the Brain’s ‘Air Traffic Control’ System: How Early Experiences Shape the Development of Executive Function. Available online: https://developingchild.harvard.edu/ (accessed on 31 May 2023).
  11. Chen, Fang Fang. 2007. Sensitivity of Goodness of Fit Indexes to Lack of Measurement Invariance. Structural Equation Modeling: A Multidisciplinary Journal 14: 464–504. [Google Scholar] [CrossRef]
  12. Chen, Hsinyi, and Jianjun Zhu. 2012. Measurement Invariance of WISC-IV across Normative and Clinical Samples. Personality and Individual Differences 52: 161–66. [Google Scholar] [CrossRef]
  13. Chen, Hsinyi, Jianjun Zhu, Yung-Kun Liao, and Timothy Z. Keith. 2020. Age and Gender Invariance in the Taiwan Wechsler Intelligence Scale for Children, Fifth Edition: Higher Order Five-Factor Model. Journal of Psychoeducational Assessment 38: 1033–45. [Google Scholar] [CrossRef]
  14. Chen, Hsinyi, Ou Zhang, Susan Engi Raiford, Jianjun Zhu, and Lawrence G. Weiss. 2015. Factor Invariance between Genders on the Wechsler Intelligence Scale for Children–Fifth Edition. Personality and Individual Differences 86: 1–5. [Google Scholar] [CrossRef]
  15. Cheung, Gordon W., and Roger B. Rensvold. 2002. Evaluating Goodness-of-Fit Indexes for Testing Measurement Invariance. Structural Equation Modeling: A Multidisciplinary Journal 9: 233–55. [Google Scholar] [CrossRef]
  16. Cohen, Jacob. 1988. Statistical Power Analysis for the Behavioral Sciences, 2nd ed. New York: Lawrence Erlbaum Associates. [Google Scholar]
  17. Coronel, Claudia Paola, Ana Betina Lacunza, and Norma Contini. 2006. Las Habilidades Cognitivas En Niños Privados Culturalmente. Resultados Preliminares de La Primera Fase de Evaluación. Revista Iberoamericana de Diagnóstico y Evaluación—E Avaliação Psicológica 2: 49–74. [Google Scholar]
  18. Diamond, Adele. 2013. Executive Functions. Annual Review of Psychology 64: 135–68. [Google Scholar] [CrossRef] [PubMed]
  19. Dimitrov, Dimiter. 2010. Testing for Factorial Invariance in the Context of Construct Validation. Measurement and Evaluation in Counseling and Development 43: 121–49. [Google Scholar] [CrossRef]
  20. Dolan, Conor V. 2000. Investigating Spearman’s Hypothesis by Means of Multi-Group Confirmatory Factor Analysis. Multivariate Behavioral Research 35: 21–50. [Google Scholar] [CrossRef]
  21. Dolan, Conor V., Roberto Colom, Francisco J. Abad, Jelte M. Wicherts, David J. Hessen, and Sophie Van De Sluis. 2006. Multi-Group Covariance and Mean Structure Modeling of the Relationship between the WAIS-III Common Factors and Sex and Educational Attainment in Spain. Intelligence 34: 193–210. [Google Scholar] [CrossRef]
  22. Dolan, Conor V., Willemijn Roorda, and Jelte M. Wicherts. 2004. Two Failures of Spearman’s Hypothesis: The GATB in Holland and the JAT in South Africa. Intelligence 32: 155–73. [Google Scholar] [CrossRef]
  23. Dombrowski, Stefan C., Marley W. Watkins, Ryan J. McGill, Gary L. Canivez, Calliope Holingue, Alison E. Pritchard, and Lisa A. Jacobson. 2021. Measurement Invariance of the Wechsler Intelligence Scale for Children, Fifth Edition 10-Subtest Primary Battery: Can Index Scores Be Compared across Age, Sex, and Diagnostic Groups? Journal of Psychoeducational Assessment 39: 89–99. [Google Scholar] [CrossRef]
  24. Elosua, Paula. 2003. Sobre la validez de los tests. Psicothema 15: 315–21. [Google Scholar]
  25. Elosua, Paula. 2005. Evaluación progresiva de la invarianza factorial entre las versiones original y adaptada de una escala de autoconcepto. Psicothema 17: 356–62. [Google Scholar]
  26. Flanagan, Dawn, and Shauna Dixon. 2014. The Cattell-Horn-Carroll Theory of Cognitive Abilities. In Encyclopedia of Special Education. Edited by Cecil Reynolds, Kimberly Vannest and Elaine Fletcher-Janzen. Hoboken: Wiley. [Google Scholar] [CrossRef]
  27. Flanagan, Dawn, and Vicent Alfonso. 2017. Essentials of WISC-V Assessment. Hoboken: Wiley. [Google Scholar]
  28. Forns, María, and J. A. Amador. 2017. Habilidades Clínicas Para Aplicar, Corregir e Interpretar Las Escalas de Inteligencia de Wechsler. Madrid: Pirámide. [Google Scholar]
  29. Garolera, Marion, and Valentina Navarro. 2020. Género y WISC en Chile: ¿mejora en el rendimiento o en la evaluación? Papeles de Investigación CEDETi-UC 13: 1–34. [Google Scholar]
  30. Giofrè, David, Katie Allen, Enrico Toffalini, and Sara Caviola. 2022. The Impasse on Gender Differences in Intelligence: A Meta-Analysis on WISC Batteries. Educational Psychology Review 34: 2543–68. [Google Scholar] [CrossRef]
  31. Gómez-Benito, Juana, M. Dolores Hidalgo, and Georgina Guilera. 2010. El Sesgo de los Instrumentos de Medición. Test Justos. Papeles del Psicólogo 31: 75–84. [Google Scholar]
  32. Grégoire, Jacques. 2020. Intellectual Differences between Boys and Girls, 35 Years of Evolution between WISC-R and WISC-V. A.N.A.E. 169: 1–8. [Google Scholar]
  33. Hernández, Ana, Cristina Aguilar, Èrica Paradell, María Muñoz, Louis-Charles Vannier, and Frédérique Vallar. 2017. The Effect of Demographic Variables on the Assessment of Cognitive Ability. Psicothema 29: 469–74. [Google Scholar] [CrossRef] [PubMed]
  34. Hu, Li-tze, and Peter Bentler. 1999. Cutoff Criteria for Fit Indexes in Covariance Structure Analysis: Conventional Criteria versus New Alternatives. Structural Equation Modeling: A Multidisciplinary Journal 6: 1–55. [Google Scholar] [CrossRef]
  35. Immekus, Jason C., and Susan J. Maller. 2010. Factor Structure Invariance of the Kaufman Adolescent and Adult Intelligence Test Across Male and Female Samples. Educational and Psychological Measurement 70: 91–104. [Google Scholar] [CrossRef]
  36. International Journal of Testing. 2017. ITC Guidelines for Translating and Adapting Tests (Second Edition). International Journal of Testing 18: 101–34. [Google Scholar] [CrossRef]
  37. Kaufman, Alan, Susan Engi Raiford, and Diane L. Coalson. 2016. Intelligent Testing with the WISC-V. Hoboken: Wiley. [Google Scholar]
  38. Leitgöb, Heinz, Daniel Seddig, Tihomir Asparouhov, Dorothée Behr, Eldad Davidov, Kim De Roover, Suzanne Jak, Katharina Meitinger, Natalja Menold, Bengt Muthén, and et al. 2023. Measurement Invariance in the Social Sciences: Historical Development, Methodological Challenges, State of the Art, and Future Perspectives. Social Science Research 110: 102805. [Google Scholar] [CrossRef]
  39. Marín, Carolina. 2021. Guía Práctica de Evaluación Psicológica Clínica. Desarrollo de Competencias. Madrid: Pirámide. [Google Scholar]
  40. McGill, Ryan J., T. Ward, and Gary L. Canivez. 2020. Use of Translated and Adapted Versions of the WISC-V: Caveat Emptor. School Psychology International 41: 276–94. [Google Scholar] [CrossRef]
  41. McGrew, Kevin S., W. Joel Schneider, Scott L. Decker, and Okan Bulut. 2023. A Psychometric Network Analysis of CHC Intelligence Measures: Implications for Research, Theory, and Interpretation of Broad CHC Scores ‘Beyond g’. Journal of Intelligence 11: 19. [Google Scholar] [CrossRef]
  42. Meredith, William. 1993. Measurement Invariance, Factor Analysis and Factorial Invariance. Psychometrika 58: 525–43. [Google Scholar] [CrossRef]
  43. Millsap, Roger E. 2011. Statistical Approaches to Measurement Invariance. New York: Routledge. Taylor and Francis Group. [Google Scholar]
  44. Ministerio Educación Chile. 2009. Decreto 170/Fija Normas Para Determinar Los Alumnos Con Necesidades Educativas Especiales Que Serán Beneficiarios de Las Subvenciones Para Educación Especial. Available online: https://www.leychile.cl/Navegar?idNorma=1012570 (accessed on 7 December 2022).
  45. Muñiz, José, Ana Hernández, and Vicente Ponsoda. 2015. Nuevas Directrices sobre el Uso de los Test: Investigación, Control de Calidad y Seguridad. Papeles del Psicólogo 36: 161–73. [Google Scholar]
  46. Muñiz, José, and Eduardo Fonseca-Pedrero. 2019. Diez pasos para la construcción de un test. Psicothema 31: 7–16. [Google Scholar] [CrossRef]
  47. Niileksela, Christopher, and Matthew Reynolds. 2019. Enduring the Tests of Age and Time: Wechsler Constructs across Versions and Revisions. Intelligence 77: 2–15. [Google Scholar] [CrossRef]
  48. Olivera-Aguilar, Margarita, Samuel H. Rikoon, Oscar Gonzalez, Yasemin Kisbu-Sakarya, and David P. MacKinnon. 2018. Bias, Type I Error Rates, and Statistical Power of a Latent Mediation Model in the Presence of Violations of Invariance. Educational and Psychological Measurement 78: 460–81. [Google Scholar] [CrossRef] [PubMed]
  49. Pauls, Franz, Monika Daseking, and Franz Petermann. 2020. Measurement Invariance Across Gender on the Second-Order Five-Factor Model of the German Wechsler Intelligence Scale for Children–Fifth Edition. Assessment 27: 1836–52. [Google Scholar] [CrossRef] [PubMed]
  50. Pedrero, Víctor. 2019. Medición de Invarianza Factorial En Pruebas de Gran Escala. In Validez de Las Evaluaciones Educacionales En Chile y Latinoamérica. Santiago: Ediciones UC, pp. 265–300. [Google Scholar]
  51. Putnick, Diane L., and Marc H. Bornstein. 2016. Measurement Invariance Conventions and Reporting: The State of the Art and Future Directions for Psychological Research. Developmental Review 41: 71–90. [Google Scholar] [CrossRef]
  52. Reynolds, Matthew R., and Timothy Z. Keith. 2017. Multi-Group and Hierarchical Confirmatory Factor Analysis of the Wechsler Intelligence Scale for Children—Fifth Edition: What Does It Measure? Intelligence 62: 31–47. [Google Scholar] [CrossRef]
  53. Rodríguez-Cancino, Marcela, María Beatriz Vizcarra, and Andrés Concha-Salgado. 2021. ¿Se Puede Evaluar a Niños Rurales con WISC-V? Explorando la Invarianza Factorial de la Inteligencia en Chile. Revista Iberoamericana de Diagnóstico y Evaluación—E Avaliação Psicológica 60: 117–31. [Google Scholar] [CrossRef]
  54. Rosas, Ricardo, and Marcelo Pizarro. 2018. WISC-V: Manual de Administración y Corrección. Santiago: Pontificia Universidad Católica de Chile, Centro de Desarrollo de Tecnologías de Inclusión & Pearson. [Google Scholar]
  55. Rosas, Ricardo, Marcelo Pizarro, Olivia Grez, Valentina Navarro, Dolly Tapia, Susana Arancibia, María Teresa Muñoz-Quezada, Boris Lucero, Claudia P. Pérez-Salas, Karen Oliva, and et al. 2022. Estandarización Chilena de la Escala Wechsler de Inteligencia para Niños—Quinta Edición. Psykhe 31: 1–23. [Google Scholar] [CrossRef]
  56. Rosas, Ricardo, Victoria Espinoza, Felipe Porflitt, and Francisco Ceric. 2019. Executive Functions Can Be Improved in Preschoolers Through Systematic Playing in Educational Settings: Evidence From a Longitudinal Study. Frontiers in Psychology 10: 2024. [Google Scholar] [CrossRef] [PubMed]
  57. Rosselli, Mónica, Esmeralda Matute, and Alfredo Ardila. 2010. Neuropsicología Del Desarrollo Infantil. Mexico: Manual Moderno. [Google Scholar]
  58. Sattler, Jerome. 2003. Child Assessment: Cognitive Applications. Mexico: Manual Moderno, vol. 1. [Google Scholar]
  59. Scheiber, Caroline. 2016. Is the Cattell–Horn–Carroll-Based Factor Structure of the Wechsler Intelligence Scale for Children—Fifth Edition (WISC-V) Construct Invariant for a Representative Sample of African–American, Hispanic, and Caucasian Male and Female Students Ages 6 to 16 Years? Journal of Pediatric Neuropsychology 2: 79–88. [Google Scholar] [CrossRef]
  60. Smith, Leann V., and Scott L. Graves. 2021. An Exploration of Gender Invariance of the WISC-V Among Black Children in an Urban School District. Contemporary School Psychology 25: 170–82. [Google Scholar] [CrossRef]
  61. Tommasi, Marco, Lina Pezzuti, Roberto Colom, Francisco J. Abad, Aristide Saggino, and Arturo Orsini. 2015. Increased Educational Level Is Related with Higher IQ Scores but Lower G-Variance: Evidence from the Standardization of the WAIS-R for Italy. Intelligence 50: 68–74. [Google Scholar] [CrossRef]
  62. Van de Schoot, Rens, Peter Lugtig, and Joop Hox. 2012. A Checklist for Testing Measurement Invariance. European Journal of Developmental Psychology 9: 486–92. [Google Scholar] [CrossRef]
  63. Van de Vijver, Fons, Lawrence G. Weiss, Donald Saklofke, Abigail Batty, and Aurelio Prifitera. 2019. A Cross-Cultural Analysis of the WISC-V. In WISC-V. Clinical Use and Interpretation, 2nd ed. London: Elsevier, pp. 223–44. [Google Scholar]
  64. Vandenberg, Robert J., and Charles E. Lance. 2000. A Review and Synthesis of the Measurement Invariance Literature: Suggestions, Practices, and Recommendations for Organizational Research. Organizational Research Methods 3: 4–70. [Google Scholar] [CrossRef]
  65. Vinet, Eugenia V., Marcela Rodríguez-Cancino, Ailin Sandoval-Dominguez, Paloma Rojas-Mora, and José L. Saiz. 2023. El Empleo de Test por Psicólogos/as Chilenos/as: Un Inquietante Panorama. Psykhe 32: 1–19. [Google Scholar] [CrossRef]
  66. Wechsler, David. 2014. WISC-V Wechsler Intelligence Scale for Children-Fifth Edition: Technical and Interpretive Manual. Bloomington: Pearson Clinical Assessment. [Google Scholar]
  67. Weiss, Lawrence G., Donald Saklofke, James Holdnack, and Aurelio Prifitera. 2019. WISC-V. Clinical Use and Interpretation, 2nd ed. London: Elsevier. [Google Scholar]
  68. Wicherts, Jelte M. 2016. The Importance of Measurement Invariance in Neurocognitive Ability Testing. The Clinical Neuropsychologist 30: 1006–16. [Google Scholar] [CrossRef]
  69. Wicherts, Jelte M., and Conor V. Dolan. 2010. Measurement Invariance in Confirmatory Factor Analysis: An Illustration Using IQ Test Performance of Minorities: Fall 2010. Educational Measurement: Issues and Practice 29: 39–47. [Google Scholar] [CrossRef]
  70. Wicherts, Jelte M., Conor V. Dolan, David J. Hessen, Paul Oosterveld, G. Caroline M. Van Baal, Dorret I. Boomsma, and Mark M. Span. 2004. Are Intelligence Tests Measurement Invariant over Time? Investigating the Nature of the Flynn Effect. Intelligence 32: 509–37. [Google Scholar] [CrossRef]
Figure 1. Hierarchical and oblique models of intelligence. Note: Figure created by the authors.
Figure 1. Hierarchical and oblique models of intelligence. Note: Figure created by the authors.
Jintelligence 11 00180 g001
Figure 2. First and second-order factor loadings and intercepts comparisons according to sex (hierarchical model). Note: Figure created by the authors.
Figure 2. First and second-order factor loadings and intercepts comparisons according to sex (hierarchical model). Note: Figure created by the authors.
Jintelligence 11 00180 g002
Figure 3. Factor loadings and intercepts comparisons according to sex (oblique model). Note: Figure created by the authors.
Figure 3. Factor loadings and intercepts comparisons according to sex (oblique model). Note: Figure created by the authors.
Jintelligence 11 00180 g003
Figure 4. First and second-order factor loading and intercept comparisons according to age group (hierarchical model). Note: Figure created by the authors.
Figure 4. First and second-order factor loading and intercept comparisons according to age group (hierarchical model). Note: Figure created by the authors.
Jintelligence 11 00180 g004
Figure 5. Factor loading and intercept comparisons according to age group (oblique model). Note: Figure created by the authors.
Figure 5. Factor loading and intercept comparisons according to age group (oblique model). Note: Figure created by the authors.
Jintelligence 11 00180 g005
Table 1. Invariance studies with WISC-V according to sociodemographic variables.
Table 1. Invariance studies with WISC-V according to sociodemographic variables.
AuthorsCountryGroupsResults
Chen et al. (2015)USAGender (male female)
n = 2200
Configural, metric, scalar and residual invariance between sexes with a second-order hierarchical pentafactorial model.
Scheiber (2016)USAEthnicity and gender (African-American, Hispanic, and Caucasian)
n = 2637
Configural, metric, and scalar invariance for the six groups according to ethnicity and sex with a second-order hierarchical pentafactorial model.
Reynolds and Keith (2017)USAAge group (11 separate age groups between 6 and 16 years)
n = 2200
Configural, metric, scalar, and residual invariance by age group with first-order, hierarchical, and bifactorial factorial models.
Pauls et al. (2020)GermanyGender (male female)
n = 1411
Configural, metric and partial scalar invariance according to sex, with a second-order hierarchical pentafactorial model.
Inequality in the Information, Figure Weights, Coding and Cancellation subtests
Smith and Graves (2021)USAGender (boys girls) among African Americans
n = 647
Configural, metric and partial scalar invariance between sexes, with an oblique model of five first-order factors.
Inequality in the Similarities and Coding subtests.
Chen et al. (2020)TaiwanGender (boys girls) and age group (6–8, 9–11, 12–14, and 15–16)
n = 1034
Configural, metric, scalar, and residual invariance between sexes and age group with a second-order hierarchical pentafactorial model.
Dombrowski et al. (2021)USASex (male female), age group (6–8, 9–11, 12–14, and 15–16), and
clinical diagnosis (ADHD, anxiety, and encephalopathy)
n = 5359
Configural, metric, and full scalar invariance for sex and clinical diagnosis groups.
Partial scalar invariance for the age group.
Invariance was tested in an oblique model of five first-order factors.
Inequality in the Fluid Reasoning subtests for the age group.
Rodríguez-Cancino et al. (2021)ChileOrigin (urban-rural)
n = 480
Configural invariance and partial metric according to origin with a second-order hierarchical pentafactorial model.
Inequality in the Similarities subtest.
Note: Table created by the authors; regarding gender and sex in Table 1, we present the same labels that authors used (male/female or boys/girls) in their respective studies.
Table 2. Distribution of the sample according to sex and age group.
Table 2. Distribution of the sample according to sex and age group.
Age Group
(in Years)
BoysGirlsMissingTotal Sample
n%n%n%n%
6–810051.5%9448.5%00.0%19426.2%
9–1111147.0%12352.1%20.8%23631.9%
12–149448.7%9850.8%10.5%19326.1%
15–165345.3%6454.7%00.0%11715.8%
Total sex35848.3%37951.2%30.5%740100%
Note: Table created by the authors.
Table 3. Distribution of subtests according to cognitive domain and reliability coefficients (internal consistency) for the Chilean version of the WISC-V.
Table 3. Distribution of subtests according to cognitive domain and reliability coefficients (internal consistency) for the Chilean version of the WISC-V.
Type of
Subtest
Cognitive Domain
Verbal Comprehension
(α = 0.943)
Visual Spatial
(α = 0.912)
Fluid Reasoning
(α = 0.945)
Working Memory
(α = 0.933)
Processing Speed
(α = 0.900)
Primary
subtest
Similarities
(SI; α = 0.921)
Block Design
(BD; α = 0.824)
Matrix Reasoning
(MR; α = 0.900)
Digit Span
(DS; α = 0.907)
Coding
(CD; α = 0.898)
Vocabulary
(VC; α = 0.888)
Visual Puzzles
(VP; α = 0.903)
Figure Weights
(FW; α = 0.941)
Picture Span
(PS; α = 0.891)
Symbol Search
(SS; α = 0.822)
Complementary
subtest
Information
(IN; α = 0.910)
Arithmetic
(AR; α = 0.900)
Letter-Number Sequencing (LN; α = 0.895)Cancellation
(CA; α = 0.645)
Comprehension
(CO; α = 0.876)
Note: Table created by the authors; α: Cronbach’s alpha.
Table 4. WISC-V descriptive statistics by sex and MANOVA results.
Table 4. WISC-V descriptive statistics by sex and MANOVA results.
BoysGirlsDifferences
Subtest/
Index
MSDSkew.Kurt.MSDSkew.Kurt.Fpη2p
BD10.252.919−0.025−0.01510.012.861−0.2720.2021.5960.2070.002
SI10.122.867−0.0380.03310.092.929−0.175−0.0270.0130.908<0.001
MR10.032.9780.4090.19110.242.7860.141−0.2650.5830.4450.001
DS10.382.8440.2380.47010.012.8220.3180.0784.4950.0340.006
CD9.752.7650.4800.18710.722.6550.1670.42221.994<.0010.030
VC10.192.8900.136−0.52010.072.981−0.125−0.6240.2880.592<0.001
FW10.202.965−0.427−0.37610.022.954−0.099−0.3110.9810.3220.001
VP10.042.9550.222−0.36610.202.8230.155−0.0360.090.765<0.001
PS10.133.023−0.085−0.14610.232.8200.155−0.2190.260.610<0.001
SS9.852.8980.4240.05510.422.9260.047−0.3296.9290.0090.010
VCI100.8514.0480.066−0.303100.4214.514−0.083−0.6250.131.718<0.001
VSI100.8414.4110.070−0.423100.5513.8790.0720.0640.358.5500.001
FRI100.6414.1940.044−0.567100.7114.2330.107−0.2310.033.855<0.001
WMI100.8213.8740.223−0.080100.0514.0640.277−0.1760.834.3620.001
PSI98.6113.7060.5980.471103.1213.520−0.110−0.26319.244<.0010.026
FSIQ100.7813.4910.141−0.084100.8713.934−0.045−0.3800.0080.930<0.001
Note: M = mean; SD = standard deviation; η2p = partial eta squared; significant effects in bold.
Table 5. Baseline CFA models results.
Table 5. Baseline CFA models results.
Model Fit Indexes
Groupχ2dfpCFITLIRMSEA
[90% CI]
SRMRAIC
Hierarchical Model
Full sample
n = 740
49.498300.0140.9900.9850.030
[0.013, 0.044]
0.02234,654.667
Boys
n = 358
49.982300.0120.9770.9660.043
[0.020, 0.064]
0.03216,887.609
Girls
n = 379
32.019300.3660.9980.9970.013
[0.000, 0.042]
0.02517,626.372
Age 6–8
n =194
41.552300.0780.9740.9610.045
[0.000, 0.075]
0.0379171.730
Age 9–11
n = 236
34.542300.2590.9940.9910.025
[0.000, 0.057]
0.03010,918.633
Age 12–14
n = 193
32.072300.4130.9980.9970.013
[0.000, 0.056]
0.0348942.556
Age 15–16
n = 117
39.451300.1150.9750.9630.052
[0.000, 0.092]
0.0465523.953
Oblique Model
Full sample
n = 740
28.263250.2960.9980.9970.013
[0.000, 0.033]
0.01534,643.432
Boys
n = 358
35.190250.0850.9880.9790.034
[0.000, 0.058]
0.02516,882.817
Girls
n = 379
23.328250.5581.0001.0000.000
[0.000, 0.038]
0.02017,627.681
Age 6–8
n = 194
23.448250.5511.0001.0000.000
[0.000, 0.053]
0.0299163.626
Age 9–11
n = 236
26.327250.3900.9980.9970.015
[0.000, 0.055]
0.02610,920.418
Age 12–14
n = 193
22.375250.6141.0001.0000.000
[0.000, 0.050]
0.0238944.859
Age 15–16
n = 117
34.680250.0940.9750.9550.058
[0.000, 0.100]
0.0415529.182
Note: χ2 = Chi-square; df = degrees of freedom; CFI = comparative fit index; TLI = Tucker Lewis index; RMSEA = root mean square error of approximation; SRMR = standardized residual root mean square; AIC = Akaike Information Criterion.
Table 6. WISC-V hierarchical multigroup CFA fit indexes and the comparison of invariance models by sex.
Table 6. WISC-V hierarchical multigroup CFA fit indexes and the comparison of invariance models by sex.
Invariance ModelModel Fit IndexesModel Comparison
χ2dfpCFITLIRMSEA
[90% CI]
SRMRAICComparisonΔχ2ΔdfΔCFIΔRMSEAΔSRMR
M0: Configural81.584600.0330.9880.9820.031
[0.009, 0.047]
0.02834,513.980------
M1: Metric
 (First-Order Loadings)
87.433650.0330.9880.9830.031
[0.009, 0.046]
0.03434,509.755M1–M05.84950.0000.0000.006
M2: Metric
 (Second-Order Loadings)
87.680690.0640.9900.9870.027
[0.000, 0.043]
0.03534,502.287M2–M10.24740.002−0.0040.001
M3: Scalar
 (Intercepts of the Indicators)
101.126740.0190.9850.9820.032
[0.013, 0.046]
0.03934,505.703M3–M213.4465−0.0050.0050.004
M4: Scalar
 (Intercepts of the First-Order Factors)
127.40178<0.0010.9730.9690.041
[0.028, 0.054]
0.04634,524.338M4–M326.2754−0.0120.0090.007
M5: Residual
 (Disturbances of First-Order Factors)
155.41087<0.0010.9630.9620.046
[0.034, 0.058]
0.05434,534.381M5–M428.0099−0.0100.0050.008
M6: Residual
 (Uniqueness of the Indicators)
171.79697<0.0010.9590.9620.046
[0.034, 0.057]
0.06434,533.061M6–M516.38610−0.0040.0000.010
Note: χ2 = Chi-square; df = degrees of freedom; CFI = comparative fit index; TLI = Tucker Lewis index; RMSEA = root mean square error of approximation; SRMR = standardized residual root mean square; AIC = Akaike Information Criterion.
Table 7. WISC-V hierarchical multigroup CFA fit indexes and comparison of invariance models by age group.
Table 7. WISC-V hierarchical multigroup CFA fit indexes and comparison of invariance models by age group.
Invariance ModelModel Fit IndexesModel Comparison
χ2dfpCFITLIRMSEA
[90% CI]
SRMRAICComparisonΔχ2ΔdfΔCFIΔRMSEAΔSRMR
M0: Configural148.3131210.0460.9860.9790.035
[0.005, 0.053]
0.03634,556.882------
M1: Metric
 (First-Order Loadings)
210.727136<0.0010.9610.9490.054
[0.040, 0.068]
0.07234,587.547M1–M062.41415−0.0250.0190.036
M1a: Metric–Partial
 (First-Order Loadings)
181.5331320.0020.9740.9650.045
[0.027, 0.060]
0.05734,568.112M1a–M033.22011−0.0120.010.021
M2: Metric–Partial
 (Second-Order Loadings)
286.790163<0.0010.9360.9290.064
[0.052, 0.076]
0.09434,607.742M2–M1105.25731−0.0380.0190.037
M3: Scalar–Partial
 (Intercepts of the Indicators)
321.900175<0.0010.9240.9220.067
[0.056, 0.079]
0.09434,616.754M3–M235.11012−0.0120.0030.000
M4: Scalar–Partial
 (Intercepts of the First-Order Factors)
337.396184<0.0010.9210.9230.067
[0.056, 0.078]
0.09434,616.174M4–M315.4969−0.0030.0000.000
M5: Residual–Partial
 (Disturbances of First-Order Factors)
292.425177<0.0010.9400.9390.059
[ 0.047, 0.071]
0.08934,588.607M5–M4−44.971−70.019−0.008−0.005
M6: Residual–Partial
 (Uniqueness of the Indicators)
397.257204<0.0010.9000.9120.072
[0.061, 0.082]
0.14434,644.369M6—M5104.83227−0.0400.0130.055
Note: χ2 = Chi-square; df = degrees of freedom; CFI = comparative fit index; TLI = Tucker Lewis index; RMSEA = root mean square error of approximation; SRMR = standardized residual root mean square; AIC = Akaike Information Criterion.
Table 8. Fit indexes of the multigroup CFA in the WISC-V oblique model and comparison of the invariance models according to sex.
Table 8. Fit indexes of the multigroup CFA in the WISC-V oblique model and comparison of the invariance models according to sex.
Invariance ModelModel Fit IndexesModel Comparison
χ2dfpCFITLIRMSEA
[90% CI]
SRMRAICComparisonΔχ2ΔdfΔCFIΔRMSEAΔSRMR
M0: Configural58.518500.1910.9960.9920.022
[0.000, 0.042]
0.02234,510.498------
M1: Metric
 (Loadings)
61.988550.2410.9960.9940.019
[0.000, 0.039]
0.02834,503.967M1–M03.47050.000−0.0030.006
M2: Scalar
 (Intercepts of the Indicators)
74.659600.0960.9920.9890.026
[0.000, 0.043]
0.03334,506.639M2–M112.6715−0.0040.0070.005
M3: Residual
 (Disturbances of First-Order Factors)
94.906700.0260.9870.9830.031
[0.011, 0.046]
0.04634,506.886M3–M220.24710−0.0050.0050.013
Note: χ2 = Chi-square; df = degrees of freedom; CFI = comparative fit index; TLI = Tucker Lewis index; RMSEA = root mean square error of approximation; SRMR = standardized residual root mean square; AIC = Akaike Information Criterion.
Table 9. Fit indexes of the multigroup CFA in the WISC-V oblique model and comparison of the invariance models according to age group.
Table 9. Fit indexes of the multigroup CFA in the WISC-V oblique model and comparison of the invariance models according to age group.
Invariance ModelModel Fit IndexesModel Comparison
χ2dfpCFITLIRMSEA
[90% CI]
SRMRAICComparisonΔχ2ΔdfΔCFIΔRMSEAΔSRMR
M0: Configural108.0711000.2730.9960.9920.021
[0.000, 0.045]
0.02934558.084------
M1: Metric
 (Loadings)
161.5431110.0010.9740.9580.050
[0.032, 0.066]
0.06434589.475M1–M053.47211−0.0220.0290.035
M1a: Metric–Partial
 (Intercepts of th Indicators)
166.947112<0.0010.9720.9540.051
[0.034, 0.067]
0.06734592.216M1a–M058.87612−0.0200.0300.038
M2: Scalar
 (Intercepts of the Indicators)
190.807124<0.0010.9660.9500.054
[0.038, 0.069]
0.06634592.167M2–M1a23.86012−0.0060.003−0.001
M3: Residual
 (Disturbances of First-Order Factors)
271.746151<0.0010.9380.9260.066
[0.053, 0.078]
0.10234624.586M3–M280.93927−0.0280.0110.036
Note: χ2 = Chi-square; df = degrees of freedom; CFI = comparative fit index; TLI = Tucker Lewis index; RMSEA = root mean square error of approximation; SRMR = standardized residual root mean square; AIC = Akaike Information Criterion.
Table 10. Summary of the reached levels of measurement invariance analyses according to sex and age group.
Table 10. Summary of the reached levels of measurement invariance analyses according to sex and age group.
SexAge Group
LevelHierarchicalObliqueHierarchicalOblique
ConfiguralYesYesYesYes
MetricYesYesPartialPartial
ScalarYesYesNoNo
ResidualYesYesNoNo
Note: Table created by the authors; “Yes” means that a certain level was reached.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodríguez-Cancino, M.; Concha-Salgado, A. WISC-V Measurement Invariance According to Sex and Age: Advancing the Understanding of Intergroup Differences in Cognitive Performance. J. Intell. 2023, 11, 180. https://doi.org/10.3390/jintelligence11090180

AMA Style

Rodríguez-Cancino M, Concha-Salgado A. WISC-V Measurement Invariance According to Sex and Age: Advancing the Understanding of Intergroup Differences in Cognitive Performance. Journal of Intelligence. 2023; 11(9):180. https://doi.org/10.3390/jintelligence11090180

Chicago/Turabian Style

Rodríguez-Cancino, Marcela, and Andrés Concha-Salgado. 2023. "WISC-V Measurement Invariance According to Sex and Age: Advancing the Understanding of Intergroup Differences in Cognitive Performance" Journal of Intelligence 11, no. 9: 180. https://doi.org/10.3390/jintelligence11090180

APA Style

Rodríguez-Cancino, M., & Concha-Salgado, A. (2023). WISC-V Measurement Invariance According to Sex and Age: Advancing the Understanding of Intergroup Differences in Cognitive Performance. Journal of Intelligence, 11(9), 180. https://doi.org/10.3390/jintelligence11090180

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop