Next Article in Journal
Stock Price Prediction Using CNN-BiLSTM-Attention Model
Previous Article in Journal
On Construction of Partially Dimension-Reduced Approximations for Nonstationary Nonlocal Problems of a Parabolic Type
Previous Article in Special Issue
Where Are We Going with Statistical Computing? From Mathematical Statistics to Collaborative Data Science
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Phi, Fei, Fo, Fum: Effect Sizes for Categorical Data That Use the Chi-Squared Statistic

by
Mattan S. Ben-Shachar
1,*,
Indrajeet Patil
2,
Rémi Thériault
3,
Brenton M. Wiernik
4 and
Daniel Lüdecke
5
1
Independent Researcher, Ramat Gan 5228555, Israel
2
Center for Humans and Machines, Max Planck Institute for Human Development, 13437 Berlin, Germany
3
Department of Psychology, Université du Québec à Montréal, Montréal, QC H2X 3P2, Canada
4
Independent Researcher, Tampa, FL 33604, USA
5
Institute of Medical Sociology, University Medical Center Hamburg-Eppendorf, 20246 Hamburg, Germany
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(9), 1982; https://doi.org/10.3390/math11091982
Submission received: 28 March 2023 / Revised: 17 April 2023 / Accepted: 20 April 2023 / Published: 22 April 2023
(This article belongs to the Special Issue Advances in Statistical Computing)

Abstract

:
In both theoretical and applied research, it is often of interest to assess the strength of an observed association. Existing guidelines also frequently recommend going beyond null-hypothesis significance testing and reporting effect sizes and their confidence intervals. As such, measures of effect sizes are increasingly reported, valued, and understood. Beyond their value in shaping the interpretation of the results from a given study, reporting effect sizes is critical for meta-analyses, which rely on their aggregation. We review the most common effect sizes for analyses of categorical variables that use the χ 2 (chi-square) statistic and introduce a new effect size—פ (Fei, pronounced “fay”). We demonstrate the implementation of these measures and their confidence intervals via the effectsize package in the R programming language.
MSC:
62-01; 62-04; 62-08; 62H17; 62P99

1. Introduction

Over the last two decades, there have been growing concerns about the so-called replication crisis in psychology and other fields [1,2]. As a result, the scientific community has paid increasing attention to the issue of replicability in science as well as to good research and statistical practices.
In this context, many have highlighted the limitations of null-hypothesis significance testing and called for more modern approaches to statistics. One such recommendation, for example, from the “New Statistics” initiative is to report effect sizes and their corresponding confidence intervals and to increasingly rely on meta-analyses to increase confidence in those estimations [3]. These recommendations are meant to complement (or even replace, according to some) null-hypothesis significance testing and would help transition toward a “cumulative quantitative discipline”.
These so-called “New Statistics” are synergistic because effect sizes are not only useful for interpreting study results in themselves but also because they are necessary for meta-analyses, which aggregate effect sizes and their confidence intervals to create a summary effect size of their own [4,5]. (The title of this paper is an allusion to the rhyme spoken by the giant in the English fairy tale Jack and the Beanstalk (“Fee-fi-fo-fum”).)
Unfortunately, popular software applications do not always offer the necessary implementations of the specialized effect sizes necessary for a given research design and their confidence intervals. In this paper, we want to focus on effect sizes for categorical data that are probably less well known than popular effect sizes like Cohen’s d or Pearson’s r [6,7]. For categorical data, d and r are inappropriate measures of an effect size. Cohen’s d refers to the standardized difference between the means of two populations, while Pearson’s correlation coefficient r measures linear correlations. Hence, both measures refer to continuous, not categorical, data.
To compare categorical data, for instance, where associations can be presented as contingency tables, several effect size metrics are available. Common effect sizes for 2-by-2 tables are odds ratios (OR), risk ratios (RR), or the phi ( ϕ ) coefficient. While phi can be interpreted similarly to a correlation coefficient, OR and RR are harder to interpret as they are not bounded between zero and one. Furthermore, RR is not symmetrical [8]. The size of the effect can change when columns and rows are exchanged. For tables with larger dimensions than 2-by-2, other effect sizes (like Cramér’s V) are available that share the property of phi of being able to be interpreted like a correlation coefficient and which are discussed later.
The observed distribution of categorical data—usually measured as multinomial variables—can also be compared to an expected distribution. Again, effect sizes to measure the strength of such associations show some limitations regarding ease of interpretation. What is missing here is an effect size whose metric is comparable to those for contingency tables.
This paper aims to review the most commonly used effect sizes for analyses of categorical variables that use the χ 2 (chi-square) test statistic and introduce a new effect size, פ (Fei, pronounced “fay”), which closes the gap of a missing effect size measure in a correlation-like metric that is appropriate for categorical data.
Importantly, we offer researchers an applied walkthrough on how to use these effect sizes in practice thanks to the effectsize package in the R programming language, which implements these measures and their confidence intervals [9,10]. The presented effectsize package closes another gap related to the aforementioned effect sizes because the uncertainty of such measures—expressed by their confidence intervals—is often not included in the output of statistical software. We cover, in turn, tests of independence (φ/phi, Cramér’s V) and tests of goodness-of-fit (Cohen’s w, Tschuprow’s T, and a new proposed effect size, פ/Fei).

2. Effect Sizes for Tests of Independence

The χ 2 test of independence between two categorical variables examines if the frequency distribution of one of the variables is dependent on the other. That is, are the two variables correlated such that, for example, members of group 1 on variable X are more likely to be members of group A on variable Y rather than evenly spread across Y variable groups A and B. Formally, the test examines how likely the observed conditional frequencies (cell frequencies) are under the null hypotheses of independence. This is done by examining the degree to which the observed cell frequencies deviate from the frequencies that would be expected if the variables were indeed independent. The test statistic for these tests is the χ 2 , which is computed as:
χ 2 = i = 1 l × k O i E i 2 E i ,
where O i are the observed frequencies and E i are the frequencies expected under independence, and l and k are the number of rows and columns, respectively, in the contingency table.
Instead of the deviations between the observed and expected frequencies, we can write χ 2 in terms of observed and expected cell probabilities and the total sample size N (since p = k / N ):
χ 2 = N × i = 1 l × k p O i p E i 2 p E i ,
where p O i are the observed cell probabilities and p E i the probabilities expected under independence.
Table 1 gives a short example in R to demonstrate whether the probability of survival is dependent on the sex of the passenger aboard the Titanic. The null hypothesis tested here is that the probability of survival is independent of the passenger’s sex.
The performed χ 2 -test is statistically significant. Thus, we can reject the hypothesis of independence. However, the output includes no effect size, and we cannot conclude the strength of the association between sex and survival.

2.1. Phi

For a 2-by-2 contingency table analysis, as the one used above, the ϕ (phi) coefficient is a correlation-like measure of effect size indicating the strength of association between the two binary variables. One possibility to compute this effect size is to recode the binary variables as dummy (“0” and “1”) variables and compute the Pearson correlation between them [11]:
ϕ = r A B
Another way to compute ϕ is by using the χ 2 statistic:
ϕ = χ 2 N = i = 1 l × k p O i p E i 2 p E i .
This value ranges between 0 (no association) and 1 (complete dependence), and its values can be interpreted the same as Pearson’s correlation coefficient. Table 2 shows the correlation coefficient and the effect size ϕ for the data shown in Table 1.
Note that ϕ cannot be negative, so we will take the absolute value of Pearson’s correlation coefficient. Also note that the effectsize package gives a one-sided confidence interval by default, to match the positive direction of the associated χ 2 test at α = 0.05 (that the association is larger than zero at a 95% confidence level).

2.2. Cramér’s V (and Tschuprow’s T)

When the contingency table is larger than 2-by-2, using χ 2 / N can produce values larger than 1, which loses its interpretability as a correlation-like effect size. Cramér showed that while for 2-by-2 the maximal possible value of χ 2 is N , for larger tables the maximal possible value for χ 2 is N × min k , l 1 [12]. Therefore, he suggested the V effect size (also sometimes known as Cramér’s phi and denoted as ϕ c ):
Cramer s   V = χ 2 N min k , l 1
where V is 1 when the columns are completely dependent on the rows or the rows are completely dependent on the columns (and 0 when rows and columns are completely independent).
Table 3 gives a short example in R to demonstrate whether the probability of survival is dependent on the person’s travel class or position aboard the Titanic. The null hypothesis tested here is that the probability of survival is independent of the travel class or position.
Tschuprow devised an alternative value, at
Tschuprow s   T = χ 2 N k 1 l 1
which is 1 only when the columns are completely dependent on the rows and the rows are completely dependent on the columns, which is only possible when the contingency table is a square [13].
For example, in Table 4, each row is dependent on the column value; that is, if we know if the food is a soy, milk, or meat product, we also know whether the food is vegan or not. However, the columns are not fully dependent on the rows: knowing the food is vegan tells us the food is soy-based; however, knowing it is not vegan does not allow us to classify the food—it can be either a milk product or a meat product.
Accordingly, as can be seen in Table 4, Cramer’s V will be 1, but Tschuprow’s T will not be:
We can generalize ϕ , V , and T to: χ 2 χ max 2 . That is, they express the square root of a proportion of the sample- χ 2 to the maximum possible χ 2 given the study design.
These coefficients can also be used for confusion matrices, which are 2-by-2 contingency tables used to assess machine learning algorithms’ classification abilities by comparing true outcome classes with the model-predicted outcome class. A popular metric is the Matthews correlation coefficient (MCC) for binary classifiers, which is often presented in terms of true and false positives and negatives but is nothing more than ϕ [14].

3. Effect Sizes for the Goodness-of-Fit Tests

These tests compare the observed distribution of a multinomial variable to the expected distribution using the same χ 2 statistic. Here, in addition, we can compute an effect size as χ 2 χ max 2 ; all we need to find is χ max 2 .

3.1. Cohen’s w

Cohen defined an effect size—w—for the goodness-of-fit test [7]:
Cohen s   w = i = 1 k p O i p E i 2 p E i = χ 2 N .
Thus, χ max 2 = N .
Unfortunately, w has an upper bound of 1 only when the variable is binomial (has two categories) and the expected distribution is uniform ( p = 1 p = 0.5 ). When the distribution is non-uniform or if there are more than two classes, then χ max 2 > N , and so w can be larger than 1 [15,16]. Examples are shown in Table 5.
Although Cohen suggested that w can also be used for such designs, we believe that this hinders the interpretation of w since it can be arbitrarily large [7].

3.2. Fei

We present here a new effect size, פ (Fei, pronounced “fay”), which normalizes goodness-of-fit χ 2 by the proper χ max 2 for non-uniform and/or multinomial variables.
The largest deviation from the expected probability distribution would occur when all observations are in the cell with the smallest expected probability. That is:
p O = 1 , if   p i = min p 0 , otherwise .
We can find E i O i 2 E i for each of these values:
p E p O 2 p E = p i 1 2 p i = 1 p i 2 p i , if   p E = min p E p i 0 2 p i = p i , otherwise .
Therefore,
i = 1 k p O i p E i 2 p E i = i = 1 k p E i min p E + 1 min p E 2 min p E = 1 min p E + 1 min p E 2 min p E = 1 min p E min p E = 1 min p E 1
So,
χ max 2 = N × i = 1 k p O i p E i 2 p E i = N × 1 min p E 1
Finally, an effect size can be derived as:
χ 2 N × 1 min p E 1
We call this effect size פ (Fei), which represents the voiceless bilabial fricative in Hebrew, keeping in line with ϕ (which in modern Greek marks the same sound) and V (which in English marks a voiced bilabial fricative; W being derived from the letter V in the modern Latin alphabet). פ will be 0 when the observed distribution perfectly matches the one expected (under the null hypothesis) and will be 1 when the sample contains only one class of observations—the one with the smallest expected probability (under the null hypothesis). That is, פ is 1 (its maximal value) only when we observe only the least expected class. When there are only two cells with uniform expected probabilities (50%), the expression N × 1 min p E 1 reduces to N and so פ = w . Table 6 shows the effect size Fei for the same vectors and distributions as seen for Cohen’s w in Table 5. As can be seen, unlike Cohen’s w, all effect size values of Fei (and their confidence intervals) are within the range from 0 to 1. (See Section 6 below for how to type the פ symbol on various computer systems.)
The computation of פ (Fei) can be achieved with the Fei() function of the effectsize package—fei(c(90, 10), p = c(0.35, 0.65)). This function here computes confidence intervals using the non-centrality parameter method (also called the “pivot method”) by finding values for the non-centrality parameter (“ncp”) of a noncentral χ 2 distribution that place the observed χ 2 test statistic at the desired probability point of the distribution (e.g., p = 0.025 and p = 0.975 for a two-sided 95% confidence interval). The two ncps (for the lower and upper bounds) are then converted back to פ using the same scale-and-root formula as with the sample χ 2 :
פ L = λ L N × 1 min p E 1
פ U = λ U N × 1 min p E 1
where λ L and λ U are the non-centrality parameters corresponding to the desired tail probabilities, and פ L and פ U are the bounds of the confidence intervals for פ (Fei) (See also the next section).

4. Simulation Study of the Distributional Form of the Fei Effect Size

In the previous section, we showed some results for the effect size פ (Fei) and its confidence intervals for different distributions of a multinomial variable. Like all effect sizes discussed in this paper, פ follows a scaled non-central χ distribution: the χ 2 statistic follows a non-central χ 2 distribution, and its square root follows a non-central χ distribution; this random variable is then scaled by a constant that is a function of the sample size and the study design. The noncentrality parameter of the non-central χ distribution can be found by applying the inverse of the scale to the population effect size. Therefore,
פ ^ noncentral   χ d f = k 1 , n c p = f 1 פ N × 1 min p E 1
where f 1 פ is the inverse of the χ to פ conversion:
f 1 פ = פ × N × 1 min p E 1 ,
פ is the population effect size, k is the number of classes, and פ ^ is the random variable of possible observed effect sizes in a random sample. This can also be formulated in terms of a non-central χ 2 distribution:
פ ^ noncentral   χ 2 d f = k 1 , n c p = g 1 פ × N × 1 min p E 1
where g 1 פ is the inverse of the χ 2 to פ conversion:
g 1 פ = פ 2 × N × 1 min p E 1 .
To validate our assumptions, we conducted a simulation study, where we simulated data of multinomial distributions for known true effect sizes of 0.1, 0.3, and 0.5, respectively. The datasets contained 500 simulations per effect size, for three different expected probabilities (same as in Table 6), and 3 different sample sizes of 50, 100, and 350, resulting in 13,500 simulated data points (500 simulations × 3 effect sizes × 3 expected probabilities × 3 different sample sizes). Figure 1 shows the results from the simulation study.
The smallest sample size is more affected by noise, and results show more variation (and less continuity) of simulation-based פ (Fei) values around the true effect sizes. For sample sizes N = 100 and N = 350, פ values closely replicate the true effect sizes and clearly follow a non-central χ distribution, indicating that Fei, like ϕ , V, T, and w, is a scaled χ value.
פ (Fei) following a non-central χ distribution also allows for power calculation. For example, if the null probabilities are [0.35, 0.65] and the alternative probabilities are [0.545, 0.455], the scaling constant is:
C = 1 min p E 1 = 1.857 ,
and the population effect size is
פ = i = 1 k p O i p E i 2 p E i 1.857 = 0.3 .
Therefore, the sample פ will follow the following distribution:
פ ^ noncentral   χ 2 d f = 1 , n c p = 0.3 2 × 1.857 × N × N × 1.857 .
One must then find the N that produces the desired power for the significance level that will be used to reject the null. For example, for a significant level of 0.01 and a power of at least 0.85, an N of at least 78 is required.
The pwr package in R provides a function (pwr.chisq.test()) that can be used to calculate the power for goodness-of-fit tests. Although the function uses Cohen’s w as input/output, פ can easily be converted to Cohen’s w (e.g., by using the fei_to_w() function from effectsize), allowing for the pwr function to be used with פ. An example can be found in the accompanying R code.

5. Conclusions

Effect sizes are essential to interpreting the magnitude of observed effects; they are frequently required in scientific journals; and they are necessary for a cumulative quantitative science relying on meta-analyses. In this paper, we have covered the mathematics and implementation in R of four different effect sizes for analyses of categorical variables that specifically use the χ 2 (chi-square) statistic. Furthermore, with our proposal of the effect size פ (Fei), we fill in the missing effect size for all cases of a χ 2 test, as can be seen in Table 7.
Thus, we now have effect sizes to accompany any sized 1-dimensional or 2-dimensional contingency tables that represent the sample’s χ 2 relative to the maximally possible χ 2 , ranging from 0 to 1, that can be easily interpreted on the scale of a correlation coefficient.

6. How to Type the פ Symbol

The Hebrew character can be inserted into documents via several methods:
  • By copying the character from https://util.unicode.org/UnicodeJsps/character.jsp?a=05E4 (access date: 9 March 2023) or similar webpages.
  • In R, by typing the string “\u05e4”.
  • In LaTeX, by typing \char”05e4 and using a Unicode-compatible compiler, such as XeTeX or LuaLaTeX.
  • In Microsoft Word, from the Hebrew character of the Symbols window (Insert → Symbol…) or by typing 05e4, followed by Alt + X on the keyboard (Windows only).
  • On Windows, using the Character Map application or by holding down Alt and typing +1508 on the numeric keypad.
  • On macOS, by enabling the Unicode Hex Input language from System Settings… → Keyboard, then typing Opt + 05e4.

Author Contributions

M.S.B.-S. conceptualized and developed the Fei effect size and its implementation in effectsize, conducted the simulation study, and drafted the paper; D.L. prepared the revised manuscript; I.P., R.T., B.M.W. and D.L. contributed to both the writing of the paper and the conception of the software. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The R code to reproduce the results from the tables and the simulation study in this article can be downloaded at https://osf.io/cg64s/ (doi: 10.17605/osf.io/cg64s).

Acknowledgments

The {effectsize} package is part of the collaborative R easystats ecosystem. Thus, we thank all members of easystats, contributors, and users alike.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Open Science Collaboration. Estimating the Reproducibility of Psychological Science. Science 2015, 349, aac4716. [Google Scholar] [CrossRef] [PubMed]
  2. Camerer, C.F.; Dreber, A.; Holzmeister, F.; Ho, T.-H.; Huber, J.; Johannesson, M.; Kirchler, M.; Nave, G.; Nosek, B.A.; Pfeiffer, T.; et al. Evaluating the Replicability of Social Science Experiments in Nature and Science between 2010 and 2015. Nat. Hum. Behav. 2018, 2, 637–644. [Google Scholar] [CrossRef] [PubMed]
  3. Cumming, G. The New Statistics: Why and How. Psychol. Sci. 2014, 25, 7–29. [Google Scholar] [CrossRef] [PubMed]
  4. Wiernik, B.M.; Dahlke, J.A. Obtaining Unbiased Results in Meta-Analysis: The Importance of Correcting for Statistical Artifacts. Adv. Methods Pract. Psychol. Sci. 2020, 3, 94–123. [Google Scholar] [CrossRef]
  5. DeGeest, D.S.; Schmidt, F.L. The Impact of Research Synthesis Methods on Industrial-Organizational Psychology: The Road from Pessimism to Optimism about Cumulative Knowledge. Res. Synth. Methods 2010, 1, 185–197. [Google Scholar] [CrossRef] [PubMed]
  6. Pearson, K., VII. Note on Regression and Inheritance in the Case of Two Parents. Proc. R. Soc. Lond. 1895, 58, 240–242. [Google Scholar] [CrossRef]
  7. Cohen, J. Statistical Power Analysis for the Behavioral Sciences, 2nd ed.; Routledge: Oxfordshire, UK, 1988; ISBN 978-0-203-77158-7. [Google Scholar]
  8. Cummings, P. The Relative Merits of Risk Ratios and Odds Ratios. Arch. Pediatr. Adolesc. Med. 2009, 163, 438. [Google Scholar] [CrossRef] [PubMed]
  9. Ben-Shachar, M.; Lüdecke, D.; Makowski, D. Effectsize: Estimation of Effect Size Indices and Standardized Parameters. J. Open Source Softw. 2020, 5, 2815. [Google Scholar] [CrossRef]
  10. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2023. [Google Scholar]
  11. Olvera Astivia, O.L. The Relationship between the Phi Coefficient and the Chi-Square Test of Association. Psychometroscar 2022. Available online: https://psychometroscar.com/2022/04/21/the-relationship-between-the-phi-coefficient-and-the-chi-square-test-of-association/ (accessed on 9 March 2023).
  12. Cramér, H. Mathematical Methods of Statistics; Princeton University Press: Princeton, NJ, USA, 1999; ISBN 978-0-691-00547-8. [Google Scholar]
  13. Tschuprow, A.A. Principles of the Mathematical Theory of Correlation; W. Hodge, Limited: London, UK, 1939. [Google Scholar]
  14. Chicco, D.; Jurman, G. The Advantages of the Matthews Correlation Coefficient (MCC) over F1 Score and Accuracy in Binary Classification Evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [PubMed]
  15. Rosenberg, M.S. A Generalized Formula for Converting Chi-Square Tests to Effect Sizes for Meta-Analysis. PLoS ONE 2010, 5, e10059. [Google Scholar] [CrossRef] [PubMed]
  16. Johnston, J.E.; Berry, K.J.; Mielke, P.W. Measures of Effect Size for Chi-Squared and Likelihood-Ratio Goodness-of-Fit Tests. Percept. Mot. Skills 2006, 103, 412–414. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison of true and simulation-based actual effect size פ (Fei) for different expected proportions, true effect sizes, and sample sizes. Histograms represent the distributions of the sample פ’s from the simulated datasets, and the plotted lines represent density functions for (scaled) non-central χ distributions for the corresponding effect sizes and sample sizes.
Figure 1. Comparison of true and simulation-based actual effect size פ (Fei) for different expected proportions, true effect sizes, and sample sizes. Histograms represent the distributions of the sample פ’s from the simulated datasets, and the plotted lines represent density functions for (scaled) non-central χ distributions for the corresponding effect sizes and sample sizes.
Mathematics 11 01982 g001
Table 1. χ 2 test of survival of Titanic passengers by sex, Titanic dataset from R.
Table 1. χ 2 test of survival of Titanic passengers by sex, Titanic dataset from R.
SexSurvivedDied
Male3671364
Female344126
χ 2 = 456.9, df = 1, p < 0.001.
Table 2. Correlation and effect size ϕ (phi) for the survival of Titanic passengers by sex, Titanic dataset from R.
Table 2. Correlation and effect size ϕ (phi) for the survival of Titanic passengers by sex, Titanic dataset from R.
Variable 1Variable 2r (95% CI) ϕ (95% CI)
Sex (male/female)Survival (survived/died)−0.46 (−0.49, −0.42)0.46 (0.42, 1.00)
Table 3. Effect size Cramér’s V for the survival of Titanic passengers by class/position, Titanic dataset from R.
Table 3. Effect size Cramér’s V for the survival of Titanic passengers by class/position, Titanic dataset from R.
Class/PositionSurvivedDied
1st203122
2nd118167
3rd178528
Crew212673
Cramér’s V = 0.29, 95% CI = 0.26, 1.00.
Table 4. Cramér’s V and Tschuprow’s T for food classes, example dataset from R.
Table 4. Cramér’s V and Tschuprow’s T for food classes, example dataset from R.
TypeProductCramér’s V (95% CI)Tschuprow’s T (95% CI)
SoyMilkMeat
Vegan47001.00 (0.81, 1.00)0.84 (0.68, 1.00)
Not-Vegan01212
Table 5. Effect size Cohen’s w for variables with different numbers of categories and distributions.
Table 5. Effect size Cohen’s w for variables with different numbers of categories and distributions.
Observed CountsExpected ProportionCohen’s w (95% CI)
90/100.5/0.50.80 (0.61, 1.00)
90/100.35/0.651.15 (0.99, 1.36)
5/10/80/50.25/0.25/0.25/0.251.27 (1.10, 1.73)
Table 6. Effect size Fei for variables with different numbers of categories and distributions.
Table 6. Effect size Fei for variables with different numbers of categories and distributions.
Observed CountsExpected ProportionFei (95% CI)
90/100.5/0.50.80 (0.64, 1.00)
90/100.35/0.650.85 (0.73, 1.00)
5/10/80/50.25/0.25/0.25/0.250.73 (0.64, 1.00)
Table 7. Effect size for χ 2 tests for differently sized contingency tables.
Table 7. Effect size for χ 2 tests for differently sized contingency tables.
TestTable SizeEffect Size
χ 2 test for independence2-by-2 ϕ
Larger than 2-by-2 V or T
(Reduces to ϕ when table is 2-by-2)
χ 2 test for goodness-of-fit2 classes
with uniform null distribution
w
More than 2 classes
and/or
non-uniform null distribution
פ
(Reduces to w when there are
2 classes with uniform null dist).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ben-Shachar, M.S.; Patil, I.; Thériault, R.; Wiernik, B.M.; Lüdecke, D. Phi, Fei, Fo, Fum: Effect Sizes for Categorical Data That Use the Chi-Squared Statistic. Mathematics 2023, 11, 1982. https://doi.org/10.3390/math11091982

AMA Style

Ben-Shachar MS, Patil I, Thériault R, Wiernik BM, Lüdecke D. Phi, Fei, Fo, Fum: Effect Sizes for Categorical Data That Use the Chi-Squared Statistic. Mathematics. 2023; 11(9):1982. https://doi.org/10.3390/math11091982

Chicago/Turabian Style

Ben-Shachar, Mattan S., Indrajeet Patil, Rémi Thériault, Brenton M. Wiernik, and Daniel Lüdecke. 2023. "Phi, Fei, Fo, Fum: Effect Sizes for Categorical Data That Use the Chi-Squared Statistic" Mathematics 11, no. 9: 1982. https://doi.org/10.3390/math11091982

APA Style

Ben-Shachar, M. S., Patil, I., Thériault, R., Wiernik, B. M., & Lüdecke, D. (2023). Phi, Fei, Fo, Fum: Effect Sizes for Categorical Data That Use the Chi-Squared Statistic. Mathematics, 11(9), 1982. https://doi.org/10.3390/math11091982

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop