These functions are convenience functions to convert F and t test statistics to partial Eta- ($$\eta$$), Omega- ($$\omega$$) Epsilon- ($$\epsilon$$) squared (an alias for the adjusted Eta squared) and Cohen's f. These are useful in cases where the various Sum of Squares and Mean Squares are not easily available or their computation is not straightforward (e.g., in liner mixed models, contrasts, etc.). For test statistics derived from lm and aov models, these functions give exact results. For all other cases, they return close approximations.
See Effect Size from Test Statistics vignette.

F_to_eta2(f, df, df_error, ci = 0.95, alternative = "greater", ...)

t_to_eta2(t, df_error, ci = 0.95, alternative = "greater", ...)

F_to_epsilon2(f, df, df_error, ci = 0.95, alternative = "greater", ...)

t_to_epsilon2(t, df_error, ci = 0.95, alternative = "greater", ...)

F_to_eta2_adj(f, df, df_error, ci = 0.95, alternative = "greater", ...)

t_to_eta2_adj(t, df_error, ci = 0.95, alternative = "greater", ...)

F_to_omega2(f, df, df_error, ci = 0.95, alternative = "greater", ...)

t_to_omega2(t, df_error, ci = 0.95, alternative = "greater", ...)

F_to_f(
f,
df,
df_error,
ci = 0.95,
alternative = "greater",
squared = FALSE,
...
)

t_to_f(t, df_error, ci = 0.95, alternative = "greater", squared = FALSE, ...)

F_to_f2(
f,
df,
df_error,
ci = 0.95,
alternative = "greater",
squared = TRUE,
...
)

t_to_f2(t, df_error, ci = 0.95, alternative = "greater", squared = TRUE, ...)

## Arguments

df, df_error Degrees of freedom of numerator or of the error estimate (i.e., the residuals). Confidence Interval (CI) level a character string specifying the alternative hypothesis; Controls the type of CI returned: "greater" (default) or "less" (one-sided CI), or "two.sided" (default, two-sided CI). Partial matching is allowed (e.g., "g", "l", "two"...). See One-Sided CIs in effectsize_CIs. Arguments passed to or from other methods. The t or the F statistics. Return Cohen's f or Cohen's f-squared?

## Value

A data frame with the effect size(s) between 0-1 (Eta2_partial, Epsilon2_partial, Omega2_partial, Cohens_f_partial or Cohens_f2_partial), and their CIs (CI_low and CI_high). (Note that for $$\omega_p^2$$ and $$\epsilon_p^2$$ it is possible to compute a negative number; even though this doesn't make any practical sense, it is recommended to report the negative number and not a 0).

## Details

These functions use the following formulae:
$$\eta_p^2 = \frac{F \times df_{num}}{F \times df_{num} + df_{den}}$$
$$\epsilon_p^2 = \frac{(F - 1) \times df_{num}}{F \times df_{num} + df_{den}}$$
$$\omega_p^2 = \frac{(F - 1) \times df_{num}}{F \times df_{num} + df_{den} + 1}$$
$$f_p = \sqrt{\frac{\eta_p^2}{1-\eta_p^2}}$$

For t, the conversion is based on the equality of $$t^2 = F$$ when $$df_{num}=1$$.

### Choosing an Un-Biased Estimate

Both Omega and Epsilon are unbiased estimators of the population Eta. But which to choose? Though Omega is the more popular choice, it should be noted that:

1. The formula given above for Omega is only an approximation for complex designs.

2. Epsilon has been found to be less biased (Carroll & Nordholm, 1975).

## Note

Adjusted (partial) Eta-squared is an alias for (partial) Epsilon-squared.

## Confidence (Compatibility) Intervals (CIs)

Unless stated otherwise, confidence (compatibility) intervals (CIs) are estimated using the noncentrality parameter method (also called the "pivot method"). This method finds the noncentrality parameter ("ncp") of a noncentral t, F, or χ2 distribution that places the observed t, F, or χ2 test statistic at the desired probability point of the distribution. For example, if the observed t statistic is 2.0, with 50 degrees of freedom, for which cumulative noncentral t distribution is t = 2.0 the .025 quantile (answer: the noncentral t distribution with ncp = .04)? After estimating these confidence bounds on the ncp, they are converted into the effect size metric to obtain a confidence interval for the effect size (Steiger, 2004).

For additional details on estimation and troubleshooting, see effectsize_CIs.

## CIs and Significance Tests

"Confidence intervals on measures of effect size convey all the information in a hypothesis test, and more." (Steiger, 2004). Confidence (compatibility) intervals and p values are complementary summaries of parameter uncertainty given the observed data. A dichotomous hypothesis test could be performed with either a CI or a p value. The 100(1 − α)% confidence interval contains all of the parameter values for which p > α for the current data and model. For example, a 95% confidence interval contains all of the values for which p > .05.

Note that a confidence interval including 0 does not indicate that the null (no effect) is true. Rather, it suggests that the observed data together with the model and its assumptions combined do not provided clear evidence against a parameter value of 0 (same as with any other value in the interval), with the level of this evidence defined by the chosen α level (Rafi & Greenland, 2020; Schweder & Hjort, 2016; Xie & Singh, 2013). To infer no effect, additional judgments about what parameter values are "close enough" to 0 to be negligible are needed ("equivalence testing"; Bauer & Kiesser, 1996).

## References

• Albers, C., & Lakens, D. (2018). When power analyses based on pilot data are biased: Inaccurate effect size estimators and follow-up bias. Journal of experimental social psychology, 74, 187-195. doi: 10.31234/osf.io/b7z4q

• Carroll, R. M., & Nordholm, L. A. (1975). Sampling Characteristics of Kelley's epsilon and Hays' omega. Educational and Psychological Measurement, 35(3), 541-554.

• Cumming, G., & Finch, S. (2001). A primer on the understanding, use, and calculation of confidence intervals that are based on central and noncentral distributions. Educational and Psychological Measurement, 61(4), 532-574.

• Friedman, H. (1982). Simplified determinations of statistical power, magnitude of effect and research sample sizes. Educational and Psychological Measurement, 42(2), 521-526. doi: 10.1177/001316448204200214

• Mordkoff, J. T. (2019). A Simple Method for Removing Bias From a Popular Measure of Standardized Effect Size: Adjusted Partial Eta Squared. Advances in Methods and Practices in Psychological Science, 2(3), 228-232. doi: 10.1177/2515245919855053

• Morey, R. D., Hoekstra, R., Rouder, J. N., Lee, M. D., & Wagenmakers, E. J. (2016). The fallacy of placing confidence in confidence intervals. Psychonomic bulletin & review, 23(1), 103-123.

• Steiger, J. H. (2004). Beyond the F test: Effect size confidence intervals and tests of close fit in the analysis of variance and contrast analysis. Psychological Methods, 9, 164-182.

eta_squared() for more details.

Other effect size from test statistic: chisq_to_phi(), t_to_d()

## Examples

# \donttest{
if (require("afex")) {
data(md_12.1)
aov_ez("id", "rt", md_12.1,
within = c("angle", "noise"),
anova_table = list(correction = "none", es = "pes")
)
}
#> ************
#> Welcome to afex. For support visit: http://afex.singmann.science/
#> - Functions for ANOVAs: aov_car(), aov_ez(), and aov_4()
#> - Methods for calculating p-values with mixed(): 'S', 'KR', 'LRT', and 'PB'
#> - 'afex_aov' and 'mixed' objects can be passed to emmeans() for follow-up tests
#> - NEWS: emmeans() for ANOVA models now uses model = 'multivariate' as default.
#> - Get and set global package options with: afex_options()
#> - Set orthogonal sum-to-zero contrasts globally: set_sum_contrasts()
#> - For example analyses see: browseVignettes("afex")
#> ************
#>
#> Attaching package: ‘afex’
#> The following object is masked from ‘package:lme4’:
#>
#>     lmer
#> Anova Table (Type 3 tests)
#>
#> Response: rt
#>        Effect    df     MSE         F  pes p.value
#> 1       angle 2, 18 3560.00 40.72 *** .819   <.001
#> 2       noise  1, 9 8460.00 33.77 *** .790   <.001
#> 3 angle:noise 2, 18 1160.00 45.31 *** .834   <.001
#> ---
#> Signif. codes:  0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘+’ 0.1 ‘ ’ 1
# compare to:
(etas <- F_to_eta2(
f = c(40.72, 33.77, 45.31),
df = c(2, 1, 2),
df_error = c(18, 9, 18)
))
#> Eta2 (partial) |       95% CI
#> -----------------------------
#> 0.82           | [0.66, 1.00]
#> 0.79           | [0.49, 1.00]
#> 0.83           | [0.69, 1.00]
#>
#> - One-sided CIs: upper bound fixed at (1).

if (require(see)) plot(etas)

if (require("lmerTest")) { # for the df_error
fit <- lmer(extra ~ group + (1 | ID), sleep)
# anova(fit)
# #> Type III Analysis of Variance Table with Satterthwaite's method
# #>       Sum Sq Mean Sq NumDF DenDF F value   Pr(>F)
# #> group 12.482  12.482     1     9  16.501 0.002833 **
# #> ---
# #> Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1

F_to_eta2(16.501, 1, 9)
F_to_omega2(16.501, 1, 9)
F_to_epsilon2(16.501, 1, 9)
F_to_f(16.501, 1, 9)
}
#>
#> Attaching package: ‘lmerTest’
#> The following object is masked from ‘package:lme4’:
#>
#>     lmer
#> The following object is masked from ‘package:stats’:
#>
#>     step
#> Cohen's f (partial) |           95% CI
#> --------------------------------------
#> 1.35                | [0.57,      Inf]
#>
#> - One-sided CIs: upper bound fixed at (Inf).

## Use with emmeans based contrasts
## --------------------------------
if (require(emmeans)) {
warp.lm <- lm(breaks ~ wool * tension, data = warpbreaks)

jt <- joint_tests(warp.lm, by = "wool")
F_to_eta2(jt$F.ratio, jt$df1, jt\$df2)
}