The concept of effect size has become very important in educational and behavioral research. However, the term, effect size, is used in many contexts. The three most common contexts are setting an effect size as a part of determining needed sample sizes to a achieve desired power level, converting various effect sizes and measures of association into a common metric for meta-analysis, and reporting the effect size post-hoc as an indication of the practical significance of group differences in experimental or quasi-experimental studies. The point this paper is to make is that arbitrary selections of effect size standards cannot be meaningful without knowledge and accounting for the critical characteristics of the situation relative to the number and size of samples or the degrees of freedom.
Monte Carlo methods were used to generate the data for this research using random normal deviates as the basis for sample means to be compared using one-way fixed-effects ANOVA. Standardized effect sizes were generated for 10,000 replications within each combination of number of groups from 2 to 12 and sample sizes of 5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 55, 60, 70, 80, 90, 100, 150, 200, 250, and 500, resulting in 2,200,000 total replications. The standardized effect size was computed as the range of means divided by the root mean square error.
The results make clear that the current arbitrary and absolute criteria proposed by Cohen are far from being sensitive to the variation in standardized effect sizes as functions of the number and size of samples. The paper also demonstrates how this result relates to the three common uses of standardized effect size.