Difference between revisions of "2.1.6 Sample size and power analysis"
(→A. Background & Definitions) |
(→A. Background & Definitions) |
||
Line 13: | Line 13: | ||
Balancing sample size, effect size and power is critical to good study design. When the power is low, only large effects can be detected, and negative results cannot be reliably interpreted. The consequences of low power are particularly dire in the search for high-impact results, when the researcher may be willing to pursue low-likelihood hypotheses for a groundbreaking discovery (see Fig. 1 in [https://www.nature.com/articles/nmeth.2738 Krzywinski & Altman 2013]). Ensuring that sample sizes are large enough to detect the effects of interest is an essential part of study design. | Balancing sample size, effect size and power is critical to good study design. When the power is low, only large effects can be detected, and negative results cannot be reliably interpreted. The consequences of low power are particularly dire in the search for high-impact results, when the researcher may be willing to pursue low-likelihood hypotheses for a groundbreaking discovery (see Fig. 1 in [https://www.nature.com/articles/nmeth.2738 Krzywinski & Altman 2013]). Ensuring that sample sizes are large enough to detect the effects of interest is an essential part of study design. | ||
− | Studies with inadequate power are a waste of research resources and arguably unethical when subjects are exposed to potentially harmful or inferior experimental conditions. Addressing this | + | Studies with inadequate power are a waste of research resources and arguably unethical when subjects are exposed to potentially harmful or inferior experimental conditions. Addressing this shortcoming is a priority — [[https://www.nature.com/documents/nr-reporting-summary.pdf the Nature Publishing Group's reporting checklist for life sciences]] includes as the first question: “How was the sample size chosen to ensure adequate power to detect a pre-specified effect size?” |
Statistical power analysis exploits the relationships among the four variables involved in statistical inference: sample size (N), significance criterion (α), effect size (ES), and statistical power. For any statistical model, these relationships are such that each is a function of the other three. | Statistical power analysis exploits the relationships among the four variables involved in statistical inference: sample size (N), significance criterion (α), effect size (ES), and statistical power. For any statistical model, these relationships are such that each is a function of the other three. |
Revision as of 06:28, 15 October 2020
UNDER CONSTRUCTION
A. Background & Definitions
Statistical power is defined as the probability of detecting as statistically significant a clinically or practically important difference of a pre-specified size, if such a difference truly exists. Formally, power is equal to 1 minus the Type II error rate (beta or ß). The Type II error rate is the probability of obtaining a non-significant result when the null hypothesis is false — in other words, failing to find a difference or relationship when one exists.
Balancing sample size, effect size and power is critical to good study design. When the power is low, only large effects can be detected, and negative results cannot be reliably interpreted. The consequences of low power are particularly dire in the search for high-impact results, when the researcher may be willing to pursue low-likelihood hypotheses for a groundbreaking discovery (see Fig. 1 in Krzywinski & Altman 2013). Ensuring that sample sizes are large enough to detect the effects of interest is an essential part of study design.
Studies with inadequate power are a waste of research resources and arguably unethical when subjects are exposed to potentially harmful or inferior experimental conditions. Addressing this shortcoming is a priority — [the Nature Publishing Group's reporting checklist for life sciences] includes as the first question: “How was the sample size chosen to ensure adequate power to detect a pre-specified effect size?”
Statistical power analysis exploits the relationships among the four variables involved in statistical inference: sample size (N), significance criterion (α), effect size (ES), and statistical power. For any statistical model, these relationships are such that each is a function of the other three.
B. Guidance & Expectations
General advice - DO:
- Whenever possible, seek professional biostatistician support to estimate sample size.
- Use power prospectively for planning future studies.
- Put science before statistics. It is easy to get caught up in statistical significance and such; but studies should be designed to meet scientific goals, and you need to keep those in sight at all times (in planning and analysis). The appropriate inputs to power/sample-size calculations are effect sizes that are deemed scientifically important, based on careful considerations of the underlying scientific (not statistical) goals of the study. Statistical considerations are used to identify a plan that is effective in meeting scientific goals – not the other way around.
- Do pilot studies. Investigators tend to try to answer all the world’s questions with one study. However, you usually cannot do a definitive study in one step. It is far better to work incrementally. A pilot study helps you establish procedures, understand and protect against things that can go wrong, and obtain variance estimates needed in determining sample size. A pilot study with 20-30 degrees of freedom for error is generally quite adequate for obtaining reasonably reliable sample-size estimates.
- Effect size should be specified on the actual scale of measurement, not on a standardized scale.
- Generate sample size estimates for a range of power and effect size values to explore the gains and and losses in power or detectable effect size due to increasing or decreasing n.
General advice - DO NOT:
- Avoid using the definition of “small,” “medium,” or “large” effect size based on Cohen's d of .20, .50, or .80, respectively. Cohen's assessments are based on an extensive survey of statistics reported in the literature in the social sciences and may not apply to other fields of science. Further, this method uses a standardized effect size as the goal. Think about it: for a “medium” effect size, you’ll choose the same n regardless of the accuracy or reliability of your instrument, or the narrowness or diversity of your subjects. Clearly, important considerations are being ignored here. “Medium” is definitely not the message!
- Retrospective power calculations should generally be avoided, because they add no new information to an analysis (i.e. avoid using observed power to interpret the results of the statistical test). You’ve got the data, did the analysis, and did not achieve “significance.” So you compute power retrospectively to see if the test was powerful enough or not. This is an empty question. Of course it wasn’t powerful enough – that’s why the result isn’t significant. Power calculations are useful for design, not analysis.
We begin by setting the values of type I error (a) and power (1 – b) to be statistically adequate: traditionally 0.05 and 0.80, respectively. We then determine n on the basis of the smallest effect we wish to measure. If the required sample size is too large, we may need to reassess our objectives or more tightly control the experimental conditions to reduce the variance.
C. Resources
Tools to sample size estimation:
- G*Power
- WISE power tutorial
- JAVA applets for power and sample size
- Computation of sample sizes @Psychometrica
Educational instruments and resources:
Useful literature (for non-statisticians):
Practical advice on sample size estimation by Russell Lenth
Useful literature:
Power in various ANOVA designs by Joel Levin
- https://www.ncbi.nlm.nih.gov/books/NBK43321/
- http://davidmlane.com/hyperstat/power.html
- http://powerandsamplesize.com/Calculators/Test-1-Mean/1-Sample-Equality
Guidelines on reporting of sample size (in vivo research):
back to Toolbox
Next item: 2.1.7 Blinding