Statistics

Statistical Power

Statistical power is the probability that a test correctly rejects a false null hypothesis.

Explanation

Statistical power measures the ability of a hypothesis test to detect an effect when one truly exists. It represents the probability of avoiding a Type II error, where researchers fail to reject a false null hypothesis. Power ranges from 0 to 1, with higher values indicating greater sensitivity to detect real differences or relationships in data. Researchers typically aim for power of 0.80 or higher, meaning an 80% chance of detecting a true effect. Power depends on four interconnected factors: the significance level (alpha), sample size, effect size, and the variability within the data. Larger sample sizes increase power, as do larger effect sizes and lower alpha levels. Statisticians, researchers, and quality control professionals use power analysis during study planning to determine the sample size needed to detect meaningful effects. Understanding power helps prevent costly studies that lack sufficient sensitivity to find important results.

Formula
Power = 1 - β, where β is the probability of Type II error
Power is calculated as one minus the Type II error rate, where Type II error is the false negative rate.

Example

A pharmaceutical company designs a trial to test if a new drug reduces cholesterol by 10 mg/dL compared to placebo. If they recruit 200 patients per group with alpha = 0.05, their power calculation shows 0.85 probability of detecting this effect. However, if they only recruit 50 patients per group, power drops to 0.45, meaning they have less than a 50% chance of finding the effect even if it truly exists. By increasing their sample to 200 per group, they gain confidence their study will detect the real benefit.

Key points
  • Power ranges from 0 to 1; 0.80 is the conventional target in most research fields
  • Type II error (β) and power are complementary: Power = 1 - β
  • Sample size, effect size, significance level, and variability all influence statistical power
  • Underpowered studies risk missing real effects, wasting resources on inconclusive research

Frequently asked questions

Why is statistical power important?
Power determines whether your study has sufficient sensitivity to detect a real effect. Low power increases the risk of Type II error, where you miss important findings. This wastes time and resources on inconclusive research. Adequate power protects your research validity and credibility.
What's the difference between power and significance level?
Significance level (alpha) is the probability of rejecting a true null hypothesis (Type I error). Power is the probability of correctly rejecting a false null hypothesis. You choose alpha before the study; power depends on alpha, sample size, and effect size.
How do I increase statistical power in my study?
You can increase power by: using a larger sample size, detecting a larger effect size, setting a higher significance level (higher alpha), or reducing data variability through better measurement methods. Sample size is the most practical lever researchers typically control.
What does 80% power mean?
Eighty percent power means you have an 80% probability of detecting a true effect if it exists, and a 20% probability of missing it (Type II error). This 80/20 balance reflects the convention that missing a true effect is four times worse than falsely claiming one.