In their Analysis article (Power failure: why small sample size undermines the reliability of neuroscience. Nature Rev. Neurosci. 14, 365–376 (2013))1, Button et al. state that insufficient statistical power owing to insufficient sample size undermines the reliability of findings in neuroscience. The authors rely on an earlier publication2 for this statement. Although it is obvious that a small sample size increases the risk of missing an existing effect, these authors question the reliability of significant findings in cases of insufficient power.

In both papers, concepts from diagnostic testing are applied to statistical testing in basic research. In this different context, a test's positive predictive value (PPV) turns into the reliability of a significant statistical effect. However, the size of the critical effect of test sensitivity (or statistical power) on the PPV (or the reliability of significant findings) depends on prevalence (or the odds of effects among tested effects) and becomes negligible if prevalence is high (Box 1).