Statistics
False Negatives
.webp)
The result of a statistical test that fails to detect a real effect, wrongly concluding that there is no significant difference between two variants, when in fact there is. In the context of CRO andA/B testing, a false negative can lead to the rejection of a variation that is nevertheless performing well, simply because the test lacked power (e.g.: sample too small, insufficient duration, MDE too low).
False negatives are often contrasted with false positives, which falsely indicate that an effect is significant when it is not. The balance between these two errors depends on the statistical parameters chosen (confidence level, power), and must be carefully calibrated to avoid erroneous decisions in optimization tests.