Sample size calculator

Minimum Detectable Effect (%) The relative minimum difference in conversion rate between the test and control group we want to observe.

Hypothesis The following values are set to their optimal defaults, but feel free to play with them to see how they affect the exposures.

One-sided Test (Recommended)
Used to determine if the test variation is better than the control (Recommended)

Two-sided Test
Used to determine if the test variation is different than the control A/B Split Ratio 0.5 is an evenly split 50/50 test Test vs. Control

Significance (α) The probability that a statistically significant difference is detected even though no actual difference exists

Range can be 0.01-0.1

Statistical Power (1 - β) The probability that the minimum detectable effect will be detected (assuming it exists)

Range can be 0.65-0.95 Minimum Detectable Effect

198

CONTROL SIZE

198

TOTAL SAMPLE SIZE

396

Setting up a proper A/B experiment involves selecting the appropriate sample size.

This ensures the test is adequately powered to detect the change while minimizing statistical noise. In statistical terms, we want to detect the minimum detectable effect (MDE) with statistical confidence.

How to use Statsig’s sample size calculator:

Determine your baseline conversion rate

This is the preexisting or expected conversion rate of your control group. For example, if you are studying the effectiveness of a new UI, the baseline conversion rate would be the percentage of people in your audience who are expected to use the desired features without being exposed to the new UI.

Choose a minimum detectable effect

This is the smallest difference in the behavior or outcome that you want to be able to consistently detect with your study. For example, if you are studying the effectiveness of a new UI, the minimum detectable effect would be the smallest increase in usage rate that you want to be able to detect.

Input your values

The calculator will use these values to determine the optimal sample size for your study. Its output will be the minimum viable test size and minimum viable control size in order to consistently achieve statistically significant results. How consistent will be based on your statistical power—an advanced input we’ll discuss below:

Advanced calculations:

The calculator is automatically set to optimal defaults, but you can adjust the advanced settings to see how they impact your results.

Hypothesis

A one-sided test (recommended) is used to determine if the test variation is better than the control. On the other hand, a two-sided test is used to determine if the test variation is simply different than the control.

A/B split ratio

The default value of 0.5 sets the A/B test to split test users versus control users 50%/50%. Standard A/B test procedure is usually 50-50, but feel free to tune this to your own special circumstances.

Significance (α)

The significance value is the probability that a statistically significant difference is detected, given no actual difference exists. This is commonly referred to as a type 1 error or a false positive.

The default value is set to 0.05, but can be set within a range of 0.01 to 0.1. Lowering this number allows for higher confidence that the difference observed isn’t due to chance, but requires larger sample sizes.

Statistical Power (1 - β)

The statistical power is the probability that the minimum detectable effect will be detected, assuming it exists. The default value is set to 0.8, but can be adjusted between 0.65 and 0.95. Again, a higher value reduces the frequency of false negatives, but requires a larger sample size.

Join the #1 experimentation community

Connect with like-minded product leaders, data scientists, and engineers to share the latest in product experimentation.