Skip to main content
A multi-variant test (sometimes written A/B/n) compares one control variant with two or more test variants in a single experiment. Instead of running several A/B tests in sequence you try every idea in parallel.

Why run multi-variant tests?

  • Speed. You discover the best setting faster because every variant is evaluated at the same time. There will be less time to gather enough samples than if you tested 2 or more variants sequentially.
  • Fair comparison. All variants share the same traffic window, removing calendar effects like holidays or seasonal campaigns.
  • Less engineering overhead. One set-up, one end date, one set of results to review.

Limitations and trade-offs

  • Traffic is divided. With more variants each one receives a smaller share of users, so it takes longer to collect enough data. This is still less time and risk than running sequential tests, however.
  • Statistical penalty. Every additional comparison increases the risk of a β€œlucky” winner. We correct for this (see below) by making the confidence threshold stricter.

Number of variants

We limit customers to 5 variants (1 control and 4 test variants). This limit is in place because a test with too many variants extends the test duration too far.

How we keep results reliable

By default, Algolia applies the Benjamini-Hochberg (BH) method. You can also choose the more conservative Bonferroni method:
MethodCorrected threshold
Benjamini-Hochberg (default)Ξ±α΅’ = (i / m) Γ— 0.05
BonferroniΞ± = 0.05 / m
where m is the number of comparisons and i the rank of the p-value when they’re sorted from smallest to largest. A comparison is marked Confident when its p-value is smaller than (or equal to) its corrected Ξ±. With Bonferroni, if there’s only one test variant (m = 1) the threshold stays at 0.05 - just like a regular two-variant test.
⌘I