ðŸŽ‰Be Part of the Best in Test Awards! ðŸŽ‰ Find out more

In this article Georgi Georgiev, statistical expert and founder of Analytics-Toolkit.com, walks you through the benefits and drawbacks of allocating unequal traffic across your A/B test variants.

Unequal allocation is most easily understood when contrasted with **equal allocation**.

Equal allocation is the default approach in which we allocate equal amounts of users to each test variant in an A/B test.

If there is one variant, and one control group, we assign 50% of users to the control group and 50% to the variant.

If there are 3 variants and one control, we would allocate 25% of users to each of them, under equal allocation.

With **unequal allocation,** one decides to allocate a smaller or greater proportion of users to either the control or the variants as a whole or to particular variants in a multivariate test (MVT).

For example, one can allocate 80% of traffic to the control and 20% of traffic to the test variant, or 40% to the control and 20% to each variant in an MVT with 3 variants.

People consider unequal allocation for different **reasons**, but the main concern is the cost/benefit analysis for the duration of the test.

With an 80/20 control/variant allocation, one risks less exposure to an inferior variant during the test, compared to a test of the same duration, but with equal 50/50 allocation.

With a 20/80 control/variant allocation, one usually aims to get more users to experience the tested variant during the test.

Some appear to be under the impression that by running a 20/80 control/variant test they can reach a conclusion faster, while others believe that running a 80/20 control/variant test means they can reach the same conclusion, but with less risk.

Below I’ll address each of these assumptions in detail.

Some people seem to think that running an A/B test with unequally distributed traffic among the variants will allow them to reach conclusions quicker.

This, however, not the case.

The reason is that statistical significance calculations and confidence interval calculations depend highly on the number of users allocated to each variant, not the test as a whole.

In a significance calculation within-group variance is compared to between-group variance. If one group contains data for significantly fewer users, the variance of whatever outcome metric you are using will be higher for that group, leading to higher uncertainty and thus higher significance level (lower confidence level).

Similarly, the confidence interval for absolute or relative difference will be wider, reflecting the greater uncertainty associated with the estimate.

For example, if during the test, the control converts at 3% and the test variant at 3.5%, with 80/20 unequal allocation we get this result:

While if the same conversion rates had been observed with 50/50 allocation, we would have almost 2 times less uncertainty (~5% instead of ~10%):

In proper terms, a relative difference as big as 20%, or bigger, would be observed with probability ~10% in the first case if there was no improvement, while in the second case the probability of observing such a difference is only about 5%. The 95% confidence interval for relative change spans from -5.3% to +∞ in the first case and from -0.3% to +∞ in the second.

Some people find it easier to visualize this in terms of confidence intervals for the conversion rate in the two groups instead of the confidence interval for the difference in conversion rates. Here are how the confidence intervals look like for the two groups with unequal vs equal allocation of 10,000 users:

As you can see, with unequal allocation we have greater accuracy for the conversion rate of the control group, compared to the accuracy for the control group with equal allocation; however, it is entirely negated by the increased inaccuracy of the variant estimation in the unequal case, versus the equal one.

There is no free lunch in life and statistics are no exception. Gaining accuracy for one group comes at the cost of reduced accuracy for the other, everything else being equal.

This is the reason why anything other than equal allocation leads to less efficient tests. **A/B tests with unequal allocation will require significantly longer time to complete compared to equal allocation. **The same, of course, holds for MVT tests.

Even after understanding this concept, many are still tempted to resort to unequal allocation if they monitor a test and discover that one variant is performing significantly worse or significantly better than another.

In such cases, practitioners may be tempted to push more traffic, or to restrict the amount of traffic allocated to a variant. This is a really bad practice, if done ad-hoc and without proper statistical methods.

It has similar effects to monitoring a test and then interpreting the results as if no monitoring (peeking) occurred: the nominal p-value no longer reflects the actual p-value, the nominal accuracy of confidence intervals and other estimates no longer holds and in it ultimately leads to a severe increase in the rate of false positives in a testing program.

First, a proper way to monitor the outcome of an A/B test during its duration is to employ a proper sequential testing method, such as AGILE. In such cases one constructs efficacy and futility stopping boundaries and only stops the test if one of the boundaries is crossed. However, most methods only hold if the allocation does not change during the entire duration of the test.

There is a class of methods for “adaptive sequential design”, which allows for changing the allocation between test variants on the go and also for dropping variants mid-way and even adding new variants to a running A/B test. They are quite more complex computationally and can be more difficult to interpret. What’s worse is that contrary to naïve intuition, such methods are **not more efficient than methods based around equal allocation**.

In fact, the opposite has been proven: for any adaptive sequential method there exists an equally-efficient or more efficient non-adaptive sequential alternative. To quote Tsiatis & Mehta’s 2003 paper “On the inefficiency of the adaptive design for monitoring clinical trials”: *“For any adaptive design, one can always construct a standard group-sequential test based on the sequential likelihood ratio test statistic that, for any parameter value in the space of alternatives will reject the null hypothesis earlier with higher probability, and, for any parameter value not in the space of alternatives will accept the null hypothesis earlier with higher probability.”*

Further confirmation can be found in Lee, Chen & Yin’s 2012 “Worth adapting? Revisiting the usefulness of outcome-adaptive randomization”: *“In summary, equal randomization maintains balanced allocation throughout the trial and reaches the specified statistical power with a smaller number of patients in the trial [compared to adaptive randomization]“.*

While unequal allocation and especially mid-test unequal allocation has an intuitive appeal, it is in fact detrimental to the efficiency of an A/B test when done properly and has negative effects on the validity of the results to other kinds of malpractices such as unaccounted peeking when done improperly.

Even if your tools allow for unequal allocation of users between test variants your best bet is to actually stick to equal allocation from the beginning till the end of an A/B test.

If some of the terminology used in this article is unfamiliar you can also refer to the Analytics-Toolkit A/B testing glossary for detailed definitions.

**Disclaimer:** *While unequal allocation can technically be beneficial with more than one treatment, if one uses a naive p-value and confidence interval calculation, the efficiency gains obtained by using procedures like the Dunnett’s correction far outweigh such gains. Given that Dunnett’s can only be applied to an equal allocation design, one should stick to equal allocation in order to achieve best use of the data.*

Subscribe

0 Comments

Oldest

Join the Best in Test awards ceremony. Submit your best tests and see who wins the testing awards.

Free For Limited Time

A primer explaining the 4 different types of tests you can run, what they mean, and how you can use each to improve your competitive testing advantage.

Free For Limited Time

One of the most debated testing topics is how large does my sample size need to be to get trustworthy test results? Some argue samples of more than 120,000 visitors per variant are needed to begin to see trustworthy test results. Ishan Goel of VWO disagrees. What does he think is needed to get trustworthy test results? Listen to this webinar recording to find out.

Please enter your login info:

Forgot Password? Click here to reset.

Don't have an account? Create one!

Create FREE Account

Enter the username or e-mail you used in your profile. A password reset link will be sent to you by email.

Already have an account? Login!

Login to your Account

Subscribe to the newsletter to get free, new A/B test case studies sent to you every other week.

Join thousands of other digital marketers from organizations like these:

Annual Plan

normally $997/yr

$

497

/yr

(billed annually )

3 team members for the price of 2

Unlimited access to all webinar, tests and resources

Exclusive offer - Get a free Conversion Audit on your site

Best Value - Save 56%

Annual Plan

normally $697/yr

$

247

/yr

(billed annually )

Save big! Enjoy 56% off by paying once, annually

Complete access to the full library of 50+ archived tests

Exclusive access to all webinars, articles, and resources

Monthly Plan

normally $79/mo

$

47

/mo

(billed monthly )

No obligation, try monthly and enjoy

Full access to the complete library of 50+ archived tests

Complete access to all webinars, articles, and resources

All are prices in $USD

A/B testing question?

Find money-making test ideas

Get actionable insights

Understand statistical significance

Other testing questions

Searching database. AI is generating your answer. This process may take a few seconds. Hang tight.

ðŸ‘‹ Use the AI-driven chatbot to answer any A/B testing question