In this article I cover the method required to calculate statistical significance for non-binomial metrics such as average revenue per user, average order value, average sessions per user, average session duration, average pages per session, and others. The focus is on A/B testing in the context of conversion rate optimization, landing page optimization and e-mail marketing optimization, but is applicable to a wider range of practical cases.

## Binomial vs. Non-binomial metrics

Let’s start with the basic definitions. A **binomial metric** is one that can only have two possible values: true or false, yes or no, present or absent, action or no action. In the context of A/B testing these are all metrics that can are expressed as “**rates**“: goal conversion rate, e-commerce transaction rate, bounce rate, and, rarely, exit rate. The goal conversion rate can measure all kinds of actions: lead form completion rate, phone call rate, add to cart rate, checkout start rate, etc.

A **non-binomial metric** is a metric in which the possible range of values is not limited to two possible states. These metrics are usually continuous, spanning from zero to plus infinity, or from minus infinity to plus infinity. These are usually “per user”, “per session” or “per order” metrics, such as: average revenue per user (ARPU), average order value (AOV), average session duration, average pages per session, average sessions per user, average products per purchase, etc.

So, in short, the major difference comes from **how the possible values are distributed**. In the binomial case they are strictly enumerated, while in the non-binomial case they can cover all real numbers. This is the main reason why can’t we use the same statistical calculators for binomial and non-binomial metrics, or, more precisely, the reason why most significance calculators and A/B testing platforms support only binomial data. Fundamentally the calculations are the same, but not according to some, who raise the additional objection that revenue-based data violates the normality assumptions, throwing away the validity of classical tests such as the t-test.

## Do revenue-based metrics violate the normality assumption?

While searching for information on how to calculate statistical significance for revenue-based metrics, one can find at least a couple of examples of blog posts and discussions in which an objection to the use of classical t-tests is raised that is, in short: revenue-based metrics do not follow a normal distribution, thus classical tests are not appropriate. Here are some graphs that are used to illustrate this point:

Many of them make the correct observation that in the case of measuring average revenue per user we need to essentially account for **two types of variance**. First – the variance of the conversion rate, that is, did the user order or not. Second – the variance in average order value. The first we know to be binomial, but the second is not. User conversion rate multiplied by the average order value gives us the average revenue per user.

Different alternatives are proposed to handle the issue, like the Mann-Whitney-Wilcoxon rank-sum test, methods from so-called robust statistics, as well as bootstrapping resampling methods. These approaches are fairly complex and while they make less or no assumptions about the distribution, they are much more difficult to interpret. While they have their place and time, I don’t think their introduction is necessary for dealing with non-binomial metrics in A/B testing.

Let me illustrate why by first bringing to your attention the following graph of a well-known family of metrics:

Can you recognize the distribution of which metric has the above shape?

If you guessed correctly, then yes, this can be any rate-based metric, such as add to cart conversion rate, e-commerce transaction conversion rate, bounce rate, etc. So, it doesn’t have a normal distribution, but why are we justified then in using classic tests that assume normality?

The reason is simple:

it is the sample statistic error distribution that matters, not the sample data distribution

Sample what? OK, let me explain…

## In-sample data distribution vs. sample statistic error distribution

Here is a simulated revenue per user distribution:

*(click for large version)*

The above is revenue data from 10,000 simulated users. On the x-axis we see the revenue bucket, on they y-axis: how many users are in each bucket. It is just like the graphs cited above: the distribution is extremely heavily skewed to the left, since the vast majority of users don’t purchase anything. About 5% of users purchased something.

Looking at the zoomed-in graph, where 0-revenue users are excluded, we see that they are, too, skewed to the left. Much more users place orders in the 20-200 range than in the 200-1000 range. The particular distribution will vary depending on the type of business – online shops selling small items versus ones that sell high-value products will have very different distributions. SaaS businesses will have data that is concentrated around specific revenue levels, corresponding to the tiers they offer. That is beyond the point, though, the data is certainly representative of a vast array of real-world scenarios and the **in-sample data distribution is certainly non-normal**.

However, significance calculations deal with sample statistics, not with individual data points. Some possible statistics in this case are: the data mean, median and mode. We usually use the average as the most actionable statistic, but the median can also be useful if we want to learn the transaction revenue that is paid by the highest number of our visitors, that is, how much does our average visitor pays. The mean in this case is $10.02. So, if we had two such samples, one with mean 10.02 and another with a mean of 10.3, statistical significance tests would measure the difference between these two values, not the difference between each and every individual user.

So, the normality requirement, or assumption, needs to hold for the **distribution of the sample statistic**, not for the distribution of the in-sample data that you see here. In other words, we are interested in the distribution of the error of the mean, not the data itself. In order to check the normality assumption, I generated 10,000 simulations with 10,000 users in each, recorded their mean and plotted it as a histogram. The x-axis is the average revenue per user range, while the y-axis is how many of the simulations ended up in that range, out of the total of 10,000.

Since all revenue values generated were random, the above plot illustrates the **random error of the mean** that would be inherent to any measurement of average revenue per user. This is how the noise in your data is shaped, which is what you need to compensate for before you can extract the signal. And its distribution is quite close to a normal curve. Of course, it is impossible to get a true normal curve with just 10,000 simulations, but the important thing is that the more simulations we run, the more the values would converge to a true normal distribution. This asymptotic convergence is what makes statistical tools work.

In other words, the Central Limit Theorem still holds, so **there is no need to use more or less exotic tools to work with continuous metrics like revenue per user**, average order value, average time on site and the like. However, just because the issue highlighted as central on most blog posts discussing this topic is not present doesn’t automatically mean that you can just use a calculator that only works with binomial data to calculate statistical significance and confidence intervals for non-binomial data.

## Calculating significance for non-binomial metrics in A/B testing

So, now that we have seen that we don’t need to worry about the normal distribution of our non-binomial metrics, what is holding us back from using just any significance calculators, same as we would for a conversion rate metric?

The unknownvarianceof the sample mean makes is impossible to use calculators that work fine with binomial data.

The variance of the sample mean can be calculated easily through a mathematical formula based on the proportion of observations in each sample and the total observations in the tested variants and control. However, for revenue and other non-binomial metrics we need to estimate the variance from the sample, before we can calculate significance. The process is fairly straightforward and looks like so:

- Extract user-level data (orders, revenue) or session-level data (session duration, pages per session) or order-level data (revenue, number of items) for the control and the variant
- Calculate the sample standard deviation of each
- Calculate the pooled standard error of the mean
- Use the SEM in any significance calculator / software that supports the specification of SEM in calculations

Different software will have different procedures for exacting user-level or session-level data. In Google Analytics you can use the user explorer report through an API tool, or a custom dimension that contains a user-ID (you can use the Google Analytics Client-ID for non-logged in users). The important thing to remember is that you need to end up with a table in which you have individual users/sessions/orders in one row, and the revenue (or another metric) for that particular user/session/order.

### 1. Estimating variance from a sample

Once you have above, the calculation is simple. You can do in Excel with a few fairly simple formulas:

So, the **standard error of the mean revenue per user** is estimated at **$0.69** from that sample. If your look at the graph for the simulated distribution of the means you will notice that the standard deviation of the mean is estimated at **$0.73** from those 10,000 means we simulated. Pretty close, meaning the estimate we drew from that one sample was a pretty good estimate indeed. If you notice the coloring of the graph, it corresponds to standard deviations in either direction. The 97.5% one-sided significance corresponds roughly to where the dark green area begins. Another way is to use an online tool, such as this standard deviation calculator, which conveniently also calculates the standard error of the mean (SEM).

### 2. Calculating significance for non-binomial metrics

In order to plan a test that has a non-binomial metric as a primary KPI you need to estimate the variance from historical data. The more – the better, but I’d suggest to not go too far back, maybe no more than a year since older data might have less predictive value than newer data due to changes in user behavior. The procedure described above should be used.

After a test is ended, you calculate SEM for all variants and the control, you can use any calculator or software tool that supports non-binomial metrics, that is, any t-test calculator that allows you to enter a standard deviation for the mean. What happens is that the pooled variance is calculated, using the combined data of all tested variants and the control, then the t-value and corresponding p-value is calculated using standard approaches. I believe the pooled variance from the test itself will have better predictive value than historical data and should therefore be preferred.

As with any other statistical significance calculation, you should not be peeking at your data or using reaching statistical significance as a stopping rule unless you are using a proper sequential testing methodology, for example the AGILE A/B testing method.

## Calculator for revenue-based and other non-binomial metrics

While there are multiple ways to calculate such data, including in Excel or R, we got feedback it would be a convenience to have all of these automated, so that the process is streamlined and faster to perform. Thus, we have just released an online calculator that allows you to copy/paste data for tens of thousands of lines of user-level, order-level, or session-level data, and it does all calculations automatically, outputting the corresponding p-value and confidence intervals for each variant. The p-values are adjusted for multiple comparisons using the Dunnett’s method. You can try it for free by signing up for Analytics-Toolkit.com and going to our statistical significance calculator, where you should chose to input **Revenue & Other** data.

Our statistical calculators now also offer the ability to estimate the **required sample size** when comparing means (difference of means), so now you have all basic tooling necessary to perform the statistical design and evaluation for a simple fixed sample size test with non-binomial data. Currently it only supports tests with one variant and control in the case of revenue, AOV, ToS and other such metrics. Future improvements may be introduced to handle more variants.

## 6 Comments

Thanks for an excellent article Georgi! I’m trying to put your method into practice using excel. However, I can’t seem to find a tutorial for applying the Dunnett’s method to the p values. Can you point me in the right direction?

Hi James,

I don’t think Excel will be up to the task, since calculation of the adjusted critical values (to avoid coding his tables, which are limited by nature) is not trivial as it involves integration over a multivariate normal distribution. There are free packages in R that can do this part of the heavy lifting for you, though I’m not sure if this will be something you can link to Excel.

How do you calculate sample size for A/B test where the metric you are comparing is non binomial, for example- revenue per visitor?

Same as with a binomial one, with the difference that you need some historical data from which to estimate the standard deviation.

hi georgi, thanks for the excellent article. i’m wondering how we go about measuring the impact of successive a/b tests. i’ll give you a concrete example.

i a/b tested a new feature on our ordering process during october 2017. i demonstrated that the average size of our orders went up about 50 cents using a t-test, where i split our A and B branches 50/50. i got my executives to agree that we were observing a 50 cent lift on a per-order basis, and we then started using this new feature on 100% of our user sessions beginning Nov 2017.

around january 2018 we developed and tested a new feature, did the same a/b test methodology, splitting our sessions 50/50. we found that the new feature provided an additional 30 cent average lift per order. so we now claimed that we were seeing an 80 cent lift over our original baseline, and our executives agreed. so we extended this new feature across 100% of our user sessions beginning Feb 2018.

this same process happened several more times over the course of the last year, each time demonstrating additional lift (sometimes not) and adding that to the total per-order lift observed to date. by the end of august 2018, we had shown a total of $1.58 average order lift since inception of the program.

now i am trying to calculate the total incremental revenue the company has gained as a result of these new features in calendar year 2018, and i’m being told by my execs that none of the claimed money is real, because compared to september 2017, our september 2018 data is not $1.58 higher per average order size (it’s only about $1 higher).

what would you do to counter this argument? i’m in a bit of a bind – as a statistician it seems obvious that a/b testing is the right way to show the lift, since there are too many other factors such as seasonality that affect order size. for example, there have been outside influences like economic changes, industry changes, product line-up changes, etc that have altered customer behavior since september 2017. so i’m being unfairly gipped out of $0.58 and that considerably alters the contribution to our company’s growth that will be credited to our team. would really appreciate your help. thanks georgi!

Hi SJ,

Interesting that you were focusing on AOV for these tests, usually one has CR as a primary metric and AOV or ARPU as secondary, or straight up ARPU as primary metric. I guess the things tested were such that they would not result in a lift in CR, but only in increase in AOV (e.g. cross-selling), so maybe CR was a secondary metric in this case?

Were you adding up the observed % lift in AOV or were you adding up the value bounding the lower limit of a one-sided XX% interval, where XX is the required level of confidence (equal to the significance threshold)? If you were adding up the observed lift, then you were being a bit too optimistic. While it is usually our best single guess, allowing for the intervals to enter into the estimation of the cumulative effect seems right as the lower bound of the interval is delimiting values you could not reject with the data at hand and your chosen level of significance. Granted, such an estimate will be conservative, but it will only be as conservative as you require it to be by setting the significance threshold, presumably based on a risk/reward calculation for each separate test.

It’s likely there are better ways to estimate the cumulative effect than simply using the lower one-sided CI bounds, but this enters into meta-analytical territory and I’m not as well-versed in those methods, yet.

Your remarks on the outside influences are correct and it is one of the big reasons why we test, instead of simply comparing YoY performance to determine the effect of changes. I would add to that any threats to generalizability that might be relevant in some of these tests (see http://blog.analytics-toolkit.com/2018/representative-samples-generalizability-a-b-testing-results/ for more on this thorny topic).

There is a white paper on the question you bring up coming soon (I was interviewed as part of its preparation) and if you want I can ping you when it is out as it is likely to be of interest.

Best,

Georgi