Go from

“I think”to“I know how much.”

### Whether you are in product, design or, growth, you probably come across the following questions:

- How could we increase our conversion rate?
- How can we improve our AARRR funnel?
- How successful has this feature launch been?
- Have these notifications increased retention?

### You can get answers to those questions by correctly designing, running, and interpreting A/B tests

All you need to get started on your testing journey is:

**Basic statistics understanding****Accurate and sufficient (>1000 monthly unique users) product data**

The approach taken is rigorous and encourages best practices but:

**It will not**demonstrate the maths behind frequentist hypothesis testing.**It will not**cover more advanced aspects such as A/B in a Bayesian context, multivariate-testing or Bonferroni correction.

Ready? Let’s get started!

## 0. Understand that split testing is a quantitative way to improve your product and that it’s easy to get it wrong

At its core, A/B testing is a way to compare two versions of a flow in your user journey in order to determine which one performs best.

To be able to get insights from your tests you will need **enough people in each group** (called *control* and *experiment* group, respectively) and **enough of a conversion difference**. Both values should be defined **before** running your test, as you’re about to see in point 3.

## 1. Select an important conversion step from your product’s funnel

Technically you can test anything, like changing the font on your terms and conditions page. However, you always want to pick the steps with the highest impact on your business outcome. Also, keep in mind the importance of **data quantity** — the higher the traffic the faster the conclusions — and of **data quality** — garbage in, garbage out.

Bad data, bad conclusions. Few data, slow conclusions.

Let’s get practical and do it for Trana, the example we’ll use for this post.

Trana is a Facebook Messenger *ChatApp* that provides extra running insights to its users. To do so it connects to their Strava account. Therefore, the moment at which we ask users to link their account is key. On top of this instinctive rationale, let’s analyze our full onboarding funnel. This can be done in a tool such as Mixpanel after having designed a tracking plan.

We can see that linking the account, going from *Onboarding Started* to *Strava Connected*, is the most impactful action both in absolute and relative terms. Got your own important step? Let’s continue!

## 2. Analyse the current *Click-Through-Probability* from this step to a desired action

The Click-through-probability (CTP) is the probability a given user clicks to get to the next step.

Its value is always between 0 and 1. It’s therefore **not** the same as the click-through-rate (CTR) which is the #clicks/#impression.

The CTP needs to be from the chosen step to a crucial action, not necessarily to the very next step, because you might otherwise end up at a local maximum. Here’s a simple example to understand this: let’s say that you now automatically redirect people from the home page to the pricing/premium page of your product. Your home to pricing conversion will be through the roof but you might want to assess whether people are in turn more likely to become premium users.

In short: optimize with the end, not the very next step, in mind.

We have seen from the full funnel that the CTP is 80%. In this case connecting their Strava account is clearly an important milestone so we can use this conversion probability as target metric.

## 3. Construct a hypothesis to kickstart your experiment

AB testing is nothing but methodologically assess whether a variation, called **experiment**, performs better than the current situation, called **control**.

An experiment starts by constructing a hypothesis. It can arise from gut-feeling, customers feedback or team discussions.

The hypothesis is the following: *By seeing how Trana can help them run smarter, users will be less reluctant to connect their Strava account*

## 4. Determine the sample size

This is the heart of A/B testing and the most jargon-heavy part of this guide. I’ll be happy to answer any question you might ask on twitter.

Running an A/B test involves creating a control and an experiment **sample**. *Sampling*, assigning each user to one variation, will be done by splitting your users into two groups.

One user should always see the same variation, meaning that the **sampling is deterministic**, not simply with a 50% probability.

We usually sample according to odd/even IDs, many tools can take care of that for you.

A common mistake is to ignore sampling, directly change your product, and compare the new conversion probabilities with old ones.

This is leading teams to bad conclusions because the people coming to your product are changing across time. For example, you can’t say that the new flow (used by people coming from ads) is worst that the old one (that was used by your friends and family) because the seeing it are way less likely to bear with your product to begin with.

Your sample size will be determined by:

- Your baseline conversion rate
- The absolute
*minimum detectable effect*(MDE/d*min*) you pick. This is the minimum difference between control and experiment conversion rates you will be able to detect. If you have a lot of weekly traffic (vs the resulting sample size) pick the smallest difference you feel is worth. If you don’t, pick a difference that results in reasonably small sample size but still seems achievable. - Alpha (α): By default keep it as 95%
- Beta (β): By default keep is as 20% which gives a 80% value for 1-β

The test will be run until **both** samples have a least 1030 observations. From the historic traffic you can assess how long that’ll be. Try to keep experiments running less than one month. The faster they run, the faster your product improves.

Compute your sample size, write it down and let the testing begin. Once you have sufficient data you can analyze the results.

## 5. Get valid insights from your results

#### 5.1 Collect the data.

Wait until both variations have sufficient sample size and proceed to the data collection.

The experiment conversion is better. It’s a first good sign but we can’t stop there. So let’s proceed to a few checks.

#### 5.2 Make sure your sampling worked as intended.

First, check that the sampling was not biased towards one of the groups. Here is the computation for our example:

#### 5.3 Check if the experiment performed significantly better.

Assessing whether the experiment is outperforming the control can be a hassle. Therefore, I built a little tool to do it, plug your figures into it to perform the three checks. Below are the details of the computation. d*min* is the minimum detectable effect, N is the sample size, and X the conversion **count**, the number of people who took the desired action.

Where d̂ ± m is the confidence interval.

Here is how to interpret results:

- d̂ – m > d
*𝑚𝑖𝑛*:

The experiment is outperforming the control (**statistical**and**practical**significance). Change flow to new version. - d̂ – m > 0:

The experiment is outperforming the control (**statistical**significance only). This is the case in**our example**. To get practical significance you can let the test run longer. You can also make a judgement call and go with your gut feeling. - d̂ – m < 0:

You can’t say that the experiment is outperforming the control. Keep the original version and generate new hypotheses.

## 6. Communicate your results.

Whatever the outcome, **always** communicate and document.

I recommend that you signal when the test starts and when it ends, using these key moments as an opportunity for stakeholders to voice their ideas and be up-to-date with the latest changes.

Also, always document and archive. You want the testing process to outlive you so that anyone taking over has a library of experiments to learn from.

Here is a Google Slides template your can use to communicate and document your results.

## 7. Bring it all together.

Here is a recap of the steps you ought to take to go from **“I think”** to **“I know how much”** :

- Select an important conversion step from your product’s funnel
- Find out the current Click-Through-Probability (CTP) from this step to a desired action
- Construct a hypothesis
- Determine the sample size
- Gather and analyse results
- Communicate and document your results

### This guide was made possible thanks to these amazing resources:

### Here’s how you can go further:

*Editor’s note: This was originally posted on Robin Rampaer’s blog, and has been reposted with permission.*