**What is Statistical Significance?**

In the context of AB testing experiments, statistical significance is how likely it is that the difference between your experiment’s control version and test version isn’t due to error or random chance.

For example, if you run a test with a 95% significance level, you can be 95% confident that the differences are real.

Calculating statistical significance is quite complex (https://en.wikipedia.org/wiki/Statistical_significance) but the good news is that TubeBuddy does it for you automatically.

Our calculations focus specifically on CTR - meaning the variables we look at are how many Impressions each version of your test got as well as how many clicks each version got. It does not take into account views, watch time or any other metrics.

If you'd like to play around with some numbers and see how significance is affected, take a look at https://neilpatel.com/ab-testing-calculator/ and plug in some numbers.

**A Simple Example**

Let's say you run a test and the original version of your video is shown to 10 people, of which 1 click to watch. Your CTR is 10%. And your variation was shown to 15 people and 2 clicked giving the variation a 12.3% CTR. Can you trust that the variation will always outperform the original? Probably not. They were only shown to 25 people total.

But let's say you do the same test and the original is shown to 100,000 and of that, 10,000 click - giving it a 10% CTR. And the variation is shown to 150,000 people and 20,000 click giving it a 12.3% CTR. Can you trust these numbers? Yes. They were shown to SO many people that there is enough evidence to show that the variation will perform better in the long run.