Statistics Definitions > Variance Sum Law

## What is the Variance Sum Law?

The **Variance Sum Law** determines the variance of a sum (or difference) when you know the variance of the component parts.

For example, suppose you ran a research project which involved sampling the weight of apples in New York orchards, and then you did a similar project on oranges in Southern California. Now, imagine you need to work with the collected data from both research projects and draw some conclusions on the weight of apples and oranges. But first, what would the variance of your new data set be?

## The Variance Sum Law – Independent Case

If your two sets are independent, like the apples and oranges example, you can use the simplest version of the variance sum law.

Var(X ± Y) = Var(X) + Var(Y).

This just states that the combined variance (or the differences) is the sum of the individual variances.

This can also be written as

So if the variance of set 1 was 2, and the variance of set 2 was 5.6, the variance of the united set would be 2 + 5.6 = 7.6.

## Sum and Difference

Note that the variance of the sum of both sets and the difference of both sets is exactly the same. This may take you by surprise on first sight, but after thinking it through you’ll realize that since Var(X-Y)=Var(X+(-Y))= Var(X) + Var(-Y) and Var(Y)= Var(-Y), Var(X+Y) must be equivalent to Var (X-Y).

## Variance Sum Law — Dependent Case

Now, what if your data sets are not independent? When you introduced dependence into your data sets the variance sum law becomes:

Cov(x,y) is the covariance of x and y.

Since the population correlation coefficient ρ is defined as = Cov(x,y) / σxσy, this can also be expressed as

When you are analyzing a sample, you’ll use Pearson’s correlation coefficient r as an estimate of ρ .