What is the Variance Sum Law?
The Variance Sum Law determines the variance of a sum (or difference) when you know the variance of the component parts.
For example, suppose you ran a research project which involved sampling the weight of apples in New York orchards, and then you did a similar project on oranges in Southern California. Now, imagine you need to work with the collected data from both research projects and draw some conclusions on the weight of apples and oranges. But first, what would the variance of your new data set be?
The Variance Sum Law- Independent Case
If your two sets are independent, like the apples and oranges example, you can use the simplest version of the variance sum law.
Var(X ± Y) = Var(X) + Var(Y).
This just states that the combined variance (or the differences) is the sum of the individual variances.
Sum and Difference
Note that the variance of the sum of both sets and the difference of both sets is exactly the same. This may take you by surprise on first sight, but after thinking it through you’ll realize that since Var(X-Y)=Var(X+(-Y))= Var(X) + Var(-Y) and Var(Y)= Var(-Y), Var(X+Y) must be equivalent to Var (X-Y).
Variance Sum Law — Dependent Case
Cov(x,y) is the covariance of x and y.
Since the population correlation coefficient ρ is defined as = Cov(x,y) / σxσy, this can also be expressed as
If you prefer an online interactive environment to learn R and statistics, this free R Tutorial by Datacamp is a great way to get started. If you're are somewhat comfortable with R and are interested in going deeper into Statistics, try this Statistics with R track.Comments are now closed for this post. Need help or want to post a correction? Please post a comment on our Facebook page and I'll do my best to help!