## Gauss Markov Theorem

The **Gauss Markov theorem **tells us that if a certain set of assumptions are met, the ordinary least squares estimate for regression coefficients gives you the *best linear unbiased estimate (BLUE)* possible.

## Gauss Markov Assumptions

There are five Gauss Markov assumptions (also called *conditions*):

**Linearity**: the parameters we are estimating using the OLS method must be themselves linear.**Random**: our data must have been randomly sampled from the population.**Non-Collinearity**: the regressors being calculated aren’t perfectly correlated with each other.**Exogeneity**: the regressors aren’t correlated with the error term.**Homoscedasticity**: no matter what the values of our regressors might be, the error of the variance is constant.

## Purpose of the Assumptions

The **Gauss Markov assumptions** guarantee the validity of ordinary least squares for estimating regression coefficients.

Checking how well our data matches these assumptions is an important part of estimating regression coefficients. When you know where these conditions are violated, you may be able to plan ways to change your experiment setup to help your situation fit the ideal Gauss Markov situation more closely.

In practice, the Gauss Markov assumptions are **rarely all met perfectly**, but they are still useful as a benchmark, and because they show us what ‘ideal’ conditions would be. They also allow us to pinpoint problem areas that might cause our estimated regression coefficients to be inaccurate or even unusable.

## The Gauss-Markov Assumptions In Algebra

We can summarize the Gauss-Markov Assumptions succinctly in algebra, by saying that a linear regression model represented by

y_{i} = x_{i}‘ β + ε_{i}

and generated by the ordinary least squares estimate is the best linear unbiased estimate (BLUE) possible if

- E{ε
_{i}} = 0, i = 1, … , N - {ε
_{1}…ε_{n}} and {x_{1},…,x_{N}} are independent - cov{ε
_{i}, ε_{j}} = 0, i, j = 1,…, N I ≠ j. - V{ε
_{1}= σ^{2}, i = 1, …N

Where:

- “ε” is the Greek letter epsilon, which measures error.
- Σ = Summation notation (“add up”)

The first of these assumptions can be read as “The expected value of the error term is zero.”. The second assumption is collinearity, the third is exogeneity, and the fourth is homoscedasticity.

## References

Anderson, Patricia. The Gauss-Markov Theorem: Study Guide. Retrieved from http://www.dartmouth.edu/~econ20pa/StudyGuide1.doc on May 20, 2018.

Lee, Q. OLS, BLUE and the Gauss Markov Theorem. Economics Society: University of Waterloo. Retrieved from http://uweconsoc.com/ols-blue-and-the-gauss-markov-theorem/ on May 20, 2018.

Troeger, Vera. Gauss-Markov Assumptions, Full Ideal Conditions of OLS. Retrieved from http://uweconsoc.com/ols-blue-and-the-gauss-markov-theorem/ on May 20, 2018.

**CITE THIS AS:**

**Stephanie Glen**. "Gauss Markov Theorem & Assumptions" From

**StatisticsHowTo.com**: Elementary Statistics for the rest of us! https://www.statisticshowto.com/gauss-markov-theorem-assumptions/

**Need help with a homework or test question? **With **Chegg Study**, you can get step-by-step solutions to your questions from an expert in the field. Your first 30 minutes with a Chegg tutor is free!

**Comments? Need to post a correction?** Please post a comment on our ** Facebook page**.