OLS Assumptions
\[
Y_i = \beta_0 + \beta_1 X_{i1} + \cdots + \beta_K X_{iK} + \epsilon_i
\]
- Random sample: \(\{ Y_i , X_{i1}, \ldots, X_{iK} \}\) is i.i.d. drawn sample
- i.i.d.: identically and independently distributed
- \(\epsilon_i\) has zero conditional mean \[
E[ \epsilon_i | X_{i1}, \ldots, X_{iK}] = 0
\]
- This implies \(Cov(X_{ik}, \epsilon_i) = 0\) for all \(k\). (or \(E[\epsilon_i X_{ik}] = 0\))
- No correlation between error term and explanatory variables.
- Large outliers are unlikely:
- The random variable \(Y_i\) and \(X_{ik}\) have finite fourth moments.
- No perfect multicollinearity:
- There is no linear relationship betwen explanatory variables.
- The OLS estimator has ideal properties (consistency, asymptotic normality, unbiasdness) under these assumptions.
- In this chapter, we study the role of these assumptions.
- In particular, we focus on the following two assumptions
- No correlation between \(\epsilon_{it}\) and \(X_{ik}\)
- No perfect multicollinearity