Economics 326 — Introduction to Econometrics II
Consider the model:
\begin{align*} \ln\left(\text{Wage}_{i}\right) = {} & \beta_{0} + \beta_{1}\text{Experience}_{i} + \beta_{2}\text{Experience}_{i}^{2} \\ & + \beta_{3}\text{PrevExperience}_{i} + \beta_{4}\text{PrevExperience}_{i}^{2} \\ & + \beta_{5}\text{Education}_{i} + U_{i}, \end{align*}
where \text{Experience} is the experience at the current job and \text{PrevExperience} is the previous experience.
We test whether, after controlling for the experience at the current job and education, previous experience has no effect on wage:
H_{0}: \beta_{3} = 0,\; \beta_{4} = 0.
We have two restrictions on the model parameters.
The alternative hypothesis is that at least one of the coefficients, \beta_{3} or \beta_{4}, is different from zero:
H_{1}: \beta_{3} \neq 0 \text{ or } \beta_{4} \neq 0.
Let T_{3} and T_{4} be the t-statistics associated with the coefficients of \text{PrevExperience} and \text{PrevExperience}^{2}:
T_{3} = \frac{\hat{\beta}_{3}}{\mathrm{se}\left(\hat{\beta}_{3}\right)} \quad \text{and} \quad T_{4} = \frac{\hat{\beta}_{4}}{\mathrm{se}\left(\hat{\beta}_{4}\right)}.
We can use T_{3} and T_{4} to test the significance of \beta_{3} and \beta_{4} separately using two size \alpha tests:
Reject H_{0,3}: \beta_{3} = 0 in favor of H_{1,3}: \beta_{3} \neq 0 when \left\lvert T_{3} \right\rvert > t_{n-k-1,1-\alpha/2}.
Reject H_{0,4}: \beta_{4} = 0 in favor of H_{1,4}: \beta_{4} \neq 0 when \left\lvert T_{4} \right\rvert > t_{n-k-1,1-\alpha/2}.
Rejecting H_{0}: \beta_{3} = 0,\, \beta_{4} = 0 in favor of H_{1}: \beta_{3} \neq 0 or \beta_{4} \neq 0 when at least one of the two coefficients is significant at level \alpha, i.e., when
\left\lvert T_{3} \right\rvert > t_{n-k-1,1-\alpha/2} \quad \text{or} \quad \left\lvert T_{4} \right\rvert > t_{n-k-1,1-\alpha/2},
is not a size \alpha test!
If A and B are two events, then (A \cap B) \subset A and therefore P(A \cap B) \leq P(A).
When \beta_{3} = \beta_{4} = 0:
\begin{aligned} & P\!\left(\text{Reject } H_{0,3} \text{ or } H_{0,4}\right) \\ &= P\!\big(\left\lvert T_{3}\right\rvert > t_{n-k-1,1-\alpha/2} \;\textbf{or}\; \left\lvert T_{4}\right\rvert > t_{n-k-1,1-\alpha/2}\big) \\ &= P\!\big(\left\lvert T_{3}\right\rvert > t_{n-k-1,1-\alpha/2}\big) \\ &\quad + P\!\big(\left\lvert T_{4}\right\rvert > t_{n-k-1,1-\alpha/2}\big) \\ &\quad - P\!\big(\left\lvert T_{3}\right\rvert > t_{n-k-1,1-\alpha/2} \;\textbf{and}\; \left\lvert T_{4}\right\rvert > t_{n-k-1,1-\alpha/2}\big) \\ &= 2\alpha - P\!\big(\text{both reject}\big) \\ &\geq \alpha. \end{aligned}
Consider the model
\begin{align*} Y_{i} = {} & \beta_{0} + \beta_{1}X_{1,i} + \ldots + \beta_{q}X_{q,i} \\ & + \beta_{q+1}X_{q+1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}. \end{align*}
We test whether the first q regressors have no effect on Y (after controlling for the other regressors).
The null hypothesis has q exclusion restrictions:
H_{0}: \beta_{1} = 0,\, \beta_{2} = 0,\, \ldots,\, \beta_{q} = 0.
The alternative hypothesis is that at least one of the restrictions in H_{0} is false:
H_{1}: \beta_{1} \neq 0 \text{ or } \beta_{2} \neq 0 \text{ or } \ldots \text{ or } \beta_{q} \neq 0.
The idea of the test is to compare the fit of the unrestricted model with that of the null-restricted model.
Let SSR_{ur} denote the Residual Sum-of-Squares of the unrestricted model:
\begin{align*} Y_{i} = {} & \beta_{0} + \beta_{1}X_{1,i} + \ldots + \beta_{q}X_{q,i} \\ & + \beta_{q+1}X_{q+1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}. \end{align*}
The restricted model given H_{0}: \beta_{1} = 0, \ldots, \beta_{q} = 0 is
Y_{i} = \beta_{0} + \beta_{q+1}X_{q+1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
The F-statistic:
F = \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)}.
F = \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)}.
Since SSR can only increase when we drop regressors,
SSR_{r} - SSR_{ur} \geq 0,
and therefore F \geq 0.
If the null restrictions are true, the excluded variables do not contribute to explaining Y (in population), so we should expect that SSR_{r} - SSR_{ur} is small and F is close to zero.
If the null restrictions are false, the imposed restrictions should substantially worsen the fit, so we should expect that SSR_{r} - SSR_{ur} is large and F is far from zero.
We should reject H_{0} when F > c where c is some positive constant.
F = \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)}.
We should reject H_{0} when F > c.
There is a probability that F > c even when H_{0} is true, so we need to choose c such that P(F > c \mid H_{0} \text{ is true}) = \alpha.
Under H_{0} and conditional on \mathbf{X}, the F-statistic has the F distribution with two parameters: the numerator df (q) and the denominator df (n - k - 1):
F \mid \mathbf{X} \sim F_{q,\, n-k-1}.
Similarly to the standard normal and t distributions, the F distribution has been tabulated and its critical values are available in statistical tables and statistical software such as R.
When H_{0} is true, conditional on \mathbf{X}:
F = \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)} \sim F_{q,\, n-k-1}.
Let F_{q,n-k-1,\tau} be the \tau-th quantile of the F_{q,n-k-1} distribution.
A size \alpha test of H_{0}: \beta_{1} = 0, \ldots, \beta_{q} = 0 against H_{1}: \beta_{1} \neq 0 or \ldots or \beta_{q} \neq 0 is
\text{Reject } H_{0} \text{ when } F > F_{q,\, n-k-1,\, 1-\alpha}.
The p-value can be found as \tau such that F = F_{q,n-k-1,1-\tau}. The p-value equals \tau.
To compute F critical values, use qf():
F_{q,\, n-k-1,\, 1-\alpha} = \texttt{qf(1 - alpha, df1 = q, df2 = n - k - 1)}.
To compute p-values from the F distribution, use pf():
\text{p-value} = 1 - \texttt{pf(F, df1 = q, df2 = n - k - 1)}.
Consider the model:
\begin{align*} \ln\left(\text{Wage}_{i}\right) = {} & \beta_{0} + \beta_{1}\text{Experience}_{i} + \beta_{2}\text{Experience}_{i}^{2} \\ & + \beta_{3}\text{PrevExperience}_{i} + \beta_{4}\text{PrevExperience}_{i}^{2} \\ & + \beta_{5}\text{Education}_{i} + U_{i}. \end{align*}
We test
H_{0}: \beta_{3} = 0,\; \beta_{4} = 0 \quad \text{against} \quad H_{1}: \beta_{3} \neq 0 \text{ or } \beta_{4} \neq 0.
q = 2.
\alpha = 0.05.
Data: wage1 from the wooldridge R package (n = 526).
The unrestricted model includes all five regressors:
fit_ur <- lm(lwage ~ Education + Experience + Experience2
+ PrevExperience + PrevExperience2,
data = wage1)
summary(fit_ur)
Call:
lm(formula = lwage ~ Education + Experience + Experience2 + PrevExperience +
PrevExperience2, data = wage1)
Residuals:
Min 1Q Median 3Q Max
-2.01561 -0.27189 -0.01607 0.27683 1.33508
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 0.2368427 0.1028700 2.302 0.021710 *
Education 0.0887704 0.0072131 12.307 < 2e-16 ***
Experience 0.0471914 0.0068074 6.932 1.23e-11 ***
Experience2 -0.0008518 0.0002472 -3.446 0.000615 ***
PrevExperience 0.0168997 0.0047331 3.571 0.000389 ***
PrevExperience2 -0.0003727 0.0001208 -3.086 0.002139 **
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Residual standard error: 0.4319 on 520 degrees of freedom
Multiple R-squared: 0.3461, Adjusted R-squared: 0.3398
F-statistic: 55.04 on 5 and 520 DF, p-value: < 2.2e-16
The residual sum-of-squares of the unrestricted model:
SSR_{ur} \approx 96.998; n - k - 1 = 526 - 5 - 1 = 520.
The restricted model drops \text{PrevExperience} and \text{PrevExperience}^{2}:
Call:
lm(formula = lwage ~ Education + Experience + Experience2, data = wage1)
Residuals:
Min 1Q Median 3Q Max
-2.07720 -0.28197 -0.02346 0.26859 1.41509
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 0.3688491 0.0908138 4.062 5.62e-05 ***
Education 0.0852822 0.0068978 12.364 < 2e-16 ***
Experience 0.0510784 0.0067937 7.518 2.43e-13 ***
Experience2 -0.0009941 0.0002463 -4.036 6.24e-05 ***
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Residual standard error: 0.4365 on 522 degrees of freedom
Multiple R-squared: 0.3294, Adjusted R-squared: 0.3256
F-statistic: 85.49 on 3 and 522 DF, p-value: < 2.2e-16
The residual sum-of-squares of the restricted model:
SSR_{r} \approx 99.463.
Computing the F-statistic manually:
The critical value F_{2,520,0.95}:
Since F \approx 6.61 > 3.01, at the 5% significance level we reject H_{0} that previous experience has no effect on wage.
The p-value:
We reject H_{0} for any \alpha > 0.00147.
linearHypothesis()Instead of running two models (restricted and unrestricted), we can use linearHypothesis() from the car package after estimating the unrestricted model.
Testing whether previous experience has no effect (\beta_{3} = 0,\, \beta_{4} = 0):
Linear hypothesis test:
PrevExperience = 0
PrevExperience2 = 0
Model 1: restricted model
Model 2: lwage ~ Education + Experience + Experience2 + PrevExperience +
PrevExperience2
Res.Df RSS Df Sum of Sq F Pr(>F)
1 522 99.463
2 520 96.998 2 2.4651 6.6075 0.001466 **
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Testing whether the experience profiles are identical (\beta_{1} = \beta_{3} and \beta_{2} = \beta_{4}):
Linear hypothesis test:
Experience - PrevExperience = 0
Experience2 - PrevExperience2 = 0
Model 1: restricted model
Model 2: lwage ~ Education + Experience + Experience2 + PrevExperience +
PrevExperience2
Res.Df RSS Df Sum of Sq F Pr(>F)
1 522 102.297
2 520 96.998 2 5.2987 14.203 9.87e-07 ***
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Let R_{ur}^{2} denote the R^{2} of the unrestricted model:
\begin{align*} Y_{i} = {} & \beta_{0} + \beta_{1}X_{1,i} + \ldots + \beta_{q}X_{q,i} \\ & + \beta_{q+1}X_{q+1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}. \end{align*}
Let R_{r}^{2} denote the R^{2} of the restricted model:
Y_{i} = \beta_{0} + \beta_{q+1}X_{q+1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
The two models have the same dependent variable and therefore the same Total Sum-of-Squares:
SST = \sum_{i=1}^{n}(Y_{i} - \bar{Y})^{2} = SST_{ur} = SST_{r}.
Since SSR/SST = 1 - R^{2}:
\begin{align*} F &= \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)} \\ &= \frac{\left(\frac{SSR_{r}}{SST} - \frac{SSR_{ur}}{SST}\right)/q}{\frac{SSR_{ur}}{SST}/(n - k - 1)} \\ &= \frac{\left((1 - R_{r}^{2}) - (1 - R_{ur}^{2})\right)/q}{(1 - R_{ur}^{2})/(n - k - 1)} \\ &= \frac{(R_{ur}^{2} - R_{r}^{2})/q}{(1 - R_{ur}^{2})/(n - k - 1)}. \end{align*}
Verification with the wage example:
Suppose we want to test H_{0}: \beta_{1} = 1 against H_{1}: \beta_{1} \neq 1 in
Y_{i} = \beta_{0} + \beta_{1}X_{1,i} + \beta_{2}X_{2,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
The restricted model is
Y_{i} = \beta_{0} + X_{1,i} + \beta_{2}X_{2,i} + \ldots + \beta_{k}X_{k,i} + U_{i},
or
Y_{i} - X_{1,i} = \beta_{0} + \beta_{2}X_{2,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
Suppose we want to test H_{0}: \beta_{1} + \beta_{2} = 1 against H_{1}: \beta_{1} + \beta_{2} \neq 1 in
Y_{i} = \beta_{0} + \beta_{1}X_{1,i} + \beta_{2}X_{2,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
The restricted model is
Y_{i} = \beta_{0} + (1 - \beta_{2})X_{1,i} + \beta_{2}X_{2,i} + \ldots + \beta_{k}X_{k,i} + U_{i},
or
Y_{i} - X_{1,i} = \beta_{0} + \beta_{2}(X_{2,i} - X_{1,i}) + \ldots + \beta_{k}X_{k,i} + U_{i}.
The F statistic can also be used for testing a single restriction.
For a single restriction, the F test and the t test lead to the same outcome because
t_{n-k-1}^{2} = F_{1,\, n-k-1}.
Consider the model
Y_{i} = \beta_{0} + \beta_{1}X_{1,i} + \ldots + \beta_{k}X_{k,i} + U_{i}.
Suppose we want to test whether none of the regressors explain Y:
\begin{align*} H_{0} &: \beta_{1} = \beta_{2} = \ldots = \beta_{k} = 0 \quad (k \text{ restrictions}), \\ H_{1} &: \beta_{j} \neq 0 \text{ for some } j = 1, \ldots, k. \end{align*}
The restricted model is Y_{i} = \beta_{0} + U_{i}, and since \hat{\beta}_{0} = \bar{Y} in this model,
SSR_{r} = \sum_{i=1}^{n}(Y_{i} - \bar{Y})^{2} = SST \quad \text{and} \quad SSR_{ur} = SSR.
The F statistic for the overall significance test is
\begin{align*} F &= \frac{(SSR_{r} - SSR_{ur})/k}{SSR_{ur}/(n - k - 1)} \\ &= \frac{(SST - SSR)/k}{SSR/(n - k - 1)} \\ &= \frac{SSE/k}{SSR/(n - k - 1)} \\ &= \frac{R^{2}/k}{(1 - R^{2})/(n - k - 1)}. \end{align*}
The F statistic for the overall significance test and its p-value are reported in the top part of R regression output:
Call:
lm(formula = lwage ~ Education + Experience + Experience2 + PrevExperience +
PrevExperience2, data = wage1)
Residuals:
Min 1Q Median 3Q Max
-2.01561 -0.27189 -0.01607 0.27683 1.33508
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 0.2368427 0.1028700 2.302 0.021710 *
Education 0.0887704 0.0072131 12.307 < 2e-16 ***
Experience 0.0471914 0.0068074 6.932 1.23e-11 ***
Experience2 -0.0008518 0.0002472 -3.446 0.000615 ***
PrevExperience 0.0168997 0.0047331 3.571 0.000389 ***
PrevExperience2 -0.0003727 0.0001208 -3.086 0.002139 **
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
Residual standard error: 0.4319 on 520 degrees of freedom
Multiple R-squared: 0.3461, Adjusted R-squared: 0.3398
F-statistic: 55.04 on 5 and 520 DF, p-value: < 2.2e-16
Individual t-tests cannot be combined to test joint hypotheses. Rejecting when at least one individual t-test rejects leads to a test with size greater than \alpha.
The F-statistic compares the fit of the unrestricted model to the restricted model:
F = \frac{(SSR_{r} - SSR_{ur})/q}{SSR_{ur}/(n - k - 1)}.
Under H_{0}, F \mid \mathbf{X} \sim F_{q,\, n-k-1}. Reject H_{0} when F > F_{q,\, n-k-1,\, 1-\alpha}.
Equivalently, in terms of R^{2}:
F = \frac{(R_{ur}^{2} - R_{r}^{2})/q}{(1 - R_{ur}^{2})/(n - k - 1)}.
For a single restriction (q = 1), F = T^{2} and the F test is equivalent to the two-sided t-test.
The overall significance test (H_{0}: \beta_{1} = \ldots = \beta_{k} = 0) is a special case with F = \frac{R^{2}/k}{(1 - R^{2})/(n - k - 1)}.