Econometrics - Solution sh.2B 2024
Econometrics - Solution sh.2B 2024
Economics Department
Third Year
Econometrics I
TA. Menna Sherif
SOLUTION OF SHEET 2B
MULTIPLE LINEAR REGRESSION MODEL
2. The linearity assumption states that the parameters and the variables have to
be linear.
False. It assumes only that the model is linear in parameters and not
necessarily in variables. For instance,
Y = 1 X e u is not linear but after the log transformation, the equation
2
3. If all the values of 𝑋𝑖 are equal, so the estimation of the parameters in the simple
regression model will be more easier.
False. Having equal values of x means that we cannot estimate the OLS
coefficients.
As we know ̂ 2 = (∑YiXi - nȲx̄) / (∑Xi2-nx̄2), since the denominator is
equal to zero, we cannot estimate ̂ 2 ; and also we cannot estimate ̂1
as it depends linearly on ̂ 2 .
1
4. In the multiple regression model if we have K independent variables, so we have K+1
parameters.
True. They are the intercept plus the coefficient of each independent
variable, but if we have a regression from the origin, the no. of
parameters will equate the no. of independent variables.
5. The multiple regression model assumes that the vector u is a stochastic vector with
expected value of zero and variance of identity matrix.
False. For a regression model, Y = X + u , u is a stochastic error term
with E(u|x)=0 and var(u | x) = 2 I that is constant variance
(homoscedasicity) and not an identity matrix.
7. The OLS residuals 𝑢̂ are uncorrelated, if this is the case for the errors 𝑢.
False. Not necessary, as I may take a sample which is characterised by
correlation between the residuals (withdrew unrepresentative sample).
8. If the regressor matrix X in the linear model 𝑌 = 𝑋𝛽 + 𝑢 is not a full rank, then
an unbiased estimator for 𝛽 does not exist.
True. If the matrix is not in a full rank, its determinant would be zero.
1
And as [𝑋 −1 = 𝑎𝑑𝑗] ,we cannot get the inverse of the matrix, hence
𝐷
we cannot estimate the parameters (ie. 𝛽𝑠 will not exist).
2
11. In the case of perfect multi-colinearity among k explanatory variables, the
rank of (X’X) is k.
False. If we have k explanatory variables, the order of (X’X) would be
(k+1)(k+1). And this matrix would be in full rank if its rank is (k+1), but
in case of perfect multicollinearity, (X’X) would not be in the full rank,
ie. Its rank would be less then (k+1), so it could be k or less.
12. The normality assumption of the disturbance term is necessary to apply least
squares method.
False. The normality assumption is only needed for constructing the
confidence intervals and hypothesis testing. In other words, it is an
important assumption for conducting inferences to the true value of
parameters.
13. Even though the error term in the linear regression model is not normally
distributed, OLS estimators are still normally distributed.
False. The normality assumption of the error term is necessary for the
normality of the OLS estimators.
Since Yi is a function of the error term Yi = 1 + 2 X i + ui , if u i is
normally distributed with zero mean and constant variance Yi is
normally distributed also.
Since ˆ1andˆ2 are linear in Yi , the estimators are also normally
distributed.
14. Even though the disturbance term in the CLRM is not normally distributed, the OLS
estimators are still unbiased.
True. The OLS estimators’ unbiasedness does not depend on the
normality of the error term. It depends on the assumption that the
expected value of the error term is equal to zero,( Estimators’
unbiasedness means that the expectation of the estimators is equal to
the true value of the parameter).
15. To have unbiased estimators, all the assumptions of OLS have to be satisfied.
False. OLS estimators’ unbiasedness depends on the assumptions that
enable us estimating the model (i.e. linear model, unstochastic
explanatory variables, sample size being greater than the no. of
parameters, variability in Xs , correct specification of the model and no
perfect multicollinearity) plus the assumption that the expectation of the
error term is equal to zero.
3
16. The variance of the estimators is equal to ϭ2(x'x)-1. True.
17. OLS estimators ̂1 and ̂ 2 are biased, consistent and efficient estimators for the
real parameters 1 and 2 .
False. OLS estimators are unbiased, consistent and efficient estimators
for the real parameters. (Define each)
19. The test of significance for the whole relation is just equal to the test of the
significance for the intercept.
False. The test of significance of the whole relation is testing the joint
significance of the slope coefficients which is different from testing the
significance of the intercept.
20. The more the standard error for the parameters increases the more we can trust
in the confidence intervals of these parameters.
False. The larger the standard error of the estimators, the less we can
trust the confidence intervals of the parameters as an inference of the
true value of the parameters.
For example,
1 = ˆ1 t[( n − no.of . parameters), 2 ] * Se( ˆ1 )
If the standard error of the estimator increases, the boundaries of the
interval also increase. Hence, our confidence on the interval reliability
will decrease.
21. A confidence coefficient of 95% means that the probability for the real
parameters to exist between the lower and upper confidence limit is 95%.
False. A confidence coefficient of 95% means that in 95 out of 100
cases, the estimates are near the true parameter.
22. In the case of multiple regression model, the relation between F and will
be negative.
False. In the case of multiple regression model, there is a positive
relationship between F and where,
𝑅2 ⁄𝑛𝑜.𝑜𝑓 𝑟𝑒𝑔𝑟𝑒𝑠𝑠𝑜𝑟𝑠
𝑓=
1−𝑅2 ⁄𝑛−𝑛𝑜.𝑜𝑓 𝑝𝑎𝑟𝑎𝑚𝑒𝑡𝑒𝑟𝑠
The higher the , the greater is the nominator, the higher is the F value.
4
23. Value is used to compare between different multiple regression models.
False. We use the adjusted -squared ( , instead of to compare
between the multiple regression models which have a different
number of independent variables as it is adjusted for the degrees of
freedom, where
n−1
̅2 = 1 − (1 − R2 )
R
n−no.of parameters
However, there are two conditions to be satisfied in order to use in
comparison:
• The models must have the same dependent variable (i.e. we can't
use to compare models with Y and lnY).
• The models must have the same sample size (n).
̂ 2 (𝑛−𝑛𝑜.𝑜𝑓 𝑝𝑎𝑟𝑎𝑚𝑒𝑡𝑒𝑟𝑠)
𝜎
𝑅2 = 1 − ∑(𝑌𝑖 −𝑌̅)
The higher the is, the greater the nominator will be, hence the lower
the would be, assuming other things holding constant.
26. If the R² differs greatly from R² adjusted, the model may contain unhelpful
predictors. True.
28. A Type I error is the probability of rejecting the null hypothesis when it is
right.
True. A Type I error is the probability of rejecting the null hypothesis
when it is right (ie. It is the probability of incorrectly rejecting the null
5
hypothesis). But a type II error is the probability of failing to reject a
null hypothesis that is wrong.
1
3. The nonlinear model 𝑌 = can be transformed into a linear model
𝛽0 +𝛽1 𝑥1 +𝜀
as ___________.
1
a) y = log β0 + x log β1 b) = log β0 + x log β1 + ε
𝑦
𝟏
c) = 𝜷𝟎 + 𝜷𝟏 𝒙𝟏 + 𝜺 d) log y = log β0 + x log β1
𝒚
4. Under the matrix notation for the classical linear regression model, 𝑌 =
𝑋𝛽 + 𝑢 with n number of observations and k the number of explanatory
variables, what are the dimensions of u?
a) n×k b) n×1 c) k×1 d) 1×1
6
7. What is the meaning of the term "heteroscedasticity"?
a) The variance of the errors is not constant
b) The variance of the dependent variable is not constant
c) The errors are not linearly independent of one another
d) The errors have non-zero mean
Part 2: The following estimated model studies the trade-off between time
spent sleeping and working and to look at other factors affecting sleep:
slêep = 3638.25 − 0.148totwork − 11.13educ + 2.20age
(112.28) (0.017) (5.88) (1.45)
2
where n = 706, R = 0.113 and the standard errors are reported between
parenthesis. Given the above information choose the correct answer.
11. If someone works five more hours per week, the predicted value of sleep
a) becomes 44.4 minutes. b) decreases by 44.4 minutes.
c) increases by 44.4 minutes. d) We cannot decide
7
(note: we must have values for all the explanatory vars to get the predicted
value of the dependent var.)
13. Given the standard error value of the variable education, we can conclude
that at 5% significance level:
a) Education has no significant impact on time spent sleeping.
b) Education has significant impact on time spent sleeping.
c) We cannot conclude if education has significant impact or not on time
spent sleeping.
d) None of the above.
−𝟏𝟏.𝟏𝟑
𝒕 = 𝟓.𝟖𝟖 = −𝟏. 𝟖𝟗 < 𝟐 , we do not reject the null
hypothesis, so var. edu is not significant
14. Degrees of freedom for the F-test for the significance of the model equals:
a) 2, 30 b) 2, 29 c) 2, 32 d) 2, 31
8
16. The 95% confidence interval for the sales elasticity given that t
critical=2.045, equals:
a) (-0.143, 2.311)
b) (1.024, 1.227)
c) (-0.961, 3.129)
d) (0.9613, 1.2067)
̂ − 𝒕𝜶/𝟐 ∗ 𝒔𝒆(𝜷
𝜷 ̂) < 𝜷 < 𝜷 ̂ + 𝒕𝜶/𝟐 ∗ 𝒔𝒆(𝜷
̂)
1.084 - 2.045 (0.06) < 𝜷 < 1.084 + 2.045 (0.06)
t= 0.0217/ 0.0218 = 0.995 < 2 , do not reject the null hypothesis so the var.
prof.marg is not significant at 95% CL.
18. For sales = 70000 and profit margin = 145, the expected value of R&D
spending will be:
a) 4.018586 b) 75878 c) 10423.17 d) 7.4488