Econ 103 2023 MidTerm Practice1 - With - Solutions
Econ 103 2023 MidTerm Practice1 - With - Solutions
Instructions:
This is a 1 hour and 10 minute exam. You are allowed to use the equations sheet provided
in the site of the course and handwritten notes.
You are not allowed to use any electronic device. No connection to the internet via WiFi or
any other method is allowed. It is not permitted to use any kind of mobile phone.
When you are finished with the exam, please turn in the exam questions.
Cheating of any form will result in a score of 0 (zero) for the exam, in addition to the normal
university disciplinary action.
Please sign below that you have read, understood, and fulfilled all of the above instructions
and conditions.
First Name
Last Name
UCLA ID #
Signature
Exam Version A
Please start solving the examinations only when you are instructed to do so.
Please stop immediately when instructed to do so.
Good Luck!
Part I (Questions based on Regression Output):
‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
wage | Coef. Std. Err. t P>|t| [95% Conf. Interval]
‐‐‐‐‐‐‐‐‐‐‐‐‐+‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
educ3| 0.05 0.0050 10.00 0.000 0.0400 0.0600
_cons | 100.00 10.0000 10.00 0.000 80.000 120.000
‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
(a) 15
(b) 50
(c) 100
(d) 150
(e) 500
Answer: d.
Page 1
Question 2. What is the marginal effect of another year of education for a person with 10
years of education?
(a) 1.5
(b) 3
(c) 5
(d) 10
(e) 15
Answer: e.
∆E(Y )
δ= = 3 · b̂2 · x20
∆x
= 3 · 0.05 · 102
= 3 · 0.05 · 100
= 3 · 5 = 15
Question 3. What is the estimated elasticity for a person with 10 years of education at his
expected wage?
(a) 0.1
(b) 0.5
(c) 1
(d) 1.5
(e) 2
Answer:c.
∆E(Y ) x0 x0
η= · =δ·
∆x
| {z } ŷ ŷ
δ
10
= 15 ·
150
=1
3
Questions 4–12 are based on the following regression output
Consider the quadratic model:
‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
wage | Coef. Std. Err. t P>|t| [95% Conf. Interval]
‐‐‐‐‐‐‐‐‐‐‐‐‐+‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
educ | ‐20.000 50.0000 ‐0.40 0.686 ‐120.0000 80.00000
educ2 | 2.000 2.0000 1.00 0.448 ‐2.0000 6.00000
_cons | 200.000 200.0000 1.00 0.422 ‐200.0000 600.0000
‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
Question 4. What is the expected wage for a person with 10 years of education?
(a) 10
(b) 20
(c) 100
(d) 200
(e) 2000
Answer: d.
Page 1
Question 5. What is the marginal effect of another year of education for a person with 10
years of education?
(a) 200
(b) 100
(c) 2
(d) 10
(e) 20
Answer: e.
∆E(Y )
δ= = b̂2 + 2 · b̂3 · x0
∆x
= −20 + 2 · 2 · 10
= −20 + 40 = 20
Question 6. What is the estimated elasticity for a person with 10 years of education at his
expected wage?
(a) 0.1
(b) 0.5
(c) 1
(d) 1.5
(e) 2
Answer: c.
∆E(Y ) x0 x0
η= · =δ·
∆x
| {z } ŷ ŷ
δ
10
= 20 ·
200
=1
5
Question 7. Parameter λ0 is defined as a linear combination of the model coefficients:
β1 β3
λ0 = − .
200 2
What is the value of the Best Linear Unbiased Predictor (BLUP) for λ0 ?
(a) 20
(b) 1
(c) 10
(d) 2
(e) 0
Answer: e.
b̂1 b̂3
λ̂ = −
200 2
200 2
= −
200 2
=1−1
=0
2 2
1 1 1 1
var(λ)
ˆ = var(
ˆ b̂1 ) + ˆ b̂3 ) + 2 ·
var( · − · cov(b
ˆ 1 , b3 )
200 2 200 2
1 2
2
2 1 2 1 1
= · 200 + ·2 +2· · − · 200
200 2 200 2
=1+1−1
=1
p √
Thus we have that se(λ)
ˆ = var(λ)
ˆ = 1=1
Question 9. What is the test statistic for the null hypothesis H0 : λ0 = 0 against the
̸ 0.
alternative H1 : λ0 =
6
(a) -1
(b) 0
(c) 1
(d) 2
(e) 10
Answer: b.
λ̂−0 0−0
Test statistic: t̂ = se(λ)
ˆ = 1 =0
Question 10. Using the critical value of tc = 2, the confidence interval for λ is given by:
(a) [−1, 0]
(b) [0, 1]
(c) [0, 2]
(d) [−2, 2]
(e) [−1, 1]
Answer: d.
Question 11. A econometrician spotted a typo in the p-values of the regression output. Let
p1 = 0.422 be the p-value associated with intercept β1 .
Let p2 = 0.686 be the p-value associated with intercept β2 (for educ).
Let p3 = 0.448 be the p-value associated with intercept β3 (for educ2)
Given the estimated coefficients, and its respective standard errors, we should have that:
Answer: e.
Note that t̂1 = t̂3 ⇒ p̂1 = p̂3
Moreover the absolute value of the t-statistic t̂2 is smaller than the value for the t-statistic t̂2 , t̂3 ,
that is: |t̂2 | = 0.4 < 1 = |t̂1 | = |t̂3 |.
Therefore we must have that: p̂2 > p̂1 = p̂3 .
7
Part II (Multiple Choice Questions that do not use Regression Output):
Question 12. Let X1 , X2 be two random variables and take values in {0, 1} and whose joint
distribution is given by:
Joint Distribution
X1 = 0 X1 = 1 fX2 (x)
X2 = 0 0.2 0.2 0.4
X2 = 1 0.3 0.3 0.6
fX1 (x) 0.5 0.5 1
(e) Cov(X1 , X2 ) ̸= 0
Answer: a.
X1 , X2 are statistically independent because P (X1 = x1 , X2 = x2 ) = P (X1 = x1 )P (X2 = x2 ) for
all x1 , x2 ∈ {0, 1}.
Independence implies that Cov(X1 , X2 ) = 0. Indeed, we have that:
Also:
P (X2 = 1, X1 = 1) 0.3
P (X2 = 1|X1 = 1) = = = 0.6 = P (X2 = 1),
P (X1 = 1) 0.5
and
P (X2 = 0, X1 = 0) 0.2
P (X2 = 0|X1 = 0) = = = 0.4 = P (X2 = 0).
P (X1 = 0) 0.5
8
Question 13. Let X1 ∼ N (1, 1) and X2 ∼ N (2, 4) be two normally distributed random
variables. Let the correlation between X1 and X2 be ρ. Mark the choice that is correct:
(d) If X1 and X2 correlate, then their covariance may take any value in the real line [−∞, ∞].
(e) We have that E (X1 − 1) · (X2 − 2) = 0 regardless if X1 , X2 correlate.
Answer: c.
Observe that Z1 , Z2 are standard normal random variables that might be potentially correlated.
Regardless of its correlation, we still have that E(Z12 + Z22 ) = E(Z12 ) + E(Z22 ).
Now observe that Z1 , Z2 are the standardized variables of X1 , X2 . Thus they have means zero and
variance one. Now when a variable X has mean zero, it is always the case that E(X 2 ) = V ar(X).
Therefore we have that:
Let’s compute the mean and variance of these variables for sake of completeness.
Variables Z1 , Z2 have mean zero because:
X1 − 1 E(X1 ) − 1 1−1
E(Z1 ) = E = = = 0,
1 1 1
X2 − 2 E(X2 ) − 2 2−2
E(Z2 ) = E = = = 0.
2 2 2
In letter (a), the addition of a constant term does not modify the variance.
In letter (b), the probabilities are different because the random variables defined as (X1 − 1)
and (X2 − 2) have mean zero but different variances.
9
In letter (d), consider two random variables X1 , X2 such that var(X1 ) > V ar(X2 ), that is, X1
has the bigger variance. Then we have that:
−1 ≤ Corr(X1 , X2 ) ≤ 1
Cov(X1 , X2 )
−1 ≤ p ≤1
V ar(X1 )V ar(X2 )
Cov(X1 , X2 )
−1 ≤ ≤1
se(X1 )se(X2 )
−se(X1 )se(X2 ) ≤ Cov(X1 , X2 ) ≤ se(X1 )se(X2 )
Note that it is clear that the covariance between two variables is bounded by ±se(X1 )se(X2 ),
which falsifies letter d. Now if V ar(X1 ) > V ar(X2 ) then se(X1 )se(X1 ) > se(X1 )se(X2 ), then it
is also true that:
In summary, it is always the case that the covariance of two variables is bounded by positive and
the negative value of the largest variance.
Cov(X1 , X2 ) p
ρ= p ⇒ Cov(X1 , X2 ) = ρ · V ar(X1 )V ar(X2 ).
V ar(X1 )V ar(X2 )
10
Question 14. Consider the following regression model:
Yi = β1 + β2 xi + ϵi ,
for i = 1, ..., N . Let ei ∼ 0, σi2 . That is, ei has a distribution whose mean is 0 and its variance
is σ 2 . Let
N N
1 X 1 X
s2x = (xi − x)2 , where x = xi . (1)
N N
i=1 i=1
A data analyst ran a regression of y on x and obtained the following estimates for β1 and β2 :
b̂1 = 4, b̂2 = .5. Define x∗i = 10 × xi . If one were to run a regression of y on x∗ the estimates b̂∗1
and b̂∗2 would be
Answer: c.
b̂2
x-transformation x∗ = c · x only changes the estimate for β2 , that is, b̂∗2 = c . the values of the
t-statistics, the inference and the R2 remain the same.
Question 15. Regarding the Simple Regression Model y = β1 +β2 ·x+e, which of the following
is FALSE?
(b) The LS estimates for the quadratic regression Y = β1 + β2 · x2 + e is not BLUE because the
relation between Y and x is not linear, so the linearity assumption is violated.
cov(x,y)
(c) In the Simple Regression Model, y = β1 + β2 · x + e, b̂2 = var(x) , where cov(x, y) is the
sample covariance and var(x) is the sample variance of x.
(d) Let b̂∗1 , b̂∗2 be estimates for β1 , β2 other than the least squares estimates b̂1 , b̂2 , then it must
be that:
XN N
X
(b̂∗1 + b̂∗2 xi − yi )2 ≥ (b̂1 + b̂2 xi − yi )2 .
i=1 i=1
(e) The sign of the covariance between b̂1 , b̂1 depends only on the sample mean of x.
Answer: b.
11
Letters a, c display the standard equations that are used to estimate parameters in the simple
regression model.
X
b̂1 , b̂2 = min (yi − β1 − β2 )2
β1 ,β2
i
PN 2
Thus the sum of the squared residuals, i=1 (b̂1 + b̂2 xi − yi ) is the lowest value that can be
obtained across all possible estimates of β1 and β2 . This means that the sum square of residuals
for any other estimates for β1 and β2 , say b̂∗1 and b̂∗2 , must be bigger than the sum of the square of
residuals computed using the LS estimates b̂1 , b̂2 . In other words, we have that
N
X N
X
(b̂∗1 + b̂∗2 xi − yi )2 > (b̂1 + b̂2 xi − yi )2 ,
i=1 i=1
for any values b̂∗1 , b̂∗2 that differ from b̂1 , b̂2 .
Letter e is correct because the formula for the covariance between estimators b1 , b2 depends on
−x̄ and other terms that are always positive.
Question 16. Let the simple regression model Y = β1 + β2 · X + e. Consider the inference
that tests the null hypothesis H0 : β2 = 0 against H1 : β2 ̸= 0 at significance level α. Which of the
statements is false?
(a) If the standard error of b̂2 decreases, then it is more likely to reject H0 , (everything else
constant).
(b) The higher the absolute value of b̂2 , the more likely it is to reject H0 (everything else con-
stant).
(c) The higher the significance level α, the more likely it is to reject H0 (everything else constant).
(d) Hypothesis H0 : β2 = 0 is not rejected whenever the value 0 belongs to its confidence interval
(with confidence level of 1 − α).
(e) The larger the sample size, the more likely it is to reject H0 , (everything else constant).
(f) The higher the p-value, the more likely you are to reject H0 .
Answer f.
12
The question requires you to interpret the equations for the estimator standard errors in the
simple regression model:
1 1
V ar(b2 ) = σ 2 · ·
N var(x)
1 var(x) + x̄2
V ar(b1 ) = σ 2 · ·
N var(x)
1 −x̄
Cov(b1 , b2 ) = σ 2 · ·
N var(x)
For a given p-value, the larger the significance level α, the more likely you are to reject the null
hypothesis.
Question 17. Let the simple regression model Y = β1 + β2 · X + e. Consider the inference
that tests the null hypothesis H0 : βk = 0 against H1 : βk ̸= 0 at significance level α for k = 1, 2.
Which of the statements is correct?
(a) It is less likely to reject the null hypothesis H0 : β1 = 0 if we do the transformation ynew = y·c.
(b) It is less likely to reject the null hypothesis H0 : β2 = 0 if we do the transformation xnew = x·c.
(c) It is less likely to reject the null hypothesis H0 : β2 = 0 if we do the transformation ynew =
y + c.
13
(d) It is less likely to reject the null hypothesis H0 : β1 = 0 if we standardize both y and x.
(e) It is less likely to reject the null hypothesis H0 : β2 = 0 if we standardize both y and x.
Answer d.
The question explores the properties of linear transformations of the dependent variable y and
the explanatory variable x. First, it is useful to revise what standardize means. The standardized
version x̃ of a variable x is obtained by subtracting x by its sample mean and dividing it by its
standard deviations, namely:
x̃ = − x̄sd(x).
x
As a consequence, x̃ has mean zero, variance one, and standard deviation one. If both x and y are
standardized, then the estimated slope b̂2 of the simple regression model is equal to the correlation
between x and y. The estimate of the intercept b̂1 is zero.
A summary of the properties of the LS estimates under linear transformations of x and y is
listed below:
The transformation xnew = x·c modifies the estimates b̂2 and se(bˆ 2 ). The estimates b̂1 , se(b)
ˆ 1,
the t-statistics for testing coefficients β1 , β2 and R2 remain the same.
The transformation ynew = y · c modifies the estimates b̂1 , se(b ˆ 2 ), σ̂ 2 . The inference,
ˆ 1 ), b̂2 , se(b
2
t-statistics for testing coefficients β1 , β2 and R remain the same.
The transformation ynew = y + c modifies the estimates b̂1 but does not modifies se(b ˆ 1 ), so
it affects its t-statistic. The transformation does not change the estimates b̂2 , se(b)
ˆ 2 , t̂2 and
R2 .
If x̄ = ȳ = 0 then the estimate b̂1 = 0, t̂1 = 0 and we never reject the null hypothesis
H0 : β1 = 0.
If we standardize both y and x, then the estimate b̂1 is zero and the estimate b̂2 is equal to
the sample correlation of x, y.
In more general terms, you must be able to read the following table:
14
(a) Note that the sample means of ỹ and x̃ are zero, that is ỹ = 0 and x̃ = 0. Thus the estimated
ˆ
value of b̃1 is given by:
ˆ ˆ
b̃1 = ỹ − b̃2 x̃ = 0.
ˆ ˆ
(b) Moreover the estimated t-statistic associated with b̃1 must be zero as t̃ˆ1 = b̃1 /se(
b b̃) = 0.
(c) Note also that a linear transformation of two random variables X, Y does not change its sample
correlation. But the goodness of fit is simply the square of the correlation between variables.
Thereby the goodness of fit of the original regression and the one that uses the transformed
variables must be the same.
(d) Lastly, note that the sample variance of each of the transformed variables is one. Thus the
covariance of the transformed variables is equal to its correlation.
Answer e.
Letter e is true because the goodness of fit, R2 can be expressed as R2 = (corr(x, y))2 . The
correlation between x and y are the same if we regress Y on X or vice-versa.
cov(x,y) cov(x,y)
Letter a is false because b̂y2 = var(x) while b̂x2 = var(y) . For b̂x2 = b̂y2 to be true, we would need
that var(x) = var(y).
Letter b is false because b̂y2 = cov(x,y) x var(y)
var(x) while 1/b̂2 = cov(x,y) . Note that one could mistakenly
think that the letter is correct by some erroneous rationale. For instance, one could think on
isolating X in the equation: Y = β1 + β2 X + ϵ. This would generated the following equation
X = ββ21 + β12 X − β12 ϵ, which suggests the wrong statement that b̂y2 = 1/b̂x2 .
A simple way to see that letter c is wrong think of a variable x such that x̄ = 0. In this case,
b̂y1 = ȳ − b̂y2 x̄ = ȳ.
On the other hand, b̂x1 = x̄ − b̂x2 ȳ = −ȳ · cov(x,y)
var(y) , which falsifies the letter. In more general terms,
15
for b̂y1 = b̂x1 we need that:
ȳ − b̂y2 x̄ = x̄ − b̂x2 ȳ
⇒ȳ + b̂x2 ȳ = x̄ + b̂y2 x̄
⇒ȳ(1 + b̂x2 ) = x̄(1 + b̂y2 )
ȳ 1 + b̂x2
⇒ =
x̄ 1 + b̂y
|{z}
Depends on Sample Means
| {z 2}
Depends on Sample Covariance
The statement is wrong because the left-hand side depends only on sample means while the
right-hand side of the equation depends on the sample covariances and we can change sample
means without changing the sample covariances by simple adding constants to x or y.
A simple way to check that Letter d is wrong consists in investigating the case where x̄ = 0,
ȳ > 0 and cov(x, y) > 0. Recall that, if x̄ = 0, then b̂y1 = ȳ and b̂x1 = −ȳ · cov(x,y)
var(y) . In this case,
b̂y1 > 0 while b̂x1 < 0 and thereby it cannot be the case that t̂y1 = t̂x1 .
(a) The lower the absolute value of the sample correlation, the larger the standard errors for
the estimators b2 and b3 .
(b) Any linear transformation of x2 changes the correlation between x2 and x3 and thereby
affects the standard error of estimator b3 .
(c) Re-scaling x2 say x∗2 = 5 · x2 changes the standard error of estimator b2 but does not change
b 2 ) = b̂∗2 /se(b
its t-statistic t̂2 = b̂2 /se(b b ∗2 ) = t̂∗2 .
(d) The estimators b2 , b3 would be most precise if the explanatory variables were equal, that is,
x2 = x3 .
(e) If the sample means x̄2 and x̄1 were zero, that is, x̄2 = x̄1 = 0, then the estimate of the
intercept is also zero, b̂1 = 0.
Answer c.
Letter c states a correct consequence of the linear transformation (see question 17)
Letter e is false because if x̄2 = x̄1 = 0, then b̂1 = ȳ.
The remaining of the items of the question explore the fact that, in multiple regression models,
the higher the correlation between explanatory variables, the higher the estimated standard error
of the leas squares estimates for β2 , β3 .
In is useful to investigate the formula for the variance of estimators b2 and b3 . The formula for
variance of estimators b2 is given by:
1 1 1
V ar(b2 ) = σ 2 · · · 2 )
N var(x2 ) (1 − r2,3
16
The interpretation of this formula is the following :
The first term (σ 2 ) means that the larger the variance of the error term, the larger the
variance of error term V ar(ei ) = E(e2i ) = σ 2 , the larger the variance of V ar(b2 ).
The second term (1/N ) means that the larger the sample size N, the smaller the variance
of V ar(b2 ).
The Third term (1/var(x2 )) means that the larger the sample variance of exploratory vari-
able x2 , the smaller the variance of V ar(b2 ).
The fourth term (1/(1 − r2,3 2 )) means that the larger the sample correlation of exploratory
The first term (σ 2 ) means that the larger the variance of the error term, the larger the
variance of error term V ar(ei ) = E(e2i ) = σ 2 , the larger the variance of V ar(b2 ).
The second term (1/N ) means that the larger the sample size N, the smaller the variance
of V ar(b2 ).
The Third term (1/var(x3 )) means that the larger the sample variance of exploratory vari-
able x3 , the smaller the variance of V ar(b2 ).
The fourth term (1/(1 − r2,3 2 )) means that the larger the sample correlation of exploratory
Note that the last term is identical for both V ar(b2 ) and V ar(b3 ). The correct answer is (a). The
most important cases when the correlation is mention in mentioned in the course are:
The increase in the sample correlation among explanatory variables (x2 , x3 ) in the mul-
tiple regression increases the variance of the estimators b2 , b3 .
The square of the sample correlation between the dependent variable y and exploratory
variable x in the simple regression is the goodness of fit R2 . Thus an increase correlation
between the dependent variable y and exploratory variable x in the simple regression
increases the R2 .
ln(Yi ) = β1 + β2 xi + ei
where x denotes annual household income (in thousands) and y denotes annual expenses on con-
sumption goods. Let ŷ be the estimated value of Y given the value of the explanatory variable x0 .
17
Consider the following estimates:
γ̂ x = b̂2 x0
γ̂ y = b̂2 ŷ
x0
γ̂ x,y = b̂2
ŷ
A general question on this topic requires you to use the information of the following non-linear
4 . 3 M O D Evariable
transformations of both the explanatory variable x and the dependent L I N G IyS :S U E S 143
2.8.3–2.8.4 and will be further discussed in Section 4.5. Note its possible shapes in
Figure 4.5(e). If b2 > 0 the function increases at an increasing rate; its slope is larger
18
for larger values of y. If b2 < 0, the function decreases, but at a decreasing rate.
3. In the linear-log model y ¼ b þ b lnðxÞ the variable x is transformed by the
The first column present a range of models.
The second column presents the function form of the model.
The third column presents the the equation for the slope.
The fourth column presents column presents the the equation for the elasticity.
The interpretation of the slope and elasticity are: The two most important estimates are the slope
and the elasticity.
Elasticity = ∆Y
∆X · X
Y : If the estimate value of the slope is η̂ then:
19