Chapter 3
Chapter 3
Introduction
Point estimation is concerned with finding a single value which we think best represents the
unknown population parameter. Suppose we want to a value which best represents the
proportion of the Zimbabwean population owing a vehicle. The best value is the sample
proportion. The population variance is best represented by the sample variance.
If we are to make some inference about some population parameters, we have to have some
observations or sample, 𝑥1 , 𝑥2 , ⋯ , 𝑥𝑛 , from the population of interest. Then based on these
values we can then find approximate probability distribution which we can then use to address
questions relating to the population of interest. The major aim of having these values is to
obtain a statistics 𝛳̂ = 𝛳̂ (𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 ) which we think is good enough to estimate the
population parameter 𝛳 . Once 𝛳 has been estimated then the underlying probability
distribution can be used for further inference about the population or probability distribution.
Definition: An estimator
There are several methods used to find these point estimates. The following section deals with
all the methods used to find these estimates.
The following methods are usually used to find estimators of population parameters.
Example
If you are asked to estimate the average household income for the population in Harare. What
statistic would you use and why?
Solution
The sample mean is the best estimate. A parameter being estimated is the population mean
therefore the sample mean would be the best natural estimate of 𝜇.
Example
Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from the uniform distribution
1
𝑓(𝑥) = , ∝≤ 𝑥 ≤ 𝛽
𝛽−∝
Solution
(a) Since all values of 𝑋 are equal to or greater than ∝, the most sensible estimator of ∝ is
̂ = 𝑚𝑖𝑛𝑖𝑚𝑢𝑚(𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 ) = 𝑋(1)
∝
Since all values of 𝑋 are equal to or less than 𝛽, the most sensible estimator of 𝛽 is 𝛽̂ =
𝑚𝑎𝑥𝑖𝑚𝑢𝑚(𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 ) = 𝑋(𝑛)
𝑛
1
𝐸(𝑋) = ∑ 𝑋𝑖
𝑛
𝑖=1
𝑛
1
𝐸(𝑋 2)
= ∑ 𝑋𝑖2
𝑛
𝑖=1
⋮
𝑛
1
𝐸(𝑋 𝑛)
= ∑ 𝑋𝑖𝑛
𝑛
𝑖=1
then solve the resulting equations for 𝜃. The solution(s) to the equations are the estimate(s)
of 𝜃.
The population parameter can be vector or single parameter. One can use moments about zero
or moments about the mean, the results obtained are the same.
Example
Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from a distribution with density function
Solution
1
(a) 𝐸(𝑋) = ∫0 𝑥. 𝜔𝑥 𝜔−1 𝑑𝑥
1
= ∫ 𝜔𝑥 𝑤 𝑑𝑥
0
𝜔
= [𝑥 𝜔+1 ]10
𝜔+1
𝜔
= 𝜔+1 is the first population moment
1
𝑋̅ = 𝑛 ∑𝑛𝑖=1 𝑋𝑖 is the first sample moment
𝑋̅
𝜔
̂=
1 − 𝑋̅
1 1
(b) 𝑋̅ = 9 ∑𝑛𝑖=1 𝑋𝑖 = 9 (4.3) = 0.4778
0.4778 0.4778
𝜔
̂= = = 0.9149
1 − 0.4778 0.5222
Exercises
1. Find the method of moments estimators of 𝜃 based on a random sample 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛
from each of the following density functions:
(a) 𝑓(𝑥, 𝜃) = 𝜃𝑥 𝜃−1 , 0 < 𝑥 < 1.
(b) 𝑓(𝑥, 𝜃) = (𝜃 + 1)𝑥 −𝜃−2 , 𝑥 > 1, 𝜃 > 0.
(c) 𝑓(𝑥, 𝜃) = 𝜃 2 𝑥𝑒 −𝜃𝑥 , 𝑥 > 0, 𝜃 > 0.
This is the most common method of estimation parameters. This is based on a function called
the likelihood function.
𝑑 𝐿(𝜃)
=0
𝑑𝜃
If one or more solutions exist, then it should be verified which ones maximises 𝐿(𝜃). The value
of 𝜃 that maximises 𝐿(𝜃) also maximises the log-likelihood, ln 𝐿(𝜃). So for computational
convenience the alternative form of the maximum likelihood equation
𝑑
𝑙𝑛𝐿(𝜃) = 0
𝑑𝜃
can be used.
Example
Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from a Poisson distribution
𝑒 −𝜇 𝜇 𝑥
𝑃(𝑥, 𝜇) = , 𝑥 = 1, 2, 3, ⋯
𝑥!
Solution
(a) The likelihood function is given by
𝑒 −𝜇 𝜇 𝑥𝑗
𝐿(𝜇) = ∏
𝑥𝑗 !
𝑒 −𝑛𝜇 𝜇 ∑ 𝑥𝑗
=
(∏ 𝑥𝑗 !)
Taking natural logarithms and then differentiate with respect to 𝜇 and set the derivative
to 0, we obtain
∑ 𝑥𝑗
𝑒 −𝑛𝜇 𝜇
𝑙𝑛 𝐿(𝜇) = 𝑙𝑛 ( (∏ 𝑥𝑗 !)
)
= 𝑙𝑛 𝑒 −𝑛𝜇 + 𝑙𝑛 𝜇 ∑ 𝑥𝑗 − 𝑙𝑛 (∏ 𝑥𝑗 !)
= −𝑛𝜇 + ∑ 𝑥𝑗 − 𝑙𝑛 (∏ 𝑥𝑗 !)
= 𝑛𝜇 + 𝑛𝑥̅ 𝑙𝑛𝜇 − 𝑙𝑛 (∏ 𝑥𝑗 !)
𝑑 ln 𝐿(𝜇) 𝑛𝑥̅
= −𝑛 + =0
𝑑𝜇 𝜇
Example
Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from an exponential distribution
1 −𝑥⁄
𝑓(𝑥, 𝜃) = 𝑒 𝜃, 0 < 𝑥 < ∞
𝜃
Find the likelihood function 𝐿(𝜃) = 𝐿(𝜃, 𝑥1 , 𝑥2 , ⋯ , 𝑥𝑛 ).
Solution
1 𝑥⁄
(a) 𝐿(𝜃) = ∏𝑛𝑗=1 (𝜃 𝑒 − 𝜃)
1 𝑛 − ∑𝑥𝑗⁄
=( ) 𝑒 𝜃
𝜃
Taking natural logarithms differentiate and equate to 0 and solve the resulting equation
for 𝜃
∑𝑥 𝑑𝑙𝑛 𝐿(𝜃) −𝑛 ∑𝑥
𝑙𝑛 𝐿(𝜃) = −𝑛𝑙𝑛𝜃 − and = + =0
𝜃 𝑑𝜃 𝜃 𝜃2
𝜃̂ = 𝑥̅
There are cases where the MLE exists but cannot be obtained as a solution of the maximum
likelihood equation. The example below illustrates such a scenario.
Example
Solution
𝐿(𝜃) = ∏ 𝑒 𝑥𝑗 −𝜃
= 𝑒 ∑ 𝑥𝑗−𝑛𝜃
= 𝑒 𝑛(𝑥̅ −𝜃)
(b) 𝐿(𝜃)
𝒙(𝟏)
(c) From the graph the likelihood attains its maximum at 𝒙(𝟏) . Thus the MLE for 𝜃 is
̂ = 𝒙(𝟏) .
𝜽
Exercises
1. Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from a geometric distribution
1 1 𝑥
𝑓(𝑥, 𝜃) = (1 − ) ( ) , 𝑥 = 1, 2, ⋯
𝜃 𝜃
Find the MLE for 𝜃.
2. Let 𝑋1 , 𝑋2 , ⋯ , 𝑋𝑛 be a random sample from a gamma distribution
1 𝑥⁄
𝑓(𝑥, 𝜃) = 𝜃2 𝑥𝑒 − 𝜃, 𝑥 > 0.
In certain types of models, the principle of least squares is very important. We assume that the
mean of the random variable 𝑌 is a linear function of 𝑝 unknown vector of parameters 𝜷 =
(𝛽1 , 𝛽2 , ⋯ , 𝛽𝑝 ) and 𝑝 factors 𝒙 = (𝑥1 , 𝑥2 , ⋯ , 𝑥𝑝 ) that can be fixed and measured without error.
We assume that 𝑉𝑎𝑟(𝑌) = 𝜎 2 , where 𝜎 2 is constant and does not depend on 𝑥𝑖 ′𝑠.
𝑛 𝑛
2
∑𝑒 = ∑( 𝑌𝑖 − 𝛽0 −𝛽1 𝑥1𝑖 − 𝛽2 𝑥2𝑖 − ⋯ 𝛽𝑝 𝑥𝑝𝑖 )2
𝑖=1 𝑖 𝑖=1
We differentiate ∑𝑛𝑖=1 𝑒 2 𝑖 with respect to all the 𝛽𝑖 ′𝑠 then solve the resulting equations for ̂
𝜷=
Example
𝑛
𝛿𝑄
= −2 ∑(𝑦𝑖 − 𝛼 + 𝛽𝑥𝑖 )
𝛿𝛼
𝑖=1
𝑛
𝛿𝑄
= −2 ∑ 𝑥𝑖 (𝑦𝑖 − 𝛼 + 𝛽𝑥𝑖 )
𝛿𝛽
𝑖=1
Set the derivatives to zero and solve the simultaneous equations for 𝛼 and 𝛽 we obtain
∑ 𝑥𝑖 𝑦𝑖 − (∑ 𝑥𝑖 )(∑ 𝑦𝑖 )
𝛽̂ =
∑ 𝑥𝑖2 − (∑ 𝑥𝑖 )2
and
∑ 𝑦𝑖 − 𝛽̂ ∑ 𝑥𝑖
𝛼̂ =
𝑛
Exercises
Consider the data on monthly Income ($) and Savings ($) given in the table below.
Income 980 1500 420 198 785 2800 3845 1890 4210
Suppose that Income (X) and savings (Y) are related through a relationship with the form
𝑦 =∝ +𝛽𝑥.