Estimation 3
Estimation 3
Ajit K Chaturvedi
This leads to
Z ∞ Z ∞
c f (y )dy = c = yf (y )dy = E {y}
−∞ −∞
2 / 12
Nonlinear MS Estimation
Now we wish to estimate y not by a constant but by a
function c(x) of the random variable x.
We need to find the function c(x) that will minimize the MS
error e:
Z ∞Z ∞
2
[y − c(x)]2 f (x, y )dxdy
e = E [y − c(x)] =
−∞ −∞
Hence e is minimum if
Z ∞
c(x) = E {y | x} = yf (y | x)dy
−∞
4 / 12
Linear MS Estimation
5 / 12
Linear MS Estimation Contd.
e = E [y − (Ax + B)]2
B = E {y − Ax} = ηy − Aηx
6 / 12
Linear MS Estimation
With B so determined, we get
where
E {(x − ηX )(y − ηY )}
r=
σx σy
e is minimum if
µ11
A = r σy /σx =
µ20
where n o
µkr = E (x − ηx )k (y − ηy )r
Inserting this A into the preceding quadratic, we obtain
µ11 2
em = σy2 (1 − r 2 ) = µ02 −
µ20
7 / 12
Terminology
8 / 12
Fundamental Note
In general, the nonlinear estimate φ(x) = E [y|x] of y in terms
of x is not a straight line and the resulting MS error
E {[y − φ(x)]2 } is smaller than the MS error em of the linear
estimate Ax + B.
However, if the random variables x and y are jointly normal,
φ(x) is a straight line with slope r σy /σx and passing through
the point (ηx , ηy ). It is expressed as:
r σy x r σy ηx
φ(x) = + ηy −
σx σx
To verify the above expression, write down the pdf f (x, y ) for
jointly normal random variables x and y. From this, obtain the
conditional pdf f (y |x) and note that this is a normal density.
Hence E [y|x] is just the value at which the argument of the
exponential term is zero.
To conclude, for normal random variables, nonlinear and linear
MS estimates are identical.
9 / 12
The Orthogonality Principle
Recall
e = E [y − (Ax + B)]2
e = E (y − ax)2
is minimum.
We get the desired result by differentiating wrt a and equating
it to 0. Thus,
E {(y − ax)x} = 0
Therefore,
E {xy}
a=
E {x2 }
There is an interesting alternate proof.
11 / 12
Alternate Proof
= E (y − ax)2 + (a − ā)2 E x2
E (y − āx)2 ≥ E (y − ax)2
12 / 12