Professional Documents
Culture Documents
Contents
Course No:
1. Random Variables and Vectors
2. Random Variables and Vectors, Intro to Estimation
3. MVU, Cramer Rao Lower Bound
4. Linear Models, Best Linear Unbiased Estimator
5. Least Squares and Maximum Likelihood Estimation
6. Introduction to Bayes Estimation
7. Linear Bayes Estimation
8. Linear Bayes Estimation, Stochastic Processes
9. Stochastic Processes
10. Wiener Filter
11. Wiener Filter
12. Least Squares Filter
Page 2
Minimum Variance Unbiased Estimator (MVU)
ˆ
let e = θ −θ then E ((θ − θˆ) 2
) = σ 2
− µ
2
e e
with µe and σe2 as the mean and variance of the estimation error e,
respectively, µe is also called the bias of the estimator.
For a minimum variance unbiased estimatior (MVU) our aim is to find
an estimator that has zero bias (µe =0), i.e. E(θˆ) = θ with minimum
variance σe2 of the error
Please note a MVU does not necessarily is the estimator with the
minimum mean squared error, in fact there might be an estimator
that is biased (µe ≠0) but has a lower mean squared error.
However, from a practical point of view, MUVs are often easier to
find that estimators minimizing the mean squared error.
Often it is even impossible to find an MVU. In this case we even have
to use additional constraints, e.g. to find best linear unbiased
estimator (BLUE)
Page 3
Minimum Variance Unbiased Estimator (MVU)
Please note that for an MVU estimator, σe2 must be minimum for every θ!!!
Taken from Kay: Fundamentals of Statistical Signal Processing, Vol 1: Estimation Theory, Prentice Hall, Upper Saddle River
Two questions:
Does the MVU exist?
How to find it?
Possible options to answer these questions:
Determining a lower bound on the estimator’s variance and
checking if there is an estimator satisfying this bound
Further restricting the estimator, e.g. to be linear
Rao-Blackwell-Lehmann-Scheffe Theorem, Maximum likelihood
estimator,…
Page 5
Cramer-Rao-Lower Bound
Page 6
Cramer-Rao-Lower Bound
∂2 ln p(x;θ ) N −1 ∂ 2 ln p( x[n];θ )
− E
2 = −∑ E 2
∂ θ n =0 ∂θ
Taken from Kay: Fundamentals of Statistical Signal Processing, Vol 1: Estimation Theory, Prentice Hall, Upper Saddle River
Page 9
Cramer-Rao-Lower Bound
1 1 N −1 2
p(x; A) = N
exp − 2 ∑( x[n] − A)
2 2 2σ n=0
(2πσ )
Taking the first derivative
∂ ln p(x; A) ∂ N
1 N −1 2 1 N −1
= − ln[(2πσ ) ] − 2 ∑( x[n] − A) = 2 ∑( x[n] − A)
2 2
∂A ∂A 2σ n=0 σ n=0
N
= 2
( x − A)
σ N −1
1
with g ( x) = x =
N
∑ x[n]
n =0
as the sample mean.
Page 10
Cramer-Rao-Lower Bound
Page 11
General CRLB for Signals in White Noise
2 2
(2πσ )
Differentiating once produces:
∂ 2 ln p(x;θ ) 1 N −1
∂s[n;θ ]
∂θ 2
= 2
σ
∑
n =0
( x[n] − s[n;θ ])
∂θ
and a second differentiation results in:
∂ 2 ln p(x;θ ) 1 N −1 ∂ 2 s[n;θ ] ∂s[n;θ ]
2
∂θ 2
=− 2
σ
∑
n =0
( x[n] − s[n;θ ])
∂θ 2
−
∂θ
Page 12
General CRLB for Signals in White Noise
Page 13
Transformation of Parameters
Page 14
Transformation of Parameters
But: be carefull!
Non-linear transformations destroy the efficiency
e.g. DC Level in WGN, Estimator for A2
◊ Square of the sample mean: x 2 might be a resonable estimator
◊ But is not even unbiased anymore
σ2
E ( x 2 ) =E 2 ( x ) + var( x ) =A2 + ≠ A2
N
On the other hand: affine (linear) transformations
∧
g (θ ) = g (θˆ) = aθˆ + b
preserve efficiency!
Page 15
CRLB for Vector Paramters
[
var(θˆi ) ≥ I −1 (θ ) ] ii
Page 16
CRLB for Vector Paramters
∂ ln p(x; θ)
= I(θ)(g (x) − θ)
∂θ
1 1 N −1 2
p(x; θ) = N
exp − 2 ∑( x[n] − A − Bn)
2 2 2σ n=0
(2πσ )
Page 18
Example: Line Fitting
and ∂ 2 ln p(x; θ) N
2
=− 2
∂A σ
∂ 2 ln p(x; θ) 1 N −1
= − 2 ∑n
∂A∂B σ n=0
∂ 2 ln p(x; θ) 1 N −1 2
2
= − 2 ∑n
∂B σ n=0
N −1
N ( N −1)
1 N ∑ n N
2
I(θ) = − 2 N −1 n=0
=
σ n N −1
2 N ( N −1) N ( N −1)(2N −1)
∑ ∑ n
n=0 n=0
2 6
Page 19
Example: Line Fitting
This means 2
2(2 N − 1)σ
var(Aˆ ) ≥
N ( N +1)
2
12 σ
var(Bˆ ) ≥
N ( N 2 −1)
Important Observations:
The CRLB of A is higher than of a single DC Level in WGN
General Result: The CRLB always increases with the number
of estimated paramters!
Page 20
Example: Line Fitting
with
N −1 N −1
2(2 N − 1) 6
Aˆ = ∑
N ( N +1) n=0
x[n] − ∑
N ( N +1) n=0
nx[n]
N −1 N −1
6 12
Bˆ = − ∑
N ( N +1) n=0
x[n] + 2 ∑
N ( N −1) n=0
nx[n]
Page 22