Professional Documents
Culture Documents
Reading:
• Chapter 3 in Kay-II,
where
X
Recall that a binary decision rule φ(x) maps |{z} to {0, 1}:
data space
1, decide H1,
φ(x) =
0, decide H0.
PFA(φ(X), θ) = E X | Θ[φ(X) | θ]
Z
= fX | Θ(x | θ) dx for θ ∈ spΘ (0) (1)
X1
Z
PFA(φ(X), θ0) = fX | Θ(x | θ0) dx
X1
Comments:
(iv) However, the perfect case where our rule is always right
and never wrong (PD = 1 and PFA = 0) cannot occur when
the conditional pdfs/pmfs fX | Θ(x | θ0) and fX | Θ(x | θ1)
overlap.
Setup:
H0 : θ = θ0 versus
H1 : θ = θ1.
where
fX | Θ(x | θ1)
Λ(x) =
fX | Θ(x | θ0)
is the likelihood ratio. The values x that satisfy Λ(x) = λ
can be allocated to either X1 or X0. To completely specify the
optimal test, we need to select
and
1, Λ(x) > λ
φL,λ(x) = 0, Λ(x) = λ .
0, Λ(x) < λ
Define
f0 = FΛ | Θ(0 | θ0) = PrΛ | Θ{Λ ≤ 0 | θ0}.
Recall that cdf FΛ | Θ(l | θ0) must be nondecreasing and right-
continuous, but may have discontinuities.
(i) When
1 − α < f0 i.e. 1 − f0 < α (7)
we select the threshold λ = 0 and apply the rule
1, Λ(x) > 0
φL,0(x) = . (8)
0, Λ(x) = 0
1 − α ≥ f0 i.e. 1 − f0 ≥ α (9)
the KKT condition (5) holds, and the test (10) is optimal.
An example of this case corresponds to λ2 and 1 − α2 in the
above figure.
1 − α ≥ f0 i.e. 1 − f0 ≥ α
where FΛ | Θ(λ− | θ0) and FΛ | Θ(λ+ | θ0) denote the left and
right limits of FΛ | Θ(λ | θ0) at l = λ. If this case happens
in practice, we can try to avoid the problem by changing
our specified α, which is anyway not God-given, but
chosen rather arbitrarily. We should pick a value of α
that satisfies the KKT condition.
α − PFA(φL,λ(x), θ0)
p =
PFA(φU,λ(x), θ0) − PFA(φL,λ(x), θ0)
PFA(φλ(x), θ0)
= PFA(φL,λ(x), θ0) + p [PFA(φU,λ(x), θ0) − PFA(φL,λ(x), θ0)] = α.
is optimal.
PFA(0) = PD(0) = 1.
dPD(τ )
= −fΛ | Θ(τ | θ1)
dτ
dPD(τ )
= −fΛ | Θ(τ | θ0)
dτ
implying
dPD(τ ) fΛ | Θ(τ | θ1)
= = τ.
dPFA(τ ) fΛ | Θ(τ | θ0)
In summary,
ROC Property 2. If the likelihood ratio is a continuous
random variable given θ, the slope of ROC at point
(PFA(τ ), PD(τ )) is equal to the threshold τ of the corresponding
likelihood-ratio test.
In particular, this result implies that the slope of ROC is
• τ = +∞ at (0, 0) and
• τ = 0 at (1, 1).
PD(τ ) ≥ PFA(τ ).
fX | µ(x | µ) = N (x | µ, C)
1
= p exp[− 21 (x − µ)T C −1 (x − µ)]
|2 π C|
fX | µ(x | µ1)
Λ(x) =
| {z } fX | µ(x | µ0)
likelihood ratio
H1
T −1 T −1
− 12 (x − µ1) C (x − µ1) + 21 (x − µ0) C (x − µ0) ≷ ln τ
i.e.
H1
T −1 1
(µ1 − µ0) C [x − (µ0 + µ1)] ≷ ln τ.
2
and, finally,
H1 4
−1
T (x) = s C T
x ≷ ln τ + 12 (µ1 − µ0)T C −1 (µ1 + µ0) = γ
4
s = µ1 − µ0.
and
T −1
γ −1 s C µ0
√ = Q (PFA) + √
T −1
s C s sT C −1 s
EE 527, Detection and Estimation Theory, # 5c 21
implying
√
PD = Q Q−1(PFA) − sT C −1 s
= Q Q−1(PFA) − d (15)
Here,
√ q
d= sT C −1 s = (µ1 − µ0)T C −1 (µ1 − µ0)
Assumptions:
fX | Θ(x[n] | θ) (16)
H0 : θ = θ0 versus
H1 : θ = θ1.
PFA,n, PD,n, n = 0, 1, . . . , N − 1.
dn 1−dn
pD(n) | Θ(dn | θ0) = PFA ,n (1 − PFA,n )
| {z }
Bernoulli pmf
N
X (dn | θ1) i
hp
D(n) | Θ
ln Λ(d) = ln
n=1
pD(n) | Θ(dn | θ0)
Nh P dn (1 − P )1−dn i H1
D,n D,n
X
= ln dn ≷ ln τ.
PFA,n (1 − PFA,n)1−dn
n=1
N
X −1
u1 = d[n].
n=0
P 1 − P H1
D D
log Λ(d) = u1 log + (N − u1) log ≷ log τ
PFA 1 − PFA
or
h P · (1 − P ) i H1 1 − P
D FA FA
u1 log ≷ log τ + N log . (17)
PFA · (1 − PD) 1 − PD
PD · (1 − PFA)
>1
PFA · (1 − PD)
H1
u1 ≷ τ 0 .
PFA(φ(X), θ))
In this context,
max PFA(φ(X), θ) (19)
θ∈spΘ (0)
is typically referred to as the size of the test φ(X). Therefore,
the condition (18) states that we focus on tests whose size is
upper-bounded by α.
PD(φUMP(X), θ) ≥ PD(φ(X), θ)
{X[n] | Θ = θ} = θ + W [n] n = 0, 1, . . . , N − 1
fX | Θ(x | 0) = N (x | 0, σ 2/N )
Note that
τ
max PFA(φ(X), θ) = PFA(φ(X), 0) = Q p =α
θ∈spΘ (0) 2
σ /N
see (18) and (19). The most powerful test is achieved if the
upper bound α in (18) is reached by equality:
r
σ2
τ= · Q−1(α). (22)
N
n X −θ τ − θ o
PrX | Θ{X > τ | θ} = PrX | Θ p >p θ
2
σ /N 2
σ /N
| {z }
standard
normal
random var.
τ −θ θ
= Q p = Q Q−1(α) − p . (23)
2
σ /N 2
σ /N
{X[n] | Θ = θ} = θ + W [n] n = 0, 1, . . . , N − 1
N −1
1 1 X h
2
i
fX | Θ(x | θ) = p ·exp − 2 (x[n]−θ) (24)
2
(2 π σ )N 2 σ n=0
and
θ0 < θ1.
Taking log etc. leads to
H1
(θ1 − θ0) x ≷ η
σ
τ = √ Q−1(α).
N
θ1 x > η.
fX | Θ(x | θ)
• there is a scalar statistic T (x) such that, for θ0 < θ1, the
likelihood ratio
fX | Θ(x | θ1)
Λ(x ; θ0, θ1) =
fX | Θ(x | θ0)
H0 : θ > θ0 versus
H1 : θ ≤ θ0.
N −1
1 Y
fX | Θ(x | θ) = N exp[−θ−1 T (x)] i(0,+∞)(x[n])
θ n=0
where
N
X −1
T (x) = x[n].
n=0
Since fX | Θ(x | θ) belongs to the one-parameter exponential
family (29) and η(θ) = −θ−1 is a monotonically increasing
function of θ. Therefore, the test
1, for T (x) ≥ λ,
φλ(x) =
0, for T (x) < λ
H0 : θ > θ0 versus
H1 : θ ≤ θ0.
1 T N −1
fT | Θ(T | θ) = N
exp(−T /θ) i(0,+∞)(T )
θ (N − 1)!
= Gamma(T | N, θ−1).
maxθ∈spΘ(1) fX | Θ(x | θ)
ΛGLR(x) = > τ.
maxθ∈spΘ(0) fX | Θ(x | θ)
fX | Θ(x | θ) = N (x | θ, σ 2/N ).
maxθ∈spΘ(1) fX | Θ(x | θ)
ΛGLR(x) = > τ.
fX | Θ(x | 0)
Now,
x = arg max fX | Θ(x | θ)
θ∈spΘ (1)
fX | Θ(x | 0) = N (x | 0, σ 2/N )
2
1 x
= p exp − 12 2
2 π σ 2/N σ /N
1
fX | Θ(x | x) = N (x | 0, σ 2/N ) = p
2 π σ 2/N
yielding
N x2
ln ΛGLR(x) = 2
.
2σ
Therefore, we accept H1 if
(x)2 > γ
or
|x| > η.
We compare this detector with the (not realizable, also called
clairvoyant) UMP detector that assumes the knowledge of the
sign of θ under H1. Assuming that the sign of θ under H1 is
known, we can construct the UMP detector, whose ROC curve
is given by
PD = Q(Q−1(PFA) − d)
p
where d = N θ2/σ 2 and θ is the value of the parameter
under H1; see (23) for the case where θ > 0 under H1. All
other detectors have PD below this upper bound.
In this case, GLR test is only slightly worse than the clairvoyant
detector (Figure 6.4 in Kay-II):
accepts H1 if
maxθ,σ2 fX | Θ,Σ2 (x | θ, σ 2)
ΛGLR(x) = >γ
maxσ2 fX | Θ,Σ2 (x | 0, σ 2)
N −1
1 1 X h i
fX | Θ,Σ2 (x | θ, σ 2) = p ·exp − 2 (x[n]−θ)2 .
(2 π σ 2)N 2 σ n=0
(30)
Here,
2 1 −N/2
max fX | Θ,Σ2 (x | θ, σ ) = · e
θ,σ 2 b12(x)]N/2
[2 π σ
1 −N/2
max fX | Θ,Σ2 (x | 0, σ 2) = · e
σ2 b02(x)]N/2
[2 π σ
where
N
1 X 2
b02(x) =
σ x [n]
N n=1
N
1 X
b12(x) =
σ (x[n] − x)2.
N n=1
Hence,
N/2
b02(x)
σ
ΛGLR(x) =
b12(x)
σ
i.e. GLR test fits data with the “best” DC-level signal θbML = x,
finds the residual variance estimate σ b12, and compares this
estimate with the variance estimate σ b02 under the null case (i.e.
b02(x) − x2.
= σ
Hence,
b02(x)
σ 1
2 ln ΛGLR(x) = N ln 2 2 = N ln 2 2
.
b0 (x) − x
σ 1 − x /bσ0 (x)
Note that
x2
0≤ 2 ≤1
σb0 (x)
and ln[1/(1 − z)] is monotonically increasing on z ∈ (0, 1).
Therefore, an equivalent test can be constructed as follows:
x2
T (x) = 2 > τ.
σ
b0 (x)