Zoom out Search Issue

[
lecture
NOTES
]
IEEE SIGNAL PROCESSING MAGAZINE [164] MARCH 2015 1053-5888/15©2015IEEE
Dave Zachariah
and Petre Stoica
Cramér–Rao Bound Analog of Bayes’ Rule
T
he estimation of multiple pa-
rameters is a common task in
signal processing. The Cra-
mér–Rao bound (CRB) sets a
statistical lower limit on the re-
sulting errors when estimating parameters
from a set of random observations. It can be
understood as a fundamental measure of
parameter uncertainty [1], [2]. As a general
example, suppose
i denotes the vector of
sought parameters and that the random ob-
servation model can be written as
,yx w=+
i
(1)
where x
i
is a function or signal parame-
terized by
i and w is a zero-mean Gauss-
ian noise vector. Then the CRB for
i has
the following notable properties:
1) For a fixed
,
i the CRB for i decreas-
es as the dimension of y increases.
2) For a fixed ,y if additional parameters
i
u
are estimated, then the CRB for
i
increases as the dimension of
i
u
increases.
3) If adding a set of observations y
u
requires estimating additional param-
eters
,
i
u
then the CRB for
i decreas-
es as the dimension of y
u
increases,
provided the dimension of
i
u
does not
exceed that of y
u
[3]. This property
implies both 1) and 2) above.
4) Among all possible distributions of
w
with a fixed covariance matrix, the CRB
for
i attains its maximum when w is
Gaussian, i.e., the Gaussian scenario is
the “worst case” for estimating
i
[4]–[6].
In this lecture note, we show a general
property of the CRB that quantifies the
interdependencies between the parameters
in
.
i The presented result is valid for more
general models than (1) and also general-
izes the result in [7] to vector parameters.
It will be illustrated via two examples.
RELEVANCE
In probability theory, the chain rule and
Bayes’ rule are useful tools to analyze the
statistical interdependence between multi-
ple random variables and to derive tractable
expressions for their distributions. In this
lecture note, we provide analogs of the
chain rule and Bayes’ rule for the CRB
associated with multiple parameters. The
results are particularly useful when esti-
mating parameters of interest in the pres-
ence of nuisance parameters.
PREREQUISITIES
The reader needs basic knowledge about
linear algebra, elementary probability the-
ory, and statistical signal processing.
PRELIMINARIES
We will consider a general scenario in which
we observe an
n 1# random vector .y Its
probability density function (pdf) ;p y
i
^h
is
parameterized by a k 1# deterministic vec-
tor .
i The goal is to estimate ,i or subvec-
tors of ,
i given .y
Let ;lpln y_
ii
^
^
h
h
denote the log-
likelihood function, and let
i
t
be any unbi-
ased estimator. Then the mean square error
(MSE) matrix E[()()]P
*
_ iiii--
i
tt
t
is
bounded from below by the inverse of
the Fisher information matrix
EJ _ -
i
,l
2
2 i
i
^h
6
@
where
2
2
i
denotes the second-
order differential or Laplacian operator with
respect to
.
i That is, ,PJ
1
*
i
i
-
t
assuming
from hereon that J
i
is nonsingular. This is
the Cramér–Rao inequality [2], [8], [9].
The determinant of the MSE matrix,
||,P
i
t
is a scalar measure of the error
magnitude. For unbiased estimators, ||P
i
t
equals the “generalized variance” of errors
[10]. By defining ,CRB J
1
_i
i
-
^h
the
generalized error variance is bounded by
.CRBP $
i
i
t
^h
In the following, we are interested in
subvectors or elements of .
i Letting i =
[],
ab
<<<
we can write the Fisher informa-
tion matrix in block form,
,,
,,
.
ll
ll
EJ
JJ
JJ
2
2
22 2
222
ab ab
ab ab
=-
=
i
ba
b
a
ab
ba b
aab
^
^^
^h
hh
h
=
>
G
H
(2)
MAIN RESULT
Let
a and b be two random vectors. Two
useful rules in probability theory are the
chain rule
() (|)()ppp,ab a b b= (3)
and Bayes’ rule
()
(|)
()
().p
p
p
p |a
a
b
ab=
b
(4)
Now consider two parameter vectors
a and .b When both are unknown, their
joint CRB bound is given by
,.CRB
J
J
J
J
1
ab =
a
ba
ab
b
-
^h
=
G
(5)
The bound for
a with known b is simply
|,
CRB J
1
ab=
a
-
^h
(6)
and the bound for a with unknown b is
J .CRB JJ J
1
1
a =-
aab
b
ba
-
-
^^
hh
(7)
[Equation (7) follows by evaluating the
inverse in (5) and extracting the upper-
left block corresponding to .
a ] Equations
(6) and (7) are the respective CRB analogs
of conditional and marginal distributions
for random variables.
Digital Object Identifier 10.1109/MSP.2014.2365593
Date of publication: 12 February 2015
Previous Page | Contents | Zoom in | Zoom out | Front Cover | Search Issue | Next Page
q
q
M
M
q
q
M
M
q
M
THE WORLD’S NEWSSTAND
®
Previous Page | Contents | Zoom in | Zoom out | Front Cover | Search Issue | Next Page
q
q
M
M
q
q
M
M
q
M
THE WORLD’S NEWSSTAND
®