Nonlinear Filters. Simon Haykin

Nonlinear Filters - Simon  Haykin


Скачать книгу
k plus 1 Baseline vertical-bar bold u Subscript 0 colon k Baseline comma bold y Subscript 0 colon k Baseline right-parenthesis Over p left-parenthesis bold y Subscript k plus 1 Baseline vertical-bar bold u Subscript 0 colon k plus 1 Baseline comma bold y Subscript 0 colon k Baseline right-parenthesis EndFraction comma EndLayout"/>

      where the normalization constant in the denominator is obtained as:

      (4.9)p left-parenthesis bold y Subscript k plus 1 Baseline vertical-bar bold u Subscript 0 colon k plus 1 Baseline comma bold y Subscript 0 colon k Baseline right-parenthesis equals integral p left-parenthesis bold y Subscript k plus 1 Baseline vertical-bar bold x Subscript k plus 1 Baseline comma bold u Subscript k plus 1 Baseline right-parenthesis p left-parenthesis bold x Subscript k plus 1 Baseline vertical-bar bold u Subscript 0 colon k Baseline comma bold y Subscript 0 colon k Baseline right-parenthesis normal d bold x Subscript k plus 1 Baseline period

       Minimum mean‐square error (MMSE) estimator(4.10) This is equivalent to minimizing the trace (sum of the diagonal elements) of the estimation‐error covariance matrix. The MMSE estimate is the conditional mean of :(4.11) where the expectation is taken with respect to the posterior, .

       Risk‐sensitive (RS) estimator(4.12) Compared to the MMSE estimator, the RS estimator is less sensitive to uncertainties. In other words, it is a more robust estimator [49].

       Maximum a posteriori (MAP) estimator(4.13)

       Minimax estimator(4.14) The minimax estimate is the medium of the posterior, . The minimax technique is used to achieve optimal performance under the worst‐case condition [50].

       The most probable (MP) estimator(4.15) MP estimate is the mode of the posterior, . For a uniform prior, this estimate will be identical to the maximum likelihood (ML) estimate.(4.16)

      The relevant portion of the data obtained by measurement can be interpreted as information. In this line of thinking, a summary of the amount of information with regard to the variables of interest is provided by the Fisher information matrix [51]. To be more specific, Fisher information plays two basic roles:

      1 It is a measure of the ability to estimate a quantity of interest.

      2 It is a measure of the state of disorder in a system or phenomenon of interest.

      The first role implies that the Fisher information matrix has a close connection to the estimation‐error covariance matrix and can be used to calculate the confidence region of estimates. The second role implies that the Fisher information has a close connection to Shannon's entropy.

      Let us consider the PDF p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis, which is parameterized by the set of parameters bold-italic theta. The Fisher information matrix is defined as:

      (4.17)bold upper F Subscript bold-italic theta Baseline equals double-struck upper E Subscript bold x Baseline left-bracket left-parenthesis nabla Subscript bold-italic theta Baseline log p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis right-parenthesis left-parenthesis nabla Subscript bold-italic theta Baseline log p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis right-parenthesis Superscript upper T Baseline right-bracket period

      This definition is based on the outer product of the gradient of log p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis with itself, where the gradient is a column vector denoted by nabla Subscript bold-italic theta. There is an equivalent definition based on the second derivative of log p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis as:

      (4.18)bold upper F Subscript bold-italic theta Baseline equals minus double-struck upper E Subscript bold x Baseline left-bracket StartFraction partial-differential squared log p Subscript bold-italic theta Baseline left-parenthesis bold x right-parenthesis Over partial-differential squared bold-italic theta EndFraction right-bracket period

      From the definition of bold upper F Subscript bold-italic theta, it is obvious that Fisher information is a function of the corresponding PDF. A relatively broad and flat PDF, which is associated with lack of predictability and high entropy, has small gradient contents and, in effect therefore, low Fisher information. On the other hand, if the PDF is relatively narrow and has sharp slopes around a specific value of bold x, which is associated with bias toward that particular value of bold x and low entropy, it has large gradient contents and therefore high Fisher information. In summary, there is a duality between Shannon's entropy and Fisher information. However, a closer look at their mathematical definitions reveals an important difference [27]:

       A rearrangement of the tuples may change the shape of the PDF curve significantly, but it does not affect the value of the summation in (2.95) or integration in (2.96), because the summation and integration can be calculated in any order. Since is not affected by local changes in the PDF curve, it can be considered as a global measure of the behavior of the corresponding PDF.

       On the other hand, such a rearrangement of points changes the slope, and therefore gradient of the PDF curve, which, in turn, changes the Fisher information significantly. Hence, the Fisher information is sensitive to local rearrangement of points and can be considered as a local measure of the behavior of the corresponding PDF.

      Both entropy (as a global measure of smoothness in the PDF) and Fisher information (as a local measure of smoothness in the PDF) can be used in a variational principle to infer about the PDF that describes the phenomenon under consideration. However, the local measure may be preferred in general [27]. This leads


Скачать книгу