Fisher Information - Distinction From The Hessian of The Entropy

Distinction From The Hessian of The Entropy

In certain cases, the Fisher Information matrix is the negative of the Hessian of the Shannon entropy. The cases where this explicitly holds is given below. A distribution's Shannon entropy

has as the negative of the entry of its Hessian:

-\frac{\partial^2}{\partial\theta_i \partial\theta_j} \mathcal{H}
= \int \left dX\,.

In contrast, the entry of the Fisher information matrix is

\mathcal{I}_{ij}(\theta)
= \int f(X; \theta) \frac{\partial \log f(X; \theta)}{\partial\theta_i} \frac{\partial \log f(X; \theta)}{\partial\theta_j} \,dX
= \int \frac{1}{f(X; \theta)} \frac{\partial f(X; \theta)}{\partial\theta_i} \frac{\partial f(X; \theta)}{\partial\theta_j} \,dX\,.

The difference between the negative Hessian and the Fisher information is

-\frac{\partial^2}{\partial\theta_i \, \partial\theta_j} \mathcal{H} - \mathcal{I}_{ij}(\theta)
= \int \frac{\partial^2 f(X; \theta)}{\partial\theta_i \, \partial\theta_j} \left(1 + \log f(X; \theta) \right) dX\,.

This extra term goes away if, instead, one considers the Hessian of the relative entropy instead of the Shannon entropy.

Read more about this topic:  Fisher Information

Famous quotes containing the words distinction and/or entropy:

    Nature has not placed us in an inferior rank to men, no more than the females of other animals, where we see no distinction of capacity, though I am persuaded if there was a commonwealth of rational horses ... it would be an established maxim amongst them that a mare could not be taught to pace.
    Mary Wortley, Lady Montagu (1689–1762)

    Just as the constant increase of entropy is the basic law of the universe, so it is the basic law of life to be ever more highly structured and to struggle against entropy.
    Václav Havel (b. 1936)