Fisher Information - Distinction From The Hessian of The Entropy

Distinction From The Hessian of The Entropy

In certain cases, the Fisher Information matrix is the negative of the Hessian of the Shannon entropy. The cases where this explicitly holds is given below. A distribution's Shannon entropy

has as the negative of the entry of its Hessian:

-\frac{\partial^2}{\partial\theta_i \partial\theta_j} \mathcal{H}
= \int \left dX\,.

In contrast, the entry of the Fisher information matrix is

\mathcal{I}_{ij}(\theta)
= \int f(X; \theta) \frac{\partial \log f(X; \theta)}{\partial\theta_i} \frac{\partial \log f(X; \theta)}{\partial\theta_j} \,dX
= \int \frac{1}{f(X; \theta)} \frac{\partial f(X; \theta)}{\partial\theta_i} \frac{\partial f(X; \theta)}{\partial\theta_j} \,dX\,.

The difference between the negative Hessian and the Fisher information is

-\frac{\partial^2}{\partial\theta_i \, \partial\theta_j} \mathcal{H} - \mathcal{I}_{ij}(\theta)
= \int \frac{\partial^2 f(X; \theta)}{\partial\theta_i \, \partial\theta_j} \left(1 + \log f(X; \theta) \right) dX\,.

This extra term goes away if, instead, one considers the Hessian of the relative entropy instead of the Shannon entropy.

Read more about this topic:  Fisher Information

Famous quotes containing the words distinction from, distinction and/or entropy:

    Let a man find himself, in distinction from others, on top of two wheels with a chain—at least in a poor country like Russia—and his vanity begins to swell out like his tyres. In America it takes an automobile to produce this effect.
    Leon Trotsky (1879–1940)

    Quadruped lions are said to be savage, only when they are hungry; biped lions are rarely sulky longer than when their appetite for distinction remains unappeased.
    Charles Dickens (1812–1870)

    Just as the constant increase of entropy is the basic law of the universe, so it is the basic law of life to be ever more highly structured and to struggle against entropy.
    Václav Havel (b. 1936)