Chernoff Bound - Matrix Chernoff Bound

Matrix Chernoff Bound

Rudolf Ahlswede and Andreas Winter introduced (Ahlswede & Winter 2003) a Chernoff bound for matrix-valued random variables.

If is distributed according to some distribution over matrices with zero mean, and if are independent copies of then for any ,


\Pr \left( \bigg\Vert \frac{1}{t} \sum_{i=1}^t M_i - \mathbf{E} \bigg\Vert_2 > \varepsilon \right) \leq d \exp \left( -C \frac{\varepsilon^2 t}{\gamma^2} \right).

where holds almost surely and is an absolute constant.

Notice that the number of samples in the inequality depends logarithmically on . In general, unfortunately, such a dependency is inevitable: take for example a diagonal random sign matrix of dimension . The operator norm of the sum of independent samples is precisely the maximum deviation among independent random walks of length . In order to achieve a fixed bound on the maximum deviation with constant probability, it is easy to see that should grow logarithmically with in this scenario.

The following theorem can be obtained by assuming has low rank, in order to avoid the dependency on the dimensions.

Read more about this topic:  Chernoff Bound

Famous quotes containing the words matrix and/or bound:

    As all historians know, the past is a great darkness, and filled with echoes. Voices may reach us from it; but what they say to us is imbued with the obscurity of the matrix out of which they come; and try as we may, we cannot always decipher them precisely in the clearer light of our day.
    Margaret Atwood (b. 1939)

    Or seen the furrows shine but late upturned,
    And where the fieldfare followed in the rear,
    When all the fields around lay bound and hoar
    Beneath a thick integument of snow.
    So by God’s cheap economy made rich
    To go upon my winter’s task again.
    Henry David Thoreau (1817–1862)