In Bayesian probability theory, the principle of maximum entropy is a prime doctrine. It states that, subject to precisely stated prior data, which must be a proposition that expresses testable information, the probability distribution which best represents the current state of knowledge is the one with largest information-theoretical entropy.
Let some precisely stated prior data or testable information about a probability distribution function be given. Consider the set of all trial probability distributions that encode the prior data. Of those, the one that maximizes the information entropy is the proper probability distribution under the given prior data.
Read more about Principle Of Maximum Entropy: History, Overview, Testable Information, Justifications For The Principle of Maximum Entropy
Famous quotes containing the words principle of, principle, maximum and/or entropy:
“The principle of fashion is ... the principle of the kaleidoscope. A new year can only bring us a new combination of the same elements; and about once in so often we go back and begin again.”
—Katharine Fullerton Gerould (18791944)
“For me chemistry represented an indefinite cloud of future potentialities which enveloped my life to come in black volutes torn by fiery flashes, like those which had hidden Mount Sinai. Like Moses, from that cloud I expected my law, the principle of order in me, around me, and in the world.... I would watch the buds swell in spring, the mica glint in the granite, my own hands, and I would say to myself: I will understand this, too, I will understand everything.”
—Primo Levi (19191987)
“Probably the only place where a man can feel really secure is in a maximum security prison, except for the imminent threat of release.”
—Germaine Greer (b. 1939)
“Just as the constant increase of entropy is the basic law of the universe, so it is the basic law of life to be ever more highly structured and to struggle against entropy.”
—Václav Havel (b. 1936)