In Bayesian probability theory, the principle of maximum entropy is a prime doctrine. It states that, subject to precisely stated prior data, which must be a proposition that expresses testable information, the probability distribution which best represents the current state of knowledge is the one with largest information-theoretical entropy.
Let some precisely stated prior data or testable information about a probability distribution function be given. Consider the set of all trial probability distributions that encode the prior data. Of those, the one that maximizes the information entropy is the proper probability distribution under the given prior data.
Read more about Principle Of Maximum Entropy: History, Overview, Testable Information, Justifications For The Principle of Maximum Entropy
Famous quotes containing the words principle of, principle, maximum and/or entropy:
“The world is not dialecticalit is sworn to extremes, not to equilibrium, sworn to radical antagonism, not to reconciliation or synthesis. This is also the principle of evil.”
—Jean Baudrillard (b. 1929)
“We rail at trade, but the historian of the world will see that it was the principle of liberty; that it settled America, and destroyed feudalism, and made peace and keeps peace; that it will abolish slavery.”
—Ralph Waldo Emerson (18031882)
“Only at his maximum does an individual surpass all his derivative elements, and become purely himself. And most people never get there. In his own pure individuality a man surpasses his father and mother, and is utterly unknown to them.”
—D.H. (David Herbert)
“Just as the constant increase of entropy is the basic law of the universe, so it is the basic law of life to be ever more highly structured and to struggle against entropy.”
—Václav Havel (b. 1936)