Definition
A Markov decision process is a 4-tuple, where
- is a finite set of states,
- is a finite set of actions (alternatively, is the finite set of actions available from state ),
- is the probability that action in state at time will lead to state at time ,
- is the immediate reward (or expected immediate reward) received after transition to state from state with transition probability .
(The theory of Markov decision processes does not actually require or to be finite, but the basic algorithms below assume that they are finite.)
Read more about this topic: Markov Decision Process
Famous quotes containing the word definition:
“... we all know the wags definition of a philanthropist: a man whose charity increases directly as the square of the distance.”
—George Eliot [Mary Ann (or Marian)
“Was man made stupid to see his own stupidity?
Is God by definition indifferent, beyond us all?
Is the eternal truth mans fighting soul
Wherein the Beast ravens in its own avidity?”
—Richard Eberhart (b. 1904)
“The man who knows governments most completely is he who troubles himself least about a definition which shall give their essence. Enjoying an intimate acquaintance with all their particularities in turn, he would naturally regard an abstract conception in which these were unified as a thing more misleading than enlightening.”
—William James (18421910)