Partially Observable Markov Decision Process - Belief Update

Belief Update

An agent needs to update its belief upon taking the action and observing . Since the state is Markovian, maintaining a belief over the states solely requires knowledge of the previous belief state, the action taken, and the current observation. The operation is denoted . Below we describe how this belief update is computed.

In, the agent observes with probability . Let be a probability distribution over the state space : denotes the probability that the environment is in state . Given, then after taking action and observing ,


b'(s') = \eta \Omega(o\mid s',a) \sum_{s\in S} T(s'\mid s,a)b(s)

where is a normalizing constant with .

Read more about this topic:  Partially Observable Markov Decision Process

Famous quotes containing the word belief:

    Some fear that if parents start listening to their own wants and needs they will neglect their children. It is our belief that children are in fact far less likely to be neglected when their parents’ needs—for support, for friendship, for decent work, for health care, for learning, for play, for time alone—are being met.
    —Wendy Coppedge Sanford. Ourselves and Our Children, by Boston Women’s Health Book Collective, introduction (1978)