Groups
A Markov Decision Process (MDP) models decision-making in situations where outcomes are partly random and partly under the control of a decision maker.
Bellman equations express how the value of a state or action equals immediate reward plus discounted value of what follows.