This document discusses value functions and Markov decision processes (MDPs). It defines value functions as estimating the long-term expected rewards from each state. It presents the Bellman equation and how it can be used to compute value functions. Finally, it introduces MDPs, which extend Markov reward processes by adding actions, and provides some examples of MDP problems like navigation and Atari games.