This document summarizes research on using fluid and diffusion approximations within approximate dynamic programming for applications in power management. Specifically, it discusses how:
1) The fluid value function provides a tight approximation to the relative value function and can be used as part of the basis for TD learning.
2) TD learning with policy improvement finds a near-optimal policy in a few iterations when applied to power management problems.
3) Fluid and diffusion models provide useful insights into the structure of optimal policies for average cost problems.
Related topics: