powered by:
MagicWare, s.r.o.

Empirical Results on Convergence and Exploration in Approximate Policy Iteration

Authors:Kaisare Niket, Georgia Institute of Technology, United States
Lee Jong Min, Georgia Institute of Technology, United States
Lee Jay H., Georgia Institute of Technology, United States
Topic:2.3 Non-Linear Control Systems
Session:Optimal Control in Nonlinear Systems I
Keywords: Optimal Control, Approximate Dynamic Programming, Cost-to-go Function, Policy Iteration, Value Iteration

Abstract

In this paper, we empirically investigate the convergence properties of policy iteration applied to the optimal control of systems with continuous state and action spaces. We demonstrate that policy iteration requires lesser iterations than value iteration to converge, but requires more function evaluations to generate cost-to-go approximations in the policy evaluation step. Two different alternatives to policy evaluation, based on iteration over simulated states and simulation of improved policies are presented. We then demonstrate that the lambda-policy iteration method (0 <= lambda <= 1) is a tradeoff between value and policy iteration. Finally, the issue of exploration to expand the coverage of the state space during offline iteration is also considered.