Performance bounds for λ policy iteration and application to the game of Tetris

Journal of Machine Learning Research(2013)

引用 32|浏览41
暂无评分
摘要
We consider the discrete-time infinite-horizon optimal control problem formalized by Markov decision processes (Puterman, 1994; Bertsekas and Tsitsiklis, 1996). We revisit the work of Bertsekas and Ioffe (1996), that introduced λ policy iteration--a family of algorithms parametrized by a parameter λ--that generalizes the standard algorithms value and policy iteration, and has some deep connections with the temporal-difference algorithms described by Sutton and Barto (1998). We deepen the original theory developed by the authors by providing convergence rate bounds which generalize standard bounds for value iteration described for instance by Puterman (1994). Then, the main contribution of this paper is to develop the theory of this algorithm when it is used in an approximate form. We extend and unify the separate analyzes developed by Munos for approximate value iteration (Munos, 2007) and approximate policy iteration (Munos, 2003), and provide performance bounds in the discounted and the undiscounted situations. Finally, we revisit the use of this algorithm in the training of a Tetris playing controller as originally done by Bertsekas and Ioffe (1996). Our empirical results are different from those of Bertsekas and Ioffe (which were originally qualified as \"paradoxical\" and \"intriguing\"). We track down the reason to be a minor implementation error of the algorithm, which suggests that, in practice, l policy iteration may be more stable than previously thought.
更多
查看译文
关键词
l policy iteration,policy iteration,approximate value iteration,approximate form,standard algorithms value,approximate policy iteration,standard bound,value iteration,performance bound,original theory,markov decision process,analysis of algorithms,artificial intelligent,reinforcement learning,temporal difference,markov decision processes,convergence rate,discrete time,stochastic optimal control,optimal control
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要