Regularizing Reinforcement Learning with State Abstraction

2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)(2018)

引用 20|浏览40
暂无评分
摘要
State abstraction in a discrete reinforcement learning setting clusters states sharing a similar optimal action to yield an easier to solve decision process. In this paper, we generalize the concept of state abstraction to continuous action reinforcement learning by defining an abstract state as a state cluster over which a near-optimal policy of simple shape exists. We propose a hierarchical reinforcement learning algorithm that is able to simultaneously find the state space clustering and the optimal sub-policies in each cluster. The main advantage of the proposed framework is to provide a straightforward way of regularizing reinforcement learning by controlling the behavioral complexity of the learned policy. We apply our algorithm on several benchmark tasks and a robot tactile manipulation task and show that we can match state-of-the-art deep reinforcement learning performance by combining a small number of linear policies.
更多
查看译文
关键词
deep reinforcement learning performance,optimal sub-policies,state space clustering,hierarchical reinforcement learning algorithm,near-optimal policy,state cluster,abstract state,continuous action reinforcement learning,similar optimal action,discrete reinforcement,state abstraction,learned policy
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要