Learning Sequential Decision Tasks For Robot Manipulation With Abstract Markov Decision Processes And Demonstration-Guided Exploration

2018 IEEE-RAS 18TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS)(2018)

引用 1|浏览19
暂无评分
摘要
Solving high-level sequential decision tasks situated on physical robots is a challenging problem. Reinforcement learning, the standard paradigm for solving sequential decision problems, allows robots to learn directly from experience, but is ill-equipped to deal with issues of scalability and uncertainty introduced by real-world tasks. We reformulate the problem representation to better apply to robot manipulation using the relations of Object-Oriented MDPs (OO-MDPs) and the hierarchical structure provided by Abstract MDPs (AMDPs). We present a relation-based AMDP formulation for solving tabletop organizational packing tasks, as well as a demonstrationguided exploration algorithm for learning AMDP transition functions inspired by state-and action-centric learning from demonstration approaches. We evaluate our representation and learning methods in a simulated environment, showing that our hierarchical representation is suitable for solving complex tasks, and that our state-and action-centric exploration biasing methods are both effective and complementary for efficiently learning AMDP transition functions. We show that the learned policy can be transferred to different tabletop organizational packing tasks, and validate that the policy can be realized on a physical system.
更多
查看译文
关键词
robot manipulation,reinforcement learning,sequential decision problems,Object-Oriented MDPs,demonstration-guided exploration algorithm,action-centric learning,action-centric exploration biasing methods,AMDP transition functions,sequential decision tasks learning,abstract Markov decision processes,relation-based AMDP
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要