Learning from outside the viability kernel: Why we should build robots that can fall with grace

2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR)(2018)

引用 1|浏览17
暂无评分
摘要
Despite impressive results using reinforcement learning to solve complex problems from scratch, in robotics this has still been largely limited to model-based learning with very informative reward functions. One of the major challenges is that the reward landscape often has large patches with no gradient, making it difficult to sample gradients effectively. We show here that the robot state-initialization can have a more important effect on the reward landscape than is generally expected. In particular, we show the counter-intuitive benefit of including initializations that are unviable, in other words initializing in states that are doomed to fail.
更多
查看译文
关键词
robotics,informative reward functions,reward landscape,patches,sample gradients,robot state-initialization,important effect,counter-intuitive benefit,including initializations,viability kernel,impressive results,reinforcement learning,complex problems
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要