Self-Adapting Recurrent Models for Object Pushing from Learning in Simulation

IROS(2020)

引用 11|浏览24
暂无评分
摘要
Planar pushing remains a challenging research topic, where building the dynamic model of the interaction is the core issue. Even an accurate analytical dynamic model is inherently unstable because physics parameters such as inertia and friction can only be approximated. Data-driven models usually rely on large amounts of training data, but data collection is time consuming when working with real robots. In this paper, we collect all training data in a physics simulator and build an LSTM-based model to fit the pushing dynamics. Domain Randomization is applied to capture the pushing trajectories of a generalized class of objects. When executed on the real robot, the trained recursive model adapts to the tracked object's real dynamics within a few steps. We propose the algorithm \emph{Recurrent} Model Predictive Path Integral (RMPPI) as a variation of the original MPPI approach, employing state-dependent recurrent models. As a comparison, we also train a Deep Deterministic Policy Gradient (DDPG) network as a model-free baseline, which is also used as the action generator in the data collection phase. During policy training, Hindsight Experience Replay is used to improve exploration efficiency. Pushing experiments on our UR5 platform demonstrate the model's adaptability and the effectiveness of the proposed framework.
更多
查看译文
关键词
accurate analytical dynamic model,physics parameters,inertia,friction,data-driven models,training data,physics simulator,LSTM-based model,pushing dynamics,Domain Randomization,pushing trajectories,trained recursive model adapts,tracked object,algorithm Recurrent Model Predictive Path Integral,state-dependent recurrent models,Deep Deterministic Policy Gradient network,model-free baseline,data collection phase,policy training,pushing experiments,object pushing,planar pushing,core issue
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要