Cooperative Control of Mobile Robots with Stackelberg Learning

IROS(2020)

引用 6|浏览37
暂无评分
摘要
Multi-robot cooperation requires agents to make decisions that are consistent with the shared goal without disregarding action-specific preferences that might arise from asymmetry in capabilities and individual objectives. To accomplish this goal, we propose a method named SLiCC: Stackelberg Learning in Cooperative Control. SLiCC models the problem as a partially observable stochastic game composed of Stackelberg bimatrix games, and uses deep reinforcement learning to obtain the payoff matrices associated with these games. Appropriate cooperative actions are then selected with the derived Stackelberg equilibria. Using a bi-robot cooperative object transportation problem, we validate the performance of SLiCC against centralized multi-agent Q-learning and demonstrate that SLiCC achieves better combined utility.
更多
查看译文
关键词
mobile robots,Stackelberg learning,multirobot cooperation,shared goal,action-specific preferences,SLiCC models,partially observable stochastic game,Stackelberg bimatrix games,deep reinforcement learning,appropriate cooperative actions,Stackelberg equilibria,bi-robot cooperative object transportation problem,multiagent Q-learning,cooperative control,payoff matrices
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要