Decentralized coordination via task decomposition and reward shaping

AAMAS(2013)

引用 1|浏览42
暂无评分
摘要
In this work, we introduce a method for decentralized coordination in cooperative multiagent multi-task problems where the subtasks and agents are homogeneous. Using the method proposed, the agents cooperate at the high level task selection using the knowledge they gather by learning subtasks. We introduce a subtask selection method for single agent multi-task MDPs and we extend the work to multiagent multi-task MDPs by using reward shaping at the subtask level to coordinate the agents. Our results on a multi-rover problem show that agents which use the combination of task decomposition and subtask based difference rewards result in significant improvement both in terms of learning speed, and converged policies.
更多
查看译文
关键词
high level task selection,subtask level,single agent multi-task mdps,subtask selection method,multi-rover problem show,cooperative multiagent multi-task problem,decentralized coordination,significant improvement,multi-task mdps,task decomposition,cooperation,reinforcement learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要