Nash Q-learning multi-agent flow control for high-speed networks

ACC'09 Proceedings of the 2009 conference on American Control Conference(2009)

引用 1|浏览4
暂无评分
摘要
For the congestion problems in high-speed networks, a multi-agent flow controller (MFC) based on Q-learning algorithm conjunction with the theory of Nash equilibrium is proposed. Because of the uncertainties and highly time-varying, it is not easy to accurately obtain the complete information for high-speed networks, especially for the multi-bottleneck case. The Nash Q-learning algorithm, which is independent of mathematic model, shows the particular superiority in high-speed networks. It obtains the Nash Q-values through trial-and-error and interaction with the network environment to improve its behavior policy. By means of learning procedures, MFCs can learn to take the best actions to regulate source flow with the features of high throughput and low packet loss ratio. Simulation results show that the proposed method can promote the performance of the networks and avoid the occurrence of congestion effectively.
更多
查看译文
关键词
nash q-learning multi-agent flow,nash q-learning algorithm,high-speed network,nash equilibrium,congestion problem,source flow,algorithm conjunction,behavior policy,multi-agent flow controller,nash q-values,flow control,bandwidth,uncertainty,learning artificial intelligence,convergence,high throughput,quality of service,mathematical model,throughput,game theory,mathematics,control systems
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要