谷歌浏览器插件
订阅小程序
在清言上使用

On Joint Offloading and Resource Allocation: A Double Deep Q-Network Approach.

IEEE transactions on cognitive communications and networking/IEEE Transactions on Cognitive Communications and Networking(2021)

引用 15|浏览6
暂无评分
摘要
Multi-access edge computing (MEC) is an important enabling technology for 5G and 6G networks. With MEC, mobile devices can offload their computationally heavy tasks to a nearby server which can be a simple node at a base station, a vehicle or another device. With the increasing number of devices, slices and multiple radio access technologies, the problem of task offloading is becoming an increasingly complex problem. Thus, traditional approaches experience limitations while machine learning algorithms emerge as promising methods. In this paper, we consider binary and partial offloading problems and aim to jointly find optimal decisions for offloading and resource allocation which maximize the number of computed bits while minimizing the energy consumption. This allows improved usage of uplink transmit power and local CPU resources. We propose the Deep Reinforcement Learning for Joint Resource Allocation and Offloading (DJROM) algorithm that uses the double deep Q-network approach and models UEs as agents. We compare the proposed approach with two other machine learning based techniques, namely, multi-agent deep Q-learning (MARL-DQL) and multi-agent deep Q network (MARL-DQN) under fixed and mobile scenarios. Our results show that, DJROM scheme enhances the efficiency better than the other compared algorithms.
更多
查看译文
关键词
Multi-access edge computing,machine learning,reinforcement learning,resource management,task offloading,5G,6G
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要