Edge Caching For Iot Transient Data Using Deep Reinforcement Learning

IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY(2020)

引用 3|浏览16
暂无评分
摘要
Connected devices generate large amount of data for IoT applicatons. Assisted by edge computing, caching IoT data at the edge nodes is considered as a promising technique for its advantage in reducing network traffic and service delay of cloud platform. However, the IoT data is characterized by transient lifetime and cache capacity that is limited by the edge nodes. As a consequence, caching policy should consider both data transiency and storage capacity of edge nodes. Inspired by the success of deep reinforcement learning (DRL) in deal with Markov Decision Process (MDP) problem in unknown environment, A DRL-based algorithm for edge caching problem is proposed in this paper. The proposed Advantage Actor Critic (A2C)-based algorithm is aimed at maximizing the long-term energy saving without knowledge of the IoT data popularity profiles. Simulation results demonstrate that the proposed DRL-based algorithm can achieve higher energy saving and cache hit ratio compared with the baseline algorithms.
更多
查看译文
关键词
Edge caching, IoT transient data, deep reinforcement learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要