TY - JOUR
T1 - Proactive Content Caching Based on Actor-Critic Reinforcement Learning for Mobile Edge Networks
AU - Jiang, Wei
AU - Feng, Daquan
AU - Sun, Yao
AU - Feng, Gang
AU - Wang, Zhenzhong
AU - Xia, Xiang Gen
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2022/6/1
Y1 - 2022/6/1
N2 - Mobile edge caching/computing (MEC) has emerged as a promising approach for addressing the drastic increasing mobile data traffic by bringing high caching and computing capabilities to the edge of networks. Under MEC architecture, content providers (CPs) are allowed to lease some virtual machines (VMs) at MEC servers to proactively cache popular contents for improving users' quality of experience. The scalable cache resource model rises the challenge for determining the ideal number of leased VMs for CPs to obtain the minimum expected downloading delay of users at the lowest caching cost. To address these challenges, in this paper, we propose an actor-critic (AC) reinforcement learning based proactive caching policy for mobile edge networks without the prior knowledge of users' content demand. Specifically, we formulate the proactive caching problem under dynamical users' content demand as a Markov decision process and propose a AC based caching algorithm to minimize the caching cost and the expected downloading delay. Particularly, to reduce the computational complexity, a branching neural network is employed to approximate the policy function in the actor part. Numerical results show that the proposed caching algorithm can significantly reduce the total cost and the average downloading delay when compared with other popular algorithms.
AB - Mobile edge caching/computing (MEC) has emerged as a promising approach for addressing the drastic increasing mobile data traffic by bringing high caching and computing capabilities to the edge of networks. Under MEC architecture, content providers (CPs) are allowed to lease some virtual machines (VMs) at MEC servers to proactively cache popular contents for improving users' quality of experience. The scalable cache resource model rises the challenge for determining the ideal number of leased VMs for CPs to obtain the minimum expected downloading delay of users at the lowest caching cost. To address these challenges, in this paper, we propose an actor-critic (AC) reinforcement learning based proactive caching policy for mobile edge networks without the prior knowledge of users' content demand. Specifically, we formulate the proactive caching problem under dynamical users' content demand as a Markov decision process and propose a AC based caching algorithm to minimize the caching cost and the expected downloading delay. Particularly, to reduce the computational complexity, a branching neural network is employed to approximate the policy function in the actor part. Numerical results show that the proposed caching algorithm can significantly reduce the total cost and the average downloading delay when compared with other popular algorithms.
KW - Actor-critic algorithm
KW - Branching neural network
KW - Mobile edge caching
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85120538485&partnerID=8YFLogxK
U2 - 10.1109/TCCN.2021.3130995
DO - 10.1109/TCCN.2021.3130995
M3 - Article
AN - SCOPUS:85120538485
SN - 2332-7731
VL - 8
SP - 1239
EP - 1252
JO - IEEE Transactions on Cognitive Communications and Networking
JF - IEEE Transactions on Cognitive Communications and Networking
IS - 2
ER -