TY - JOUR
T1 - Joint Computation Offloading and Resource Allocation for D2D-Assisted Mobile Edge Computing
AU - Jiang, Wei
AU - Feng, Daquan
AU - Sun, Yao
AU - Feng, Gang
AU - Wang, Zhenzhong
AU - Xia, Xiang Gen
N1 - Publisher Copyright:
© 2008-2012 IEEE.
PY - 2023/5/1
Y1 - 2023/5/1
N2 - Computation offloading via device-to-device communications can improve the performance of mobile edge computing by exploiting the computing resources of user devices. However, most proposed optimization-based computation offloading schemes lack self-adaptive abilities in dynamic environments due to time-varying wireless environment, continuous-discrete mixed actions, and coordination among devices. The conventional reinforcement learning based approaches are not effective for solving an optimal sequential decision problem with continuous-discrete mixed actions. In this paper, we propose a hierarchical deep reinforcement learning (HDRL) framework to solve the joint computation offloading and resource allocation problem. The proposed HDRL framework has a hierarchical actor-critic architecture with a meta critic, multiple basic critics and actors. Specifically, a combination of deep Q-network (DQN) and deep deterministic policy gradient (DDPG) is exploited to cope with the continuous-discrete mixed action spaces. Furthermore, to handle the coordination among devices, the meta critic acts as a DQN to output the joint discrete action of all devices and each basic critic acts as the critic part of DDPG to evaluate the output of the corresponding actor. Simulation results show that the proposed HDRL algorithm can significantly reduce the task computation latency compared with baseline offloading schemes.
AB - Computation offloading via device-to-device communications can improve the performance of mobile edge computing by exploiting the computing resources of user devices. However, most proposed optimization-based computation offloading schemes lack self-adaptive abilities in dynamic environments due to time-varying wireless environment, continuous-discrete mixed actions, and coordination among devices. The conventional reinforcement learning based approaches are not effective for solving an optimal sequential decision problem with continuous-discrete mixed actions. In this paper, we propose a hierarchical deep reinforcement learning (HDRL) framework to solve the joint computation offloading and resource allocation problem. The proposed HDRL framework has a hierarchical actor-critic architecture with a meta critic, multiple basic critics and actors. Specifically, a combination of deep Q-network (DQN) and deep deterministic policy gradient (DDPG) is exploited to cope with the continuous-discrete mixed action spaces. Furthermore, to handle the coordination among devices, the meta critic acts as a DQN to output the joint discrete action of all devices and each basic critic acts as the critic part of DDPG to evaluate the output of the corresponding actor. Simulation results show that the proposed HDRL algorithm can significantly reduce the task computation latency compared with baseline offloading schemes.
KW - Computation offloading
KW - deep reinforcement learning
KW - device-to-device
KW - mobile edge computing
KW - resource allocation
UR - http://www.scopus.com/inward/record.url?scp=85134197337&partnerID=8YFLogxK
U2 - 10.1109/TSC.2022.3190276
DO - 10.1109/TSC.2022.3190276
M3 - Article
AN - SCOPUS:85134197337
SN - 1939-1374
VL - 16
SP - 1949
EP - 1963
JO - IEEE Transactions on Services Computing
JF - IEEE Transactions on Services Computing
IS - 3
ER -