TY - GEN
T1 - A-DDPG
T2 - 29th IEEE/ACM International Symposium on Quality of Service, IWQOS 2021
AU - He, Nan
AU - Yang, Song
AU - Li, Fan
AU - Trajanovski, Stojan
AU - Kuipers, Fernando A.
AU - Fu, Xiaoming
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/6/25
Y1 - 2021/6/25
N2 - The efficacy of Network Function Virtualization (NFV) depends critically on (1) where the virtual network functions (VNFs) are placed and (2) how the traffic is routed. Unfortunately, these aspects are not easily optimized, especially under time-varying network states with different quality of service (QoS) requirements. Given the importance of NFV, many approaches have been proposed to solve the VNF placement and traffic routing problem. However, those prior approaches mainly assume that the state of the network is static and known, disregarding real-time network variations. To bridge that gap, in this paper, we formulate the VNF placement and traffic routing problem as a Markov Decision Process model to capture the dynamic network state transitions. In order to jointly minimize the delay and cost of NFV providers and maximize the revenue, we devise a customized Deep Reinforcement Learning (DRL) algorithm, called A-DDPG, for VNF placement and traffic routing in a real-time network. A-DDPG uses the attention mechanism to ascertain smooth network behavior within the general framework of network utility maximization (NUM). The simulation results show that A-DDPG outperforms the state-of-the-art in terms of network utility, delay, and cost.
AB - The efficacy of Network Function Virtualization (NFV) depends critically on (1) where the virtual network functions (VNFs) are placed and (2) how the traffic is routed. Unfortunately, these aspects are not easily optimized, especially under time-varying network states with different quality of service (QoS) requirements. Given the importance of NFV, many approaches have been proposed to solve the VNF placement and traffic routing problem. However, those prior approaches mainly assume that the state of the network is static and known, disregarding real-time network variations. To bridge that gap, in this paper, we formulate the VNF placement and traffic routing problem as a Markov Decision Process model to capture the dynamic network state transitions. In order to jointly minimize the delay and cost of NFV providers and maximize the revenue, we devise a customized Deep Reinforcement Learning (DRL) algorithm, called A-DDPG, for VNF placement and traffic routing in a real-time network. A-DDPG uses the attention mechanism to ascertain smooth network behavior within the general framework of network utility maximization (NUM). The simulation results show that A-DDPG outperforms the state-of-the-art in terms of network utility, delay, and cost.
KW - Network function virtualization
KW - deep reinforcement learning
KW - placement
KW - routing
UR - http://www.scopus.com/inward/record.url?scp=85115406035&partnerID=8YFLogxK
U2 - 10.1109/IWQOS52092.2021.9521285
DO - 10.1109/IWQOS52092.2021.9521285
M3 - Conference contribution
AN - SCOPUS:85115406035
T3 - 2021 IEEE/ACM 29th International Symposium on Quality of Service, IWQOS 2021
BT - 2021 IEEE/ACM 29th International Symposium on Quality of Service, IWQOS 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 25 June 2021 through 28 June 2021
ER -