TY - GEN
T1 - Hierarchical caching via deep reinforcement learning
AU - Sadeghi, Alireza
AU - Wang, Gang
AU - Giannakis, Georgios B.
N1 - Publisher Copyright:
© 2020 IEEE
PY - 2020/5
Y1 - 2020/5
N2 - Wireless and wireline networks, such as Internet, cellular, and content delivery networks are to serve end-user file requests proactively. To this aim, by storing anticipated highly popular files during off-peak periods, and fetching them to end-users during on-peak instances, these networks smoothen out the load fluctuations on the back-haul links. In this context, several practical networks comprise a parent caching node connected to multiple leaf nodes to serve end-user file requests. To model the two-way interactive influence between caching decisions at the parent and leaf nodes, a reinforcement learning formulation is put forth in this work. Furthermore, to endow with scalability so that the algorithm can effectively handle the curse of dimensionality, a deep reinforcement learning approach is also developed. Our novel caching policy relies on a deep Q-network to enforce the parent node with ability to learn-and-adapt to unknown policies of leaf nodes as well as spatio-temporal dynamic evolution of file requests, results in remarkable caching performance, as corroborated through numerical tests.
AB - Wireless and wireline networks, such as Internet, cellular, and content delivery networks are to serve end-user file requests proactively. To this aim, by storing anticipated highly popular files during off-peak periods, and fetching them to end-users during on-peak instances, these networks smoothen out the load fluctuations on the back-haul links. In this context, several practical networks comprise a parent caching node connected to multiple leaf nodes to serve end-user file requests. To model the two-way interactive influence between caching decisions at the parent and leaf nodes, a reinforcement learning formulation is put forth in this work. Furthermore, to endow with scalability so that the algorithm can effectively handle the curse of dimensionality, a deep reinforcement learning approach is also developed. Our novel caching policy relies on a deep Q-network to enforce the parent node with ability to learn-and-adapt to unknown policies of leaf nodes as well as spatio-temporal dynamic evolution of file requests, results in remarkable caching performance, as corroborated through numerical tests.
KW - Content delivery
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85091190588&partnerID=8YFLogxK
U2 - 10.1109/ICASSP40776.2020.9054485
DO - 10.1109/ICASSP40776.2020.9054485
M3 - Conference contribution
AN - SCOPUS:85091190588
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
SP - 3532
EP - 3536
BT - 2020 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2020 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2020 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2020
Y2 - 4 May 2020 through 8 May 2020
ER -