TY - JOUR
T1 - Weighted double deep Q-network based reinforcement learning for bi-objective multi-workflow scheduling in the cloud
AU - Li, Huifang
AU - Huang, Jianghang
AU - Wang, Binyang
AU - Fan, Yushun
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
PY - 2022/4
Y1 - 2022/4
N2 - As a promising distributed paradigm, cloud computing provides a cost-effective deploying environment for hosting scientific applications due to its provisioning elastic, heterogeneous resources in a pay-per-use model. More and more applications modeled as workflows are being moved to the cloud, and time and cost become important for workflow execution. However, scheduling workflows is still a challenge due to their large-scale and complexity, as well as the cloud’s dynamic characteristics and different quotations. In this work, we propose a Weighted Double Deep Q-Network-based Reinforcement Learning algorithm (WDDQN-RL) for scheduling multiple workflows to obtain near-optimal solutions in a relatively short time with both makespan and cost minimized. Specifically, we first introduce a dynamic coefficient-based adaptive balancing method into WDDQN to improve the accuracy of the target value estimation by making a trade-off between Deep Q-Network (DQN) overestimation and Double Deep Q-Network (DDQN) underestimation. Second, pointer network-based agents and a two-level scheduling strategy are designed, where pointer networks are used to process a variable candidate task set in the first-level and one selected task is fed to agents in the second-level for allocating resources. Third, we present a dynamic sensing mechanism by adjusting the model’s attention to each individual objective for increasing the diversity of solutions while guaranteeing their quality. Experimental results show that our algorithm outperforms the benchmarking approaches in various indicators.
AB - As a promising distributed paradigm, cloud computing provides a cost-effective deploying environment for hosting scientific applications due to its provisioning elastic, heterogeneous resources in a pay-per-use model. More and more applications modeled as workflows are being moved to the cloud, and time and cost become important for workflow execution. However, scheduling workflows is still a challenge due to their large-scale and complexity, as well as the cloud’s dynamic characteristics and different quotations. In this work, we propose a Weighted Double Deep Q-Network-based Reinforcement Learning algorithm (WDDQN-RL) for scheduling multiple workflows to obtain near-optimal solutions in a relatively short time with both makespan and cost minimized. Specifically, we first introduce a dynamic coefficient-based adaptive balancing method into WDDQN to improve the accuracy of the target value estimation by making a trade-off between Deep Q-Network (DQN) overestimation and Double Deep Q-Network (DDQN) underestimation. Second, pointer network-based agents and a two-level scheduling strategy are designed, where pointer networks are used to process a variable candidate task set in the first-level and one selected task is fed to agents in the second-level for allocating resources. Third, we present a dynamic sensing mechanism by adjusting the model’s attention to each individual objective for increasing the diversity of solutions while guaranteeing their quality. Experimental results show that our algorithm outperforms the benchmarking approaches in various indicators.
KW - Cloud computing
KW - Multi-objective workflow scheduling
KW - Reinforcement learning
KW - Weighted double deep Q-networks
UR - http://www.scopus.com/inward/record.url?scp=85118409341&partnerID=8YFLogxK
U2 - 10.1007/s10586-021-03454-6
DO - 10.1007/s10586-021-03454-6
M3 - Article
AN - SCOPUS:85118409341
SN - 1386-7857
VL - 25
SP - 751
EP - 768
JO - Cluster Computing
JF - Cluster Computing
IS - 2
ER -