TY - JOUR
T1 - Free Market of Multi-Leader Multi-Follower Mobile Crowdsensing
T2 - An Incentive Mechanism Design by Deep Reinforcement Learning
AU - Zhan, Yufeng
AU - Liu, Chi Harold
AU - Zhao, Yinuo
AU - Zhang, Jiang
AU - Tang, Jian
N1 - Publisher Copyright:
© 2002-2012 IEEE.
PY - 2020/10/1
Y1 - 2020/10/1
N2 - The explosive increase of mobile devices with built-in sensors such as GPS, accelerometer, gyroscope and camera has made the design of mobile crowdsensing (MCS) applications possible, which create a new interface between humans and their surroundings. Until now, various MCS applications have been designed, where the task initiators (TIs) recruit mobile users (MUs) to complete the required sensing tasks. In this paper, deep reinforcement learning (DRL) based techniques are investigated to address the problem of assigning satisfactory but profitable amount of incentives to multiple TIs and MUs as a MCS game. Specifically, we first formulate the problem as a multi-leader and multi-follower Stackelberg game, where TIs are the leaders and MUs are the followers. Then, the existence of the Stackelberg Equilibrium (SE) is proved. Considering the challenge to compute the SE, a DRL based Dynamic Incentive Mechanism (DDIM) is proposed. It enables the TIs to learn the optimal pricing strategies directly from game experiences without knowing the private information of MUs. Finally, numerical experiments are provided to illustrate the effectiveness of the proposed incentive mechanism compared with both state-of-the-art and baseline approaches.
AB - The explosive increase of mobile devices with built-in sensors such as GPS, accelerometer, gyroscope and camera has made the design of mobile crowdsensing (MCS) applications possible, which create a new interface between humans and their surroundings. Until now, various MCS applications have been designed, where the task initiators (TIs) recruit mobile users (MUs) to complete the required sensing tasks. In this paper, deep reinforcement learning (DRL) based techniques are investigated to address the problem of assigning satisfactory but profitable amount of incentives to multiple TIs and MUs as a MCS game. Specifically, we first formulate the problem as a multi-leader and multi-follower Stackelberg game, where TIs are the leaders and MUs are the followers. Then, the existence of the Stackelberg Equilibrium (SE) is proved. Considering the challenge to compute the SE, a DRL based Dynamic Incentive Mechanism (DDIM) is proposed. It enables the TIs to learn the optimal pricing strategies directly from game experiences without knowing the private information of MUs. Finally, numerical experiments are provided to illustrate the effectiveness of the proposed incentive mechanism compared with both state-of-the-art and baseline approaches.
KW - Incentive mechanism
KW - deep reinforcement learning
KW - multi-leader multi-follower mobile crowdsensing
KW - stackelberg equilibrium
UR - http://www.scopus.com/inward/record.url?scp=85090983182&partnerID=8YFLogxK
U2 - 10.1109/TMC.2019.2927314
DO - 10.1109/TMC.2019.2927314
M3 - Article
AN - SCOPUS:85090983182
SN - 1536-1233
VL - 19
SP - 2316
EP - 2329
JO - IEEE Transactions on Mobile Computing
JF - IEEE Transactions on Mobile Computing
IS - 10
M1 - 8758205
ER -