Finding the Equilibrium for Continuous Constrained Markov Games under the Average Criteria

Xiaofeng Jiang, Shuangwu Chen, Jian Yang*, Han Hu, Zhenliang Zhang

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

4 引用 (Scopus)

摘要

For Markov game with cost constraints and continuous actions, the local constraint of single-decision maker is the interacted result of joint actions taken by the other decision makers, and is usually eliminated by imposing penalties on the undesired states and policies, which may suffer from the failure of penalties as the game policy changes and the nonexistence of the mixed policies. In this article, a framework of the actor-critic deep neural network is utilized to solve this problem. The actor network establishes the continuous pure policy to replace the mixed policy, and the critic network converts the global interacted results into a local performance potential. The local search for a constrained equilibrium average objective is converted into an unconstrained minimax optimization. Based on the equivalent conversion, the optimality function of the local action is given to evaluate the influence of the single decision maker's action on the global system. The proposed algorithm simultaneously iterates the local constraint multiplier and policy along opposite directions, and a typical congestion control numerical result in the emerging Internet of Things shows the efficiency.

源语言英语
文章编号8972586
页(从-至)5399-5406
页数8
期刊IEEE Transactions on Automatic Control
65
12
DOI
出版状态已出版 - 12月 2020

指纹

探究 'Finding the Equilibrium for Continuous Constrained Markov Games under the Average Criteria' 的科研主题。它们共同构成独一无二的指纹。

引用此