TY - GEN
T1 - Multi-Attacker Multi-Defender Target Guarding Game Using Hierarchical Reinforcement Learning
AU - Dong, Shaoqian
AU - Huang, Yi
N1 - Publisher Copyright:
© 2025 Technical Committee on Control Theory, Chinese Association of Automation.
PY - 2025
Y1 - 2025
N2 - This study proposes a Hierarchical Reinforcement Learning (HRL) framework for multi-agent target guarding in dynamic environments with respectively coordinated attackers and defenders. The framework decomposes defender actions into patrol, pursuit, and encirclement subtasks, with a high-level multi-head attention mechanism dynamically allocating subtasks based on global observations. Defenders trained with Multi-Agent Proximal Policy Optimization (MAPPO) execute a low-level policy to engage subtasks. Customized reward functions promote collision avoidance, target protection, and coordination: patrol rewards optimize circular surveillance, pursuit rewards minimize distances to attackers, and encirclement rewards enhance cooperation. Attackers employ evasion tactics to breach defenses. Simulations in 2D environments demonstrate effective subtask transitions and coordinated interceptions, validating the framework's robustness against nonstationary interactions.
AB - This study proposes a Hierarchical Reinforcement Learning (HRL) framework for multi-agent target guarding in dynamic environments with respectively coordinated attackers and defenders. The framework decomposes defender actions into patrol, pursuit, and encirclement subtasks, with a high-level multi-head attention mechanism dynamically allocating subtasks based on global observations. Defenders trained with Multi-Agent Proximal Policy Optimization (MAPPO) execute a low-level policy to engage subtasks. Customized reward functions promote collision avoidance, target protection, and coordination: patrol rewards optimize circular surveillance, pursuit rewards minimize distances to attackers, and encirclement rewards enhance cooperation. Attackers employ evasion tactics to breach defenses. Simulations in 2D environments demonstrate effective subtask transitions and coordinated interceptions, validating the framework's robustness against nonstationary interactions.
KW - Hierarchical Reinforcement learning
KW - Multi-Agent Systems
KW - Target Guarding Game
UR - https://www.scopus.com/pages/publications/105020281648
U2 - 10.23919/CCC64809.2025.11178586
DO - 10.23919/CCC64809.2025.11178586
M3 - Conference contribution
AN - SCOPUS:105020281648
T3 - Chinese Control Conference, CCC
SP - 2743
EP - 2748
BT - Proceedings of the 44th Chinese Control Conference, CCC 2025
A2 - Sun, Jian
A2 - Yin, Hongpeng
PB - IEEE Computer Society
T2 - 44th Chinese Control Conference, CCC 2025
Y2 - 28 July 2025 through 30 July 2025
ER -