Research on Multi-Agent Region Coverage Search Based on Multi-Task Reinforcement Learning

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Cluster search and area coverage technologies are widely applied in both military and civilian fields. This paper divides the task area into grids and studies the multi-agent coverage search algorithm in unknown environments based on multi-task reinforcement learning. The main focus is on accelerating the training of existing schemes on the basis of the multi-agent area coverage search algorithm based on reinforcement learning. A multi-task reinforcement learning training method based on a hard parameter sharing network model is proposed. The area coverage search problem is decomposed into coverage, search, and obstacle avoidance problems, and corresponding network update strategies are designed. Through comparative experiments, it is verified that the multi-task reinforcement learning training method can effectively improve the convergence speed of the algorithm and increase the robustness of the model.

Original languageEnglish
Title of host publicationProceedings of the 44th Chinese Control Conference, CCC 2025
EditorsJian Sun, Hongpeng Yin
PublisherIEEE Computer Society
Pages5685-5690
Number of pages6
ISBN (Electronic)9789887581611
DOIs
Publication statusPublished - 2025
Externally publishedYes
Event44th Chinese Control Conference, CCC 2025 - Chongqing, China
Duration: 28 Jul 202530 Jul 2025

Publication series

NameChinese Control Conference, CCC
ISSN (Print)1934-1768
ISSN (Electronic)2161-2927

Conference

Conference44th Chinese Control Conference, CCC 2025
Country/TerritoryChina
CityChongqing
Period28/07/2530/07/25

Keywords

  • covering search tasks
  • multi-agent
  • multi-task reinforcement learning
  • unknown environment

Fingerprint

Dive into the research topics of 'Research on Multi-Agent Region Coverage Search Based on Multi-Task Reinforcement Learning'. Together they form a unique fingerprint.

Cite this