Releasing source locating based on Multi-Agent Reinforcement Learning with reward function designed by maximum entropy

Zhi Pu Wang, Guang Rong Zeng, Lie Wei Deng, Wang Cao, Yao Guo

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Citation (Scopus)

Abstract

This paper is focused on locating the actual releasing source in the environment of multiple disturbance sources. The actual releasing source is located with multiple mobile sensors. In an attempt to avoid mobile sensors falling into the disturbance releasing source and gather at the actual releasing source quickly, an improved Multi-Agent Reinforcement Learning (MARL) with novel designed reward function is applied to guide the movement of mobile sensors. To ensure finding the actual releasing source with maximum releasing concentration, the reward function is designed based on maximum entropy (ME). Finally, MARL with reward function designed by ME and normal MARL are simulated and compared to verify the efficiency and advantage of this method.

Original languageEnglish
Title of host publicationProceedings of the 41st Chinese Control Conference, CCC 2022
EditorsZhijun Li, Jian Sun
PublisherIEEE Computer Society
Pages4688-4693
Number of pages6
ISBN (Electronic)9789887581536
DOIs
Publication statusPublished - 2022
Externally publishedYes
Event41st Chinese Control Conference, CCC 2022 - Hefei, China
Duration: 25 Jul 202227 Jul 2022

Publication series

NameChinese Control Conference, CCC
Volume2022-July
ISSN (Print)1934-1768
ISSN (Electronic)2161-2927

Conference

Conference41st Chinese Control Conference, CCC 2022
Country/TerritoryChina
CityHefei
Period25/07/2227/07/22

Keywords

  • Maximum entropy
  • Multi-Agent Reinforcement Learning
  • Releasing source locating

Fingerprint

Dive into the research topics of 'Releasing source locating based on Multi-Agent Reinforcement Learning with reward function designed by maximum entropy'. Together they form a unique fingerprint.

Cite this