Reinforcement Learning-Based Pathfinding for Multiple UAVs Facing Abrupt Hazardous Areas

  • Qizhen Wu
  • , Lei Chen*
  • , Kexin Liu
  • , Jinhu Lu
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Planning feasible paths for multiple unmanned aerial vehicles (UAVs) amidst abrupt hazardous areas is a critical safety challenge, where existing methods often lack safety guarantees and uncertainty handling. To address this, we propose a novel multi-agent reinforcement learning (MARL) approach for the UAV pathfinding problem. Our method ensures rapid responsiveness and adherence to safety constraints through the integration of a control barrier function, guaranteeing safe replanning even during sudden route changes. To overcome the potential inefficiency of purely reactive safety, we introduce a probabilistic neural network that quantifies hazard uncertainty, enhancing the anticipation of sudden dangers. Finally, to utilize swarm intelligence for mutual risk avoidance, the approach incorporates neighbors’ observations using a proximity-weighted mean-field mechanism, allowing each UAV to consider the impact of this aggregated information in its planning. Extensive simulations show that our method achieves a planning success rate surpassing 90% in transient environments, outperforming traditional planners and other MARL baselines. Real-world experiments further validate the approach’s adaptability, demonstrating its practical value for safety-critical missions.

Original languageEnglish
JournalIEEE Transactions on Automation Science and Engineering
DOIs
Publication statusAccepted/In press - 2026
Externally publishedYes

Keywords

  • Unmanned aerial vehicle
  • artificial intelligence
  • deep reinforcement learning
  • multi-agent system
  • pathfinding

Fingerprint

Dive into the research topics of 'Reinforcement Learning-Based Pathfinding for Multiple UAVs Facing Abrupt Hazardous Areas'. Together they form a unique fingerprint.

Cite this