Abstract
Planning feasible paths for multiple unmanned aerial vehicles (UAVs) amidst abrupt hazardous areas is a critical safety challenge, where existing methods often lack safety guarantees and uncertainty handling. To address this, we propose a novel multi-agent reinforcement learning (MARL) approach for the UAV pathfinding problem. Our method ensures rapid responsiveness and adherence to safety constraints through the integration of a control barrier function, guaranteeing safe replanning even during sudden route changes. To overcome the potential inefficiency of purely reactive safety, we introduce a probabilistic neural network that quantifies hazard uncertainty, enhancing the anticipation of sudden dangers. Finally, to utilize swarm intelligence for mutual risk avoidance, the approach incorporates neighbors’ observations using a proximity-weighted mean-field mechanism, allowing each UAV to consider the impact of this aggregated information in its planning. Extensive simulations show that our method achieves a planning success rate surpassing 90% in transient environments, outperforming traditional planners and other MARL baselines. Real-world experiments further validate the approach’s adaptability, demonstrating its practical value for safety-critical missions.
| Original language | English |
|---|---|
| Journal | IEEE Transactions on Automation Science and Engineering |
| DOIs | |
| Publication status | Accepted/In press - 2026 |
| Externally published | Yes |
Keywords
- Unmanned aerial vehicle
- artificial intelligence
- deep reinforcement learning
- multi-agent system
- pathfinding
Fingerprint
Dive into the research topics of 'Reinforcement Learning-Based Pathfinding for Multiple UAVs Facing Abrupt Hazardous Areas'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver