TY - GEN
T1 - Real-Time Obstacle Avoidance for Magnetic Microswimmers Based on Proximal Policy Optimization
AU - Yang, Haotian
AU - Niu, Zhenyang
AU - Hu, Jincheng
AU - Wu, Anping
AU - Li, Wenbo
AU - Nie, Ruhao
AU - Zhong, Shihao
AU - Hou, Yaozhen
AU - Wang, Huaping
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Magnetic helical microswimmers represent a promising technology for biomedical applications such as minimally invasive surgery and targeted drug delivery. However, autonomous navigation in dynamic and unstructured environments characteristic of biological systems remains a critical challenge. Traditional path planning algorithms struggle with real-time adaptation to unpredictable environmental changes and dynamic obstacles. This paper proposes a novel Deep Reinforcement Learning approach tailored for autonomous obstacle avoidance in magnetic helical microswimmers. The navigation problem is framed as a Markov Decision Process, and a physics-informed virtual environment is developed to precisely simulate the dynamics of microswimmers and their interactions with obstacles. A Deep Reinforcement Learning agent is developed through Proximal Policy Optimization with domain randomization to enhance policy robustness and generalization across diverse conditions. Comprehensive evaluation through simulations and real-world experiments demonstrates superior performance across static, fluctuating, and flowing obstacle scenarios. Our Deep Reinforcement Learning-based controller achieves navigation success rates of 75% in environments with 30 static obstacles and 60% with 30 dynamic obstacles, significantly outperforming traditional navigation methods in both success rate and path efficiency. The results validate the approach's potential for practical biomedical applications requiring autonomous microrobot navigation in complex in vivo environments.
AB - Magnetic helical microswimmers represent a promising technology for biomedical applications such as minimally invasive surgery and targeted drug delivery. However, autonomous navigation in dynamic and unstructured environments characteristic of biological systems remains a critical challenge. Traditional path planning algorithms struggle with real-time adaptation to unpredictable environmental changes and dynamic obstacles. This paper proposes a novel Deep Reinforcement Learning approach tailored for autonomous obstacle avoidance in magnetic helical microswimmers. The navigation problem is framed as a Markov Decision Process, and a physics-informed virtual environment is developed to precisely simulate the dynamics of microswimmers and their interactions with obstacles. A Deep Reinforcement Learning agent is developed through Proximal Policy Optimization with domain randomization to enhance policy robustness and generalization across diverse conditions. Comprehensive evaluation through simulations and real-world experiments demonstrates superior performance across static, fluctuating, and flowing obstacle scenarios. Our Deep Reinforcement Learning-based controller achieves navigation success rates of 75% in environments with 30 static obstacles and 60% with 30 dynamic obstacles, significantly outperforming traditional navigation methods in both success rate and path efficiency. The results validate the approach's potential for practical biomedical applications requiring autonomous microrobot navigation in complex in vivo environments.
UR - https://www.scopus.com/pages/publications/105030500255
U2 - 10.1109/CBS65871.2025.11267633
DO - 10.1109/CBS65871.2025.11267633
M3 - Conference contribution
AN - SCOPUS:105030500255
T3 - 2025 IEEE International Conference on Cyborg and Bionic Systems, CBS 2025
SP - 13
EP - 18
BT - 2025 IEEE International Conference on Cyborg and Bionic Systems, CBS 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2025 IEEE International Conference on Cyborg and Bionic Systems, CBS 2025
Y2 - 17 October 2025 through 19 October 2025
ER -