A Path Planning Method Based on Deep Reinforcement Learning with Improved Prioritized Experience Replay for Human-Robot Collaboration

Deyu Sun, Jingqian Wen, Jingfei Wang, Xiaonan Yang*, Yaoguang Hu

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Owing to its ability to integrate human flexibility with robotic automation, human-robot collaboration possesses tremendous potential in intelligent manufacturing. A quintessential characteristic of this collaboration is the necessity for robotic arms to cooperate with humans in a dynamically changing environment, wherein humans could be considered as dynamic obstacles. One of the significant challenges in human-robot collaboration is the development of obstacle avoidance strategies for robotic path planning within dynamically changing environments. The inability of traditional two-dimensional path planning methods to handle high-dimensional spaces, therefore, many researchers have turned their attention to deep reinforcement learning, and many deep reinforcement learning methods have been applied to robotic arm path planning. However, most deep reinforcement learning models for robotic arm path planning require a significant amount of training time to achieve convergence. In this study, we introduce an algorithm that synergizes Soft Actor-Critic (SAC) with an improved version of Prioritized Experience Replay (PER)—SAC-iPER. We prioritizes experiences based on task-rewards, employing metrics such as time consumption and collision occurrences, in addition to task completion, to rank experiences. This reward-based ordering significantly boosts the learning process in both speed and quality. The results of this study significantly enhanced the training efficiency of deep reinforcement learning models for robotic arm path planning within human-robot collaboration, paving the way for the development of more efficient human-robot collaborative systems.

Original languageEnglish
Title of host publicationHuman-Computer Interaction - Thematic Area, HCI 2024, Held as Part of the 26th HCI International Conference, HCII 2024, Proceedings
EditorsMasaaki Kurosu, Ayako Hashizume
PublisherSpringer Science and Business Media Deutschland GmbH
Pages196-206
Number of pages11
ISBN (Print)9783031604119
DOIs
Publication statusPublished - 2024
EventHuman Computer Interaction thematic area of the 26th International Conference on Human-Computer Interaction, HCII 2024 - Washington, United States
Duration: 29 Jun 20244 Jul 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14685 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

ConferenceHuman Computer Interaction thematic area of the 26th International Conference on Human-Computer Interaction, HCII 2024
Country/TerritoryUnited States
CityWashington
Period29/06/244/07/24

Keywords

  • deep reinforcement learning
  • human-robot collaboration
  • PER
  • SAC

Fingerprint

Dive into the research topics of 'A Path Planning Method Based on Deep Reinforcement Learning with Improved Prioritized Experience Replay for Human-Robot Collaboration'. Together they form a unique fingerprint.

Cite this