Meta-learning for dynamic multi-robot task scheduling

Peng Song, Huaiyu Chen, Kaixin Cui, Junzheng Wang, Dawei Shi*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

In this work, we investigate the problem of dynamic task scheduling for multi-robot systems, in which a large number of robots collaborate to achieve a multi-objective optimization goal in transportation, rescue, etc. Considering the dynamic characteristics of tasks and robots in industrial scenarios, a reinforcement learning scheduling algorithm based on a meta-learning framework is proposed, which learns to interact with the environment to obtain an optimal solution. A DenseNet-like deep Q-network is designed to mine high level features of a state matrix, whose size changes dynamically with the scenario settings. By optimizing network parameters in inner and outer meta learning loops, the Q-network learns from the experience of multiple scheduling scenarios and obtains a generalized initialization parameter, which can be fine-tuned online to adapt to a new multi-robot system. The effectiveness of the proposed meta-scheduling approach is illustrated by numerical simulations in 9 different multi robot scenarios, achieving a 11.0% higher objective score and a 63.9% reduction in training time compared with a standard deep Q-Learning algorithm.

Original languageEnglish
Article number107109
JournalComputers and Operations Research
Volume182
DOIs
Publication statusPublished - Oct 2025
Externally publishedYes

Keywords

  • Meta learning
  • Multi-robot system
  • Reinforcement learning
  • Task scheduling

Cite this