Abstract
In this work, we investigate the problem of dynamic task scheduling for multi-robot systems, in which a large number of robots collaborate to achieve a multi-objective optimization goal in transportation, rescue, etc. Considering the dynamic characteristics of tasks and robots in industrial scenarios, a reinforcement learning scheduling algorithm based on a meta-learning framework is proposed, which learns to interact with the environment to obtain an optimal solution. A DenseNet-like deep Q-network is designed to mine high level features of a state matrix, whose size changes dynamically with the scenario settings. By optimizing network parameters in inner and outer meta learning loops, the Q-network learns from the experience of multiple scheduling scenarios and obtains a generalized initialization parameter, which can be fine-tuned online to adapt to a new multi-robot system. The effectiveness of the proposed meta-scheduling approach is illustrated by numerical simulations in 9 different multi robot scenarios, achieving a 11.0% higher objective score and a 63.9% reduction in training time compared with a standard deep Q-Learning algorithm.
| Original language | English |
|---|---|
| Article number | 107109 |
| Journal | Computers and Operations Research |
| Volume | 182 |
| DOIs | |
| Publication status | Published - Oct 2025 |
| Externally published | Yes |
Keywords
- Meta learning
- Multi-robot system
- Reinforcement learning
- Task scheduling
Fingerprint
Dive into the research topics of 'Meta-learning for dynamic multi-robot task scheduling'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver