An efficient data-driven framework of hybrid dynamics for real-time modeling and control of continuum robots

Research output: Contribution to journalArticlepeer-review

Abstract

Continuum robots exhibit exceptional compliance and adaptability. However, their intrinsic nonlinear dynamics poses significant challenges for real-time control under dynamic conditions. This paper presents a new data-driven framework of hybrid dynamics (HD) for modeling and control of continuum robots. This framework decomposes the nonlinear dynamics of a continuum robot into the nonlinear statics of global deformation and the linear dynamics of relative vibration, and leads to two sub-models of global deformation statics (GDS) and relative motion dynamics (RMD) via different data-driven approaches. The first sub-model uses a deep neural network to predict nonlinear static deformations, while the second sub-model utilizes dynamics mode decomposition for relative vibration compensation. The nonlinear integration of the two sub-models establishes the physics-embedded data-driven model with reduced complexity. Combining quasi-static positioning and model-predictive vibration suppression leads to design of a hierarchical dynamic controller. The experiments of a cable-driven continuum robot demonstrate precise trajectory tracking and effective vibration suppression under payload and high-speed conditions. This data-driven modeling and control framework balances computational efficiency and control performance, enabling practical advances of continuum robots under complex working scenarios.

Original languageEnglish
Article number106361
JournalMechanism and Machine Theory
Volume220
DOIs
Publication statusPublished - Apr 2026

Keywords

  • Continuum robots
  • Data-driven modeling
  • Dynamic control
  • Hybrid dynamics framework
  • Vibration suppression

Fingerprint

Dive into the research topics of 'An efficient data-driven framework of hybrid dynamics for real-time modeling and control of continuum robots'. Together they form a unique fingerprint.

Cite this