Skip to main navigation Skip to search Skip to main content

Data-Informed Residual Reinforcement Learning for High-Dimensional Robotic Tracking Control

  • Cong Li
  • , Fangzhou Liu*
  • , Yongchao Wang
  • , Martin Buss
  • *Corresponding author for this work
  • National University of Defense Technology
  • Ningbo Institute of Intelligent Equipment Technology Company Ltd
  • Xidian University
  • Technical University of Munich

Research output: Contribution to journalArticlepeer-review

Abstract

The learning inefficiency of reinforcement learning (RL) from scratch hinders its practical application toward continuous robotic tracking control, especially for high-dimensional robots. This article proposes a data-informed residual reinforcement learning (DR-RL)-based robotic tracking control scheme applicable to robots with high dimensionality. The proposed DR-RL methodology outperforms common RL methods regarding sample efficiency and scalability. Specifically, we first decouple the original robot into low-dimensional robotic subsystems; and further utilize one-step backward data to construct incremental subsystems that are equivalent model-free representations of the aforementioned decoupled robotic subsystems. The formulated incremental subsystems allow for parallel learning to relieve computation load and offer us mathematical descriptions of robotic movements for conducting theoretical analysis. Then, we apply DR-RL to learn the tracking control policy, a combination of incremental base policy and incremental residual policy, under a parallel learning architecture. The incremental residual policy uses the guidance from the incremental base policy as the learning initialization and further learns from interactions with environments to endow the tracking control policy with adaptability toward dynamically changing environments. Our proposed DR-RL-based tracking control scheme is developed with rigorous theoretical analysis of system stability and weight convergence. The effectiveness of our proposed method is validated numerically on a 7-DoF KUKA iiwa robot manipulator and experimentally on a 3-DoF robot manipulator that would fail for other counterpart RL methods.

Original languageEnglish
Pages (from-to)1681-1691
Number of pages11
JournalIEEE/ASME Transactions on Mechatronics
Volume30
Issue number3
DOIs
StatePublished - 2025

Keywords

  • Residual reinforcement learning (RL)
  • parallel learning
  • robotic tracking control

Fingerprint

Dive into the research topics of 'Data-Informed Residual Reinforcement Learning for High-Dimensional Robotic Tracking Control'. Together they form a unique fingerprint.

Cite this