Skip to main navigation Skip to search Skip to main content

A Visual Reinforcement Learning-Based Separate Primitive Policy for Peg-in-Hole Tasks

  • Zichun Xu
  • , Zhaomin Wang
  • , Yuntao Li
  • , Lei Zhuang
  • , Zhiyuan Zhao
  • , Guocai Yang
  • , Jingdong Zhao*
  • *Corresponding author for this work
  • Harbin Institute of Technology
  • UBTECH Robotics Corp. Ltd.
  • Shenzhen Academy of Robotics
  • Shandong University

Research output: Contribution to journalArticlepeer-review

Abstract

For peg-in-hole tasks, humans rely on binocular visual perception to locate the peg above the hole surface and then proceed with insertion. This letter draws insights from this behavior to enable agents to learn efficient assembly strategies through visual reinforcement learning. Hence, we propose a Separate Primitive Policy (S2P) to learn how to derive location and insertion actions simultaneously. S2P is compatible with model-free reinforcement learning algorithms. Ten insertion tasks featuring different polygons are developed as benchmarks for evaluations. Simulation experiments show that S2P can boost the sample efficiency and success rate even with force constraints. Real-world experiments are also performed to verify the feasibility of S2P. Ablations are finally given to discuss the generalizability of S2P and some factors that affect its performance.

Original languageEnglish
Pages (from-to)3748-3755
Number of pages8
JournalIEEE Robotics and Automation Letters
Volume11
Issue number3
DOIs
StatePublished - 2026

Keywords

  • Visual reinforcement learning
  • peg-in-hole
  • sim2real

Fingerprint

Dive into the research topics of 'A Visual Reinforcement Learning-Based Separate Primitive Policy for Peg-in-Hole Tasks'. Together they form a unique fingerprint.

Cite this