Skip to main navigation Skip to search Skip to main content

A Time-Aggregated Model-Free RL Algorithm for Optimal Containment Control of MASs

  • Xiongtao Shi
  • , Yanjie Li*
  • , Chenglong Du
  • , Weihua Gui
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

In this brief, the optimal containment control problem for a class of unknown nonlinear multi-agent systems (MASs) is studied via a time-aggregated (TA) model-free reinforcement learning (RL) algorithm. First, based on the idea of TA, the control policy is updated only when the system visits a finite subset of the state space. Thus, the control is event-triggered and not time-triggered. On this basis, a model-free TA-based value iteration (TA-VI) algorithm is proposed to learn the optimal control protocol. Since the finite important states are considered and the control is event-triggered, this algorithm requires fewer updating times and fewer computation than the conventional optimal containment control. Moreover, the TA-VI algorithm eliminates requirements on the function approximator and state discretization, which allows a strict convergence analysis via the mathematical induction method. Finally, simulation results are given to show the feasibility and superiority of the proposed algorithm.

Original languageEnglish
Pages (from-to)3393-3397
Number of pages5
JournalIEEE Transactions on Circuits and Systems II: Express Briefs
Volume71
Issue number7
DOIs
StatePublished - 2024
Externally publishedYes

Keywords

  • Time-aggregated
  • model-free learning
  • optimal containment control
  • reinforcement learning
  • value iteration

Fingerprint

Dive into the research topics of 'A Time-Aggregated Model-Free RL Algorithm for Optimal Containment Control of MASs'. Together they form a unique fingerprint.

Cite this