Abstract
In this brief, the optimal containment control problem for a class of unknown nonlinear multi-agent systems (MASs) is studied via a time-aggregated (TA) model-free reinforcement learning (RL) algorithm. First, based on the idea of TA, the control policy is updated only when the system visits a finite subset of the state space. Thus, the control is event-triggered and not time-triggered. On this basis, a model-free TA-based value iteration (TA-VI) algorithm is proposed to learn the optimal control protocol. Since the finite important states are considered and the control is event-triggered, this algorithm requires fewer updating times and fewer computation than the conventional optimal containment control. Moreover, the TA-VI algorithm eliminates requirements on the function approximator and state discretization, which allows a strict convergence analysis via the mathematical induction method. Finally, simulation results are given to show the feasibility and superiority of the proposed algorithm.
| Original language | English |
|---|---|
| Pages (from-to) | 3393-3397 |
| Number of pages | 5 |
| Journal | IEEE Transactions on Circuits and Systems II: Express Briefs |
| Volume | 71 |
| Issue number | 7 |
| DOIs | |
| State | Published - 2024 |
| Externally published | Yes |
Keywords
- Time-aggregated
- model-free learning
- optimal containment control
- reinforcement learning
- value iteration
Fingerprint
Dive into the research topics of 'A Time-Aggregated Model-Free RL Algorithm for Optimal Containment Control of MASs'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver