Szczegóły publikacji

Opis bibliograficzny

Reinforcement-learning-based fuzzy bipartite consensus for multiagent systems: a novel scaling off-policy learning scheme / Jing Wang, Qing Yang, Jinde Cao, Leszek RUTKOWSKI, Hao Shen // IEEE Transactions on Cybernetics ; ISSN  2168-2267 . — 2025 — vol. 55 no. 9, s. 4491–4501. — Bibliogr. s. 4500–4501, Abstr. — Publikacja dostępna online od: 2025-06-04. — L. Rutkowski - dod. afiliacja: Systems Research Institute of the Polish Academy of Sciences, Warsaw, Poland

Autorzy (5)

Słowa kluczowe

bipartite consensusscaling policy iteration PI algorithmTakagi Sugeno T S fuzzy modelMASsBCmultiagent systems

Dane bibliometryczne

ID BaDAP165162
Data dodania do BaDAP2025-12-22
Tekst źródłowyURL
DOI10.1109/TCYB.2025.3572104
Rok publikacji2025
Typ publikacjiartykuł w czasopiśmie
Otwarty dostęptak
Czasopismo/seriaIEEE Transactions on Cybernetics

Abstract

The bipartite consensus (BC) issue for nonlinear multiagent systems (NMASs) with unknown system dynamics information is investigated in this article. Initially, the dynamics of NMASs are represented using the Takagi-Sugeno (T-S) fuzzy model. Subsequently, to achieve distributed control, a minmax game policy is introduced, where each agent aims to minimize its performance index while its neighbors attempt to maximize it. Consequently, the BC problem for NMASs is reformulated as a zero-sum game, transforming the controller design into solving a set of game algebraic Riccati equations (GAREs). To solve such equations, a novel scaling off-policy iteration (PI) algorithm is proposed. The key features of the proposed learning algorithm can be outlined in three main aspects: 1) during the learning process, the reliance on system dynamics is relaxed; 2) compared with the PI method, the requirement for initial admissible control policies is eliminated; and 3) a more rapid convergence speed is achieved than traditional value iteration. Finally, the effectiveness and advantages of the proposed method are validated through a simulation example and a series of comparative experiments. © 2025 IEEE.

Publikacje, które mogą Cię zainteresować

artykuł
#150855Data dodania: 9.1.2024
Fuzzy $H_\infty$ control of discrete-time nonlinear Markov jump systems via a novel hybrid reinforcement Q-learning method / Jing Wang, Jiacheng Wu, Hao Shen, Jinde Cao, Leszek RUTKOWSKI // IEEE Transactions on Cybernetics ; ISSN 2168-2267. — 2023 — vol. 53 no. 11, s. 7380–7391. — Bibliogr. s. 7390–7391, Abstr. — Publikacja dostępna online od: 2022-11-23. — L. Rutkowski - dod. afiliacja: Systems Research Institute, Polish Academy of Sciences, Warsaw
artykuł
#165622Data dodania: 23.2.2026
$H_{\infty}$ control for interconnected systems with unknown system dynamics: a two-stage reinforcement learning method / Jinxu Liu, Hao Shen, Jing Wang, Jinde Cao, Leszek RUTKOWSKI // IEEE Transactions on Automation Science and Engineering ; ISSN  1545-5955 . — 2025 — vol. 22, s. 6388–6397. — Bibliogr. s. 6396–6397, Abstr. — Publikacja dostępna online od: 2024-08-21. — L. Rutkowski - dod. afiliacja: Polish Academy of Sciences, Warsaw, Poland