中文版 | English
Title

Policy gradient adaptive dynamic programming for nonlinear discrete-time zero-sum games with unknown dynamics

Author
Corresponding AuthorZhao, Bo
Publication Years
2023
DOI
Source Title
ISSN
1432-7643
EISSN
1433-7479
Abstract
A novel policy gradient (PG) adaptive dynamic programming method is developed to deal with nonlinear discrete-time zero-sum games with unknown dynamics. To facilitate the implementation, a policy iteration algorithm is established to approximate the iterative Q-function, as well as the control and disturbance policies via three neural network (NN) approximators, respectively. Then, the iterative Q-function is exploited to update the control and disturbance policies via PG method. To stabilize the training process and improve the data usage efficiency, the experience replay technique is applied to train the weight vectors of the three NNs by using mini-batch empirical data from replay memory. Furthermore, the convergence in terms of the iterative Q-function is proved. Simulation results of two numerical examples are provided to show the effectiveness of the proposed method.
Keywords
URL[Source Record]
Indexed By
Language
English
SUSTech Authorship
Others
Funding Project
Beijing Natural Science Foundation[4212038] ; National Natural Science Foundation of China["61973330","62073085"] ; Open Research Project of the State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences[20210108] ; Open Research Project of the Key Laboratory of Industrial Internet of Things & Networked Control, Ministry of Education[2021FF10]
WOS Research Area
Computer Science
WOS Subject
Computer Science, Artificial Intelligence ; Computer Science, Interdisciplinary Applications
WOS Accession No
WOS:000915638100001
Publisher
ESI Research Field
COMPUTER SCIENCE
Data Source
Web of Science
Citation statistics
Cited Times [WOS]:0
Document TypeJournal Article
Identifierhttp://kc.sustech.edu.cn/handle/2SGJ60CL/430992
DepartmentDepartment of Mechanical and Energy Engineering
Affiliation
1.Beijing Normal Univ, Sch Syst Sci, Beijing 100875, Peoples R China
2.Chongqing Univ Posts & Telecommun, Key Lab Ind Internet Things & Networked Control, Minist Educ, Chongqing 400065, Peoples R China
3.Southern Univ Sci & Technol, Dept Mech & Energy Engn, Shenzhen 518055, Peoples R China
4.Univ Illinois, Dept Elect & Comp Engn, Chicago, IL 60607 USA
Recommended Citation
GB/T 7714
Lin, Mingduo,Zhao, Bo,Liu, Derong. Policy gradient adaptive dynamic programming for nonlinear discrete-time zero-sum games with unknown dynamics[J]. SOFT COMPUTING,2023.
APA
Lin, Mingduo,Zhao, Bo,&Liu, Derong.(2023).Policy gradient adaptive dynamic programming for nonlinear discrete-time zero-sum games with unknown dynamics.SOFT COMPUTING.
MLA
Lin, Mingduo,et al."Policy gradient adaptive dynamic programming for nonlinear discrete-time zero-sum games with unknown dynamics".SOFT COMPUTING (2023).
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Export to Excel
Export to Csv
Altmetrics Score
Google Scholar
Similar articles in Google Scholar
[Lin, Mingduo]'s Articles
[Zhao, Bo]'s Articles
[Liu, Derong]'s Articles
Baidu Scholar
Similar articles in Baidu Scholar
[Lin, Mingduo]'s Articles
[Zhao, Bo]'s Articles
[Liu, Derong]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Lin, Mingduo]'s Articles
[Zhao, Bo]'s Articles
[Liu, Derong]'s Articles
Terms of Use
No data!
Social Bookmark/Share
No comment.

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.