| Literature DB >> 17278575 |
Asma Al-Tamimi, Murad Abu-Khalaf, Frank L Lewis.
Abstract
In this correspondence, adaptive critic approximate dynamic programming designs are derived to solve the discrete-time zero-sum game in which the state and action spaces are continuous. This results in a forward-in-time reinforcement learning algorithm that converges to the Nash equilibrium of the corresponding zero-sum game. The results in this correspondence can be thought of as a way to solve the Riccati equation of the well-known discrete-time H(infinity) optimal control problem forward in time. Two schemes are presented, namely: 1) a heuristic dynamic programming and 2) a dual-heuristic dynamic programming, to solve for the value function and the costate of the game, respectively. An H(infinity) autopilot design for an F-16 aircraft is presented to illustrate the results.Mesh:
Year: 2007 PMID: 17278575 DOI: 10.1109/tsmcb.2006.880135
Source DB: PubMed Journal: IEEE Trans Syst Man Cybern B Cybern ISSN: 1083-4419