| Literature DB >> 34715533 |
Abstract
In this paper, a complementary learning scheme for experience transference of unknown continuous-time linear systems is proposed. The algorithm is inspired in the complementary learning properties that exhibit the hippocampus and neocortex learning systems via the striatum. The hippocampus is modelled as pattern-separated data of a human optimized controller. The neocortex is modelled as a Q-reinforcement learning algorithm which improves the hippocampus control policy. The complementary learning (striatum) is designed as an inverse reinforcement learning algorithm which relates the hippocampus and neocortex learning models to seek and transfer the weights of the hidden expert's utility function. Convergence of the proposed approach is analysed using Lyapunov recursions. Simulations are given to verify the proposed approach.Entities:
Keywords: Batch least squares; Complementary learning; Gradient-descent rule; Hippocampus and neocortex learning systems; Inverse reinforcement learning; Q-learning
Mesh:
Year: 2021 PMID: 34715533 DOI: 10.1016/j.neunet.2021.10.009
Source DB: PubMed Journal: Neural Netw ISSN: 0893-6080