Literature DB >> 28559954

Route searching based on neural networks and heuristic reinforcement learning.

Fengyun Zhang1,2, Shukai Duan1,2, Lidan Wang1,2.   

Abstract

In this paper, an improved and much stronger RNH-QL method based on RBF network and heuristic Q-learning was put forward for route searching in a larger state space. Firstly, it solves the problem of inefficiency of reinforcement learning if a given problem's state space is increased and there is a lack of prior information on the environment. Secondly, RBF network as weight updating rule, reward shaping can give an additional feedback to the agent in some intermediate states, which will help to guide the agent towards the goal state in a more controlled fashion. Meanwhile, with the process of Q-learning, it is accessible to the underlying dynamic knowledge, instead of the need of background knowledge of an upper level RBF network. Thirdly, it improves the learning efficiency by incorporating the greedy exploitation strategy to train the neural network, which has been testified by the experimental results.

Entities:  

Keywords:  Greedy exploitation; Heuristic reinforcement learning; Neural network; Route searching

Year:  2017        PMID: 28559954      PMCID: PMC5430242          DOI: 10.1007/s11571-017-9423-7

Source DB:  PubMed          Journal:  Cogn Neurodyn        ISSN: 1871-4080            Impact factor:   5.082


  11 in total

1.  Online learning of shaping rewards in reinforcement learning.

Authors:  Marek Grześ; Daniel Kudenko
Journal:  Neural Netw       Date:  2010-01-11

2.  A Fast Adaptive Tunable RBF Network For Nonstationary Systems.

Authors:  Hao Chen; Yu Gong; Xia Hong; Sheng Chen
Journal:  IEEE Trans Cybern       Date:  2015-10-28       Impact factor: 11.448

3.  Impulsive Effects and Stability Analysis on Memristive Neural Networks With Variable Delays.

Authors:  Shukai Duan; Huamin Wang; Lidan Wang; Tingwen Huang; Chuandong Li
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2016-01-05       Impact factor: 10.451

4.  Heuristically-accelerated multiagent reinforcement learning.

Authors:  Reinaldo A C Bianchi; Murilo F Martins; Carlos H C Ribeiro; Anna Helena Reali Costa
Journal:  IEEE Trans Cybern       Date:  2014-02       Impact factor: 11.448

5.  Analysis of stability of neural network with inhibitory neurons.

Authors:  Yan Liu; Rubin Wang; Zhikang Zhang; Xianfa Jiao
Journal:  Cogn Neurodyn       Date:  2009-10-30       Impact factor: 5.082

6.  Computational models of reinforcement learning: the role of dopamine as a reward signal.

Authors:  R D Samson; M J Frank; Jean-Marc Fellous
Journal:  Cogn Neurodyn       Date:  2010-03-21       Impact factor: 5.082

7.  Equilibrium analysis and phase synchronization of two coupled HR neurons with gap junction.

Authors:  Haixia Wang; Qingyun Wang; Qishao Lu; Yanhong Zheng
Journal:  Cogn Neurodyn       Date:  2012-09-28       Impact factor: 5.082

8.  The neural basis of human error processing: reinforcement learning, dopamine, and the error-related negativity.

Authors:  Clay B Holroyd; Michael G H Coles
Journal:  Psychol Rev       Date:  2002-10       Impact factor: 8.934

9.  A neural network model of attention-modulated neurodynamics.

Authors:  Yuqiao Gu; Hans Liljenström
Journal:  Cogn Neurodyn       Date:  2007-10-02       Impact factor: 5.082

10.  Memristor-based cellular nonlinear/neural network: design, analysis, and applications.

Authors:  Shukai Duan; Xiaofang Hu; Zhekang Dong; Lidan Wang; Pinaki Mazumder
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2014-07-21       Impact factor: 10.451

View more
  2 in total

1.  Adaptive sparse coding based on memristive neural network with applications.

Authors:  Xun Ji; Xiaofang Hu; Yue Zhou; Zhekang Dong; Shukai Duan
Journal:  Cogn Neurodyn       Date:  2019-05-04       Impact factor: 5.082

2.  Design exploration predicts designer creativity: a deep learning approach.

Authors:  Yu-Cheng Liu; Chaoyun Liang
Journal:  Cogn Neurodyn       Date:  2020-01-19       Impact factor: 5.082

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.