Literature DB >> 31500931

Improved robustness of reinforcement learning policies upon conversion to spiking neuronal network platforms applied to Atari Breakout game.

Devdhar Patel1, Hananel Hazan2, Daniel J Saunders2, Hava T Siegelmann2, Robert Kozma3.   

Abstract

Deep Reinforcement Learning (RL) demonstrates excellent performance on tasks that can be solved by trained policy. It plays a dominant role among cutting-edge machine learning approaches using multi-layer Neural networks (NNs). At the same time, Deep RL suffers from high sensitivity to noisy, incomplete, and misleading input data. Following biological intuition, we involve Spiking Neural Networks (SNNs) to address some deficiencies of deep RL solutions. Previous studies in image classification domain demonstrated that standard NNs (with ReLU nonlinearity) trained using supervised learning can be converted to SNNs with negligible deterioration in performance. In this paper, we extend those conversion results to the domain of Q-Learning NNs trained using RL. We provide a proof of principle of the conversion of standard NN to SNN. In addition, we show that the SNN has improved robustness to occlusion in the input image. Finally, we introduce results with converting full-scale Deep Q-network to SNN, paving the way for future research to robust Deep RL applications.
Copyright © 2019 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  Atari; Deep learning; Reinforcement learning; Robustness; Spiking neural networks

Mesh:

Year:  2019        PMID: 31500931     DOI: 10.1016/j.neunet.2019.08.009

Source DB:  PubMed          Journal:  Neural Netw        ISSN: 0893-6080


  4 in total

1.  Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning.

Authors:  Sooyoung Jang; Hyung-Il Kim
Journal:  Sensors (Basel)       Date:  2022-08-04       Impact factor: 3.847

2.  Solving the spike feature information vanishing problem in spiking deep Q network with potential based normalization.

Authors:  Yinqian Sun; Yi Zeng; Yang Li
Journal:  Front Neurosci       Date:  2022-08-25       Impact factor: 5.152

3.  Training Spiking Neural Networks for Reinforcement Learning Tasks With Temporal Coding Method.

Authors:  Guanlin Wu; Dongchen Liang; Shaotong Luan; Ji Wang
Journal:  Front Neurosci       Date:  2022-08-17       Impact factor: 5.152

4.  Training spiking neuronal networks to perform motor control using reinforcement and evolutionary learning.

Authors:  Daniel Haşegan; Matt Deible; Christopher Earl; David D'Onofrio; Hananel Hazan; Haroon Anwar; Samuel A Neymotin
Journal:  Front Comput Neurosci       Date:  2022-09-30       Impact factor: 3.387

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.