Literature DB >> 33501138

From Semantics to Execution: Integrating Action Planning With Reinforcement Learning for Robotic Causal Problem-Solving.

Manfred Eppe1, Phuong D H Nguyen1, Stefan Wermter1.   

Abstract

Reinforcement learning is generally accepted to be an appropriate and successful method to learn robot control. Symbolic action planning is useful to resolve causal dependencies and to break a causally complex problem down into a sequence of simpler high-level actions. A problem with the integration of both approaches is that action planning is based on discrete high-level action- and state spaces, whereas reinforcement learning is usually driven by a continuous reward function. Recent advances in model-free reinforcement learning, specifically, universal value function approximators and hindsight experience replay, have focused on goal-independent methods based on sparse rewards that are only given at the end of a rollout, and only if the goal has been fully achieved. In this article, we build on these novel methods to facilitate the integration of action planning with model-free reinforcement learning. Specifically, the paper demonstrates how the reward-sparsity can serve as a bridge between the high-level and low-level state- and action spaces. As a result, we demonstrate that the integrated method is able to solve robotic tasks that involve non-trivial causal dependencies under noisy conditions, exploiting both data and knowledge.
Copyright © 2019 Eppe, Nguyen and Wermter.

Entities:  

Keywords:  causal puzzles; hierarchical architecture; neural networks; planning; reinforcement learning; robotics

Year:  2019        PMID: 33501138      PMCID: PMC7805615          DOI: 10.3389/frobt.2019.00123

Source DB:  PubMed          Journal:  Front Robot AI        ISSN: 2296-9144


  4 in total

1.  Do new caledonian crows solve physical problems through causal reasoning?

Authors:  A H Taylor; G R Hunt; F S Medina; R D Gray
Journal:  Proc Biol Sci       Date:  2009-01-22       Impact factor: 5.349

Review 2.  State representation learning for control: An overview.

Authors:  Timothée Lesort; Natalia Díaz-Rodríguez; Jean-Frano Is Goudou; David Filliat
Journal:  Neural Netw       Date:  2018-08-04

3.  Human-level control through deep reinforcement learning.

Authors:  Volodymyr Mnih; Koray Kavukcuoglu; David Silver; Andrei A Rusu; Joel Veness; Marc G Bellemare; Alex Graves; Martin Riedmiller; Andreas K Fidjeland; Georg Ostrovski; Stig Petersen; Charles Beattie; Amir Sadik; Ioannis Antonoglou; Helen King; Dharshan Kumaran; Daan Wierstra; Shane Legg; Demis Hassabis
Journal:  Nature       Date:  2015-02-26       Impact factor: 49.962

4.  Open-Ended Learning: A Conceptual Framework Based on Representational Redescription.

Authors:  Stephane Doncieux; David Filliat; Natalia Díaz-Rodríguez; Timothy Hospedales; Richard Duro; Alexandre Coninx; Diederik M Roijers; Benoît Girard; Nicolas Perrin; Olivier Sigaud
Journal:  Front Neurorobot       Date:  2018-09-25       Impact factor: 2.650

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.