| Literature DB >> 33501200 |
Andreas B Martinsen1, Anastasios M Lekkas1,2, Sébastien Gros1, Jon Arne Glomsrud3, Tom Arne Pedersen3.
Abstract
We present a reinforcement learning-based (RL) control scheme for trajectory tracking of fully-actuated surface vessels. The proposed method learns online both a model-based feedforward controller, as well an optimizing feedback policy in order to follow a desired trajectory under the influence of environmental forces. The method's efficiency is evaluated via simulations and sea trials, with the unmanned surface vehicle (USV) ReVolt performing three different tracking tasks: The four corner DP test, straight-path tracking and curved-path tracking. The results demonstrate the method's ability to accomplish the control objectives and a good agreement between the performance achieved in the Revolt Digital Twin and the sea trials. Finally, we include an section with considerations about assurance for RL-based methods and where our approach stands in terms of the main challenges.Entities:
Keywords: approximate dynamic programming (ADP); autonomous ships; dynamic positioning (DP); model-based adaptive control; optimal control; reinforcement learning; system identification; trajectory tracking
Year: 2020 PMID: 33501200 PMCID: PMC7806118 DOI: 10.3389/frobt.2020.00032
Source DB: PubMed Journal: Front Robot AI ISSN: 2296-9144