| Literature DB >> 35391942 |
Ashkan Zehfroosh1, Herbert G Tanner1.
Abstract
This paper offers a new hybrid probably approximately correct (PAC) reinforcement learning (RL) algorithm for Markov decision processes (MDPs) that intelligently maintains favorable features of both model-based and model-free methodologies. The designed algorithm, referred to as the Dyna-Delayed Q-learning (DDQ) algorithm, combines model-free Delayed Q-learning and model-based R-max algorithms while outperforming both in most cases. The paper includes a PAC analysis of the DDQ algorithm and a derivation of its sample complexity. Numerical results are provided to support the claim regarding the new algorithm's sample efficiency compared to its parents as well as the best known PAC model-free and model-based algorithms in application. A real-world experimental implementation of DDQ in the context of pediatric motor rehabilitation facilitated by infant-robot interaction highlights the potential benefits of the reported method.Entities:
Keywords: human-robot interaction; markov decision process; probably approximately correct; reinforcement learning; sample complexity
Year: 2022 PMID: 35391942 PMCID: PMC8982074 DOI: 10.3389/frobt.2022.797213
Source DB: PubMed Journal: Front Robot AI ISSN: 2296-9144
The ddq Algorithm.
|
|
FIGURE 1The grid-world example.
FIGURE 2The actual optimal policy in the grid-world example.
Average # of samples for reaching 4ɛ optimality.
| Algorithms | # Of samples |
|---|---|
| Delayed Q-learning | 6622 |
| R-max | 6727 |
|
| 5960 |
FIGURE 3A family of difficult-to-learn mdps (Strehl et al., 2009).
FIGURE 4The number of samples required by the Mormax algorithm.
FIGURE 5The required samples by UCB Q-learning algorithm.
The best possible performance on learning mdp M.
| Algorithms | # Of samples | # Of model resolution |
|---|---|---|
| Mormax | 7770 | 12.06 |
| UCB Q-learning | 8097 | 0 (model-free) |
|
| 5662 | 3.76 |
FIGURE 6Instance of play-based child-robot social interaction. Two robots are visible in the scene: a small humanoid nao, and a differential-drive small mobile robot toy Dash.
FIGURE 7MDP model for the game of chase between a mobile robot and an infant.
Accumulated rewards for the Dash robot. The “in” condition corresponds to the infant wearing the full-body-weight support mechanism (see Figure 6) and the “out” condition represents completely unassisted infant motion. The last two highlighted rows give outcomes on the reward obtained through the optimal policy learned by ddq. The 95% confidence interval for the accumulated rewards is [0.02893.4197] with a P-value of 0.047 7.
| Session # | “In” condition | “Out” condition |
|---|---|---|
| 1 | 1.081 0 | 5.063 2 |
| 2 | 2.142 8 | 1.836 7 |
| 3 | 2.043 3 | 1.793 4 |
| 4 | 2.663 5 | 2.468 3 |
| 5 | 4.258 0 | 5.938 5 |
| 6 | 3.458 6 | 2.544 1 |
| 7 |
|
|
| 8 |
|
|