| Literature DB >> 27835670 |
Xiaoting Ji1, Yifeng Niu1, Lincheng Shen1.
Abstract
This paper presents a robust satisficing decision-making method for Unmanned Aerial Vehicles (<span class="Chemical">UAVs) executing complex missions in an uncertain environment. Motivated by the info-gap decision theory, we formulate this problem as a novel robust satisficing optimization problem, of which the objective is to maximize the robustness while satisfying some desired mission r<span class="Chemical">equirements. Specifically, a new info-gap based Markov Decision Process (IMDP) is constructed to abstract the uncertain UAV system and specify the complex mission requirements with the Linear Temporal Logic (LTL). A robust satisficing policy is obtained to maximize the robustness to the uncertain IMDP while ensuring a desired probability of satisfying the LTL specifications. To this end, we propose a two-stage robust satisficing solution strategy which consists of the construction of a product IMDP and the generation of a robust satisficing policy. In the first stage, a product IMDP is constructed by combining the IMDP with an automaton representing the LTL specifications. In the second, an algorithm based on robust dynamic programming is proposed to generate a robust satisficing policy, while an associated robustness evaluation algorithm is presented to evaluate the robustness. Finally, through Monte Carlo simulation, the effectiveness of our algorithms is demonstrated on an UAV search mission under severe uncertainty so that the resulting policy can maximize the robustness while reaching the desired performance level. Furthermore, by comparing the proposed method with other robust decision-making methods, it can be concluded that our policy can tolerate higher uncertainty so that the desired performance level can be guaranteed, which indicates that the proposed method is much more effective in real applications.Entities:
Mesh:
Year: 2016 PMID: 27835670 PMCID: PMC5105955 DOI: 10.1371/journal.pone.0166448
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1The space of possible values of uncertain variables.
Fig 2Robust satisficing decision-making framework.
Fig 3The solution schemes for the min-max robust decision-making method and the robust satisficing decision-making method.
(A) Solution scheme for the min-max robust decision-making method. (B) Solution scheme for the robust satisficing decision-making method.
Fig 4The monotonicity relationship between the highest worst-case LSP and the uncertainty level, and the maximum robustness.
Fig 5The mission space with atomic propositions.
Fig 6The transition probability for action ‘up’.
Fig 7The original trajectory of the robust satisficing policy through forward simulation.
Fig 8The trajectory of the robust satisficing policy removing overlaps.
Fig 9The robustness curve for the robust satisficing policy π*.
Fig 10The maximum robustness curve.
Robustness for different desired probability levels of satisfying LTL specifications.
| robustness of | 1.00 | 0.83 | 0.83 | 0.82 | 0.81 | 0.78 |
| maximum robustness | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 0.94 |
| robustness of | 0.74 | 0.69 | 0.60 | 0.40 | 0.14 | 0.00 |
| maximum robustness | 0.82 | 0.72 | 0.62 | 0.40 | 0.18 | 0.00 |
Desired probability level vs. true probability of satisfying the LTL specifications.
| DLSP | 0.90 | 0.80 | 0.70 | 0.60 |
|---|---|---|---|---|
| robust satisficing policy | ||||
| robustness | 0.40 | 0.62 | 0.72 | 0.82 |
| true LSP | 0.955 | 0.947 | 0.788 | 0.613 |
Fig 11The trajectory of the min-max robust policy π through forward simulations.
Fig 12The trajectory of the robust satisficing policy π through forward simulations.
Fig 13The histograms representing the true LSPs for the min-max robust policy π and the robust satisficing policy π at different uncertainty levels.
Fig 14The curves representing the true LSPs for the min-max robust policy π and the robust satisficing policy π at different uncertainty levels.
The true LSPs for the min-max robust policy π and the robust satisficing policy π at different uncertainty levels.
| LSP for | 1 | 0.97 | 0.97 | 0.86 | 0.80 | 0.59 |
| LSP for | 1 | 1 | 0.96 | 0.95 | 0.94 | 0.92 |
| LSP for | 0.48 | 0.43 | 0.14 | 0.02 | 0.00 | 0 |
| LSP for | 0.80 | 0.59 | 0.45 | 0.16 | 0.04 | 0 |
Fig 15The trajectory of the RDM robust policy π through forward simulations.
Fig 16The trajectory of the RDM robust policy π through forward simulations.
Fig 17The curves representing the true LSPs for the robust satisficing policy π and the RDM robust policies π and π at different uncertainty levels.
The true LSPs for the robust satisficing policy π and the RDM robust policies π and π at different uncertainty levels.
| LSP for | 1 | 1 | 0.96 | 0.95 | 0.94 | 0.92 |
| LSP for | 1 | 1 | 1 | 0.99 | 0.99 | 0.95 |
| LSP for | 0.85 | 0.81 | 0.81 | 0.75 | 0.77 | 0.75 |
| LSP for | 0.80 | 0.59 | 0.45 | 0.16 | 0.04 | 0 |
| LSP for | 0.76 | 0.69 | 0.29 | 0.21 | 0.05 | 0 |
| LSP for | 0.70 | 0.61 | 0.59 | 0.58 | 0.51 | 0.46 |