| Literature DB >> 24747417 |
Xudan Xu1, J Jim Zhu2, Ping Zhang3.
Abstract
This paper studies a Non-convex State-dependent Linear Quadratic Regulator (NSLQR) problem, in which the control penalty weighting matrix [Formula: see text] in the performance index is state-dependent. A necessary and sufficient condition for the optimal solution is established with a rigorous proof by Euler-Lagrange Equation. It is found that the optimal solution of the NSLQR problem can be obtained by solving a Pseudo-Differential-Riccati-Equation (PDRE) simultaneously with the closed-loop system equation. A Comparison Theorem for the PDRE is given to facilitate solution methods for the PDRE. A linear time-variant system is employed as an example in simulation to verify the proposed optimal solution. As a non-trivial application, a goal pursuit process in psychology is modeled as a NSLQR problem and two typical goal pursuit behaviors found in human and animals are reproduced using different control weighting [Formula: see text]. It is found that these two behaviors save control energy and cause less stress over Conventional Control Behavior typified by the LQR control with a constant control weighting [Formula: see text], in situations where only the goal discrepancy at the terminal time is of concern, such as in Marathon races and target hitting missions.Entities:
Mesh:
Year: 2014 PMID: 24747417 PMCID: PMC3991650 DOI: 10.1371/journal.pone.0094925
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Figure 1Three Different Behaviors: GGB, SMB and CCB.
Figure 2Trajectories of and .
The Values of Parameters in Simulation Case 01.
|
|
|
|
|
|
|
| |
|
|
| 0.413 | 0.5 |
|
|
| 0.237653 |
|
|
| 0.417 | 0.5 |
|
|
| 0.237672 |
|
|
| 0.421 | 0.5 |
|
|
| 0.237695 |
*Diff. RE: .
Figure 3System Behavior of Simulation Case 01: a) Control Energy Penalty R vs. Goal Discrepancy x; b) Goal Discrepancy x.
The Values of Parameters in Simulation Case 02.
|
|
|
|
| |
|
|
| 1 | 1 | 23.5 |
|
|
| 10 | 10 | 0 |
|
|
|
|
| 2 |
|
|
| 0.000001 | 0.000001 | 0.000001 |
|
|
| 1 | 1 | 1 |
|
|
| 2.831 | 0.903 | 6.841 |
Simulation Results of Three Goal Pursuit Behaviors.
|
|
|
|
| |
|
|
| 0.064 | 0.064 | 0.064 |
|
|
| 10.000 | 10.000 | 0.000 |
|
|
| 59.847 | 49.183 | 29.680 |
|
|
| 96.937 | 95.770 | 95.180 |
|
|
| 9.977 | 9.730 | 13.207 |
|
|
| 1.651 | 1.643 | 3.421 |
Figure 4Three Goal Pursuit Behaviors: a) Control Energy Penalty R vs. Time; b) Control Energy Penalty R vs. Goal Discrepancy x; c) Feedback Gain K; d) Goal Discrepancy x; e) Control Effort u vs. Time; f) Control Effort u vs. Goal Discrepancy x.