| Literature DB >> 31766236 |
Wenlei Liu1, Sentang Wu1, Zhongbo Wu2, Xiaolong Wu3.
Abstract
The novel contribution of this paper is to propose an incremental pose map optimization for monocular vision simultaneous localization and mapping (SLAM) based on similarity transformation, which can effectively solve the scale drift problem of SLAM for monocular vision and eliminate the cumulative error by global optimization. With the method of mixed inverse depth estimation based on a probability graph, the problem of the uncertainty of depth estimation is effectively solved and the robustness of depth estimation is improved. Firstly, this paper proposes a method combining the sparse direct method based on histogram equalization and the feature point method for front-end processing, and the mixed inverse depth estimation method based on a probability graph is used to estimate the depth information. Then, a bag-of-words model based on the mean initialization K-means is proposed for closed-loop feature detection. Finally, the incremental pose map optimization method based on similarity transformation is proposed to process the back end to optimize the pose and depth information of the camera. When the closed loop is detected, global optimization is carried out to effectively eliminate the cumulative error of the system. In this paper, indoor and outdoor environmental experiments are carried out using open data sets, such as TUM and KITTI, which fully proves the effectiveness of this method. Closed-loop detection experiments using hand-held cameras verify the importance of closed-loop detection. This method can effectively solve the scale drift problem of monocular vision SLAM and has strong robustness.Entities:
Keywords: bag-of-words; histogram equalization; incremental pose map; monocular vision SLAM; probability graph; similarity transformation; sparse direct method
Year: 2019 PMID: 31766236 PMCID: PMC6891346 DOI: 10.3390/s19224945
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1System schematic diagram of incremental pose map optimization for monocular vision SLAM based on similarity transformation.
Figure 2Probability diagram of Gauss-uniform mixed probability distribution model.
Figure 3A schematic diagram of the generation process of a K-ary tree dictionary using the Thiessen polygon.
Figure 4Schematic diagram of the incremental pose map optimization method.
Figure 5Comparison of TUM and EuRoC data sets before and after optimization. (a) Before TUM data set optimization; (b) After TUM data set optimization; (c) Before EuRoC data set optimization; (d) After EuRoC dataset optimization.
Comparison of indoor experimental errors.
| TUM Data Set | EuRoC Data Set | |||
|---|---|---|---|---|
| ORB-SLAM2 | Incremental Pose SLAM | ORB-SLAM2 | Incremental Pose SLAM | |
| Root mean square error (RMSE) | 0.41 m | 0.34 m | 0.86 m | 0.75 m |
| Mean error | 0.40 m | 0.34 m | 0.76 m | 0.66 m |
| Median error | 0.41 m | 0.35 m | 0.81 m | 0.69 m |
| Standard deviation of error | 0.05 m | 0.04 m | 0.40 m | 0.36 m |
| Minimum error | 0.31 m | 0.26 m | 0.06 m | 0.06 m |
| Maximum error | 0.51 m | 0.44 m | 1.50 m | 1.36 m |
| Average tracking time (s) | 0.037 s | 0.038 s | 0.042 s | 0.043 s |
Figure 6Experimental trajectory image of the KITTI dataset. (a) Comparison of GPS and ORB_SLAM2; (b) Comparison of GPS and incremental pose SLAM.
Comparison of outdoor experimental errors.
| ORB-SLAM2 | Incremental Pose SLAM | |
|---|---|---|
| Root mean square error (RMSE) | 10.52 m | 6.25 m |
| Mean error | 10.21 m | 5.98 m |
| Median error | 10.11 m | 6.02 m |
| Standard deviation of error | 4.68 m | 2.01 m |
| Minimum error | 3.89 m | 2.13 m |
| Maximum error | 18.98 m | 10.63 m |
Figure 7The experiment of the hand-held camera with or without closed-loop detection. (a) Comparison of the IMU and no-closed loop; (b) Comparison of IMU and closed-loop.
Figure 8Precision-recall curves achieved by ORB, LBD, and proposed method. (a) Based on the data set of TUM; (b) Based on the data set of KITTI; (c) Based on the data set of EuRoC.