| Literature DB >> 31470677 |
Shujun Ma1, Xinhui Bai2, Yinglei Wang2, Rui Fang2.
Abstract
The fusion of visual and inertial odometry has matured greatly due to the complementarity of the two sensors. However, the use of high-quality sensors and powerful processors in some applications is difficult due to size and cost limitations, and there are also many challenges in terms of robustness of the algorithm and computational efficiency. In this work, we present VIO-Stereo, a stereo visual-inertial odometry (VIO), which jointly combines the measurements of the stereo cameras and an inexpensive inertial measurement unit (IMU). We use nonlinear optimization to integrate visual measurements with IMU readings in VIO tightly. To decrease the cost of computation, we use the FAST feature detector to improve its efficiency and track features by the KLT sparse optical flow algorithm. We also incorporate accelerometer bias into the measurement model and optimize it together with other variables. Additionally, we perform circular matching between the previous and current stereo image pairs in order to remove outliers in the stereo matching and feature tracking steps, thus reducing the mismatch of feature points and improving the robustness and accuracy of the system. Finally, this work contributes to the experimental comparison of monocular visual-inertial odometry and stereo visual-inertial odometry by evaluating our method using the public EuRoC dataset. Experimental results demonstrate that our method exhibits competitive performance with the most advanced techniques.Entities:
Keywords: localization; simultaneous localization and mapping; state estimation; visual-inertial SLAM
Year: 2019 PMID: 31470677 PMCID: PMC6749198 DOI: 10.3390/s19173747
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1A block diagram illustrating our Visual-Inertial optimization framework.
Figure 2IMU pre-integration diagram.
Figure 3Mismatch accuracy of different types of cameras.
Figure 4The tracking of the visual front end: (a) left image; (b) right image.
Figure 5Trajectory compared with VINS in (a) MH_03_median, (b) MH 05 difficult, (c) V2_01_easy, and (d) V2_02_median.
Figure 6Translation error plot: (a) MH_03_median, (b) MH_05_median.
CPU load and memory utilization statistics of the algorithms (%).
| Sequence | CPU Load | Memory Utilization | ||
|---|---|---|---|---|
| Ours | VINS-Mono | Ours | VINS-Mono | |
| V1_01_easy |
| 43.14 |
| 34.09 |
| V1_02_medium |
| 36.78 |
| 33.22 |
| V1_03_difficult | 36.72 |
|
| 34.75 |
| V2_01_easy | 42.32 |
|
| 32.02 |
| V2_02_medium | 45.88 |
|
| 34.48 |
| V2_03_difficult | / | 40.76 |
| 34.18 |
| MH_01_easy | 40.59 |
|
| 34.08 |
| MH_02_easy |
| 39.08 |
| 34.11 |
| MH_03_medium |
| 40.69 |
| 35.04 |
| MH_04_difficult | 44.02 |
|
| 34.77 |
| MH_05_difficult | 45.38 |
|
| 37.8 |
| Average | 40.48 |
|
| 34.41 |
The numbers in bold indicate the lower computational cost of the algorithms.
RMSE of the algorithms (m).
| Sequence | Ours | VINS-Mono | S-MSCKF | OKVIS |
|---|---|---|---|---|
| V1_01_easy |
| 0.043 | 0.065 | 0.208 |
| V1_02_medium | 0.051 |
| 0.154 | 0.19 |
| V1_03_difficult | 0.105 |
| 0.281 | 0.195 |
| V2_01_easy |
| 0.062 | 0.069 | 0.172 |
| V2_02_medium |
| 0.121 | 0.147 | 0.181 |
| V2_03_difficult | / |
| / | 0.327 |
| MH_01_easy |
| 0.066 | 0.215 | 0.169 |
| MH_02_easy |
| 0.081 | 0.226 | 0.193 |
| MH_03_medium |
| 0.045 | 0.202 | 0.285 |
| MH_04_difficult | 0.134 |
| 0.323 | 0.386 |
| MH_05_difficult |
| 0.095 | 0.224 | 0.455 |
| Average |
| 0.081 | 0.19 | 0.251 |
The numbers in bold indicate the smallest values of RMSE of the algorithms.