| Literature DB >> 27529256 |
Kaichang Di1, Qiang Zhao2,3, Wenhui Wan4, Yexin Wang5, Yunjun Gao6.
Abstract
In the study of SLAM problem using an RGB-D camera, depth information and visual information as two types of primary measurement data are rarely tightly coupled during refinement of camera pose estimation. In this paper, a new method of RGB-D camera SLAM is proposed based on extended bundle adjustment with integrated 2D and 3D information on the basis of a new projection model. First, the geometric relationship between the image plane coordinates and the depth values is constructed through RGB-D camera calibration. Then, 2D and 3D feature points are automatically extracted and matched between consecutive frames to build a continuous image network. Finally, extended bundle adjustment based on the new projection model, which takes both image and depth measurements into consideration, is applied to the image network for high-precision pose estimation. Field experiments show that the proposed method has a notably better performance than the traditional method, and the experimental results demonstrate the effectiveness of the proposed method in improving localization accuracy.Entities:
Keywords: Kinect; RGB-D camera; SLAM; bundle adjustment; projection model
Year: 2016 PMID: 27529256 PMCID: PMC5017450 DOI: 10.3390/s16081285
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Flowchart of our method.
Figure 2Spatial relationship of Kinect cameras.
Internal parameters of depth camera and RGB camera. The unit for f, f, x0, y0 is pixel.
| Depth | 519.95 | 519.55 | 315.82 | 238.71 | 0.04810 | 0.19281 | 0.0 | 0.00458 | 0.00014 |
| RGB | 584.35 | 584.33 | 317.97 | 252.80 | 0.10585 | 0.27096 | 0.0 | 0.00504 | 0.00166 |
External parameters of depth camera and RGB camera.
| −0.00079 | −0.00084 | −0.00541 | |
| −25.59983 | 0.16700 | −0.40571 |
Figure 3Features detection and matching results. (a) 2D feature detection and matching; (b) 3D feature detection and matching.
Figure 4Illustration of projection model of RGB-D camera.
Figure 5The RGB-D camera mounted on the moving platform.
Figure 6Typical RGB images and corresponding depth images acquired in two experiments. (a) Typical images acquired in a tunnel in Experiment I; (b) Typical images acquired in an outdoor field in Experiment II. Images in the first row were captured by the RGB camera, and images in the second row were captured by the depth camera. The areas, which are out of the imaging range (0.5 m to 4 m) or without reflected infrared light, are shown in black in the depth images. The middle image of the first row in (a) shows one of the control points (inside the red circle).
Localization results of experiment I with a ground truth length of 46.97 m.
| Calculated Length (m) | Error | |
|---|---|---|
| Our method | 45.82 | 2.45% |
| Traditional method | 44.99 | 4.22% |
Figure 7Overhead views of the 3D mapping results in Experiment I. The left figure is the result of the traditional method. The right is the result of the proposed method in this paper.
Localization results of experiment II with a total length of 183.5 m.
| Closure Error(m) | Error | |
|---|---|---|
| Our method | 4.56 | 2.48% |
| Traditional method | 7.05 | 3.84% |
Figure 8Estimated rover paths from the two BA methods. Red and blue curves represent the estimated trajectory using the traditional method and the proposed method, respectively.
Figure 9Overhead view of the mapping result of the whole scene in experiment II using our method. The two insets are detailed 3D views of the areas in the two rectangles.