| Literature DB >> 27843754 |
Mohammad O A Aqel1, Mohammad H Marhaban2, M Iqbal Saripan3, Napsiah Bt Ismail4.
Abstract
Accurate localization of a vehicle is a fundamental challenge and one of the most important tasks of mobile robots. For autonomous navigation, motion tracking, and obstacle detection and avoidance, a robot must maintain knowledge of its position over time. Vision-based odometry is a robust technique utilized for this purpose. It allows a vehicle to localize itself robustly by using only a stream of images captured by a camera attached to the vehicle. This paper presents a review of state-of-the-art visual odometry (VO) and its types, approaches, applications, and challenges. VO is compared with the most common localization sensors and techniques, such as inertial navigation systems, global positioning systems, and laser sensors. Several areas for future research are also highlighted.Entities:
Keywords: Global positioning system; Image stream; Inertial navigation system; Localization sensors; Visual odometry
Year: 2016 PMID: 27843754 PMCID: PMC5084145 DOI: 10.1186/s40064-016-3573-7
Source DB: PubMed Journal: Springerplus ISSN: 2193-1801
Fig. 1Visual odometry [Aqel et al. 2016]
Fig. 2Wheel odometry with an optical encoder [Pololu Corporation 2016]
Fig. 3Inertial navigation system. a Block diagram of INS. b Miniature INS [SBG Systems 2016]
Fig. 4Global positioning system [Aboelmagd et al. 2013]. a GPS satellite constellation. b Concept of positioning by trilateration (red dot represents user’s position)
Fig. 5Real-time differential global positioning system
Comparison of commonly used localization sensors
| Sensor/technology | Advantages | Disadvantages |
|---|---|---|
| Wheel odometry | Simple to determine position/orientation | Position drift due to wheel slippage |
| INS | Provides both position and orientation using 3-axis accelerometer and gyroscope | Position drift (position estimation requires second-order integral) |
| GPS/GNSS | Provides absolute position with known value of error | Unavailable in indoor, underwater, and closed areas |
| Ultrasonic sensor | Provides a scalar distance measurement from sensor to object | Reflection of signal wave is dependent on material or orientation of obstacle surface |
| Laser sensor | Similar to sonar sensors but has higher accuracy and scan rate | Reflection of signal wave is dependent on material or orientation of obstacle surface |
| Optical camera | Images store a huge meaningful information | Requires image-processing and data-extraction techniques |
Fig. 6Different types of camera used in VO systems. a Stereo camera [courtesy of VOLTRIUM]. b Stereo omnidirectional [courtesy of Occam]. c Monocular camera [courtesy of Microsoft]. d Monocular omnidirectional [courtesy of Occam]
Comparison between types of cameras used for VO
| Type of VO camera | Pros | Cons |
|---|---|---|
| Monocular | Low cost and easy deployment | Suffer from image scale uncertainty |
| Stereo | Image scale and depth information is easy to be retrieved | More expensive and needs more calibration effort than monocular cameras |
| Omnidirectional | Provides very wide field of vision (FOV) (up to 360° FOV) | Complex system |
Fig. 7Flowchart of visual odometry system algorithm
Summary table of some VO works in literature
| Reference | Camera type | Approach | VO estimation accuracy | Limitations |
|---|---|---|---|---|
| Gonzalez et al. ( | Two monocular cameras: downward-facing camera for displacement and front-facing camera for orientation estimation | Appearance-based approach (NCC template matching) | Error <3% of the total travelling distance and <8° average orientation error | False matches due to shadows and blur at velocity >1.5 m/s |
| Van Hamme et al. ( | Monocular camera | Feature-based approach (inverse perspective projection and Kalman filter for Tracking of features in the ground plane) | >8.5% translation error (for 800 m) | Significant rotational bias on some estimated trajectory segments due to non-planarity of the road environment in those segments |
| Scaramuzza and Siegwart ( | Omnidirectional camera | Hybrid approach (tracking SIFT feature points from ground plane to estimate translation. Image appearance similarity measure (NCC, Manhattan and Euclidean distance) was used to estimate the rotation of the car) | Error is <2% of the distance traversed | Unavoidable visual odometry drift and deviation due to road humps that violate the planar motion assumption |
| Nistér et al. ( | Stereo camera | Feature-based approach (Detection of features independently in all frames and only allowed matches between salient features) | 1.63% error over 380 m of the distance traversed | No mention |
| Howard ( | Stereo camera | Feature-based approach (Feature matching and employing stereo range data for inlier detection) | 0.25% error over 400 m of the distance traversed | Self-Shadow leads to false-matches |
| Nourani-Vatani and Borges ( | Monocular camera | Appearance-based approach (NCC multi-template matching which selects best template based on entropy) | Error <5% of total travelling distance | Deficiency in dealing with scale variance at uneven surfaces |
| Yu et al. ( | Monocular camera | Appearance-based approach (NCC rotated template matching) | 1.38% distance error and 2.8° heading error | Cannot deal with image scale variance, shadows and blur |
| Nagatani et al. ( | Telecentric camera (which maintains the same field of ground area view, regardless of variation in camera height from ground | Appearance-based approach (cross correlation template matching) | <3% error indoor experiment | Cannot estimate the camera height from ground variations |
| Zhang et al. ( | Monocular camera | Feature-based approach [tracking of features using Lucas Kanade Tomasi (LKT)] | Error is <1% of the distance traversed | Image scale uncertainty at complicated ground conditions for example loose soil floors |