| Literature DB >> 26690439 |
Jing Chen1, Ruochen Cao2, Yongtian Wang3,4.
Abstract
Wide-area registration in outdoor environments on mobile phones is a challenging task in mobile augmented reality fields. We present a sensor-aware large-scale outdoor augmented reality system for recognition and tracking on mobile phones. GPS and gravity information is used to improve the VLAD performance for recognition. A kind of sensor-aware VLAD algorithm, which is self-adaptive to different scale scenes, is utilized to recognize complex scenes. Considering vision-based registration algorithms are too fragile and tend to drift, data coming from inertial sensors and vision are fused together by an extended Kalman filter (EKF) to achieve considerable improvements in tracking stability and robustness. Experimental results show that our method greatly enhances the recognition rate and eliminates the tracking jitters.Entities:
Keywords: VLAD; mobile augmented reality; sensor fusion; sensor-aware scene recognition
Year: 2015 PMID: 26690439 PMCID: PMC4721767 DOI: 10.3390/s151229847
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Framework overview.
Figure 2An example of 3D reconstruction of scenes.
Figure 3Framework of GVLAD method.
Figure 4A novel methodology to introduce geo-based GVLAD.
Figure 5Relative coordinate systems.
Figure 6Examples of our database images.
Figure 7Geometry clustering result.
The recognition accuracy.
| Method | K = 64 | K = 128 | K = 256 |
|---|---|---|---|
| VLAD | 0.778 | 0.794 | 0.806 |
| Geo-based VLAD | 0.814 | 0.833 | 0.847 |
| GVLAD | 0.875 | 0.893 | 0.897 |
| Geo-based GVLAD | 0.922 | 0.933 | 0.934 |
Figure 8Recognition results.
Figure 9Effect of de-noising. (a) Static; (b) Random moving.
Figure 10Re-projection error of hybrid tricking method. (a) Rotate along Y-axis; (b) Move backwards and forwards.
Figure 11Motion estimation results when vision measurements are unavailable in some frames. (a) Rotation; (b) Translation.
Figure 12Tracking effects.
The average computation time.
| Step | Time (ms) | |
|---|---|---|
| Initialization phase | Feature Extraction | 88.4 |
| Feature Matching | 3.4 | |
| Tracking phase | Optical Flow Tracking | 17.1 |
| PROSAC | 2.5 | |
| Pose Estimation | 6.4 | |
| Sensor Fusion (Prediction) | 0.5 | |
| Sensor Fusion (Correction) | 1.4 | |
| Render latency | 0.5 | |