| Literature DB >> 31687438 |
Giulio Marin1, Gianluca Agresti1, Ludovico Minto1, Pietro Zanuttigh1.
Abstract
Time-of-Flight (ToF) sensors and stereo vision systems are two of the most diffused depth acquisition devices for commercial and industrial applications. They share complementary strengths and weaknesses. For this reason, the combination of data acquired from these devices can improve the final depth estimation accuracy. This paper introduces a dataset acquired with a multi-camera system composed by a Microsoft Kinect v2 ToF sensor, an Intel RealSense R200 active stereo sensor and a Stereolabs ZED passive stereo camera system. The acquired scenes include indoor settings with different external lighting conditions. The depth ground truth has been acquired for each scene of the dataset using a line laser. The data can be used for developing fusion and denoising algorithms for depth estimation and test with different lighting conditions. A subset of the data has already been used for the experimental evaluation of the work "Stereo and ToF Data Fusion by Learning from Synthetic Data".Entities:
Keywords: Active stereo; Data fusion; Depth estimation; Stereo vision; Time-of-Flight
Year: 2019 PMID: 31687438 PMCID: PMC6820107 DOI: 10.1016/j.dib.2019.104619
Source DB: PubMed Journal: Data Brief ISSN: 2352-3409
Fig. 1Multi-sensors arrangement.
Fig. 2Images of the same checkerboard acquired from the three sensors during the calibration process.
Fig. 3Color and IR images acquired by the RealSense R200 on the considered 10 scenes.
Fig. 4Data acquired with the considered acquisition system. The left and the right views are provided for the ZED system. The Kinect v2 data are the RGB image, the depth map and the amplitude image related to the ToF acquisition. The RealSense R200 data are the RGB image and the left and the right IR views of its stereo system.
Fig. 5RGB and IR images acquired respectively with the color and left IR cameras of the RealSense R200 device. These recordings are acquired under 4 different external illumination conditions. L1 stands for “no external light”; L2 stands for “regular lighting”; L3 stands for “stronger light”; L4 stands for “use of an additional incandescent light source”.
Specifications Table
| Subject area | |
| More specific subject area | |
| Type of data | |
| How data was acquired | |
| Data format | |
| Experimental factors | |
| Experimental features | |
| Data source location | |
| Data accessibility | |
| Related research article | Agresti, G., Minto, L., Marin, G., & Zanuttigh, P. (2019). Stereo and ToF Data Fusion by Learning from Synthetic Data. |
Multiple acquisitions of the same scene with different sensors are provided: we provide 10 real world static scenes acquired with an active IR stereo camera, a passive stereo camera and a ToF sensor. Each scene comes with 4 different lighting conditions allowing to evaluate how the illumination affects depth estimation algorithms. Additionally, the ToF amplitude and color data from the depth sensors are also provided allowing to exploit them in depth refinement algorithms. Accurate depth ground truth, acquired with a line laser, is provided for all the scenes (only a very few datasets have ToF depth with ground truth, this allows to evaluate the algorithms and machine learning based approaches). Accurate calibration information allows to reproject the data to a common reference system, a fundamental step for being able to test data fusion strategies. The data can be used by researchers working on both depth estimation and data fusion. Part of the data has been used in previous research works, allowing to perform experimental comparison. Suitable for testing fusion and denoising algorithms for depth estimation in different lighting conditions. |