Literature DB >> 31406905

KFuji RGB-DS database: Fuji apple multi-modal images for fruit detection with color, depth and range-corrected IR data.

Jordi Gené-Mola1, Verónica Vilaplana2, Joan R Rosell-Polo1, Josep-Ramon Morros2, Javier Ruiz-Hidalgo2, Eduard Gregorio1.   

Abstract

This article contains data related to the research article entitle "Multi-modal Deep Learning for Fruit Detection Using RGB-D Cameras and their Radiometric Capabilities" [1]. The development of reliable fruit detection and localization systems is essential for future sustainable agronomic management of high-value crops. RGB-D sensors have shown potential for fruit detection and localization since they provide 3D information with color data. However, the lack of substantial datasets is a barrier for exploiting the use of these sensors. This article presents the KFuji RGB-DS database which is composed by 967 multi-modal images of Fuji apples on trees captured using Microsoft Kinect v2 (Microsoft, Redmond, WA, USA). Each image contains information from 3 different modalities: color (RGB), depth (D) and range corrected IR intensity (S). Ground truth fruit locations were manually annotated, labeling a total of 12,839 apples in all the dataset. The current dataset is publicly available at http://www.grap.udl.cat/publicacions/datasets.html.

Entities:  

Keywords:  Depth cameras; Fruit detection; Fruit reflectance; Fuji apple; Multi-modal dataset; RGB-D

Year:  2019        PMID: 31406905      PMCID: PMC6685673          DOI: 10.1016/j.dib.2019.104289

Source DB:  PubMed          Journal:  Data Brief        ISSN: 2352-3409


Specifications table First dataset for fruit detection that contains 3 different modalities: color, depth and range corrected IR intensity. The presented dataset could be used in the development and training of fruit detection systems with applications in yield prediction, yield mapping and automated harvesting. Compilation of this database allows fusing RGB-D and radiometric information obtained with Kinect v2 for fruit detection.

Data

The KFuji RGB-DS database contains a total of 967 multi-modal images of Fuji apples on trees and the corresponding ground truth fruit location annotations. Each image contains data from three different modalities: color (RGB), depth (D), and range-corrected IR intensity (S). Fig. 1 illustrates three selected images from de dataset, showing ground truth annotations and the modalities that composes each image.
Fig. 1

Selection of 3 multi-modal images and the corresponding ground truth fruit locations (red bounding boxes). Each image column corresponds to a different image modality: RGB, S and D, respectively.

Selection of 3 multi-modal images and the corresponding ground truth fruit locations (red bounding boxes). Each image column corresponds to a different image modality: RGB, S and D, respectively. This dataset was built to be used for training, validation and benchmarking of fruit detection algorithms using RGB-D sensors. For instance, in Ref. [1], the deep convolutional neural network Faster R-CNN [2] was used to detect and localize fruits from the presented dataset. Images are 548 × 373px and were saved in three different files: RGBhr (high resolution color image): Raw color image. These images are saved in 8-bit JPG files. RGBp (projected color image): Projection of the color 3D point cloud onto the camera focal plane. The RGBp and the D-S modalities are obtained following the same procedure, allowing the comparison between these modalities for fruit detection. These images are saved in 8-bit JPG files. DS (depth and range-corrected IR image): Projection of the range-corrected IR 3D point cloud onto the camera focal plane. The D channel corresponds to the depth values, while the S channel corresponds to the range-corrected IR intensity values. These modalities are saved in a unique 64-bit MAT file. S and D data were normalized between 0 and 255 –like RGB images-to achieve similar mean and variance between channels. This normalization allows a faster learning convergence of machine learning algorithms (such as deep convolutional neural networks). All images were manually annotated with rectangular bounding boxes, labelling a total of 12,839 apples in all the dataset. Annotations are provided in XLM and CSV formats, where each row corresponds to an apple annotation, giving the following information: item, topleft-x, topleft-y, width, height, label id.

Experimental design, materials, and methods

The data acquisition was carried out in a commercial Fuji apple orchard (Malus domestica Borkh. cv. Fuji), three weeks before harvesting (85 BBCH growth stage [3]). The RGB-D sensors used were two Microsoft Kinect v2 (Microsoft, Redmond, WA, USA), which are composed by an RGB camera and a time-of-flight (ToF) depth sensor. For each capture, the sensor provides a 3D point cloud with RGB and backscattered IR intensity data, and a raw RGB image. Due to the performance of the depth sensor drops under direct sunlight exposure [4], data was acquired at night using artificial lighting. Pre-processing of data was carried out to build the multi-modal images with pixel-wise correspondence between channels. Fig. 2 shows an outline of the data preparation steps. To overcome the IR signal attenuation, the IR intensity data was range-corrected (Fig. 2a) following the methodology described in Ref. [1]. Then the acquired 3D point clouds were projected onto the camera focal plane (Fig. 2b), generating the RGB, range-corrected IR and depth projected images. These images were geometrically wrapped and registered (Fig. 2c) with RGBhr so that different image modalities have pixel-wise correspondence. Finally, to reduce the number of fruits per image, and considering that fruit size is small compared with the image size, each capture was split into 9 images of 548 × 373 px (Fig. 2d).
Fig. 2

Data preparation outline.

Data preparation outline.

Specifications table

Subject areaMachine learning, computer vision, deep learning, agronomy
More specific subject areaImage fusion, Precision agriculture.
Type of dataMulti-modal images with color (RGB), depth (D), and range-corrected IR intensity (S).
How data was acquiredThe images were acquired using Microsoft Kinect v2.
Data formatRaw images: JPGRaw point clouds: MATPre-processed images: JPG (color channels) and MAT (depth and range-corrected IR channels)Annotations: CSV and XLM.
Experimental factorsDifferent image modalities have been registered to have pixel-wise correspondence between image channels.
Experimental featuresAll captures were carried out during the night, using artificial lighting.
Data source locationData were acquired in Tarassó Farm, a commercial apple field located in Agramunt, Catalonia, Spain (E: 336297 m N: 4623494 m 31 N 312 m a.s.l., UTM31T - ETRS89).
Data accessibilityhttp://www.grap.udl.cat/publicacions/datasets.html
Related research articleGené-Mola J, Vilaplana V, Rosell-Polo J.R, Morros J.R, Ruiz-Hidalgo J, Gregorio E. Multi-modal Deep Learning for Fruit Detections Using RGB-D Cameras and their Radiometric Capabilites. Computers and Electronics in Agriculture (2018) 162, 689–698.https://doi.org/10.1016/j.compag.2019.05.016[1]
Value of the data

First dataset for fruit detection that contains 3 different modalities: color, depth and range corrected IR intensity.

The presented dataset could be used in the development and training of fruit detection systems with applications in yield prediction, yield mapping and automated harvesting.

Compilation of this database allows fusing RGB-D and radiometric information obtained with Kinect v2 for fruit detection.

  1 in total

1.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.

Authors:  Shaoqing Ren; Kaiming He; Ross Girshick; Jian Sun
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-06-06       Impact factor: 6.226

  1 in total
  2 in total

1.  Toward Joint Acquisition-Annotation of Images with Egocentric Devices for a Lower-Cost Machine Learning Application to Apple Detection.

Authors:  Salma Samiei; Pejman Rasti; Paul Richard; Gilles Galopin; David Rousseau
Journal:  Sensors (Basel)       Date:  2020-07-27       Impact factor: 3.576

2.  Augmentation Method for High Intra-Class Variation Data in Apple Detection.

Authors:  Huibin Li; Wei Guo; Guowen Lu; Yun Shi
Journal:  Sensors (Basel)       Date:  2022-08-23       Impact factor: 3.847

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.