| Literature DB >> 26230701 |
Thuy Tuong Nguyen1, David C Slaughter2, Nelson Max3, Julin N Maloof4, Neelima Sinha5.
Abstract
Camera-based 3D reconstruction of physical objects is one of the most popular computer vision trends in recent years. Many systems have been built to model different real-world subjects, but there is lack of a completely robust system for plants. This paper presents a full 3D reconstruction system that incorporates both hardware structures (including the proposed structured light system to enhance textures on object surfaces) and software algorithms (including the proposed 3D point cloud registration and plant feature measurement). This paper demonstrates the ability to produce 3D models of whole plants created from multiple pairs of stereo images taken at different viewing angles, without the need to destructively cut away any parts of a plant. The ability to accurately predict phenotyping features, such as the number of leaves, plant height, leaf size and internode distances, is also demonstrated. Experimental results show that, for plants having a range of leaf sizes and a distance between leaves appropriate for the hardware design, the algorithms successfully predict phenotyping features in the target crops, with a recall of 0.97 and a precision of 0.89 for leaf detection and less than a 13-mm error for plant size, leaf size and internode distance.Entities:
Keywords: 3D feature extraction; 3D reconstruction; plant phenotyping; point cloud; stereo vision; structured light
Mesh:
Substances:
Year: 2015 PMID: 26230701 PMCID: PMC4570338 DOI: 10.3390/s150818587
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Mechanical structure (a) of the 3D reconstruction system: (b) the arc holding ten Canon EOS Rebel T3 cameras; (c) a pair of cameras where the second camera is upside down relative to the first one; (d) structured light devices and their power adapters with a relay controlled by a digital I/O control NI USB-6008; (e) a turn-table that rotates the plant 360 degrees; and (f) the target plant.
Figure 2Random-dot pattern projected onto surfaces and plants: (a) the dot pattern printed on a transparency film (in this figure, it is under a transparent window); (b) an image of the pattern projected (with maximum brightness) on a piece of white paper; (c) an image of the pattern projected on a black curtain; (d) the pattern projected on a cabbage plant; (e) cucumber plant; and (f) a compound leaf from a big tomato plant.
Figure 3Block diagram representation of the software algorithm.
Figure 4Segmentation of a cabbage plant and its soil from the background. The plant (a) without and (b) with using structured light; segmentation results of (c) without and (d) with using structured light; note that (c,d) are the results of binary segmentation in which the plant and its soil are necessarily isolated from the black background; and (c) indicates segmentation failure.
Figure 5Results of the block matching (BM) (first row) and belief propagation (BP) (second row) algorithms with the use of disparity bilateral filtering (DBF) and structured light (SL): (a) BM result without SL; (b) BM + DBF result without SL; (c) BM result with SL; (d) BM + DBF result with SL; (e) BP result without structured light (SL); (f) BP + DBF result without SL; (g) BP result with SL; and (h) BP + DBF result with SL. The BM + DBF algorithm with SL is used in our system.
Figure 6Estimation of the depth resolution of our stereo system as a function of the target distance from 1–1.82 m away from the camera: (a) Plot of disparities versus actual distances and (b) Plot of depth resolutions versus actual distances.
Figure 7Point clouds from different view angles (top row) and their merged point cloud and surface reconstruction results (bottom row).
Figure 83D model of a cucumber plant (a) and its leaf detection result (b), where each part of the plant is color-coded and the bounding boxes of leaves are shown.
Figure 9Illustration of internode distance estimation, where leaf centers are projected onto the plant’s principal axis. (The original internode image is from http://pixgood.com/internode-plant.html).
Figure 103D model results of eight cabbage plants (a–h), eight cucumber plants (i–p) and three compound leaves from tomato plants (q–s).
List of the 19 plants used for the experiments.
| Plant | Height | No. of Leaves | Brief Description | |
|---|---|---|---|---|
| Cabbage 1 | (a) | 114 | 4 | Good leaf shape |
| Cabbage 2 | (b) | 150 | 4 | 1 vertically-long leaf |
| Cabbage 3 | (c) | 140 | 4 | 1 small and 2 curved leaves |
| Cabbage 4 | (d) | 114 | 4 | Long branches |
| Cabbage 5 | (e) | 130 | 4 | 2 overlapped leaves |
| Cabbage 6 | (f) | 139 | 3 | Long and thin branches |
| Cabbage 7 | (g) | 105 | 3 | 1 leaf attaches to plant stem |
| Cabbage 8 | (h) | 229 | 2 | 1 curved leaf |
| Cucumber 1 | (i) | 242 | 3 | Tall, big leaves |
| Cucumber 2 | (j) | 117 | 4 | 2 brown-textured-surface leaves |
| Cucumber 3 | (k) | 131 | 3 | 2 brown-textured-surface leaves |
| Cucumber 4 | (l) | 115 | 2 | 1 small leaf |
| Cucumber 5 | (m) | 113 | 1 | Good leaf shape |
| Cucumber 6 | (n) | 123 | 2 | 1 small leaf |
| Cucumber 7 | (o) | 132 | 2 | 1 leaf attaches to plant stem |
| Cucumber 8 | (p) | 116 | 2 | 1 yellow-textured-surface leaf |
| Tomato 1 | (q) | 192 * | 6 | Long and curved leaves |
| Tomato 2 | (r) | 253 * | 8 | Long and curved leaves |
| Tomato 3 | (s) | 269 * | 8 | Long and curved leaves |
+ Unit in mm; * length of the compound leaf.
Algorithm parameters used for the experiments.
| GPU-Based Stereo Matching | Point Cloud Registration | 3D Feature Extraction * | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Plant segmentation | Spatial win-size | 11 | Registration | Max distance | 25 | Clustering | Tolerance | 0.03 | ||
| Color win-size | 7 | Max iteration | 10 | Min cluster size | 4000 | |||||
| Min segment size | 10 | Outlier rejection | 25 | Max cluster size | 10 | |||||
| Threshold | 10 | Poisson surface reconstruction | Octree depth | 12 | Leaf detection | Size threshold | 0.005 | |||
| Stereo block matching | No. of disparities | 256 | Solver divide | 7 | Direction threshold | 0.7 | ||||
| Win-size | 17 | Samples/node | 1 | Ratio: leaf location | 0.25 | |||||
| Texture threshold | 60 | Surface offset | 1 | * Parameters vary depending on leaf shape | ||||||
| Bilateral filter | Filter size | 41 | Face removal w.r.t. edge length | 0.05 | ||||||
| No. of iterations | 10 | Noise removal w.r.t. No. of faces | 25 | |||||||
Average accuracy in plant phenotype estimation from 3D reconstruction.
| Plant Features | Cabbage | Cucumber | Tomato | Average | |
|---|---|---|---|---|---|
| Leaf height | Error (mm) | 6.86 | 5.08 | 10.16 | 6.6 |
| % error * | 5.58% | 4.36% | 4.36% | 4.87% | |
| Leaf width | Error (mm) | 5.08 | 4.83 | 5.33 | 5.08 |
| % error * | 4.16% | 3.9% | 2.31% | 3.76% | |
| Internode distance | Error (mm) | 9.65 | 7.87 | 21.34 | 10.92 |
| % error * | 7.67% | 6.3% | 8.49% | 7.28% | |
* Percentage of error over plant height.
Figure 11Evaluation of the number of detected leaves/leaflets in terms of precision and recall, from the 3D reconstructed cabbage, cucumber and tomato plants.
Figure 12Errors of plant height (calculated by differentiating the ground truth from the estimated one) of the cabbage (a) and cucumber plants (b). Notice that plant height was not determined for tomato compound leaves, because these leaves were imaged individually and are parts of a plant.
Figure 13Comparison between without and with using structured light (SL) on textured leaves. The disparity result of using SL is slightly better than that without using SL in the regions of less natural textures (marked by red rectangles and ellipses). Note that the colorized disparity images are presented here, instead of grayscale ones, for better illustration of the differences.
Figure 14Comparison of the percentage of error in plant height estimation for plants having different leaf sizes, leaf shapes and numbers of leaves. Errors for leaf length, leaf width and internode distance were considered in order to understand which types of plant leaves yield higher errors. From left to right, top to bottom: plants having big leaves versus plants having small leaves, plants having curved versus flat leaves, plants having many versus fewer leaves and plants having long-shaped versus round-shaped leaves.
Comparison of various camera-based 3D reconstruction systems for plants.
| Study | Camera System | Camera View | Measures | Environment | Techniques | Accuracy | Processing Time |
|---|---|---|---|---|---|---|---|
| Alenya, 2011 [ | ToF and color cameras; robot arm | Multiview for leaf modeling | Leaf size | Indoor | Depth-aided color segmentation, quadratic surface fitting, leaf localization | Square fitting error: 2 cm2 | 1 min for 3D leaf segmentation |
| Chene, 2012 [ | Kinect camera | Top view | Leaf azimuth | Indoor | Maximally stable extremal regions-based leaf segmentation | Detection accuracy 68%; azimuth error 5% | n/a |
| Heijden, 2012 [ | ToF and color cameras | Single view | Leaf size and angle | Greenhouse (for large pepper plants) | Edge-based leaf detection, locally weighted scatterplot smoothing-based surface reconstruction | Leaf height correlation 0.93; leaf area correlation 0.83 | 3 min for image recording, hours for the whole process |
| Paproki, 2012 [ | High-resolution SLR camera, with 3D modeling software [ | Multiview for full 3D reconstruction | Plant height, leaf width and length | Indoor | Constrained region growing, tubular shape fitting-based stem segmentation, planar-symmetry and normal clustering-based leaf segmentation, pair-wise matching-based temporal analysis | Plant height error 9.34%; leaf width error 5.75%; leaf length error 8.78% | 15 min for 3D reconstruction [ |
| Azzari, 2013 [ | Kinect camera | Top view | Plant height, base diameter | Outdoor | Canopy structure extraction | Correlation: 0.97 | n/a |
| Ni, 2014 [ | 2 low-resolution stereo cameras with 3D modeling software [ | Multiview for full 3D reconstruction | Plant height and volume, leaf area | Indoor | Utilizing VisualSFM software [ | n/a | n/a |
| Song, 2014 [ | Two stereo cameras; ToF camera | Single view | Leaf area (foreground leaves only) | Greenhouse (for large plants) | Dense stereo with localized search, edge-based leaf detection, locally weighted scatterplot smoothing-based surface reconstruction | Error: 9.3% | 5 min for the whole process |
| Polder, 2014 [ | 3D light-field camera | Single view | Leaf and fruit detection | Greenhouse (for large tomato plants) | Utilizing 3D light-field camera to output a pixel to pixel registered color image and depth map | n/a | n/a |
| Rose, 2015 [ | High-resolution SLR camera, with 3D modeling software [ | Multiview for full 3D reconstruction | Plant height, leaf area, convex hull | Indoor | Utilizing Pix4Dmapper software [ | Correlation: 0.96 | 3 min for data acquisition, 20 min for point cloud generation, 5 min for manual scaling, 10 min for error removal |
| Andujar, 2015 [ | 4 Kinect cameras with 3D modeling software [ | Multiview for semi-full 3D reconstruction | Plant height, leaf area, biomass | Outdoor | Utilizing Skanect software [ | Correlation: plant height 0.99, leaf area 0.92, biomass 0.88 | n/a |
| Our system | 10 high-resolution SLR cameras organized into 5 stereo pairs; 2 structured lights | Multiview for full 3D reconstruction | Plant height, leaf width and length, internode distance | Indoor | Texture creation using structured lights, mean shift-based plant segmentation, stereo block matching, disparity bilateral filtering, ICP-based point cloud registration, Poisson surface reconstruction, plant feature extraction | Leaf detection accuracy 97%; plant height error 8.1%, leaf width error 3.76%, leaf length error 4.87%, internode distance error 7.28% | 4 min for the whole process |