| Literature DB >> 35401137 |
Ping Jiang1, Junji Oaki1, Yoshiyuki Ishihara1, Junichiro Ooga1, Haifeng Han1, Atsushi Sugahara1, Seiji Tokura1, Haruna Eto1, Kazuma Komoda1, Akihito Ogawa1.
Abstract
Deep learning has been widely used for inferring robust grasps. Although human-labeled RGB-D datasets were initially used to learn grasp configurations, preparation of this kind of large dataset is expensive. To address this problem, images were generated by a physical simulator, and a physically inspired model (e.g., a contact model between a suction vacuum cup and object) was used as a grasp quality evaluation metric to annotate the synthesized images. However, this kind of contact model is complicated and requires parameter identification by experiments to ensure real world performance. In addition, previous studies have not considered manipulator reachability such as when a grasp configuration with high grasp quality is unable to reach the target due to collisions or the physical limitations of the robot. In this study, we propose an intuitive geometric analytic-based grasp quality evaluation metric. We further incorporate a reachability evaluation metric. We annotate the pixel-wise grasp quality and reachability by the proposed evaluation metric on synthesized images in a simulator to train an auto-encoder-decoder called suction graspability U-Net++ (SG-U-Net++). Experiment results show that our intuitive grasp quality evaluation metric is competitive with a physically-inspired metric. Learning the reachability helps to reduce motion planning computation time by removing obviously unreachable candidates. The system achieves an overall picking speed of 560 PPH (pieces per hour).Entities:
Keywords: bin picking; deep learning; grasp planning; graspability; suction grasp
Year: 2022 PMID: 35401137 PMCID: PMC8987443 DOI: 10.3389/fnbot.2022.806898
Source DB: PubMed Journal: Front Neurorobot ISSN: 1662-5218 Impact factor: 2.650
Figure 1Problem statement: (A) Picking robot; (B) Suction hand; (C) Grasp pose.
Figure 2System diagram.
Figure 3Data generation pipeline: (A) Dataset generation flow; (B) Cluttered scene generation; (C) Graspable surface detection; (D) Grasp quality evaluation; (E) Robot reachability evaluation.
Figure 4The architecture of SG-U-Net++.
Figure 5Experiment object set.
Inference precision.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| Grasp quality | Dex-Net | 91.9 | 91.0 | 88.7 | 84.2 |
| SQ-U-Net++ | 99.8 | 99.6 | 99.2 | 97.5 | |
| Reachability | SQ-U-Net++ | 95.8 | 91.1 | 80.7 | 61.2 |
Experiment results.
|
|
|
|
|
|---|---|---|---|
|
|
|
| |
| Dex-Net 4.0 Suction | 91.5 | 0.60 | 2.91 |
| (FC-GQCNN-4.0-SUCTION) | |||
| SQ-U-Net++ Policy1 | 94.6 |
| 1.71 |
| (grasp quality only) | |||
| SQ-U-Net++ Policy2 |
| 0.17 |
|
| (grasp quality+reachability) |
Bold values indicates the best performance among three methods in the Table. For the success rate, the higher the better. For the cost (computation time) of grasp planning and motion planning, the shorter the better.
Figure 6Example of Dex-Net grasp prediction that is farther from the center of mass of the object.
Figure 7Example of grasps predicted by Dex-Net and Policy 1 that are unreachable or difficult to reach.