| Literature DB >> 35626463 |
Yanzi Kong1,2,3,4, Feng Zhu1,2,3, Haibo Sun1,2,3,5, Zhiyuan Lin1,2,3,4, Qun Wang1,2,3,4.
Abstract
View planning (VP) is a technique that guides the adjustment of the sensor's postures in multi-view perception tasks. It converts the perception process into active perception, which improves the intelligence and reduces the resource consumption of the robot. We propose a generic VP system for multiple kinds of visual perception. The VP system is built on the basis of the formal description of the visual task, and the next best view is calculated by the system. When dealing with a given visual task, we can simply update its description as the input of the VP system, and obtain the defined best view in real time. Formal description of the perception task includes the task's status, the objects' prior information library, the visual representation status and the optimization goal. The task's status and the visual representation status are updated when data are received at a new view. If the task's status has not reached its goal, candidate views are sorted based on the updated visual representation status, and the next best view that can minimize the entropy of the model space is chosen as the output of the VP system. Experiments of view planning for 3D recognition and reconstruction tasks are conducted, and the result shows that our algorithm has good performance on different tasks.Entities:
Keywords: active perception; entropy reduction; information expression; next best view; view planning
Year: 2022 PMID: 35626463 PMCID: PMC9141229 DOI: 10.3390/e24050578
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.738
Figure 1Structure of the generic view planning system. Ability of sensors, objects’ prior knowledge, optimization goal of view planning and the expected status of a perception task are set before the perception process. With data receiving at the calculated views, real-time task status and the status of perception result are updated. View planning is based on these formal descriptions and the NBV is determined for the next iteration of data collection.
Figure 2Features of different classes of objects. The prism has features and , while the pyramid has features and . The rough shape of an instance can be defined by the parameters of its features.
Figure 3The update flow chart of each state. Boxes are basic description modules of our system and arrows represent the data streams when a new measurement is received.
Figure 4Positional relationship of local features. The green surfaces are detected by definite prior information, the pink one is predicted by strong prior information, and the yellow is predicted by weak prior information.
Figure 5Voxels’ status in the model space.
Figure 6Transformation between the existence states of a feature.
Figure 7Prior knowledge library of the objects for reconstruction.
Figure 8Prior knowledge library of the tanks for recognition.
Comparison of each model for reconstruction. T-S, GN and GNO achieve better performance.
| Models | m1 | m2 | m3 | m4 | m5 | m6 | m7 | m8 | m9 | m10 | m11 | m12 | m13 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
| OA | 8 | 20 | 20 | 8 | 8 | 8 | 5 | 9 | 9 | 20 | 10 | 8 | 17 |
| UV | 8 | 20 | 13 | 8 | 8 | 8 | 5 | 9 | 9 | 20 | 14 | 8 | 15 |
| RSV | 6 | 17 | 4 | 5 | 9 | 6 | 4 | 16 | 8 |
| 11 | 13 | 12 |
| RSE | 6 | 16 | 6 |
| 8 |
| 4 | 14 | 12 | 11 | 8 | 14 | 20 |
| PC | 5 | 9 | 5 | 5 |
| 5 | 3 | 6 |
| 8 | 5 | 16 | 13 |
| AF | 9 | 10 | 5 | 20 | 17 | 20 | 18 | 7 | 20 | 10 | 9 | 20 | 20 |
| AE |
| 8 | 7 | 6 | 6 | 8 | 5 | 6 | 8 | 10 | 7 | 14 | 15 |
| T-S | 5 |
| 4 | 5 |
|
| 5 | 6 |
|
|
| 8 | 10 |
| GN | 5 | 8 |
| 5 |
| 5 | 3 |
|
| 7 | 5 | 9 |
|
| GNO | 8 | 8 |
| 6 |
| 6 |
| 6 | 8 |
| 5 |
|
|
Figure 9Average number of views needed by each method in reconstruction experiments.
Figure 10Number of views needed in recognition experiments.