| Literature DB >> 34945867 |
Sheng Zeng1, Guohua Geng1, Mingquan Zhou1.
Abstract
Automatically selecting a set of representative views of a 3D virtual cultural relic is crucial for constructing wisdom museums. There is no consensus regarding the definition of a good view in computer graphics; the same is true of multiple views. View-based methods play an important role in the field of 3D shape retrieval and classification. However, it is still difficult to select views that not only conform to subjective human preferences but also have a good feature description. In this study, we define two novel measures based on information entropy, named depth variation entropy and depth distribution entropy. These measures were used to determine the amount of information about the depth swings and different depth quantities of each view. Firstly, a canonical pose 3D cultural relic was generated using principal component analysis. A set of depth maps obtained by orthographic cameras was then captured on the dense vertices of a geodesic unit-sphere by subdividing the regular unit-octahedron. Afterwards, the two measures were calculated separately on the depth maps gained from the vertices and the results on each one-eighth sphere form a group. The views with maximum entropy of depth variation and depth distribution were selected, and further scattered viewpoints were selected. Finally, the threshold word histogram derived from the vector quantization of salient local descriptors on the selected depth maps represented the 3D cultural relic. The viewpoints obtained by the proposed method coincided with an arbitrary pose of the 3D model. The latter eliminated the steps of manually adjusting the model's pose and provided acceptable display views for people. In addition, it was verified on several datasets that the proposed method, which uses the Bag-of-Words mechanism and a deep convolution neural network, also has good performance regarding retrieval and classification when dealing with only four views.Entities:
Keywords: Bag-of-Words; cultural relic; information entropy; viewpoint selection
Year: 2021 PMID: 34945867 PMCID: PMC8700342 DOI: 10.3390/e23121561
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Representative view of the horse.
Figure 2In the left (a), 258 cameras are placed on the vertices of the geodesic unit sphere to obtain the depth map for the cow model. The middle (b) and right (c) are the views with maximum depth variation entropy and depth distribution entropy, respectively.
Figure 3The whole pipeline of Representative View Selection of Jishou. Initial selected views from each group are shown in (a). The viewpoints in the order of marks 1, 2, 3 and 4 in (b) were obtained according to the scattered viewpoint selection rules. The final four selected views are shown in (c).
Figure 4Generation of the threshold word histogram.
Figure 5Comparison of optimal view methods and scattered view selection results.
Comparing 7 shape descriptors on the PSB test set with base classification.
| CM-BOF | LFD | Ours | REXT | SHD | GEDT | I2 | D2 | |
|---|---|---|---|---|---|---|---|---|
| 1-NN (%) | 73.1 | 65.7 | 65.3 | 60.2 | 55.6 | 60.3 | 39.4 | 31.1 |
| 1-Tier (%) | 47.0 | 38.0 | 36.0 | 32.7 | 30.9 | 31.3 | 20.8 | 15.8 |
| 2-Tier (%) | 59.8 | 48.7 | 46.9 | 43.2 | 41.1 | 40.7 | 27.9 | 23.5 |
| DCG (%) | 72.0 | 64.3 | 56.0 | 60.1 | 58.4 | 23.7 | 45.3 | 43.4 |
Figure 6Precision-recall curves of four different views.
Comparison of different data sets under four view selections.
| McGill (10 Class) | McGill (8 Class) | |||||||
|---|---|---|---|---|---|---|---|---|
| Number of views | 1 | 4 | 6 | 18 | 1 | 4 | 6 | 18 |
| 1-Tier (%) | 68.9 | 70.7 | 71.8 | 73.0 | 74.1 | 78.6 | 78.6 | 80.5 |
| 2-Tier (%) | 83.2 | 83.8 | 84.2 | 84.7 | 88.7 | 91.5 | 91.4 | 92.3 |
| DCG (%) | 83.4 | 83.9 | 84.3 | 84.7 | 85.8 | 87.1 | 87.0 | 87.5 |
Comparison of classification accuracy on different views using MVCNN.
| Number of Views | 4 | 8 | 12 |
|---|---|---|---|
| Classification (Overall accuracy) | 85.1% | 86.3% | 92.5% |
| Classification (Mean accuracy) | 81.7% | 83.1% | 88.9% |