| Literature DB >> 33286094 |
Pingping Liu1,2,3, Lida Shi4, Zhuang Miao1, Baixin Jin1, Qiuzhan Zhou5.
Abstract
Convolutional neural networks (CEntities:
Keywords: Euclidean distance; deep metric learning; image retrieval; relative entropy
Year: 2020 PMID: 33286094 PMCID: PMC7516778 DOI: 10.3390/e22030321
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Each rectangle represents a feature descriptor obtained after the convolution of the neural network. Different colors of small squares represent different feature intensities in the descriptor. The Euclidean distance () between descriptor n and descriptor q is , while the Euclidean distance () between descriptor p and descriptor q is . D1 is approximately equal to D2, but the internal spatial distribution of p and n is obviously different.
Figure 2The core idea of relative distribution entropy (RDE)-loss. D represents the Euclidean distance between two image descriptors. RDE represents the relative distribution entropy between two image descriptors. RDE-distance is a new metric that combines Euclidean distance with relative distributed entropy. We called it the relative distribution entropy weighted distance, which can enhance the discrimination of image descriptors.
Figure 3Convolutional neural network (CNN) network architecture.
Figure 4Training process using relative distribution entropy contrastive loss.
Figure 5Training process using relative distribution entropy triplet loss.
Experimental results of hyperparameter comparison in the relative distribution entropy contrastive loss function. The best results would be highlighted in bold.
| Network |
|
| Oxford5k | Oxford5k(W) | Pairs6k | Pairs6k(W) |
|---|---|---|---|---|---|---|
| AlexNet | 0.50 | 10 | 58.10 | 67.60 | 71.64 | 79.60 |
| 0.75 | 20 | 60.87 | 67.19 | 75.33 | 79.43 | |
| 0.85 | 25 | 60.79 | 67.93 |
| 79.59 | |
|
|
|
|
| 75.29 |
| |
| 1.00 | 50 | 60.27 | 67.72 | 74.88 | 80.10 | |
| VGG | 0.85 | 30 | 84.62 | 87.83 | 82.40 | 88.01 |
| 0.85 | 100 | 76.18 | 83.04 | 81.71 | 87.11 | |
|
|
|
|
|
|
|
Comparative results of different pooling methods on AlexNet. The best results would be highlighted in bold.
| Net | Oxford5k | Oxford5k(W) | Pairs6k | Pairs6k(W) |
|---|---|---|---|---|
| SPoC [ | 41.83 | 55.34 | 55.49 | 68.61 |
| MAC [ | 47.50 | 55.95 | 62.16 | 71.30 |
| GeM [ |
|
|
|
|
Performance comparison of relative distribution entropy triplet loss and triplet loss. The best results would be highlighted in bold.
| Loss | Network | Oxford5k | Oxford5k(W) | Pairs6k | Pairs6k(W) |
|---|---|---|---|---|---|
| Triplet loss [ | VGG | 81.48 | 82.80 | 82.79 | 84.78 |
| Ours |
|
|
|
| |
| Triplet loss [ | ResNet | 81.49 | 85.33 | 87.70 | 91.11 |
| Ours |
|
|
|
|
Comparison of our method with the state-of-art image retrieval methods. The best results would be highlighted in bold.
| Net | Method | F-tuned | Oxford5k | Oxford105k | Pairs6k | Pairs106k |
|---|---|---|---|---|---|---|
| VGG | MAC [ | no | 56.4 | 47.8 | 72.3 | 58.0 |
| SPoC [ | no | 68.1 | 61.1 | 78.2 | 68.4 | |
| Crow [ | no | 70.8 | 65.3 | 79.7 | 72.2 | |
| R-MAC [ | no | 66.9 | 61.6 | 83.0 | 75.7 | |
| BoW-CNN [ | yes | 73.9 | 59.3 | 82.0 | 64.8 | |
| NetVLAD [ | yes | 71.6 | - | 79.7 | - | |
| Fisher [ | yes | 81.5 | 76.6 | 82.4 | - | |
| R-MAC [ | yes | 83.1 | 78.6 | 87.1 | 79.7 | |
| GeM [ | yes | 87.9 | 83.3 | 87.7 |
| |
| ours | yes |
|
|
| 79.9 | |
| Res | R-MAC [ | no | 69.4 | 63.7 | 85.2 | 77.8 |
| GeM [ | yes | 87.8 | 84.6 | 92.7 |
| |
| ours | yes |
|
|
| 86.3 | |
| Re-ranking(R) and Query Expansion(QE) | ||||||
| VGG | Crow + QE [ | no | 74.9 | 70.6 | 84.8 | 79.4 |
| R-MAC+R+QE [ | no | 77.3 | 73.2 | 86.5 | 79.8 | |
| BoW-CNN+R+QE [ | no | 78.8 | 65.1 | 84.8 | 64.1 | |
| R-MAC+QE [ | yes | 89.1 | 87.3 | 91.2 | 86.8 | |
| GeM+ | yes | 91.9 | 89.6 | 91.9 |
| |
| ours | yes |
|
|
| 87.3 | |
| Res | R-MAC+QE [ | no | 78.9 | 75.5 | 89.7 | 85.3 |
| R-MAC+QE [ | yes | 90.6 | 89.4 | 96.0 | 93.2 | |
| GeM+QE [ | yes | 91.0 | 89.5 | 95.5 | 91.9 | |
| ours | yes |
|
|
|
| |