| Literature DB >> 34945976 |
Xiaojun Lu1, Libo Zhang1, Lei Niu1, Qing Chen1, Jianping Wang1.
Abstract
In the era of big data, it is challenging to efficiently retrieve the required images from the vast amount of data. Therefore, a content-based image retrieval system is an important research direction to address this problem. Furthermore, a multi-feature-based image retrieval system can compensate for the shortage of a single feature to a certain extent, which is essential for improving retrieval system performance. Feature selection and feature fusion strategies are critical in the study of multi-feature fusion image retrieval. This paper proposes a multi-feature fusion image retrieval strategy with adaptive features based on information entropy theory. Firstly, we extract the image features, construct the distance function to calculate the similarity using the information entropy proposed in this paper, and obtain the initial retrieval results. Then, we obtain the precision of single feature retrieval based on the correlation feedback as the retrieval trust and use the retrieval trust to select the effective features automatically. After that, we initialize the weights of selected features using the average weights, construct the probability transfer matrix, and use the PageRank algorithm to update the initialized feature weights to obtain the final weights. Finally, we calculate the comprehensive similarity based on the final weights and output the detection results. This has two advantages: (1) the proposed strategy uses multiple features for image retrieval, which has better performance and more substantial generalization than the retrieval strategy based on a single feature; (2) compared with the fixed-feature retrieval strategy, our method selects the best features for fusion in each query, which takes full advantages of each feature. The experimental results show that our proposed method outperforms other methods. In the datasets of Corel1k, UC Merced Land-Use, and RSSCN7, the top10 retrieval precision is 99.55%, 88.02%, and 88.28%, respectively. In the Holidays dataset, the mean average precision (mAP) was 92.46%.Entities:
Keywords: feature fusion; image retrieval; information entropy; pagerank
Year: 2021 PMID: 34945976 PMCID: PMC8700127 DOI: 10.3390/e23121670
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1The proposed retrieval system framework.
Retrieval performance comparison in the Holidays dataset.
| Features | Dimensions | ||||||||
|---|---|---|---|---|---|---|---|---|---|
| 500 | 400 | 350 | 300 | 250 | 200 | 150 | 100 | 50 | |
| Color | 0.0848 | 0.0874 | 0.0876 | 0.0871 | 0.0869 | 0.0870 | 0.0870 | 0.0890 | 0.0857 |
| InceptionResNetV2 | 0.7092 | 0.7342 | 0.7306 | 0.7322 | 0.7324 | 0.7238 | 0.7121 | 0.6842 | 0.5762 |
| MobileNetV2 | 0.7199 | 0.7357 | 0.7468 | 0.7527 | 0.7459 | 0.7414 | 0.7289 | 0.7086 | 0.6175 |
| NASNetLarge | 0.3109 | 0.3232 | 0.3305 | 0.3302 | 0.3392 | 0.3525 | 0.3497 | 0.3378 | 0.2774 |
| ResNet50 | 0.7820 | 0.7903 | 0.7936 | 0.8013 | 0.8046 | 0.7953 | 0.7908 | 0.7533 | 0.6672 |
| VGG16 | 0.7128 | 0.7384 | 0.7432 | 0.7553 | 0.7508 | 0.7333 | 0.7295 | 0.6991 | 0.6046 |
| VGG19 | 0.7107 | 0.7444 | 0.7489 | 0.7495 | 0.7567 | 0.7406 | 0.7244 | 0.6833 | 0.5987 |
| Xception | 0.7401 | 0.7566 | 0.7637 | 0.7623 | 0.7696 | 0.7589 | 0.7511 | 0.7082 | 0.6355 |
| AlexNet | 0.6808 | 0.7007 | 0.7116 | 0.7241 | 0.7212 | 0.7172 | 0.7007 | 0.6681 | 0.5805 |
| GIST | 0.2552 | 0.2640 | 0.2804 | 0.2778 | 0.2909 | 0.2841 | 0.2923 | 0.2773 | 0.2324 |
| LBP | 0.3244 | 0.3244 | 0.3244 | 0.3244 | 0.3244 | 0.3272 | 0.3124 | 0.2909 | 0.2617 |
| DenseNet | 0.7548 | 0.7847 | 0.7963 | 0.8047 | 0.8047 | 0.7960 | 0.7851 | 0.7598 | 0.6687 |
| EfficientNetB0 | 0.7613 | 0.7740 | 0.7850 | 0.7873 | 0.7944 | 0.7795 | 0.7680 | 0.7364 | 0.6299 |
| AVG | 0.9018 | 0.9028 | 0.9114 | 0.9170 | 0.9132 | 0.9143 | 0.9125 | 0.9075 | 0.8981 |
| Ours | 0.9194 | 0.9188 | 0.9246 | 0.9245 | 0.9178 | 0.9239 | 0.9216 | 0.9150 | 0.9050 |
Figure 2The relationship between retrieval accuracy and selected ratio.
Figure 3The relationship between retrieval accuracy and final weights.
Comparison of retrieval results in the Holidays dataset.
| Method | Ours | [ | [ | [ | [ | [ | [ |
|---|---|---|---|---|---|---|---|
| mAP | 0.9246 | 0.9120 | 0.8845 | 0.8800 | 0.8720 | 0.8690 | 0.8552 |
Figure 4Precision in the Corel-1k dataset.
Precision for different numbers of returned images in the Corel-1k dataset.
| Features | Top10 | Top20 | Top30 | Top40 | Top50 |
|---|---|---|---|---|---|
| Color | 0.2995 | 0.2840 | 0.2788 | 0.2686 | 0.2613 |
| InceptionResNetV2 | 0.9115 | 0.9007 | 0.8823 | 0.8709 | 0.8570 |
| MobileNetV2 | 0.9175 | 0.9117 | 0.9040 | 0.8949 | 0.8795 |
| NASNetLarge | 0.9335 | 0.9255 | 0.9115 | 0.9061 | 0.8814 |
| ResNet50 | 0.2310 | 0.2482 | 0.2363 | 0.2226 | 0.2315 |
| VGG16 | 0.8970 | 0.8777 | 0.8593 | 0.8436 | 0.8278 |
| VGG19 | 0.8785 | 0.8697 | 0.5493 | 0.8364 | 0.8140 |
| Xception | 0.9155 | 0.9135 | 0.8993 | 0.8775 | 0.8495 |
| AlexNet | 0.2847 | 0.2847 | 0.2847 | 0.2847 | 0.2847 |
| GIST | 0.1162 | 0.1162 | 0.1162 | 0.1162 | 0.1162 |
| LBP | 0.1245 | 0.1245 | 0.1245 | 0.1245 | 0.1245 |
| DenseNet | 0.8642 | 0.8642 | 0.8642 | 0.8642 | 0.8642 |
| EfficientNetB0 | 0.8957 | 0.8957 | 0.8957 | 0.8957 | 0.8957 |
| AVG | 0.9935 | 0.9890 | 0.9876 | 0.9861 | 0.9851 |
| Ours | 0.9955 | 0.9910 | 0.9895 | 0.9872 | 0.9856 |
Figure 5The relationship between retrieval accuracy and selected ratio.
Figure 6The relationship between retrieval accuracy and final weights.
Comparison of search results in the Corel-1k dataset.
| Method | Ours | [ | [ | [ | [ | [ | [ |
|---|---|---|---|---|---|---|---|
| Precision | 0.9955 | 0.9709 | 0.8650 | 0.8550 | 0.8250 | 0.8040 | 0.8000 |
Precision for different numbers of returned images in the UC Merced Land Use dataset.
| Features | Top10 | Top20 | Top30 | Top40 | Top50 |
|---|---|---|---|---|---|
| Color | 0.1102 | 0.0996 | 0.0937 | 0.0879 | 0.0876 |
| InceptionResNetV2 | 0.5369 | 0.4825 | 0.4372 | 0.4032 | 0.3744 |
| MobileNetV2 | 0.6152 | 0.5582 | 0.5084 | 0.4747 | 0.4417 |
| NASNetLarge | 0.6026 | 0.5499 | 0.5035 | 0.4677 | 0.4355 |
| ResNet50 | 0.1740 | 0.1579 | 0.1438 | 0.1356 | 0.1308 |
| VGG16 | 0.5552 | 0.5015 | 0.4600 | 0.4347 | 0.4095 |
| VGG19 | 0.5480 | 0.5008 | 0.4586 | 0.4292 | 0.4035 |
| Xception | 0.5588 | 0.4997 | 0.4533 | 0.4224 | 0.3930 |
| AlexNet | 0.2038 | 0.2038 | 0.2038 | 0.2038 | 0.2038 |
| GIST | 0.1033 | 0.1033 | 0.1033 | 0.1033 | 0.1033 |
| LBP | 0.1269 | 0.1269 | 0.1269 | 0.1269 | 0.1269 |
| DenseNet | 0.5359 | 0.5359 | 0.5359 | 0.5359 | 0.5359 |
| EfficientNetB0 | 0.5369 | 0.5369 | 0.5369 | 0.5369 | 0.5369 |
| AVG | 0.8683 | 0.8117 | 0.7773 | 0.7368 | 0.6989 |
| Ours | 0.8802 | 0.8334 | 0.7879 | 0.7447 | 0.7067 |
Precision for different numbers of returned images in the RSSCN7 dataset.
| Features | Top10 | Top20 | Top30 | Top40 | Top50 |
|---|---|---|---|---|---|
| Color | 0.3019 | 0.3323 | 0.3253 | 0.3113 | 0.3026 |
| InceptionResNetV2 | 0.5391 | 0.5011 | 0.4739 | 0.4509 | 0.4370 |
| MobileNetV2 | 0.5623 | 0.5189 | 0.4877 | 0.4649 | 0.4495 |
| NASNetLarge | 0.5369 | 0.5025 | 0.4763 | 0.4461 | 0.4114 |
| ResNet50 | 0.3566 | 0.3390 | 0.3285 | 0.3226 | 0.3015 |
| VGG16 | 0.5316 | 0.4932 | 0.4608 | 0.4436 | 0.4278 |
| VGG19 | 0.5417 | 0.5063 | 0.4848 | 0.4664 | 0.4340 |
| Xception | 0.5353 | 0.4984 | 0.4667 | 0.4575 | 0.4495 |
| AlexNet | 0.2586 | 0.2586 | 0.2586 | 0.2586 | 0.2586 |
| GIST | 0.1829 | 0.1829 | 0.1829 | 0.1892 | 0.1892 |
| LBP | 0.2031 | 0.2031 | 0.2031 | 0.2031 | 0.2031 |
| DenseNet | 0.4225 | 0.4225 | 0.4225 | 0.4225 | 0.4225 |
| EfficientNetB0 | 0.4244 | 0.4244 | 0.4244 | 0.4244 | 0.4244 |
| AVG | 0.8651 | 0.8083 | 0.7683 | 0.7291 | 0.6851 |
| Ours | 0.8828 | 0.8326 | 0.7838 | 0.7372 | 0.6856 |
Figure 7Two sets of relations in the last two datasets.
Examples of image retrieval in different databases.
| Database | Holidays | Corel1k | UC Merced Land Use | RSSCN7 |
|---|---|---|---|---|
| Retrieval image |
|
|
|
|
| Results and similarity | ||||