| Literature DB >> 34068795 |
Artur Grudzień1, Marcin Kowalski1, Norbert Pałka1.
Abstract
This paper reports on a new approach to face verification in long-wavelength infrared radiation. Two face images were combined into one double image, which was then used as an input for a classification based on neural networks. For testing, we exploited two external and one homemade thermal face databases acquired in various variants. The method is reported to achieve a true acceptance rate of about 83%. We proved that the proposed method outperforms other studied baseline methods by about 20 percentage points. We also analyzed the issue of extending the performance of algorithms. We believe that the proposed double image method can also be applied to other spectral ranges and modalities different than the face.Entities:
Keywords: convolutional neural networks; face verification; long-wavelength infrared radiation
Year: 2021 PMID: 34068795 PMCID: PMC8126239 DOI: 10.3390/s21093301
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Gallery of thermal face images from (a) PROTECT dataset, (b) CARL dataset, (c) In-House (FLIR A65), (d) and In-House (FLIR P640).
Parameters of the algorithms used in the study.
| Name of Method | Parameters |
|---|---|
| Histogram of Oriented Gradients | cell size: 30 × 30 pixels |
| Local Binary Pattern | cell size: 30 × 30 pixels |
| Local Derivative Pattern | cell size: 30 × 30 pixels |
Figure 2Scheme of combining two images into genuine and impostor classes.
Figure 3Scheme of the verification through identification method.
Percentage share of each database in the appropriate training and test dataset for local descriptor and CNNs models.
| Training Datasets | ||||
|---|---|---|---|---|
| Number of Datasets | IOEA [%] | IOEP [%] | PROT [%] | CARL [%] |
| 1 | 13.04 | 23.91 | 34.78 | 28.27 |
| 2 | 9.78 | 22.83 | 31.52 | 35.87 |
| 3 | 11.96 | 19.57 | 36.96 | 31.51 |
| 4 | 11.96 | 23.91 | 36.96 | 27.17 |
| 5 | 10.87 | 26.09 | 34.78 | 28.26 |
|
| ||||
|
|
|
|
|
|
| 1 | 10.00 | 17.50 | 35.00 | 37.50 |
| 2 | 17.50 | 20.00 | 42.50 | 20.00 |
| 3 | 12.50 | 27.50 | 30.00 | 30.00 |
| 4 | 12.50 | 17.50 | 30.00 | 40.00 |
| 5 | 15.00 | 12.50 | 35.00 | 37.50 |
TAR results for the local descriptor and CNN models with the decision function based on metrics.
| Local Descriptor Methods | |||
|---|---|---|---|
| Algorithm of Feature Extraction | Distance Function | TAR @ FAR 1% | TAR @ FAR 0.1% |
| HOG | Spearman | 48.46 ± 4.11 | 36.22 ± 4.50 |
|
|
|
|
|
| LDP | Spearman | 35.25 ± 1.83 | 17.88 ± 2.51 |
|
| |||
|
|
|
|
|
| DenseNet-201 | Spearman | 53.15 ± 4.67 | 34.13 ± 6.16 |
| GoogLeNet | Euclidean | 48.18 ± 2.51 | 26.85 ± 2.97 |
| InceptionResNet-v2 | Spearman | 45.01 ± 4.86 | 26.49 ± 5.19 |
| Inception-v3 | Spearman | 46.33 ± 3.97 | 29.71 ± 4.08 |
| ResNet-18 | Euclidean | 53.60 ± 2.04 | 33.04 ± 2.49 |
| ResNet-50 | Euclidean | 49.56 ± 4.01 | 29.92 ± 3.48 |
| ResNet-101 | Spearman | 49.50 ± 4.73 | 30.73 ± 5.51 |
| VGG16 | Euclidean | 44.08 ± 2.06 | 26.19 ± 2.55 |
| VGG19 | Euclidean | 48.68 ± 3.27 | 28.69 ± 6.17 |
TAR results for the local descriptor and CNN models with SVM.
| Local Descriptors Methods | ||
|---|---|---|
| Algorithm of Feature Extraction | TAR @ FAR 1% | TAR @ FAR 0.1% |
|
|
|
|
| LBP | 5.08 ± 6.81 | 0.86 ± 1.21 |
| LDP | 5.95 ± 3.16 | 1.65 ± 1.19 |
|
| ||
| AlexNet | 6.35 ± 3.12 | 0.36 ± 0.18 |
| DenseNet-201 | 4.03 ± 0.72 | 0.10 ± 0.10 |
|
|
|
|
| InceptionResNet-v2 | 2.72 ± 1.09 | 0.25 ± 0.25 |
| Inception-v3 | 5.15 ± 2.96 | 0.32 ± 0.38 |
| ResNet-18 | 6.42 ± 1.74 | 1.01 ± 0.55 |
| ResNet-50 | 4.39 ± 2.45 | 0.32 ± 0.26 |
| ResNet-101 | 4.90 ± 1.95 | 0.44 ± 0.92 |
| VGG16 | 5.74 ± 3.27 | 0.99 ± 0.95 |
| VGG19 | 5.85 ± 2.50 | 1.24 ± 1.15 |
Figure 4ROC curves for individual databases: (A) CNNs with metrics, (B) CNNs with SVM, (C) local descriptor methods with metrics, and (D) local descriptor methods with SVM.
Percentage share of each database in the training and testing datasets for the VTI method.
| Training Datasets | ||||
|---|---|---|---|---|
| Number of Datasets | IOEA [%] | IOEP [%] | PROT [%] | CARL [%] |
| 1 | 11.96 | 20.65 | 34.78 | 32.61 |
| 2 | 13.04 | 21.74 | 36.96 | 28.26 |
| 3 | 10.87 | 20.65 | 35.87 | 32.61 |
| 4 | 10.87 | 26.09 | 34.78 | 28.26 |
| 5 | 15.22 | 20.65 | 36.96 | 27.17 |
|
| ||||
|
|
|
|
|
|
| 1 | 12.50 | 25.00 | 35.00 | 27.50 |
| 2 | 10.00 | 22.50 | 30.00 | 37.50 |
| 3 | 15.00 | 25.00 | 32.50 | 27.50 |
| 4 | 15.00 | 12.50 | 35.00 | 37.50 |
| 5 | 5.00 | 25.00 | 30.00 | 40.00 |
TAR results for the VTI method.
| Neural Network Model | TAR @ FAR 1% | TAR @ FAR 0.1% |
|---|---|---|
| AlexNet | 61.57 ± 14.89 | 21.14 ± 10.09 |
| DenseNet-201 | 78.16 ± 7.29 | 11.01 ± 24.63 |
| GoogLeNet | 58.00 ± 12.06 | 4.71 ± 10.53 |
| InceptionResNet-v2 | 64.66 ± 9.81 | 9.76 ± 21.83 |
|
|
|
|
| ResNet-18 | 59.92 ± 9.06 | 20.51 ± 13.44 |
| ResNet-50 | 57.89 ± 9.01 | 8.24 ± 11.46 |
| ResNet-101 | 58.28 ± 10.23 | 15.42 ± 14.91 |
| VGG16 | 63.64 ± 11.57 | 17.23 ± 15.91 |
| VGG19 | 60.25 ± 6.76 | 19.82 ± 18.33 |
Figure 5ROC curves for individual databases for the VTI method.
Test results for different sample image and test image positions for the VTI method for the Inception-v3 model.
| Full Augmentation | ||
|---|---|---|
| Location of Images | TAR @ FAR 1% | TAR @ FAR 0.1% |
| “12” | 83.92 ± 6.96 | 49.24 ± 19.08 |
| “21” | 85.33 ± 11.53 | 52.81 ± 15.69 |
|
| ||
|
|
|
|
| “21” | 83.39 ± 8.94 | 51.46 ± 13.04 |
Figure 6The best ROC curves for five face verification methods.