| Literature DB >> 35271074 |
Fadi Boutros1,2, Naser Damer1,2, Kiran Raja3, Florian Kirchbuchner1, Arjan Kuijper1,2.
Abstract
This work addresses the challenge of building an accurate and generalizable periocular recognition model with a small number of learnable parameters. Deeper (larger) models are typically more capable of learning complex information. For this reason, knowledge distillation (kd) was previously proposed to carry this knowledge from a large model (teacher) into a small model (student). Conventional KD optimizes the student output to be similar to the teacher output (commonly classification output). In biometrics, comparison (verification) and storage operations are conducted on biometric templates, extracted from pre-classification layers. In this work, we propose a novel template-driven KD approach that optimizes the distillation process so that the student model learns to produce templates similar to those produced by the teacher model. We demonstrate our approach on intra- and cross-device periocular verification. Our results demonstrate the superiority of our proposed approach over a network trained without KD and networks trained with conventional (vanilla) KD. For example, the targeted small model achieved an equal error rate (EER) value of 22.2% on cross-device verification without KD. The same model achieved an EER of 21.9% with the conventional KD, and only 14.7% EER when using our proposed template-driven KD.Entities:
Keywords: biometrics; knowledge distillation; periocular verification
Mesh:
Year: 2022 PMID: 35271074 PMCID: PMC8914924 DOI: 10.3390/s22051921
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Overview of the proposed template-driven KD approach for periocular verification based on the ResNet architecture. Note that both the template loss and KD output loss contribute to the distillation process.
Figure 2Sample images from the periocular database employed.
The inference time (in milliseconds) and the number of trainable parameters (in millions (m)) for each of the evaluated models.
| Model | Inference Time | No. Trainable Parameters |
|---|---|---|
| ReseNet-110 | 0.015 ms | 1.8 m |
| ResNet-18 | 0.007 ms | 11.2 m |
| ResNet-34 | 0.008 ms | 21.3 m |
Figure 3The achieved ROC curves for different experimental settings. For each experimental setting, the number next to the model label is the achieved AUC. Note the improvement in the ResNet-110 verification performance using our proposed template-driven KD approach.
Performance obtained for different experimental settings. The first three rows of the table present the achieved result for the three evaluated models (without using KD), where the smallest model (ResNet-100) performed the worst. The next three rows of the table present the achieved verification performance by including KD in the training process using ResNet-18 as a teacher with conventional KD loss, and both the template-driven KD with MSE and cosine (COS) embedding loss. The last three rows present the achieved KD verification performance using ResNet-34 as a teacher with conventional KD, and both the template-driven KD with MSE and cosine embedding loss. The enhanced performance by the proposed method in comparison to the conventional KD is illustrated over all experimental setups. The lowest EER, FMR10 and FMR100 for each of the verification scenarios (iPhone, Nokia and Cross-Smartphone) achieved by ResNet-100 are in bold.
| Model | Teacher | iPhone | Nokia | Cross-Smartphone | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| EER | FMR10 | FMR100 | EER | FMR10 | FMR100 | EER | FMR10 | FMR100 | ||
| ResNet-18 | - | 0.0251 | 0.0032 | 0.0557 | 0.0304 | 0.0108 | 0.0776 | 0.1104 | 0.1216 | 0.5171 |
| ResNet-34 | - | 0.0273 | 0.0071 | 0.0431 | 0.0218 | 0.0033 | 0.0434 | 0.0871 | 0.0767 | 0.3267 |
| ResNet-110 | - | 0.0883 | 0.0839 | 0.1984 | 0.0694 | 0.0473 | 0.2298 | 0.2216 | 0.3945 | 0.6932 |
| ResNet-110_KD18 | ResNet-18 | 0.0713 | 0.0547 | 0.1401 | 0.0471 | 0.0198 | 0.1284 | 0.2187 | 0.3619 | 0.6872 |
| ResNet-110_KD18MSE | ResNet-18 | 0.0632 | 0.0500 | 0.1401 | 0.0353 | 0.0114 | 0.0763 | 0.1753 | 0.2539 | 0.5752 |
| ResNet-110_KD18COS | ResNet-18 | 0.0702 | 0.0571 | 0.1171 |
|
| 0.0724 | 0.1587 | 0.2215 | 0.5051 |
| ResNet-110_KD34 | ResNet-34 | 0.0651 | 0.0530 | 0.1153 | 0.0413 | 0.0227 | 0.0844 | 0.1661 | 0.2321 | 0.5494 |
| ResNet-110_KD34MSE | ResNet-34 |
|
|
| 0.0375 | 0.0193 |
|
|
|
|
| ResNet-110_KD34COS | ResNet-34 | 0.0775 | 0.0711 | 0.1161 | 0.0457 | 0.0265 | 0.0962 | 0.1725 | 0.2580 | 0.5368 |