| Literature DB >> 34804243 |
Abstract
The coronavirus disease (COVID-19) is an unparalleled crisis leading to a huge number of casualties and security problems. In order to reduce the spread of coronavirus, people often wear masks to protect themselves. This makes face recognition a very difficult task since certain parts of the face are hidden. A primary focus of researchers during the ongoing coronavirus pandemic is to come up with suggestions to handle this problem through rapid and efficient solutions. In this paper, we propose a reliable method based on occlusion removal and deep learning-based features in order to address the problem of the masked face recognition process. The first step is to remove the masked face region. Next, we apply three pre-trained deep Convolutional Neural Networks (CNN), namely VGG-16, AlexNet, and ResNet-50, and use them to extract deep features from the obtained regions (mostly eyes and forehead regions). The Bag-of-features paradigm is then applied to the feature maps of the last convolutional layer in order to quantize them and to get a slight representation comparing to the fully connected layer of classical CNN. Finally, Multilayer Perceptron (MLP) is applied for the classification process. Experimental results on Real-World-Masked-Face-Dataset show high recognition performance compared to other state-of-the-art methods.Entities:
Keywords: COVID-19; Deep learning; Face recognition; Masked face
Year: 2021 PMID: 34804243 PMCID: PMC8591434 DOI: 10.1007/s11760-021-02050-w
Source DB: PubMed Journal: Signal Image Video Process ISSN: 1863-1703 Impact factor: 1.583
Fig. 1Overview of the proposed method
Fig. 22D Face rotation
Fig. 3(1): Masked face. (2): Sampling the masked face image into 100 regions of the same size. (3): Cropping filter
Fig. 4VGG-16 network architecture introduced in [27]
Fig. 5AlexNet network architecture introduced in [15]
Fig. 6ResNet-50 network architecture introduced in [10]. The extracted DRF are shown
Fig. 7Pairs of face images from RMFRD dataset: face images without a mask (up) and with a mask (down)
Fig. 8Masked faces from SMFRD dataset
Recognition performance on RMFRD dataset using four codebook sizes
| Method | Size 1 | Size 2 | Size 3 | Size 4 |
|---|---|---|---|---|
| term vectors | 50 | 60 | 70 | 100 |
| VGG-16 | Model | |||
| Conv5 FM1 14 | 88.5% | 89.2% | 87.1% | 87.5% |
| Conv5 FM2 14 | 90.8% | 87.4% | 87.2% | 88.0% |
| Conv5 FM3 14 | 91.0% | 90.1% | 89.8% | |
| AlexNet | Model | |||
| Conv5 FM 13 | 84.3% | 85.7% | 85.9% | 86.6% |
| ResNet-50 | Model | |||
| Conv5 FM 7 | 87.4% | 87.9% | 89.5% | 89.3% |
Recognition performance on SMFRD dataset using four codebook sizes
| Method | Size 1 | Size 2 | Size 3 | Size 4 |
|---|---|---|---|---|
| term vectors | 50 | 60 | 70 | 100 |
| VGG-16 | Model | |||
| Conv5 FM1 14 | 82.4% | 83.7% | 84.5% | 84.7% |
| Conv5 FM2 14 | 83.1% | 83.5% | 85.0% | 85.4% |
| Conv5 FM3 14 | 81.7% | 81.3% | 84.4% | 85.6 |
| AlexNet | Model | |||
| Conv5 FM 13 | 83.7% | 83.9% | 84.2% | 86.0% |
| ResNet-50 | Model | |||
| Conv5 FM 7 | 83.5% | 84.7% |
| 88.5% |
Performance comparison with state-of-the-art methods
| Method | Dataset | Technique | Masks | Accuracy |
|---|---|---|---|---|
| Luttrell et al. [ | RMFRD | TL | yes | 85.7% |
| Hariri et al. [ | RMFRD | Covariance | yes | 84.6% |
| Almabdy et al. [ | RMFRD | CNN+SVM | yes | 87.0% |
| RMFRD | CNN+BoF | yes | ||
| Luttrell et al. [ | SMFRD | TL | yes | 83.3% |
| Hariri et al. [ | SMFRD | Covariance | yes | 83.8% |
| Almabdy et al. [ | SMFRD | CNN+SVM | yes | 86.1% |
| SMFRD | CNN+BoF | yes |
Training and testing time on the RMFRD dataset in milliseconds
| Method | AlexNet | VGG-16 |
|---|---|---|
| Almabdy et al. [ | train:550 | train:930 |
| test:34 | test:120 | |
| train: 308 | Train:605 | |
| test:21 | test:84 |