| Literature DB >> 34764581 |
Yande Li1, Kun Guo1, Yonggang Lu1, Li Liu2.
Abstract
The global epidemic of COVID-19 makes people realize that wearing a mask is one of the most effective ways to protect ourselves from virus infections, which poses serious challenges for the existing face recognition system. To tackle the difficulties, a new method for masked face recognition is proposed by integrating a cropping-based approach with the Convolutional Block Attention Module (CBAM). The optimal cropping is explored for each case, while the CBAM module is adopted to focus on the regions around eyes. Two special application scenarios, using faces without mask for training to recognize masked faces, and using masked faces for training to recognize faces without mask, have also been studied. Comprehensive experiments on SMFRD, CISIA-Webface, AR and Extend Yela B datasets show that the proposed approach can significantly improve the performance of masked face recognition compared with other state-of-the-art approaches.Entities:
Keywords: Attentional mechanism; Cropping-based approach; Masked face recognition
Year: 2021 PMID: 34764581 PMCID: PMC7847808 DOI: 10.1007/s10489-020-02100-9
Source DB: PubMed Journal: Appl Intell (Dordr) ISSN: 0924-669X Impact factor: 5.086
Fig. 1Cases for MFR. Case 2 and case 3 are two special cases and case 4 is conventional face recognition
Fig. 2Diagram of Convolutional Block Attention Module
Fig. 3Examples of cropped images at different cropping proportions
Fig. 4Framework of proposed attention ResNet in this paper. The CBAM module is adopted in every convolution block in ResNet50
The comparison of different attention modules on Masked-Webface dataset
| Cases | Baseline(%) | BAM(%) | SE(%) | CBAM(%) |
|---|---|---|---|---|
| Case1 | 85.925 | 87.725 | 85.526 | |
| Case2 | 41.925 | 45.292 | 44.122 | |
| Case3 | 43.329 | 43.721 | 43.444 | 43.689 |
| Case4 | 94.211 | 94.327 | 94.376 | 94.411 |
Bold entries are the necessary findings
The comparison of different attention modules on AR dataset and Extend Yela B Dataset
| Baseline(%) | BAM(%) | SE(%) | CBAM(%) | |
|---|---|---|---|---|
| AR | 97.600 | 97.800 | 97.600 | 98.400 |
| Yela B | 99.474 | 99.342 | 99.342 | 99.474 |
Test results of different test categories on Masked-LFW dataset
| Categories | 300 | 400 | 500 | 600 | 700 | 800 | 900 | 1000 | 3000 |
|---|---|---|---|---|---|---|---|---|---|
| CBAM(%) | 96.8153 | 94.958 | 91.8644 | 91.0364 | 90.3981 | 89.4422 | 88.1295 | 87.1753 | 82.8648 |
The MFR performance at different cropping proportions with CBAM
| Cropping | Case1 | Case2 | Case3 |
|---|---|---|---|
| proportion(*L) | Acc(%) | Acc(%) | Acc(%) |
| 0.4 | 80.066 | 80.122 | 74.929 |
| 0.5 | 82.152 | 82.796 | 76.962 |
| 0.55 | 83.371 | 83.906 | 79.12 |
| 0.6 | 84.293 | 85.507 | 81.25 |
| 0.65 | 85.263 | 85.976 | 82.01 |
| 85.461 | |||
| 0.75 | 87.292 | 86.211 | 81.439 |
| 0.8 | 89.163 | 85.843 | 81.359 |
| 0.85 | 89.31 | 85.162 | 80.101 |
| 84.497 | 76.348 | ||
| 0.95 | 91.01 | 82.969 | 75.912 |
| 1 | 90.95 | 80.01 | 74.606 |
| 1.1 | 90.212 | 77.212 | 71.421 |
| 1.2 | 89.492 | 73.012 | 66.437 |
| Uncut | 88.5 | 46.828 | 43.689 |
Bold entries are the necessary findings
The MFR performance at different cropping proportions without attention
| Cropping | Case1 | Case2 | Case3 |
|---|---|---|---|
| proportion(*L) | Acc(%) | Acc(%) | Acc(%) |
| 0.4 | 79.423 | 77.426 | 74.162 |
| 0.5 | 81.547 | 79.206 | 75.124 |
| 0.55 | 82.169 | 79.461 | 77.426 |
| 0.6 | 83.962 | 81.427 | 80.012 |
| 0.65 | 83.991 | 82.529 | 80.602 |
| 84.101 | |||
| 0.75 | 85.264 | 82.112 | 80.042 |
| 0.8 | 85.921 | 82.011 | 79.429 |
| 0.85 | 86.796 | 81.864 | 78.926 |
| 81.519 | 75.42 | ||
| 0.95 | 86.977 | 80.974 | 74.194 |
| 1 | 86.539 | 80.112 | 73.12 |
| 1.1 | 86.32 | 76.421 | 69.521 |
| 1.2 | 86.101 | 73.112 | 64.421 |
| Uncut | 85.925 | 41.925 | 43.329 |
Bold entries are the necessary findings
Fig. 5The cropping performance at different proportions with CBAM module (left) and without attention(right) on Masked-Webface dataset
Performance comparison between our approach and state-of-the-art approaches
| Cases | Baseline(%) | BAM(%) | SE(%) | CBAM(%) | Cosface(%) | Arcface(%) | PDSN(%) | Ours(%) |
|---|---|---|---|---|---|---|---|---|
| Case1 | 85.925 | 87.725 | 85.526 | 88. 5 | 87.906 | 88.01 | 91.421 | |
| Case2 | 41.925 | 45.292 | 44.122 | 46.828 | 45.372 | 45.494 | 69.426 | |
| Case3 | 43.329 | 43.721 | 43.444 | 43.689 | 43.551 | 44.012 | 62.914 | |
| Case4 | 94.211 | 94.327 | 94.376 | 94.411 | 94.761 | 94.794 | 92.612 |
Bold entries are the necessary findings
Fig. 6Performance comparison of MFR with different approaches on Masked-Webface dataset
Fig. 7Examples of class activation maps of different approaches