| Literature DB >> 35765303 |
Ricardo Mar-Cupido1, Vicente García1, Gilberto Rivera1, J Salvador Sánchez2.
Abstract
The use of face masks in public places has emerged as one of the most effective non-pharmaceutical measures to lower the spread of COVID-19 infection. This has led to the development of several detection systems for identifying people who do not wear a face mask. However, not all face masks or coverings are equally effective in preventing virus transmission or illness caused by viruses and therefore, it appears important for those systems to incorporate the ability to distinguish between the different types of face masks. This paper implements four pre-trained deep transfer learning models (NasNetMobile, MobileNetv2, ResNet101v2, and ResNet152v2) to classify images based on the type of face mask (KN95, N95, surgical and cloth) worn by people. Experimental results indicate that the deep residual networks (ResNet101v2 and ResNet152v2) provide the best performance with the highest accuracy and the lowest loss.Entities:
Keywords: COVID-19; Deep learning; Face mask; Recognition; Transfer learning
Year: 2022 PMID: 35765303 PMCID: PMC9222491 DOI: 10.1016/j.asoc.2022.109207
Source DB: PubMed Journal: Appl Soft Comput ISSN: 1568-4946 Impact factor: 8.263
Fig. 1Types of face masks commonly worn during the COVID-19 pandemic (from left to right, they are KN95 mask, N95 mask, cloth mask, surgical mask, and without mask).
Fig. 2Sample images from the databases for faces with and without a mask.
Fig. 3Framework of the proposed methodology.
Fig. 4Region of interest (ROI).
Confusion matrix for multi-class problems.
| Predicted | |||||
|---|---|---|---|---|---|
| Actual | Total | ||||
| ⋮ | ⋮ | ⋮ | ⋮ | ⋮ | |
| Total | … | ||||
Accuracy and loss for the recognition models.
| Accuracy | Loss | |
|---|---|---|
| NasNetMobile | 0.9324 | 0.2160 |
| MobileNetv2 | 0.9324 | 0.2131 |
| ResNet101v2 | 0.9800 | 0.0734 |
| ResNet152v2 | 0.9737 | 0.0808 |
Confusion matrix of the pre-trained deep transfer learning models (true class labels on the rows and predicted class labels on the columns).
Recall, precision, specificity and F1-score for the ResNet101v2 model.
| Recall | Precision | Specificity | F1-score | |
|---|---|---|---|---|
| KN95 | 0.9793 | 0.9895 | 0.9974 | 0.9844 |
| N95 | 0.9828 | 0.9794 | 0.9948 | 0.9811 |
| Cloth | 0.9655 | 0.9825 | 0.9957 | 0.9739 |
| Surgical | 0.9897 | 0.9696 | 0.9922 | 0.9795 |
| Without | 0.9828 | 0.9794 | 0.9948 | 0.9811 |
| Macro-avg | 0.9802 | 0.9801 | 0.9950 | 0.9802 |
Fig. 5Examples of images misclassified by the ResNet101v2 model (text means ‘true classpredicted class’).
Comparison of our proposal against related works.
| Reference | Techniques | Classification problem | # Classes | Results |
|---|---|---|---|---|
| This paper | ResNet101v2 | Multi-class | 5 | Accuracy = 98.00% |
| InceptionV3 | Two-class | 2 | Accuracy = 100% | |
| SRCNet | Multi-class | 3 | Accuracy = 98.70% | |
| ResNet50 and YOLOv2 | Two-class | 2 | Average Precision = 81% | |
| YOLOv3 and Faster R-CNN (ResNet101-FPN) | Two-class | 2 | Average Precision = 62% | |
| ResNet50 and SVM | Two-class | 2 | Accuracy = 100% | |
| Faster R-CNN and InceptionV2 | Two-class | 2 | F1-score = 94.19% | |
| VGG16 | Multi-class | 7 | Accuracy = 99.29% | |
| SE-YOLOv3 | Multi-class | 3 | Average Precision = 73.7% | |
| CSPDarkNet53, PANet and YOLOv4 | Multi-class | 3 | Average Precision = 98.30%, F1-score = 96.7% | |
| ResNet50 | Two-class | 2 | Accuracy = 98.20% | |
| YOLOv3 | Two-class | 2 | Average confidence = 97.00% | |
| MobileNetv2 | Two-class | 2 | Accuracy = 96.85% | |
| MobileNetv2 and Single Shot Detector | Two-class | 2 | Accuracy = 91.70% | |
| MobileNetv2 | Two-class | 2 | Accuracy= 98.00% | |
| MobileNetv2 | Two-class | 2 | Accuracy = 98.00% | |
| InceptionV3 | Two-class | 2 | Accuracy = 98.00% | |
| Single Shot Multibox Detector and MobileNetv2 | Two-class | 2 | Accuracy = 92.64%, F1-score = 93.00% | |
| YOLOv4 | Two-class | 2 | Average Precision = 88%, F1-score = 99.54% | |
| MobileNetv2 | Two-class | 2 | Accuracy = 81.74% | |
| Multigraph CN-VGG16 | Two-class | 2 | Accuracy = 97.9% | |
| MobileNetv2, DenseNet121, NASNet | Two-class | 2 | F1-score = 99.40% | |
| MobileNetv2-SVM | Two-class | 2 | Accuracy = 97.11% | |
| VGG-16 | Multi-class | 3 | Accuracy = 99.81% |