| Literature DB >> 35890840 |
Sandeep Kumar1, Shilpa Rani2, Arpit Jain3, Chaman Verma4, Maria Simona Raboaca5,6, Zoltán Illés4, Bogdan Constantin Neagu7.
Abstract
Nowadays, the demand for soft-biometric-based devices is increasing rapidly because of the huge use of electronics items such as mobiles, laptops and electronic gadgets in daily life. Recently, the healthcare department also emerged with soft-biometric technology, i.e., face biometrics, because the entire data, i.e., (gender, age, face expression and spoofing) of patients, doctors and other staff in hospitals is managed and forwarded through digital systems to reduce paperwork. This concept makes the relation friendlier between the patient and doctors and makes access to medical reports and treatments easier, anywhere and at any moment of life. In this paper, we proposed a new soft-biometric-based methodology for a secure biometric system because medical information plays an essential role in our life. In the proposed model, 5-layer U-Net-based architecture is used for face detection and Alex-Net-based architecture is used for classification of facial information i.e., age, gender, facial expression and face spoofing, etc. The proposed model outperforms the other state of art methodologies. The proposed methodology is evaluated and verified on six benchmark datasets i.e., NUAA Photograph Imposter Database, CASIA, Adience, The Images of Groups Dataset (IOG), The Extended Cohn-Kanade Dataset CK+ and The Japanese Female Facial Expression (JAFFE) Dataset. The proposed model achieved an accuracy of 94.17% for spoofing, 83.26% for age, 95.31% for gender and 96.9% for facial expression. Overall, the modification made in the proposed model has given better results and it will go a long way in the future to support soft-biometric based applications.Entities:
Keywords: Alex-Net; U-Net; age; face detection; face spoofing; facial expression; gender
Mesh:
Year: 2022 PMID: 35890840 PMCID: PMC9317232 DOI: 10.3390/s22145160
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1General diagram of Authentication [4].
Literature work on existing technology.
| Ref. | Author and Year | Dataset | Pre-Processing | Segment-ation | Method | Classification | Performance/Accuracy |
|---|---|---|---|---|---|---|---|
| [ | ErnoMakinen, 2008 | IIM, FERET | Y | N | Gender | SVM | 83.38% accuracy with automatic alignment method. |
| [ | Y. Jiang, 2014 | Mixed FERET, CAS-PEAL dataset | N | N | Gender | JFLDNNs | Accuracy is 89.63% on mixes dataset, which is good compared to CNN, LBP and CNN + LBP. |
| [ | Huajie Chen, 2006 | - | N | N | Gender | SVM and AAM | Classifier performance has been improved by adding pseudo examples |
| [ | B. Kabasakal, 2018 | LFW, IMDB and WIKI | Y | N | Gender | Google Net DNN, SVM | The performance of Google Net is 94.76%, which is better than SVM |
| [ | M. Mayo, 2008 | --- | N | N | Gender | SVMLinear, SVMQuad, RF200 | The accuracy of SVMQuad is 92.5%. |
| [ | M. Shabanian, 2019 | 317 MRI data from NDA | Y | N | Age | 3DCNN | Achieved sensitivity by 99% and specificity by98.3%. |
| [ | M. F. Aydogdu, 2017 | MORPH | Y | N | Age | 6 layer CNN, ResNet18 and ResNet-34 | The performance of ResNet18 and ResNet-34 is better for the age classification problem. |
| [ | T. Zheng, 2017 | MORPH | Y | N | Age | Deep probabilities CNN | Proposed methods have better accuracy compared to state of theart methods. |
| [ | S. Chen, 2017 | MORPH album 2 | Y | N | Age | SVM multi-class CNN | The proposed method has fewer errors in estimation compared to existing methods |
| [ | Fenker S. | Own dataset of 630 images | Y | N | Feature extraction and age prediction | __ | 69% of accuracy has been achieved. |
| [ | Karen Hollingsworth et al., 2009 | Own database collected at University | Y | N | Age prediction using IRIS | __ | Achieved 70% accuracy |
| [ | LVQNet et al., 2011 | CASIA-IrisV1 | Y | N | Age prediction using IRIS | CNN | LVQNet required 31 iterations for better results |
| [ | Salah EddineBekhouche et al., 2015 | Groups | N | N | Age and gender classification | SVM with non-linear kernel | Accuracy of age and gender classification is 88.8% 79.1%. |
| [ | X. Liu et al., 2017 | Adience, CAS-PEAL | N | N | Age and gender classification | Google Net | Accuracy of age and gender classification is about 98%. |
| [ | M. R. Dileep et al., 2016 | 1000 greyscale facial image | N | N | Age and gender classification | FFANN | Accuracy of age and gender classification is about 95%. |
| [ | K. Hu et al., 2018 | MRI image | N | Y | segmentation | U-Net | The performance of the proposed method is higher compared to other methods |
| [ | Sepidehsadat Hosseini et al., 2019 | Web face Morph II, FG-Net | N | N | Age and gender, facial expression classification | GF-CapsNet | Performance of Proposed GF- CapsNet is better than plain CNN for age, gender |
| [ | Ayesha Gurnani et al. | Adience, AffectNet | N | N | Age and gender, facial expression classification | SAF-BAGE | Performance is better comparatively. |
| [ | Zitong Yu et al., 2020 | OULU-NPU dataset, CASIA MFSD to Replay-Attack datasets | N | N | Face anti spoofing | CDCN | 0.2% ACER in Protocol- 1 of OULU-NPU dataset, |
| [ | Jiun-Da Lin et al., 2022 | Mask Dataset | N | N | Face anti spoofing | ArcFace Classifier (AC) | Performance of the proposed method is better than existing systems. |
| [ | Weihua Liu et al., 2021 | MIP-2D and MIP-3D, CASIA-SURF | N | Y | Face anti spoofing | D-Net, M-Net | ACER of 0.1071% outperforms all three comparative models with ACER of 0.4152%, 0.3425%, 0.1102% |
Figure 2Flow Chart of Proposed Methodology.
Figure 3Five-Layer U-Net based Architecture.
Convolution Layers output of U-Net based Architecture.
| Phases | Input | CNN Layers | Filter | Output Image | Sampling Type | Stride |
|---|---|---|---|---|---|---|
| 1 | 128 × 128 × 3 | 2 | 16- 3 × 3 | 128 × 128 × 16 | down | 2 |
| 2 | 128 × 128 × 16 | 2 | 32- 2 × 2 | 64 × 64 × 32 | down | 2 |
| 3 | 64 × 64 × 32 | 2 | 2 × 2 | 32 × 32 × 32 | down | 2 |
| 3 | 32 × 32 × 32 | 2 | 64- 3 × 3 | 32 × 32 × 64 | down | 2 |
| 4 | 32 × 32 × 64 | 2 | 2 × 2 | 16 × 16 × 64 | down | 2 |
| 4 | 16 × 16 × 64 | 2 | 128- 3 × 3 | 16 × 16 × 128 | down | 2 |
| 5 | 16 × 16 × 128 | 2 | 2 × 2 | 8 × 8 × 128 | down | 2 |
| 5 | 8 × 8 × 128 | 2 | 256- 3 × 3 | 8 × 8 × 256 | up | 2 |
| 5 | 8 × 8 × 256 | 2 | - | 16 × 16 × 128 | up | 2 |
| 5 | 16 × 16 × 128 | 2 | - | 16 × 16 × 256 | up | 2 |
| 5 | 16 × 16 × 256 | 2 | 128- 3 × 3 | 16 × 16 × 128 | up | 2 |
| 6 | 16 × 16 × 128 | - | 32 × 32 × 64 | upsampling to 32 × 32 × 64 | 2 | |
| 6 | 32 × 32 × 64 | - | 32 × 32 × 128 | 2 | ||
| 7 | 32 × 32 × 128 | 2 | 64- 3 × 3 | 32 × 32 × 64 | 2 | |
| 7 | 32 × 32 × 64 | - | 64 × 64 × 32 | upsampling to 64 × 64 × 32 | 2 | |
| 64 × 64 × 32 | 64 × 64 × 64 | 2 | ||||
| 8 | 64 × 64 × 64 | 2 | 32- 3 × 3 | 64 × 64 × 32 | upsampling to128 × 128 × 16 | 2 |
| 8 | 128 × 128 × 16 | - | 128 × 128 × 32 | 2 | ||
| 9 | 128 × 128 × 32 | 2 | 16- 3 × 3 | 128 × 128 × 16 | 128 × 128 × 1 | 2 |
Figure 4Max-pooling Structure of U-Net/Alex-Net-based Architecture.
Figure 5Real-time face detection with the processing time.
Figure 6Cropped Image.
Figure 7Five-Layer Alex-Net-based Architecture.
Figure 8Feature Extraction from cropped Image.
Figure 9Results of Real-time face detection.
Figure 10Overall GUI of Proposed Work.
Details of Alex-Net Classifier.
| Parameter Name | Spoofing | Age and Gender | Facial Expression |
|---|---|---|---|
| Epochs | 35 | 45 | 40 |
| Learning Rate | 0.001 | 0.001 | 0.001 |
| Droupout | 0.35 | 0.40 | 0.35 |
| Batch Size | 128 | 128 | 128 |
Details of Benchmark/Standard Datasets.
| Sr. No. | Dataset Name | Remarks |
|---|---|---|
| 1 | NUAA | Subjects = 15, Real Image = 5000, Fake Image = 7500 |
| 2 | CASIA | Subjects = 50, Real Image = 500, Fake Image = 450 |
| 3 | Adience | Total Image = 26,500, 8 Types of Categories |
| 4 | IOG | Total Image = 5100, 7 Types of Categories |
| 5 | CK+ | Subject = 123, Sequence = 593, Total Images = 10,600 |
| 6 | JAFFE | Total Image: 213, 7 Facial Expression |
Accuracy Comparison for Face Spoofing on NUAA and CASIA.
| Sr. No. | References | Accuracy on NUAA | References | Accuracy on CASIA |
|---|---|---|---|---|
| 1 | Bruno et al. [ | 82.9% | Bruno et al. [ | 83% |
| 2 | Zhen et al. [ | 86.9% | Zhen et al. [ | 88.2% |
| 3 | Jukka et al. [ | 82.3% | Duc-Tien et al. [ | 91.1% |
| 4 | Yaman et al. [ | 77.51% | G. Desmon et al. [ | 91.2% |
| 5 | Proposed Methodology | 91.1% | Proposed Methodology | 92.71% |
Figure 11Accuracy Comparison for Face Spoofing on NUAA.
Figure 12Accuracy Comparison for Face Spoofing on CASIA.
Accuracy Comparison for Age Classification on Adienceand IOG.
| Sr. No. | References | Accuracy on Adience | References | Accuracy on IOG |
|---|---|---|---|---|
| 1 | Gil et al. [ | 49.6 ± 4.7% | A. Ouafi et al. [ | 55.9% |
| 2 | Eran et al. [ | 44.9 ± 2.7% | C. Shan et al. [ | 60.1% |
| 3 | Afshin et al. [ | 60.9 ± 4.1% | M. Awad et al. [ | 62.9% |
| 4 | Zakariya et al. [ | 60.12% | R Enbar et al. [ | 67.1% |
| 5 | Proposed Methodology | 83.26 ± 4.3% | Proposed Methodology | 76.3% |
Figure 13Accuracy Comparison for Age Classification on Adience.
Figure 14Accuracy Comparison for Age Classification on IOG.
Accuracy Comparison for Gender Classification on Adienceand IOG.
| Sr. No. | Method | Accuracy on Adience | Method | Accuracy on IOG |
|---|---|---|---|---|
| 1 | Gil et al. [ | 86.9 ± 1.6% | J. L. Navarro et al. [ | 91.9% |
| 2 | Eran et al. [ | 78.1 ± 1.4% | K. Bowyer et al. [ | 92.8% |
| 3 | Zukang et al. [ | 79.13% | C. A. Perez et al. [ | 94.5% |
| 4 | Tal et al. [ | 79.7 ± 0.7% | A. Albial et al. [ | 96.7% |
| 5 | Proposed Methodology | 93.01 ± 2.03% | Proposed Methodology | 94.91% |
Figure 15Accuracy Comparison for Gender Classification on Adience.
Figure 16Accuracy Comparison for Gender Classification on IOG.
Accuracy Comparison for Facial Expression CK+.
| Expression | H. R. Wu et al. [ | A. Routray et al. [ | Proposed Accuracy (%) |
|---|---|---|---|
| Anger | 61.09 | 88.6 | 92.7 |
| Disgust | 58.61 | 94.41 | 96.7 |
| Happiness | 74.21 | 93.9 | 95.6 |
| Sadness | 68.81 | 95.09 | 96.07 |
| Surprise | 85.22 | 97.17 | 94.31 |
Accuracy Comparison for Facial Expression JAFFE.
| Expression | H. R. Wu et al. [ | H. R. Wu et al. [ | Proposed Accuracy (%) |
|---|---|---|---|
| Anger | 63.01 | 99.9 | 96.41 |
| Disgust | 58.51 | 87.3 | 94.13 |
| Fear | 61.9 | 94.01 | 95.43 |
| Happiness | 76.41 | 97.01 | 96.73 |
| Sadness | 68.81 | 78.11 | 95.07 |
| Surprise | 83.91 | 97.26 | 96.51 |
Figure 17Accuracy Comparison for Facial Expression CK+.
Figure 18Accuracy Comparison for Facial Expression JAFFE.