| Literature DB >> 36093471 |
Aimee Booysens1, Serestina Viriri1.
Abstract
Biometrics is the recognition of a human using biometric characteristics for identification, which may be physiological or behavioral. The physiological biometric features are the face, ear, iris, fingerprint, and handprint; behavioral biometrics are signatures, voice, gait pattern, and keystrokes. Numerous systems have been developed to distinguish biometric traits used in multiple applications, such as forensic investigations and security systems. With the current worldwide pandemic, facial identification has failed due to users wearing masks; however, the human ear has proven more suitable as it is visible. Therefore, the main contribution is to present the results of a CNN developed using EfficientNet. This paper presents the performance achieved in this research and shows the efficiency of EfficientNet on ear recognition. The nine variants of EfficientNets were fine-tuned and implemented on multiple publicly available ear datasets. The experiments showed that EfficientNet variant B8 achieved the best accuracy of 98.45%.Entities:
Mesh:
Year: 2022 PMID: 36093471 PMCID: PMC9451992 DOI: 10.1155/2022/3514807
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Diagram of the outer ear.
Summary of biometric characteristics.
| Biometric identifier | Biometric type | Distinctiveness | Permanence | Collectability | Performance | Acceptability |
|---|---|---|---|---|---|---|
| DNA | Physiological | High | High | Low | High | Low |
| Ear | Physiological | Medium | High | Medium | Medium | High |
| Face | Physiological | Low | Medium | High | Low | High |
| Facial | Physiological | High | Low | High | Medium | High |
| Fingerprint | Physiological | High | High | Medium | High | Medium |
| Gait | Behavioral | Low | Low | High | Low | High |
| Hand geometry | Physiological | Medium | Medium | High | Medium | Medium |
| Hand vein | Physiological | Medium | Medium | Medium | Medium | Medium |
| Iris | Physiological | High | High | Medium | High | Low |
| Keystroke | Behavioral | Low | Low | Medium | Low | Medium |
| Odor | Physiological | High | High | Low | Low | Medium |
| Palm print | Physiological | High | High | Medium | High | Medium |
| Retina | Physiological | High | Medium | Low | High | Low |
| Signature | Behavioral | Low | Low | High | Low | High |
| Voice | Combination of physiological and behavioral | Low | Low | Medium | Low | High |
Summary of the related works.
| Author | Dataset | Accuracy | Summary |
|---|---|---|---|
| Emeršič et al. [ | NA | 30 | It was a handcrafted feature extraction method, such as LBP and patterns of oriented edge magnitudes (POEM), and CNN-based feature extraction methods were used to obtain the ear identification |
|
| |||
| Tian and Mu [ | AMI, WPUT, IITD, and UERC | 70.58, 67.01, 81.98, and 57.75 | This system used deep convolutional neural network (CNN) to ear recognition. There were occlusions like no earrings, headsets, or similar occlusions |
|
| |||
| Raveane et al. [ | NA | 98 | This system used variable conditions, and this could also be because of the odd shape of the human ears and changing lighting conditions |
|
| |||
| Zhang and Mu [ | Notre Dame Biometrics database and University of Beira Interior Ear dataset | 100 and 98.22 | This system contained large occlusions, scale, and pose variation |
|
| |||
| Kohlakala and Coetzer [ | Mathematical Analysis of Images Ear database and Indian Institute of Technology Delhi Ear database | 99.2 and 96.06 | It is used to classify ears in either the foreground or background of the image. The binary contour image applied the matching for feature extraction, and this was done by implementing a Euclidean distance measure, which had a ranking to verify for authentication |
|
| |||
| Tomczyk and Szczepaniak [ | NA | NA | It shows the published experimental results that the approach did the rotation equivalence property to detect rotated structures |
|
| |||
| Hammam et al. [ | Three ear datasets but not stated | 22 | The paper took seven performing handcrafted descriptors to extract the discriminating ear image. They then took the extracted ear and trained it using support vector machines (SVM) to learn a suitable model |
|
| |||
| Alkababji and Mohammed [ | NA | 97.8 | It used the principal component analysis (PCA) and a genetic algorithm for feature reduction and selection |
|
| |||
| Jamil et al. [ | Very underexposed or overexposed database | 97 | They considered that their work was the first to test the performance of CNN on very underexposed or overexposed images |
|
| |||
| Hansley et al. [ | UERC challenge | NA | This was done using handcrafted descriptors, which were fused to improve recognition |
Summary of datasets.
| Database | Year | Number of subjects | Number of images | Left ear count | Right ear count | Total ears | Image size | Country | Sides | |
|---|---|---|---|---|---|---|---|---|---|---|
| 1 | Institute of Technology Delhi Ear Database (IIT Delhi-I) [ | 2007 | 121 | 471 | 471 | 471 | 272 × 204 | India | Right | |
| Institute of Technology Delhi Ear Database (IIT Delhi-II) [ | NA | 221 | 793 | 793 | 793 | 272 × 204 | India | Right | ||
|
| ||||||||||
| 2 | The University of Science and Technology Beijing (USTB ear I) [ | 2002 | 60 | 185 | 185 | 185 | Varied | China | Right | |
| The University of Science and Technology Beijing (USTB ear II) [ | 2004 | 77 | 308 | 308 | 308 | Varied | China | Right | ||
|
| ||||||||||
| 3 | The Annotated Web Ears database (AWE) [ | 2016 | 100 | 1000 | 500 | 500 | 1000 | Varied | Slovenia | Both |
| The Annotated Web Ears database extended (AWE extend) [ | 2017 | 346 | 4104 | 2052 | 2052 | 4104 | Varied | Slovenia | Both | |
|
| ||||||||||
| 4 | Mathematical Analysis of Images Ear database (AMI) [ | NA | 106 | 700 | 420 | 280 | 700 | 492 × 702 | Spain | Both |
|
| ||||||||||
| 5 | The West Pomeranian University of Technology Ear database (WPUTE) [ | 2010 | 501 | 2071 | 829 | 1242 | 2071 | Varied | Poland | Both |
|
| ||||||||||
| 6 | Unconstrained Ear Recognition Challenge database (UERC) [ | 2017 | 3706 | 11804 | 5902 | 5902 | 11804 | Varied | Slovenia | Both |
|
| ||||||||||
| 7 | EarVN1.0 [ | 2018 | 164 | 28412 | 14206 | 14206 | 28412 | Varied and low resolution | Vietnam | Both |
|
| ||||||||||
| 8 | The In-the-Wild Ear database (ITWE) [ | 2015 | 55 | 605 | 424 | 181 | 605 | Varied | Slovenia | Both |
|
| ||||||||||
| 9 | The Carreira-Perpinan (CP) [ | 1995 | 17 | 102 | 102 | 102 | Varied | NA | Left | |
|
| ||||||||||
| 10 | The University of Beira Ear Database (UBEAR) [ | 2011 | 126 | 4430 | 2215 | 2215 | 4430 | 1280 × 960 | Mozambique | Both |
|
| ||||||||||
| 11 | Indian Institute of Technology Kanpur (IITK) [ | 2011 | 801 | 190 | 95 | 95 | 190 | Varied | India | Both |
| 12 | The Forensic Ear Identification Database (FEARID) [ | 2005 | 1229 | 1229 | 615 | 614 | 1229 | Varied | UK, Italy, and Netherlands | Both |
|
| ||||||||||
| 13 | University of Notre Dame (UND) [ | 2006 | 3480 | 952 | 952 | 952 | Varied | France | Left | |
|
| ||||||||||
| 14 | The Face Recognition Technology database $FERET) [ | 2010 | 9427 | 4745 | 3796 | 949 | 4745 | Varied | Spain | Both |
|
| ||||||||||
| 15 | The Pose, Illumination and Expression (PIE) [ | 2002 | 40000 | 68 | 34 | 34 | 68 | Varied | USA | Both |
|
| ||||||||||
| 16 | The XM2VTS Ear database [ | NA | 2360 | 295 | 89 | 206 | 295 | 720 × 576 | UK | Both |
|
| ||||||||||
| 17 | The West Virginia University (WVU) [ | 2006 | 460 | 402 | 402 | 402 | Varied | USA | Left | |
Figure 2Examples of original ear images. (a) Example of a 2D profile image for a female. (b) Example of a 2D profile image for a male. (c) Example of a facial image for a female. (d) Example of a facial image for a male.
Figure 3Examples of extracted ear images. (a) Example of ear extracted from 2D profile image for a female. (b) Example of ear extracted from 2D profile image for a male. (c) Example of ear extracted from facial image for a female. (d) Example of ear extracted from facial image for a male.
Figure 4Block structure of the proposed model.
Figure 5Accuracy for the ear dataset of each EfficientNet. (a) Accuracy for EfficientNet B0. (b) Accuracy for EfficientNet B1. (c) Accuracy for EfficientNet B2. (d) Accuracy for EfficientNet B3. (e) Accuracy for EfficientNet B4. (f) Accuracy for EfficientNet B5. (g) Accuracy for EfficientNet B6. (h) Accuracy for EfficientNet B7. (i) Accuracy for EfficientNet B8.
Figure 6Loss for the ear dataset of each EfficientNet. (a) Loss for EfficientNet B0. (b) Loss for EfficientNet B1. (c) Loss for EfficientNet B2. (d) Loss for EfficientNet B3. (e) Loss for EfficientNet B4. (f) Loss for EfficientNet B5. (g) Loss for EfficientNet B6. (h) Loss for EfficientNet B7. (i) Loss for EfficientNet B8.
Performance of EfficientNet models.
| Epoch | EfficientNet B0 | EfficientNet B1 | EfficientNet B2 | EfficientNet B3 | EfficientNet B4 | EfficientNet B5 | EfficientNet B6 | EfficientNet B7 | EfficientNet B8 | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | Accuracy | Loss | |
| 1 | 95 | 5 | 97 | 3 | 96 | 4 | 97 | 3 | 98 | 2 | 98 | 2 | 98 | 2 | 99 | 1 | 97 | 3 |
| 10 | 97 | 3 | 96 | 4 | 95 | 5 | 97 | 3 | 96 | 4 | 98 | 2 | 97 | 3 | 96 | 4 | 99 | 1 |
| 20 | 97 | 3 | 96 | 4 | 96 | 4 | 96 | 4 | 98 | 2 | 98 | 2 | 99 | 1 | 99 | 1 | 100 | 0 |
| 30 | 95 | 5 | 96 | 4 | 97 | 3 | 97 | 3 | 98 | 2 | 98 | 2 | 99 | 1 | 97 | 3 | 99 | 1 |
| 40 | 97 | 3 | 97 | 3 | 96 | 4 | 96 | 4 | 97 | 3 | 96 | 4 | 97 | 3 | 99 | 1 | 100 | 0 |
| 50 | 97 | 3 | 96 | 4 | 95 | 5 | 96 | 4 | 96 | 4 | 98 | 2 | 98 | 2 | 96 | 4 | 98 | 2 |
| 60 | 97 | 3 | 96 | 4 | 97 | 3 | 96 | 4 | 96 | 4 | 96 | 4 | 99 | 1 | 99 | 1 | 97 | 3 |
| 70 | 97 | 3 | 97 | 3 | 96 | 4 | 96 | 4 | 96 | 4 | 96 | 4 | 98 | 2 | 96 | 4 | 98 | 2 |
| 80 | 97 | 3 | 97 | 3 | 96 | 4 | 96 | 4 | 98 | 2 | 97 | 3 | 98 | 2 | 99 | 1 | 99 | 1 |
| 90 | 97 | 3 | 97 | 3 | 96 | 4 | 96 | 4 | 97 | 3 | 96 | 4 | 99 | 1 | 98 | 2 | 98 | 2 |
| 100 | 95 | 5 | 97 | 3 | 96 | 4 | 96 | 4 | 96 | 4 | 97 | 3 | 99 | 1 | 96 | 4 | 98 | 2 |
Figure 7Proposed method compared with the related studies.
Proposed method compared with the related studies.
| Authors | Result |
|---|---|
| Emeršič et al. [ | 30 |
| Tian and Mu [ | 69.33 |
| Raveane et al. [ | 98 |
| Zhang and Mu [ | 99.11 |
| Kohlakala and Coetzer [ | 95.63 |
| Tomczyk and Szczepaniak [ | NA |
| Alshazly et al. [ | 22 |
| Alkababji and Mohammed [ | 97.8 |
| Jamil et al. [ | 97 |
| Hansley et al. [ | NA |
| Average of our work | 97.07 |