| Literature DB >> 29748495 |
Muhammad Arsalan1, Rizwan Ali Naqvi2, Dong Seop Kim3, Phong Ha Nguyen4, Muhammad Owais5, Kang Ryoung Park6.
Abstract
The recent advancements in computer vision have opened new horizons for deploying biometric recognition algorithms in mobile and handheld devices. Similarly, iris recognition is now much needed in unconstraint scenarios with accuracy. These environments make the acquired iris image exhibit occlusion, low resolution, blur, unusual glint, ghost effect, and off-angles. The prevailing segmentation algorithms cannot cope with these constraints. In addition, owing to the unavailability of near-infrared (NIR) light, iris recognition in visible light environment makes the iris segmentation challenging with the noise of visible light. Deep learning with convolutional neural networks (CNN) has brought a considerable breakthrough in various applications. To address the iris segmentation issues in challenging situations by visible light and near-infrared light camera sensors, this paper proposes a densely connected fully convolutional network (IrisDenseNet), which can determine the true iris boundary even with inferior-quality images by using better information gradient flow between the dense blocks. In the experiments conducted, five datasets of visible light and NIR environments were used. For visible light environment, noisy iris challenge evaluation part-II (NICE-II selected from UBIRIS.v2 database) and mobile iris challenge evaluation (MICHE-I) datasets were used. For NIR environment, the institute of automation, Chinese academy of sciences (CASIA) v4.0 interval, CASIA v4.0 distance, and IIT Delhi v1.0 iris datasets were used. Experimental results showed the optimal segmentation of the proposed IrisDenseNet and its excellent performance over existing algorithms for all five datasets.Entities:
Keywords: convolutional neural network (CNN); iris recognition; iris segmentation; semantic segmentation; visible light and near-infrared light camera sensors
Year: 2018 PMID: 29748495 PMCID: PMC5981870 DOI: 10.3390/s18051501
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Comparisons between the proposed and existing methods for iris segmentation.
| Type | Methods | Strength | Weakness |
|---|---|---|---|
|
| Iris localization by circular HT [ | These methods show a good estimation of the iris region in ideal cases | These types of methods are not very accurate for non-ideal cases or visible light environments |
| Integro-differential operator [ | |||
| Iris localization by gradient on iris-sclera boundary points [ | A new idea to use a gradient to locate iris boundary | The gradient is affected by eyelashes and true iris boundary is not found | |
| The two-stage method with circular moving window [ | Pupil based on dark color approximated simply by probability | Calculating gradient in a search way is time-consuming | |
| Radial suppression-based edge detection and thresholding [ | Radial suppression makes the case simpler for the iris edges | In non-ideal cases, the edges are not fine to estimate the boundaries | |
| Adaptive thresholding and first derivative-based iris localization [ | Simple way to obtain the boundary based on the gray level in ideal cases | One threshold cannot guarantee good results in all cases | |
|
| Two-circular edge detector assisted with AdaBoost eye detector [ | Closed eye, eyelash and eyelid detection is executed to reduce error | The method is affected by pupil/eyelid detection error |
| Curve fitting and color clustering [ | Upper and lower eyelid detections are performed to reduce the error | The empirical threshold is set for eyelid and eyelash detection, and still true boundary is not found | |
| Graph-cut-based approach for iris segmentation [ | Eyelashes are removed using Markov random field to reduce error | A separate method for each eyelash, pupil, iris detection is time-consuming | |
| Rotated ellipse fitting method combined with occlusion detection [ | Ellipse fitting gives a good approximation for the iris with reduced error | Still, the iris and other boundaries are considered as circular | |
| Three model fusion-based method assisted with Daugman’s method [ | Simple integral derivative as a base for iris boundaries is a quite simple way | High-score fitting is sensitive in ideal cases, and can be disturbed by similar RGB pixels in the image | |
|
| Geodesic active contours, Chan–Vese and new pressure force active contours [ | These methods iteratively approximate the true boundaries in non-ideal situations | In these methods, many iterations are required for accuracy, which takes much processing time |
|
| Region growing with integro-differential constellation [ | Both iris and non-iris regions are identified along with reflection removal to reduce error | The rough boundary is found first and then a boundary refinement process is performed separately |
| Region growing with binary integrated intensity curve-based method [ | Eyelash and eyelid detection is performed along with iris segmentation | The region growing process starts with the pupil circle, so the case of visible light images where the pupil is not clear can cause errors | |
| Watershed BIRD with seed selection [ | Limbus boundary detection is performed to separate sclera, eyelashes, and eyelid pixels from iris | Watershed transform shares the disadvantage of over-segmentation, so circle fitting is used further | |
|
| HCNNs and MFCNs [ | This approach shows the lower error than existing methods for non-ideal cases | The similar parts to iris regions can be incorrectly detected as iris points |
| Two-stage iris segmentation method using deep learning and modified HT [ | Better accuracy due to CNN, which is just applied inside the ROI defined in the first stage | Millions of 21 × 21 images are needed for CNN training and pre-processing required to improve the image | |
| IrisDenseNet for iris segmentation | Accurately find the iris boundaries without pre-processing with better information gradient flow. With robustness for high-frequency areas such as eyelashes and ghost regions | Due to dense connectivity, the mini-batch size should be kept low owing to more time required for training |
Figure 1Flowchart of the proposed method.
Figure 2Overview of the proposed method.
Figure 3Dense connectivity within the dense block by feature concatenation.
Figure 4Overall connectivity diagram of IrisDenseNet dense encoder–decoder.
IrisDenseNet connectivity and output feature map size of each dense block (Conv, BN, and ReLU represent convolutional layer, batch normalization layer, and rectified linear unit layer, respectively. Cat, B-Conv, and Pool indicate concatenation layer, bottleneck convolution layer, and pooling layer, respectively) (Here, dense blocks 1 and 2 have the same number of convolution layers, and dense blocks 3, 4, and 5 have the same number of convolution layers) (Convolutional layers with “*” mean that these layers include BN and ReLU. Transition layers are a combination of max-pooling and B-Conv)
| Block | Name/Size | No. of Filters | Output Feature Map Size |
|---|---|---|---|
|
| Conv-1_1*/3 × 3 × 3 | 64 | 300 × 400 × 64 |
| Conv-1_2*/3 × 3 × 64 | 64 | ||
| Cat-1 | - | 300 × 400 × 128 | |
|
| B-Conv-1/1 × 1 | 64 | 300 × 400 × 64 |
| Pool-1/2 × 2 | - | 150 × 200 × 64 | |
|
| Conv-2_1*/3 × 3 × 64 | 128 | 150 × 200 × 128 |
| Conv-2_2*/3 × 3 × 128 | 128 | ||
| Cat-2 | - | 150 × 200 × 256 | |
|
| B-Conv-2/1 × 1 | 128 | 150 × 200 × 128 |
| Pool-2/2 × 2 | - | 75 × 100 × 128 | |
|
| Conv-3_1*/3 × 3 × 128 | 256 | 75 × 100 × 256 |
| Conv-3_2*/3 × 3 × 256 | 256 | ||
| Cat-3 | - | 75 × 100 × 512 | |
| Conv-3_3*/3 × 3 × 256 | 256 | 75 × 100 × 256 | |
| Cat-4 | - | 75 × 100 × 768 | |
|
| B-Conv-3/1 × 1 | 256 | 75 × 100 × 256 |
| Pool-3/2 × 2 | - | 37 ×50 × 256 | |
|
| Conv-4_1*/3 × 3 × 256 | 512 | 37 ×50 × 512 |
| Conv-4_2*/3 × 3 × 512 | 512 | ||
| Cat-5 | - | 37 ×50 × 1024 | |
| Conv-4_3*/3 × 3 × 512 | 512 | 37 ×50 × 512 | |
| Cat-6 | - | 37 ×50 × 1536 | |
|
| B-Conv-4/1 × 1 | 512 | 37 ×50 × 512 |
| Pool-4/2 × 2 | - | 18 × 25 × 512 | |
|
| Conv-5_1*/3 × 3 × 512 | 512 | 18 × 25 × 512 |
| Conv-5_2*/3 × 3 × 512 | 512 | ||
| Cat-7 | - | 18 × 25 × 1024 | |
| Conv-5_3*/3 × 3 × 512 | 512 | 18 × 25 × 512 | |
| Cat-8 | - | 18 × 25 × 1536 | |
|
| B-Conv-5/1 × 1 | 512 | 18 × 25 × 512 |
| Pool-5/2 × 2 | - | 9 × 12 × 512 |
Figure 5Noisy iris challenge evaluation part-II (NICE-II) sample images with corresponding ground truths.
Figure 6Difference in frequency between iris and non-iris classes. (a) NICE-II original input image. (b) Difference in frequency of iris and non-iris pixels in NICE-II training dataset.
Figure 7Training accuracy and loss curves from (a) 1st-fold cross-validation and (b) 2nd-fold cross-validation.
Comparative accuracies according to various augmentation methods.
| Method |
|
|---|---|
| Excessive data augmentation | 0.00729 |
| Data augmentation by changing the contrast and brightness of iris image | 0.00761 |
| Proposed data augmentation in | 0.00695 |
Figure 8Examples of NICE-II good segmentation results obtained by IrisDenseNet. (a) Original image. (b) Ground-truth image. (c) Segmentation result obtained by IrisDenseNet (The false positive and negative errors are shown in green and red, respectively. The true positive case is shown in black).
Figure 9Examples of incorrect iris segmentation by our method. (a) Original input images. (b) Ground-truth images. (c) Segmentation results (The false positive and negative errors are presented as green and red, respectively. The true positive case is shown in black).
Comparisons of the proposed method with previous methods using NICE-II dataset.
| Method |
|
|---|---|
| Luengo-Oroz et al. [ | 0.0305 |
| Labati et al. [ | 0.0301 |
| Chen et al. [ | 0.029 |
| Jeong et al. [ | 0.028 |
| Li et al. [ | 0.022 |
| Tan et al. [ | 0.019 |
| Proença et al. [ | 0.0187 |
| de Almeida [ | 0.0180 |
| Tan et al. [ | 0.0172 |
| Sankowski et al. [ | 0.016 |
| Tan et al. [ | 0.0131 |
| Haindl et al. [ | 0.0124 |
| Zhao et al. [ | 0.0121 |
| Arsalan et al. [ | 0.0082 |
| SegNet-Basic [ | 0.00784 |
| Proposed IrisDenseNet | 0.00695 |
Figure 10Mobile iris challenge evaluation (MICHE-I) sample images with corresponding ground truths.
Figure 11The institute of automation, Chinese academy of sciences (CASIA) v4.0 interval sample images with corresponding ground truths.
Figure 12CASIA v4.0 distance sample images with corresponding ground truths.
Figure 13IIT Delhi (IITD) v1.0 sample images with corresponding ground truths.
Figure 14Examples of correct segmentation results in MICHE-I database by our method. (a) Original image. (b) Ground-truth image. (c) Segmentation result by IrisDenseNet (The false and negative errors are presented as green and red, respectively. The true positive case is presented as black).
Figure 15Examples of correct segmentation results in CASIA v4.0 interval database by the proposed method. (a) Original image. (b) Ground-truth image. (c) Segmentation result by IrisDenseNet (The false positive and negative errors are presented as green and red, respectively. The true positive case is presented as black).
Figure 16Examples of correct segmentation results in CASIA v4.0 distance database by the proposed method. (a) Original image (b) Ground-truth image. (c) Segmentation result by IrisDenseNet (The false positive and negative errors are presented as green and red, respectively. The true positive case is presented as black).
Figure 17Examples of correct segmentation results in IITD database by the proposed method. (a) Original image. (b) Ground-truth image. (c) Segmentation result by IrisDenseNet (The false positive and negative errors are presented as green and red, respectively. The true positive case is presented as black).
Figure 18Examples of MICHE-I incorrect segmentation results by our method. (a) Original image. (b) Ground-truth image. (c) Segmentation result by IrisDenseNet (The false positive and negative errors are presented as green and red, respectively. The true positive case is presented as black).
Comparison of the proposed method with previous methods using MICHE-I dataset based on NICE-I evaluation protocol.
| Method |
| |
|---|---|---|
| Hu et al. [ | Sub-dataset by iPhone5 | 0.0193 |
| Sub-dataset by Galaxy S4 | 0.0192 | |
| Arsalan et al. [ | Sub-dataset by iPhone5 | 0.00368 |
| Sub-dataset by Galaxy S4 | 0.00297 | |
| Sub-dataset by Galaxy Tab2 | 0.00352 | |
| SegNet-Basic [ | Sub-dataset by iPhone5 | 0.0025 |
| Sub-dataset by Galaxy S4 | 0.0027 | |
| Sub-dataset by Galaxy Tab2 | 0.0029 | |
| Proposed IrisDenseNet | Sub-dataset by iPhone5 | 0.0020 |
| Sub-dataset by Galaxy S4 | 0.0022 | |
| Sub-dataset by Galaxy Tab2 | 0.0021 | |
Comparison of the proposed method with previous methods using CASIA v4.0 distance database based on NICE-I evaluation protocol.
| Method |
|
|---|---|
| Tan et al. [ | 0.0113 |
| Liu et al. [ | 0.0108 |
| Tan et al. [ | 0.0081 |
| Zhao et al. [ | 0.0068 |
| Liu et al. [ | 0.0059 |
| SegNet-Basic [ | 0.0044 |
| Proposed IrisDenseNet | 0.0034 |
Comparison of the proposed method with previous methods using CASIA v4.0 interval and IITD databases based on the RPF-measure evaluation protocol. A smaller value of σ and a higher value of µ show better performance. (unit: %) (The resultant values of GST [85], Osiris [86], WAHET [87], IFFP [88], CAHT [89], Masek [90], IDO [25], and IrisSeg [84] are referred from [84]).
| DB | Method | R | P | F | |||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
| ||
|
| GST [ | 85.19 | 18 | 89.91 | 7.37 | 86.16 | 11.53 |
| Osiris [ | 97.32 | 7.93 | 93.03 | 4.95 | 89.85 | 5.47 | |
| WAHET [ | 94.72 | 9.01 | 85.44 | 9.67 | 89.13 | 8.39 | |
| IFFP [ | 91.74 | 14.74 | 83.5 | 14.26 | 86.86 | 13.27 | |
| CAHT [ | 97.68 | 4.56 | 82.89 | 9.95 | 89.27 | 6.67 | |
| Masek [ | 88.46 | 11.52 | 89 | 6.31 | 88.3 | 7.99 | |
| IDO [ | 71.34 | 22.86 | 61.62 | 18.71 | 65.61 | 19.96 | |
| IrisSeg [ | 94.26 | 4.18 | 92.15 | 3.34 | 93.1 | 2.65 | |
| SegNet-Basic [ | 99.60 | 0.66 | 91.86 | 2.65 | 95.55 | 1.40 | |
| Proposed Method | 97.10 | 2.12 | 98.10 | 1.07 | 97.58 | 0.99 | |
|
| GST [ | 90.06 | 16.65 | 85.86 | 10.46 | 86.6 | 11.87 |
| Osiris [ | 94.06 | 6.43 | 91.01 | 7.61 | 92.23 | 5.8 | |
| WAHET [ | 97.43 | 8.12 | 79.42 | 12.41 | 87.02 | 9.72 | |
| IFFP [ | 93.92 | 10.62 | 79.76 | 11.42 | 85.83 | 9.54 | |
| CAHT [ | 96.8 | 11.2 | 78.87 | 13.25 | 86.28 | 11.39 | |
| Masek [ | 82.23 | 18.74 | 90.45 | 11 .85 | 85.3 | 15.39 | |
| IDO [ | 51.91 | 15.32 | 52.23 | 14.85 | 51.17 | 13.26 | |
| IrisSeg [ | 95.33 | 4.58 | 93.70 | 5.33 | 94.37 | 3.88 | |
| SegNet-Basic [ | 99.68 | 0.51 | 92.53 | 2.05 | 95.96 | 1.04 | |
| Proposed Method | 98.0 | 1.56 | 97.16 | 1.40 | 97.56 | 0.84 | |
Figure 19SegNet-Basic last 64 channel (the 449th to 512th channels) features before the 4th max-pooling (Pool-4).
Figure 20IrisDenseNet last 64 channel (the 449th to 512th channels) features for the 4th dense block before the 4th max-pooling (Pool-4 of Table 2).
Figure 21Comparison of iris thin boundary. Segmentation results obtained by (a) IrisDenseNet and (b) SegNet.
Figure 22Comparisons of pupil boundary detection. Segmentation results obtained by (a) IrisDenseNet and (b) SegNet.
Figure 23Comparisons of iris detection affected by ghost effect. Segmentation results obtained by (a) IrisDenseNet and (b) SegNet.