| Literature DB >> 31284418 |
Pedro Lopes Silva1, Eduardo Luz1, Gladston Moreira2, Lauro Moraes1, David Menotti3.
Abstract
Multimodal systems are a workaround to enhance the robustness and effectiveness of biometric systems. A proper multimodal dataset is of the utmost importance to build such systems. The literature presents some multimodal datasets, although, to the best of our knowledge, there are no previous studies combining face, iris/eye, and vital signals such as the Electrocardiogram (ECG). Moreover, there is no methodology to guide the construction and evaluation of a chimeric dataset. Taking that fact into account, we propose to create a chimeric dataset from three modalities in this work: ECG, eye, and face. Based on the Doddington Zoo criteria, we also propose a generic and systematic protocol imposing constraints for the creation of homogeneous chimeric individuals, which allow us to perform a fair and reproducible benchmark. Moreover, we have proposed a multimodal approach for these modalities based on state-of-the-art deep representations built by convolutional neural networks. We conduct the experiments in the open-world verification mode and on two different scenarios (intra-session and inter-session), using three modalities from two datasets: CYBHi (ECG) and FRGC (eye and face). Our multimodal approach achieves impressive decidability of 7.20 ± 0.18, yielding an almost perfect verification system (i.e., Equal Error Rate (EER) of 0.20% ± 0.06) on the intra-session scenario with unknown data. On the inter-session scenario, we achieve a decidability of 7.78 ± 0.78 and an EER of 0.06% ± 0.06. In summary, these figures represent a gain of over 28% in decidability and a reduction over 11% of the EER on the intra-session scenario for unknown data compared to the best-known unimodal approach. Besides, we achieve an improvement greater than 22% in decidability and an EER reduction over 6% in the inter-session scenario.Entities:
Keywords: Doddington Zoo; ECG; biometric systems; chimeric dataset; deep learning; deep representation; eye; face; multimodal biometrics
Year: 2019 PMID: 31284418 PMCID: PMC6651892 DOI: 10.3390/s19132968
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Example of the commercial off-the person mobile equipment used to capture the ECG signal. Source: https://www.alivecor.com/en/.
Figure 2Example of the device and the ECG from the CYBHi dataset. (a) Example of the device used to get the ECG signal. Source: [20]. (b) Example of the ECG shape extracted from the CYBHi dataset.
Figure 3Example of FRGC face images. Source: [24].
Figure 4Recognition process of a chimeric individual.
Figure 5The process to combine modalities and create the chimeric individual is a stochastic process, respecting the constraints presented in Section 4.3. The figure illustrates the process of creating chimeric individuals of one specific Doddington Zoo category.
Figure 6Example of the crops made from a random image of the FRGC dataset. Source: [24].
Figure 7CNN proposed in [10] aiming at ECG recognition for the verification task.TIFS ECG
Figure 8Periocular Region Recognition (PRR)-256 proposed in [2] aiming at periocular recognition for the verification task.VGG
Comparison of Proença [26] and our approach in the FRGC dataset ran 30 times. The mean and standard deviation are reported.
| Method | Decidability | EER |
|---|---|---|
| Proença [ | 2.97 ± 0.04 | - |
| Proposed approach | 3.62 ± 0.05 | 5.11% ± 0.23 |
Resulting EER and decidability from 30 executions in verification mode of all modalities’ combinations explored by different rules. (con = simple concatenation; min = minimal; mul = multiplication). The best results are highlighted in red. There was no statistical significance among the figures in red; however, they were significantly better than the other figures.
| Method | Rule | Intra-Session in Training Data | Intra-Session in Test Data | Inter-Session | |||
|---|---|---|---|---|---|---|---|
| Decidability | EER (%) | Decidability | EER (%) | Decidability | EER (%) | ||
| ECG | - | 6.61 ± 0.19 | 0.94 ± 0.14 | 5.73 ± 0.21 | 2.16 ± 0.26 | 5.70 ± 0.93 | 1.92 ± 1.23 |
| Eye | - | 5.01 ± 0.10 | 1.43 ± 0.18 | 3.62 ± 0.12 | 5.16 ± 0.47 | 4.32 ± 0.34 | 3.05 ± 0.84 |
| Face | - | 5.46 ± 0.12 | 0.67 ± 0.12 | 4.71 ± 0.12 | 1.57 ± 0.24 | 5.37 ± 0.57 | 0.93 ± 0.46 |
| ECG + Eye | min | 7.24 ± 0.11 | 0.46 ± 0.08 | 6.54 ± 0.20 | 1.16 ± 0.15 | 6.35 ± 0.66 | 1.00 ± 0.49 |
| mult | 6.49 ± 0.10 | 0.08 ± 0.02 | 5.97 ± 0.10 | 0.49 ± 0.10 | 6.10 ± 0.22 | 0.28 ± 0.19 | |
| sum | 7.62 ± 0.12 | 0.07 ± 0.02 | 5.86 ± 0.16 | 0.58 ± 0.11 | 6.51 ± 0.48 | 0.30 ± 0.17 | |
| con | 7.55 ± 0.19 | 0.43 ± 0.08 | 6.41 ± 0.23 | 1.34 ± 0.15 | 6.50 ± 0.97 | 1.06 ± 0.73 | |
| ECG + Face | min | 6.89 ± 0.15 | 0.62 ± 0.10 | 6.18 ± 0.21 | 1.41 ± 0.17 | 6.01 ± 0.75 | 1.21 ± 0.56 |
| mult | 7.44 ± 0.11 | 0.08 ± 0.03 | 6.94 ± 0.11 | 0.36 ± 0.08 | 6.97 ± 0.39 | 0.21 ± 0.14 | |
| sum | 8.42 ± 0.18 | 0.04 ± 0.02 | 7.30 ± 0.18 | 0.20 ± 0.06 | 7.78 ± 0.78 | 0.10 ± 0.09 | |
| con | 7.33 ± 0.21 | 0.44 ± 0.10 | 6.31 ± 0.24 | 1.31 ± 0.16 | 6.41 ± 1.04 | 1.06 ± 0.71 | |
| Eye + Face | min | 5.99 ± 0.10 | 0.48 ± 0.06 | 4.38 ± 0.10 | 2.10 ± 0.22 | 5.29 ± 0.43 | 1.10 ± 0.42 |
| mult | 5.86 ± 0.11 | 0.19 ± 0.04 | 4.72 ± 0.10 | 1.76 ± 0.26 | 5.38 ± 0.27 | 0.73 ± 0.36 | |
| sum | 6.66 ± 0.11 | 0.18 ± 0.04 | 4.60 ± 0.12 | 1.80 ± 0.27 | 5.81 ± 0.52 | 0.72 ± 0.35 | |
| con | 6.56 ± 0.11 | 0.17 ± 0.05 | 4.67 ± 0.11 | 1.27 ± 0.21 | 5.93 ± 0.60 | 0.50 ± 0.29 | |
| ECG + Eye + Face | min | 7.35 ± 0.11 | 0.36 ± 0.07 | 6.63 ± 0.19 | 1.06 ± 0.12 | 6.47 ± 0.60 | 0.82 ± 0.37 |
| mult | 5.63 ± 0.10 | 0.01 ± 0.00 | 5.28 ± 0.09 | 0.18 ± 0.04 | 5.47 ± 0.13 | 0.06 ± 0.06 | |
| sum | 8.73 ± 0.12 | 0.01 ± 0.01 | 6.31 ± 0.15 | 0.24 ± 0.05 | 7.56 ± 0.59 | 0.07 ± 0.06 | |
| con | 7.95 ± 0.22 | 0.26 ± 0.06 | 6.68 ± 0.24 | 0.90 ± 0.11 | 6.92 ± 1.07 | 0.65 ± 0.44 | |
Figure 9Distribution impostor/genuine curves from the inter-session scenario. The fusion is at the score level using the sum rule.