| Literature DB >> 32150911 |
Ahmed E Khorshid1, Ibrahim N Alquaydheb1, Fadi Kurdahi1, Roger Piqueras Jover2, Ahmed Eltawil1,3.
Abstract
In this paper, we propose and validate using the Intra-body communications channel as a biometric identity. Combining experimental measurements collected from five subjects and two multi-layer tissue mimicking materials' phantoms, different machine learning algorithms were used and compared to test and validate using the channel characteristics and features as a biometric identity for subject identification. An accuracy of 98.5% was achieved, together with a precision and recall of 0.984 and 0.984, respectively, when testing the models against subject identification over results collected from the total samples. Using a simple and portable setup, this work shows the feasibility, reliability, and accuracy of the proposed biometric identity, which allows for continuous identification and verification.Entities:
Keywords: body area networks; channel gain/attenuation; channel modeling; galvanic coupling; intra-body communications; phantoms; tissue mimicking materials; ultralow power systems
Mesh:
Year: 2020 PMID: 32150911 PMCID: PMC7085539 DOI: 10.3390/s20051421
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Experimental setup, where miniVNA (Vector Network Analyzer) is used to measure the channel gain/attenuation profile for different transmitter (TX)–receiver nodes (RX) separations and configurations.
Figure 2The constructed five-tissue layers arm phantom model in a measurement’s setup scenario (connected to the miniVNA).
Figure 3Intra-Body Communication (IBC) channel gain, for a human subject vs for a phantom, when varying the distance between the transmitter and the receiver to be at 10, 15, and 20 cm and separation between electrodes of each node is 5 cm.
Performance metrics for TX–RX (Transmitter–Receiver) of 10 cm features (632 features per case).
| Classifier | Accuracy | Precision | Recall | F-Measure |
|---|---|---|---|---|
|
| 98.8372% | 0.989 | 0.988 | 0.988 |
|
| 95.5% | 0.962 | 0.955 | 0.955 |
|
| 100% | 1 | 1 | 1 |
|
| 96% | 0.962 | 0.960 | 0.959 |
|
| 92% | 0.939 | 0.920 | 0.923 |
Performance metrics for TX–RX of 15 cm features (632 features per case).
| Classifier | Accuracy | Precision | Recall | F-Measure |
|---|---|---|---|---|
|
| 91.9811% | 0.920 | 0.920 | 0.920 |
|
| 90.566% | 0.909 | 0.906 | 0.907 |
|
| 91.9811% | 0.921 | 0.920 | 0.920 |
|
| 91.9811% | 0.921 | 0.920 | 0.920 |
|
| 81.6038% | 0.827 | 0.816 | 0.818 |
Performance metrics for TX–RX of 20 cm features (632 features per case).
| Classifier | Accuracy | Precision | Recall | F-Measure |
|---|---|---|---|---|
|
| 71.6102% | 0.740 | 0.716 | 0.691 |
|
| 80.5085% | 0.838 | 0.805 | 0.812 |
|
| 83.8983% | 0.750 | 0.581 | 0.655 |
|
| 79.661% | 0.793 | 0.797 | 0.782 |
|
| 78.3898% | 0.606 | 0.645 | 0.625 |
Performance metrics for all TX–RX as different cases (632 features per case).
| Classifier | Accuracy | Precision | Recall | F-Masure |
|---|---|---|---|---|
|
| 85.9756% | 0.844 | 0.609 | 0.708 |
|
| --- | ---- | ---- | ----- |
|
| 89.4817% | 0.894 | 0.895 | 0.894 |
|
| 87.0427% | 0.870 | 0.870 | 0.868 |
|
| 82.3171% | 0.833 | 0.823 | 0.826 |
Performance metrics for all TX–RX as different cases (1896 features per case).
| Classifier | Accuracy | Precision | Recall | F-Measure |
|---|---|---|---|---|
|
| 97.9695% | 0.981 | 0.980 | 0.980 |
|
| ---- | - | - | - |
|
| 96.9543% | 0.970 | 0.970 | 0.970 |
|
| 97.9695% | 0.981 | 0.980 | 0.980 |
|
| 92.3858% | 0.941 | 0.924 | 0.926 |
Performance metrics for the power bin approach (0.5 MHz bin size).
| Classifier | Accuracy | Precision | Recall | F-Measure |
|---|---|---|---|---|
|
| 98.4127% | 0.984 | 0.984 | 0.984 |
|
| 94.9206% | 0.963 | 0.949 | 0.952 |
|
| 97.7778% | 0.978 | 0.978 | 0.978 |
|
| 87.619% | 0.892 | 0.876 | 0.863 |
|
| 47.9365% | 0.750 | 0.479 | 0.512 |
Figure 4Accuracy for different classifiers versus power bin size.
Figure 5Accuracy for different classifiers versus the number of features.
Summary for the different features selection approaches and best performance result for each.
| Description | Recommended Model and Performance | Comments | |
|---|---|---|---|
|
| Using the magnitude of channel response at different frequencies as features, considering only the best separation distance | Naïve Bayes (98.8372%) | - Achieves high accuracy with simple models, in case of the 10 cm separation. |
|
| Same as approach 1, but all the separation distances are used as different training/test cases for the same run (number of train/test cases are 3× that of approach 1) | KNN | - Simple models, yet least accuracy among all approaches |
|
| Same as approach one, but channel response at different frequencies and separation distances are all combined as features (3× number of features compared to approach 1 and 2) | Naïve Bayes | - More complex models (3× number of features) and more computational resources needed. |
|
| Use of power Bins as features (integration of power across a frequency spectrum (bin)) | Naïve Bayes (98.4127%) | - High accuracy |