| Literature DB >> 36081089 |
Navneet Singh1, Sangho Choe1, Rajiv Punmiya1, Navneesh Kaur2.
Abstract
Location-based indoor applications with high quality of services require a reliable, accurate, and low-cost position prediction for target device(s). The widespread availability of WiFi received signal strength indicator (RSSI) makes it a suitable candidate for indoor localization. However, traditional WiFi RSSI fingerprinting schemes perform poorly due to dynamic indoor mobile channel conditions including multipath fading, non-line-of-sight path loss, and so forth. Recently, machine learning (ML) or deep learning (DL)-based fingerprinting schemes are often used as an alternative, overcoming such issues. This paper presents an extreme gradient boosting-based ML indoor localization scheme, simply termed as XGBLoc, that accurately classifies (or detects) the positions of mobile devices in multi-floor multi-building indoor environments. XGBLoc not only effectively reduces the RSSI dataset dimensionality but trains itself using structured synthetic labels (also termed as relational labels), rather than conventional independent labels, that classify such complex and hierarchical indoor environments well. We numerically evaluate the proposed scheme on the publicly available datasets and prove its superiority over existing ML or DL-based schemes in terms of classification and regression performance.Entities:
Keywords: RSSI fingerprints; WiFi; XGBoost; classification; hyper-parameter tuning; indoor localization; labeling; regression
Year: 2022 PMID: 36081089 PMCID: PMC9459943 DOI: 10.3390/s22176629
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Summary of recent related works. [Note] P Dataset: public dataset used, H tuning: hyperparameter tuning required, B/F Classification: building and floor classification.
| Ref. | P Dataset | Labels | Techniques Used | H Tuning | B/F Classification |
|---|---|---|---|---|---|
| [ | ✓ | NRL | RNN | Extensive | ✓ |
| [ | ✓ | NRL | RNN | Extensive | ✓ |
| [ | ✓ | NRL | SAE, CNN | Extensive | ✓ |
| [ | ✓ | NRL | DNN | Extensive | ✓ |
| [ | ✓ | NRL | 2D Radio map, CNN | Extensive | ✓ |
| [ | ✓ | NRL | m-CELL, 2D CNN | Extensive | ✓ |
|
| ✓ |
|
|
| ✓ |
Figure 1Conventional non-relational labeling (NRL) [left] and proposed relational labeling (RL) [right] for a multi-building multi-floor environment example.
Figure 2Architecture of proposed system model XGBLoc.
Elements of fingerprints in UJIIndoorLoc.
| Elements of Fingerprints ( | Description |
|---|---|
|
| |
|
| Longitudinal values of location in meters. |
|
| Latitudinal values of location in meters. |
|
| Floor ID. |
|
| Building ID. |
Symbols used for defining objective function of proposed ML model.
| Symbols | Description |
|---|---|
|
| Fingerprint dataset. |
|
| Total number of samples in dataset. |
|
| Total number of dimmensions/WAPs. |
|
| RSSI vector at |
|
| Position of |
|
| Coordinates of the |
|
| Total number of trees. |
|
| Predicted position of |
|
| Predicted coordinates of the |
|
| Loss function. |
|
| Regularization term of the |
XGBLoc hyperparameters.
| Hyperparameter | Value | Description |
|---|---|---|
| learning_rate or eta | 0.3 | Weighting factor for learning in gradient boosting. |
| gamma | 1 | Minimum loss reduction needed to render partition |
| max_depth | 6 | Maximum depth of tree. |
| colsample_bytree | 1 | Subsample ratio of columns when constructing each |
| lambda | 1 | L2 regularization term on weights. |
| loss function | multi:softprob | Multiclass classification problem. |
| n_estimators | 100 | Number of trees to be generated. |
| scale_pos_weight | 1 | Control the balance of positive and negative weights. |
| booster | gbtree | Use tree based model. |
| tree_method | gpu_hist | GPU implementation of faster histogram optimized |
| Subsample | 1 | Subsample ratio of the training samples. |
Figure 3Sparsity in RSSI values.
Figure 4Explained variance ratio (EVR) vs number of principal components for the UJIIndoorLoc dataset.
Effect of explained variance ratio (EVR) on localization performance with different input and output specification.
| Explained Variance Ratio (EVR) | 0.7 | 0.8 | 0.9 |
|---|---|---|---|
| Classification Accuracy | 98% | 98.6% | 99.2% |
| 2-D Mean position error ( | 5.2 m | 4.93 m | 5.01 m |
Figure 5Classification accuracy vs. learning rate vs. n_estimators.
Effect of hyperparameter tuning on localization performance. [Note] lr: learning_rate, md: max_depth, cb: colsample_bytree, lf: loss function, ne: n_estimators.
| Task | Hyperparameter | Output | |||||||
|---|---|---|---|---|---|---|---|---|---|
| lr | gamma | md | cb | lambda | lf | ne | Subsample | ||
|
| 0.1 | 1 | 6 | 0.9 | 0.8 | multi:softprob | 500 | 0.8 |
|
|
| 0.1 | 0 | 10 | 0.8 | 0.9 | reg:squarederror | 1000 | 0.8 |
|
Classification results of existing and proposed schemes on the UJIIndoorLoc dataset.
| WiFi Fingerprint-Based Schemes | Classification Accuracy | |
|---|---|---|
| Building | Floor | |
| MOSAIC | 98.5% | 93.83% |
| 1-KNN | 100% | 89.95% |
| 13-KNN | 100% | 95.17% |
| DNN | 100% | 91.97% |
| 2D-DNN | 100% | 95.64% |
| Scalable DNN | 99.5% | 91.26% |
| CNNLoc | 100% | 96.03% |
|
|
|
|
Regression results of existing and proposed schemes on the UJIIndoorLoc dataset.
| WiFi Fingerprint-Based Schemes | 2-D Average Positioning Error (m) |
|---|---|
| KNN | 7.9 |
| WKNN | 6.2 |
| HybLoc | 6.46 |
| RF | 10.2 |
| CNNLoc | 11.78 |
| CCpos | 12.4 |
|
|
|
Classification results of existing and proposed schemes on the Tampere dataset.
| WiFi Fingerprint-Based Schemes | Classification Results |
|---|---|
| Weighted Centroid | 83.18% |
| Log-Gaussian Probability | 85.30% |
| RSS Clustering | 90.79% |
| UJI KNN | 92.97% |
| RTLS@UM | 90.03% |
| Rank-based | 86.48% |
| Coverage Area-based | 86.56% |
| CNNLoc | 94.12% |
|
|
|
Regression results of existing and proposed schemes on the Alcala dataset.
| WiFi Fingerprint-Based Schemes | 2-D Average Positioning Error (m) |
|---|---|
| KNN | 2.62 |
| WKNN | 2.27 |
| SVM | 6.71 |
| RF | 2.53 |
| CNNLoc | 4.62 |
| CCpos | 1.05 |
|
|
|