| Literature DB >> 27304957 |
Victor Garcia-Font1, Carles Garrigues2, Helena Rifà-Pous2.
Abstract
In many countries around the world, smart cities are becoming a reality. These cities contribute to improving citizens' quality of life by providing services that are normally based on data extracted from wireless sensor networks (WSN) and other elements of the Internet of Things. Additionally, public administration uses these smart city data to increase its efficiency, to reduce costs and to provide additional services. However, the information received at smart city data centers is not always accurate, because WSNs are sometimes prone to error and are exposed to physical and computer attacks. In this article, we use real data from the smart city of Barcelona to simulate WSNs and implement typical attacks. Then, we compare frequently used anomaly detection techniques to disclose these attacks. We evaluate the algorithms under different requirements on the available network status information. As a result of this study, we conclude that one-class Support Vector Machines is the most appropriate technique. We achieve a true positive rate at least 56% higher than the rates achieved with the other compared techniques in a scenario with a maximum false positive rate of 5% and a 26% higher in a scenario with a false positive rate of 15%.Entities:
Keywords: anomaly detection; information security; outlier detection; smart cities; support vector machines; wireless sensor networks
Year: 2016 PMID: 27304957 PMCID: PMC4934294 DOI: 10.3390/s16060868
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Pipeline of the simulation and the experimental process.
Figure 2Schema and topology of the simulated WSN. The layout of the sensor nodes (i.e., nodes 1-10) reproduces the layout of real sound meters deployed in Barcelona over a 140 m × 140 m terrain. The topology and the base station (i.e., node 0) location are setup ad-hoc for the simulation.
Figure 3Size of the dataset partitions. The validation and test (val/test) datasets are partitioned in the same manner and contain the same number of samples of each attack type.
Metrics.
| True positive rate (tpr) | |
| False positive rate (fpr) | |
| F-score | |
Figure 4Results using the test dataset with samples of all the attacks filtering the features according to the three feature vector definitions with a very restrictive PFPR. The plots show the metrics f-score (f), the false positive rate (fpr) and the true positive rate (tpr). The captions below each plot indicate the feature vector definition used in each case.
Results sorted by TPR using test dataset (b) with samples of all the attacks.
| FV | PFPR | Technique | F-score | FPR | TPR |
|---|---|---|---|---|---|
| FV3 | very restrictive | ocsvm | 0.872 | 0.033 | 0.798 |
| FV3 | restrictive | ocsvm | 0.857 | 0.033 | 0.774 |
| FV2 | very restrictive | ocsvm | 0.853 | 0.024 | 0.762 |
| FV2 | restrictive | ocsvm | 0.853 | 0.024 | 0.762 |
| FV3 | permissive | ocsvm | 0.843 | 0.030 | 0.750 |
| FV1 | very restrictive | ocsvm | 0.6 | 0.708 | 0.729 |
| FV1 | restrictive | ocsvm | 0.599 | 0.696 | 0.723 |
| FV2 | permissive | ocsvm | 0.809 | 0.024 | 0.696 |
| FV1 | permissive | ocsvm | 0.583 | 0.681 | 0.690 |
| FV2 | permissive | hierarchical clustering | 0.665 | 0.211 | 0.552 |
| FV2 | permissive | mahalanobis | 0.670 | 0.149 | 0.542 |
| FV2 | restrictive | mahalanobis | 0.655 | 0.098 | 0.511 |
| FV2 | permissive | lofactor | 0.641 | 0.149 | 0.507 |
| FV3 | permissive | hierarchical clustering | 0.616 | 0.220 | 0.495 |
| FV2 | very restrictive | mahalanobis | 0.645 | 0.048 | 0.487 |
| FV3 | permissive | mahalanobis | 0.621 | 0.149 | 0.484 |
| FV2 | restrictive | lofactor | 0.631 | 0.098 | 0.484 |
| FV3 | restrictive | mahalanobis | 0.598 | 0.098 | 0.448 |
| FV2 | very restrictive | lofactor | 0.601 | 0.048 | 0.44 |
| FV3 | permissive | lofactor | 0.569 | 0.149 | 0.428 |
| FV3 | restrictive | hierarchical clustering | 0.545 | 0.140 | 0.401 |
| FV3 | restrictive | lofactor | 0.547 | 0.098 | 0.395 |
| FV3 | very restrictive | mahalanobis | 0.535 | 0.048 | 0.374 |
| FV2 | restrictive | hierarchical clustering | 0.517 | 0.098 | 0.366 |
| FV3 | very restrictive | lofactor | 0.514 | 0.048 | 0.354 |
| FV1 | permissive | hierarchical clustering | 0.394 | 0.158 | 0.265 |
| FV3 | very restrictive | hierarchical clustering | 0.340 | 0.054 | 0.210 |
| FV2 | very restrictive | hierarchical clustering | 0.311 | 0.071 | 0.191 |
| FV1 | restrictive | hierarchical clustering | 0.258 | 0.101 | 0.156 |
| FV1 | permissive | lofactor | 0.251 | 0.149 | 0.154 |
| FV1 | restrictive | lofactor | 0.195 | 0.098 | 0.113 |
| FV1 | permissive | mahalanobis | 0.124 | 0.149 | 0.071 |
| FV1 | very restrictive | hierarchical clustering | 0.122 | 0.057 | 0.067 |
| FV1 | very restrictive | lofactor | 0.117 | 0.048 | 0.064 |
| FV1 | restrictive | mahalanobis | 0.112 | 0.098 | 0.062 |
| FV1 | very restrictive | mahalanobis | 0.046 | 0.048 | 0.024 |
Results of several cases exceeding the PFPR. Cases where PFPR
| FV | Attack | PFPR | Technique | F-score | FPR | TPR |
|---|---|---|---|---|---|---|
| FV2 | Selective forwarding 30% | very restrictive | ocsvm | 0.811 | 0.117 | 0.762 |
| FV2 | Selective forwarding 30% | very restrictive | lofactor | 0.218 | 0.048 | 0.125 |
| FV2 | Selective forwarding 30% | very restrictive | mahalanobis | 0.598 | 0.048 | 0.437 |
| FV2 | Selective forwarding 30% | very restrictive | hierarchical clustering | 0.003 | 0.071 | 0.002 |
| FV2 | Selective forwarding 50% | very restrictive | ocsvm | 0.82 | 0.054 | 0.732 |
| FV2 | Selective forwarding 50% | very restrictive | lofactor | 0.502 | 0.048 | 0.343 |
| FV2 | Selective forwarding 50% | very restrictive | mahalanobis | 0.609 | 0.048 | 0.449 |
| FV2 | Selective forwarding 50% | very restrictive | hierarchical clustering | 0.003 | 0.071 | 0.002 |
| FV2 | Selective forwarding 30% | restrictive | ocsvm | 0.811 | 0.117 | 0.762 |
| FV2 | Selective forwarding 30% | restrictive | lofactor | 0.348 | 0.098 | 0.221 |
| FV2 | Selective forwarding 30% | restrictive | mahalanobis | 0.613 | 0.098 | 0.464 |
| FV2 | Selective forwarding 30% | restrictive | hierarchical clustering | 0.111 | 0.098 | 0.062 |
Mean of the standard deviation of all the features of the training dataset (a) and the test dataset (b) with all the attacks for each feature vector definition.
| FV | Dataset | Std. Mean |
|---|---|---|
| FV1 | training dataset (a) | 0.48 |
| FV1 | test dataset (b) | 0.45 |
| FV2 | training dataset (a) | 0.39 |
| FV2 | test dataset (b) | 0.60 |
| FV3 | training dataset (a) | 0.57 |
| FV3 | test dataset (b) | 0.79 |