| Literature DB >> 31100879 |
Nikolaos Sideris1,2, Georgios Bardis3, Athanasios Voulodimos4, Georgios Miaoulis5, Djamchid Ghazanfarpour6.
Abstract
The constantly increasing amount and availability of urban data derived from varying sources leads to an assortment of challenges that include, among others, the consolidation, visualization, and maximal exploitation prospects of the aforementioned data. A preeminent problem affecting urban planning is the appropriate choice of location to host a particular activity (either commercial or common welfare service) or the correct use of an existing building or empty space. In this paper, we propose an approach to address these challenges availed with machine learning techniques. The proposed system combines, fuses, and merges various types of data from different sources, encodes them using a novel semantic model that can capture and utilize both low-level geometric information and higher level semantic information and subsequently feeds them to the random forests classifier, as well as other supervised machine learning models for comparisons. Our experimental evaluation on multiple real-world data sets comparing the performance of several classifiers (including Feedforward Neural Networks, Support Vector Machines, Bag of Decision Trees, k-Nearest Neighbors and Naïve Bayes), indicated the superiority of Random Forests in terms of the examined performance metrics (Accuracy, Specificity, Precision, Recall, F-measure and G-mean).Entities:
Keywords: decision support system; machine learning; random forests; urban planning
Year: 2019 PMID: 31100879 PMCID: PMC6567884 DOI: 10.3390/s19102266
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Visualization of a geoquery by the proposed system.
Figure 2Altered point of view visualization of a geoquery by our system.
Figure 3Functional Block Diagram of the proposed System.
Figure 4Euclidian distance vs routable road distance.
Figure 5Feature extraction: distance from nearest atm.
Summary of prediction results.
| Actual Class | |||
|---|---|---|---|
| YES ( | NO ( | ||
|
| YES ( | TP | FP |
| NO ( | FN | TN | |
Figure 6Architecture and graph of Mean Square Error (MSE) plot as varied for 1st layer number of neurons.
Results for Artificial Neural Networks.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| 1 | 0.790 | 0.84 | 0.82 | 0.74 | 0.78 | 0.79 |
| 2 | 0.740 | 0.92 | 0.88 | 0.56 | 0.68 | 0.72 |
| 3 | 0.660 | 0.72 | 0.68 | 0.60 | 0.64 | 0.66 |
| 4 | 0.640 | 0.67 | 0.65 | 0.61 | 0.63 | 0.64 |
| 5 | 0.640 | 0.68 | 0.65 | 0.60 | 0.63 | 0.64 |
| 6 | 0.630 | 0.78 | 0.69 | 0.48 | 0.56 | 0.61 |
| 7 | 0.675 | 0.65 | 0.67 | 0.70 | 0.68 | 0.67 |
| 8 | 0.675 | 0.78 | 0.72 | 0.57 | 0.64 | 0.67 |
Figure 7Optimization of hyperparameters box and sigma.
Results for SVMs.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| 1 | 0.85 | 0.91 | 0.90 | 0.78 | 0.83 | 0.84 |
| 2 | 0.85 | 0.90 | 0.89 | 0.80 | 0.84 | 0.85 |
| 3 | 0.77 | 0.81 | 0.79 | 0.73 | 0.76 | 0.77 |
| 4 | 0.77 | 0.80 | 0.78 | 0.73 | 0.76 | 0.76 |
| 5 | 0.78 | 0.87 | 0.84 | 0.69 | 0.76 | 0.77 |
| 6 | 0.80 | 0.90 | 0.88 | 0.70 | 0.78 | 0.79 |
| 7 | 0.78 | 0.83 | 0.81 | 0.73 | 0.77 | 0.78 |
| 8 | 0.80 | 0.90 | 0.88 | 0.70 | 0.78 | 0.79 |
Figure 8Behavior of Random Forest classifier with different number of features.
Results for optimized Random Forests.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| 1 | 0.96 | 0.98 | 0.98 | 0.93 | 0.95 | 0.95 |
| 2 | 0.94 | 0.98 | 0.98 | 0.89 | 0.93 | 0.93 |
| 3 | 0.86 | 0.93 | 0.92 | 0.79 | 0.85 | 0.86 |
| 4 | 0.92 | 0.93 | 0.93 | 0.91 | 0.92 | 0.92 |
| 5 | 0.93 | 0.92 | 0.92 | 0.94 | 0.93 | 0.93 |
| 6 | 0.93 | 0.92 | 0.92 | 0.93 | 0.93 | 0.92 |
| 7 | 0.92 | 0.92 | 0.92 | 0.92 | 0.92 | 0.92 |
| 8 | 0.86 | 0.92 | 0.91 | 0.80 | 0.85 | 0.86 |
Results of KNN.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| 1 | 0.73 | 0.92 | 0.87 | 0.53 | 0.66 | 0.70 |
| 2 | 0,71 | 0.94 | 0.89 | 0.48 | 0.62 | 0.67 |
| 3 | 0.72 | 0.90 | 0.84 | 0.53 | 0.65 | 0.69 |
| 4 | 0.69 | 0.92 | 0.85 | 0.46 | 0.60 | 0.65 |
| 5 | 0.71 | 0.92 | 0.86 | 0.50 | 0.63 | 0.68 |
| 6 | 0.68 | 0.91 | 0.83 | 0.44 | 0.58 | 0.63 |
| 7 | 0.68 | 0.93 | 0.86 | 0.43 | 0.57 | 0.63 |
| 8 | 0.69 | 0.87 | 0.80 | 0.51 | 0.62 | 0.67 |
Results of Naïve Bayes.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| 1 | 0.77 | 0.85 | 0.82 | 0.68 | 0.74 | 0.76 |
| 2 | 0.74 | 0.85 | 0.81 | 0.63 | 0.71 | 0.73 |
| 3 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
| 4 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
| 5 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
| 6 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
| 7 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
| 8 | 0.73 | 0.78 | 0.76 | 0.68 | 0.72 | 0.73 |
Average metrics for all datasets of all classifiers.
| Dataset | Accuracy | Specificity | Precision | Recall | F1 Measure | G-Mean |
|---|---|---|---|---|---|---|
| MLP | 0.681 | 0.755 | 0.719 | 0.608 | 0.655 | 0.674 |
| SVM | 0.799 | 0.865 | 0.846 | 0.733 | 0.784 | 0.796 |
| KNN | 0.699 | 0.914 | 0.850 | 0.485 | 0.617 | 0.665 |
| Naive Bayes | 0.736 | 0.798 | 0.770 | 0.674 | 0.718 | 0.733 |
| Bag of Decision trees | 0.651 | 0.601 | 0.636 | 0.700 | 0,666 | 0.649 |
| Random Forest | 0.913 | 0.938 | 0.934 | 0.889 | 0.910 | 0.912 |
Figure 9Comparison of Precision of all Classifiers.
Figure 10Comparison of Accuracy of all Classifiers.
Figure 11Comparison of Recall of all Classifiers.
Figure 12Comparison of F1 measure of all Classifiers.