| Literature DB >> 35591037 |
Noureen Zafar1,2, Irfan Ul Haq1, Jawad-Ur-Rehman Chughtai1, Omair Shafiq3.
Abstract
With the advent of the Internet of Things (IoT), it has become possible to have a variety of data sets generated through numerous types of sensors deployed across large urban areas, thus empowering the notion of smart cities. In smart cities, various types of sensors may fall into different administrative domains and may be accessible through exposed Application Program Interfaces (APIs). In such setups, for traffic prediction in Intelligent Transport Systems (ITS), one of the major prerequisites is the integration of heterogeneous data sources within a preprocessing data pipeline resulting into hybrid feature space. In this paper, we first present a comprehensive algorithm to integrate heterogeneous data obtained from sensors, services, and exogenous data sources into a hybrid spatial-temporal feature space. Following a rigorous exploratory data analysis, we apply a variety of deep learning algorithms specialized for time series geospatial data and perform a comparative analysis of Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), Convolutional Neural Network (CNN), and their hybrid combinations. The hybrid LSTM-GRU model outperforms the rest with Root Mean Squared Error (RMSE) of 4.5 and Mean Absolute Percentage Error (MAPE) of 6.67%.Entities:
Keywords: CNN; GRU; ITS; IoT; LSTM
Mesh:
Year: 2022 PMID: 35591037 PMCID: PMC9099662 DOI: 10.3390/s22093348
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Hybrid Feature Space.
| Nature of Attributes | Data Type |
|---|---|
| Day | integer |
| Hour | integer |
| Startnode | integer |
| Endnode | integer |
| aggminutes | 15 min time interval |
| Weather | char |
| maxspeed-real | integer |
| aggSpeed | integer |
| Holiday | boolean |
Figure 1Urban traffic speed prediction based on hybrid LSTM–GRU model.
Hyperparameters configuration for XGBoost, ANN, KNN, MLP, and Hybrid LSTM–GRU.
| Model | Hyperparameters | Values |
|---|---|---|
| XGBOOST | objective | linear |
| n-estimators | 4000 | |
| ANN | input dimension | 10 |
| activation function | relu | |
| loss function |
| |
| optimizere | adam | |
| epoch | 100 | |
| batch-size | 512 | |
| KNN | K | 20 |
| loss |
| |
| MLP | activation function | relu |
| loss function |
| |
| hidden-layer-size | 100 | |
| optimizer | SGD | |
| learning rate | 0.001 | |
|
| Batch Size | 512 |
| Learning Rate | 0.001 | |
| No of epochs | 10 | |
| No of Hidden Layers | 04 | |
| Hidden Units | 256 | |
| Dropout Ratio | 0.2 | |
| Activation Function | tanh | |
| Output-Units | 1 | |
| Output-Type | Single Label | |
| Output-Layer-Activation-Function | linear | |
| Optimizer | Adam | |
| Loss Function | mean squared error |
Figure 2Speed Performance index variation on weekdays.
Figure 3Speed performance index variation on weekends.
Feature selection through correlation feature selection technique.
| Features | Scores |
|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Figure 4Feature selection using mutual information regression feature selection technique.
Figure 5Heat map of hybrid feature space.
Figure 6Evaluation metrics of prediction on test data.
Performance metrics of deep learning models.
| Model |
|
|
|
|---|---|---|---|
|
| 4.86 | 2.13 | 6.95 |
|
| 5.05 | 2.29 | 7.7 |
|
| 30.3 | 25.96 | 64.10 |
|
| 4.7 | 23.9 | 7.9 |
|
| 5.89 | 3.53 | 11.47 |
|
| 4.6 | 2.08 | 6.85 |
|
|
|
|
|
|
| 5.1 | 2.4 | 8.4 |
|
| 7.8 | 4.5 | 14.6 |
Figure 7Evaluation metrics of prediction on test data.