| Literature DB >> 35402904 |
Lakshmi Babu Saheer1, Ajay Bhasy1, Mahdi Maktabdar1, Javad Zarrin1.
Abstract
Monitoring, predicting, and controlling the air quality in urban areas is one of the effective solutions for tackling the climate change problem. Leveraging the availability of big data in different domains like pollutant concentration, urban traffic, aerial imagery of terrains and vegetation, and weather conditions can aid in understanding the interactions between these factors and building a reliable air quality prediction model. This research proposes a novel cost-effective and efficient air quality modeling framework including all these factors employing state-of-the-art artificial intelligence techniques. The framework also includes a novel deep learning-based vegetation detection system using aerial images. The pilot study conducted in the UK city of Cambridge using the proposed framework investigates various predictive models ranging from statistical to machine learning and deep recurrent neural network models. This framework opens up possibilities of broadening air quality modeling and prediction to other domains like vegetation or green space planning or green traffic routing for sustainable urban cities. The research is mainly focused on extracting strong pieces of evidence which could be useful in proposing better policies around climate change.Entities:
Keywords: aerial view image recognition; climate change mitigation; cost effective modeling; machine learning and deep learning algorithms; regression based prediction algorithms; urban air quality; urban vegetation detection
Year: 2022 PMID: 35402904 PMCID: PMC8993228 DOI: 10.3389/fdata.2022.822573
Source DB: PubMed Journal: Front Big Data ISSN: 2624-909X
Pollutants monitored in Cambridge city.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| Gonville place | ✓ | ✓ | ✓ | ✓ | ✓ |
| Montague road | ✓ | X | ✓ | ✓ | ✓ |
| Newmarket road | X | ✓ | ✓ | ✓ | ✓ |
| Parker street | ✓ | X | ✓ | ✓ | ✓ |
Weather data variables and their units in the data set.
|
|
|
|---|---|
| Temperature | Degree celsius (°C) |
| Dew point temperature | Degree celsius (°C) |
| Pressure | Millibar (mBar) |
| Wind speed | Knots (kts) |
| Wind direction | Direction (South-East, East etc) |
| Sunshine hours | Hours (hrs) |
| Rain | Millimeter (mm) |
| Maximum wind speed | Knots (kts) |
Figure 1The impact of brightness, saturation, and normalization processes on tree detection performance. Normalized images are exhibiting a noticeably better detection accuracy.
Figure 2Tree recognition framework: self supervised training approach.
Tree crown recognition results on images from Cambridge city.
|
|
|
|---|---|
| Untrained DeepForest | 0.28mAP |
| Proposed model (Self-supervised model) | 0.89mAP |
| Weinstein et al. ( | 0.61mAP |
Figure 3Positive results with self supervised model (trees in red bounding boxes).
Figure 4Some examples of undesirable results with self supervised model (trees in red bounding boxes).
Figure 5AirQuality framework: modeling pollutant concentration with weather and vegetation.
Figure 6Yearly trends: pollutant concentration trends in the recent years.
Figure 7Daily NO2 trends: concentration trends across four different locations.
Figure 8Correlation between Wind Speed and PM10.
Figure 9Correlation between vegetation and pollutants (PM10 and NO2).
Features introduced as a part of feature engineering.
|
|
|
|
|---|---|---|
| Weekend | Float (0/1) | Indicates whether the date is weekend or not |
| Weekday | Float (0/1) | Indicates whether the date is weekday or not |
| Season | String | The name of season derived from the date |
| HourCos,HourSin | Float | Since hour is a cyclic variable converted it to trigonometric functions Cos and Sin |
| MonthCos,MonthSin | Float | Since month is a cyclic variable converted it to trigonometric functions Cos and Sin |
| NO2MA10 | Float | 10 day Moving average of the concentration |
| NO2MA20 | Float | 20 Day Moving average of the concentration |
| 100mTrees | Float | Number of trees within 100 m of the sensor |
| 250mTrees | Float | Number of trees within 200 m of the sensor |
| 500mTrees | Float | Number of trees within 500 m of the sensor |
| 1000mTrees | Float | Number of trees within 1,000 m of the sensor |
Figure 10Correlation of the engineered features with regards to PM10 and NO2.
Long short term memory (LSTM) Hyperparameters optimized during model training.
|
|
|
|---|---|
| Number of LSTM layers | 2 to 5 |
| Number of Neurons per layer | 32 to 512 with stepsize 32 |
| Learning rate | 1e-2, 1e-3 and 1e-4 |
| Rate for dropout | between 0 and 0.5 |
| Dropout option | True or False |
| Loss functions | MSE, MAE |
| Activation functions | Tanh, Linear, Relu, Sigmoid |
Experimental results for NO2, location-1, and Parker Street.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| ARIMA | Without trees or extra features |
| 64.2431 | 8.0151 | 0.7804 | 29.2145 |
| Linear regression | With trees | 7.1151 | 86.9743 | 9.3260 | 0.6911 | 39.0577 |
| Linear regression | Without trees | 5.7071 | 59.4619 | 7.7111 | 0.7888 | 28.6116 |
| LinearSVR | With trees | 7.0114 | 85.4898 | 9.2460 | 0.6963 | 37.7552 |
| LinearSVR | Without trees | 5.7071 | 59.4619 | 7.7111 | 0.7888 | 28.6116 |
| Polynomial regression | With Trees | 5.7845 |
| 7.6947 |
| 30.7340 |
| Polynomial regression | Without trees | 5.7071 |
| 7.7111 |
| 28.6116 |
|
|
|
| 59.6007 |
| 0.7883 |
|
|
|
|
| 59.4619 |
| 0.7888 |
|
| SVR With RBF Kernel | With trees | 5.8804 | 62.1121 | 7.8811 | 0.7794 | 32.5764 |
| SVR With RBF Kernel | Without trees | 5.7071 | 59.4619 | 7.7111 | 0.7888 | 28.6116 |
| PF-SVR with RBF Kernel | With trees | 5.8081 | 61.4881 | 7.8414 | 0.7816 | 31.7735 |
| PF-SVR with RBF Kernel | Without trees | 5.7071 | 59.4619 | 7.7111 | 0.7888 | 28.6116 |
| LSTM | With trees | 6.8184 | 87.6507 | 9.3621 | 0.6822 | 30.6485 |
| LSTM | Without trees | 6.2056 | 71.3269 | 8.4455 | 0.7414 | 28.3280 |
linearSVR, SVR with linear kernel; polynomial SVR, SVR with polynomial kernel; PF-SVR, SVR using polynomial features; SVR with RBF kernel, SVR with radial basis function (RBF) kernel; PF-SVR with RBF kernel, SVR using polynomial features and RBF kernel; LSTM, long short term memory; MAE, mean absolute error; MSE, mean squared error; RMSE, root MSE; R2, R-squared error; MAPE, mean absolute percentage error; SVR, support vector regression. Bold values are indicate the overall best performing models.
Experimental results for PM10, location-1, and Parker Street.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| ARIMA | Without Trees or Extra Features | 3.8532 | 40.9184 | 6.3967 | 0.6375 | 59.7081 |
| Linear Regression | With Trees | 4.6605 | 40.9548 | 6.3996 | 0.4913 | 29.2084 |
| Linear Regression | Without Trees | 4.6605 | 40.9548 | 6.3996 | 0.4913 | 29.2084 |
| LinearSVR | With Trees | 4.4460 | 39.5361 | 6.2878 | 0.5089 | 27.0797 |
| LinearSVR | Without Trees | 4.4407 | 39.5027 | 6.2851 | 0.5093 | 27.0429 |
| Polynomial Regression | With Trees | 4.0152 | 32.2899 | 5.6824 | 0.5989 | 25.1360 |
| Polynomial Regression | Without Trees | 4.1887 | 34.1941 | 5.8476 | 0.5752 | 26.5016 |
| Polynomial SVR Regression | With Trees | 4.0677 | 33.2232 | 5.7640 | 0.5873 | 25.1261 |
| Polynomial SVR Regression | Without Trees | 4.0601 | 33.1445 | 5.7571 | 0.5883 | 25.0372 |
| SVR With RBF Kernel | With Trees | 3.9584 |
|
|
|
|
| SVR With RBF Kernel | Without Trees | 3.9584 |
|
|
|
|
| PF-SVR with RBF Kernel | With Trees |
| 32.2535 | 5.6792 | 0.5993 | 25.0009 |
| PF-SVR with RBF Kernel | Without Trees | 3.9689 | 32.6011 | 5.7097 | 0.5950 | 25.0535 |
| LSTM | With Trees | 6.9262 | 79.6629 | 8.9254 | 0.0852 | 46.8802 |
| LSTM | Without Trees | 7.3973 | 94.0812 | 9.6995 | 0.0803 | 49.0300 |
LinearSVR, SVR with linear kernel; polynomial SVR, SVR with polynomial kernel; PF-SVR, SVR using polynomial features; SVR with RBF kernel, SVR with radial basis function (RBF) Kernel; PF-SVR with RBF kernel, SVR using polynomial features and RBF kernel; LSTM, long short term memory; MAE, mean absolute error; MSE, mean squared error; RMSE, root MSE; R2, R-squared error; MAPE, mean absolute percentage error, SVR, support vector regression. Bold values are indicate the overall best performing models.
Experimental results for PM10, location-2, Gonville Place.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| ARIMA | Without Trees or Extra Features | 3.5494 | 30.2370 | 5.4988 | 0.6762 | 23.1247 |
| Linear Regression | With Trees | 4.1726 | 31.8227 | 5.6411 | 0.5735 | 27.7313 |
| Linear Regression | Without Trees | 4.1711 | 31.8322 | 5.6420 | 0.5733 | 27.7040 |
| LinearSVR | With Trees | 4.1327 | 32.3228 | 5.6853 | 0.5668 | 26.8300 |
| LinearSVR | Without Trees | 4.1337 | 32.3442 | 5.6871 | 0.5665 | 26.8376 |
| Polynomial Regression | With Trees | 3.7524 |
|
| 0.6486 | 24.7285 |
| Polynomial Regression | Without Trees | 3.7730 | 26.5354 | 5.1512 | 0.6443 | 24.8110 |
| Polynomial SVR Regression | With Trees | 3.7695 | 26.7237 | 5.1695 | 0.6418 | 24.3154 |
| Polynomial SVR Regression | Without Trees | 3.7736 | 26.8158 | 5.1783 | 0.6406 | 24.4013 |
| SVR With RBF Kernel | With Trees |
| 26.5390 | 5.1516 |
| 24.2396 |
| SVR With RBF Kernel | Without Trees |
| 26.5239 | 5.1501 |
|
|
| PF-SVR with RBF Kernel | With Trees | 3.7497 | 27.2717 | 5.2222 | 0.6345 | 24.1598 |
| PF-SVR with RBF Kernel | Without Trees | 3.7679 | 27.5793 | 5.2516 | 0.6303 | 24.2984 |
| LSTM | With Trees | 4.9231 | 45.8927 | 6.7744 | 0.4292 | NA |
| LSTM | Without Trees | 4.7463 | 44.2022 | 6.6484 | 0.4503 | NA |
LinearSVR, SVR with linear kernel; Polynomial SVR, SVR with polynomial Kernel; PF-SVR, SVR using Polynomial features; SVR with RBF kernel, SVR with radial basis function (RBF) kernel; PF-SVR with RBF kernel, SVR using polynomial features and RBF kernel; LSTM, long short term memory; MAE, mean absolute error; MSE, mean squared error; RMSE, root MSE; R2, R-squared error; MAPE, mean absolute percentage error; SVR, support vector regression. Bold values are indicate the overall best performing models.
Experimental results for NO2, location-2, Gonville Place.
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| ARIMA | Without trees or extra features | 4.9797 | 51.3185 | 7.1636 | 0.7516 | 31.9924 |
| Linear regression | With trees | 6.8627 | 77.8650 | 8.8241 | 0.6172 | 40.6186 |
| Linear regression | Without trees | 6.8627 | 77.8650 | 8.8241 | 0.6172 | 40.6186 |
| LinearSVR | With trees | 6.5645 | 73.6806 | 8.5837 | 0.6378 | 37.8003 |
| LinearSVR | Without trees | 6.5697 | 73.7699 | 8.5889 | 0.6373 | 37.8582 |
| Polynomial regression | With trees | 5.8845 | 57.6263 | 7.5912 | 0.7167 | 34.9102 |
| Polynomial regression | Without trees | 5.7924 | 56.6969 | 7.5297 | 0.7213 | 33.9497 |
|
| With Trees |
|
| 7.4938 |
|
|
| Polynomial SVR regression | Without trees | 5.6841 | 56.3862 | 7.5091 | 0.7228 | 32.6777 |
| SVR With RBF Kernel | With trees | 5.8322 | 56.9608 | 7.5472 | 0.7200 | 36.4192 |
| SVR With RBF Kernel | Without trees | 5.8322 | 56.9608 | 7.5472 | 0.7200 | 36.4192 |
| PF-SVR with RBF Kernel | With trees | 5.7556 |
|
| 0.7263 | 36.0017 |
| PF-SVR with RBF Kernel | Without trees | 5.7361 | 55.7099 | 7.4639 | 0.7261 | 35.6012 |
| LSTM | Without trees | 5.0542 | 51.1251 | 7.1501 | 0.7484 | 23.0866 |
| LSTM | With trees | 5.2241 | 52.6281 | 7.2545 | 0.7410 | 25.0298 |
LinearSVR, SVR with linear kernel; polynomial SVR, SVR with polynomial kernel; PF-SVR, SVR using polynomial features; SVR with RBF kernel, SVR with radial basis function (RBF) kernel; PF-SVR with RBF kernel, SVR using polynomial features and RBF kernel; LSTM, long short term memory; MAE, mean absolute error; MSE, mean squared error; RMSE, root MSE; R2, R-squared error; MAPE, mean absolute percentage error; SVR, support vector regression. Bold values are indicate the overall best performing models.