| Literature DB >> 35591195 |
Abdallah Namoun1, Burhan Rashid Hussein2, Ali Tufail2, Ahmed Alrehaili1,3, Toqeer Ali Syed1, Oussama BenRhouma1.
Abstract
With the increase in urbanization and smart cities initiatives, the management of waste generation has become a fundamental task. Recent studies have started applying machine learning techniques to prognosticate solid waste generation to assist authorities in the efficient planning of waste management processes, including collection, sorting, disposal, and recycling. However, identifying the best machine learning model to predict solid waste generation is a challenging endeavor, especially in view of the limited datasets and lack of important predictive features. In this research, we developed an ensemble learning technique that combines the advantages of (1) a hyperparameter optimization and (2) a meta regressor model to accurately predict the weekly waste generation of households within urban cities. The hyperparameter optimization of the models is achieved using the Optuna algorithm, while the outputs of the optimized single machine learning models are used to train the meta linear regressor. The ensemble model consists of an optimized mixture of machine learning models with different learning strategies. The proposed ensemble method achieved an R2 score of 0.8 and a mean percentage error of 0.26, outperforming the existing state-of-the-art approaches, including SARIMA, NARX, LightGBM, KNN, SVR, ETS, RF, XGBoosting, and ANN, in predicting future waste generation. Not only did our model outperform the optimized single machine learning models, but it also surpassed the average ensemble results of the machine learning models. Our findings suggest that using the proposed ensemble learning technique, even in the case of a feature-limited dataset, can significantly boost the model performance in predicting future household waste generation compared to individual learners. Moreover, the practical implications for the research community and respective city authorities are discussed.Entities:
Keywords: IoT; ensemble learning; forecasting; machine learning; meta learner; smart cities; smart waste management; solid waste generation; time-series
Mesh:
Substances:
Year: 2022 PMID: 35591195 PMCID: PMC9104882 DOI: 10.3390/s22093506
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Summary of main research weaknesses/gaps in smart waste management.
| Article | Strengths | Weaknesses |
|---|---|---|
| A multi-layer LoRaWAN infrastructure |
A LoRaWAN-based architecture with three layers of implementation is proposed. Fill level measurements, vandalism detection, video surveillance, user privacy, and security. Machine-learning-based image detection mechanism. |
The study does not predict fill levels based on waste generation patterns. The route prediction and optimization mechanisms are not discussed. The study does not focus on household waste generation or prediction. |
| Use of IoT to enable solid waste management [ |
An architecture is presented to monitor the fill levels of smart bins. A GUI-enabled interface is provided for authorities to monitor the fill level. A prototype is implemented and tested. |
No fill-level predictions. Route selection and prediction are not presented. Waste generation patterns are not explored. |
| Prediction of plastic waste generation using SHAP [ |
A neural network (ANN) model is proposed to anticipate the generation of plastic waste. Shapley additive explanation (i.e., SHAP) analysis is applied to interpret the outcomes. |
The study focuses on plastic waste generation only. The exclusive target of the study is the European Union. The study does not focus on household waste generation patterns. |
| Forecasting of regional waste generation [ |
Demographic and socioeconomic variables are used to predict MSW generation and diversion. A real dataset from 2001 to 2014 is utilized. A decision tree and artificial neural network are used to forecast annual waste generation. |
The analysis in the study is based on a dataset that is limited to Canada. The specific validity of the analysis is not proven on different datasets or regions with different variables such as demographics, age, income level, etc. |
| Application of deep neuroevolution for the prediction of waste generation [ |
Waste generation predictions are made based on real-world conditions. Authors utilize a deep neuroevolutionary algorithm that helps to automatically design a recurrent neural network for predicting waste filling levels from waste containers. Authors validate their approach by utilizing real-world scenarios and implementations. |
The study focuses on individual metrics that may not reflect the actual performance improvements. Parameters related to waste generation is not discussed, such as demographics, social economics, age, etc. |
Figure 1The proposed method for the prediction of solid waste generation (in model optimization phase: M = model, P = prediction).
A conceptual comparison of single prediction models for smart waste management.
| Algorithm | Advantages | Disadvantages | |
|---|---|---|---|
| LightGBM | Family of gradient boosting framework that uses a tree-based algorithm |
Fast and efficient. Low memory usage. Better performance than most boosting algorithms. Works well with a large dataset. |
Prone to overfitting on a smaller dataset. |
| KNN | Selects k-closest examples from the dataset |
Intuitive and straightforward to understand. Constantly changes with the new data. |
Does not scale well with a large dataset. Sensitive to outliers. |
| SVR | Selects the best hyperplane that maximizes separability between classes |
It is effective in high dimensional spaces. The algorithm is memory efficient. |
Does not scale well when using an extensive dataset. |
| ETS | An ensemble of decision trees that performs a random split of the node |
Extremely fast compared to the RF. Works better than the RF when having a large dataset. |
The model may have high variance as it performs random splits. |
| RF | An ensemble of decision trees that uses a greedy approach to achieve the best split |
Reduces the variance of the model and overfitting. Gives information about essential features from the data. |
It is computationally expensive compared to other algorithms. |
| XGBoosting | An ensemble of decision trees that uses a gradient boosting framework |
Can handle missing data with its in-built features. Works well on small and medium-sized data. |
Very sensitive to outliers. Does not scale well when using massive datasets. |
| ANN | Inspired by the human brain, ANN consists of several input and output layers that can extract patterns from given data |
Ability to learn and model complex relationships. Robust to noisy training data. |
Works only on numerical data. Hard to obtain the optimal results. |
Search space configuration for hyperparameter optimization using Optuna.
| Algorithm | Search Space |
|---|---|
| SVR | Kernel = [‘rbf’,’poly’,’linear’,’sigmoid’] |
| XGboost | max_depth = int (4, 12) |
| LightGBM | reg_alpha = log uniform (1 × 10−3, 10.0) |
| RF | n_estimators =_int (low = 100, high = 1000) |
| ANN | learning_rate_init = float (0.0001, 0.1, step = 0.005) |
| KNN | n_neighbors = int (1, 30) |
| Extreme Trees | n_estimators = int (low = 100, high = 1000) |
Performance comparison between the proposed ensemble method and existing ML approaches.
| Model | MAE | MSE | RMSE | MAPE | R2 Score |
|---|---|---|---|---|---|
| SARIMA [ | 0.191 | 0.061 | 0.247 | 0.99 | −0.708 |
| NARX neural network [ | 0.0850 | 0.014 | 0.119 | 0.834 | 0.612 |
| LightGBM [ | 0.0796 | 0.0114 | 0.1066 | 0.3557 | 0.7462 |
| KNN [ | 0.0652 | 0.0112 | 0.1060 | 0.2816 | 0.7489 |
| SVR [ | 0.0927 | 0.0138 | 0.1174 | 0.4015 | 0.6918 |
| ETS [ | 0.0817 | 0.0118 | 0.1085 | 0.3490 | 0.7368 |
| RF [ | 0.0850 | 0.0128 | 0.1131 | 0.3822 | 0.7143 |
| XGBoosting [ | 0.0898 | 0.0148 | 0.1215 | 0.4044 | 0.6700 |
| ANN [ | 0.0895 | 0.0141 | 0.1186 | 0.3784 | 0.6854 |
| Ensemble (Average ensemble) | 0.073 | 0.010 | 0.098 | 0.331 | 0.787 |
| Proposed Ensemble (Meta model) |
|
|
|
|
|
Figure 2Hyperparameter optimization for the SVR model.
Figure 3Hyperparameter optimization for the XGBoosting model.
Figure 4Hyperparameter optimization for the XGBoosting model.
Figure 5Hyperparameter optimization for the RF model.
Figure 6Hyperparameter optimization for the ANN model.
Figure 7Hyperparameter optimization for the KNN model.
Figure 8Hyperparameter optimization for the ETS model.