| Literature DB >> 35795740 |
Fahad Kamal Alsheref1, Ibrahim Eldesouky Fattoh2, Waleed M Ead1.
Abstract
Competent employees are a rare commodity for great companies. The problem of maintaining good employees with experience threatens the owners of companies. The issue of employee attrition can cost employers a lot as it takes a lot to compensate for their expertise and efficiency. For this reason, in this research, we present an automated model that can predict employee attrition based on different predictive analytical techniques. These techniques have been applied with different pipeline architectures to select the best champion model. Also, an autotuning approach has been implemented to calculate the best combination of hyper parameters to build the champion model. Finally, we propose an ensemble model for selecting the most efficient model subject to different assessments measures. The results of the proposed model show that no model up until now could be considered ideal and perfect for each case of business context. Yet, our chosen model was pretty much optimal as per our requirements and adequately satisfied the intended goal.Entities:
Mesh:
Year: 2022 PMID: 35795740 PMCID: PMC9251085 DOI: 10.1155/2022/7728668
Source DB: PubMed Journal: Comput Intell Neurosci
Overview of machine learning methods used for the prediction of employee turnover.
| Ref. | Machine learning method | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| DT | RF | GBT | XGB | LR | SVM | NN | LDA | NB | KNN | AdaBoost | |
| Alao and Adeyem | Yes | – | – | – | – | – | – | – | – | – | |
| Sisodia et al. | Yes | Yes | Yes | Yes | Yes | ||||||
| Zhao et al. | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | |
| Yadav et al. | Yes | Yes | Yes | Yes | Yes | ||||||
| Alduayj, S. S. and Rajpoot, K. (2018) | Yes | Yes | Yes | ||||||||
| Falluchi et al. | Yes | Yes | Yes | Yes | Yes | Yes | |||||
Figure 1Steps of the proposed model.
Parameter values used in the preprocessing step.
| Parameters | Parameter values |
|---|---|
| Maximum class level | 20 |
| Interval cut-off | 20 |
| Maximum missing percentage | 50 |
| Partitioning method | Stratify |
| Partitioning ratios | 60 : 30 : 10 |
| Imputation (missing values) method | Count for categorical variable |
| Mean for interval variables |
Figure 2Mlp Neural Network.
Figure 3Cumulative lift value for used algorithms.
Figure 4Lift value for used algorithms.
Figure 5Sensitivity value for used algorithms.
Figure 6Accuracy value for used algorithms.
Figure 7F1 score value for used algorithms.
Gini coefficient.
| Partition | GB | NN | Forest | Ensemble |
|---|---|---|---|---|
| Train | −0.0335 | 0.6527 | 0.4671 | 0.5715 |
| Validate | 0.5078 | 0.7936 | 0.9826 | 0.8161 |
| Test | 0.3030 | 0.7704 | 0.6377 | 0.7326 |
Misclassification rate.
| Partition | GB | NN | Forest | Ensemble |
|---|---|---|---|---|
| Train | 0.1633 | 0.1428 | 0.1701 | 0.1701 |
| Validate | 0.1609 | 0.1043 | 0.1088 | 0.1247 |
| Test | 0.1609 | 0.0861 | 0.1406 | 0.1337 |
Average square error.
| Partition | GB | NN | Forest | Ensemble |
|---|---|---|---|---|
| Train | 0.138 | 0.10120 | 0.11953 | 0.113693 |
| Validate | 0.129 | 0.07701 | 0.06321 | 0.08596 |
| Test | 0.132 | 0.075362 | 0.107621 | 0.094548 |