| Literature DB >> 31998200 |
Graziella Orrù1, Merylin Monaro2, Ciro Conversano1, Angelo Gemignani1, Giuseppe Sartori2.
Abstract
Recent controversies about the level of replicability of behavioral research analyzed using statistical inference have cast interest in developing more efficient techniques for analyzing the results of psychological experiments. Here we claim that complementing the analytical workflow of psychological experiments with Machine Learning-based analysis will both maximize accuracy and minimize replicability issues. As compared to statistical inference, ML analysis of experimental data is model agnostic and primarily focused on prediction rather than inference. We also highlight some potential pitfalls resulting from adoption of Machine Learning based experiment analysis. If not properly used it can lead to over-optimistic accuracy estimates similarly observed using statistical inference. Remedies to such pitfalls are also presented such and building model based on cross validation and the use of ensemble models. ML models are typically regarded as black boxes and we will discuss strategies aimed at rendering more transparent the predictions.Entities:
Keywords: cross-validation; machine learning; machine learning in psychological experiments; machine learning in psychometrics; replicability
Year: 2020 PMID: 31998200 PMCID: PMC6966768 DOI: 10.3389/fpsyg.2019.02970
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
FIGURE 110-fold cross validation.
ML analysis conducted on 186 participants tested with the MCM III.
| Naive Bayes | 67% | 65% | 66% | 1% |
| Logistic | 75% | 62% | 58% | 17% |
| SVM | 74% | 70% | 67% | 7% |
| Knn | 79% | 70% | 64% | 15% |
| OneR | 70% | 62% | 67% | 3% |
| CART | 93% | 62% | 61% | 32% |
| Random forest | 100% | 66% | 64% | 36% |
| Neural network | 96% | 66% | 69% | 27% |
| (Averaging) | 81.6% | 65.4% | 65.3% | 0.1% |
| (12.7) | (3.33) | (3.37) | ||
| Ensemble learner | 80.6% | 67.7% | 69.4% | 1.7% |
Different machine learning models trained using 10-fold cross validation.
| Fold 1 = 68% | Fold 2 = 73% | Fold 3 = 68% | Fold 4 = 66% | 5% |
| Fold 2 = 69% | Fold 1 = 65% | Fold 3 = 66% | Fold 4 = 66% | 4% |
| Fold 3 = 69% | Fold 1 = 63% | Fold 2 = 73% | Fold 4 = 66% | 6% |
| Fold 4 = 65% | Fold 1 = 65% | Fold 2 = 66% | Fold 3 = 63% | 2% |
| Fold 1 = 63% | Fold 2 = 70% | Fold 3 = 71% | Fold 4 = 69% | 8% |
| Fold 2 = 69% | Fold 1 = 66% | Fold 3 = 66% | Fold 4 = 61% | 8% |
| Fold 3 = 69% | Fold 1 = 70% | Fold 2 = 69% | Fold 4 = 61% | 8% |
| Fold 4 = 65% | Fold 1 = 74% | Fold 2 = 67% | Fold 3 = 72% | 9% |
| Fold 1 = 62% | Fold 2 = 69% | Fold 3 = 67% | Fold 4 = 58% | 7% |
| Fold 2 = 72% | Fold 1 = 66% | Fold 3 = 64% | Fold 4 = 67% | 8% |
| Fold 3 = 71% | Fold 1 = 69% | Fold 2 = 67% | Fold 4 = 56% | 15% |
| Fold 4 = 63% | Fold 1 = 66% | Fold 2 = 71% | Fold 3 = 64% | 8% |
| Fold 1 = 65% | Fold 2 = 67% | Fold 3 = 69% | Fold 4 = 61% | 5% |
| Fold 2 = 69% | Fold 1 = 64% | Fold 3 = 65% | Fold 4 = 63% | 6% |
| Fold 3 = 68% | Fold 1 = 65% | Fold 2 = 74% | Fold 4 = 60% | 8% |
| Fold 4 = 63% | Fold 1 = 71% | Fold 2 = 69% | Fold 3 = 72% | 9% |