| Literature DB >> 31393894 |
Bennett Kleinberg1,2, Arnoud Arntz2, Bruno Verschuere2.
Abstract
PURPOSE: Verbal credibility assessments examine language differences to tell truthful from deceptive statements (e.g., of allegations of child sexual abuse). The dominant approach in psycholegal deception research to date (used in 81% of recent studies that report on accuracy) to estimate the accuracy of a method is to find the optimal statistical separation between lies and truths in a single dataset. However, this method lacks safeguards against accuracy overestimation. METHOD &Entities:
Mesh:
Year: 2019 PMID: 31393894 PMCID: PMC6687387 DOI: 10.1371/journal.pone.0220228
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Accuracies for latent discriminant analysis without cross-validation on simulated data for 8, 12 and 19 predictors with increasing n.
The averaged accuracy is displayed on the y-axis and the increased sample size for the simulation on the x-axis.
Fig 2Accuracy differences (y-axis) between traditional training set optimisation and leave-one-out cross-validation compared to independent test set validation for increasing sample sizes (x-axis).
The dashed horizontal grey lines indicate the upper and lower boundary of the [-0.05; +0.05] stability corridor. The vertical coloured line indicates the sample size points-of-stability for the training set optimisation technique. Inset plot: accuracy difference scores zoomed in for sample size between 40 and 240.
Illustration how the dominant practice (linear discriminant analysis with training set optimisation) can lead to an erroneous conclusion.
| Training set optimisation (current practice) | Leave-one-out cross-validation | |
|---|---|---|
| Accuracy estimate | 61.54% | 51.28% |
| Conclusion | Significantly | No better than chance classification. |
Suggestions for the improvement of the accuracy estimation in the predictive analysis in verbal credibility assessment research.
| Remedy | Key Advantage | Key challenge | Safeguard |
|---|---|---|---|
| Validation on an independent sample | - Allows for robust claims regarding the generalizability of findings | Resource intensive (new data collection) | - Pre-registration of the classification algorithm |
| Cross-validation | - Easy to implement (no new data collection needed, often default setting in statistical software) | - Might still capitalise on idiosyncrasies of the sample | - Pre-registration of cross-validation procedure |
| Larger sample sizes | - Solidifies conclusions on statistical inferences and prediction metrics | - Resource intensive | - Preregistration of sample size justification |