| Literature DB >> 35600326 |
Emil Rijcken1,2, Uzay Kaymak1, Floortje Scheepers3, Pablo Mosteiro2, Kalliopi Zervanou4,5, Marco Spruit2,4,5.
Abstract
The clinical notes in electronic health records have many possibilities for predictive tasks in text classification. The interpretability of these classification models for the clinical domain is critical for decision making. Using topic models for text classification of electronic health records for a predictive task allows for the use of topics as features, thus making the text classification more interpretable. However, selecting the most effective topic model is not trivial. In this work, we propose considerations for selecting a suitable topic model based on the predictive performance and interpretability measure for text classification. We compare 17 different topic models in terms of both interpretability and predictive performance in an inpatient violence prediction task using clinical notes. We find no correlation between interpretability and predictive performance. In addition, our results show that although no model outperforms the other models on both variables, our proposed fuzzy topic modeling algorithm (FLSA-W) performs best in most settings for interpretability, whereas two state-of-the-art methods (ProdLDA and LSI) achieve the best predictive performance.Entities:
Keywords: electronic health records; explainability; information extraction; interpretability; natural language processing; psychiatry; text classification; topic modeling
Year: 2022 PMID: 35600326 PMCID: PMC9114871 DOI: 10.3389/fdata.2022.846930
Source DB: PubMed Journal: Front Big Data ISSN: 2624-909X
Figure 1Visual representation of the modeling pipeline per algorithm.
Figure 2Model interpretability vs. predictive performance per trained model, as measured by the AUC_20 (only the top 20 words per topic are considered.).
Figure 3Two graphs showing how a model's interpretability relates to its predictive performance, as measured by the AUK per trained model. (A) Shows predictions based on a topics first 20 words only, while (B) takes an entire topics' distribution into account.
Figure 4The effect of number of topics on the Interpretability—each data point is a mean score based on 10 runs.
Figure 5The effect of number of topics on the coherence, based on the top 20 words (A) and diversity (B)—each data point is a mean score based on 10 runs.
Figure 6The effect of number of topics on AUC. (A) Shows predictions based on 20 words only and (B) shows predictions based on the entire topic distribution—each data point is a mean score based on 10 runs.
Figure 7Interpretability vs. AUC_20 zoomed in FLSA_W_FCM and FLSA_W_GK shows a positive correlation between interpretability and predictive performance - each data point represents a trained model.