| Literature DB >> 33955834 |
Meng Ji1, Yanmeng Liu1, Mengdan Zhao1, Ziqing Lyu1,2, Boren Zhang1, Xin Luo3,4, Yanlin Li3,4, Yin Zhong5,6.
Abstract
BACKGROUND: Improving the understandability of health information can significantly increase the cost-effectiveness and efficiency of health education programs for vulnerable populations. There is a pressing need to develop clinically informed computerized tools to enable rapid, reliable assessment of the linguistic understandability of specialized health and medical education resources. This paper fills a critical gap in current patient-oriented health resource development, which requires reliable and accurate evaluation instruments to increase the efficiency and cost-effectiveness of health education resource evaluation.Entities:
Keywords: PEMAT; health education; machine learning; patient-oriented; understandability evaluation
Year: 2021 PMID: 33955834 PMCID: PMC8138706 DOI: 10.2196/28413
Source DB: PubMed Journal: JMIR Med Inform
Natural language features relevant to Patient Education Materials Assessment Tool guidelines.
| Evaluation criteria in the Patient Education Materials Assessment Tool | Language features | Machine learning evaluation | |
|
| |||
|
| The material makes its purpose completely evident. | A1, X1, X2, X7 | Information evidentness |
|
| The material does not include information that distracts from its purpose. | B1, B3 | Relevance to education purpose |
|
| |||
|
| Medical terms are used only to familiarize audience with the terms. | B2 | Domain knowledge |
|
| |||
|
| The material does not expect the user to perform calculations. | N1, N2, N3 | Numeracy demand |
|
| |||
|
| The material presents information in a logical sequence. | Z5, Z6, Z7 | Logical sequence |
Differences between easy and difficult medical texts derived by the Mann-Whitney U test.
| Variables | Easy texts, mean (SD) score | Difficult texts, mean (SD) score | Mann-Whitney | |
| A1 | 14.09 (14.52) | 10.10 (13.13) | 97905.00 | <.001 |
| X1 | 0.42 (3.89) | 0.18 (1.49) | 120325.50 | .02 |
| X2 | 10.41 (11.26) | 6.57 (9.14) | 89487.50 | <.001 |
| X7 | 3.24 (5.41) | 1.79 (3.12) | 103350.50 | <.001 |
| B1 | 17.10 (31.14) | 15.69 (21.78) | 117882.50 | .12 |
| B2 | 15.04 (21.53) | 24.68 (34.04) | 99536.50 | <.001 |
| B3 | 9.25 (14.30) | 12.80 (18.02) | 103338.00 | <.001 |
| N1 | 5.74 (8.54) | 5.42 (6.51) | 123009.00 | .66 |
| N2 | 0.21 (0.70) | 0.21 (0.70) | 123284.50 | .52 |
| N3 | 5.73 (9.38) | 4.77 (5.60) | 120978.50 | .38 |
| Z5 | 133.63 (118.93) | 122.77 (119.05) | 108744.00 | <.001 |
| Z6 | 4.13 (5.28) | 3.01 (5.01) | 100719.00 | <.001 |
| Z7 | 4.22 (4.62) | 2.10 (4.03) | 81063.50 | <.001 |
A1 in difficult texts.
| A1 | Keyword concordances |
| limited to | infections in humans are |
| strains | There are five |
| containment | Previous outbreaks had been limited to remote areas allowing initial |
| combined | This outbreak was unprecedented in scale, being larger than all other outbreaks |
| spread | The virus |
| boundaries | The virus spread across multiple international |
| isolated | Seven other countries had minor outbreaks with nonsustained transmission or |
| events | This article aims to summarize the |
Figure 1Mean receiver operating characteristic curve for the 5 machine learning algorithms. C5: C5 decision tree; LR: logistic regression; MLP: multilayer perceptron; ROC: receiver operating characteristic; RF: random forest; XGB: extreme gradient boosting.
Performance of the 5 machine learning models on predicting language understandability of the health texts for international students in tertiary education.
| Algorithm | Area under the receiver operating characteristic curve, mean (SD) | Sensitivity, mean (SD) | Specificity, mean (SD) | Accuracy, mean (SD) |
| Extreme gradient boosting | 0.979 (0.006) | 0.947 (0.011) | 0.944 (0.011) | 0.945 (0.01) |
| Random forest | 0.967 (0.033) | 0.924 (0.034) | 0.885 (0.094) | 0.904 (0.064) |
| Multilayer perceptron | 0.946 (0.006) | 0.897 (0.006) | 0.893 (0.014) | 0.895 (0.008) |
| C5.0 decision tree | 0.981 (0.005) | 0.95 (0.009) | 0.941 (0.023) | 0.945 (0.014) |
| Logistic regression | 0.804 (0.002) | 0.837 (0.009) | 0.627 (0.016) | 0.732 (0.004) |
Results of the pairwise comparison of the model predictive performance by two-tailed t test.
| Pair number | Comparison | AUCa difference | Sensitivity difference | Specificity difference | Accuracy difference | |||||||
| Mean (SD) | Mean (SD) | Mean (SD) | Mean (SD) | |||||||||
| Pair 1 | XGBb vs RFc | 0.013 (0.034) | .44 | 0.023 (0.034) | .21 | 0.059 (0.089) | .21 | 0.041 (0.062) | .21 | |||
| Pair 2 | XGB vs MLPd | 0.034 (0.007) | <.001e | 0.049 (0.009) | <.001e | 0.051 (0.013) | .001e | 0.050 (0.010) | <.001e | |||
| Pair 3 | XGB vs C5.0 | –0.001 (0.006) | .66 | –0.003 (0.008) | .40 | 0.003 (0.022) | .76 | 0.000 (0.015) | >.99 | |||
| Pair 4 | XGB vs LRf | 0.175 (0.004) | <.001e | 0.109 (0.006) | <.001e | 0.317 (0.021) | <.001e | 0.213 (0.012) | <.001e | |||
| Pair 5 | RF vs MLP | 0.021 (0.036) | .27 | 0.026 (0.037) | .19 | –0.008 (0.095) | .86 | 0.009 (0.066) | .77 | |||
| Pair 6 | RF vs C5.0 | –0.014 (0.029) | .34 | –0.026 (0.032) | .15 | –0.056 (0.076) | .18 | –0.041 (0.054) | .17 | |||
| Pair 7 | RF vs LR | 0.163 (0.033) | <.001e | 0.086 (0.034) | .005e | 0.258 (0.090) | .003e | 0.172 (0.062) | .003e | |||
| Pair 8 | MLP vs C5.0 | –0.035 (0.008) | .001e | –0.052 (0.011) | <.001e | –0.048 (0.023) | .01 | –0.050 (0.016) | .002e | |||
| Pair 9 | MLP vs LR | 0.142 (0.005) | <.001e | 0.060 (0.007) | <.001e | 0.266 (0.015) | <.001e | 0.163 (0.008) | <.001e | |||
| Pair 10 | C5.0 vs LR | 0.177 (0.004) | <.001e | 0.112 (0.010) | <.001e | 0.314 (0.020) | <.001e | 0.213 (0.014) | <.001e | |||
aAUC: area under the receiver operating characteristic curve.
bXGB: extreme gradient boosting.
cRF: random forest.
dMLP: multilayer perceptron.
eSignificant at the adjusted .005 significance level using Bonferroni correction.
fLR: logistic regression.
Mean decrease in the area under the receiver operating characteristic curve of the 5 machine learning algorithms.
| Feature | Predictor variable | Extreme gradient boosting (%) | Random forest (%) | Deep neural networks (%) | C5.0 decision tree (%) | Logistic regression (%) |
| General and abstract terms | A1 | 0.62 | 2.17 | 0.35 | 1.75 | 1.47 |
| Psychological actions, states, processes | X1 | 0.45 | 1.31 | 0.32 | 0.92 | 0.17 |
| Mental actions and processes | X2 | 1.09 | 2.91 | 1.42 | 1.65 | 2.07 |
| Wanting, planning, and | X7 | 0.19 | 1.94 | 0.49 | 0.79 | 0.13 |
| Anatomy and physiology | B1 | 0.75 | 2.97 | 1.15 | 1.19 | 0.50 |
| Health and disease | B2 | 0.99 | 3.16 | 1.19 | 1.39 | 1.23 |
| Medicines and medical | B3 | 0.69 | 1.87 | 1.92 | 2.79 | 0.93 |
| Numbers | N1 | 1.12 | 0.47 | 0.62 | 1.89 | 0.00 |
| Mathematics | N2 | 0.39 | 4.14 | 0.09 | 0.55 | 0.33 |
| Measurement | N3 | 0.15 | 1.27 | 0.89 | 2.12 | 0.10 |
| Grammatical bin | Z5 | 0.65 | 3.37 | 1.85 | 1.12 | 1.47 |
| Negative | Z6 | 0.35 | 1.81 | 1.32 | 1.95 | 0.77 |
| If | Z7 | 0.82 | 2.27 | 2.32 | 2.65 | 3.93 |
Figure 2The impact of different linguistic features on the machine learning algorithms. AUC: area under the receiver operating characteristic curve; C5: C5 decision tree; LR: logistic regression; NN: neural networks; RF: random forest; XGB: extreme gradient boosting. A1: general and abstract terms; X1: psychological actions, states, and processes; X2: mental actions and processes; X7: wanting, planning, and choosing; B1: anatomy and physiology; B2: health and disease; B3: medicines and medical treatment; N1: numbers; N2: mathematics; N3: measurement; Z5: grammatical bin; Z6: negative; Z7: if.
Impact of the different dimensions on the area under the curves of the algorithms.
| Evaluation dimensions | Understandability | Predictor variable | Extreme gradient boosting (%) | Random forest (%) | Deep neural networks (%) | C5.0 decision tree (%) | Logistic regression (%) |
| Dimension 1 | Information evidentness | A1, X1, X2, X7 | 2.35 | 8.33 | 2.58 | 5.11 | 3.84 |
| Dimension 2 | Relevance to education purpose | B1, B3 | 1.44 | 4.84 | 3.07 | 3.98 | 1.43 |
| Dimension 3 | Domain knowledge | B2 | 0.99 | 3.16 | 1.19 | 1.39 | 1.23 |
| Dimension 4 | Numeracy demand | N1, N2, N3 | 1.66 | 5.88 | 1.60 | 4.56 | 0.43 |
| Dimension 5 | Logical sequence | Z5, Z6, Z7 | 1.82 | 7.45 | 5.49 | 5.72 | 6.17 |