| Literature DB >> 34330259 |
Kun Zeng1, Yibin Xu1, Ge Lin2, Likeng Liang3, Tianyong Hao4.
Abstract
BACKGROUND: Eligibility criteria are the primary strategy for screening the target participants of a clinical trial. Automated classification of clinical trial eligibility criteria text by using machine learning methods improves recruitment efficiency to reduce the cost of clinical research. However, existing methods suffer from poor classification performance due to the complexity and imbalance of eligibility criteria text data.Entities:
Keywords: Clinical trial; Eligibility criteria classification; Ensemble learning; Focal loss; Metric learning
Year: 2021 PMID: 34330259 PMCID: PMC8323220 DOI: 10.1186/s12911-021-01492-z
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 2.796
Fig. 1The framework of the ensemble learning-based model consists of a preprocessing layer, a single model layer integrating 5 pre-trained models including BERT, XLNet, ERNIE, RoBERTa, and ELECTRA, as well as an ensemble layer containing Soft Voting to output prediction result
Fig. 2The overall architecture of the single models
Fig. 3Histogram distributions of the training set, validation set, and test set. X-axis represents counts of data instances and Y-axis represents categories
Examples of eligibility criteria text and corresponding annotated categories
| Eligibility criteria text | Category |
|---|---|
| 年龄 > 80岁 (Age > 80) | Age |
| 近期颅内或椎管内手术史 (recent intracranial or spinal canal surgery) | Therapy or surgery |
| 血糖 < 2.7 mmol/L (Blood glucose < 2.7 mmol/L) | Laboratory examinations |
| 性别不限,年龄18 ~ 70岁 (unlimited gender, age 18–70) | Multiple |
| 合并造血系统或恶性肿瘤等严重原发性疾病(complicated with serious primary disease such as hematopoietic system or malignant tumor) | Disease |
| 其他研究者认为不适合参加本研究的患者 (patients that unsuitable for this study considered by other investigators) | Researcher decision |
| 预期生存超过12周 (expected survival over 12 weeks) | Life expectancy |
| 男、女不限 (male or female) | Gender |
The performances of our model and baseline models on the same dataset
| Model | Accuracy | Precision | Recall | Macro F1 |
|---|---|---|---|---|
| TextCNN | 0.8256 | 0.8074 | 0.7538 | 0.7696 |
| TextRNN | 0.8094 | 0.7262 | 0.7369 | 0.7258 |
| TextRCNN | 0.8256 | 0.7894 | 0.7678 | 0.7704 |
| FastText | 0.8116 | 0.7732 | 0.7268 | 0.7385 |
| Transformer | 0.7934 | 0.7545 | 0.6469 | 0.6721 |
| BERT | 0.8385 | 0.8055 | 0.7980 | 0.7973 |
| XLNet | 0.8508 | 0.8164 | 0.8011 | 0.803 |
| ERNIE | 0.8382 | 0.8035 | 0.7969 | 0.7952 |
| RoBERTa | 0.8439 | 0.7929 | 0.8215 | 0.7992 |
| ELECTRA | 0.8324 | 0.7935 | 0.791 | 0.7862 |
| Our model | 0.850 | 0.825 | 0.821 | 0.8167 |
Performance comparison of all single models with or without metric learning using macro F1 score (margin parameter m = 0.1)
| Model | Without metric learning | With metric learning | Increase rate (%) |
|---|---|---|---|
| BERT | 0.7880 | 0.7973 | 1.18 |
| XLNet | 0.7983 | 0.8030 | 0.59 |
| RoBERTa | 0.7951 | 0.7992 | 0.52 |
| ERNIE | 0.7865 | 0.7952 | 1.11 |
| ELECTRA | 0.7758 | 0.7862 | 1.34 |
Performance comparison of all single models with cross entropy loss or focal loss functions using macro F1 score
| Model | Cross entropy loss | Focal loss |
|---|---|---|
| BERT | 0.7902 | 0.7973 |
| XLNet | 0.7987 | 0.8030 |
| RoBERTa | 0.7959 | 0.7992 |
| ERNIE | 0.7868 | 0.7952 |
| ELECTRA | 0.7804 | 0.7862 |
Fig. 4Performance of single models based on BERT and XLNet pre-training models under different percentages of data volume