Literature DB >> 35468085

Using Natural Language Processing and Machine Learning to Preoperatively Predict Lymph Node Metastasis for Non-Small Cell Lung Cancer With Electronic Medical Records: Development and Validation Study.

Danqing Hu1, Shaolei Li2, Huanyao Zhang1, Nan Wu2, Xudong Lu1.   

Abstract

BACKGROUND: Lymph node metastasis (LNM) is critical for treatment decision making of patients with resectable non-small cell lung cancer, but it is difficult to precisely diagnose preoperatively. Electronic medical records (EMRs) contain a large volume of valuable information about LNM, but some key information is recorded in free text, which hinders its secondary use.
OBJECTIVE: This study aims to develop LNM prediction models based on EMRs using natural language processing (NLP) and machine learning algorithms.
METHODS: We developed a multiturn question answering NLP model to extract features about the primary tumor and lymph nodes from computed tomography (CT) reports. We then combined these features with other structured clinical characteristics to develop LNM prediction models using machine learning algorithms. We conducted extensive experiments to explore the effectiveness of the predictive models and compared them with size criteria based on CT image findings (the maximum short axis diameter of lymph node >10 mm was regarded as a metastatic node) and clinician's evaluation. Since the NLP model may extract features with mistakes, we also calculated the concordance correlation between the predicted probabilities of models using NLP-extracted features and gold standard features to explore the influence of NLP-driven automatic extraction.
RESULTS: Experimental results show that the random forest models achieved the best performances with 0.792 area under the receiver operating characteristic curve (AUC) value and 0.456 average precision (AP) value for pN2 LNM prediction and 0.768 AUC value and 0.524 AP value for pN1&N2 LNM prediction. And all machine learning models outperformed the size criteria and clinician's evaluation. The concordance correlation between the random forest models using NLP-extracted features and gold standard features is 0.950 and improved to 0.984 when the top 5 important NLP-extracted features were replaced with gold standard features.
CONCLUSIONS: The LNM models developed can achieve competitive performance using only limited EMR data such as CT reports and tumor markers in comparison with the clinician's evaluation. The multiturn question answering NLP model can extract features effectively to support the development of LNM prediction models, which may facilitate the clinical application of predictive models. ©Danqing Hu, Shaolei Li, Huanyao Zhang, Nan Wu, Xudong Lu. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 25.04.2022.

Entities:  

Keywords:  algorithm; decision making; electronic medical records; forest modeling; lung cancer; lymph node metastasis prediction; machine learning; natural language processing; non–small cell lung cancer; prediction models

Year:  2022        PMID: 35468085      PMCID: PMC9086872          DOI: 10.2196/35475

Source DB:  PubMed          Journal:  JMIR Med Inform


Introduction

Lung cancer remains the leading cause of cancer death worldwide, representing approximately 1 in 5 (18.0%) cancer deaths [1]. Non–small cell lung cancer (NSCLC) accounts for about 84% of lung cancer cases, and its 5-year relative survival rate is only 25.0% [2], making it one of the biggest threats to human health. Staging of NSCLC is a process to determine the extent of the cancer and is critical to prognosis evaluation and treatment decision making [3,4]. The TNM stage classification [5] is the most widely used staging method in clinical practice; it describes the anatomic extent of a tumor from 3 aspects (ie, T for extent of the primary tumor, N for involvement of lymph nodes, M for distant metastases). For patients with resectable NSCLC, preoperative confirmed N2 (a type of N stage) lymph node metastasis (LNM) indicates neoadjuvant therapy should be given before surgery to achieve the best clinical practice [3]. Currently, various advanced noninvasive diagnostic modalities are available for N staging like chest computed tomography (CT) and positron emission tomography–computed tomography (PET-CT). In clinical practice, clinicians commonly use a size criterion (ie, the maximum short axis diameter of lymph node >10 mm on CT scan) to discriminate LNM from benign nodes and yield 55% sensitivity [6]. Another criterion is the maximum standardized uptake value (SUVmax) of lymph node >2.5 on PET-CT scan, which has an 81% sensitivity [7]. Invasive methods such as mediastinoscopy and endobronchial ultrasound-guided transbronchial needle aspiration have better diagnostic abilities than noninvasive methods. However, these methods are mainly for lymph nodes with indications and not suitable for patients with severe comorbidities, so they are not routinely used in clinical practice [8]. One study analyzed data from 9 clinical trials and found nearly 38% of patients were misclassified in comparison with their pathological N staging [9]. Therefore, new reliable LNM prediction methods are required to alleviate this clinical dilemma. For precise staging, researchers explored using statistical analysis or machine learning methods to learn nontrivial knowledge between the comprehensive patient features and LNM status [8,10-16]. Recently, with the rapid development of hospital information systems, a large volume of electronic medical records (EMR) has become available, and it contains almost all clinical features about patients. However, some important features are recorded in the narratives in free text, such as the size of the tumor and lymph node, tumor density, pleural indentation, etc, which hinders their direct use. Manual extraction is time-consuming and error-prone. So, one big challenge is how to extract this information effectively to support subsequent tasks like LNM prediction [17]. A review by Garg et al [18] found studies in which users were automatically prompted to use the system achieved better performance in comparison with those in which users were required to actively initiate the system. The finding implicitly indicates that the duplicative data entry activity may explain why the predictive models are not widely adopted in the clinic despite their potential to improve diagnostic accuracy. Furthermore, with the prevalence of machine learning models, more features are required for analysis, making the clinical application of the models more difficult [19-21]. Natural language processing (NLP) offers the opportunity to automatically extract information to support the application of predictive models [17,22]. Many studies used rule-based, machine learning, or deep learning methods to extract the cancer-related information from free-text EMR data [22-29], but only a few included further elaboration on how to exploit the extracted information. Chen et al [30] extracted information from various clinical notes including CT reports and operative notes to calculate the Cancer of the Liver Italian Program score. Martinez et al [31] extracted information from pathology reports to calculate the TNM and Australian clinicopathological stage of colorectal cancer. Castro et al [32] developed an NLP system for automated breast imaging reporting and data system (BI-RADS) categories extraction from breast radiology reports. Bozkurt et al [33,34] developed an information extraction pipeline to extract information from mammography reports to predict the malignancy of breast cancer. Sui et al [35] constructed an NLP-based feature generalizing to extract features from free-text EMR data and provided the stage of lung cancer using a Bayesian reasoning network. Yuan et al [36] used NLP tools to extract multiple features from EMRs to estimate survival for patients with lung cancer. Although many studies have explored how to extract the cancer-related information from various types of free-text narratives and some also exploit the extracted information for cancer risk evaluation, diagnosis, and pathological staging, few studies exploit the extracted information from radiological reports for preoperative LNM prediction, especially for NSCLC. In this study, we aim to use EMR data to develop LNM prediction models for NSCLC patients. We first developed a multiturn question answering NLP model to extract the features from CT reports and then combined these features with other clinical characteristics to develop the predictive models. Since the NLP model may produce imperfect extraction results, we also conducted experiments to compare the predicted probabilities between models using NLP-extracted features and gold standard features.

Methods

Patients

We retrospectively analyzed EMR data of 794 patients who underwent surgical resection for NSCLC with systematic mediastinal lymphadenectomy at the Department of Thoracic Surgery II of Peking University Cancer Hospital from 2010 to 2018. All patients underwent contrast-enhanced chest CT images within 2 months before surgical resection. We excluded the patients with preoperative chemotherapy or radiotherapy. The collected EMR includes demographic information, medical history, CT reports, preoperative serum tumor markers, and pathology reports, which can be analyzed to develop the prediction model. For each patient, we also collected the clinical staging that clinicians evaluated before surgery as the baseline to compare with the LNM prediction models.

Ethics Approval

This study was approved by the Ethics Committee of Peking University Cancer Hospital (2019KT59).

Clinical and Pathological LNM Evaluation

In this study, all included patients underwent systematic mediastinal lymphadenectomy during surgical resection. The lymph node tissues were examined by pathologists, and the metastasis results were recorded in the postoperative pathology reports. We reviewed the pathology reports to determine the LNM status and label the pathological N (pN) stage (pN0/pN1/pN2) for each patient based on the 8th edition TNM stage classification [5] as the gold standard. We also used the size criterion (ie, the maximum short axis diameter of lymph node >10 mm on CT scan as positive) to label the clinical N (cN) stages (cN0/cN1/cN2) based on the CT-reported lymph node size. Moreover, we collected the cN stages, which were determined preoperatively by a thoracic surgeon using all available patient data including the information used in this study. The thoracic surgeon has 10 years of experience in lung cancer surgery. The cN stages determined by the size criterion and the thoracic surgeon were regarded as the baselines.

NLP Feature Extraction

As one of the most important preoperative examinations, CT reports record valuable information about the tumors and lymph nodes, which is of paramount importance for staging. However, the free-text nature of CT reports makes it difficult to understand and analyze them using computer programs. In our previous work [27], we developed an information extraction system composed of named entity recognition, relation classification, and postprocessing modules to extract valuable information in a pipeline manner. However, in this pipeline, the subsequent tasks would be influenced by the outputs of former tasks, which may affect the performance of the whole system. Therefore, to alleviate this problem, we applied a multiturn question answering (MTQA) [37] approach to extract information from CT reports in this study. Using the MTQA strategy, we can encode the relation into the question query and jointly model entity and relation in a natural question answering way. Specifically, we first defined 10 questions related to the primary tumor and lymph nodes. All questions are listed in Table 1. Note that there are 2 types of questions (ie, head entity questions and tail entity question templates). In the model training stage, we inserted the annotated head entities into the slots in the tail entity question templates as the tail entity questions. We then used 2 special tokens (ie, CLS and SEP) to concatenate the questions and sentences in the reports as the inputs and annotated entities as the answers to conduct the bidirectional encoder representations from transformers (BERT) model training. In the model test stage, we first concatenated the head entity questions and sentences in the reports as the inputs and applied the trained MTQA model to extract the head entities (ie, tumor and lymph node). If there were any head entities recognized, we inserted the extracted head entities into the slots in the tail entity question templates as the tail entity questions and combined them with sentences in the reports as the inputs to drive the tail entity extraction. A case of the MTQA application is shown in Figure 1. Finally, the extracted head and tail entities are organized as triples, and a rule-based postprocessing algorithm proposed in the previous work [27] is used to process the triples to obtain the standardized NLP-extracted features. Furthermore, the NLP-extracted features were manually reviewed and corrected by a clinician based on the report contents as the gold standard features. In this study, we used BERT [38], an advanced pretrained language representation model, to tag the answer for each question.
Table 1

Questions and entity types for natural language processing–extracted features.

Question (Chinese)Question (English)Answer notationEntity type
Head entity question
原发肿物的相关描述是什么?What is the description about the primary tumor?Head1Tumor
淋巴结的相关描述是什么?What is the description about the lymph nodes?Head2Lymph node
Tail entity question template
Head1 位于什么地方?Where is Head1 located?Tail1Location
Head1 的大小是多少?What is the size of Head1?Tail2Size
Head1 的形状是什么?What is the shape of Head1?Tail3Shape
Head1 的密度是什么?What is the density of Head1?Tail4Density
与Head1 相关的胸膜侵犯的描述是什么?What is the description about the pleura invasion related to Head1?Tail5Pleura
与Head1 相关的血管侵犯的描述是什么?What is the description about the vessel invasion related to Head1?Tail6Vessel
Head2 位于什么地方?Where is Head2 located?Tail7Location
Head2 的大小是多少?What is size of Head2?Tail8Size
Figure 1

A case of multiturn question answering application. BERT: bidirectional encoder representations from transformers.

Questions and entity types for natural language processing–extracted features. A case of multiturn question answering application. BERT: bidirectional encoder representations from transformers.

LNM Prediction

Six machine learning algorithms were applied to develop the LNM prediction models, including logistic regression (LR) [39], L2-logistic regression (L2-LR) [40], random forest (RF) [41], LightGBM (LGBM) [42], support vector machine (SVM) [43], and artificial neural network (ANN) [44]. LR is the conventional classification method, and L2-LR is the LR with the L2 regularization for parameters. RF and LGBM are ensemble methods but with different ways to combine the weak decision trees. SVM is a classical algorithm that constructs hyperplanes in a high- or infinite-dimensional space to classify samples. ANN is a supervised learning algorithm that can learn nonlinear functions between features and targets. LR and L2-LR have good interpretability because the predicted results can be calculated by a simple linear function and a sigmoid transformation. RF and LGBM are also interpretable, in which they can provide the feature importance.

Experimental Setup

In this study, we used the Whole Word Masking version of BERT [45] pretrained on the Chinese Wikipedia corpus as the tagging model in the MTQA. An additional 359 annotated CT reports from our previous work were used to develop and evaluate the MTQA model. We randomly split 70% of CT reports as the training set, 10% as the validation set, and 20% as the test set. A total of 100 of these reports were each annotated by 2 biomedical informatics engineers to calculate the interannotator agreement score using the kappa score. Pipeline methods with bidirectional long short-term memory (BiLSTM) and BERT were selected as the baseline. To obtain the NLP-extracted features for LNM prediction, the MTQA model developed on the 359 reports was used to process the 794 CT reports of included patients. Subsequently, the NLP-extracted features were manually reviewed and corrected by a clinician as the gold standard features. Univariate analysis was performed using the Mann-Whitney U test for continuous features and Pearson chi-square test for categorical features. P<.05 was considered statistically significant. To obtain robust experimental results, a 10-fold cross-validation strategy was first performed on the total data set. The 10-fold cross-validation randomly split the data set into 10 subsets. Each subset was considered as the independent test set and the remaining 9 subsets were considered as the training set. During each fold, a 5-fold cross-validation was applied on the training set to find the optimal hyperparameters for the machine learning algorithms by a grid search. When the optimal hyperparameters were selected, we retrained the prediction model on the training set and tested it on the test set to obtain the final predictive performance. Using this strategy, we can ensure that the test set is always invisible during the model training and hyperparameter tuning and obtain the predicted probability for each case. The hyperparameter spaces are as follows: LR: tol ∈ {1e–3, 1e–4, 1e–5}, max_iter ∈ {500, 1000} L2-LR: C ∈ {10, 1, 0.1}, tol ∈ {1e–3, 1e–4, 1e–5}, max_iter ∈ {500, 1000} RF: n_estimators ∈ {50, 100, 200}, max_depth ∈ {2, 3}, min_samples_leaf ∈ {1, 2} LGBM: n_estimators ∈ {50, 100, 200}, max_depth ∈ {2, 3}, num_leaves ∈ {20, 31, 50}, min_child_samples ∈ {1, 2, 3}, reg_alpha ∈ {2, 3} SVM: C ∈ {10, 1, 0.1, 0.01}, kernel ∈ {‘linear,’ ‘rbf,’ ‘poly’}, tol ∈ {1e–3, 1e–4, 1e–5} ANN: hidden_layer_sizes ∈ {5, 10, 30}, learning_rate ∈ {1e–2, 1e–3, 1e–4}, alpha ∈ {1e–3, 1e–4, 1e–5} We applied the receiver operating characteristic (ROC) curve to evaluate the diagnostic performances of the machine learning models. Besides the ROC curve, we also used the precision-recall (PR) curve to test the models because the ROC curve pays attention to sensitivity and specificity but ignores precision. The mean area under the receiver operating characteristic curve (AUC) and average precision (AP) values with standard derivations were calculated based on the 10-fold cross-validation results. We also drew the ROC curves and PR curves to compare with the size criterion (maximum short axis diameter of lymph node >10 mm on CT) and the clinician’s evaluation. All LNM prediction models were developed using the Scikit-learn 0.24.1 and LightGBM 3.2.0 Python packages. All statistical analyses were conducted using SciPy 1.6.2 Python package.

Results

Patient Characteristics

Table 2 shows the characteristics of all 794 patients. Univariate analysis was performed for all collected features, and 13.2% (105/794) of patients had pN2 LNM. Sex, age, drinking history, family history, and disease history are not significantly associated with the pN2. The pN2 occurred more frequently in smokers (P=.04). The long and short axis diameters of the tumor in pN2 patients are significantly larger than those in pN0 and pN1 patients (both P<.001). Patients with solid nodules are more likely to have pN2 (P<.001). Other morphological characteristics of tumor-like lobulation and pleural indentation are more likely to occur in pN2 patients (P=.006 and P=.003, respectively), but spiculation and vessel invasion present no significant differences between pN2 and other patients. Using 10 mm as the size criterion, the maximum long and short axis diameters of the hilar and mediastinal lymph nodes show significant differences between the 2 groups (P=.008, P<.001, P<.001, and P<.001, respectively). Among all 6 serum tumor biomarkers, carcinoembryonic antigen (CEA), carbohydrate antigen 12-5 (CA125), and neuron-specific enolase (NSE) show significant differences between the 2 groups (P<.001, P<.001, and P=.048, respectively).
Table 2

Patient characteristics.

Total (n=794)LNMa statusP value
pN2b (n=105)pN0c or pN1d (n=689)
Age (years), mean (SD)60.92 (51.48 to 70.36)60.87 (51.87 to 69.86)60.93 (51.42 to 70.44).45
Sex, n (%) e.06
Male39762335
Female39743354
Smoking history, n (%) .04
Yes33755282
No45750407
Drinking history, n (%) .94
Yes18325158
No61180531
Family history, n (%) .32
Yes13714123
No65791566
Hypertension, n (%) .18
Yes23237195
No56268494
Diabetes, n (%) .25
Yes841569
No71090620
Pulmonary tuberculosis, n (%) .33
Yes33231
No761103658
Cardiovascular disease, n (%) .06
Yes36927
No75896662
Cerebrovascular disease, n (%) .35
Yes29623
No76599666
Tumor locationf, n (%) .22
RULg24927222
RMLh59455
RLLi15018132
LULj18531154
LLLk12621105
Other25421
TLAf,l, median (IQR)2.61 (1.20 to 4.01)3.02 (1.64 to 4.39)2.55 (1.15 to 3.94)<.001
TSAf,m, median (IQR)2.03 (0.88 to 3.18)2.38 (1.27 to 3.48)1.98 (0.83 to 3.13)<.001
Spiculationf, n (%) .08
Yes25542213
No53963476
Lobulationf, n (%) <.001
Yes21148163
No58357526
Tumor densityf, n (%) <.001
pGGOn1240124
mGGOo96393
Solid nodule574102472
Vessel invasionf, n (%) .87
Yes52646
No74299643
Pleural indentationf, n (%) .001
Yes40670336
No38835353
HLNLAf,p, n (%) .008
>10 mm14830118
≤10 mm64675571
HLNSAf,q, n (%) <.001
>10 mm661947
≤10 mm72886642
MLNLAf,r, n (%) <.001
>10 mm19150141
≤10 mm60355548
MLNSAf,s, n (%) <.001
>10 mm722745
≤10 mm72278644
CEAt, median (IQR)5.31 (–6.66 to 17.27)12.66 (–8.44 to 33.76)4.18 (–5.17 to 13.54)<.001
CA199u, median (IQR)14.41 (–3.24 to 32.06)15.80 (–5.08 to 36.68)14.20 (–2.90 to 31.29).47
CA125v, median (IQR)14.46 (0.03 to 28.90)19.88 (–5.56 to 45.32)13.64 (1.96 to 25.32)<.001
NSEw, median (IQR)15.81 (8.85 to 22.78)16.26 (10.19 to 22.33)15.75 (8.66 to 22.83).048
Cyfra211x, median (IQR)3.20 (–0.23 to 6.62)3.55 (–0.64 to 7.75)3.14 (–0.15 to 6.43).06
SCCAgy, median (IQR)0.96 (–0.16 to 2.08)1.18 (–0.62 to 2.99)0.93 (–0.04 to 1.90).14

aLNM: lymph node metastasis.

bpN2: pathological N stage 2.

cpN0: pathological N stage 0.

dpN1: pathological N stage 1.

eNot applicable.

fFeatures recorded in computed tomography reports.

gRUL: right upper lobe.

hRML: right middle lobe.

iRLL: right lower lobe.

jLUL: left upper lobe.

kLLL: left lower lobe.

lTLA: tumor long axis.

mTSA: tumor short axis

npGGO: pure ground glass opacity.

omGGO: mixed ground glass opacity.

pHLNLA: hilar lymph node long axis.

qHLNSA: hilar lymph node short axis.

rMLNLA: mediastinal lymph node long axis.

sMLNSA: mediastinal lymph node short axis.

tCEA: carcinoembryonic antigen.

uCA199: carbohydrate antigen 19-9.

vCA125: carbohydrate antigen 12-5.

wNSA: neuron-specific enolase.

xCyfra211: cytokeratin 19-fragments.

ySCCAg: squamous cell carcinoma antigen.

Patient characteristics. aLNM: lymph node metastasis. bpN2: pathological N stage 2. cpN0: pathological N stage 0. dpN1: pathological N stage 1. eNot applicable. fFeatures recorded in computed tomography reports. gRUL: right upper lobe. hRML: right middle lobe. iRLL: right lower lobe. jLUL: left upper lobe. kLLL: left lower lobe. lTLA: tumor long axis. mTSA: tumor short axis npGGO: pure ground glass opacity. omGGO: mixed ground glass opacity. pHLNLA: hilar lymph node long axis. qHLNSA: hilar lymph node short axis. rMLNLA: mediastinal lymph node long axis. sMLNSA: mediastinal lymph node short axis. tCEA: carcinoembryonic antigen. uCA199: carbohydrate antigen 19-9. vCA125: carbohydrate antigen 12-5. wNSA: neuron-specific enolase. xCyfra211: cytokeratin 19-fragments. ySCCAg: squamous cell carcinoma antigen.

Performance of pN2 LNM Prediction Models

As preoperative confirmed N2 indicating neoadjuvant therapy should be given before surgery, we first developed machine learning models to predict the pN2 LNM. We regarded the pN2 patients as positive and pN0 and pN1 patients as negative to train the predictive models. To obtain reliable models, we used the gold standard features instead of NLP-extracted features in this section. Table 3 shows the performances of all models. The RF model achieved the highest averaged AUC value with 0.792 and the LGBM model achieved the highest averaged AP value with 0.457 while all models’ 95% CI are overlapping with each other. The LR obtained a competitive performance in comparison with ANN and SVM. The L2-LR did not obtain improvements in AUC value and AP value compared with the LR. To compare with the size criterion and clinician’s evaluation, we used the probabilities predicted during the 10-fold cross-validation to draw the ROC and PR curves. Figure 2 shows the ROC curves and PR curves of pN2 prediction models and the results of the size criterion and clinician’s evaluation. From Figure 2 we can notice all the ROC curves and PR curves are above the points of size criterion and clinician’s evaluation, which indicates the developed pN2 prediction models not only have better discriminative ability than the diagnostic size criterion used in the clinical practice but also may exceed the clinician in pN2 LNM evaluation.
Table 3

Performances of pN2 lymph node metastasis prediction models.

ModelAUCaAPb
MeanSD95% CIMeanSD95% CI
LRc0.7780.0410.747-0.8090.4420.0750.385-0.499
L2-LRd0.7680.0380.739-0.7960.4130.0720.359-0.467
ANNe0.7690.0510.730-0.8080.4340.0950.363-0.506
SVMf0.7710.0710.718-0.8250.4530.0840.389-0.516
RFg0.7920.0420.760-0.8250.4560.0750.399-0.512
LGBMh0.7870.0440.755-0.8200.4570.1010.381-0.534

aAUC: area under the receiver operating characteristic curve.

bAP: average precision.

cLR: logistic regression.

dL2-LR: L2-logistic regression.

eANN: artificial neural network.

fSVM: support vector machine.

gRF: random forest.

hLGBM: LightGBM.

Figure 2

The receiver operating characteristic curve (A) and precision-recall curves (B) of pN2 prediction models.

Performances of pN2 lymph node metastasis prediction models. aAUC: area under the receiver operating characteristic curve. bAP: average precision. cLR: logistic regression. dL2-LR: L2-logistic regression. eANN: artificial neural network. fSVM: support vector machine. gRF: random forest. hLGBM: LightGBM. The receiver operating characteristic curve (A) and precision-recall curves (B) of pN2 prediction models.

Performance of pN1&N2 LNM Prediction Models

Besides predicting pN2 LNM, we also developed machine learning models to predict the pN1&N2 LNM by regarding patients with pN1 or pN2 LNM as positive. The model training and evaluation processes are the same as pN2 LNM prediction. Table 4 shows the performances of the machine learning models for pN1&N2 LNM prediction. LGBM obtained the highest averaged AUC value with 0.771. The RF model achieved a comparable performance in comparison with LGBM. As in pN2 prediction, LGBM and RF obtained better predictive performances than other models. Figure 3 shows the ROC curves and PR curves of pN1&N2 LNM prediction models. The curves of the machine learning models are also all above the points of the size criterion and clinician’s evaluation.
Table 4

Performances of pN1&N2 lymph node metastasis prediction models.

ModelAUCaAPb
MeanSD95% CIMeanSD95% CI
LRc0.7400.0350.714-0.7660.4670.0580.423-0.510
L2-LRd0.7360.0440.704-0.7690.4650.0580.422-0.509
ANNe0.7340.0470.698-0.7700.4790.0870.413-0.545
SVMf0.7350.0230.717-0.7520.4740.0470.439-0.509
LGBMg0.7680.0300.745-0.7910.5240.0440.491-0.557
RFh0.7710.0260.752-0.7910.5240.0570.481-0.567

aAUC: area under the receiver operating characteristic curve.

bAP: average precision.

cLR: logistic regression.

dL2-LR: L2-logistic regression.

eANN: artificial neural network.

fSVM: support vector machine.

gRF: random forest.

hLGBM: LightGBM.

Figure 3

The receiver operating characteristic curve (A) and precision-recall curves (B) of pN1&N2 prediction models.

Performances of pN1&N2 lymph node metastasis prediction models. aAUC: area under the receiver operating characteristic curve. bAP: average precision. cLR: logistic regression. dL2-LR: L2-logistic regression. eANN: artificial neural network. fSVM: support vector machine. gRF: random forest. hLGBM: LightGBM. The receiver operating characteristic curve (A) and precision-recall curves (B) of pN1&N2 prediction models.

Feature Importance

Among all machine learning models, the LR, L2-LR, RF, and LGBM can provide the feature importance. Table 5 shows the top 10 important features of LR, L2-LR, RF, and LGBM for pN2 LNM prediction. The features were ranked by averaging the weights of models developed from 10-fold cross validation. Note that the LR and L2-LR models provide weights with signs, so we used the absolute values to rank the features. Because the weight magnitudes from different models vary greatly, we used the averaged rankings of features, but not the averaged weights, to find the most important features among the 4 types of models. The CEA is ranked as the most important feature to increase the risk of pN2 LNM by all models. Features recorded in CT reports account for at least half of the top 10 important features, indicating these features are of great importance for pN2 LNM prediction.
Table 5

Top 10 important features for pN2 lymph node metastasis prediction.

RankLRaL2-LRbRFcLGBMdAll
FeatureWeightFeatureWeightFeatureWeightFeatureWeight
1pGGOe,f–10.383CEAg3.530CEA0.229CEA46.0CEA
2CEA6.010CA125h3.067CA1250.094Age23.3Solid nodulef
3CA1254.728pGGOf–1.799Solid nodulef0.094Solid nodulef18.8CA125
4Solid nodulef3.683Solid nodulef1.773MLNSAf,i0.073TLAf,j17.6Age
5TLAf–2.701Age–1.315MLNLAf,k0.072TSAf,l15.1MLNLAf
6Age–1.908SCCAgm0.944TLAf0.054CA12513.3TLAf
7SCCAg1.763MLNLAf0.896TSAf0.048Cyfra211n12.9pGGOf
8mGGOf,o1.759Pleural indentationf0.836Cyfra2110.038NSEp12.7SCCAg
9RMLf,q–1.729Cardiovascular disease0.807SCCAg0.037MLNLAf11.6Lobulationf
10TSAf1.601Lobulationf0.725Lobulationf0.036SCCAg9.0TSAf

aLR: logistic regression.

bL2-LR: L2-logistic regression.

cRF: random forest.

dLGBM: LightGBM.

epGGO: pure ground glass opacity.

fFeatures recorded in computed tomography reports.

gCEA: carcinoembryonic antigen.

hCA125: carbohydrate antigen 12-5.

iMLNSA: mediastinal lymph node short axis.

jTLA: tumor long axis.

kMLNLA: mediastinal lymph node long axis.

lTSA: tumor short axis.

mSCCAg: squamous cell carcinoma antigen.

nCyfra211: cytokeratin 19-fragments.

omGGO: mixed ground glass opacity.

pNSE: neuron-specific enolase.

qRML: right middle lobe.

Top 10 important features for pN2 lymph node metastasis prediction. aLR: logistic regression. bL2-LR: L2-logistic regression. cRF: random forest. dLGBM: LightGBM. epGGO: pure ground glass opacity. fFeatures recorded in computed tomography reports. gCEA: carcinoembryonic antigen. hCA125: carbohydrate antigen 12-5. iMLNSA: mediastinal lymph node short axis. jTLA: tumor long axis. kMLNLA: mediastinal lymph node long axis. lTSA: tumor short axis. mSCCAg: squamous cell carcinoma antigen. nCyfra211: cytokeratin 19-fragments. omGGO: mixed ground glass opacity. pNSE: neuron-specific enolase. qRML: right middle lobe.

NLP-Extracted Features Versus Gold Standard Features

In this study, we applied the MTQA model to extract important features from CT reports to support the development of LNM prediction models. In this section, we first conduct experiments to explore the effectiveness of the MTQA model on feature extraction and then analyze the influence of imperfect extraction results on LNM prediction. We used an additional 359 annotated CT reports to develop the MTQA model. The interannotator agreement score was 0.937 based on the 100 reports annotated by 2 annotators. Table 6 shows the performances of the MTQA model and the pipeline models on the test set. We can notice that the BERT-MTQA model achieved significant improvement compared with the pipeline models.
Table 6

Performance of the multiturn question answering model and baseline models.

FeatureBiLSTMa-pipelineBERTb-pipelineBERT-MTQAc
PdReFfPRFPRF
Tumor density0.8820.6250.7320.8890.6670.7620.9380.9380.938
MLNLAg1.0000.6400.7801.0000.7200.8371.0000.9600.980
TLAh0.9670.8920.9280.9840.9380.9610.9840.9540.969
Lobulation0.8890.5330.6670.9090.6670.7691.0000.8670.929
TSAi0.9670.8920.9280.9840.9380.9610.9840.9540.969
MLNSAj1.0000.7500.8571.0000.7500.8571.0000.9380.968
Pleural indentation0.9310.8180.8710.9640.8180.8851.0000.8480.918
Tumor location0.9840.8970.9380.9680.8970.9310.9850.9850.985
Spiculation1.0000.7270.8421.0000.7730.8721.0001.0001.000
Vessel invasion1.0000.1110.2001.0000.2220.3641.0000.5560.714
HLNLAk1.0000.7780.8751.0000.8330.9091.0001.0001.000
HLNSAl1.0000.7500.8571.0000.7500.8571.0001.0001.000
Average0.9680.7010.7900.9750.7480.8300.9910.9170.948

aBiLSTM: bidirectional long short-term memory.

bBERT: bidirectional encoder representations from transformers.

cMTQA: multiturn question answering.

dP: precision.

eR: recall.

fF: F1 score.

gMLNLA: mediastinal lymph node long axis.

hTLA: tumor long axis.

iTSA: tumor short axis.

jMLNSA: mediastinal lymph node short axis.

kHLNLA: hilar lymph node long axis.

lHLNSA: hilar lymph node short axis.

Table 7 illustrates the performance of the BERT-MTQA model on the 794 CT reports of included patients. We can notice that the accuracy values of all extracted features are higher than 0.90. The F1 scores are higher than 0.90 except for lobulation, tumor density, vessel invasion, and hilar lymph node long axis. For the NLP-extracted features ranked in the top 10 important features, the mediastinal lymph node long axis (MLNLA), tumor long axis (TLA), and tumor short axis (TSA) obtained good accuracy values and F1 scores, but the F1 scores of tumor density and lobulation are not higher than 0.90.
Table 7

Performance of the multiturn question answering model for feature extraction.

FeatureAccuracyPrecisionRecallF1 score
Tumor density0.9400.8750.9150.893
MLNLAa0.9650.9270.9270.927
TLAb0.9740.9740.9740.974
Lobulation0.9230.9930.7160.832
TSAc0.9720.9720.9720.972
MLNSAd0.9860.9180.9310.924
Pleural indentation0.9170.9030.9380.920
Tumor location0.9940.9900.9900.990
Spiculation0.9790.9880.9450.966
Vessel invasion0.9820.9320.7880.854
HLNLAe0.9651.0000.8110.896
HLNSAf0.9860.9820.8480.911

aMLNLA: mediastinal lymph node long axis.

bTLA: tumor long axis.

cTSA: tumor short axis.

dMLNSA: mediastinal lymph node short axis.

eHLNLA: hilar lymph node long axis.

fHLNSA: hilar lymph node short axis.

In this study, the MTQA model generates imperfect extractions, which may influence the subsequent application. To analyze the influence on the pN2 LNM prediction, we calculated the Pearson correlation between the predicted probabilities of models using NLP-extracted features and gold standard features. Moreover, we also replaced the NLP-extracted feature with the gold standard feature one by one according to their importance in Table 5 to explore the changes in the consistency. Figure 4 shows the concordance correlations of the pN2 LNM prediction models. The RF model obtained a high concordance correlation with 0.950 when using all NLP-extracted features in comparison with using gold standard features, and the correlation increased to 0.984 when replacing top 5 important NLP-extracted features. The correlation values of the LR, L2-LR, LGBM, and SVM models were more influenced by using the NLP-extracted features. With the replacement of gold standard features, the correlation values gradually increased and exceeded 0.950. The ANN model did not achieve a good concordance correlation even when the top 5 important NLP-extracted features were replaced.
Figure 4

Concordance correlation values between pN2 prediction models using complete and partial gold standard features. LR: logistic regression; L2-LR: L2-logistic regression; RF: random forest; LGBM: LightGBM; SVM: support vector machine; ANN: artificial neural network: NLP: natural language processing; pGGO: pure ground glass opacity; MLNLA: mediastinal lymph node long axis; TLA: tumor long axis; TSA: tumor short axis.

Performance of the multiturn question answering model and baseline models. aBiLSTM: bidirectional long short-term memory. bBERT: bidirectional encoder representations from transformers. cMTQA: multiturn question answering. dP: precision. eR: recall. fF: F1 score. gMLNLA: mediastinal lymph node long axis. hTLA: tumor long axis. iTSA: tumor short axis. jMLNSA: mediastinal lymph node short axis. kHLNLA: hilar lymph node long axis. lHLNSA: hilar lymph node short axis. Performance of the multiturn question answering model for feature extraction. aMLNLA: mediastinal lymph node long axis. bTLA: tumor long axis. cTSA: tumor short axis. dMLNSA: mediastinal lymph node short axis. eHLNLA: hilar lymph node long axis. fHLNSA: hilar lymph node short axis. Concordance correlation values between pN2 prediction models using complete and partial gold standard features. LR: logistic regression; L2-LR: L2-logistic regression; RF: random forest; LGBM: LightGBM; SVM: support vector machine; ANN: artificial neural network: NLP: natural language processing; pGGO: pure ground glass opacity; MLNLA: mediastinal lymph node long axis; TLA: tumor long axis; TSA: tumor short axis.

Discussion

Principal Findings

In this study, we explored the feasibility of using EMR to develop machine learning models to predict LNM for patients with NSCLC. The important features about the primary tumor and lymph nodes were extracted from the CT reports using NLP technique to support the model development. To the best of our knowledge, this is the first study to use NLP technique to extract features to build preoperative LNM prediction models for patients with NSCLC. Experimental results indicate that the RF model achieved the best performances with 0.792 AUC value and 0.456 AP value for pN2 LNM prediction. All machine learning models outperformed the size criterion and clinician’s evaluation. Among all models, the LR, L2-LR, RF, and LGBM provide the feature importance to show the connections between the patient features and LNM status. CEA, tumor density, CA125, MLNLA, TLA, lobulation, and TSA were ranked in the top 10 important features by the machine learning models, which was consistent with the results of univariate analysis. Squamous cell carcinoma antigen (SCCAg) was also identified as a top 10 important feature by the models, although univariate analysis did not show significance. However, SCCAg has been proved to be associated with LNM in esophageal squamous cell carcinoma [46], anus squamous cell carcinoma [47], oral-cavity squamous cell carcinoma [48], and cervical squamous cell carcinoma [49]. It is also a poor prognostic factor of lung squamous cell carcinoma and upgrading the patient stage is recommended [50,51]. Surprisingly, TLA was identified as an important feature with negative weight by the LR model, which means the longer the TLA is, the lower the risk of pN2 LNM the patient may have. The result is contrary to the result of univariate analysis and may be caused by multicollinearity or interactions between the features [52]. In the L2-LR model, the TLA was not ranked in the top 10 important features, indicating the L2 regularization can indeed reduce the influence of multicollinearity and improve the interpretability of the model [53]. In addition, other features like right middle lobe cardiovascular disease also suffered interpretability problems, which may be hard to accept in clinical practice. Therefore, more robust interpretable machine learning algorithms are needed to make accurate predictions while giving more reasonable explanations. In this study, we innovatively extracted features from CT reports and used them to develop LNM prediction models. The concordance correlations between the predicted probabilities of models using NLP-extracted features, partially NLP-extracted features, and gold standard features indicate that the automatically developed models can obtain similar predictive results to those of models using gold standard features. This finding implicitly indicates it is possible to build models using a large amount of unstructured data and update them automatically. More importantly, it can also reduce the burden of manual feature extraction to improve the usability of the prediction models in clinical practice.

Limitations

Although the experimental results show that machine learning models using CT reports, demographic information, medical history, and biomarker data can achieve better performances than the size criterion and clinician’s evaluation on the collected data, external validation is still needed to further prove the effectiveness and generalization of the NLP and LNM prediction models. Note that the writing styles of CT reports from different medical centers may vary greatly, which poses a huge challenge to the NLP model developed using the CT reports from a single medical center. Transfer learning is a proper strategy to solve the problem by fine-tuning the model to adapt to CT reports from other centers. Overall, multicenter data is necessary to develop a more robust and generalizable NLP and LNM prediction model. Furthermore, many studies have proved that there are deep features or radiomics features related to LNM in the CT images [54-60]. Clinicians cannot recognize these with the naked eye, so these features may provide extra information about the metastasis status. In the future, we will extract the image features and combine them with the features in this study to develop more robust, accurate multimodal LNM prediction models.

Conclusions

In this study, we used NLP and machine learning methods to develop the LNM prediction models for patients with NSCLC using EMRs. The RF model achieved the best performance with 0.792 AUC value and 0.456 AP value for pN2 prediction and 0.768 AUC value and 0.524 AP value for pN1&N2 prediction. All machine learning models outperformed the size criterion and clinician’s evaluation. Furthermore, the experimental results indicate that the NLP model can effectively extract features from CT reports to support the automatic development and update of the LNM prediction model and may facilitate the application of models in clinical practice.
  46 in total

1.  Development and validation of a clinical prediction model for N2 lymph node metastasis in non-small cell lung cancer.

Authors:  Kezhong Chen; Fang Yang; Guanchao Jiang; Jianfeng Li; Jun Wang
Journal:  Ann Thorac Surg       Date:  2013-08-30       Impact factor: 4.330

2.  A prediction model for pathologic N2 disease in lung cancer patients with a negative mediastinum by positron emission tomography.

Authors:  Farhood Farjah; Feiran Lou; Camelia Sima; Valerie W Rusch; Nabil P Rizk
Journal:  J Thorac Oncol       Date:  2013-09       Impact factor: 15.609

3.  FDG-PET/CT imaging for mediastinal staging in patients with potentially resectable non-small cell lung cancer.

Authors:  Mia Schmidt-Hansen; David Raymond Baldwin; Javier Zamora
Journal:  JAMA       Date:  2015-04-14       Impact factor: 56.272

4.  Automatic abstraction of imaging observations with their characteristics from mammography reports.

Authors:  Selen Bozkurt; Jafi A Lipson; Utku Senol; Daniel L Rubin
Journal:  J Am Med Inform Assoc       Date:  2014-10-28       Impact factor: 4.497

5.  Information extraction from multi-institutional radiology reports.

Authors:  Saeed Hassanpour; Curtis P Langlotz
Journal:  Artif Intell Med       Date:  2015-10-03       Impact factor: 5.326

6.  Correlation Between Squamous Cell Carcinoma Antigen Level and the Clinicopathological Features of Early-Stage Cervical Squamous Cell Carcinoma and the Predictive Value of Squamous Cell Carcinoma Antigen Combined With Computed Tomography Scan for Lymph Node Metastasis.

Authors:  Dianbo Xu; Danbo Wang; Shuo Wang; Ye Tian; Zaiqiu Long; Xuemei Ren
Journal:  Int J Gynecol Cancer       Date:  2017-11       Impact factor: 3.437

7.  DeepPhe: A Natural Language Processing System for Extracting Cancer Phenotypes from Clinical Records.

Authors:  Guergana K Savova; Eugene Tseytlin; Sean Finan; Melissa Castine; Timothy Miller; Olga Medvedeva; David Harris; Harry Hochheiser; Chen Lin; Girish Chavan; Rebecca S Jacobson
Journal:  Cancer Res       Date:  2017-11-01       Impact factor: 12.701

8.  Methods for staging non-small cell lung cancer: Diagnosis and management of lung cancer, 3rd ed: American College of Chest Physicians evidence-based clinical practice guidelines.

Authors:  Gerard A Silvestri; Anne V Gonzalez; Michael A Jantz; Mitchell L Margolis; Michael K Gould; Lynn T Tanoue; Loren J Harris; Frank C Detterbeck
Journal:  Chest       Date:  2013-05       Impact factor: 9.410

9.  Prognostic factors based on clinicopathological data among the patients with resected peripheral squamous cell carcinomas of the lung.

Authors:  Tomonari Kinoshita; Takashi Ohtsuka; Tai Hato; Taichiro Goto; Ikuo Kamiyama; Atsushi Tajima; Katsura Emoto; Yuichiro Hayashi; Mitsutomo Kohno
Journal:  J Thorac Oncol       Date:  2014-12       Impact factor: 15.609

10.  Performance of a Machine Learning Algorithm Using Electronic Health Record Data to Identify and Estimate Survival in a Longitudinal Cohort of Patients With Lung Cancer.

Authors:  Qianyu Yuan; Tianrun Cai; Chuan Hong; Mulong Du; Bruce E Johnson; Michael Lanuti; Tianxi Cai; David C Christiani
Journal:  JAMA Netw Open       Date:  2021-07-01
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.