Literature DB >> 33381748

Natural language processing systems for pathology parsing in limited data environments with uncertainty estimation.

Anobel Y Odisho1, Briton Park2, Nicholas Altieri2, John DeNero3, Matthew R Cooperberg1,4, Peter R Carroll1, Bin Yu2,3,5.   

Abstract

OBJECTIVE: Cancer is a leading cause of death, but much of the diagnostic information is stored as unstructured data in pathology reports. We aim to improve uncertainty estimates of machine learning-based pathology parsers and evaluate performance in low data settings.
MATERIALS AND METHODS: Our data comes from the Urologic Outcomes Database at UCSF which includes 3232 annotated prostate cancer pathology reports from 2001 to 2018. We approach 17 separate information extraction tasks, involving a wide range of pathologic features. To handle the diverse range of fields, we required 2 statistical models, a document classification method for pathologic features with a small set of possible values and a token extraction method for pathologic features with a large set of values. For each model, we used isotonic calibration to improve the model's estimates of its likelihood of being correct.
RESULTS: Our best document classifier method, a convolutional neural network, achieves a weighted F1 score of 0.97 averaged over 12 fields and our best extraction method achieves an accuracy of 0.93 averaged over 5 fields. The performance saturates as a function of dataset size with as few as 128 data points. Furthermore, while our document classifier methods have reliable uncertainty estimates, our extraction-based methods do not, but after isotonic calibration, expected calibration error drops to below 0.03 for all extraction fields.
CONCLUSIONS: We find that when applying machine learning to pathology parsing, large datasets may not always be needed, and that calibration methods can improve the reliability of uncertainty estimates.
© The Author(s) 2020. Published by Oxford University Press on behalf of the American Medical Informatics Association.

Entities:  

Keywords:  cancer; information extraction; machine learning; natural language processing; pathology; prostate cancer

Year:  2020        PMID: 33381748      PMCID: PMC7751177          DOI: 10.1093/jamiaopen/ooaa029

Source DB:  PubMed          Journal:  JAMIA Open        ISSN: 2574-2531


INTRODUCTION

An estimated 1.8 million Americans will be diagnosed with cancer in 2020. In nearly all cases, diagnosis is made via tissue analysis, described in detail in a pathology report, which is stored in most electronic medical record systems as unstructured free text. Without manual data abstraction, these important details are unavailable for scalable and algorithmic approaches for case identification, risk stratification, prognostication, treatment selection, clinical trial screening, and surveillance., Moreover, access to these data in structured formats can drive algorithmic personalized treatment strategies based on pathologic information. For nearly 50 years investigators have worked to develop natural language processing (NLP) algorithms to extract these details from pathology reports., However, only a limited number of categorical data elements are typically extracted and model outputs often lack reliable uncertainty estimates, limiting the clinical applicability of these systems, only 10% of which have been reported to be in real-world use. Parsing pathology reports has traditionally been approached using rule-based methods. However, designing rules is labor intensive and requires deep involvement of clinical experts. The complexity and conflicts between rules grow rapidly as the number of rules increases, and as the underlying documents shift, rules quickly become ineffective. NLP has been applied to pathology report information extraction with promising results, using both classic NLP (boosting over a bag-of-n-grams representation of the document) and deep learning approaches (convolutional, recurrent, and hierarchical attention networks)., While most work focuses on classification tasks involving fields with a small number of labels (such as histology or margin status), Li and Martinez investigate categorical fields as well as numeric fields such as the tumor size and the number of lymph nodes examined. Furthermore, many other information extraction tasks and methods have been applied to pathology reports, such as Coden et al which creates a knowledge representation model to represent cancer disease characteristics; Si and Roberts which uses a frame-based representation to extract information from clinical narratives focusing on cancer diagnosis, cancer therapeutic procedure, and tumor description; Xu et al which considers attribute detection as a sequence labeling problem; and Oliwa et al uses NLP to classify gastrointestinal pathology reports into internal and external reports and uses Named Entity Recognition to label accession number, location, date, and sub-labels. Despite these developments, there has been comparatively little effort in understanding 2 additional important criteria that are the basis for reproducibility and real-world use. The first is evaluating performance as a function of training data size, which informs practitioners about how much data they may need to deploy similar systems. Creating an annotated corpus is costly and time-consuming, and accurate assessment of necessary sample size can aid deployment. Second, accurate uncertainty estimates for the predicted results are critical for clinical deployment, as different uses have varying acceptability thresholds. Having accurate uncertainty estimates means that for all cases where the model score outputs a probability p, it is correct p percent of the time. An example of a model with inaccurate uncertainty estimates would be one that gives a predicted probability of correctness of 90% on all examples but is actually only correct 10% of the time. Accurate uncertainty estimates are important for deployment, as lower certainty may be acceptable if the results are used for initial screening with manual verification to follow, but higher certainty is required for a clinical decision support system. Resources can be directed to verification for cases of high uncertainty, supplanting the need for full manual abstraction. The source code for this project will be made available under an open source license to facilitate adoption of NLP tools in cancer pathology.

OBJECTIVE

Our objective was to investigate 2 practical issues that arise when deploying machine learning-based information extraction systems to pathology reports, using prostate cancer as a test case. First, we evaluate the performance of models as a function of dataset size for tasks that involve categorical values, such as histologic grade or presence of lymphovascular invasion, as well as numeric values, such as tumor size. Second, we describe an approach to model calibration and calculation of uncertainty estimates for each prediction and assessing the quality of the model’s uncertainty estimates. We address these gaps in the literature to guide practitioners as they implement these systems in real-world settings.

MATERIALS AND METHODS

Data sources

We used a corpus of 3232 free-text pathology reports for patients that had undergone radical prostatectomy for prostate cancer at the University of California, San Francisco (UCSF) from 2001 to 2018, which were extracted from UCSF’s electronic health record (Epic Systems, Verona, WI, United States). For each document, annotations for 17 pathologic features, such as Gleason scores, margin status, extracapsular extension, and seminal vesicle invasion were extracted (Table 1) in the Urologic Outcomes Database, which is a prospective database that contains clinical and demographic information about patients treated for urologic cancer. Since 2001, data have been manually abstracted by trained abstractors under an institutional review board (IRB) approved protocol. This study was separately approved by the IRB.
Table 1.

Data elements extracted from pathology reports

Data elementsDescription
Document classifier algorithm fields
 Gleason GradePrimary, secondary, tertiaryHistologic grading of tumor aggressiveness based on the Gleason grading system. Each specimen is assigned a primary, secondary, and occasionally a tertiary score, each of which are whole numbers from 1 to 5
 Tumor histologic typePrimary histologic type, such as acinar adenocarcinoma, ductal adenocarcinoma, and small cell neuro-endocrine carcinoma
 Cribriform patternWhether the cells exhibit a cribriform growth pattern (Gleason 4 only)
 Treatment effectIndicator whether there is evidence of a prior treatment, such as hormone treatment or radiation therapy
 Margin status for tumorTo evaluate surgical margins, the entire prostate surface is inked after removal. The surgical margins are designated as “negative” if the tumor is not present at the inked margin and “positive” if tumor is present at the inked margin
 Margin status for benign glandsThe benign margins are designated as “positive” if there are benign prostate glands present at the inked margin and “negative” otherwise
 Perineural invasionWhether cancer cells were seen surrounding or tracking along a nerve fiber within the prostate
 Seminal vesicle invasionInvasion of tumor into the seminal vesicle
 Extraprostatic extensionPresence of tumor beyond the prostatic capsule
 Lymph node statusWhether tumor was identified in lymph nodes
Token extractor algorithm fields
 Pathologic stage classification  T (primary tumor)  N (regional lymph nodes)  M (distant metastasis)Based on American Joint Committee on Cancer TNM staging system for prostate cancer. Based on the edition used in each report (5th–8th edition)
 Tumor volumeAmount of tumor identified in prostate specimen (cubic centimeters)
 Prostate weightOverall weight of the prostate (g)
Data elements extracted from pathology reports The full corpus was divided into 4 parts, 64% training, 16% validation, 10% test, and 10% true test. We looked at the true test only while compiling results. In order to handle our diverse set of fields, we used 2 separate information extraction methods. For categorical fields, we used a document-based classification method which has been previously applied to information extraction from pathology reports., For fields with a large number of possible values (such as numeric quantities), we used a sequence labeling task to extract individual tokens from the document. We applied our methods to the full training dataset as well as randomly selected subsets of 16, 32, 64, 128, and 256 reports. All models are implemented in using scikit-learn and pytorch., Detailed explanation of the pre-processing pipeline and dataset statistics are presented in the Supplementary Material.

Document classifier methods

For categorical data fields, such as the presence of lymphovascular invasion, we treat it as a document classification problem. These fields have between 2 and 6 possible classes (Table 1). We apply classical methods, such as logistic regression, random forests, support-vector machines (SVMs), and adaptive boosting (AdaBoost) on bag-of-n-gram features, as well as deep learning methods, such as convolutional neural networks and long short-term memory networks.

Token extractor methods

Many critical clinical data elements, such as tumor volume, are not suited for classification because they are not categorical in nature. In order to broaden the variety of data fields extracted from the reports, we employ an additional approach which we refer to as token extractor methods. These methods are well-suited to extract numerical quantities from a document (such as the estimated tumor volume or the patient’s medical record number, Table 1). For these fields, we take each token’s surrounding context of k words represented as a bag-of-n-grams as the primary features. We additionally append the token type encoded as a vector to the bag-of-n-grams context vector. The token type vector specifies whether a particular token is an ordinary word, a numeric value, or a hybrid of the 2. These features are used to predict whether or not the token is the token we aim to extract using logistic regression, AdaBoost, or random forest methods. Unlike the document classifier methods, we excluded SVMs and deep learning methods for the token classifier due to the impractical computational requirements for our compute resources. Because our labeled data did not contain the location information of the token of interest within the document, we labeled all tokens that matched our label as a positive example at the time of training. At test time for each token, we compute the score under the model that this token should be extracted and then choose the token with the largest score as our final prediction. This token extraction method is applied to the following fields: pathologic T, N, and M stage, prostate weight, and tumor volume. For additional details regarding the pathologic stage field, we refer the reader to the supplementary material. We would like to give a comparison with a related but slightly different information extraction task of Named Entity Recognition (NER), which classifies named entities in text into categories. Like token extraction, this too is a sequence labeling task. In NER, this involves labeling each token into a predefined category and in our case, for a given field, we label each token with a 0 or 1 as to whether or not it is the desired token for this field and document. As a clarifying example for the distinction between the tasks, an NER system with procedure as a predefined category would label all mentions of procedures in a pathology report as the procedures class. However, this is not what we want, as pathologists will often discuss multiple procedures in a report, but we are interested in only the specific procedure that resected the tumor.

Dataset size and performance

We investigate the performance over varying data-regimes, since for informaticists who wish to build a machine learning parser on their data, a critical question is the quantity of data points needed for adequate performance and which methods are most likely to perform well. We fixed the training set size to 16, 32, 64, 128, and 256 reports, which were randomly drawn from the full training set and averaged the results over 5 random draws.

Evaluation metrics

For each field, we report the weighted F1 score of the classifier, which is the weighted sum of the F1 scores for each class in the field, where the term for each class is weighted by the portion of true instances of the class. In the Supplementary Tables S1–S4, we report the micro F1 and macro F1 to better compare to existing literature. For token extractor models, we compute the accuracy of whether the token extracted from the report was correct.

Hyperparameter tuning

To tune hyperparameters for the classification models on the full data, we used random search with a validation set to tune each method. For each model, we randomly select 20 model-specific hyperparameter configurations, train the model on the training set, and obtain weighted F1 scores on the validation set. The model with the hyperparameter configuration with the highest score is used to obtain results on the test set. To tune hyperparameters for the classification models in the low data regimes (training on ≤256 reports), we used random search across 20 configurations of hyperparameters but with 4-fold cross-validation to calculate weighted F1 scores. For extractor models, we used random search with 20 hyperparameter configurations and 4-fold cross-validation for both the full data and low data regimes. We chose 4-fold cross-validation as it provided a good balance between performance and computational cost in preliminary experiments. For more details regarding hyperparameter ranges for different models, we refer the reader to the Supplementary Materials.

Calibration of systems

To support multiple use cases for the outputs of our model, it is desirable to estimate the model’s uncertainty reflecting the true probability of correctness for each predicted value. For example, values that have a low probability of being correct can be flagged for manual verification, or results can be limited to only those with a high probability of being correct. More rigorously, for a model f and data distribution X ideally we would like a function such that One common definition of the discrepancy between the model’s predicted probability of correctness and its true probability of correctness is given by the expected calibration error (ECE) which is the expected difference between the models confidence and its true probability of being correct.  where is the model’s prediction for a datapoint point x, is the true value, and (x) is the model’s predicted probability of being correct for point x. However, this is typically not able to be measured in practice, for example iftakes on a continuous set of values, so instead is discretized into bins and the ECE is defined as follows: Where is the mth bin, is the average accuracy of the model in bin m, and is the average value of (x) of the model in bin m. To improve the calibration of our system, we apply isotonic regression. In the binary case, it takes the confidence of the models output of the positive class and fits a monotonic function where the x-axis represents the model’s confidence score and the y-axis represents whether or not the model was correct. In the multivariate case, the calibration method attempts to calibrate the probability estimate of each class. It does this by first calibrating the probability of each class in a one-vs-all setting, then after fitting, estimating the probabilities by normalizing the one-vs-all probability for each class.

Error analysis

To understand the potential failure modes of our models, for each field we manually analyzed 10 errors randomly chosen in our test set split of the best models in Table 2 by comparing the model output and annotated label with the text of the report to check the source of the error. If there were fewer than 10 errors for a field, we analyzed all the model’s errors.
Table 2.

Weighted F1 scores for classification fields and mean accuracy for token extractor fields on full training data sample (n = 2066)

Data elementsLogistic regressionAdaBoost classifierRandom forestSVMCNNLSTMMajority class accuracy
Gleason grade—primary0.9780.9710.9410.9320.9810.6280.709
Gleason grade—secondary0.9580.9430.9130.9120.9680.5760.467
Gleason grade—tertiary0.9230.9300.8440.8860.9300.7410.901
Tumor histology0.9890.9950.9950.9930.9950.9940.991
Cribriform pattern0.9630.9810.9630.9680.9870.9660.997
Treatment effect0.9810.9790.9810.9810.9810.9730.985
Tumor margin status0.9410.9530.8880.9180.9500.6300.799
Benign margin status0.9770.9750.9720.9810.9780.9670.997
Perineural invasion0.9440.9780.9380.9290.9720.6130.771
Seminal vesicle invasion0.9430.9740.9400.9650.9760.7840.904
Extraprostatic extension0.9540.9530.8820.9390.9610.7780.712
Lymph node status0.9830.9520.9830.9730.9860.8240.570
Mean weighted F1 across classification models0.9610.9650.9370.9480.9720.7900.817
T stage0.9510.9540.948
N stage0.9540.9540.948
M stage0.9720.9690.969
Estimate tumor volume0.6050.7650.873
Prostate weight0.8460.8550.914
Mean accuracy for token extractor models0.8660.8990.930

CNN, convolutional neural network; LSTM, long short-term memory neural network; SVM, support vector machine.

Weighted F1 scores for classification fields and mean accuracy for token extractor fields on full training data sample (n = 2066) CNN, convolutional neural network; LSTM, long short-term memory neural network; SVM, support vector machine. If the error was a result of an incorrect label in our original data set, it was named as an annotation error. Model errors occurred when the model extracted the incorrect value for a certain field. Next, an error was classified as a report anomaly if there was something wrong with the raw text of the report, such as if the sentences of a report were repeated many times in the text or there was internal inconsistency in the report. Lastly, the evaluation error means that the extracted value was correct but the evaluation method incorrectly penalized the model such as if the correct extracted token was 2 for volume of tumor and the model extracted the token “2-cm” for example.

RESULTS

Document classifier performance

We calculated the weighted F1-score for each data field using the true test set (Table 2). When working with the full training corpus (n = 2066), convolutional networks perform the best (mean weighted F1 0.972 across all 12 clinical data elements). However, we see that the best non-deep-learning method is not far behind with AdaBoost having a weighted F1 score of 0.965.

Token extractor performance

For token extraction, we measure the accuracy of extracting the correct token from each document (Table 2). In greater detail, we choose the most probable token over all tokens in the document and compare this to the ground truth. We observe that random forests perform the best out of all the methods with a mean accuracy of 0.883 across 5 fields.

Performance as a function of dataset size

We see in (Table 3) for the classification fields, the classical machine learning methods (logistic regression, SVM, AdaBoost, and random forests) clearly outperform the deep learning methods on average, likely due to the small amount of training data available. The results also show that 128 reports are needed for the best methods to achieve a 0.90 weighted F1 on average across all classification fields. For the token extractor fields, the results seem to plateau at 64 reports. Our experiments show that a training set size in the thousands is not always needed to adequately extract structured data from these pathology reports, an important observation for practitioners weighing the cost of creating an annotated dataset.
Table 3.

Mean weighted F1 score ± standard deviation for classification models for classification models and mean accuracy ± standard deviation for token extractor models on increasing numbers of reports (n) after 5 trials

Model n = 16 n = 32 n = 64 n = 128 n = 256
Classification models (mean weighted F1 score across all classification fields ± SD)
 Logistic0.781 ± 0.1750.846 ± 0.1170.875 ± 0.0900.911 ± 0.0590.934 ± 0.041
 AdaBoost0.829 ± 0.1400.878 ± 0.1000.907 ± 0.0660.928 ± 0.0490.945 ± 0.034
 Random forest0.795 ± 0.1690.835 ± 0.1280.867 ± 0.1010.882 ± 0.0880.901 ± 0.070
 SVM0.738 ± 0.2140.763 ± 0.2090.786 ± 0.1940.842 ± 0.1120.860 ± 0.140
 CNN0.720 ± 0.2250.790 ± 0.1630.851 ± 0.1220.893 ± 0.0860.935 ± 0.055
 LSTM0.688 ± 0.2050.729 ± 0.1870.743 ± 0.2030.739 ± 0.2140.739 ± 0.212
Token extractor models (mean accuracy across all token extractor fields ± SD)
 Logistic0.844 ± 0.0850.897 ± 0.0790.892 ± 0.0960.902 ± 0.0870.896 ± 0.092
 Adaptive boost0.877 ± 0.0970.892 ± 0.0800.890 ± 0.0840.896 ± 0.0820.890 ± 0.092
 Random forest0.897 ± 0.1800.898 ± 0.0640.915 ± 0.0540.920 ± 0.0410.924 ± 0.038

CNN, convolutional neural network; LSTM, long short-term memory neural network; SVM, support vector machine.

Mean weighted F1 score ± standard deviation for classification models for classification models and mean accuracy ± standard deviation for token extractor models on increasing numbers of reports (n) after 5 trials CNN, convolutional neural network; LSTM, long short-term memory neural network; SVM, support vector machine.

Effect of calibration

We apply calibration to 2 of our models. For the classification model, we apply isotonic calibration to boosting and for the extractor model we apply isotonic regression to the random forest model. For the extractor case, we treat the probability of the token with the highest probability as the confidence score of the model. We fit our isotonic regression calibration methods on the development test set and evaluate the ECE on the test set, binning our uncertainty estimates (x) into bins of width 0.1 (Table 4). Additional experiments investigating the ECE as a function of the bin size, which we include in Supplementary Figures S1 and S2, show that while the average ECE increased, the difference in the average ECE between the smallest bin size (4) and the largest (64) was less than 0.02 for both classification and extraction tasks.
Table 4.

Upper: classifier accuracy and expected calibration error for boosting before and after isotonic calibration and Lower: expected calibration error for random forest model before and after isotonic calibration

Data elementsWeighted-F1ECEIsotonic weighted-F1Isotonic ECE
Classification calibration
 Gleason grade—primary0.950.030.930.03
 Gleason grade—secondary0.940.080.920.14
 Gleason grade—tertiary0.910.050.910.03
 Tumor histology0.990.0090.990.007
 Cribriform pattern0.9950.0070.9950.017
 Treatment effect0.990.0070.990.003
 Tumor margin status0.960.150.940.013
 Benign margin status0.9940.0070.9950.019
 Perineural invasion0.950.260.960.02
 Seminal vesicle invasion0.9870.160.970.02
 Extraprostatic extension0.960.120.960.01
 Lymph node status0.960.040.980.01

Data elementsECEIsotonic ECE

Extractor calibration
 T stage0.1550.016
 N stage0.1440.013
 M stage0.0070.005
 Estimated volume of tumor0.2210.021
 Prostate weight0.2780.033
Upper: classifier accuracy and expected calibration error for boosting before and after isotonic calibration and Lower: expected calibration error for random forest model before and after isotonic calibration We find that for most classifications fields, the model had expected calibration scores less than 0.1 and that isotonic regression generally improves upon this. Since for each class the one-vs-all probabilities are calibrated, the calibrated model’s predictions may differ from the original model if it is not a binary classification problem, so in addition to the ECE of the model, we list the weighted F1 score of the calibrated model. Conversely, extractor models are not well calibrated out of the box in general, but surprisingly, by only using the probability of the token with greatest probability, performing isotonic regression on this single value is enough to achieve sub 0.05 ECEs. We also examined when the model was most overconfident, where we look for examples with high estimated probabilities of being correct, but which were nevertheless wrong. We found the most overconfident example in each field and observed that in 10 of the 15 examples the algorithm was correct and the label was actually incorrect. The most common type of evaluation error for the token extractor occurred when the model extracted the right token, but the evaluation method did not correctly score the model (Supplementary Table S6). For example, if the label for the estimated volume of tumor was 2 (in cm) and the model extracted the token “2-cm”, the model would be penalized. The most common type of report anomaly occurred when the text in the report was repeated. For example, in one case, each sentence in the report was repeated 3 times. This was an issue in the raw text of the report and was not an aberration in preprocessing. Overall, error analysis shows that the scores given for the models are likely underestimates and the models actually perform better than the raw results show. For a comprehensive breakdown of errors, we refer the reader to Supplementary Table S5. Because the pathologic stage errors are highly correlated (due to the fact that the different types of stages are encoded in the same token in the text), only the results for the pathologic T-stage are shown.

DISCUSSION

We have investigated several practical issues in the clinical deployment of a machine learning-based pathology parsing system and developed a system that can accurately parse prostate reports across a variety of fields and provide reliable per-label uncertainty estimates. Furthermore, we evaluated the number of samples required for adequate performance to guide outside practitioners who are considering using a learning-based parsers for their datasets. The dual classification/extraction approach to our pipeline was developed to accommodate a larger variety of data fields. Yala et al applied boosting across twenty binary fields on 17 000 labeled breast cancer reports and observed strong performance with F1 scores above 0.9 for many fields. Gao et al applied hierarchical attention networks to predict tumor site and grade from pathology reports within the NCI-SEER dataset and noted improvement in micro-F1 (up to 0.2 greater) compared to baselines across 2 fields (primary site and histologic grade) for a dataset of lung and breast cancer pathology reports. Much of the previous work does not attempt to extract all relevant data fields since they rely primarily on document classification methods which cannot handle continuous values, such as tumor size or prostate weight or perform the related but slightly different task of NER. Although Li and Martinez (2010) attempt to extract data fields based on numeric values using a hierarchical prediction method, the final prediction step relies on a rule-based method that has no clear way to be calibrated. Furthermore, while our 2 methods are not run on the same fields, our algorithm appears to have higher performance in general. Our solution is developing a sequence tagging algorithm that extracts tokens corresponding to the desired values directly, as well as employing classifier methods to extract categorical data fields. Each method is also capable of outputting a score that can be directly calibrated using isotonic regression. One limitation of our extraction methods is that we only consider simple bag-of-n-grams-based representations and it would be interesting to see how sample efficiency or calibration errors change under a deep learning approach. Second, we investigated the necessary number of reports needed for accurate classification for our pathology reports by varying the size of the training set of reports from 16 to 256 across both classification and extraction. While others have performed sample efficiency analysis of NLP algorithms across many tasks, to our knowledge, this has not been investigated for the important application of clinical information extraction from pathology reports, with the exception of Yala et al. who plot dataset size vs performance over only one method (boosting) and over fields that only take 2 values. Overall, we found that only 128 labeled reports were needed for the best methods for classification and only 64 for the token extractor, a small number compared to the dataset sizes used in prior work. It is important for practitioners who have a smaller dataset to understand approximately how much performance to expect from a machine learning-based approach as it can be expensive and time-consuming to create a large corpus of annotated documents. We hope this encourages more groups to explore these approaches, as large datasets may not always be required. Our study is limited by focusing on a single cancer from one institution, and further work can assess generalizability to other cancers and sites. Of note, there was heterogeneity in report structure and style over 20 years. Another important observation is that the classical statistical learning methods outperformed deep learning methods by a large margin when fewer than 256 data points were available, while deep learning only slightly outperformed logistic regression when using all 2066 reports in the training set. This suggests deep learning only adds marginal value and the complexity of the problem, at least for the reports we worked with, is more suited to classical methods. Finally, we investigated the reliability of uncertainty estimates of the model, which to the authors’ knowledge, has not been investigated in other pathology information extraction work. Knowing which reports are likely to be incorrect can decrease the time needed to manually verify extracted data and filter uncertain predictions for tasks like clinical research with small populations, where each predicted value may have a large impact on conclusions. Through our calibration work, we observed that the classification model was typically well calibrated without any modification, whereas our token extraction algorithm was not. However, by just using the probability of the selected token, isotonic regression was a very effective calibration solution. We furthermore investigated when the model is most likely to be overconfident and found that two-thirds of these errors were due to incorrect annotation labels, not incorrect algorithm outputs.

CONCLUSION

Creating annotated datasets and reliable systems are serious practical concerns when developing and deploying biomedical information extraction systems due to the high cost of creating annotations and the impact of errors on patients outcomes. We show when applying machine learning to pathology parsing, accurate results can be obtained using relatively small annotated datasets and calibration methods can improve the reliability of per-label uncertainty estimates.

FUNDING

Partial support is gratefully acknowledged from ARO (W911NF1710005), NSF (DMS-1613002 and IIS 1741340), the Center for Science of Information (CSoI), a US NSF Science and Technology Center, under grant agreement CCF-0939370, and the Bakar Computational Health Sciences Institute, University of California, San Francisco.

AUTHOR CONTRIBUTIONS

BP and NA implemented, tested, and validated the experiments. All authors were involved in designing and developing the study and writing the paper.

CONFLICT OF INTEREST

There are no competing interests. Click here for additional data file.
  20 in total

1.  Pattern-based information extraction from pathology reports for cancer registration.

Authors:  Giulio Napolitano; Colin Fox; Richard Middleton; David Connolly
Journal:  Cancer Causes Control       Date:  2010-07-23       Impact factor: 2.506

2.  Obtaining Knowledge in Pathology Reports Through a Natural Language Processing Approach With Classification, Named-Entity Recognition, and Relation-Extraction Heuristics.

Authors:  Tomasz Oliwa; Steven B Maron; Leah M Chase; Samantha Lomnicki; Daniel V T Catenacci; Brian Furner; Samuel L Volchenboum
Journal:  JCO Clin Cancer Inform       Date:  2019-08

3.  Building gold standard corpora for medical natural language processing tasks.

Authors:  Louise Deleger; Qi Li; Todd Lingren; Megan Kaiser; Katalin Molnar; Laura Stoutenborough; Michal Kouril; Keith Marsolo; Imre Solti
Journal:  AMIA Annu Symp Proc       Date:  2012-11-03

4.  A Frame-Based NLP System for Cancer-Related Information Extraction.

Authors:  Yuqi Si; Kirk Roberts
Journal:  AMIA Annu Symp Proc       Date:  2018-12-05

5.  Cancer statistics, 2020.

Authors:  Rebecca L Siegel; Kimberly D Miller; Ahmedin Jemal
Journal:  CA Cancer J Clin       Date:  2020-01-08       Impact factor: 508.702

6.  Automated Extraction of Grade, Stage, and Quality Information From Transurethral Resection of Bladder Tumor Pathology Reports Using Natural Language Processing.

Authors:  Alexander P Glaser; Brian J Jordan; Jason Cohen; Anuj Desai; Philip Silberman; Joshua J Meeks
Journal:  JCO Clin Cancer Inform       Date:  2018-12

7.  The CLEF corpus: semantic annotation of clinical text.

Authors:  Angus Roberts; Robert Gaizauskas; Mark Hepple; Neil Davis; George Demetriou; Yikun Guo; Jay Kola; Ian Roberts; Andrea Setzer; Archana Tapuria; Bill Wheeldin
Journal:  AMIA Annu Symp Proc       Date:  2007-10-11

8.  Using machine learning to parse breast pathology reports.

Authors:  Adam Yala; Regina Barzilay; Laura Salama; Molly Griffin; Grace Sollender; Aditya Bardia; Constance Lehman; Julliette M Buckley; Suzanne B Coopey; Fernanda Polubriaginof; Judy E Garber; Barbara L Smith; Michele A Gadd; Michelle C Specht; Thomas M Gudewicz; Anthony J Guidi; Alphonse Taghian; Kevin S Hughes
Journal:  Breast Cancer Res Treat       Date:  2016-11-08       Impact factor: 4.872

9.  Applying a deep learning-based sequence labeling approach to detect attributes of medical concepts in clinical text.

Authors:  Jun Xu; Zhiheng Li; Qiang Wei; Yonghui Wu; Yang Xiang; Hee-Jin Lee; Yaoyun Zhang; Stephen Wu; Hua Xu
Journal:  BMC Med Inform Decis Mak       Date:  2019-12-05       Impact factor: 2.796

10.  Developing a manually annotated clinical document corpus to identify phenotypic information for inflammatory bowel disease.

Authors:  Brett R South; Shuying Shen; Makoto Jones; Jennifer Garvin; Matthew H Samore; Wendy W Chapman; Adi V Gundlapalli
Journal:  BMC Bioinformatics       Date:  2009-09-17       Impact factor: 3.169

View more
  3 in total

1.  Electronic case report forms generation from pathology reports by ARGO, automatic record generator for onco-hematology.

Authors:  Gian Maria Zaccaria; Vito Colella; Simona Colucci; Felice Clemente; Fabio Pavone; Maria Carmela Vegliante; Flavia Esposito; Giuseppina Opinto; Anna Scattone; Giacomo Loseto; Carla Minoia; Bernardo Rossini; Angela Maria Quinto; Vito Angiulli; Luigi Alfredo Grieco; Angelo Fama; Simone Ferrero; Riccardo Moia; Alice Di Rocco; Francesca Maria Quaglia; Valentina Tabanelli; Attilio Guarini; Sabino Ciavarella
Journal:  Sci Rep       Date:  2021-12-10       Impact factor: 4.379

2.  Improving natural language information extraction from cancer pathology reports using transfer learning and zero-shot string similarity.

Authors:  Briton Park; Nicholas Altieri; John DeNero; Anobel Y Odisho; Bin Yu
Journal:  JAMIA Open       Date:  2021-09-30

3.  Automatic Classification of Cancer Pathology Reports: A Systematic Review.

Authors:  Thiago Santos; Amara Tariq; Judy Wawira Gichoya; Hari Trivedi; Imon Banerjee
Journal:  J Pathol Inform       Date:  2022-01-20
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.