Literature DB >> 35721398

Automatic information extraction from childhood cancer pathology reports.

Hong-Jun Yoon1, Alina Peluso1, Eric B Durbin2, Xiao-Cheng Wu3, Antoinette Stroup4, Jennifer Doherty5, Stephen Schwartz6, Charles Wiggins7, Linda Coyle8, Lynne Penberthy9.   

Abstract

Objectives: The International Classification of Childhood Cancer (ICCC) facilitates the effective classification of a heterogeneous group of cancers in the important pediatric population. However, there has been no development of machine learning models for the ICCC classification. We developed deep learning-based information extraction models from cancer pathology reports based on the ICD-O-3 coding standard. In this article, we describe extending the models to perform ICCC classification. Materials and
Methods: We developed 2 models, ICD-O-3 classification and ICCC recoding (Model 1) and direct ICCC classification (Model 2), and 4 scenarios subject to the training sample size. We evaluated these models with a corpus consisting of 29 206 reports with age at diagnosis between 0 and 19 from 6 state cancer registries.
Results: Our findings suggest that the direct ICCC classification (Model 2) is substantially better than reusing the ICD-O-3 classification model (Model 1). Applying the uncertainty quantification mechanism to assess the confidence of the algorithm in assigning a code demonstrated that the model achieved a micro-F1 score of 0.987 while abstaining (not sufficiently confident to assign a code) on only 14.8% of ambiguous pathology reports. Conclusions: Our experimental results suggest that the machine learning-based automatic information extraction from childhood cancer pathology reports in the ICCC is a reliable means of supplementing human annotators at state cancer registries by reading and abstracting the majority of the childhood cancer pathology reports accurately and reliably.
© The Author(s) 2022. Published by Oxford University Press on behalf of the American Medical Informatics Association.

Entities:  

Keywords:  cancer pathology reports; information extraction; machine learning; pediatric cancer

Year:  2022        PMID: 35721398      PMCID: PMC9202570          DOI: 10.1093/jamiaopen/ooac049

Source DB:  PubMed          Journal:  JAMIA Open        ISSN: 2574-2531


INTRODUCTION

Cancer is the leading cause of death by disease in American children ages 0–19 years. Each year, nearly 16 000 children in the United States and over 300 000 children globally are diagnosed with cancer., Analysis of the population-level data for childhood cancers will increase the understanding of the factors that cause cancer and shed light on factors that may help protect against cancer, thus providing evidence to guide public health recommendations and identify and develop improved treatments. Further rapid identification and classification of cases might be used to enhance enrollment for clinical trials for ultra-rare pediatric cancers, thereby enabling access to state-of-the-art care to a wider set of childhood cancer patients. Cancer pathology reports are an excellent resource for such studies. A pathology report is a medical document written by a pathologist that contains the diagnosis determined by examining cells and tissues under a microscope. The reports include information about the topography (site of origin) and the morphology (histology and behavior) of the tumor. Automatic information extraction is the machine learning (ML)-based method that abstracts the findings using the standardized codes. We have been researching automatic information extraction and abstraction from cancer pathology reports based on the International Classification of Diseases for Oncology, third edition (ICD-O-3) coding standard. Recent advances in AI have enabled us to establish robust natural language processing and text comprehension algorithms, which could help mitigate the overhead of manually curating data. We have demonstrated that the deep learning (DL) models exhibited state-of-the-art performance compared against traditional ML-based and rule-based approaches. However, to the best of our knowledge, the efficacy of using the existing automatic information extraction models for childhood cancer pathology reports has not been studied. The difficulties of applying the existing automatic information extraction model to the cancer pathology reports from childhood cancers originated from the differences of prevalent cancer types between adult and pediatric cancer cases. The most prevalent adult cancers are breast, lung and bronchus, prostate, and colorectal cancers. In contrast, the most common childhood cancers are leukemia, lymphoma, and tumors of the central nervous system. Cancers more prevalent in adults are underrepresented among the children, and—likewise—the cancers that are more prevalent in children are underrepresented among the adults. Because the number of adult cancers is substantially higher than the number of childhood cancers, the ML model trained using the entire cancer pathology data corpus is likely biased toward adult cancers. Consequently, this discrepancy may manifest downstream as a classification performance decrease for the childhood cancer model. Also, studies,, suggest that the classification of childhood cancers should be based on morphology rather than topography. The ICD-O-3 is designed to categorize primarily by the site of origin, which is suitable for representing adult cancers. The International Classification of Childhood Cancer (ICCC), developed under the auspices of the International Agency of Research of Cancer (IARC), the International Association of Cancer Registries, and the Société Internationale d’Oncologie Pédiatrique (SIOP), is designed to emphasize the histology of tumors and leverages a combination of site and histology to characterize and classify childhood cancers. The information extraction from childhood cancer pathology reports should emphasize the morphology rather than the primary site of origin for these cancers. The present study aims to develop an optimal ML model for automatic information extraction for pediatric cancer pathology reports based on ICCC coding and to establish a high-precision model by applying the uncertainty quantification (UQ) mechanism, which is critical for state cancer registries. To that end, this article (1) developed a model of automatic information extraction from childhood cancer pathology reports based on ICCC, which—to the best of our knowledge—is the first AI/ML model for pediatric cancers; (2) presents results of a model trained on a large volume (29 206 cases) of pediatric cancer cases from 6 state cancer registries; (3) optimized the model for classifying childhood cancer pathology reports; and (4) describes the model calibration using UQ to support human annotators with high precision.

MATERIALS AND METHODS

Data sources

This study’s data set consisted of unstructured text in pathology reports from 6 state cancer registries: the Kentucky Cancer Registry (KCR), Louisiana Tumor Registry (LTR), New Jersey State Cancer Registry (NJSCR), New Mexico Tumor Registry (NMTR), Seattle Cancer Registry (SCR), and the Utah Cancer Registry (UCR). KCR, LTR, NMTR, SCR, and UCR participate in the National Cancer Institute’s (NCI’s) Surveillance, Epidemiology, and End Results (SEER) program. The study was executed according to the institutional review board protocol DOE000619, approved by the US Department of Energy (DOE) Institutional Review Board on April 6, 2021 (initial approval on September 23, 2016). From the data of millions of e-path reports from the cancer registries, we selected cases in which a cancer patient was diagnosed before they were 20 years old. The gold standard of the abstraction of information extracted from e-path reports is the Cancer/Tumor/Case (CTC) database, which stores all diagnostic, staging, and treatment information for reportable cancers in the SEER Data Management System. Notably, the CTC provides the abstraction of the e-path reports in terms of the ICD-O-3, such as primary cancer site, laterality, histology, and behavior, but does not contain ICCC coding of the cases. Instead, the ICCC codes are being recoded from the ICD-O-3.

ICCC recoding

The NCI SEER provides tables that list ICCC codes and their corresponding ICD-O-3 site and histology codes. Notably, a few slight modifications were made to the coding standards to allow for new and expanded coding of cancers. The latest ICCC recode standard is the “ICCC, Third Edition, based on ICD-O-3/IARC 2017,” which we used in our studies. We chose the ICCC regular recoding as our truth labeling and inference protocol, which consists of 12 main groups and 46 subgroups. Table 1 lists the codes and descriptions for the 12 main groups and 46 subgroups.
Table 1.

ICCC (a) main and (b) subgroup codes and definitions based on the ICCC third edition

(a)
CodeDescription
01Leukemias, myeloproliferative, and myelodysplastic diseases
02Lymphomas and reticuloendothelial neoplasms
03CNS and miscellaneous intracranial and intraspinal neoplasms
04Neuroblastoma and other peripheral nervous cell tumors
05Retinoblastoma
06Renal tumors
07Hepatic tumors
08Malignant bone tumors
09Soft tissue and other extraosseous sarcomas
10Germ cell tumors, trophoblastic tumors, and neoplasms of gonads
11Other malignant epithelial neoplasms and malignant melanomas
12Other and unspecified malignant neoplasms
999Not classified by SEER or in situ
ICCC (a) main and (b) subgroup codes and definitions based on the ICCC third edition

Childhood cancer pathology report data corpus

The total number of childhood cancer pathology reports in our data corpus is 29 206 from 11 274 patients. Figure 1 illustrates the number of cases per each ICCC code. Leukemias (01) and lymphomas (02) represent more than half of all childhood cancers. Leukemia is the most prevalent cancer in children, and this finding is consistent with existing research. Lymphoid leukemia (011) is the most prevalent leukemia and represents more than 25% of childhood cancer cases. Among lymphomas, non-Hodgkin lymphoma (022) is the most prevalent type. Note that the figure illustrates the severe class imbalance within this data set.
Figure 1.

Number of childhood cancer pathology reports by ICCC main and subgroup codes.

Number of childhood cancer pathology reports by ICCC main and subgroup codes. Figure 2 illustrates the number of cases per ICCC code sorted by a patient’s age at diagnosis. Leukemia (01) is more common in younger patients (ages 0–4) but diminishes among older children. Similar patterns occurred for patients with neuroblastoma (04). In contrast, the incidence of lymphomas (02) was highest among adolescents. Germ cell tumors (10) and other malignant epithelial tumors and melanomas (11) are also most common among young adolescents. Note that the number of cancer cases between ages 5 and 11 is considerably lower than for the other age groups. The observations and findings are consistent with the reports and statistics from other studies,, which implies that the data set from the 6 population-based registries in SEER included in this study reflects the real-world situation.
Figure 2.

Number of childhood cancer pathology reports by ICCC main codes and age at diagnosis.

Number of childhood cancer pathology reports by ICCC main codes and age at diagnosis.

ML models for text classification

TextCNN, is one of the most successful and widely used convolutional neural network (CNN) models for text comprehension and classification. It consists of 3 parts: word embedding, 1D convolution, and a fully connected decision layer. Word embedding is a learned representation of terms and words to map a set of words onto vectors of numerical representations with the same semantic meaning and similar observation. The 1D convolution layer has a series of 1D convolution filters that have latent representations to articulate the features in the word vectors of documents. The features found are passed to the fully connected layer to make inferences. MT-CNN extends TextCNN by adding a multitask learning (MTL) mechanism to the decision layer. A classifier learns multiple tasks simultaneously and finds an optimal latent representation to solve a series of related tasks. The MTL helps find more generalized solutions than single-task models, thus yielding higher task performance. We have successfully developed an MT-CNN model for automatic information extraction based on ICD-O-3 and verified that the CNN model has competitive task performance while exhibiting prompt training and inference time.

Automatic information extraction based on ICCC

The following subsections describe 2 models that we designed and tested for this study along with 2 scenarios for each model.

Model 1: ICD-O-3 classification then ICCC recoding approach

The first model involves the classification of ICD-O-3. Generally, the ICCC coding is a recoding based on the site, histology, and behavior from the ICD-O-3 codes. Therefore, the automatic information extraction from childhood cancer pathology reports can be accomplished using the existing classification model for cancer patients of all ages. This approach saves the time and effort required to develop a new model for classifying cancer pathology reports based exclusively on the ICCC coding. However, some factors may cause a decrease in classification accuracy. First, ICCC includes 47 codes, whereas ICD-O-3 consists of more than 300 site codes and 600 histology codes. Designing and training an ML/DL model with that many labels could be overly complex and prone to error. Second, as stated earlier, certain cancer types are more prevalent in adults than in children and vice versa; moreover, cancer is more prevalent in adults, generally. Consequently, the model trained on the entire corpus of cancer patients could be skewed more toward the reports from adult patients. We developed 2 scenarios to evaluate if we can achieve better accuracy by limiting the scope to pathology reports of childhood cancers. Model 1(a): ICD-O-3 classification model is trained by cancer pathology reports from all age groups and then recoded for ICCC Model 1(b): ICD-O-3 classification model is trained by the childhood cancer pathology reports only and then recoded for ICCC

Model 2: Direct ICCC classification approach

The second model involves classifying ICCC codes directly from the cancer pathology reports. The new model was trained on ICCC data and may have higher classification accuracy because it only deals with 47 classes, whereas the ICD-O-3-based models must contend with numerous class labels. This approach required us to train and deploy another ML model specifically for childhood cancer pathology reports, which required extra effort and resources. In addition, we conducted a further study based on the consensus recommendation from the Childhood Cancer Data Initiative’s (CCDI’s) advisory group that the upper age limit of diagnosis be up to 39 years old for certain childhood/pediatric cancers. Table 2 lists the cancer types to be regarded as pediatric cancers at this upper age limit per CCDI’s suggestion. Note that 1055 cases fall into these categories, which is a relatively small number.
Table 2.

List of cancers diagnosed between ages 20 and 39 that could be regarded as pediatric cancers per CCDI and the number of cases that are augmented to the training of Model 2(b)

CodeDescriptionNo. of cases
061Nephroblastoma and other nonepithelial renal tumors26
071Hepatoblastoma and mesenchymal tumors of the liver11
081Osteosarcoma301
082Chondrosarcoma232
083Ewing tumors and related sarcoma of the bone193
084Other specified malignant bone tumors90
085Unspecified malignant bone tumors23
091Rhabdomyosarcomas179
List of cancers diagnosed between ages 20 and 39 that could be regarded as pediatric cancers per CCDI and the number of cases that are augmented to the training of Model 2(b) We developed 2 scenarios to quantify the effect of augmenting the data per CCDI’s recommendation. Model 2(a): ICCC classification model is trained by the childhood cancer pathology reports Model 2(b): ICCC classification model is trained by the childhood cancer pathology reports with augmentation suggested by CCDI Figure 3 illustrates the architectures of the 2 models (and 4 scenarios) that we designed and evaluated.
Figure 3.

Model architecture for ICCC classification from childhood cancer pathology reports. (A) Model 1: ICD-O-3 classification then ICCC recoding. (B) Model 2: direct ICCC classification.

Model architecture for ICCC classification from childhood cancer pathology reports. (A) Model 1: ICD-O-3 classification then ICCC recoding. (B) Model 2: direct ICCC classification.

Uncertainty quantification

The purpose of automatic information extraction is to either assist human observers with a second opinion or automate coding where feasible to enable humans to focus on cases that are more complex or challenging. To this end, the most important feature that the model should possess to achieve the objective is a highly reliable and accurate decision from the model. Inaccurate second opinions (from the model) may distract human observers and even degrade the process’ performance. If the model’s decision is incorrect, then review is needed, which limits the efficiency and benefit of using an automated process. UQ, which is vital to the process, assigns a confidence estimate to the machine-assigned code to allow a human to determine whether additional review is necessary, thus minimizing human labor. In this article, we propose a post-training threshold approach based on the estimation of a confidence score from the softmax-predicted probabilities in the validation set (rather than the training set). Let be the softmax-predicted probabilities for the classification labels. The confidence score is estimated as the conditional distribution of a correct classification via the Bayes theorem for a binary variable (in which being correct and incorrect are mutually exclusive outcomes): The marginal probabilities and are called and are estimated as the corresponding relative frequencies (ie, the total number of correct or incorrect decisions divided by the total number of cases). The conditional probabilities and are estimated from the data by modeling the conditional quantile functions and for a selected percentile point, .

Experimental setup

We designed a comparison study to determine the classification accuracies of the models described above. F1 scores, a widely accepted metric for information retrieval, are used as the performance benchmark. Because of the severe class imbalance of the data corpus, we employed both micro-averaged and macro-averaged F1 scores. The micro-F1 is weighted equally to the individual cases, whereas the macro-F1 is weighted equally to the class label. If the ML model was favorable to the prevalent class labels but did not work well with the samples from minor classes, then the macro-F1 score would be lower than the micro-F1 score. To compensate for the limited availability of childhood cancer pathology reports in the data corpus, we chose to perform 10-fold cross-validation tests. We utilized the StratifiedKFold function available in the scikit-learn library. The TextCNN and MT-CNN model training used the Keras/TensorFlow platform.,

RESULTS

We performed 2 experiments: the first experiment aimed to develop an optimal classification model for information extraction based on ICCC, and the second experiment aimed to establish a reliable model by adopting the UQ mechanism.

Developing an optimal classification model

Table 3 lists the associated task performance for classifying ICCC codes from the childhood cancer pathology reports for both micro-averaged and macro-averaged F1 scores. Model 2 recorded substantially higher scores than Model 1, which implies that the models trained by the ICCC codes performed better than the models for classifying ICD-O-3 codes and recoding them to ICCC. Macroscores (0.701–0.843) showed more improvement than microscores (0.882–0.936), which indicates that Model 2 performed better for the underrepresented class labels. Model 1(b), trained only on the childhood cancer cases, performed slightly better than Model 1(a), which incorporated adult cancer cases. However, the difference was negligible. Performance differences between Models 2(a) and 2(b) were also negligible.
Table 3.

Classification accuracy scores per each ICCC code in F1 metric

Code1(a)1(b)2(a)2(b)# casesUQ# UQ
0110.940.950.970.9780420.997171
0120.880.910.930.9316940.981450
0130.900.890.940.932940.98248
0140.620.570.710.661370.9261
0150.330.490.760.752120.92127
0210.940.940.960.9615300.991380
0220.750.800.870.8819930.961498
0230.840.860.900.908010.98641
0240.960.960.980.973161.00299
0250.070.000.150.17250.2514
0310.890.890.930.933910.99333
0320.880.900.920.9314410.991238
0330.900.900.930.928030.99681
0340.520.580.760.762780.93156
0350.820.850.910.928810.98704
0360.000.240.600.61360.9016
0410.960.970.980.9816391.001558
0420.640.550.730.74261.0012
0500.920.950.990.99711.0068
0610.970.970.980.987361.00694
0620.870.890.940.95990.9987
0710.960.960.980.973340.99315
0720.890.870.920.92870.9873
0810.960.970.980.986171.00581
0820.800.530.810.80400.9335
0830.790.830.870.896270.97433
0840.730.450.760.85300.9820
0850.160.230.380.60250.8210
0910.930.930.960.958400.99764
0920.710.740.830.821590.97102
0930.830.000.290.5060.001
0940.720.730.830.847720.97508
0950.570.580.770.782250.96128
1010.700.720.840.851550.99105
1020.770.830.860.871520.98101
1030.940.940.960.966881.00619
1040.560.650.800.84440.9624
1050.440.360.800.73181.006
1110.640.730.800.76160.778
1120.990.990.990.9911121.001096
1130.920.890.950.93460.9940
1140.950.890.980.974270.99392
1150.590.590.860.77231.009
1160.880.920.960.967151.00639
1210.590.580.880.82750.9944
1220.000.090.580.54201.003
9990.750.720.900.895080.97400
Micro-F10.8820.8960.9350.93629 2060.98724 892
Macro-F10.7010.7190.8370.84329 2060.93524 892

Note: Column 1(a) is the scores from Model 1(a), 1(b) is from Model 1(b), 2(a) is from Model 2(a), and 2(b) is from Model 2(b), “# cases” is the number of classified cases in the data corpus, UQ is the scores from Model 2(b) but without abstained cases based on the softmax UQ, and “# UQ” is the number of classified cases by the UQ model. Micro-averaged and macro-averaged F1 scores are at the bottom of the table.

Classification accuracy scores per each ICCC code in F1 metric Note: Column 1(a) is the scores from Model 1(a), 1(b) is from Model 1(b), 2(a) is from Model 2(a), and 2(b) is from Model 2(b), “# cases” is the number of classified cases in the data corpus, UQ is the scores from Model 2(b) but without abstained cases based on the softmax UQ, and “# UQ” is the number of classified cases by the UQ model. Micro-averaged and macro-averaged F1 scores are at the bottom of the table. The classification accuracy for each ICCC code was analyzed further, and the results are listed in Table 3. Overall, Model 2 recorded higher scores across the ICCC codes. The difference was higher for the underrepresented ICCC codes (eg, 015: Model 1[a] 0.33, Model 2[b] 0.75) than for the more prevalent codes/types (eg, 001: Model 1[a] 0.94, Model 2[b] 0.97). Also, the models performed better for the prevalent ICCC codes (eg, 021 [1530 cases]: Model 2[b] 0.96) than for the minor ones (eg, 025 [25 cases]: Model 2[b] 0.17). However, there were no significant differences between Models 1(a) and 1(b) or between Models 2(a) and 2(b).

Establishing a reliable model

The UQ was applied to Model 2(b), which recorded the highest accuracy score. The abstention classifier was tuned to abstain on cases with confidence scores that were associated with softmax-predicted probabilities lower than 0.9. With the UQ mechanism, Model 2(b) achieved a micro-F1 score of 0.987 and a macro-F1 score of 0.935, and the model discarded only 14.8% of cases from the data corpus. Table 3 lists the accuracy scores for each ICCC code. With the softmax-based UQ, we achieved high accuracy scores across all ICCC labels while maintaining a low abstention rate. Still, there were a few ICCC code outliers in the table for which the model with UQ did not achieve high accuracy scores: unspecified lymphomas (025), unspecified malignant bone tumors (085), Kaposi sarcoma (093), and adrenocortical carcinomas (111).

DISCUSSION

In terms of classification accuracy, the results in Table 3 support the argument that training the models for classifying ICCC codes showed improved accuracy and reliability over the method of using existing ICD-O-3 classification models and then performing ICCC recoding. Presumably, the lower performance on Model 1 is caused by the complexity of the ICD-O-3 classifications, which consist of more than 300 class labels associated with subsites and more than 600 class labels with histology. Moreover, in our data corpus, many of the ICD-O-3 class labels are severely underrepresented. In contrast, the ICCC consists of only 47 class labels. Note that even the ICCC distribution remains imbalanced because of the high prevalence of certain cancer types (leukemias and lymphomas) and the occurrence of “ultra-rare” pediatric tumors. However, the severity of this imbalance is substantially smaller than for the ICD-O-3 system as applied to pediatric cancers. Performance degradation in information extraction models caused by underrepresented class labels in the data corpus is a critical issue in developing algorithms for automation in cancer surveillance. There is no definitive way to increase the sample size of cancer pathology reports from rare cancers given the rarity of pediatric tumors in general (16 000 cases in the United States per year). One might suggest special ML techniques, such as data synthesis, but it is well known that synthesizing free-form text data is not a trivial task. One possible solution is to augment the training corpus with the reports from the subjects of age 20–39, as some ICCC sites are considered pediatric tumors even when occurring in adolescents or young adults. The results listed in Table 3 show the effects of adding those reports based on expert consensus in the CCDI community. The class labels that already had more than 100 samples in the corpus did not benefit from the augmentation because the model had already achieved high classification accuracy scores for those labels. However, substantial improvements were made for code 084—Other specified malignant bone tumors (from 0.76∼F1 to 0.85∼F1) and for code 085—Unspecified malignant bone tumors (from 0.38∼F1 to 0.60∼F1). Those 2 labels had a relatively small number of samples in the corpus: 30 cases for code 084 and 25 cases for code 085. Adding 90 cases for code 084 and 23 cases for code 085 increased the chance of learning features for correct decisions for those class labels. Note, however, that Model 1(b) recorded higher accuracy scores than Model 1(a), which implies that simply adding adult cancer pathology reports did not improve accuracy. This makes sense because the corpus contained many more adult cancers than childhood cancers, and the prevalent adult cancers (eg, breast cancer) are rare in childhood. Thus, this simple addition may not improve the ML model’s performance for childhood cancer cases, which leads to the following question: can we improve the accuracy score if we curate the augmented data set by undersampling the breast and lung cancers or by downselecting the cancer types that are more prevalent in children? This would make an excellent future research topic. This solution of expanding the corpus based on codes, combined with expanding the training corpus by bringing in additional registries, might serve as a partial but nontrivial solution to reduce the class imbalance. The application of the softmax-based UQ mechanism was successful. We demonstrated that Model 2(b) with the softmax-based UQ achieved 0.987—a nearly perfect micro-F1 score—while it abstained in only 14.8% of the cases. This result implies that the system could serve human annotators at state cancer registries. The model can process more than 85.2% of the childhood cancer pathology reports with confidence. Further manual review is needed for only 14.8% of the incoming data, which indicates that the model is highly reliable and potentially ready to use. Several factors contribute to abstention of cases, such as the case difficulty, incompleteness of information on the data samples, or not enough information supplied to the classification model owing to a lack of training samples. One clue is that the abstention rate (ie, the number of abstained cases divided by the total number of cases in the class label) is considerably lower for the prevalent classes. For example, the abstention rate of code 011 was 0.108 (871/8042), whereas the abstention rate for code 014 was 0.555 (76/137). Increasing performance for those underrepresented classes is key to achieving a more reliable model. Some questions remain. The choice of 0.9 as the threshold of probability for correct decision-making is entirely arbitrary and unlikely to be sufficiently robust to meet the high-quality standards of the surveillance community. The threshold value is subject to the tolerance level of wrong decisions by the state cancer registries and will determine the credibility of the data products from the registries. Follow-up statistical analyses of this threshold are required.

CONCLUSION

In this article, we described our study of the classification of childhood cancer pathology reports in terms of the ICCC coding and established an automatic information extraction system for processing a massive volume of pathology reports suitable for state cancer registries. We tested the 2 models: (1) reusing the existing model for extracting ICD-O-3 codes and recoding them into ICCC and (2) developing a new model for extracting ICCC coding. We also experimented with a softmax-based UQ algorithm to evaluate model performance when discarding the minimum amount of ambiguous cases. Our findings suggest that the model for classifying ICCC coding with UQ is suitable for alleviating the workload of human annotators at state cancer registries.

FUNDING

This work was supported in part by the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) program established by DOE and the NCI of the National Institutes of Health. This work was performed under the auspices of DOE by Argonne National Laboratory under Contract DE-AC02-06-CH11357, Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344, Los Alamos National Laboratory under Contract DE-AC5206NA25396, and Oak Ridge National Laboratory under Contract DE-AC05-00OR22725.

AUTHOR CONTRIBUTIONS

H-JY carried out the problem conception, implementation, validation tests, and drafted manuscript. AP carried out the design of uncertainty quantification algorithm. EBD, X-CW, AS, JD, SS, CW, and LC carried out the data curation. EBD, X-CW, and LP helped to draft the manuscript. All authors read and approved the final manuscript.
  11 in total

1.  International Classification of Childhood Cancer, third edition.

Authors:  Eva Steliarova-Foucher; Charles Stiller; Brigitte Lacour; Peter Kaatsch
Journal:  Cancer       Date:  2005-04-01       Impact factor: 6.860

2.  Cancer Statistics, 2021.

Authors:  Rebecca L Siegel; Kimberly D Miller; Hannah E Fuchs; Ahmedin Jemal
Journal:  CA Cancer J Clin       Date:  2021-01-12       Impact factor: 508.702

3.  Childhood and adolescent cancer statistics, 2014.

Authors:  Elizabeth Ward; Carol DeSantis; Anthony Robbins; Betsy Kohler; Ahmedin Jemal
Journal:  CA Cancer J Clin       Date:  2014-01-31       Impact factor: 508.702

4.  Classifying cancer pathology reports with hierarchical self-attention networks.

Authors:  Shang Gao; John X Qiu; Mohammed Alawad; Jacob D Hinkle; Noah Schaefferkoetter; Hong-Jun Yoon; Blair Christian; Paul A Fearn; Lynne Penberthy; Xiao-Cheng Wu; Linda Coyle; Georgia Tourassi; Arvind Ramanathan
Journal:  Artif Intell Med       Date:  2019-10-15       Impact factor: 5.326

Review 5.  The international classification of childhood cancer.

Authors:  E Kramárová; C A Stiller
Journal:  Int J Cancer       Date:  1996-12-11       Impact factor: 7.396

6.  Childhood cancer.

Authors:  R W Miller; J L Young; B Novakovic
Journal:  Cancer       Date:  1995-01-01       Impact factor: 6.860

7.  Limitations of Transformers on Clinical Text Classification.

Authors:  Shang Gao; Mohammed Alawad; M Todd Young; John Gounley; Noah Schaefferkoetter; Hong Jun Yoon; Xiao-Cheng Wu; Eric B Durbin; Jennifer Doherty; Antoinette Stroup; Linda Coyle; Georgia Tourassi
Journal:  IEEE J Biomed Health Inform       Date:  2021-09-03       Impact factor: 7.021

8.  International incidence of childhood cancer, 2001-10: a population-based registry study.

Authors:  Eva Steliarova-Foucher; Murielle Colombet; Lynn A G Ries; Florencia Moreno; Anastasia Dolya; Freddie Bray; Peter Hesseling; Hee Young Shin; Charles A Stiller
Journal:  Lancet Oncol       Date:  2017-04-11       Impact factor: 41.316

9.  Scalable deep text comprehension for Cancer surveillance on high-performance computing.

Authors:  John X Qiu; Hong-Jun Yoon; Kshitij Srivastava; Thomas P Watson; J Blair Christian; Arvind Ramanathan; Xiao C Wu; Paul A Fearn; Georgia D Tourassi
Journal:  BMC Bioinformatics       Date:  2018-12-21       Impact factor: 3.169

10.  Automatic extraction of cancer registry reportable information from free-text pathology reports using multitask convolutional neural networks.

Authors:  Mohammed Alawad; Shang Gao; John X Qiu; Hong Jun Yoon; J Blair Christian; Lynne Penberthy; Brent Mumphrey; Xiao-Cheng Wu; Linda Coyle; Georgia Tourassi
Journal:  J Am Med Inform Assoc       Date:  2020-01-01       Impact factor: 4.497

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.