Literature DB >> 34101764

Development and validation of artificial intelligence to detect and diagnose liver lesions from ultrasound images.

Thodsawit Tiyarattanachai1, Terapap Apiparakoon2, Sanparith Marukatat3, Sasima Sukcharoen4, Nopavut Geratikornsupuk5, Nopporn Anukulkarnkusol6, Parit Mekaroonkamol4, Natthaporn Tanpowpong7, Pamornmas Sarakul8, Rungsun Rerknimitr9, Roongruedee Chaiteerakij9.   

Abstract

Artificial intelligence (AI) using a convolutional neural network (CNN) has demonstrated promising performance in radiological analysis. We aimed to develop and validate a CNN for the detection and diagnosis of focal liver lesions (FLLs) from ultrasonography (USG) still images. The CNN was developed with a supervised training method using 40,397 retrospectively collected images from 3,487 patients, including 20,432 FLLs (hepatocellular carcinomas (HCCs), cysts, hemangiomas, focal fatty sparing, and focal fatty infiltration). AI performance was evaluated using an internal test set of 6,191 images with 845 FLLs, then externally validated using 18,922 images with 1,195 FLLs from two additional hospitals. The internal evaluation yielded an overall detection rate, diagnostic sensitivity and specificity of 87.0% (95%CI: 84.3-89.6), 83.9% (95%CI: 80.3-87.4), and 97.1% (95%CI: 96.5-97.7), respectively. The CNN also performed consistently well on external validation cohorts, with a detection rate, diagnostic sensitivity and specificity of 75.0% (95%CI: 71.7-78.3), 84.9% (95%CI: 81.6-88.2), and 97.1% (95%CI: 96.5-97.6), respectively. For diagnosis of HCC, the CNN yielded sensitivity, specificity, and negative predictive value (NPV) of 73.6% (95%CI: 64.3-82.8), 97.8% (95%CI: 96.7-98.9), and 96.5% (95%CI: 95.0-97.9) on the internal test set; and 81.5% (95%CI: 74.2-88.8), 94.4% (95%CI: 92.8-96.0), and 97.4% (95%CI: 96.2-98.5) on the external validation set, respectively. CNN detected and diagnosed common FLLs in USG images with excellent specificity and NPV for HCC. Further development of an AI system for real-time detection and characterization of FLLs in USG is warranted.

Entities:  

Year:  2021        PMID: 34101764      PMCID: PMC8186767          DOI: 10.1371/journal.pone.0252882

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Hepatocellular carcinoma (HCC) is the fourth leading cause of cancer death worldwide [1]. Screening abdominal ultrasonography (USG) has been shown to be cost-effective in reducing mortality from hepatocellular carcinoma (HCC) by 37% [2-5]. However, worldwide surveillance rates remain low, ranging from 6.7–28.0% [5-9]. One significant barrier to timely HCC screening is inaccessibility to high-quality ultrasound with interpreting radiologists, particularly in rural areas [10]. Developing an artificial intelligence (AI)-assisted USG image analysis system may potentially facilitate USG screening programs, increase the surveillance rate and improve the survival of HCC patients. AI systems have shown potential in facilitating radiologic image interpretation [11]. Abdominal USG is one of the most challenging imaging modalities in the field of AI-based medical image analysis for several reasons. First, the quality of USG images varies among different devices and operators [12]. Second, USG images have a low signal-to-noise ratio making the identification of small lesions from the background difficult. Additionally, a single abdominal USG image often contains several organ structures, often including the liver, gallbladder, kidney, bile duct and blood vessels. The position and orientation of these structures in USG images are not consistent and standardized as with CT or MRI images, therefore, differentiating true lesions from normal structures and pseudo-lesions can be challenging. Although previous studies on AI neural networks demonstrated 88–96% accuracy in the diagnosis of focal liver lesions (FLLs) in still USG images, the size of the training datasets were small with only internal tests being performed [13-15]. Whether these AI systems can be applied in other clinical settings has yet to be investigated. In the current study, we used a large number of off-line USG images to develop an AI-assisted USG image analysis system for detection and diagnosis of various FLLs including HCC, cyst, hemangioma, focal fatty sparing (FFS), and focal fatty infiltration (FFI). To strengthen generalizability of our AI system, we evaluated its performance on images from both an internal test set and external validation datasets (i.e. images from different hospitals using different machines and different sonographers).

Materials and methods

Dataset

This retrospective study was approved by the Institutional Review Board of the Faculty of Medicine, Chulalongkorn University (IRB No. 423/61 and 646/62). Data was collected upon approval from the director and/or ethics committee of King Chulalongkorn Memorial Hospital, Bangkok, Thailand; Mahachai Hospital, Samut Sakhon, Thailand; and Queen Savang Vadhana Memorial Hospital, Chonburi, Thailand. Requirement for informed consent was waived due to the retrospective nature of this study. All ultrasound examinations were de-identified and analyzed anonymously. Images from upper abdominal USG performed between 2010 and 2019 were retrospectively retrieved from the Picture Archiving and Communication System (PACS) of three different hospitals. All data were still images taken as snapshots during ultrasound. They had been stored in Digital Imaging and Communications in Medicine (DICOM) format. All images were acquired using curvilinear transducers and allocated into 3 datasets: training set, internal test set and external validation set. The training set and the internal test set were retrieved from the same patient batch at the main study site, King Chulalongkorn Memorial Hospital, Bangkok, Thailand. All images from this batch were randomly allocated in a 9:1 ratio of the training set to the internal test set. Allocation design ensured that all images from the same patient were assigned to the same set making the image sets independent of each other without any duplicated patients. The external validation set was acquired from Mahachai Hospital, Samut Sakhon, Thailand and Queen Savang Vadhana Memorial Hospital, Chonburi, Thailand to further validate the performance of the AI system. The external validation images were completed by different sonographers using a variety of USG machine models. We included USG studies with all ranges of image qualities from new and older machines to ensure that the AI system can be generalized to other datasets. A total of 17 different ultrasound machine models were included in this study (S1 Table in ). Five of the most commonly encountered liver lesions, including HCCs, cysts, hemangiomas, FFSs and FFIs were selected for this study () [16, 17]. The definitive diagnoses of FLLs were verified using pathology and/or MRI/CT reports. Pathology reports were reviewed first. If not available, MRI and CT reports were then considered. Exclusion criteria were USG studies without further investigation for definitive diagnoses of FLLs and USG studies in which the lesion characteristics were altered by prior treatments. It is noted that in each USG study, there were images with and without FLLs. The normal images without FLLs, which were randomly selected in a 1:1 ratio, were used as negative controls for training the AI system to learn to distinguish FLLs from normal organ structures. An equal number of both types of images facilitated the training process for AI to correctly detect FLLs while minimizing false positivity. In contrast, for the internal test set and external validation set, all negative control images were included in order to replicate the real-world situation in which rare instances of FLLs emerge among a vast number of images showing normal liver and other normal organs. Example images of HCC (1a), cyst (1b), hemangioma (1c), FFS (1d), and FFI (1e) included in the study. Left panels show original images inputted into the AI system. Right panels show AI-outputted bounding boxes around each lesion along with the predicted diagnoses and its confidence value for prediction. Since some patients had more than 1 USG study and some studies had more than one image containing FLLs, the following protocol was used to select and include images in the dataset. For the training set, we included all FLL images of all USG studies of each patient in order to diversify images for the AI training. By contrast, in the internal test set and external validation set, we included up to 2 images with FLLs per study and up to 2 studies per patient. For the USG study having >1 image with FLLs, 2 images containing different FLLs were randomly selected. If there were >1 image containing the same FLL, 2 images taken at different probe positions were randomly selected. If there were >1 image with the same FLL taken at identical probe position, only 1 image was randomly selected.

AI system architecture

The AI framework used in this study was a convolutional neural network (CNN) [18]. CNNs are currently the preferred technique for several types of image analyses due to its structured layering characteristic that can detect complex features of the input images, where the shallow layers detect simple features such as dots and lines and the deeper layers detect more complex features, such as curves and loops [18]. In the present study, we adopted a CNN architecture called RetinaNet [19] which takes an image as input and creates a set of bounding boxes surrounding the FLL along with its class (predicted diagnosis) and its confidence in predicting that particular diagnosis. Confidence value range from 0 to 1, with a value of 1 being the most confident. The confidence threshold can be adjusted according to clinical relevance; for example, the confidence threshold may be lowered to increase the detection rate for HCC if needed in a certain patient population. The overall performance of the CNN, therefore, varies by different confidence thresholds. In this study, the confidence threshold was selected such that the F2 score was optimized on a tuning set, which was a subset of the training set (Details in S3 Appendix and S8 Fig in ). The selected confidence threshold was then used in both the internal test set and external validation cohorts. Since each diagnosis was independent, it was possible for RetinaNet to output multiple diagnoses for a single lesion. This approach resembled the usual practice of reporting differential diagnoses of FLLs by radiologists.

Ultrasound image preprocessing

During image preprocessing, all patient identification information and the peripheral areas in the USG images were cropped out. We identified the coordinates of fan-shaped USG region by ‘Sequence of Ultrasound Regions’ DICOM header, in order to ensure that the cropped image contained only the fan-shaped USG region where annotations and dimension measurements had been cropped out. We also removed markers which were made by sonographers in some images (S1 Appendix in ). The images were then resized to 1333 pixels wide and 800 pixels tall before inputted into the CNN.

AI system development process

Training phase

A supervised training method was implemented to train the AI system. In order to generate an image training dataset, pathology and/or MRI/CT reports were reviewed by experienced sonographers to identify labels, which were the locations and definitive diagnoses of FLLs in each USG image [20]. A hepatologist (R.C.) subsequently verified the labels to ensure their accuracy. Images in the training set were fed into the AI system to train it to predict the location and diagnosis of the FLLs (). RetinaNet codes were adopted from an open-source repository [21, 22]. The codes were then modified and optimized for analyzing USG images. In this work, RetinaNet was composed of backbone ResNet50 and the detection and diagnosis heads. The backbone ResNet50 extracted the hierarchy of features, and the detection and diagnosis heads subsequently processed these features and outputted locations and diagnoses of FLLs [23]. The training was done in two main steps. First, the backbone ResNet50 was trained on a publicly-available image dataset called Microsoft Common Objects in Context (MS-COCO), which comprises 330,000 images of 1.5 million object instances [24]. Subsequently, the whole CNN, both backbone and heads, was fine-tuned on our USG images in the training set. The CNN was trained for 500,000 iterations (25 epochs × 20,000 steps per epoch) on USG images. The initial learning rate was 0.0001. To enable the CNN to recognize diverse configurations of images and to maximize the number of training images, image augmentation was performed by horizontal translation, vertical translation, rotation, scaling, horizontal flipping, motion blur, contrast, brightness, hue and saturation adjustment at each iteration [25]. The training hyperparameters are shown in the S8 Table in . During training, a tuning set was used to monitor performance of the CNN. We selected an epoch that optimized mean average precision [26] on the tuning set for final evaluation on the internal test set and the external validation set.

Evaluation phase

The performance of the developed AI system was evaluated first on the internal test set, then on the external validation set.

Performance evaluation metrics

We separately evaluated detection and diagnosis, the two primary tasks of the CNN. Evaluation of detection rates and diagnosis performance were performed on a per-lesion basis. The definitions of the evaluation metrics are described below.

Detection task

An FLL was counted as correctly detected if the CNN generated a bounding box around it and the box overlapped with the true location of FLL, which was assessed using Intersection-over-Union (IoU). In this study, an IoU of greater than 0.2 was a cut-off for a correct detection by the CNN (S2 Appendix and S1 Fig in ). We opted to use this cutoff because FLLs in USG images often have indistinct boundaries, especially for FFSs and FFIs (S2 Fig in ). The detection rate was calculated by dividing the number of FLLs correctly detected by the number of total FLLs. Detection rates stratified by ground truth diagnoses were also evaluated. In contrast, a false positive detection was counted when the AI system outputted a bounding box on an area that did not contain FLLs (e.g. liver parenchyma, normal organ structures, etc.). Evaluation of false positive detections was performed on a per-image basis.

Diagnosis task

We used the following metrics to evaluate AI diagnostic performance: where TP, TN, FP and FN are the number of true positive, true negative, false positive and false negative diagnoses, respectively. We used a “one-versus-all” method to evaluate diagnostic performance for each FLL diagnosis [27]. For example, when evaluating diagnostic performance for HCC, other diagnoses were counted as a single non-HCC class: where sensitivity is the diagnostic sensitivity for HCC. TP is the number of true positive diagnoses for HCC, where the definitive diagnosis is HCC and the AI system correctly diagnosed the lesion as HCC. FN is the number of false negative diagnoses for HCC, where the definitive diagnosis is HCC, but the AI system falsely diagnosed the lesion as either cyst, hemangioma, FFS or FFI. In cases where multiple diagnoses reached the confidence threshold and hence were predicted by the AI system, only the diagnosis with the highest confidence value was selected as the AI prediction.

Calculation of overall detection rate and overall diagnostic performance

After calculating detection and diagnostic performance metrics for each definitive diagnosis of FLLs, we pooled the performance results from the 5 FLL diagnoses into an overall performance result. Because the numbers of each FLL diagnosis in our dataset were imbalanced, overall performance, including overall detection rate, overall diagnostic sensitivity and specificity, were pooled by an unweighted average, to minimize the effect of imbalanced number of FLL diagnoses. For example, where DR is the overall detection rate. DR, DR, DR, DR and DR are the detection rates for HCC, cyst, hemangioma, FFS and FFI, respectively. where sens is the overall diagnostic sensitivity. sens, sens, sens, sens and sens are the diagnostic sensitivities for HCC, cyst, hemangioma, FFS and FFI, respectively.

Statistical analysis

Performance of the CNN was reported by detection rates, false positive detection rates, diagnostic sensitivities, specificities, accuracies, positive predictive values (PPVs), and negative predictive values (NPVs) with 95% confidence intervals (95% CI). Detection and diagnostic performance of each type of FLL as well as overall performance for all FLL diagnoses were calculated. Performance on the internal test set and external validation set were compared using two-tailed z-test for difference of proportion. Python version 3.7 (Python Software Foundation, Delaware, USA) and IBM SPSS Statistics for Windows, version 22 (SPSS Inc., Chicago, Ill., USA) were used for data analyses. A p-value of <0.05 was considered statistically significant.

Results

Baseline characteristics

A total of 40,397 images with 20,432 FLLs were included in the training set, while 6,191 images with 845 FLLs and 18,922 images with 1,195 FLLs were included in the internal test set and external validation set, respectively. Baseline characteristics of each dataset is described in . aKing Chulalongkorn Memorial Hospital, Bangkok, Thailand bMahachai Hospital, Samut Sakhon, Thailand cQueen Savang Vadhana Memorial Hospital, Chonburi, Thailand

Performance of the CNN

Performance of CNN in detection and diagnosis on the internal test set and external validation set are summarized in . aKCMH, King Chulalongkorn Memorial Hospital, Bangkok, Thailand bMahachai Hospital, Samut Sakhon, Thailand cQueen Savang Vadhana Memorial Hospital, Chonburi, Thailand *P-value for two-tailed z-test for difference of proportion, comparing performance results in the internal test set and pooled external validation set. P-value of <0.05 was considered statistically significant. †Clopper-Pearson confidence interval was calculated for performance value at boundaries (i.e. 0% and 100%) Detection rates, diagnostic sensitivities and specificities are shown in percentages. 95% confidence intervals are shown in parenthesis.

Lesion detection performance

On the internal test set, the CNN had an overall lesion detection rate of 87.0% (95%CI: 84.3–89.6). The median IoU was 0.788 (range: 0.202–0.978) (S3 Fig in ), suggesting an exceptional agreement between the predicted and true location of the FLL. Compared to the internal test set, the overall detection rate in the pooled external validation set was significantly lower (75.0% (95%CI: 71.7–78.3), p < 0.001), with the median IoU of 0.781 (range: 0.201–0.970) (S3 Fig in ). The false positive detection rate was 3.7% (226/6191) and 5.1% (970/18922) in the internal test set and external validation set, respectively. The images with false positive detections were reviewed. Blood vessel in the liver was the most common falsely identified structure as FLLs (12.3%, 147/1196), followed by heterogeneous background liver parenchyma (7.4%, 88/1196), renal cysts (6.8%, 81/1196), inferior vena cava (3.4%, 41/1196) and splenic lesions (2.8%, 33/1196) (S2 Table and S4 Fig in ). Likewise, 114 and 273 images with false negative detection in the internal test set and the external validation set were reviewed. Characteristics for incorrect detection included being a small lesion <1 cm (27.4%), having an uncommon location of that particular diagnosis (8.0%), lesion with atypical characteristics (7.8%), ill-defined lesion (7.5%), and lesion obscured by shadow artifacts or not completely visible (6.2%) (S3 Table and S5 Fig in ).

Diagnostic performance

After detection of a lesion, the AI algorithm identified and diagnosed the lesion as one of five diagnoses (HCC, cyst, hemangioma, FFS, FFI) of FLLs. On the internal test set, the CNN had overall sensitivity, specificity, accuracy, PPV and NPV of 83.9% (95%CI: 80.3–87.4), 97.1% (95%CI: 96.5–97.7), 95.4% (95%CI: 94.8–96.1), 83.6% (95%CI: 80.1–87.1), and 97.2% (95%CI: 96.6–97.8), respectively, for classifying any FLLs. For the diagnosis of HCC, CNN had a sensitivity of 73.6% (95%CI: 64.3–82.8), specificity of 97.8% (95%CI: 96.7–98.9), accuracy of 94.9% (95%CI: 93.3–96.5), PPV of 82.1% (95%CI: 73.5–90.6), and NPV of 96.5% (95%CI: 95.0–97.9). The sensitivity and specificity for diagnosing other FLLs ranged from 69.0% to 98.0% and 95.0% to 98.3%, respectively (). The overall performance of the CNN in diagnosing any FLLs in the external validation set was similar to that of the internal test set, with the sensitivity, specificity, accuracy, PPV and NPV of 84.9% (95%CI: 81.6–88.2), 97.1% (95%CI: 96.5–97.6), 95.3% (95%CI: 94.7–95.9), 81.9% (95%CI: 78.4–85.4), and 97.1% (95%CI: 96.6–97.7), respectively. In subgroup analyses of each type of FLL, the diagnostic performance in the external validation set was also comparable to the performance of the internal test set as displayed in . Confusion matrix for classification results in the internal test set and external validation set is shown in . After reviewing misclassified images, we found that the most common cause was atypical characteristics of FLLs (30.1%, 56/186) (S4 Table and S6 Fig in ).

Subgroup analyses

The AI system detection and diagnostic performance was further stratified by FLL sizes (S5 Table in ) and background liver parenchyma (cirrhosis vs. non-cirrhosis) (S6 Table in ). As expected, diagnostic sensitivities for HCC increased by size. Sensitivities of HCC sizes of < 2 cm, 2–3 cm, and > 3 cm were 23.5% (95%CI: 3.4–43.7), 77.3% (95%CI: 59.8–94.8) and 89.6% (95%CI: 80.9–98.2) in the internal test set and 50.0% (95%CI: 30.0–70.0), 84.2% (95%CI: 67.8–100) and 92.3% (95%CI: 85.8–98.8) in the external validation set, respectively. Additionally, detection rates of HCC in cirrhosis subgroup were lower than in non-cirrhosis subgroup, i.e. 79.5% (95%CI: 67.6–91.5) vs 89.7% (95%CI: 81.8–97.5) in the internal test set and 72.0% (95%CI: 63.2–80.8) vs 94.7% (95%CI: 87.6–100) in the external validation set, respectively.

Discussion

The CNN developed in our study using an advanced structured AI learning system demonstrated a consistently high diagnostic performance on USG images from both an internal test set and an external validation set. It achieved overall diagnostic sensitivity and specificity of 83.9% and 97.1% on the internal test set and 84.9% and 97.1% on the external validation set. Regarding detection task, our AI system was able to detect 85.3% of HCCs in the internal test set and 78.3% in the external validation set (p = 0.16). However, averaging all included FLL diagnoses, the detection rate of the external validation set was significantly lower than the internal test set (75.0% vs 87.0%; p <0.001). Factors that may be at play include the increased heterogeneity of image characteristics from different ultrasound machine models in the external validation set, compared to the training set (S1 Table in ). This finding underscores the importance of image diversity in the training dataset. To enhance practicality, we propose to train the AI system with additional USG videos which contain numerous image frames to better detect FLLs. For the diagnosis task, the performance results were consistent between the internal test set and the external validation set. The AI system achieved overall sensitivities of 83.9% and 84.9%, and specificities of 97.1% and 97.1% on the internal test set and external validation set, respectively. Our AI system had lower sensitivity for FLL diagnosis than the sensitivities of 93.8%-98.8% shown in previous studies, with comparable specificities of 94.3–98.9% in the previous reports [13-15]. The lower sensitivity may have been due to the wider spectrum of FLL diagnoses and characteristics. In the two previous studies, only HCCs, cysts and hemangiomas were selected for testing [14, 15]. In the current study, FFSs and FFIs were additionally included as both diagnoses are encountered frequently in liver cancer surveillance settings with prevalence rates of FFS and FFI previously reported as 6.3% and 9.2%, respectively [17, 28]. Misclassifications of FLLs by the AI system may be explained by the fact that different types of FLLs can appear very similar on USG images. Moreover, some lesions may have atypical characteristics. We found that HCCs and hemangiomas were sometimes interchangeably misdiagnosed (). This may be because our sample contained a considerable number of hemangiomas with atypical characteristics (18.8% of all hemangiomas) with 11.8% of hemangiomas appearing as hypoechoic lesions in fatty liver background and 7% of hemangiomas as giant hemangioma with heterogeneous echogenicity in contrast to typical well-defined round hyperechoic lesion (S6 Fig in ). This is supported by our findings that diagnostic sensitivity of HCC increased when the size of lesion increased, while diagnostic sensitivities of hemangioma decreased when the size of lesion increased. We specifically had designed our AI system to output diagnoses of detected FLLs as differential diagnoses. This should be clinically useful as physicians will be able to decide what is the most likely diagnosis of FLL by incorporating the AI diagnosis together with their clinical information. We further analyzed whether HCC appeared in the top-k predicted differential diagnoses. Top-1 (equal to diagnostic sensitivity reported in the main results section), top-2 and top-3 sensitivities for diagnosing HCC were 73.6%, 90.8% and 96.6%, respectively in the internal test set and 81.5%, 89.0% and 93.6%, respectively in the external validation set (S7 Table in ). This provides evidence that the AI system can characterize HCC with low miss rate. The unique approach of our study is the development and testing of an AI system that can both detect and diagnose FLLs from USG still images. This novel AI system could automatically detect and classify FLLs without the need for human help for guiding the location of FLLs. Images with all ranges of qualities were included that help strengthen our findings on the practicality of using this AI method. We found that the CNN was able to handle such variation reasonably well. We believe that with more data, the performance of the AI system could be further improved. The AI development flow can be divided into the following stages: 1) pre-clinical stage using single-site retrospective dataset, 2) validation on external cohorts, and 3) evaluating usefulness of AI systems in real clinical settings by prospective or randomized-controlled trial study designs [29]. In this study, we validated the performance on external validation cohorts (i.e. 2nd stage of AI development flow) with satisfactory results. Currently, our AI system works off-line on still USG images. Since the ultimate goal is to implement an AI system in clinical practice, we are now incorporating USG videos as training materials to leverage our AI system to perform real-time analysis while a USG procedure is being performed.

Conclusion

Given the structured training framework, the CNN has shown good performance for the detection and diagnosis of FLLs in USG images. HCCs can be detected and diagnosed with satisfactory performance. To fulfill our goal of assisting in the detection and diagnosis of FLLs during USG performed by non-radiologists, an AI system for real-time detection and analysis is warranted. (PDF) Click here for additional data file.

Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present. 29 Mar 2021 PONE-D-20-36551 Development and validation of artificial intelligence to detect and diagnose liver lesions from ultrasound images PLOS ONE Dear Dr. Chaiteerakij, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by May 13 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Khanh N.Q. Le Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: No ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: In this manuscript, the authors developed a convolutional neural network model for detection and classification of focal liver lesions in diagnostic ultrasound images. The current study had a larger training dataset and external testing data compared to existing literature on this topic. There are a lot of important information and data in the supporting document. Overall, the study is interesting and thorough; however, there are a number of issues that should be addressed. 1. Ultrasound image preprocessing, CNN model architecture and training process should be described in the Methods section. 2. Were all ultrasound images acquired with a curvilinear transducer? Were they all still images or from a video clip? Were the images resized to certain matrix size? 3. Besides markers on the images, were there any texts such as dimension measurement and annotations? 4. If some failure was due to dark images, could image intensity scaling be added into the image preprocessing step? 5. Consistent and standardized naming would be recommended. There was internal test which was sometimes called internal validation. It could be confusing with the validation in training process. 6. Could the training and internal test datasets from the same patients and same lesion? They should be different. 7. It’s not clear if the ground truth diagnosis labelling was based on MR or CT, or only pathology reports. 8. IoU cutoff of 0.2 appears to be loose. Was there any more justification on this cutoff value? Could IoU be optimized in the model? 9. How was the diagnostic accuracy calculated? Was it area under the ROC curve or something else? Without clarification, the first paragraph in the Discussion could be misleading. Reviewer #2: In this work, the authors proposed a CNN for focal liver lesion detection from ultrasonography still images. Internal and external validation datasets were used to validate the proposed method. The whole study is complete. However, a lot of details were missing, which need further modifications. Major comments: 1. The authors mentioned ‘detection’ and ‘diagnosis’ performance in the paper. However, it is hard to understand what ‘diagnosis’ metrics mean. ‘The AI system was… by one-versus-all method’. This paragraph is very difficult to understand. The authors need to rewrite the diagnosis task part. 2. The training phase section is not detailed enough. What’s the training epoch number for the model? Did the authors use validation datasets to choose the training epoch? Minor comments: 1. Grammar errors exist throughout the paper, which need the authors to further address, e.g. ‘unweighted average method’. 2. The reference number should be put ahead of the period sign, not after. 3. Table 3. It is better to present it as a heat map, instead of a single table, to better visualize the results. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 16 Apr 2021 Reviewer #1: 1. Ultrasound image preprocessing, CNN model architecture and training process should be described in the Methods section. Response: We have added the CNN model architecture, ultrasound image preprocessing and training process in the Methods section on page 11-13 as follows: AI system architecture The AI framework used in this study was a convolutional neural network (CNN). CNNs are currently the preferred technique for several types of image analyses due to its structured layering characteristic that can detect complex features of the input images, where the shallow layers detect simple features such as dots and lines and the deeper layers detect more complex features, such as curves and loops. In the present study, we adopted a CNN architecture called RetinaNet which takes an image as input and creates a set of bounding boxes surrounding the FLL along with its class (predicted diagnosis) and its confidence in predicting that particular diagnosis. Confidence value range from 0 to 1, with a value of 1 being the most confident. The confidence threshold can be adjusted according to clinical relevance; for example, the confidence threshold may be lowered to increase the detection rate for HCC if needed in a certain patient population. The overall performance of the CNN, therefore, varies by different confidence thresholds. In this study, the confidence threshold was selected such that the F2 score was optimized on a tuning set, which was a subset of the training set (Details in S3 Appendix and S8 Fig). The selected confidence threshold was then used in both the internal test set and external validation cohorts. Since each diagnosis was independent, it was possible for RetinaNet to output multiple diagnoses for a single lesion. This approach resembled the usual practice of reporting differential diagnoses of FLLs by radiologists. Ultrasound image preprocessing During image preprocessing, all patient identification information and the peripheral areas in the USG images were cropped out. We identified the coordinates of fan-shaped USG region by ‘Sequence of Ultrasound Regions’ DICOM header, in order to ensure that the cropped image contained only the fan-shaped USG region where annotations and dimension measurements had been cropped out. We also removed markers which were made by sonographers in some images (S1 Appendix). The images were then resized to 1333 pixels wide and 800 pixels tall before inputted into the CNN. AI system development process Training phase A supervised training method was implemented to train the AI system. In order to generate an image training dataset, pathology and/or MRI/CT reports were reviewed by experienced sonographers to identify labels, which were the locations and definitive diagnoses of FLLs in each USG image. A hepatologist (R.C.) subsequently verified the labels to ensure their accuracy. Images in the training set were fed into the AI system to train it to predict the location and diagnosis of the FLLs (Fig 2). RetinaNet codes were adopted from an open-source repository. The codes were then modified and optimized for analyzing USG images. In this work, RetinaNet was composed of backbone ResNet50 and the detection and diagnosis heads. The backbone ResNet 50 extracted the hierarchy of features, and the detection and diagnosis heads subsequently processed these features and outputted locations and diagnoses of FLLs. The training was done in two main steps. First, the backbone ResNet50 was trained on a publicly-available image dataset called Microsoft Common Objects in Context (MS-COCO), which comprises 330,000 images of 1.5 million object instances. Subsequently, the whole CNN, both backbone and heads, was fine-tuned on our USG images in the training set. The CNN was trained for 500,000 iterations (25 epochs × 20,000 steps per epoch) on USG images. The initial learning rate was 0.0001. To enable the CNN to recognize diverse configurations of images and to maximize the number of training images, image augmentation was performed by horizontal translation, vertical translation, rotation, scaling, horizontal flipping, motion blur, contrast, brightness, hue and saturation adjustment at each iteration. The training hyperparameters are shown in the S8 Table. During training, a tuning set was used to monitor performance of the CNN. We selected an epoch that optimized mean average precision on the tuning set for final evaluation on the internal test set and the external validation set. 2. Were all ultrasound images acquired with a curvilinear transducer? Were they all still images or from a video clip? Were the images resized to certain matrix size? Response: All ultrasound images were acquired with a curvilinear transducer. They were all still images. The images were resized to 1333 pixels wide and 800 pixels tall before inputted into the CNN. We have added sentences in the Methods section as follows: In the Dataset subsection (page 9): “All data were still images taken as snapshots during ultrasound. They had been stored in Digital Imaging and Communications in Medicine (DICOM) format. All images were acquired using curvilinear transducers and allocated into 3 datasets: training set, internal test set and external validation set.” In the Ultrasound image preprocessing subsection (page 12): “The images were then resized to 1333 pixels wide and 800 pixels tall before inputted into the CNN.” 3. Besides markers on the images, were there any texts such as dimension measurement and annotations? Response: There were dimension measurement and annotation texts in the periphery areas of the ultrasound images. These texts were cropped out during the image preprocessing step. We have added this explanation in the Methods section under Ultrasound image preprocessing subsection on page 12 as follows: “We identified the coordinates of fan-shaped USG region by ‘Sequence of Ultrasound Regions’ DICOM header, in order to ensure that the cropped image contained only the fan-shaped USG region where annotations and dimension measurements had been cropped out.” 4. If some failure was due to dark images, could image intensity scaling be added into the image preprocessing step? Response: We thank the reviewer for this insightful question. We have acknowledged this issue during our experiments. We tried using histogram equalization technique to scale image intensity. Unfortunately, the image quality was minimally improved by this technique. We found that brightness and contrast augmentation during the training step was more helpful in enabling the CNN to handle images with variation in light/dark conditions. Hyperparameters for brightness and contrast augmentation are shown in Table S8. 5. Consistent and standardized naming would be recommended. There was internal test which was sometimes called internal validation. It could be confusing with the validation in training process. Response: We apologize for the inconsistency. We have changed the word “internal validation set” to “internal test set” throughout the manuscript. 6. Could the training and internal test datasets from the same patients and same lesion? They should be different. Response: We apologize for the unclear explanation. We divided the dataset into the training and the internal test set based on patient level, i.e., images from the same patient were allocated to the same set. We have add sentences in the Method section, under the Dataset subsection on page 9 as follows: “Allocation design ensured that all images from the same patient were assigned to the same set making the image sets independent of each other without any duplicated patients.” 7. It’s not clear if the ground truth diagnosis labelling was based on MR or CT, or only pathology reports. Response: We again apologize for the unclear writing in the original manuscript. The pathology reports were considered first for ground truth diagnosis. If the pathology report was not available, MRI and CT reports were then considered. We have added this explanation to the Method section, under Dataset subsection on page 10 as follows: “The definitive diagnoses of FLLs were verified using pathology and/or MRI/CT reports. Pathology reports were reviewed first. If not available, MRI and CT reports were then considered.” 8. IoU cutoff of 0.2 appears to be loose. Was there any more justification on this cutoff value? Could IoU be optimized in the model? Response: In this study, we opted to use the IoU cutoff of 0.2 because FLLs in USG images often have indistinct boundary, especially for FFSs and FFIs. Nonetheless, our model achieved overall high IoU. In the internal test set, the median IoU was 0.788 (range: 0.202 – 0.978). In the external validation set, the median IoU was 0.781 (range: 0.201 – 0.970). Distribution of IoU values is summarized in Fig S3 and examples of images with different IoU values are illustrated in Fig S2. We have added sentences to explain the reason for choosing the IoU cutoff of 0.2 in the Method section under the Performance evaluation metrics subsection on page 14 as follows: “ In this study, an IoU of greater than 0.2 was a cut-off for a correct detection by the CNN (S2 Appendix and S1 Fig). We opted to use this cutoff because FLLs in USG images often have indistinct boundaries, especially for FFSs and FFIs (S2 Fig).” 9. How was the diagnostic accuracy calculated? Was it area under the ROC curve or something else? Without clarification, the first paragraph in the Discussion could be misleading. Response: We apologize for the unclear explanation. We calculated the accuracy by the following formula: (TP+TN)/(TP+TN+FP+FN) where TP, TN, FP and FN are the number of true positive, true negative, false positive and false negative classifications, respectively. We have provided the formula used for calculating sensitivity, specificity, accuracy, positive predictive value and negative predictive value in the Methods section under Performance evaluation metrics subsection on page 14. Additionally, to avoid ambiguity to the readers and to be consistent with the main results shown in Table 2, we have revised the first paragraph of the Discussion section on page 23 as follows: “ The CNN developed in our study using an advanced structured AI learning system demonstrated a consistently high diagnostic performance on USG images from both an internal test set and an external validation set. It achieved overall diagnostic sensitivity and specificity of 83.9% and 97.1% on the internal test set and 84.9% and 97.1% on the external validation set.” Reviewer #2: Major comments: 1. The authors mentioned ‘detection’ and ‘diagnosis’ performance in the paper. However, it is hard to understand what ‘diagnosis’ metrics mean. ‘The AI system was… by one-versus-all method’. This paragraph is very difficult to understand. The authors need to rewrite the diagnosis task part. Response: We apologize for the unclear writing. We have rewritten the Diagnosis task subsection to describe the diagnosis metrics and explain the “one-versus-all” method on page xx as follows: We used the following metrics to evaluate AI diagnostic performance: sensitivity=TP/(TP+FN) specificity=TN/(TN+FP) accuracy=(TP+TN)/(TP+TN+FP+FN) positive predictive value=TP/(TP+FP) negative predictive value=TN/(TN+FN) where TP, TN, FP and FN are the number of true positive, true negative, false positive and false negative diagnoses, respectively. We used a “one-versus-all” method to evaluate diagnostic performance for each FLL diagnosis. For example, when evaluating diagnostic performance for HCC, other diagnoses were counted as a single non-HCC class: 〖sensitivity〗_HCC=〖TP〗_HCC/(〖TP〗_HCC+〖FN〗_HCC ) where 〖sensitivity〗_HCC is the diagnostic sensitivity for HCC. 〖TP〗_HCC is the number of true positive diagnoses for HCC, where the definitive diagnosis is HCC and the AI system correctly diagnosed the lesion as HCC. 〖FN〗_HCC is the number of false negative diagnoses for HCC, where the definitive diagnosis is HCC, but the AI system falsely diagnosed the lesion as either cyst, hemangioma, FFS or FFI. 2. The training phase section is not detailed enough. What’s the training epoch number for the model? Did the authors use validation datasets to choose the training epoch? Response: We apologize for not providing the details of the training phase in the main manuscript. The training epoch number of the model was 25 (20,000 steps per epoch). We used a validation set to choose the training epoch. In our manuscript, the validation set was called ‘tuning set’ to avoid confusion with the external validation set. We have added details regarding the training of the AI system in the Method section, under the Training phase subsection on page 12-13 as follows: RetinaNet codes were adopted from an open-source repository. The codes were then modified and optimized for analyzing USG images. In this work, RetinaNet was composed of backbone ResNet50 and the detection and diagnosis heads. The backbone ResNet 50 extracted the hierarchy of features, and the detection and diagnosis heads subsequently processed these features and outputted locations and diagnoses of FLLs. The training was done in two main steps. First, the backbone ResNet50 was trained on a publicly-available image dataset called Microsoft Common Objects in Context (MS-COCO), which comprises 330,000 images of 1.5 million object instances. Subsequently, the whole CNN, both backbone and heads, was fine-tuned on our USG images in the training set. The CNN was trained for 500,000 iterations (25 epochs × 20,000 steps per epoch) on USG images. The initial learning rate was 0.0001. To enable the CNN to recognize diverse configurations of images and to maximize the number of training images, image augmentation was performed by horizontal translation, vertical translation, rotation, scaling, horizontal flipping, motion blur, contrast, brightness, hue and saturation adjustment at each iteration. The training hyperparameters are shown in the S8 Table. During training, a tuning set was used to monitor performance of the CNN. We selected an epoch that optimized mean average precision on the tuning set for final evaluation on the internal test set and the external validation set. Minor comments: 1. Grammar errors exist throughout the paper, which need the authors to further address, e.g. ‘unweighted average method’. Response: We apologize for the grammar errors. We have sent the revised manuscript to our institution’s English editing service for correcting grammar errors. Additionally, we have added a subsection called ‘Calculation of overall detection rate and overall diagnostic performance’ in the Method section under Performance evaluation metrics subsection on page 15. This added subsection describes how we calculated overall performance, as follows: Calculation of overall detection rate and overall diagnostic performance After calculating detection and diagnostic performance metrics for each definitive diagnosis of FLLs, we pooled the performance results from the 5 FLL diagnoses into an overall performance result. Because the numbers of each FLL diagnosis in our dataset were imbalanced, overall performance, including overall detection rate, overall diagnostic sensitivity and specificity, were pooled by an unweighted average, to minimize the effect of imbalanced number of FLL diagnoses. For example, 〖DR〗_overall=(〖DR〗_HCC+〖DR〗_cyst+〖DR〗_hemangioma+〖DR〗_FFS+〖DR〗_FFI)/5 where 〖DR〗_overall is the overall detection rate. 〖DR〗_HCC, 〖DR〗_cyst, 〖DR〗_hemangioma, 〖DR〗_FFS and 〖DR〗_FFI are the detection rates for HCC, cyst, hemangioma, FFS and FFI, respectively. 〖sens〗_overall=(〖sens〗_HCC+〖sens〗_cyst+〖sens〗_hemangioma+〖sens〗_FFS+〖sens〗_FFI)/5 where 〖sens〗_overall is the overall diagnostic sensitivity. 〖sens〗_HCC, 〖sens〗_cyst, 〖sens〗_hemangioma, 〖sens〗_FFS and 〖sens〗_FFI are the diagnostic sensitivities for HCC, cyst, hemangioma, FFS and FFI, respectively. 2. The reference number should be put ahead of the period sign, not after. Response: We thank the reviewer for this comment. We have moved the reference number ahead of the period sign, as suggested by the reviewer. 3. Table 3. It is better to present it as a heat map, instead of a single table, to better visualize the results. Response: We thank you the reviewer for this suggestion. Table 3 has been revised and presented as a heatmap accordingly. Submitted filename: response_letter_r10.docx Click here for additional data file. 25 May 2021 Development and validation of artificial intelligence to detect and diagnose liver lesions from ultrasound images PONE-D-20-36551R1 Dear Dr. Chaiteerakij, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Khanh N.Q. Le Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: No ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: (No Response) Reviewer #2: The authors have addressed my comments. it is an interesting work in the area of ultrasound imaging. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No 31 May 2021 PONE-D-20-36551R1 Development and validation of artificial intelligence to detect and diagnose liver lesions from ultrasound images Dear Dr. Chaiteerakij: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Khanh N.Q. Le Academic Editor PLOS ONE
Table 1

Number of USG images from 3 participating hospitals, along with allocation of images for AI training and performance evaluation.

Training setaInternal test setaExternal validation cohorts
Cohort 1bCohort 2cPooled
Number of patients3487385311625936
Total images40397619156241329818922
Total images with FLLs182398013447341078
Number of lesions (%)
    Total20432 (100)845 (100)360 (100)835 (100)1195 (100)
    HCC2414 (11.8)102 (12.1)34 (9.4)104 (12.5)138 (11.5)
    Cyst6600 (32.3)215 (25.4)130 (36.1)87 (10.4)217 (18.2)
    Hemangioma5374 (26.3)217 (25.7)60 (16.7)202 (24.2)262 (21.9)
    FFS5110 (25.0)264 (31.2)120 (33.3)404 (48.4)524 (43.8)
    FFI934 (4.6)47 (5.6)16 (4.4)38 (4.6)54 (4.5)
Median sizes in cm (IQR)
    Total1.6 (1.7)1.6 (1.6)1.5 (1.3)1.8 (1.7)1.7 (1.6)
    HCC3.7 (5.5)3.3 (5.8)2.3 (6.6)3.9 (4.4)3.9 (5.5)
    Cyst1.4 (1.5)1.0 (0.8)1.0 (0.7)1.2 (0.9)1.1 (0.9)
    Hemangioma1.2 (1.2)1.4 (1.1)1.9 (1.1)1.5 (1.5)1.6 (1.4)
    FFS1.7 (1.1)1.8 (1.4)1.9 (1.4)1.7 (1.3)1.8 (1.3)
    FFI2.5 (2.5)2.4 (3.5)1.7 (1.0)2.4 (2.7)2.1 (2.7)
Total images without FLLs22158539052801256417844

aKing Chulalongkorn Memorial Hospital, Bangkok, Thailand

bMahachai Hospital, Samut Sakhon, Thailand

cQueen Savang Vadhana Memorial Hospital, Chonburi, Thailand

Table 2

Performance of the AI system on the internal test set and external validation cohorts.

Internal test setaExternal validation cohortsP*
Cohort 1bCohort 2cPooled
Overall
Detection rate87.0 (84.3–89.6)80.3 (74.8–85.8)73.9 (69.9–78.0)75.0 (71.7–78.3)<0.001
Diagnostic sensitivity83.9 (80.3–87.4)84.6 (79.0–90.2)85.7 (81.7–89.6)84.9 (81.6–88.2)0.69
Diagnostic specificity97.1 (96.5–97.7)97.2 (96.3–98.2)97.0 (96.3–97.7)97.1 (96.5–97.6)0.98
HCC
Detection rate85.3 (78.4–92.2)91.2 (81.6–101)74.0 (65.6–82.5)78.3 (71.4–85.2)0.16
Diagnostic sensitivity73.6 (64.3–82.8)74.2 (58.8–89.6)84.4 (76.3–92.5)81.5 (74.2–88.8)0.19
Diagnostic specificity97.8 (96.7–98.9)96.1 (93.7–98.5)93.6 (91.5–95.7)94.4 (92.8–96.0)0.55
Cyst
Detection rate89.3 (85.2–93.4)76.9 (69.7–84.2)85.1 (77.6–92.5)80.2 (74.9–85.5)0.008
Diagnostic sensitivity97.9 (95.9–99.9)91.0 (85.4–96.6)98.6 (96.0–100)94.3 (90.8–97.7)0.07
Diagnostic specificity98.3 (97.2–99.4)97.8 (95.8–99.9)98.7 (97.7–99.7)98.5 (97.6–99.4)0.99
Hemangioma
Detection rate93.5 (90.3–96.8)78.3 (67.9–88.8)79.7 (74.2–85.2)79.4 (74.5–84.3)<0.001
Diagnostic sensitivity80.8 (75.4–86.2)74.5 (62.0–86.9)67.7 (60.5–74.9)69.2 (63.0–75.5)0.006
Diagnostic specificity95.0 (93.2–96.9)97.9 (96.1–99.7)96.2 (94.4–98.0)96.8 (95.5–98.1)0.12
FFS
Detection rate77.3 (72.2–82.3)80.0 (72.8–87.2)67.6 (63.0–72.1)70.4 (66.5–74.3)0.03
Diagnostic sensitivity98.0 (96.1–99.9)100 (96.2–100)98.5 (97.1–100)98.9 (97.9–100)0.41
Diagnostic specificity96.9 (95.5–98.4)95.8 (92.9–98.6)98.5 (97.2–99.8)97.5 (96.2–98.9)0.53
FFI
Detection rate89.4 (80.5–98.2)75.0 (53.8–96.2)63.2 (47.8–78.5)66.7 (54.1–79.3)0.004
Diagnostic sensitivity69.0 (55.1–83.0)83.3 (62.2–100)79.2 (62.9–95.4)80.6 (67.6–93.5)0.60
Diagnostic specificity97.4 (96.2–98.6)98.5 (97.1–100)98.1 (97.0–99.2)98.3 (97.4–99.1)0.97

aKCMH, King Chulalongkorn Memorial Hospital, Bangkok, Thailand

bMahachai Hospital, Samut Sakhon, Thailand

cQueen Savang Vadhana Memorial Hospital, Chonburi, Thailand

*P-value for two-tailed z-test for difference of proportion, comparing performance results in the internal test set and pooled external validation set. P-value of <0.05 was considered statistically significant.

†Clopper-Pearson confidence interval was calculated for performance value at boundaries (i.e. 0% and 100%)

Detection rates, diagnostic sensitivities and specificities are shown in percentages. 95% confidence intervals are shown in parenthesis.

Table 3

Confusion matrix for classification results on internal test set and external validation set.

    Internal test set
Definitive diagnosis
HCCCystHemangiomaFFSFFITotal
Predicted diagnosis by AIHCC64291278
Cyst3188420197
Hemangioma132164110190
FFS50102001216
FFI201602947
Total8719220320442728
    Pooled external validation set
Definitive diagnosis
HCCCystHemangiomaFFSFFITotal
Predicted diagnosis by AIHCC8833821132
Cyst4164700175
Hemangioma12214426166
FFS3553650378
FFI101402944
Total10817420836936895
  17 in total

1.  Focal fatty infiltration of the liver: analysis of prevalence and CT findings in children and young adults.

Authors:  B F Kammen; P Pacharn; R F Thoeni; Y Lu; A Qayyum; F Coakly; C A Gooding; R C Brasch
Journal:  AJR Am J Roentgenol       Date:  2001-11       Impact factor: 3.959

2.  Hepatocellular carcinoma surveillance rates in commercially insured patients with noncirrhotic chronic hepatitis B.

Authors:  D S Goldberg; A Valderrama; R Kamalakar; S S Sansgiry; S Babajanyan; J D Lewis
Journal:  J Viral Hepat       Date:  2015-01-12       Impact factor: 3.728

3.  Assessing Radiology Research on Artificial Intelligence: A Brief Guide for Authors, Reviewers, and Readers-From the Radiology Editorial Board.

Authors:  David A Bluemke; Linda Moy; Miriam A Bredella; Birgit B Ertl-Wagner; Kathryn J Fowler; Vicky J Goh; Elkan F Halpern; Christopher P Hess; Mark L Schiebler; Clifford R Weiss
Journal:  Radiology       Date:  2019-12-31       Impact factor: 11.105

4.  Surveillance for hepatocellular carcinoma in a Medicaid cirrhotic population.

Authors:  Lena B Palmer; Michael D Kappelman; Robert S Sandler; Paul H Hayashi
Journal:  J Clin Gastroenterol       Date:  2013-09       Impact factor: 3.062

5.  Classification of focal liver lesions on ultrasound images by extracting hybrid textural features and using an artificial neural network.

Authors:  Yoo Na Hwang; Ju Hwan Lee; Ga Young Kim; Yuan Yuan Jiang; Sung Min Kim
Journal:  Biomed Mater Eng       Date:  2015       Impact factor: 1.300

6.  Impact of surveillance for hepatocellular carcinoma on survival in patients with compensated cirrhosis.

Authors:  Ju Dong Yang; Ajitha Mannalithara; Andrew J Piscitello; John B Kisiel; Gregory J Gores; Lewis R Roberts; W Ray Kim
Journal:  Hepatology       Date:  2018-05-09       Impact factor: 17.425

7.  Randomized controlled trial of screening for hepatocellular carcinoma.

Authors:  Bo-Heng Zhang; Bing-Hui Yang; Zhao-You Tang
Journal:  J Cancer Res Clin Oncol       Date:  2004-07       Impact factor: 4.553

8.  Early detection, curative treatment, and survival rates for hepatocellular carcinoma surveillance in patients with cirrhosis: a meta-analysis.

Authors:  Amit G Singal; Anjana Pillai; Jasmin Tiro
Journal:  PLoS Med       Date:  2014-04-01       Impact factor: 11.069

9.  Prevalence of benign focal liver lesions: ultrasound investigation of 45,319 hospital patients.

Authors:  Tanja Eva-Maria Kaltenbach; Phillip Engler; Wolfgang Kratzer; Suemeyra Oeztuerk; Thomas Seufferlein; Mark Martin Haenle; Tilmann Graeter
Journal:  Abdom Radiol (NY)       Date:  2016-01

10.  Rate of hepatocellular carcinoma surveillance remains low for a large, real-life cohort of patients with hepatitis C cirrhosis.

Authors:  Sally Ann Tran; An Le; Changqing Zhao; Joseph Hoang; Lee Ann Yasukawa; Susan Weber; Linda Henry; Mindie H Nguyen
Journal:  BMJ Open Gastroenterol       Date:  2018-03-20
View more
  4 in total

1.  Artificial intelligence for identification of focal lesions in intraoperative liver ultrasonography.

Authors:  Ido Nachmany; Niv Pencovich; Yiftach Barash; Eyal Klang; Adar Lux; Eli Konen; Nir Horesh; Ron Pery; Nadav Zilka; Rony Eshkenazy
Journal:  Langenbecks Arch Surg       Date:  2022-09-07       Impact factor: 2.895

2.  The feasibility to use artificial intelligence to aid detecting focal liver lesions in real-time ultrasound: a preliminary study based on videos.

Authors:  Thodsawit Tiyarattanachai; Terapap Apiparakoon; Sanparith Marukatat; Sasima Sukcharoen; Sirinda Yimsawad; Oracha Chaichuen; Siwat Bhumiwat; Natthaporn Tanpowpong; Nutcha Pinjaroen; Rungsun Rerknimitr; Roongruedee Chaiteerakij
Journal:  Sci Rep       Date:  2022-05-11       Impact factor: 4.996

3.  Performance of Deep Learning-Based Algorithm for Detection of Pediatric Intussusception on Abdominal Ultrasound Images.

Authors:  Zheming Li; Chunze Song; Jian Huang; Jing Li; Shoujiang Huang; Baoxin Qian; Xing Chen; Shasha Hu; Ting Shu; Gang Yu
Journal:  Gastroenterol Res Pract       Date:  2022-08-12       Impact factor: 1.919

Review 4.  Artificial intelligence in liver ultrasound.

Authors:  Liu-Liu Cao; Mei Peng; Xiang Xie; Gong-Quan Chen; Shu-Yan Huang; Jia-Yu Wang; Fan Jiang; Xin-Wu Cui; Christoph F Dietrich
Journal:  World J Gastroenterol       Date:  2022-07-21       Impact factor: 5.374

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.