Literature DB >> 34481459

Automatic identification of suspicious bone metastatic lesions in bone scintigraphy using convolutional neural network.

Yemei Liu1, Pei Yang1, Yong Pi2, Lisha Jiang1, Xiao Zhong1, Junjun Cheng1, Yongzhao Xiang1, Jianan Wei2, Lin Li1, Zhang Yi2, Huawei Cai3, Zhen Zhao4.   

Abstract

BACKGROUND: We aimed to construct an artificial intelligence (AI) guided identification of suspicious bone metastatic lesions from the whole-body bone scintigraphy (WBS) images by convolutional neural networks (CNNs).
METHODS: We retrospectively collected the 99mTc-MDP WBS images with confirmed bone lesions from 3352 patients with malignancy. 14,972 bone lesions were delineated manually by physicians and annotated as benign and malignant. The lesion-based differentiating performance of the proposed network was evaluated by fivefold cross validation, and compared with the other three popular CNN architectures for medical imaging. The average sensitivity, specificity, accuracy and the area under receiver operating characteristic curve (AUC) were calculated. To delve the outcomes of this study, we conducted subgroup analyses, including lesion burden number and tumor type for the classifying ability of the CNN.
RESULTS: In the fivefold cross validation, our proposed network reached the best average accuracy (81.23%) in identifying suspicious bone lesions compared with InceptionV3 (80.61%), VGG16 (81.13%) and DenseNet169 (76.71%). Additionally, the CNN model's lesion-based average sensitivity and specificity were 81.30% and 81.14%, respectively. Based on the lesion burden numbers of each image, the area under the receiver operating characteristic curve (AUC) was 0.847 in the few group (lesion number n ≤ 3), 0.838 in the medium group (n = 4-6), and 0.862 in the extensive group (n > 6). For the three major primary tumor types, the CNN-based lesion identifying AUC value was 0.870 for lung cancer, 0.900 for prostate cancer, and 0.899 for breast cancer.
CONCLUSION: The CNN model suggests potential in identifying suspicious benign and malignant bone lesions from whole-body bone scintigraphic images.
© 2021. The Author(s).

Entities:  

Keywords:  Artificial intelligence; Bone metastasis; Bone scintigraphy; Convolutional neural network

Mesh:

Year:  2021        PMID: 34481459      PMCID: PMC8417997          DOI: 10.1186/s12880-021-00662-9

Source DB:  PubMed          Journal:  BMC Med Imaging        ISSN: 1471-2342            Impact factor:   1.930


Background

Bone metastasis commonly appears in the advanced stages of cancers [1-4]. It seriously affects the survival quality of patients due to the occurrence of adverse skeletal-related events [2, 5, 6]. The early diagnosis of bone metastasis is beneficial to make appropriate and timely treatment of metastatic bone disease, which can improve the quality of survival [7-10]. Even after the advent of single-photon emission computed tomography combined with computed tomography (SPECT/CT), whole-body bone scintigraphy (WBS) is a standard method to survey the existence and extent of bone metastasis [11]. However, the image resolution and the specificity of WBS are lacking [12]. And the interpretation of WBS is an experience-dependent work and the diagnostic agreement of inter-observer is not satisfactory [13]. Previously, we had proposed an automated diagnostic system of bone metastasis based on multi-view bone scans using an attention-augmented deep neural network [14, 15]. While it achieved considerable accuracy in the patient-based diagnosis from WBS images, a definitive diagnosis for suspicious bone metastatic lesions is still crucial for pragmatic decisions, such as precise bone biopsy, bone surgery and external beam radiotherapy [16]. Thus, a new artificial intelligence (AI) model with lesion-based diagnosis from the WBS image is more valuable for the clinic. Therefore, we fed a fully annotated WBS images dataset to construct a new AI model and evaluated its lesion-based performance in automatic diagnosing suspicious bone metastatic lesions.

Methods

This retrospective single-center study was approved by the Institutional Ethics Committee of West China Hospital of Sichuan University. The written informed consent was waived from the Institutional Ethics Committee of West China Hospital of Sichuan University.

Data resource

The WBS images of patients who were identified lung cancer, prostate cancer and breast cancer were retrieved from our hospital database within the period from Feb. 2012 to Apr. 2019. The WBS was performed using two gamma cameras (GE Discovery NM/CT 670 and Philips Precedence 16 SPECT/CT). The patient received 555 to 740 MBq of technetium-99 m methylene diphosphonate (99mTc-MDP; purchased from Syncor Pharmaceutical Co., Ltd, Chengdu, China) by intravenous injection, and the anterior and posterior views WBS images were obtained approximately 3 h post-injection. The gamma cameras were equipped with low-energy, high-resolution, parallel-hole collimators. The scan speed was 16–20 cm/min, and the matrix size was 512 × 1024. Energy peak was centered at 140 keV with 15% to 20% windows. The visible bone lesion in WBS images was manually delineated by human experts and annotated into malignant and benign according to the following criteria [17, 18]: Malignant: bone lesion with increased 99mTc-MDP were identified as malignant (1) when computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography-computed tomography (PET/CT), etc. presented bone destruction; (2) when it appeared newly but couldn’t be ruled out as malignant in follow-up bone scan; (3) when it presented flare phenomenon; (4) when it enlarged and thickened significantly after at least 3 months follow-up. Benign: bone lesion with increased 99mTc-MDP were identified as benign (1) when CT, MRI and PET/CT, etc. demonstrated fracture, bone cyst, osteogeny, osteophyte, bone bridge, degenerative osteoarthrosis; (2) when it appeared around the bone joint; (3) when it confirmed as trauma. The diagram of manual delineation and annotation was shown in Fig. 1. Additionally, the patient-based WBS image was assigned to malignant once a lesion was identified as malignant. Finally, from the 3352 patients, 14,972 visible bone lesions were identified as benign or malignant. According to the total number of lesions per WBS image [19], we divided all cases into three groups: few lesions group: 1–3 lesions; medium lesions group: 4–6 lesions; extensive lesions group: > 6 lesions.
Fig. 1

The diagram of manual annotation in WBS image. All visible bone lesions were delineated and annotated as benign and malignant. Red areas represent malignant lesions, while green areas represent benign lesions

The diagram of manual annotation in WBS image. All visible bone lesions were delineated and annotated as benign and malignant. Red areas represent malignant lesions, while green areas represent benign lesions

Model architecture

We implemented 2D CNN to automatically identification of bone metastatic lesions. Our network is based on the architecture of ResNet50 [20]. The CNN model was pre-trained on ImageNet, and fine-tuned on our own dataset. Before training the network, a pre-processing step was performed for data curation. The WBS and corresponding lesion mask were resized to 512 × 256. Considering the diagnosis of bone lesions was tremendously correlated to the location and burden extent, we stacked the full-sized images and the corresponding lesion mask on channel, instead of only inputting ROI of lesions. The data consisted of the original WBS image, the corresponding lesion mask and the qualitative of the lesion was used for CNN training. The fivefold cross validation was performed for evaluating the ability of the trained network model to achieve the qualitative task of bone scan lesions. Additionally, three state-of-the-art CNNs that included Inception V3 [21], VGG16 [22] and DenseNet169 [23] were compared with the proposed network. The developed network was implemented using PyTorch [24], and trained using Adam [25] as the optimizer with a learning rate of 0.001 for 300 epochs. The mini-batch size was fixed 8. During the training process, random horizontal flipping with a probability of 0.5 was applied to the input to increase the diversity of the data. The detailed network architecture is shown in Fig. 2.
Fig. 2

Architecture of convolutional neural network for AI model

Architecture of convolutional neural network for AI model

Statistical analysis

The performance of AI was evaluated using diagnostic sensitivity, specificity, accuracy, positive predictive value (PPV), negative predictive value (NPV) and the area under the receiver operating characteristic curve (AUC). The Chi-square test was performed to compare differences in the AI performance between different number of lesions and different primary tumor types. The confusion matrix showed the numbers of true positive, true negative, false positive and false negative. All analyses were conducted using statistical software SPSS22.0 (SPSS Inc, Chicago, Illinois, USA). P values less than 0.05 were considered statistically significant.

Results

Baseline characteristics of patients

3352 cancer patients (Age: 61.61 ± 12.69y; Gender: 1758 males and 1594 females) were retrospectively included in the study and 43.85% of all patients presented bone metastasis. A total of 14,972 visible bone lesions were recognized in all WBS images and 51.23% of them were identified metastasis. The lesion-based metastasis rate was 50.13% in lung cancer, 57.39% in prostate cancer, and 44.61% in breast cancer, respectively. The detailed information was listed in Table 1.
Table 1

The summary of patient-based and lesion-based analysis in all WBS images

Lung cancerProstate cancerBreast cancerTotal
Patient-based
Total1253101710823352
Malignant5674664371470
Benign6865516451882
Metastasis rate45.25%45.82%40.39%43.85%
Lesion-based
Total49375623441214,972
Malignant2475322719687670
Benign2462239624446812
Metastasis rate50.13%57.39%44.61%51.23%
The summary of patient-based and lesion-based analysis in all WBS images

The performance of the proposed network

After fivefold cross validation, the CNN model demonstrated an average sensitivity, specificity, accuracy, PPV and NPV for all visible bone lesions were 81.30%, 81.14%, 81.23%, 81.89% and 80.61%, respectively. When compared with the other three start-of-art CNNs, our proposed network achieved the best accuracy in identification the bone lesions at bone scintigraphy (Tables 2, 3).
Table 2

The fivefold cross validation results of the proposed network

Fold 1Fold 2Fold 3Fold 4Fold 5
Number of patients (benign/malignant)669 (376/293)669 (376/293)671 (376/295)669 (376/293)674 (378/296)
Number of lesions (benign/malignant)2986 (1505/1481)2952 (1389/1563)3084 (1468/1616)2897 (1475/1422)3053 (1465/1588)
Sensitivity83.1983.1179.8976.7983.5
Specificity78.0782.5882.782.7879.59
Accuracy80.6182.8681.2379.8481.62
PPV78.8784.383.5681.1381.6
NPV82.5181.2978.8878.7281.65
Table 3

The comparison of the proposed network and other three networks

SensitivitySpecificityAccuracyPPVNPV
Our model81.3081.1481.2381.8980.61
InceptionV377.2984.0280.6183.5378.00
VGG1678.7383.5181.1383.3979.14
DenseNet16967.9085.7376.7183.1772.16
The fivefold cross validation results of the proposed network The comparison of the proposed network and other three networks

Subgroup analysis of proposed network

Based on the number of lesions per image, we found that the AI model reached the highest sensitivity (89.56%, P < 0.001), accuracy (82.79%, P = 0.018) and PPV (87.37%, P < 0.001) in the extensive lesions group as shown in Table 4. Whereas, the highest specificity (89.41%, P < 0.001) and NPV (86.76%, P < 0.001) of the AI model were captured in few lesions group. We also calculated the AUC to evaluate the diagnostic performance of the AI model, which was 0.847 in the few lesions group, 0.838 in the medium lesions group, and 0.862 in the extensive lesions group. And the confusion matrix directly demonstrated the true labels and predicted labels in the three groups (Fig. 3).
Table 4

The lesion-based diagnostic performance of AI model in testing cohort and the comparison of the AI performance among few, medium and extensive lesion groups

Group for number of lesionsχ2P value
FewMediumExtensive
Sensitivity58.6364.3489.56163.41 < 0.001
Specificity89.4185.2462.85108.69 < 0.001
Accuracy81.7878.0382.798.060.018
PPV64.8969.6787.3783.70 < 0.001
NPV86.7682.0467.9349.24 < 0.001

Chi-square test was performed to compare the performance of AI model among different groups of the number of lesions. Few lesions group: 1–3 lesions per image; Medium lesions group: 4–6 lesions per image; Extensive lesions group: > 6 lesions per image

Fig. 3

The confusion matrix of few lesions group (A), medium lesions group (B), extensive lesions group (C). The ROC of the three groups in the lesion-based diagnosis (D)

The lesion-based diagnostic performance of AI model in testing cohort and the comparison of the AI performance among few, medium and extensive lesion groups Chi-square test was performed to compare the performance of AI model among different groups of the number of lesions. Few lesions group: 1–3 lesions per image; Medium lesions group: 4–6 lesions per image; Extensive lesions group: > 6 lesions per image The confusion matrix of few lesions group (A), medium lesions group (B), extensive lesions group (C). The ROC of the three groups in the lesion-based diagnosis (D) The detailed results based on the primary tumor types were shown in Table 5, the results demonstrated the highest diagnostic sensitivity (84.66%, P = 0.002) in the prostate cancer group. Albeit slightly higher accuracy (82.30%) in the prostate cancer group, there was no statistical significance (P = 0.209) comparing with the lung cancer group (79.40%) and breast cancer group (81.82%). The specificity in lung cancer (82.52%), prostate cancer (79.07%) and breast cancer (81.78%) group also did not indicate statistical significance between each other (P = 0.354). Furthermore, the AUC was 0.870 for lung cancer, 0.900 for prostate cancer, 0.899 for breast cancer. The confusion matrix directly demonstrated the true labels and predicted labels in the three groups (Fig. 4).
Table 5

The lesion-based diagnostic performance of AI model in the testing cohort and the comparison of the AI performance among lung, prostate and breast cancers

Group for primary tumor typesχ2P value
Lung cancerProstate cancerBreast cancer
Sensitivity76.1684.6681.6512.880.002
Specificity82.5279.0781.782.080.354
Accuracy79.4082.3081.823.130.209
PPV81.1284.4378.336.520.038
NPV77.7079.5285.058.850.012

Chi-square test was performed to compare the performance of AI model among different tumor types

Fig. 4

The confusion matrix of lung cancer group (A), prostate cancer group (B), breast cancer group (C). The ROC of the three groups in the lesion-based diagnosis (D)

The lesion-based diagnostic performance of AI model in the testing cohort and the comparison of the AI performance among lung, prostate and breast cancers Chi-square test was performed to compare the performance of AI model among different tumor types The confusion matrix of lung cancer group (A), prostate cancer group (B), breast cancer group (C). The ROC of the three groups in the lesion-based diagnosis (D) Additionally, we also evaluated the lesion-based diagnostic performance of the AI model according to the different number of lesions per image (few, medium and extensive lesions group) in lung cancer, prostate cancer and breast cancer, respectively. The results were supported as Additional file 1: Table 1 and Additional file 2: Figs. 1, 2, and 3.

Discussion

The definitive identification of abnormal bone lesions is beneficial to proper personalized treatment and subserves the patients who were suffering from advanced malignant cancers [26]. However, the precise differentiation of suspicious bone lesions is still tricky based on the WBS images only [27]. In light of the superiority of artificial intelligence in image feature extraction and big data analysis, we developed a new AI model using the 2D CNN to explore the potential for automatically definitive identification of suspicious bone metastatic lesions from WBS images. In general, our AI model achieved moderate performance in the identification of suspicious bone lesions with a sensitivity of 81.30% and specificity of 81.14%. We found that AI indicated significantly higher accuracy in the extensive-lesions group (n > 6, accuracy = 82.79%) than that in the few-lesions group (n ≤ 3, accuracy = 81.78%, p < 0.05) and medium-lesions group (n = 4–6, accuracy = 78.03%, p < 0.05), this might be beneficial from the deep neural network which imitating human thinking model. Originally, classification of every single lesion is judged independently, regardless of the other lesions that appeared in the same image. However, nuclear medicine physicians usually take other lesions and additional cues into account when determining one single lesion itself. For example, an isolated lesion without other nearby lesions would be more difficult to assert benign or malignant, while multiple lesions that occur within a narrow region would be more likely malignant. We input corresponding lesion masks to the CNN and take the whole WBS image into account, and this might be a possible reason for the improved accuracy of the extensive-lesions group. Previous studies also reported AI for bone lesion identification from WBS images. The authors used a ladder network to pre-train a nerual network with an unlabeled dataset [28]. On the metastasis classification task, It reached a sensitivity of 0.657 and a specificity of 0.857. Another similar study also build a model to detect and identify bone metastasis from bone scintigraphy images through negative mining, pre-training, the convolutional neural network, and deep learning [29]. The mean lesion-based sensitivity and precision rates for bone metastasis classification were 0.72 and 0.90, respectively. In our study, the lesion-based sensitivity, specificity and precision values for metastasis classification were 0.813, 0.811 and 0.819, respectively. It is difficult to compare the difference of algorithms, all studies have used in-house datasets of a gold standard and these datasets were not open. We were not able to try other datasets using our algorithm. Therefore, the performances reported by other researchers can only be used as references, rather than for objective comparison. It is worth mentioning that the aforementioned AI was focused on the chest image instead of the whole body. This strategy excluded the influence from keen osteoarthritis, degenerative changes of lumbar/cervical vertebrae, but it was limited to analyzing the metastases in other regions such as the pelvis, sacrum, iliac joints and other distant lesions. Addittionaly, we stacked the WBS and the corresponding lesion mask in channel and input it into the network. Thus, this CNN approach could select any suspicious bone lesion that needs to be input manually and obviate missed lesion detection and wrong lesion detection. Three common kinds of primary cancers were investigated in this study. The different sensitivity among primary cancer types seemed to be affiliated to osteoblastic and osteolytic activity. The highest sensitivity appeared in the prostate cancer group and it is consistent with other former studies [17]. The probable reason is due to the typical osteoblastic metastasis principally in prostate cancer, though it is also associated with the osteoclastic process and bone resorption [30]. On the other hand, lung cancer and breast cancer group showed more significant osteolytic changes and corresponding mild radioactivity in lesions [31, 32]. Generally, our AI model achieved a moderate accuracy, sensitivity and specificity in the lesion-based diagnosis of WBS images, the false-positive lesions and false-negative lesions still could not be avoided. It is limited to the substantive character and specificity of 99mTc-MDP imaging technology. Most pathological bone conditions, whether of infectious, traumatic, neoplastic or other origin could demonstrate as an increased radioactive signal in WBS images [33]. There are still several limitations in the current study. Firstly, since it is impossible to obtain the pathological result of each lesion, we made the “gold labels” based on the patients’ medical records, the follow-up bone scans, CT, MRI, PET/CT images, etc., which may not be totally correct for every lesion. Secondly, the labeled lesions on WBS images were all visible, which means only the “hotspots” were included, whereas some “cold lesions” were missed. Then, at present, this AI model was constructed by those non-quantitative images, the indraught of anatomical localization parameter and quantitative index might further improve the property, all of which would be paid attention in our future studies. Even though the AI model is not always correct, it still can be used by nuclear medicine physicians for assisting the bone lesions analysis and the final interpretation of an examination, especially for the patients who could not be performed SPECT/CT timely due to the poverty of resource devices.

Conclusions

The AI model based on CNN reached a moderate lesion-based performance in the diagnosis of suspicious bone metastatic lesions from WBS images. Even though the AI model is not always correct, it could serve as an effective auxiliary tool for diagnosis and guidance in patients with suspicious bone metastatic lesions in daily clinical practice. Additional file 1. The lesion-based diagnostic performance of the AI model according to the different number of lesions per image (few, medium and extensive lesions group) in lung cancer, prostate cancer and breast cancer. Additional file 2: Fig. S1. The confusion matrix of few lesions group (A), medium lesions group (B), extensive lesions group (C) in lung cancer group. The ROC of the three groups in the lesion-based diagnosis (D). Additional file 3: Fig. S2. The confusion matrix of few lesions group (A), medium lesions group (B), extensive lesions group (C) in prostate cancer group. The ROC of the three groups in the lesion-based diagnosis (D). Additional file 4: Fig. S3. The confusion matrix of few lesions group (A), medium lesions group (B), extensive lesions group (C) in breast cancer group. The ROC of the three groups in the lesion-based diagnosis (D).
  25 in total

1.  Clinical benefit of early treatment with bone-modifying agents for preventing skeletal-related events in patients with genitourinary cancer with bone metastasis: A multi-institutional retrospective study.

Authors:  Takuya Owari; Makito Miyake; Yasushi Nakai; Shunta Hori; Mitsuru Tomizawa; Kazuki Ichikawa; Takuto Shimizu; Kota Iida; Shoji Samma; Yusuke Iemura; Hitoshi Momose; Chihiro Omori; Takeshi Otani; Masaomi Kuwada; Shuya Hirao; Nobuo Oyama; Yoshinori Nakagawa; Yoshiki Hayashi; Nobumichi Tanaka; Kiyohide Fujimoto
Journal:  Int J Urol       Date:  2019-03-18       Impact factor: 3.369

2.  Predictors of skeletal-related events in non-small cell lung cancer patients with bone metastases.

Authors:  Jong-Mu Sun; Jin Seok Ahn; Soohyeon Lee; Jung A Kim; Jeeyun Lee; Yeon Hee Park; Hee Chul Park; Myung-Ju Ahn; Yong Chan Ahn; Keunchil Park
Journal:  Lung Cancer       Date:  2010-07-03       Impact factor: 5.705

3.  Skeletal metastases in non-small cell lung cancer: a retrospective study.

Authors:  Asuka Tsuya; Takayasu Kurata; Kenji Tamura; Masahiro Fukuoka
Journal:  Lung Cancer       Date:  2007-04-23       Impact factor: 5.705

Review 4.  Dynamic process of prostate cancer metastasis to bone.

Authors:  Kwanchanit Tantivejkul; Linda M Kalikin; Kenneth J Pienta
Journal:  J Cell Biochem       Date:  2004-03-01       Impact factor: 4.429

5.  Frequency and distribution pattern of distant metastases in breast cancer patients at the time of primary presentation.

Authors:  Christoph Schneider; Mathias K Fehr; Rolf A Steiner; Daniela Hagen; Urs Haller; Daniel Fink
Journal:  Arch Gynecol Obstet       Date:  2002-11-15       Impact factor: 2.344

Review 6.  Bone metastases: When and how lung cancer interacts with bone.

Authors:  Ilaria Roato
Journal:  World J Clin Oncol       Date:  2014-05-10

7.  Deep learning based automatic segmentation of metastasis hotspots in thorax bone SPECT images.

Authors:  Qiang Lin; Mingyang Luo; Ruiting Gao; Tongtong Li; Zhengxing Man; Yongchun Cao; Haijun Wang
Journal:  PLoS One       Date:  2020-12-03       Impact factor: 3.240

8.  The EANM practice guidelines for bone scintigraphy.

Authors:  T Van den Wyngaert; K Strobel; W U Kampen; T Kuwert; W van der Bruggen; H K Mohan; G Gnanasegaran; R Delgado-Bolton; W A Weber; M Beheshti; W Langsteger; F Giammarile; F M Mottaghy; F Paycha
Journal:  Eur J Nucl Med Mol Imaging       Date:  2016-06-04       Impact factor: 9.236

9.  Lesion-Based Bone Metastasis Detection in Chest Bone Scintigraphy Images of Prostate Cancer Patients Using Pre-Train, Negative Mining, and Deep Learning.

Authors:  Da-Chuan Cheng; Te-Chun Hsieh; Kuo-Yang Yen; Chia-Hung Kao
Journal:  Diagnostics (Basel)       Date:  2021-03-15
View more
  1 in total

Review 1.  Emerging Applications of Deep Learning in Bone Tumors: Current Advances and Challenges.

Authors:  Xiaowen Zhou; Hua Wang; Chengyao Feng; Ruilin Xu; Yu He; Lan Li; Chao Tu
Journal:  Front Oncol       Date:  2022-07-19       Impact factor: 5.738

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.