Literature DB >> 30417017

Tumor Segmentation in Contrast-Enhanced Magnetic Resonance Imaging for Nasopharyngeal Carcinoma: Deep Learning with Convolutional Neural Network.

Qiaoliang Li1, Yuzhen Xu1, Zhewei Chen1, Dexiang Liu2,3, Shi-Ting Feng4, Martin Law5, Yufeng Ye2,3, Bingsheng Huang1.   

Abstract

OBJECTIVES: To evaluate the application of a deep learning architecture, based on the convolutional neural network (CNN) technique, to perform automatic tumor segmentation of magnetic resonance imaging (MRI) for nasopharyngeal carcinoma (NPC).
MATERIALS AND METHODS: In this prospective study, 87 MRI containing tumor regions were acquired from newly diagnosed NPC patients. These 87 MRI were augmented to >60,000 images. The proposed CNN network is composed of two phases: feature representation and scores map reconstruction. We designed a stepwise scheme to train our CNN network. To evaluate the performance of our method, we used case-by-case leave-one-out cross-validation (LOOCV). The ground truth of tumor contouring was acquired by the consensus of two experienced radiologists.
RESULTS: The mean values of dice similarity coefficient, percent match, and their corresponding ratio with our method were 0.89±0.05, 0.90±0.04, and 0.84±0.06, respectively, all of which were better than reported values in the similar studies.
CONCLUSIONS: We successfully established a segmentation method for NPC based on deep learning in contrast-enhanced magnetic resonance imaging. Further clinical trials with dedicated algorithms are warranted.

Entities:  

Mesh:

Year:  2018        PMID: 30417017      PMCID: PMC6207874          DOI: 10.1155/2018/9128527

Source DB:  PubMed          Journal:  Biomed Res Int            Impact factor:   3.411


1. Introduction

Head and neck cancer (HNC), especially nasopharyngeal carcinoma (NPC), is an aggressive cancer type with high incidence rate in Southern China [1]. The cancer incidence data collected in Guangxi and Guangdong show that nasopharyngeal cancer is the fourth most common cancer for males [2]. External beam radiation therapy is the primary therapy to this cancer. The 3-year local control rate for NPC after therapy is higher than 80% and the 3-year overall survival rate is up to 90% [3]. Noninvasive medical imaging is of great importance to determine the tumor volume for successful radiation treatment planning [3, 4]. Dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI), a functional noninvasive imaging modality, plays a key role in the studies of cancer by providing information about physiological characteristics in tissues. Studies have concluded that DCE-MRI is useful in differentiating tumors from normal tissues in NPC [4]. Accurate segmentation of NPC tumors from DCE-MRI is important for the radiotherapy treatment planning and prognosis evaluation. However, the accuracy of tumor segmentation in DCE-MRI is affected by some imaging factors such as low spatial resolution, poor signal-to-noise ratio, partial volume effect, and the intensity changes during perfusion [5]. There have been many studies performed to automatically segment NPC tumors from medical images. Zhou et al. [6] performed NPC tumor segmentation in MR images by using Semi-Fuzzy C-means with the percent match (PM) values close to 0.87. Zhou et al. [7] performed NPC tumor segmentation in MRI by using the two-class support vector machine (SVM) method with PM values close to 0.79. Huang et al. [8] performed semisupervised NPC lesion extraction in MR images by using spectral clustering-based method with the positive predictive value up to 0.71. Huang et al. [9] performed NPC tumor segmentation by using Bayesian classifiers and SVM method with average specificity of 0.93. The above-mentioned methods were all conventional machine learning techniques that require subjective feature extraction and selection. Deep learning (DL) technique, such as convolutional neural network (CNN), has recently emerged as a powerful tool in solving the challenges aforementioned, which detects low-level features such as shape and texture information autonomously from small patches of the input images and then combines these features into high-level features for the image processing tasks such as classification, segmentation, and detection without the subjective feature extraction and selection [10, 11]. Deep learning techniques perform even better in generalization with new datasets [12]. To the best of our knowledge, DL with CNN technique in tumor segmentation has recently attracted research interest [13, 14]. Wang et al. [15] performed NPC tumor segmentation in MR images by using deep convolutional neural networks; however, the average Jaccard similarity coefficient (JSC) value was less than 0.8. In the current study, we reported an automatic and accurate segmentation method based on the CNN architecture with dynamic contrast-enhanced MRI.

2. Materials and Methods

2.1. CE-MRI and Preprocessing

Twenty-nine newly diagnosed NPC patients from August 2010 to April 2013 were included from the First Affiliated Hospital, Sun Yat-Sen University. This study was approved by the local institutional review board of Sun Yat-Sen University. Written informed consent was obtained from each patient before the MRI scan. PVE could severely affect the images whenever the tumor size is less than 3 times the full width at half maximum (FWHM) of the reconstructed image resolution [16]. Thus, the patients with lymph nodes or lesions smaller than 1 cm were excluded in the current study to avoid possible partial volume effects (PVE), according to the advice from the radiologists. Imaging of DCE-MRI was performed in the primary tumor region including the retropharyngeal nodes with regional nodal metastasis, in with a 3.0-T MRI system (Magnetom Trio, Siemens) with the field of view of 22cm×22cm×6cm (AP×RL×FH), a flip angle of 15°, and scanning time of 6 minutes and 47 seconds, resulting in 65 dynamic images. The contrast agent gadolinium-diethylenetriamine pentaacetic acid (Gd-DTPA) (Omniscan; Nycomed, Oslo, Norway) was injected intravenously as a bolus into the blood at around the 8th dynamic acquisition using a power injector system (Spectris Solaris, MedRad, USA) and a 25 mL saline flush at a rate of 3.5 mL/sec was immediately followed. The dose of Gd-DTPA was 0.1 mmol per body weight in kg of the patient. The matrix of the 65 reconstructed dynamic image was 144×144×20×65. The ground truth was manually contoured in ImageJ (National Institutes of Health, Bethesda, MD) with the consensus between two experienced radiologists (Dr. Yufeng Ye, 13 years' experience, and Dr. Dexiang Liu, 18 years' experience in Radiology) who were blind to this study. Since tumors were mostly enhanced at the 35th scan of our DCE-MRI, this scan from each patient was used for training and testing our DL model, and we only selected the scanned images containing the tumor area. There were a total of 87 slices of CE-MRI acquired from each of the 29 patients. To fulfill the requirement of large number of data in training the DL model, we augmented the 87 MRI to more than 60,000 slices of images by using the following methods [17], namely, rotating each slice between -10 degrees and 10 degrees with an interval of 2 degrees to augment each slice to 11 slices, changing the image contrast with an embedded Matlab function, Imadjust, to adjust the image contrast automatically to produce 33 extra different slices from one single slice and adding Gaussian noise to the images with a power of 1×10−8 to produce 2 different additional slices from each slice. Totally we augmented the images by 11×33×2=726 times for each patient's CE-MRI set to give a total of 63126 (87x726) slices. These augmented images were then normalized by performing Z-score translation [18], in which the image intensity value in each voxel was normalized by the mean intensity of this image.

2.2. CNN Network

The CNN network included two phases of feature representation and scores map reconstruction. In the feature representation phase, the network consisted of 2 Pool-Conv-ReLu blocks (P1-P2) and 4 Conv-ReLu blocks (C1-C4) (see Figure 1). A Pool-Conv-ReLu block included one pooling layer (Pool), one convolution layer (Conv), and one rectified linear units (ReLu) layer, while a Conv-ReLu block consisted of one convolution layer and one ReLu layer. The convolution layer detected local features from the input images and the ReLu layer accelerated the convergence. The pooling layer was designed for reducing the dimension of feature maps and network parameters. The input images with a matrix size of 144×144 were transformed into the feature maps of matrix size of 36×36 in the feature representation phase.
Figure 1

The schematic diagram of the proposed convolutional neural network (CNN) structure. The proposed CNN network includes two phases of feature representation and scores map reconstruction. The feature representation phase consists of 2 Pool-Conv-ReLu blocks and 3 Con-ReLu blocks, while the scores map reconstruction phase consists of 2 deconv-concat-conv blocks. The output of each layer is a three-dimensional matrix with size of h×w×d, where h and w are the length and width of the scores map, respectively, and d is the feature dimension. a×a indicates the matrix size of the convolution kernels. Conv: convolution, Relu: rectified linear units, Pool: pooling, Deconv: deconvolution.

In the scores map reconstruction phase (D1-D2, Ct1-Ct2, C5-C6), the images were reconstructed from the 36×36 feature maps. Two deconvolution layers (D1-D2) were applied to reconstruct an output image with a matrix size of 144×144. Since some image details could be missing in this reconstruction from the 36×36 feature maps, the fine features obtained from the previous feature representation phase were combined with the scores map to allow the integration of local and global multilevel contextual information. A concatenate layer was then used for the information connection. Then a convolution layer was applied for information fusion and the final reconstruction. The detailed parameters of the CNN network are shown in Table 1.
Table 1

Detailed parameters of the CNN network.

Block Layer Kernel size Stride Pad Output size
Input Data ----1441441
Conv-Relu 1 Conv151214414480
Relu1---14414480
Pool-Conv-Relu 1 Pool122-727280
Conv25127272120
Relu2---7272120
Pool-Conv-Relu 2 Pool222-3636120
Conv39143636500
Relu3---3636500
Conv-Relu 2 Conv41103636300
Relu4---3636300
Conv-Relu 3 Conv53113636100
Relu5---3636100
Conv-Relu 4 Conv631136361
Relu6---36361
Deconv1 Deconv142172721
Concatenate1 Concat1---7272121
Conv-Relu 5 Conv711072721
Relu7---72721
Deconv2 Deconv24211441441
Concatenate2 Concat2---14414481
Conv6 Conv81101441441
Output Flatten1---1441441

2.3. Model Training and Model-Based Segmentation

A stepwise training scheme was used to train the DL CNN network. Firstly, we trained the network in the feature representation phase and a 36×36 scores' map was obtained. Based on this 36×36' scores map, we next trained the deconvolution layer, the concatenate layer, and the convolution layer, resulting in a 72×72 scores' map. Finally, based on the network parameters and output feature maps acquired in the second step, we reconstructed the images with matrix size of 144×144 scores' map. In the training process, the weights were optimized in each iteration. The weight of a Gaussian distribution with mean of 0 and standard deviation of 1 was used in the convolution kernel at the initialization step. The training parameters were as follows: basic learning rate: 1×10−7, step size: 1x105, gamma: 0.1, momentum: 0.9, weight decay: 5x10-4. It took 52 hours for a complete training procedure with a NVIDIA GeForce GTX 980 GPU equipped on an Intel Core i7 3.5 GHz computer. We used the trained model in the segmentation tasks of NPC tumor lesions in the testing dataset. The testing images were input into the trained model. A score map representing the tumor region of the NPC tumor was acquired for each input image.

2.4. Tumor Segmentation

We used the testing dataset to make forward propagation and evaluated the segmentation performance based on the trained model. Parameters of recall, precision, and dice similarity coefficient (DSC) were given bywhere true positive (TP) denotes the correctly identified tumor area and false positive (FP) denotes the tumor area, but the area is normal tissue in ground truth and false negative (FN) denotes normal tissue but the pixel isolated is tumor area in ground truth. And those are the results for each patient. For the comparisons with other published results, values of corresponding ratio (CR), percent match (PM) [7], and Jaccard similarity coefficient (JSC) [15] were also calculated asThe model validation technique of leave-one-out cross-validation (LOOCV) was used such that, in one repetition, the images of 28 patients were used as the training dataset (which were then augmented to >60000 images), and the images of the remaining patient were used as the testing dataset. After each patient's images in these 87 images were tested, the mean and variance of DSC, recall, CR, PM, and JSC were calculated to evaluate the segmentation performance of our method.

3. Results

Table 2 tabulates the tumor volumes as segmented by the radiologist (the golden standard) and by the proposed automatic segmentation method together with DSC, CR, PM, recall, and JSC. These values were calculated for each patient, not for each lesion. Table 3 shows the comparison of segmentation performance in terms of DSC, CR, and PM between our current results of DL CNN network and those of published results using other models. The mean DSC with our method for 29 patients was 0.89±0.05, and the range was 0.80-0.95. The mean PM with our method for 29 patients was 0.90±0.04 with a range of 0.71-0.92, which was higher compared to the mean PM of the value less than 0.9 in other studies. The mean CR was 0.84±0.06 and the range was 0.83-0.96, while the mean CR was 0.72 in similar studies using other algorithms [7, 15, 19].
Table 2

The segmentation performance of all cases.

Patients numberVolume of current DL method (cm3)Volume obtained by the two readers (cm3)Percent matchCorresponding ratioDice similarity coefficientRecallJaccard similarity coefficient
17.88.00.840.740.830.810.70
25.26.20.900.760.820.750.70
35.16.40.890.720.800.720.66
418.518.40.880.830.890.890.79
57.06.50.860.820.890.920.80
611.910.60.850.820.890.940.81
711.911.90.830.740.830.820.71
811.010.20.850.820.890.920.80
95.96.70.900.780.840.790.73
1014.817.10.860.710.800.750.66
114.44.70.960.910.930.910.87
1217.317.90.940.890.920.900.85
134.54.30.880.840.900.920.82
1446.745.00.930.910.950.970.90
155.85.70.880.820.880.890.79
168.89.20.900.830.880.870.79
175.76.80.920.790.840.780.73
1810.711.20.940.890.920.900.85
1910.610.90.950.910.940.930.88
208.68.90.930.880.910.900.84
2113.313.90.930.870.910.890.83
2213.713.60.950.920.950.950.90
236.36.50.900.830.880.870.79
2416.216.50.950.920.940.940.89
2518.217.70.940.920.950.960.91
2611.411.10.910.870.920.930.85
2710.811.00.880.800.870.860.76
2817.117.90.950.900.930.910.86
2912.012.20.940.900.930.920.87
Mean±Std--0.90±0.040.84±0.060.89±0.050.88 ±0.070.81 ±0.07
Table 3

Comparison of segmentation performance among our convolutional neural network model and other models. N.A.: not available.

Dice similarity coefficientCorresponding ratioPercent match
StudyAlgorithmMean ± SDRangeMean ± SDRangeMean ± SDRange
Current studyConvolutional neural network0.89±0.050.80-0.950.84±0.060.71-0.920.90±0.040.83-0.96
Zhou at al. [7]Support vector machineN.AN.A0.72±0.060.58~0.850.79±0.070.65-0.91
Huang et al. [14]Hidden Markov random fieldN.AN.A0.720.44-0.910.850.64-1.00
Wang et al. [15]Deep Convolutional Neural NetworksN.A-0.80N.AN.AN.A-0.90
Figure 2 shows the segmentation with high accuracy, in which the DSC, CR, and PM were 0.941, 0.915, and 0.950, respectively, showing good accordance between segmentation results using our current DL CNN network and ground truth.
Figure 2

NPC tumor segmentation with high accuracy using the current deep learning method with convolutional neural network. (a) The original image. (b) Ground truth (white line). (c) Segmentation from our deep learning method result (white line) with the dice similarity coefficient = 0.941, corresponding ratio = 0.915, and percent match = 0.950.

Figure 3 shows a less accurate segmentation result as obtained by our current DL CNN model with values of DSC, CR, and PM being 0.797, 0.731, and 0.937, respectively, showing slight difference between segmentation results using our current DL CNN network and ground truth.
Figure 3

NPC tumor segmentation with less accuracy using the current deep learning method with convolutional neural network. (a) The original image. (b) Ground truth (white line). (c) Segmentation from our deep learning method result (white line) with the dice similarity coefficient = 0.797, corresponding ratio = 0.731, and percent match = 0.937.

4. Discussion

Based on the CNN technique, we achieved a supervised segmentation method for NPC tumors in CE-MRI with high accuracy of mean DSC being 0.89. The performance was also robust with a low standard deviation of 0.05 for DSC among the results of different images. For comparison with the other studies, we calculated CR and PM. The mean values of CR and PM achieved with our method were 0.84 and 0.90, respectively. Compared with similar studies in literature, results of CR and PM in our study are more superior, indicating better accuracy in tumor segmentation with our current CNN technique than with other models with highest mean CR and PM of 0.72 and 0.90, respectively [7, 15, 19]. This may indicate that our method has indeed improved the automatic segmentation accuracy. Firstly, the improvement may lie in the application of CNN to extract the image features automatically and objectively. In our model, the low-level features were combined into high-level features with semantic information through convolutions (Figure 1). By iterations through the back propagation algorithm, we highlighted the characteristics associated with the targeted area and gradually suppressed irrelevant features [12]. In this way, our model can extract the most useful features and achieve better segmentation results. Secondly, in our designed network architecture, we fused the different feature maps at feature representation phase and scores map reconstruction phase for the final reconstruction. As shown in Figure 4(a), which was acquired in the reconstruction phase, the tumor location and shape are roughly visible; however, they are unclear. Through the fusion of this feature map and the fine-feature map acquired in feature representation phase, we may fix the problem of information loss in the reconstruction process. As shown in Figure 4(b), we finally had better segmentation through the reconstruction from the fused feature maps.
Figure 4

A feature map fusion. (a) Scores map acquired in the reconstruction phase and (b) scores map reconstructed from the fused feature maps, indicating that score map reconstruction using fused maps is better than that during reconstruction phase.

There is space to further improve the accuracy and effectiveness of our current model. As shown in Table 3, our method resulted in less accurate DSC results of 0.80 in some cases. For further improvement, we may include T2 weighted images, since T2 weighted images are widely used in the manual contouring of tumor regions. Therefore we would expect to have better performance with both the DCE-MRI with T2 weighted images. We applied the Z-score translation in preprocessing to normalize the DCE-MRI [18]. However, some information could be lost during this normalization. Therefore, we may investigate an appropriate method of normalization to avoid the loss of intrinsic image features. Importantly, we may improve our network architecture, such as the depth of our network, for the direct training of the 3D images and the incorporation of time domain information from the dynamic scans. In future studies, it is expected to further improve our method and the segmentation results with these ideas.

5. Conclusion

A robust segmentation method for NPC tumor based on deep learning convolutional neural network and CE-MRI has been established. The tumors can be segmented successfully in seconds with high accuracy. This automatic segmentation method may be time-effective in tumor contouring for routine radiotherapy treatment planning. Future studies may aim to improve the segmentation accuracy and efficiency with more training data and optimized network structure, thus helping clinicians improve the segmentation results in the clinical practice of NPC.
  10 in total

Review 1.  Partial-volume effect in PET tumor imaging.

Authors:  Marine Soret; Stephen L Bacharach; Irène Buvat
Journal:  J Nucl Med       Date:  2007-05-15       Impact factor: 10.057

2.  Semi-supervised nasopharyngeal carcinoma lesion extraction from magnetic resonance images using online spectral clustering with a learned metric.

Authors:  Wei Huang; Kap Luk Chan; Yan Gao; Jiayin Zhou; Vincent Chong
Journal:  Med Image Comput Comput Assist Interv       Date:  2008

3.  Extranodal spread in the neck: MRI detection on the basis of pixel-based time-signal intensity curve analysis.

Authors:  Misa Sumi; Takashi Nakamura
Journal:  J Magn Reson Imaging       Date:  2011-04       Impact factor: 4.813

Review 4.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

5.  Multimodal Neuroimaging Feature Learning With Multimodal Stacked Deep Polynomial Networks for Diagnosis of Alzheimer's Disease.

Authors:  Jun Shi; Xiao Zheng; Yan Li; Qi Zhang; Shihui Ying
Journal:  IEEE J Biomed Health Inform       Date:  2017-01-19       Impact factor: 5.772

6.  Segmentation and visualization of nasopharyngeal carcinoma using MRI.

Authors:  Jiayin Zhou; Tuan-Kay Lim; Vincent Chong; Jing Huang
Journal:  Comput Biol Med       Date:  2003-09       Impact factor: 4.589

7.  Nasopharyngeal carcinoma segmentation via HMRF-EM with maximum entropy.

Authors: 
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2015-08

8.  Dynamic contrast-enhanced magnetic resonance imaging for characterising nasopharyngeal carcinoma: comparison of semiquantitative and quantitative parameters and correlation with tumour stage.

Authors:  Bingsheng Huang; Chun-Sing Wong; Brandon Whitcher; Dora Lai-Wan Kwong; Vincent Lai; Queenie Chan; Pek-Lan Khong
Journal:  Eur Radiol       Date:  2013-02-02       Impact factor: 5.315

9.  Long-term treatment outcome of recurrent nasopharyngeal carcinoma treated with salvage intensity modulated radiotherapy.

Authors:  Yi-Jun Hua; Fei Han; Li-Xia Lu; Hai-Qiang Mai; Xiang Guo; Ming-Huang Hong; Tai-Xiang Lu; Chong Zhao
Journal:  Eur J Cancer       Date:  2012-07-25       Impact factor: 9.162

10.  Brain tumor segmentation with Deep Neural Networks.

Authors:  Mohammad Havaei; Axel Davy; David Warde-Farley; Antoine Biard; Aaron Courville; Yoshua Bengio; Chris Pal; Pierre-Marc Jodoin; Hugo Larochelle
Journal:  Med Image Anal       Date:  2016-05-19       Impact factor: 8.545

  10 in total
  11 in total

Review 1.  Artificial Intelligence in radiotherapy: state of the art and future directions.

Authors:  Giulio Francolini; Isacco Desideri; Giulia Stocchi; Viola Salvestrini; Lucia Pia Ciccone; Pietro Garlatti; Mauro Loi; Lorenzo Livi
Journal:  Med Oncol       Date:  2020-04-22       Impact factor: 3.064

2.  A convolutional neural network combined with positional and textural attention for the fully automatic delineation of primary nasopharyngeal carcinoma on non-contrast-enhanced MRI.

Authors:  Lun M Wong; Qi Yong H Ai; Darren M C Poon; Macy Tong; Brigette B Y Ma; Edwin P Hui; Lin Shi; Ann D King
Journal:  Quant Imaging Med Surg       Date:  2021-09

3.  Convolutional Neural Network Intelligent Segmentation Algorithm-Based Magnetic Resonance Imaging in Diagnosis of Nasopharyngeal Carcinoma Foci.

Authors:  Deli Wang; Zheng Gong; Yanfen Zhang; Shouxi Wang
Journal:  Contrast Media Mol Imaging       Date:  2021-08-13       Impact factor: 3.161

4.  A Collaborative Dictionary Learning Model for Nasopharyngeal Carcinoma Segmentation on Multimodalities MR Sequences.

Authors:  Haiyan Wang; Guoqiang Han; Haojiang Li; Guihua Tao; Enhong Zhuo; Lizhi Liu; Hongmin Cai; Yangming Ou
Journal:  Comput Math Methods Med       Date:  2020-08-28       Impact factor: 2.238

5.  Convolutional Neural Network-Processed MRI Images in the Diagnosis of Plastic Bronchitis in Children.

Authors:  Xiaoqun Chen; Rong Lu; Feng Zhao
Journal:  Contrast Media Mol Imaging       Date:  2021-09-13       Impact factor: 3.161

6.  Evaluation of deep learning-based multiparametric MRI oropharyngeal primary tumor auto-segmentation and investigation of input channel effects: Results from a prospective imaging registry.

Authors:  Kareem A Wahid; Sara Ahmed; Renjie He; Lisanne V van Dijk; Jonas Teuwen; Brigid A McDonald; Vivian Salama; Abdallah S R Mohamed; Travis Salzillo; Cem Dede; Nicolette Taku; Stephen Y Lai; Clifton D Fuller; Mohamed A Naser
Journal:  Clin Transl Radiat Oncol       Date:  2021-10-16

7.  Multiscale Local Enhancement Deep Convolutional Networks for the Automated 3D Segmentation of Gross Tumor Volumes in Nasopharyngeal Carcinoma: A Multi-Institutional Dataset Study.

Authors:  Geng Yang; Zhenhui Dai; Yiwen Zhang; Lin Zhu; Junwen Tan; Zefeiyun Chen; Bailin Zhang; Chunya Cai; Qiang He; Fei Li; Xuetao Wang; Wei Yang
Journal:  Front Oncol       Date:  2022-03-18       Impact factor: 6.244

8.  Lightweight Compound Scaling Network for Nasopharyngeal Carcinoma Segmentation from MR Images.

Authors:  Yi Liu; Guanghui Han; Xiujian Liu
Journal:  Sensors (Basel)       Date:  2022-08-05       Impact factor: 3.847

9.  Fully-Automated Segmentation of Nasopharyngeal Carcinoma on Dual-Sequence MRI Using Convolutional Neural Networks.

Authors:  Yufeng Ye; Zongyou Cai; Bin Huang; Yan He; Ping Zeng; Guorong Zou; Wei Deng; Hanwei Chen; Bingsheng Huang
Journal:  Front Oncol       Date:  2020-02-19       Impact factor: 6.244

10.  Predictive Analytics in Spine Oncology Research: First Steps, Limitations, and Future Directions.

Authors:  Elie Massaad; Nida Fatima; Muhamed Hadzipasic; Christopher Alvarez-Breckenridge; Ganesh M Shankar; John H Shin
Journal:  Neurospine       Date:  2019-12-31
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.