Literature DB >> 31871484

Differentiation of Recurrence from Radiation Necrosis in Gliomas Based on the Radiomics of Combinational Features and Multimodality MRI Images.

Quan Zhang1, Jianyun Cao1,2, Junde Zhang2, Junguo Bu2, Yuwei Yu1, Yujing Tan2, Qianjin Feng1, Meiyan Huang1.   

Abstract

PURPOSE: To classify radiation necrosis versus recurrence in glioma patients using a radiomics model based on combinational features and multimodality MRI images.
METHODS: Fifty-one glioma patients who underwent radiation treatments after surgery were enrolled in this study. Sixteen patients revealed radiation necrosis while 35 patients showed tumor recurrence during the follow-up period. After treatment, all patients underwent T1-weighted, T1-weighted postcontrast, T2-weighted, and fluid-attenuated inversion recovery scans. A total of 41,284 handcrafted and 24,576 deep features were extracted for each patient. The 0.623 + bootstrap method and the area under the curve (denoted as 0.632 + bootstrap AUC) metric were used to select the features. The stepwise forward method was applied to construct 10 logistic regression models based on different combinations of image features.
RESULTS: For handcrafted features on multimodality MRI, model 7 with seven features yielded the highest AUC of 0.9624, sensitivity of 0.8497, and specificity of 0.9083 in the validation set. These values were higher than the accuracy of using handcrafted features on single-modality MRI (paired t-test, p < 0.05, except sensitivity). For combined handcrafted and AlexNet features on multimodality MRI, model 6 with six features achieved the highest AUC of 0.9982, sensitivity of 0.9941, and specificity of 0.9755 in the validation set. These values were higher than the accuracy of using handcrafted features on multimodality MRI (paired t-test, p < 0.05).
CONCLUSIONS: Handcrafted and deep features extracted from multimodality MRI images reflecting the heterogeneity of gliomas can provide useful information for glioma necrosis/recurrence classification.
Copyright © 2019 Quan Zhang et al.

Entities:  

Mesh:

Year:  2019        PMID: 31871484      PMCID: PMC6913337          DOI: 10.1155/2019/2893043

Source DB:  PubMed          Journal:  Comput Math Methods Med        ISSN: 1748-670X            Impact factor:   2.238


1. Introduction

Gliomas are the most common and aggressive brain tumors in adults and have an approximate 5-year survival rate of 10% in their highest grade (e.g., glioblastoma multiforme) [1]. The conventional therapy for gliomas is surgery followed by conventional radiotherapy/chemotherapy [1, 2]. However, this combinatory therapy usually leads to radiation necrosis, which is the most common side effect in gliomas within 2 years after treatment [3, 4]. Unfortunately, the period of occurrence of radiation necrosis is also the peak period of glioma recurrence [4]. Clinically, the methods used to distinguish between glioma recurrence and necrosis are follow-up, biopsy, and surgical resection [5]. Given that the treatment protocols of glioma necrosis and recurrence are quite different [6, 7], finding a fast and noninvasive way to differentiate glioma necrosis from recurrence is important. Radiomics [8] is widely used as a noninvasive method to classify lesions into recurrence or necrosis [9, 10]. In the current radiomics models, good classification results are achieved by using handcrafted features (e.g., intensity and texture features). However, handcrafted features are shallow and low-ordered; as such, they may not fully characterize tumor heterogeneity and, in fact, could limit the potential of the radiomics model applied [11]. To solve this problem, a number of studies have proposed the use of deep features [12-15]. In these studies, improvements in performance were observed by incorporating deep features into the radiomics model of interest. Thus, exploiting potential tumor heterogeneity by using deep features is expected to provide a new and effective point of view from which to improve glioma necrosis and recurrence classification. Frequent monitoring is required for cancer patients; among the imaging methods available, MRI is consistently the preferred technique [16]. Different MRI modalities, such as magnetic resonance spectroscopy [17, 18], T1-weighted postcontrast (T1C) imaging [19], and diffusion-weighted imaging [20], are used to differentiate glioma necrosis from recurrence. However, most previous studies employ image information from single-modality MRI. Moreover, during follow-up, the most commonly used scans for glioma patients are T1-weighted (T1), T1C, T2-weighted (T2), and fluid-attenuated inversion recovery (FLAIR) images. Single-modality MRI provides partial information, whereas multimodality MRI comprehensively characterizes tissues [21, 22]. Therefore, combining different MRI modalities can enhance the tumor discriminatory power of the technology and reveal the degree of tumor infiltration [21, 23, 24]. Figure 1 shows that the heterogeneities of glioma recurrence and necrosis in different MRI modalities are different.
Figure 1

MRI diagnostic images of two patients with glioma. (a–d) Recurrent image of a 52-year-old patient with astrocytoma 1 year after radiotherapy. (e–h) Necrotic image of a 54-year-old patient with oligodendroglioma 6 months after radiotherapy. (a), (b), (c), and (d) and (e), (f), (g), and (h), respectively, show T1C, T1, T2, and FLAIR images. The inside of the red line shows the edge of the lesion.

In this study, we proposed a novel radiomics model for distinguishing necrosis versus glioma recurrence. The contributions of our study were as follows. First, multimodality MRI images were used in this research. Different MRI modalities could reveal different parts of the tumor area [22]. Therefore, the accuracy of glioma identification could be improved by using multimodality MRI images [25]. Second, deep features were combined with handcrafted features in this study to classify glioma necrosis versus recurrence. The powerful ability of deep features has been verified in previous studies [13-15]. Moreover, to the best of our knowledge, previous studies have not combined multimodality MRI images and deep features for classifying glioma necrosis versus recurrence. Therefore, the proposed method might be a valuable tool for distinguishing glioma necrosis from recurrence.

2. Materials and Methods

2.1. Study Population and MRI Images

This retrospective study was supported by the ethics committee of hospital and written informed consent was waived. In this study, the diagnosis of glioma recurrence and necrosis was confirmed by two neuroradiologists with work experiences of over 9 and 20 years. Patients were included on the basis of the following criteria: (1) pathologically confirmed that glioma recurrence or necrosis occurred after radiotherapy; (2) all glioma patients' recurrence or necrosis after radiotherapy was confirmed by imaging and clinical follow-up (follow-up time > 6 months); (3) all MRI images (T1, T1C, T2, and FLAIR images) of glioma necrosis and recurrence used must be confirmed at a follow-up of no less than 6 months after radiotherapy; and (4) glioma patients without pathologic diagnosis excluded the possibility of pseudoprogression based on follow-up. If the follow-up time was not enough, it was difficult for the neuroradiologist to judge whether the patient has recurrence or pseudoprogression. Such glioma patients were not accepted. The exclusion criteria were as follows: (1) patients with recurrent glioma without radiotherapy; (2) follow-up time of less than 6 months for glioma patients; and (3) glioma patients without four modalities of MRI images. A total of 51 patients (16 necrosis and 35 recurrences) were enrolled in this study. The clinical characteristics of all patients are summarized in Table 1.
Table 1

Clinical characteristics of glioma patients.

CharacteristicTypeValue
SexMale24 (47%)
Female29 (53%)
AgeMean47.6 (10–74)
HistologyGlioblastoma12 (23.5%)
Astrocytoma14 (27.5%)
Ependymoma3 (5.9%)
Mixed glioma22 (43.1%)
GradeHigh (III-IV)32 (62.7%)
Low (I-II)19 (37.3)
Recurrence or necrosisRecurrence35 (68.6%)
Necrosis16 (31.4%)
Time intervalMean1.8 years
Tumor locationFrontal lobe21 (41.2%)
Temporal lobe22 (43.1%)
Cerebellum2 (3.9%)
Occipital lobe3 (5.9%)
Parietal lobe3 (5.9%)

Time interval refers to the time point from first radiotherapy to diagnosis of necrosis or recurrence. The grade corresponds to the pathological outcome of patients' first surgery.

MRI images were obtained by using 3.0 T MRI machines (Philips, Achieva). The MRI protocols for the four modalities are listed in Table 2. All MRI scans were obtained in the axial plane. Figure 1 illustrates an example of the four modalities of MRI images of glioma recurrence and necrosis.
Table 2

MRI protocols for four MRI modalities.

ImageSlice thickness (mm)TR (ms)TE (ms)FAMatrixAcquisition time (s)
T2630008090°376 × 26972
T1620002090°284 × 184102
T1C62504.680°332 × 24679.5
FLAIR61100012590°288 × 14988

2.2. Overview of the Proposed Method

The overall framework of the proposed method is shown in Figure 2. The method consists of three fundamental steps: (1) an image preprocessing step that obtains tumor regions, (2) a feature extraction step that extracts handcrafted and deep features, and (3) an analysis step that combines univariate and multivariate analyses. This combination allows the selection of features and construction of a prediction model for glioma necrosis and recurrence classification.
Figure 2

Overall framework of the proposed method.

2.2.1. Image Preprocessing

We used the linear registration function in FSL5.0.9 (http://fsl.fmrib.ox.ac.uk) to register T1, T2, and FLAIR images to T1C images and then applied ITK-SNAP software (http://www.itk-snap.org) to manually segment the tumor region for each patient. All manual segmentations of the tumor region were drawn slice-by-slice by a neuroradiologist with over 9 years of experience in neuroradiology. To avoid mistakes, another senior neuroradiologist with 20 years of experience in brain tumor diagnosis confirmed the final tumor region. All segmentations were drawn on T1C images, covered the entire tumor (avoid cystic changes, edema, and blood vessels), and used to extract handcrafted and deep features.

2.2.2. Feature Extraction

The methodology used to extract handcrafted features from the tumor region and texture extraction parameters is described in the Supplementary Information. A total of 4 nontexture and 41,280 texture parameter features (10,320 features from each image of each MRI modality) were extracted for each patient. We selected AlexNet [26] and Inception v3 [27], which were pretrained on approximately 1.2 million images from the ImageNet Dataset. Features were obtained by forward propagating an MRI slice through the network and extracting the deep features. The architectures of both networks are illustrated in the Supplementary Information. In our glioma dataset, the number of glioma lesion slices and the size of the glioma lesion area in each slice were different for different patients. If we used all slices of a glioma lesion for each patient, the feature dimension would be varied across different patients. Therefore, a 3D bounding box was used to extract the tumor region and then to extract the axial slice with the largest tumor area from this box. We also extracted the front and back slices of the extracted slice. Finally, the three axial slices were combined as an RGB volume. AlexNet and Inception v3 network were a continuous convolution and pooling operation on the input images. Compared with the shallower layer of the two networks, more heterogeneous information of tumors can be extracted in the deeper layer. Moreover, the penultimate layer of the deep network had been used to extract deep features in some recent studies [15, 28, 29], and good performance could be achieved, which indicated the effectiveness of the deep features extracted from the deeper layers. Therefore, for AlexNet, the RGB volume was converted into a size of 227 × 227 × 3 as the input, and the penultimate layer (FC7) was used as the output. A total of 16,384 AlexNet features (4,096 features from each image of each MRI modality) were found. For Inception v3, the RGB volume was adjusted to a size of 299 × 299 × 3 as the final input, and the average pooling layer was used as the output. Therefore, a total of 8,192 Inception v3 features (2048 features from each image of each MRI modality) were drawn.

2.2.3. Univariate and Multivariable Analyses

Univariate associations between feature sets (4 nontexture features, 41,280 texture features, 16,384 AlexNet features, and 8,192 Inception v3 features) and glioma necrosis or recurrence were assessed using Spearman's rank correlation (r). Given the existence of multiple comparisons, Bonferroni correction was also applied. The significance level was set to p=α/K, where K is the number of comparisons and α is the significance level set to 0.05. In multivariate analysis, our goal is to find a linear combination of interesting features so that whether the output is necrosis or recurrence for the new input data could be properly judged. Therefore, the prediction model was constructed by using a logistic regression model:where x is the j th input variable (image features) of i th patient x and a={a ∈ ℝ : j=1,2,…, p} is the regression coefficient of the model for a total of N patients. We employed the 0.623 + bootstrap method and the area under the curve (denoted as 0.632 + bootstrap AUC) metric to estimate which model learned from our dataset could best classify glioma recurrence and necrosis on a new sample. Before presenting the estimation method, we provide a brief symbol introduction. Let our image dataset be denoted as x={x, i=1,2,…, N}. We construct a bootstrap sample x={x : i=1,2,…, N} with N patients randomly drawn with replacement from x. An original sample that does not appear in the bootstrap sample was defined as x(0). The generation of a large number B (B = 1000) of randomly drawn bootstrap samples x (b = 1, 2,…, B) is used to estimate a statistical quantity of interest on the unknown true population distribution. Note that the probability of selecting a positive (necrosis group class) is made equal to the probability of selecting a negative (recurrence group class) each time by drawing x from x; this approach is called “imbalance-adjusted bootstrap resampling.” Prediction models were constructed for three different types of initial feature sets: (1) four sets of single-modality handcrafted features (T1, T1C, T2, and FLAIR), including 10,320 texture and 4 nontexture features for each modality; (2) multimodality handcrafted features, including 41,280 texture and 4 nontexture features; and (3) two sets of multimodality deep features, including 16,384 AlexNet features and 8,192 Inception v3 features. However, the number of feature sets being too large, many redundant and irrelevant features would cause overfitting. Therefore, feature reduction was applied to reduce feature dimensionality. We then used the 0.632 + bootstrap method AUC metric to select features to construct different orders of regression models. Finally, we selected the optimal model from the constructed models for classification. We provide details of each step in the following sections. (1) Feature Reduction. The feature set reduction was performed by a stepwise forward feature selection scheme in order to create reduced feature sets containing 25 different features from larger initial sets, a procedure carried out using the gain equation: where and . In equation (2), r(x, y) is Spearman's rank correlation coefficient between each feature j and the output vector y={y ∈ {0 : recurrence, and 1 : necrosis} : i=1,2,…, N}, MIC(x, x) is the maximal information coefficient between features c and j [30], d is the selected feature for the reduced feature set, and D is the feature that has not been removed from the initial feature set. Parameters γ, δ, and δ are set to 0.5, 0.5, and 0, respectively. During each iteration, a new feature with the largest gain value is selected for the reduced feature set, after which a new gain can be calculated based on the features reserved from the initial feature set using imbalance-adjusted bootstrap resampling (B = 1,000). The final selected 25 features are arranged in descending order according to their gain value because the gain equation uses rs varying over the whole feature set. (2) Feature Selection. After feature reduction, we obtained 25 features for each of the three initial feature sets. We then combined the deep features obtained from AlexNet and Inception v3 features with the handcrafted features from multimodality MRI images to obtain 50 fusion features (denoted as fusion AlexNet and fusion Inception v3 features). Using the reduced feature sets, stepwise forward feature selection was performed by maximizing the 0.632 + bootstrap AUC. The order of the regression model was set from 1 to 10, where the order value is the number of features to be selected. For a given model order, 25 independent experiments were conducted for each of the three initial feature sets and 50 independent experiments were performed for each of the two fusion feature sets. In each independent experiment, different features from the reduced set were assigned as a different “starter.” For each given starter, 1,000 logistic regression models were first created for the remaining features by using imbalance-adjusted bootstrap resampling. Then, the single feature maximizing the 0.632 + bootstrap AUC, defined in equation (3), was selected. This process was repeated up to order 10, after which the combination of features yielding the highest 0.632 + bootstrap AUC for each model was identified. To clarify, we used 1-order regression model as an example. For the 25 selected features, each of them could be utilized to construct a 1-order regression model, and thus, there were 25 1-order regression models. For a given 1-order regression model, we first resampled the sample 1000 times to get 1000 pairs of training sets and validation sets. Then, 1000 experiments were carried out and the average of AUC of 1000 experiments was obtained for the model. Therefore, for a 1-order regression model, we got 25 averaged AUC values. For the 2–10 order features, the same process was performed as above. Finally, we selected the order corresponding to the maximum averaged AUC value as the final feature combination for the classification of glioma necrosis versus recurrence.where (3) Classification Model Construction. After feature selection, the optimal combination of features was obtained for models of different orders. Imbalance-adjusted bootstrap resampling was performed for all models once more, and the 0.632 + bootstrap AUC of these models was calculated to select the optimal model. To construct the final prediction model, the coefficients of the optimal combination of features were calculated as follows:where a(x, y) is the coefficient of feature j and can be calculated by solving the logistic regression model in equation (1), p is the model order, and j = 0 refers to the offset of model y(x). Using the calculated , the output of the optimal model y(x) could be obtained by using equation (1), and the final prediction score could be defined as We employed equation (6) to convert the response of the model into a probabilistic form of output.

2.3. Implementation

In this study, all experiments were implemented on a standard personal computer with a single-thread Intel (R) Xeon (R) E5-2667 v4 3.2 GHz processor. The MATLAB 2017b packages used to analyze the radiomics data were available at https://cn.mathworks.com/matlabcentral/fileexchange/51948-radiomics. AlexNet and Inception v3 were pretrained and included in MATLAB 2017b. Since our dataset was relatively small, the 0.632 + bootstrap resampling method has been used. The principle of this method was to resample all data. For a sample, the selected probability was assumed to be n, and the not being selected probability was 1−n. For the whole data, if we resampled n times, the not being selected probability of a sample was (1 − (1/n)). When n was large enough, lim(1 − (1/n)) ≈ 1/e=0.368 could be obtained, and thus, the selected probability of a sample was 1−0.368 = 0.632. In our experiments, we resampled the original dataset 1000 times and obtained 1000 training sets, so some samples of the original dataset might have appeared multiple times in a training set, and those did not appear eventually formed a validation set. Therefore, after resampling, 1000 pairs of training sets and validation sets could be generated. The number of samples in a training set was 51 × 0.632 ≈ 32, and the number of samples in a validation set was 51 × 0.368 ≈ 19. Classification performance was first evaluated in the training set and then confirmed in the validation set.

3. Univariate and Multivariable Results

The rs values between the features and glioma recurrence versus necrosis, along with the corresponding p values, are listed in Table 3. In the table, only nontexture features and a portion of the texture and deep features are listed. Other detailed rs results of handcrafted features are provided in the Supplementary Information. In Table 3, for handcrafted features, such as gray-level run-length matrix (GLRLM)-HGRE and gray-level size zone matrix (GLSZM)-(HGZE, SZLGE, and SZHGE), extracted from the T1C, T2, and FLAIR images, have a slightly higher correlation with glioma recurrence versus necrosis. For deep features, a certain Spearman's rank correlation with glioma necrosis versus recurrence is observed.
Table 3

r between features (portion of the handcrafted and deep features) and glioma recurrence versus necrosis (p=α/K, α = 0.05, and K = 176, 4,096, and 2,048 for handcrafted, AlexNet, and Inception v3 features, respectively).

TypeFeatureModality r s p value
NontextureVolumeT10.03730.7949
T2
FLAIR
T1C
SizeT10.01720.9045
T2
FLAIR
T1C
SolidityT10.01150.9363
T2
FLAIR
T1C
EccentricityT1−0.01720.9045
T2
FLAIR
T1C
GLRLMHGRET10.32730.0190
T2−0.33310.0169
FLAIR−0.31870.0226
T1C0.45940.0007
GLSZMHGZET10.37900.0061
T2−0.45080.0009
FLAIR−0.48520.0003
T1C0.47380.0004
SZLGET10.38760.0049
T2−0.37900.0061
FLAIR−0.46520.0006
T1C−0.39620.0040
SZHGET10.41630.0024
T2−0.34460.0133
FLAIR−0.47380.0004
T1C0.36180.0091
AlexNetF7_618T1C0.5656 0.00001
F7_1394T10.51680.0001
F7_2793FLAIR0.48230.0003
F7_3501T20.44210.0012
Inception v3avg_pool_663T10.5770 0.0000093
avg_pool__469T1C0.54830.000031
avg_pool_827FLAIR0.38760.005
avg_pool_774T20.46510.000584

For deep feature names, the first character indicates the layer of the CNN and the second character represents the neuron. For example, F7_618 was extracted from a T1C image and taken from the 618th neuron of fully connected layer 7.

Figure 3 shows the prediction performance of the proposed method for the estimation of multivariable models with optimal feature combinations, which were obtained for each model order of the three original feature sets (including T1, T1C, T2, FLAIR, multimodality, AlexNet, and Inception v3 feature sets) and the two fusion feature sets (including fusion AlexNet and fusion Inception v3 feature sets). In Figure 3, the multimodality handcrafted features yielded the highest AUC of 0.9624, sensitivity of 0.8497, and specificity of 0.9083 in model 7 of validation set compared to the single-modality handcrafted features (paired t-test, p < 0.05, except sensitivity). Model 6 with six features (four handcrafted and two AlexNet features) yielded the highest AUC of 0.9982 in the validation set. Details of the classification accuracy of the optimal model on the training and validation sets are shown in Table 4. The selected features and response map of the optimal model for each feature set are given in the Supplemental Information.
Figure 3

Estimation of the classification performance of multivariable models constructed from T1C, T2, T1, FLAIR, multimodality, AlexNet, Inception v3, fusion AlexNet, and fusion Inception v3 images using optimal features in the training set (a) and validation set (b) for the model orders 1–10. The optimal degrees of freedom were separately found in terms of the maximum 0.632 + bootstrap AUC for each model order. Error bars represent the standard error of the mean at the 95% confidence interval.

Table 4

Mean ± standard deviations of the evaluation metrics with different features in the training and validation sets. The results of deep features from each column are shown in bold. The p values of paired t-tests among different features in the validation set are listed in the lower half of the table. Calculations of the sensitivity and specificity of handcrafted and deep feature sets are provided in the Supplementary Information.

Training setValidation set
TypeAUCSeSpAccAUCSeSpAcc
FLAIR0.9429 ± 0.00370.7936 ± 0.01290.8738 ± 0.00440.8598 ± 0.00360.9271 ± 0.00470.7826 ± 0.01570.8421 ± 0.00620.8304 ± 0.0052
T1C0.8980 ± 0.00530.6912 ± 0.01170.8455 ± 0.00530.8094 ± 0.00430.8771 ± 0.00650.7153 ± 0.01570.8032 ± 0.00720.7854 ± 0.006
T10.9783 ± 0.00170.8687 ± 0.01030.9284 ± 0.00340.9179 ± 0.00290.9696 ± 0.00240.8529 ± 0.01080.9077 ± 0.0050.8960 ± 0.0043
T20.9182 ± 0.00380.8109 ± 0.01150.8290 ± 0.00440.8264 ± 0.00370.8994 ± 0.00490.8019 ± 0.01440.7905 ± 0.00610.7909 ± 0.0051
Multimodality0.9722 ± 0.00290.8849 ± 0.01090.9190 ± 0.00350.9172 ± 0.00330.9624 ± 0.00380.8497 ± 0.01330.9083 ± 0.0052.8960 ± 0.0047
AlexNet0.9995 ± 0.00020.9996 ± 0.00040.9870 ± 0.00150.9892 ± 0.0012 0.9993 ± 0.0003 0.9994 ± 0.0006 0.9801 ± 0.0022 0.9833 ± 0.0018
Inception v30.9941 ± 0.00120.9913 ± 0.00340.9615 ± 0.00390.9669 ± 0.0033 0.9914 ± 0.0017 0.9884 ± 0.0042 0.9436 ± 0.0054 0.9509 ± 0.0047
Fusion AlexNet0.9988 ± 0.00050.9957 ± 0.00210.9838 ± 0.0020.9860 ± 0.0017 0.9982 ± 0.0007 0.9941 ± 0.0028 0.9755 ± 0.0029 0.9786 ± 0.0025
Fusion Inception v30.9992 ± 0.00040.9933 ± 0.00250.9863 ± 0.00190.9874 ± 0.0017 0.9988 ± 0.0006 0.9907 ± 0.0034 0.9793 ± 0.0028 0.9809 ± 0.0025
Single-modality handcrafted features compared to multimodality handcrafted features (p values)
T15.35 × 10−391.71 × 10−271.56 × 10−286.93 × 10−59
T23.57 × 10−160.18326.0 × 10−261.22 × 10−22
T1C3.10 × 10−230.029.71 × 10−171.68 × 10−22
FLAIR0.030.48630.020.0099
Deep features compared to multimodality handcrafted features (p values)
AlexNet3.88 × 10−1386.40 × 10−1178.37 × 10−2087.99 × 10−301
Inception v33.60 × 10−991.72 × 10−1033.97 × 10−981.32 × 10−162
Fusion AlexNet1.56 × 10−1345.76 × 10−1091.47 × 10−1935.44 × 10−277
Fusion Inception v32.81 × 10−1351.02 × 10−1022.94 × 10−1901.49 × 10−269
AlexNet features compared to Inception v3 and fusion AlexNet features, respectively (p values)
Inception v31.35 × 10−188.17 × 10−88.21 × 10−312.41 × 10−33
Fusion AlexNet0.011.11 × 10−40.080.02
Fusion Inception v3 features compared to Inception v3 and fusion AlexNet features (p values)
Inception v32.18 × 10−140.881.81 × 10−252.09 × 10−23
Fusion AlexNet0.89130.030.420.9663

Se: sensitivity; Sp: specificity; Acc: accuracy. “—” in the table indicates the item was not calculated to correspond to paired t-test value in the training set.

4. Discussion

In this study, we proposed a novel radiomics model that is expected to support the classification of glioma recurrence versus necrosis. In the proposed method, MRI images of four modalities (i.e., T1, T2, T1C, and FLAIR images) are used to extract handcrafted and deep features. Fifty-one cases of glioma necrosis and recurrence were applied to validate the proposed method. More importantly, we have obtained the highest classification accuracy on the validation set by using fusion AlexNet features from the perspective of classification accuracy and interpretability of features. Therefore, the proposed method might be a valuable tool for distinguishing glioma necrosis from recurrence. We employed radiomics to distinguish glioma necrosis from recurrence. We first evaluated the performance of handcrafted features extracted from multimodality and single-modality MRI images. Table 4 shows that the classification accuracy of multimodality MRI is higher than that of single-modality MRI (paired t-test p < 0.05, except for AUC and sensitivity in T1 modality), which reveals the usefulness of employing different MRI modalities for the current task. However, the extraction methods for handcrafted features are similar for different types of lesions [31-36] and, thus, may limit the potential of radiomics [11]. In order to better reflect the heterogeneity of tumors and improve the performance of radiomics, some scholars have proposed the use of deep features. Antropova et al. [13] used the VGG19 model trained on natural images to extract the deep features of breast cancer and combined it with handcrafted features for the classification of breast cancer, and the results were greatly improved compared with handcrafted features; Decuyper et al. [12] used the trained VGG11 network to extract deep features for different inputs (ROI or the whole image) and classify the grade of glioma into two categories. In addition, Oikonomou et al. [8] show that the combination of handcrafted and deep features leads to the highest performance in lung cancer survival prediction using Random Forest and Naive Bayes classifiers. This successful application of deep features [12-14] confirms the validity of using deep features in our study. We employed two CNNs (AlexNet and Inception v3) to extract deep features from multimodality MRI images to evaluate the effectiveness of these images for the classification work. In this study, we only extracted features from a given layer of a CNN rather than fine-tuning or training from scratch. Doing so can save computational time [13] and avoid the difficulty in designing CNN. Table 4 reveals that the classification accuracy of using AlexNet and Inception v3 features is higher than that of employing handcrafted features (paired t-test p < 0.0001). This finding further illustrates the usefulness of deep features in the classification of glioma necrosis versus recurrence. Table 4 also demonstrates that the classification accuracy of using AlexNet features is higher than that of using Inception v3 features (paired t-test p < 0.0001). The high performance of AlexNet may be due to its simple structure. Complex networks are designed for a specific task; thus, the generalized performance of a complex network is poorer than that of a simple network [13, 37]. Regardless of the performance of deep features, they are less interpretable than handcrafted features, which include tumor shape, volume, texture, and other descriptive features. Therefore, combining deep and handcrafted features provides more information on the object of interest. Finally, we built a six-order model based on the combination of AlexNet and multimodality handcrafted features, ultimately obtaining an AUC of 0.9982, a sensitivity of 0.9941, and a specificity of 0.9755. Table 5 shows that the classification results of the proposed method outperform the results of recently published papers. Here, only the results of different methods in the literature are listed. Direct comparison of the performances of different methods is unreasonable because various datasets and methods for extracting features and building classifiers are used among studies. Nonetheless, the proposed method shows the highest AUC, specificity, and accuracy among the methods surveyed for the classification problem, thereby implying its advantage for classifying glioma recurrence and necrosis.
Table 5

Comparison of the classifying results of glioma necrosis versus recurrence.

YearTypeRecurrence/necrosisAUCSeSpAcc
Tsuyuguchi et al. [38]2004PET6/5 1.00 0.60.82
Ozsunar et al. [39]2010PET/MRI (DSCE-CBV, and ASL)28/70.94
Rani et al. [1]2018SPECT/MRI (T1, T2, FLAIR, and DWI)18/100.920.92
Takenaka et al. [40]2014PET34/160.9250.9120.875
Jena et al. [41]2016PET/MRI (FLAIR, T2, DWI, MRS, and EPI)19/70.97
Jena et al. [42]2017PET/MRI (T1, T2, FLAIR, DWI, PWI/EPI, and MRS)25/100.935
Fusion AlexNetMRI (T1, T2, T1C, and FLAIR)35/16 0.9982 0.9941 0.9755 0.9786

Se: sensitivity; Sp: specificity; Acc: accuracy; DSCE-CBV: dynamic susceptibility contrast-enhanced cerebral blood volume; ASL: arterial spin-labeled; DWI: diffusion-weighted imaging; EPI: perfusion EPI; PWI: perfusion-weighted imaging; MRS: magnetic resonance spectroscopy.

The proposed method presents certain limitations. First, the correlations among features were ignored. Despite finding that these correlations could contribute to the model, high-dimensional features were, in general, difficult to handle. Second, there were tens of thousands of features. However, these high-dimensional features were reduced by a stepwise forward feature selection scheme that reduces each of the initial feature sets to only 25 different features through the gain equation (2). Third, the dataset used in this study was relatively small, which was also the problem identified in previous studies [1, 38–42]. Therefore, a large patient cohort was necessary to create a more robust model. In this study, the bootstrap resampling method was used due to the small size of the dataset. We first resampled the whole sample 1000 times and then got 1000 training and validation sets, respectively. For each run of the 1000 experiments, we measured the AUC, sensitivity, and specificity for the training and the validation sets, respectively. Results show that all measurements of the training and the validation sets for each experiment were above 0.8 after combining the deep and the handcrafted features. In order to intuitively show the classification results of glioma necrosis versus recurrence, we used six-order AlexNet deep features and handcrafted features to conduct 1000 experiments as an example. Figure 4 shows the AUC values of the classification of glioma necrosis versus recurrence. The x-axis represented the number of the experiments, and the y-axis was the AUC values of the training and validation sets, respectively, measured by each experiment. It can be seen from Figure 4 that the variation tendency of the AUC of the validation sets was the same as that of the training sets, and the difference between the AUC values of the training and validation sets was very small, which indicated that overfitting on the proposed method was alleviated in this study.
Figure 4

The AUC values of 1000 pairs of the training and validation sets in the classification of glioma necrosis versus recurrence. The x-axis represented the number of the experiments and the y-axis was the AUC values of the training and validation sets, respectively, measured by each experiment.

In conclusion, finding a noninvasive and accurate method to classify glioma recurrence versus necrosis is clinically significant. In this study, we explored a novel method by combining deep and handcrafted features extracted from multimodality MRI images to improve the classification accuracy of glioma recurrence versus necrosis. Classification models based on objective and quantitative handcrafted and deep features can be useful for precision medicine and improve the treatment strategies used for glioma necrosis and recurrence.
  31 in total

Review 1.  Malignant glioma: ESMO clinical recommendations for diagnosis, treatment and follow-up.

Authors:  R Stupp; F Roila
Journal:  Ann Oncol       Date:  2009-05       Impact factor: 32.976

2.  MR spectroscopy for differentiation of recurrent glioma from radiation-induced changes.

Authors:  Ajit H Goenka; Atin Kumar; Raju Sharma
Journal:  AJR Am J Roentgenol       Date:  2009-12       Impact factor: 3.959

3.  Brain tumor segmentation based on local independent projection-based classification.

Authors:  Meiyan Huang; Wei Yang; Yao Wu; Jun Jiang; Wufan Chen; Qianjin Feng
Journal:  IEEE Trans Biomed Eng       Date:  2014-05-19       Impact factor: 4.538

4.  Differentiation of residual/recurrent gliomas from postradiation necrosis with arterial spin labeling and diffusion tensor magnetic resonance imaging-derived metrics.

Authors:  Ahmed Abdel Khalek Abdel Razek; Lamiaa El-Serougy; Mohamed Abdelsalam; Gada Gaballa; Mona Talaat
Journal:  Neuroradiology       Date:  2017-12-07       Impact factor: 2.804

5.  Differentiation of Recurrent/Residual Glioma From Radiation Necrosis Using Semi Quantitative 99mTc MDM (Bis-Methionine-DTPA) Brain SPECT/CT and Dynamic Susceptibility Contrast-Enhanced MR Perfusion: A Comparative Study.

Authors:  Nisha Rani; Baljinder Singh; Narendra Kumar; Paramjit Singh; Puja Panwar Hazari; Harpreet Singh; Ganesh Kumar; Bishan Radotra; Mukesh Kumar; Anish Bhattacharya; Rajnish Sharma; Adarsh Kohli; Rajesh Chhabra; Sunil Kumar Gupta; Anil Kumar Mishra
Journal:  Clin Nucl Med       Date:  2018-03       Impact factor: 7.794

6.  A deep feature fusion methodology for breast cancer diagnosis demonstrated on three imaging modality datasets.

Authors:  Natalia Antropova; Benjamin Q Huynh; Maryellen L Giger
Journal:  Med Phys       Date:  2017-08-12       Impact factor: 4.071

7.  Treatment of radiation-induced nervous system injury with heparin and warfarin.

Authors:  M J Glantz; P C Burger; A H Friedman; R A Radtke; E W Massey; S C Schold
Journal:  Neurology       Date:  1994-11       Impact factor: 9.910

8.  Multiparametric tissue characterization of brain neoplasms and their recurrence using pattern classification of MR images.

Authors:  Ragini Verma; Evangelia I Zacharaki; Yangming Ou; Hongmin Cai; Sanjeev Chawla; Seung-Koo Lee; Elias R Melhem; Ronald Wolf; Christos Davatzikos
Journal:  Acad Radiol       Date:  2008-08       Impact factor: 3.173

9.  Hyperbaric oxygen therapy for radiation-induced brain injury in children.

Authors:  P J Chuba; P Aronin; K Bhambhani; M Eichenhorn; L Zamarano; P Cianci; M Muhlbauer; A T Porter; J Fontanesi
Journal:  Cancer       Date:  1997-11-15       Impact factor: 6.860

10.  Glioma recurrence versus radiation necrosis? A pilot comparison of arterial spin-labeled, dynamic susceptibility contrast enhanced MRI, and FDG-PET imaging.

Authors:  Yelda Ozsunar; Mark E Mullins; Kenneth Kwong; Fred H Hochberg; Christine Ament; Pamela W Schaefer; R Gilberto Gonzalez; Michael H Lev
Journal:  Acad Radiol       Date:  2010-01-12       Impact factor: 3.173

View more
  8 in total

Review 1.  Machine learning in neuro-oncology: toward novel development fields.

Authors:  Vincenzo Di Nunno; Mario Fordellone; Giuseppe Minniti; Sofia Asioli; Alfredo Conti; Diego Mazzatenta; Damiano Balestrini; Paolo Chiodini; Raffaele Agati; Caterina Tonon; Alicia Tosoni; Lidia Gatto; Stefania Bartolini; Raffaele Lodi; Enrico Franceschi
Journal:  J Neurooncol       Date:  2022-06-28       Impact factor: 4.506

Review 2.  A Survey of Radiomics in Precision Diagnosis and Treatment of Adult Gliomas.

Authors:  Peng Du; Hongyi Chen; Kun Lv; Daoying Geng
Journal:  J Clin Med       Date:  2022-06-30       Impact factor: 4.964

3.  A deep learning-based radiomics approach to predict head and neck tumor regression for adaptive radiotherapy.

Authors:  Shohei Tanaka; Noriyuki Kadoya; Yuto Sugai; Mariko Umeda; Miyu Ishizawa; Yoshiyuki Katsuta; Kengo Ito; Ken Takeda; Keiichi Jingu
Journal:  Sci Rep       Date:  2022-05-27       Impact factor: 4.996

4.  Differentiation of recurrent glioblastoma from radiation necrosis using diffusion radiomics with machine learning model development and external validation.

Authors:  Yae Won Park; Dongmin Choi; Ji Eun Park; Sung Soo Ahn; Hwiyoung Kim; Jong Hee Chang; Se Hoon Kim; Ho Sung Kim; Seung-Koo Lee
Journal:  Sci Rep       Date:  2021-02-03       Impact factor: 4.379

5.  Application of radiomics feature captured from MRI for prediction of recurrence for glioma patients.

Authors:  Canyu Liu; Yujiao Li; Xiang Xia; Jiazhou Wang; Chaosu Hu
Journal:  J Cancer       Date:  2022-01-04       Impact factor: 4.207

6.  A Radiomics Model for Predicting Early Recurrence in Grade II Gliomas Based on Preoperative Multiparametric Magnetic Resonance Imaging.

Authors:  Zhen-Hua Wang; Xin-Lan Xiao; Zhao-Tao Zhang; Keng He; Feng Hu
Journal:  Front Oncol       Date:  2021-09-02       Impact factor: 6.244

7.  Multimodal MRI-Based Radiomic Nomogram for the Early Differentiation of Recurrence and Pseudoprogression of High-Grade Glioma.

Authors:  Hui Jing; Fan Yang; Kun Peng; Danlei Qin; Yexin He; Guoqiang Yang; Hui Zhang
Journal:  Biomed Res Int       Date:  2022-09-30       Impact factor: 3.246

Review 8.  Machine learning imaging applications in the differentiation of true tumour progression from treatment-related effects in brain tumours: A systematic review and meta-analysis.

Authors:  Abhishta Bhandari; Ravi Marwah; Justin Smith; Duy Nguyen; Asim Bhatti; Chee Peng Lim; Arian Lasocki
Journal:  J Med Imaging Radiat Oncol       Date:  2022-05-22       Impact factor: 1.667

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.