Literature DB >> 33920221

Radiomic Feature Reduction Approach to Predict Breast Cancer by Contrast-Enhanced Spectral Mammography Images.

Raffaella Massafra1, Samantha Bove2, Vito Lorusso3, Albino Biafora4, Maria Colomba Comes1, Vittorio Didonna1, Sergio Diotaiuti5, Annarita Fanizzi1, Annalisa Nardone6, Angelo Nolasco3, Cosmo Maurizio Ressa7, Pasquale Tamborra1, Antonella Terenzio8, Daniele La Forgia9.   

Abstract

Contrast-enhanced spectral mammography (CESM) is an advanced instrument for breast care that is still operator dependent. The aim of this paper is the proposal of an automated system able to discriminate benign and malignant breast lesions based on radiomic analysis. We selected a set of 58 regions of interest (ROIs) extracted from 53 patients referred to Istituto Tumori "Giovanni Paolo II" of Bari (Italy) for the breast cancer screening phase between March 2017 and June 2018. We extracted 464 features of different kinds, such as points and corners of interest, textural and statistical features from both the original ROIs and the ones obtained by a Haar decomposition and a gradient image implementation. The features data had a large dimension that can affect the process and accuracy of cancer classification. Therefore, a classification scheme for dimension reduction was needed. Specifically, a principal component analysis (PCA) dimension reduction technique that includes the calculation of variance proportion for eigenvector selection was used. For the classification method, we trained three different classifiers, that is a random forest, a naïve Bayes and a logistic regression, on each sub-set of principal components (PC) selected by a sequential forward algorithm. Moreover, we focused on the starting features that contributed most to the calculation of the related PCs, which returned the best classification models. The method obtained with the aid of the random forest classifier resulted in the best prediction of benign/malignant ROIs with median values for sensitivity and specificity of 88.37% and 100%, respectively, by using only three PCs. The features that had shown the greatest contribution to the definition of the same were almost all extracted from the LE images. Our system could represent a valid support tool for radiologists for interpreting CESM images.

Entities:  

Keywords:  breast cancer; computer-automated diagnosis (CADx); contrast-enhanced spectral mammography (CESM); feature extraction; feature reduction; principal component analysis (PCA)

Year:  2021        PMID: 33920221      PMCID: PMC8070152          DOI: 10.3390/diagnostics11040684

Source DB:  PubMed          Journal:  Diagnostics (Basel)        ISSN: 2075-4418


1. Introduction

Reducing the breast cancer mortality rate in the population of women worldwide is greatly impacted by early diagnosis using mammography [1,2]. In recent years, new mammography techniques such as contrast-enhanced spectral mammography (CESM) and tomosynthesis have contributed to further improve the performance of mammography, even in less “readable” types of breasts due to their high density [3,4]. Specifically, the CESM represents a unique and very particular methodology in the clinical and scientific field due to the considerable amount of information derived from morphological images and contrast dynamics obtained simultaneously on the same breast and the same region of interest. While CESM has the same type of indications as DCE-MRI due to the radiation dose and non-uniform availability of the territory, CESM is currently used according to the guidelines as an alternative to DCE-MRI in cases where this is not feasible (for absolute contraindications or related) or not available [5]. In MRI the performance of the human reader can be reduced by the presence of marked background parenchymal enhancement (BPE), defined as the normal background impregnation of the breast but with very different intensity and distribution in relation to age, hormonal phase, menopausal state and any therapies in progress [6]. In these circumstances, diagnostic aids to reporting and fusion-imaging systems in different positions can support the diagnosis [7,8,9]. These aspects are also expected in CESM. Nowadays, clinicians are aided in the interpretation of these images by automated support systems, known as computer-aided detection/diagnosis (CAD), which is software used in clinical medicine to suggest diagnoses and treatments based on the clinical data and algorithms for their interpretation. Although various models of CAD systems for breast diseases using mammographic images have been developed in the past years [10,11,12,13,14,15], the literature is limited with regard to CAD systems for breast lesions using CESM images. Most of the recent literature includes comparative studies on the diagnostic performance of human readers on CESM images with respect to those of mammographic and MR images [3,16,17,18]. However, interest in this new instrumentation is strong and some support tools have recently been proposed for characterizing breast lesions [19,20,21,22,23] and predicting the response to neoadjuvant therapy [24,25]. For this reason, we have developed an accurate computer-aided diagnosis tool to classify breast cancer lesions based on radiomic analysis of CESM images. In our previous work, we proposed a preliminary radiomics analysis aimed to explore the usefulness of quantitative information extracted from CESM images, to understand the behavior of each different set of well-known textural features automatically extracted from CESM images, and to compare them with each other [21,23]. In these works, an important role was played by the feature selection processes used to describe and characterize the regions of interest (ROIs) identified by our expert radiologists. In order to reduce the dataset dimensionality, starting from the initial feature set, a sub-set of these features, which were characterized by their high discriminating power, was selected by filtering techniques (i.e., statistical tests) for more manageable data processing [26]. Then, we selected the most important features by developing different approaches to feature selection, such as embedded and wrapper methods. The feature selection techniques can influence the process and accuracy of cancer classification. Indeed, although the subset of features identified was the one with the greatest discriminating power with respect to the various evaluation criteria considered, the feature selection techniques involve a natural and inevitable loss of information due to the exclusion of features from the original dataset. For this reason, before proceeding with the exploration of new features useful for increasing the accuracy of the classification performance, we considered it appropriate to evaluate a different dimensionality reduction technique that does not produce information losses but only a reduction of the noise inherent in the dataset. Therefore, we evaluated a classification scheme that includes dimension reduction. Specifically, firstly, a principal component analysis (PCA) dimension reduction method that includes the calculation of the proportion of variance for eigenvector selection was used. Following, the ROIs classification was performed by training different binary classifiers on a subset identified by sequential supervised learning procedures.

2. Materials and Methods

2.1. Materials

2.1.1. CESM Examination

The CESM is based on a dual-energy exposure after a single injection of an iodinated contrast medium (CM), which produces three images: a low-energy (LE), a high-energy (HE) and a third image, defined as a recombined (RC) image that is obtained thanks to the digital subtraction of the LE image from the HE one. Finally, two images are displayed by the radiologist on the reporting monitor for diagnostic purposes: the LE, which can be superimposed on a standard 2D mammography, which allows morphological surveys, and the RC, which provides information on the tumor neoangiogenesis of the breast. An example of CESM images is shown in Figure 1, which includes a LE image (a), a HE image (b) and their RC image (c).
Figure 1

Images produced by contrast-enhanced spectral mammography (CESM instrumentation). Typical example of low energy (a), high energy (b), and recombined (c) images [17]. The white arrow points out a suspicious lesion.

For all CESM exams, a modified Full-Field Digital Mammography (FFDM) system derived from a standard Senographe Essential (GE Healthcare) was used. First, the breast with no pathology was imaged, then the breast with the suspected lesion. Both craniocaudal (CC) and mediolateral oblique (MLO) views were collected. All of the images obtained were in DICOM format and were evaluated by a dedicated radiologist with more than 10 years of experience in reading mammography and breast MR images and trained in reading contrast-enhanced images.

2.1.2. Experimental Dataset

The study was pre-approved by the Scientific Board of the Istituto Tumori “Giovanni Paolo II” of Bari, Italy. As this is a retrospective study, the anonymized images of patients who had given consent to the use of data for scientific purposes, as required by our Regulations, were acquired. Once we had selected images from 53 patients aged between 37 and 76 years (with a mean of 52.2 ± 10.1 years), showing a positive result according to CESM, a radiologist from our Institute dedicated to senologic diagnostics manually identified the ROIs with a box on the image by using the reporting tool. Some patients had more than one, so a total of 58 ROIs were identified according to the BIRADS classification [27]: lesions belonging to BIRADS 2 and 3 classes were considered as benign, while lesions belonging to BIRADS 4 and 5 classes were labeled as malignant. Then, the histological diagnosis based on bioptic sampling established that 15 ROIs contained benign lesions while 43 ROIs included malignant ones.

2.2. Methods

As summarized in Figure 2, after the radiologist manually segmented a region of interest (ROI) from each LE and RC image, a large feature set consisting of five different kinds was extracted, such as points and corners of interest, textural and statistical features from both the original ROIs and also the ones obtained by a Haar decomposition and a gradient image implementation. Subsequently, principal component analysis (PCA) dimension reduction method that includes the calculation of variance proportion for eigenvector selection was used from each of five features set. Finally, different binary classifiers were trained to discriminate benign and malignant ROIs by developing a sequential forward feature selection algorithm that selected feature sub-sets first individually, and then simultaneously. MATLAB R2017a (MathWorks, Inc., Natick, MA, USA) software was used for all analyses.
Figure 2

Schematic overview of the CESM images classification process. First, five sets of features were automatically extracted from each region of interest (ROI), then a principal component analysis was performed on each feature set. Finally, three binary classifiers were trained and their performances were evaluated on 100 ten-fold cross-validation rounds.

2.2.1. Feature Extraction

Starting from each ROI extracted from both original LE and RC images, five feature sets were automatically extracted to make the lesion classification more objective and operator independent. We then started to mathematically define a digital image [28]. An image can be defined as a two-dimensional function mapping the spatial coordinates and into a value representing the intensity of gray level of the image at that point. When , , and are all finite and discrete quantities, we call the image a digital image. A digital image is composed of a finite number of elements called pixels, each one having a particular position and gray intensity value. The section of the real plane spanned by the coordinates of the image is called the spatial domain. A very common representation for a digital image is a two-dimensional array represented as an numerical matrix. Therefore, given a digital image represented by a matrix, it was possible to extract statistical and textural features and, moreover, to identify points, edges and corners of interest. As shown in Figure 3, in this work we used five different extracted feature subsets, which we named as follows: STAT set, COUNT set, GRAD set, HAAR set and GLCM set [21,23].
Figure 3

Scheme of the feature extraction process.

The STAT set consists of 22 statistical features that describe the distribution of the ROI gray levels, measuring the likelihood of observing a gray value in random positions in the image, without taking into account the spatial information [29]. From each LE and RC original ROI, we extracted the following features: mean, standard deviation and their ratio, variance, skewness, entropy, relative smoothness, kurtosis, minimum and maximum values of gray-level and their difference. The COUNT set contains a total of 10 features that describe the points, edges and corners of interest. In this work, we used five known learning algorithms: Scale invariant feature transform (SIFT) algorithm [30,31], which detects and describes image local features Minimum eigenvalue algorithm, which underlies the Shi-Tomasi corner detection algorithm [32] for identifying the corners of an object Features from the accelerate segment test (FAST) algorithm [33,34], which is another corner detection method Binary robust invariant scalable key-points (BRISK) method [35], which combines the SIFT and the FAST algorithms to feature detection, descriptor composition and key-points matching Maximally stable external regions (MSER) algorithm [36], which is a method of blob detection in images whose aim consists of finding correspondence between image elements from two images with different viewpoints. The GRAD set is formed by 24 features and contains some previously defined statistical features, that is, mean, variance, skewness, entropy, relative smoothness and kurtosis, extracted from the gradient’s magnitude and direction of each LE and RC original ROI. The gradient of an image is represented as a two-component vector (- and -derivative) defined at each pixel [28]. These can be computed by the convolution with a kernel, such as the Sobel or Prewitt operator, since the image is a discrete function for which the derivatives are not defined. For each vector, the magnitude Gmag shows how quickly the intensity of each pixel is changing in the neighborhood of the pixel in the direction of the gradient, while the direction Gdir represents the orientation of greatest intensity change in the neighborhood of the pixel . The gradient can be approximated by convolving a kernel, in this work a Sobel kernel, with the original image [28]. The importance of calculating the gradient image lies in the two pieces of information it provides: the magnitude, which is a measure of how quickly the image is changing and the direction, which illustrates the direction in which the image is changing most rapidly. The HAAR set contains 96 features, the same statistical features previously computed in the GRAD set, but this time extracted from each sub-ROI obtained by decomposing each LE and RC original ROI thanks to the Haar wavelet transform [28,37]. This technique obtains multi-resolution representations of images, which are very effective for analyzing the information content of images due to the dependence of the texture on the scale at which an image is analyzed. Particularly, once Haar two-dimensional scaling and wavelets functions are computed, these can be used as filters in order to decompose the image into four bands. First the image is low-pass filtered and downscaled in order to obtain the low low band, then it is high-pass filtered in the three different directions in order to obtain the three types of detail images: horizontal , vertical and diagonal . The operations can then be repeated on the band using the identical filters, as shown in Figure 4.
Figure 4

Haar decomposition: (a) first level and (b) second level decomposition.

Finally, the GLCM set is formed by a total of 312 features, which represent the spatial relationship that on average links the gray levels of the image to each other. Texture analysis returns textural variables, which are any geometric and repetitive arrangement of gray levels. The textural information of an image can be described by second-order variables computed on the gray-level co-occurrence matrix (GLCM) ( represents the number of intensity values in the gray-scale image, which is usually reduced from 256 to 8 at the beginning of the algorithm to reduce the computational cost) [38,39]. Given a gray-level image f, the first step in building the matrix consists in defining a specific spatial relationship between pixels. This relationship, known as offset, is the distance D between a pixel of interest and its neighbors with respect to a specific direction, identified by an angle . Thus, the element in GLCM matrix represents the number of times the combination of level and occurs in two pixels in the image, which satisfy the relationship given by the offset [28]. Thus, for each sub-ROI previously decomposed by the Haar transform only at first level, the co-occurence matrices are extracted in the four possible directions by choosing the parameter D equal to 2 in order to evaluate the relationship between the gray levels of the pixels immediately close in the image. Finally, the following features are extracted from both LE and RC sub-ROIs: contrast, correlation, cluster prominence, cluster shade, dissimilarity, energy, entropy, homogeneity, sum average, sum variance, sum entropy, difference entropy and normalized inverse difference moment.

2.2.2. Principal Component Analysis

The feature extraction procedures returned five feature sets. The next step consisted of exploring the discriminating power of these sets to identify a sub-set of significant features. To do this, a feature reduction was performed through PCA [40]. The central idea of PCA is to reduce the dimensionality of a dataset consisting of a large number of interrelated variables while retaining as much as possible of the variation present in the data set. This is achieved by performing a linear transformation of the features that projects the original ones into a new Cartesian system, where the variables are sorted in descending order with respect to the overall variance percentage explained. In this work, a PCA was performed for each feature set. The first step consists of representing the data set as a matrix. In particular, if is the number of observations, in this case ROIs, and the number of variables, a matrix is obtained. Then, the raw data in the matrix have to be standardized so that each variable contributes equally to the analysis. Thus, a new matrix is computed. The next step consists of processing the correlation matrix of , a matrix whose elements are the correlation coefficients between the variables. In general, the -th principal component is such that: where is the -th largest eigenvalue of and is the corresponding eigenvector. Now, a sub-set of principal components has to be selected to replace the elements of by a much smaller number of PCs. There are different rules for deciding how many PCs should be retained in order to account for most of the variation in , without serious information loss [40]. In this work, we adopted the explained variance criterion: the required number of PCs is the smallest value of for which the chosen percentage, in this case , is exceeded. In particular, since the principal components are sorted in descending order with respect to their variance, it is sufficient to select the components whose summed variance exceeds the selected threshold value.

2.2.3. Classification Model

Once the significant PCs were selected for each feature set, three different classification models were trained to discriminate the ROIs into benign and malignant, first on each PCs’ sub-set, then on the set containing all the selected PCs. We trained a random forest (RF) classifier [41], a naïve Bayes (NB) classifier [42] and a logistic regression (GLM) [43] in a procedure of PCs stepwise forward selection [44]. The RF algorithm is among the most popular machine learning algorithms because it generally provides good predictive performance combined with low over-fitting. In particular, the tree-based strategy used by RF naturally ranks by how well they improve the purity of the node: nodes with the greatest decrease in impurity are at the start of the trees, while nodes with the least decrease in impurity occur at the end of trees measured by Gini’s diversity index. The NB algorithm is a probabilistic approach based on Bayes’ theorem with an assumption of independence among predictors. In simple terms, a naive Bayes classifier assumes that the presence of a particular feature in a class is unrelated to the presence of any other feature. The GLM is a statistical model that uses a logistic function to model a binary variable. This regression transforms its output using the sigmoid function to return a probability value that can then be mapped to the two classes under question. The stepwise forward selection algorithm identifies the best sub-set of PCs by sequentially adding a PC to the set selected in the previous steps. Starting from a model composed by a single PC that showed the highest median AUC [45] on 100 ten-fold cross-validation rounds, we iteratively added the PC that allowed us to obtain the highest classification performances in terms of median value AUC. The process was repeated until all variables were included into the model. The performance of each classification was evaluated on 100 ten-fold cross-validation rounds in order to obtain the variability of the experimental results. Once the best model was selected for each method and dataset, we compared the classification performances of these prediction models also in terms of: where TP and TN stand for true positive (number of true malignant ROIs identified) and true negative (number of true benign ROIs identified) cases, while FP (number of benign ROIs identified as malignant) and FN (number of malignant ROIs identified as benign) are the false positive and false negative ones, respectively. Specifically, the above values were calculated to identify the optimal threshold by means of Youden’s index on ROC curves [46], an index able to solve dataset unbalance problems (15 benign and 43 malignant).

3. Results

The aim of this study was to devise a prediction model that was successful in discriminating benign and malignant ROIs. First, we reduced the initial dataset dimensionality through a principal component analysis, and we obtained one sub-set of discriminant principal components for each initial feature set. Then, we used three different classification algorithms on both the individual obtained sub-set and on the complete sub-set of significant principal components.

3.1. Principal Component Analysis

We performed a PCA for each standardized set of features, then we adopted the explained variance criterion to select a sub-set of discriminant principal components. We chose a threshold value equal to . The STAT set required 4 PCs, the GRAD set was represented by a set of 9 PCs, the COUNT set was replaced by a set of 3 PCs, the sub-set linked to the HAAR set was formed by a total of 19 PCs, and the GLCM set needed 11 PCs, as shown in Figure 5.
Figure 5

Overview of PC’s cumulative variance for each feature set.

3.2. Classification Performances

Once the PCs’ subsets were extracted, we trained the random forest algorithm, the naïve Bayes method and the logistic regression in the procedure of PCs stepwise forward selection, first on each PCs’ sub-set, then on the set containing all the selected PCs. Concerning the individual PCs’ sub-sets, the best models were obtained by training either the RF or the GLM classifier on sub-sets containing two PCs at most (Table 1).
Table 1

Classification performances of the best models obtained from the individual PCs’ sets calculated on 100 ten-fold cross-validation rounds. For each set of features, the performance measures of the different classification algorithms implemented are reported in correspondence with the best PC combination. with the best PC combination. The third column shows the PCs selected by the sequential forward selection algorithm. The best results are highlighted in bold.

PCs’ SetClassifierPCs Best CombinationAUC (%)Acc (%)Sens (%)Spec (%)
STATRF1 + 278.2977.5981.4073.33
NB181.7174.1467.4493.33
GLM 1 83.49 74.41 67.44 93.33
GRAD RF 1 + 5 85.31 81.03 79.07 93.33
NB1 + 276.5975.8676.7473.33
GLM1 + 5 + 2 + 983.1074.1465.12100
COUNTRF1 + 366.8262.0758.140.8
NB2 + 164.8860.3450.0086.67
GLM 1 + 3 75.66 79.31 88.37 53.33
HAAR RF 2 + 12 94.65 87.93 86.05 100
NB1 + 3 + 16 + 19 + 15 + 1486.5184.4887.2180.00
GLM1 + 3 + 9 + 19 + 16 + 8 + 1283.7277.5974.4293.33
GLCM RF 2 + 1 86.40 81.03 79.07 86.67
NB2 + 4 + 1 + 11 + 10 + 975.5075.8672.0986.67
GLM2 + 4 + 1 + 9 + 11 + 1082.3387.9393.0273.33
Since the PCs were obtained as a linear combination of the starting variables, we were interested in understanding which of these features contributed most and positively to the computation of the PCs, which determined the best classification model for each set. We selected the variables shown in Table 2, excluding the ones characterized by coefficients close to zero compared to the those of the chosen variables.
Table 2

Overview of the features that were important in the computation of the selected PCs on 100 ten-fold cross-validation rounds. The factors, LE and RC identify the features extracted from the LE images and the RC images, respectively.

SetPCImportant Features
STAT1RC_EntropyRC_StdRC_Max-MinRC_RelativeSmoothnessRC_Variance
GRAD1RC_Mean_GmagLE_Mean_GmagRC_Entropy_GmagRC_RelativeSmoothness_GmagRC_Variance_Gmag
5LE_Kurtosis_GmagLE_Skewness_GdirRC_Skewness_GdirRC_Kurtosis_GdirRC_Entropy_Gdir
COUNT1RC_FastRC_BriskLE_BriskLE_Fast
3LE_SiftRC_SiftLE_MSERRC_MinimumEigenvalue
HAAR2LE_RelativeSmoothness_HL2LE_ RelativeSmoothness_LH2LE_ RelativeSmoothness_HL1LE_Entropy_LL1LE_RelativeSmoothness_HH2
12LE_Skewnes_LL1RC_Skewness_LH2LE_Kurtosis_HL1LE_Skewness_HH1
GLCM2RC_SumEntropy_HH1 dir2RC_Entropy_HH1 dir2RC_Entropy_HH1 dir3RC_Entropy_HH1 dir4RC_SumEntropy_HH1 dir4
1LE_SumAverage_HH1 dir3LE_SumAverage_HH1 dir1LE_SumAverage_HH1 dir4LE_SumAverage_HH1 dir2
As regards the STAT set, the best model had a median AUC of 83.49% and was obtained by training the GLM on the sub-set containing only the first PC, which was determined based on the contributions of the variables’ entropy, standard deviation, range, relative smoothness and variance, all computed on the RC images. The best model obtained with the GRAD set’s PCs reached a median AUC of 85.31% and included the RF classifier and the sub-set containing the first and the fifth PCs. These were estimated considering the variables’ mean, entropy, relative smoothness, variance, kurtosis and skewness computed on both the gradient’s magnitude and the gradient’s direction extracted from the RC and the LE images. Concerning the COUNT set, the GLM was the best classifier, achieving a median AUC of 75.66% when trained on the sub-set including the first and the third PCs. The most important features in the computation of these two PCs were the ones calculated on both the LE and the RC images by all five algorithms. As far as the HAAR set, the best model was obtained by training the RF classifier on the sub-set containing the second and the twelfth PCs. This model, with a median AUC of 94.65%, was the best, even compared to the best performing models of each set. The above-mentioned PCs were calculated mainly thanks to the contributions of the variables’ relative smoothness, entropy, skewness and kurtosis, which were all extracted from the LE images, except one extracted from the RC images, after the Haar decomposition. Finally, with regards to the GLCM set, a median AUC of 86.40% was reached with the RF classifier when trained on the sub-set including the first and the second PCs. The features that contributed most to their computation were entropy, sum entropy and sum average; the first were two extracted from the RC images while the third was extracted from the LE images. As regards the models obtained by training the three classifiers on the complete set of PCs previously computed, the best was the one obtained with the RF classifier (Table 3).
Table 3

Classification performance of the best models obtained from the complete set of PCs calculated on 100 ten-fold cross-validation rounds. The best result is highlighted in bold.

ClassifierBest ModelAUC (%)Acc (%)Sens (%)Spec (%)
RF H2 + G1 + H12 95.66 90.52 88.37 100
NBS1 + GL10 + G2 + GL11 + H16 + H3 + H1988.9989.6693.0280
GLMS1 + G2 + G9 + S3 + H8 + GL2 + GL190.0884.4881.40100
Indeed, while the NB classifier reached a median AUC of 88.99% and the GLM achieved a median AUC of 90.08%, the model obtained by training the RF classifier on the set, including the second and the twelfth PCs belonging to the HAAR set and the first PC belonging to the GRAD set, reached a median AUC of 95.66% and a median accuracy of 90.52%. This model is also characterized by a better classification performance than the performance of the best model obtained by considering the single PCs’ sets, and reached a sensitivity of 88.37% and a specificity of 100%.

4. Discussion

Although its use is still not very widespread in the area, CESM is a very interesting technique in our opinion, and due to its intrinsic characteristics, it is particularly suitable for the analysis of images using radiomics. The opportunity to simultaneously analyze similar morphological images on the same mammographic and dynamic images with contrast medium for the evaluation of neoangiogenesis provides the system with a remarkable and varied series of information that can achieve excellent results. Moreover, the absence of a consistent number of studies on this method and on these issues, in our opinion, makes this work even more precious. In this work, we proposed an automated support system able to characterize and discriminate breast lesions as benign/malignant. We extracted 58 ROIs from 53 CESM images and for each ROI we determined five set of features: the STAT set, which included statistical features extracted from the original ROIs, the GRAD and the HAAR set that comprised statistical measures extracted from the ROIs’ manipulations by filters and wavelet functions, respectively, the GLCM set that included textural features and the COUNT set, which comprised information about points and corners of interest. Subsequently, each above-mentioned feature set was replaced by the set containing the related discriminant principal components. In particular, for each set a principal component analysis was performed and a sub-set of principal components was selected by setting a variance explained threshold value of 0.8. Then, all the PCs’ sets obtained were used, first individually and then simultaneously, to train three different classifiers combined with a stepwise forward algorithm. The classification performances were evaluated and compared in terms of accuracy, sensitivity and specificity and AUC values on 100 ten-fold cross-validation rounds. The best model among the ones developed on the single PCs’ sets turned out to be the model obtained by training the RF classifier on the HAAR PCs’ sub-set including the second and the twelfth PCs. It was observed that using only these two variables led to a median AUC value of 94.95%, a median accuracy of 87.93%, a sensitivity of 86.05% and a specificity of 100%. On the other hand, it was observed that adding the first GRAD set’s PC to the previous model led to better results. Indeed, this new model reached a median AUC value of 95.66%, a median accuracy of 90.52%, a sensitivity of 88.37% and a specificity of 100%. In both cases, the specificity value highlights that these prediction models are able to correctly identify all the benign ROIs. Thus, the sequential forward selection algorithm allowed us to obtain a very well performing classification model by selecting only three PCs: the second and the twelfth PCs from the HAAR set and the first PC from the GRAD set. The above-mentioned HAAR PCs were estimated thanks to the contributions of the variables, mainly calculated on LE images, only one of them was extracted from the RC images. On the other hand, the features that contributed most to the calculation of the first GRAD PC were all extracted from the LE images. We compared the performance of the proposed approach with respect to the literature, and in this work, we improved the classification performances obtained with the CAD system developed in our previous work [21] (Table 4). Indeed, by using principal component analysis as a feature selection technique, instead of a backward feature selection algorithm combined with a naïve Bayes classifier, we reduced the number of discriminant variables and at the same time we obtained better results in terms of sensitivity and specificity (in the previous work we obtained a sensitivity of 87.5% and a specificity of 91.7%).
Table 4

Benign vs. malignant breast lesion classification evaluated on CESM images: comparison of the performance results of the proposed models in the literature.

ArticleNo. of ROIsFeaturesClassifierPerformance (%)
Patel et al. [19]50 SVMAUC: 95Acc: 90Sens: 88Spec: 92
Perek et al. [20]129 MultimodalNetworkAUC: 89Sens: 100Spec: 66
Fanizzi et al. [21]4812RandomForestAUC: 93.1Acc: 87.5Sens: 87.5Spec: 91.7
Losurdo et al. [23]5510SVMAcc: 80.91Sens: 90.28Spec: 71.55
Best proposed model582RandomForestAUC: 95.66Acc: 90.52Sens: 88.37Spec: 100
Moreover, compared to state of the art models, our model seems to perform better. In [19], the authors reached a sensitivity of 88% and a specificity of 92% by training a SVM classifier on a feature set extracted from 50 lesions manually segmented by radiologists. Finally, we want to emphasize the importance of radiomic analysis for the characterization of benign and malignant ROIs and the achievement of more balanced values for sensitivity and specificity. Specifically, we want to compare our results with the ones obtained in [20] (a sensitivity of 100% and a specificity of 66%) with the use of only textural descriptors provided by the radiologist combined with CESM pixel information extracted directly from the images.

5. Conclusions

With the aim of improving the number of early diagnoses of breast cancer, in this work we proposed an automated expert system for discriminating benign and malignant ROIs. We proposed the use of a principal component analysis combined with machine learning techniques in order to select the optimal subset for characterizing breast regions and classifying them. Particularly, we trained three binary classifiers on an increasing number of features sorted by their diagnostic power, evaluated in terms of AUC. Our model’s performance represents a step forward compared to our previous work, both in regard to the greater accuracy achieved in the classification of benign and malignant lesions and the smaller number of variables used to obtain these results.
  19 in total

1.  Novel network architecture and learning algorithm for the classification of mass abnormalities in digitized mammograms.

Authors:  Brijesh Verma
Journal:  Artif Intell Med       Date:  2007-11-09       Impact factor: 5.326

2.  Index for rating diagnostic tests.

Authors:  W J YOUDEN
Journal:  Cancer       Date:  1950-01       Impact factor: 6.860

3.  Six-year prospective evaluation of second-look US with volume navigation for MRI-detected additional breast lesions.

Authors:  Alfonso Fausto; Marco Bernini; Daniele La Forgia; Annarita Fanizzi; Marco Marcasciano; Luca Volterrani; Donato Casella; Maria Antonietta Mazzei
Journal:  Eur Radiol       Date:  2018-10-15       Impact factor: 5.315

4.  Breast MRI background parenchymal enhancement as an imaging bridge to molecular cancer sub-type.

Authors:  Giuseppe Dilorenzo; Michele Telegrafo; Daniele La Forgia; Amato Antonio Stabile Ianora; Marco Moschetta
Journal:  Eur J Radiol       Date:  2019-02-15       Impact factor: 3.528

5.  Microcalcification detection in full-field digital mammograms: A fully automated computer-aided system.

Authors:  T M A Basile; A Fanizzi; L Losurdo; R Bellotti; U Bottigli; R Dentamaro; V Didonna; A Fausto; R Massafra; M Moschetta; P Tamborra; S Tangaro; D La Forgia
Journal:  Phys Med       Date:  2019-06-14       Impact factor: 2.685

6.  Diagnostic performance of contrast-enhanced dual-energy spectral mammography (CESM): a retrospective study involving 644 breast lesions.

Authors:  María Del Mar Travieso-Aja; Daniel Maldonado-Saluzzi; Pedro Naranjo-Santana; Claudia Fernández-Ruiz; Wilsa Severino-Rondón; Mario Rodríguez Rodríguez; Víctor Vega Benítez; Octavio Pérez-Luzardo
Journal:  Radiol Med       Date:  2019-06-27       Impact factor: 3.469

7.  Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries.

Authors:  Freddie Bray; Jacques Ferlay; Isabelle Soerjomataram; Rebecca L Siegel; Lindsey A Torre; Ahmedin Jemal
Journal:  CA Cancer J Clin       Date:  2018-09-12       Impact factor: 508.702

8.  Mammography: an update of the EUSOBI recommendations on information for women.

Authors:  Francesco Sardanelli; Eva M Fallenberg; Paola Clauser; Rubina M Trimboli; Julia Camps-Herrero; Thomas H Helbich; Gabor Forrai
Journal:  Insights Imaging       Date:  2016-11-16

9.  Annual Report to the Nation on the Status of Cancer, part I: National cancer statistics.

Authors:  Kathleen A Cronin; Andrew J Lake; Susan Scott; Recinda L Sherman; Anne-Michelle Noone; Nadia Howlader; S Jane Henley; Robert N Anderson; Albert U Firth; Jiemin Ma; Betsy A Kohler; Ahmedin Jemal
Journal:  Cancer       Date:  2018-05-22       Impact factor: 6.860

10.  Logistic Regression for Machine Learning in Process Tomography.

Authors:  Tomasz Rymarczyk; Edward Kozłowski; Grzegorz Kłosowski; Konrad Niderla
Journal:  Sensors (Basel)       Date:  2019-08-02       Impact factor: 3.576

View more
  13 in total

1.  Prediction of Breast Cancer Histological Outcome by Radiomics and Artificial Intelligence Analysis in Contrast-Enhanced Mammography.

Authors:  Antonella Petrillo; Roberta Fusco; Elio Di Bernardo; Teresa Petrosino; Maria Luisa Barretta; Annamaria Porto; Vincenza Granata; Maurizio Di Bonito; Annarita Fanizzi; Raffaella Massafra; Nicole Petruzzellis; Francesca Arezzo; Luca Boldrini; Daniele La Forgia
Journal:  Cancers (Basel)       Date:  2022-04-25       Impact factor: 6.575

2.  Sensitivity of Contrast-Enhanced Breast MRI vs X-ray Mammography Based on Cancer Histology, Tumor Grading, Receptor Status, and Molecular Subtype: A Supplemental Analysis of 2 Large Phase III Studies.

Authors:  Jan Endrikat; Gilda Schmidt; Daniel Haverstock; Olaf Weber; Zuzana Jirakova Trnkova; Jörg Barkhausen
Journal:  Breast Cancer (Auckl)       Date:  2022-04-19

3.  Comparison of Contrast-Enhanced Spectral Mammography and Contrast-Enhanced MRI in Screening Multifocal and Multicentric Lesions in Breast Cancer Patients.

Authors:  Lei Feng; Lei Sheng; Litao Zhang; Na Li; Yuanzhong Xie
Journal:  Contrast Media Mol Imaging       Date:  2022-04-06       Impact factor: 3.009

4.  Contrast-Enhanced Spectral Mammography-Based Prediction of Non-Sentinel Lymph Node Metastasis and Axillary Tumor Burden in Patients With Breast Cancer.

Authors:  Xiaoqian Wu; Yu Guo; Yu Sa; Yipeng Song; Xinghua Li; Yongbin Lv; Dong Xing; Yan Sun; Yizi Cong; Hui Yu; Wei Jiang
Journal:  Front Oncol       Date:  2022-05-06       Impact factor: 5.738

5.  An Invasive Disease Event-Free Survival Analysis to Investigate Ki67 Role with Respect to Breast Cancer Patients' Age: A Retrospective Cohort Study.

Authors:  Raffaella Massafra; Samantha Bove; Daniele La Forgia; Maria Colomba Comes; Vittorio Didonna; Gianluca Gatta; Francesco Giotta; Agnese Latorre; Annalisa Nardone; Gennaro Palmiotti; Davide Quaresmini; Lucia Rinaldi; Pasquale Tamborra; Alfredo Zito; Alessandro Rizzo; Annarita Fanizzi; Vito Lorusso
Journal:  Cancers (Basel)       Date:  2022-04-28       Impact factor: 6.575

Review 6.  Radiomics in Triple Negative Breast Cancer: New Horizons in an Aggressive Subtype of the Disease.

Authors:  Camil Ciprian Mireștean; Constantin Volovăț; Roxana Irina Iancu; Dragoș Petru Teodor Iancu
Journal:  J Clin Med       Date:  2022-01-26       Impact factor: 4.241

7.  Deep learning analysis of contrast-enhanced spectral mammography to determine histoprognostic factors of malignant breast tumours.

Authors:  Caroline Dominique; Françoise Callonnec; Anca Berghian; Diana Defta; Pierre Vera; Romain Modzelewski; Pierre Decazes
Journal:  Eur Radiol       Date:  2022-01-29       Impact factor: 7.034

8.  Special Issue "Advancement in Breast Diagnostic and Interventional Radiology".

Authors:  Graziella Di Grezia
Journal:  Diagnostics (Basel)       Date:  2022-01-17

9.  Architectural Distortion-Based Digital Mammograms Classification Using Depth Wise Convolutional Neural Network.

Authors:  Khalil Ur Rehman; Jianqiang Li; Yan Pei; Anaa Yasin; Saqib Ali; Yousaf Saeed
Journal:  Biology (Basel)       Date:  2021-12-23

10.  Optimal Breast Density Characterization Using a Three-Dimensional Automated Breast Densitometry System.

Authors:  Reika Yoshida; Takenori Yamauchi; Sadako Akashi-Tanaka; Misaki Matsuyanagi; Kanae Taruno; Terumasa Sawada; Akatsuki Kokaze; Seigo Nakamura
Journal:  Curr Oncol       Date:  2021-12-14       Impact factor: 3.677

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.