Literature DB >> 33920219

Semantic Cardiac Segmentation in Chest CT Images Using K-Means Clustering and the Mathematical Morphology Method.

Beanbonyka Rim1, Sungjin Lee1, Ahyoung Lee2, Hyo-Wook Gil3, Min Hong4.   

Abstract

Whole cardiac segmentation in chest CT images is important to identify functional abnormalities that occur in cardiovascular diseases, such as coronary artery disease (CAD) detection. However, manual efforts are time-consuming and labor intensive. Additionally, labeling the ground truth for cardiac segmentation requires the extensive manual annotation of images by the radiologist. Due to the difficulty in obtaining the annotated data and the required expertise as an annotator, an unsupervised approach is proposed. In this paper, we introduce a semantic whole-heart segmentation combining K-Means clustering as a threshold criterion of the mean-thresholding method and mathematical morphology method as a threshold shifting enhancer. The experiment was conducted on 500 subjects in two cases: (1) 56 slices per volume containing full heart scans, and (2) 30 slices per volume containing about half of the top of heart scans before the liver appears. In both cases, the results showed an average silhouette score of the K-Means method of 0.4130. Additionally, the experiment on 56 slices per volume achieved an overall accuracy (OA) and mean intersection over union (mIoU) of 34.90% and 41.26%, respectively, while the performance for the first 30 slices per volume achieved an OA and mIoU of 55.10% and 71.46%, respectively.

Entities:  

Keywords:  K-Means clustering; chest CT scans; image processing; mathematical morphology method; silhouette score; whole cardiac segmentation

Mesh:

Year:  2021        PMID: 33920219      PMCID: PMC8070040          DOI: 10.3390/s21082675

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

Cardiovascular disease (CVD) has been reported as one of the leading causes of death globally and occurs due to functional abnormalities in the heart and blood vessels [1]. In 2016, according to the World Health Organization (WHO), about 17.9 million people died from CVDs, which is equivalent to 31% of all global deaths (mainly from stroke and heart attack) [1]. For instance, one of the CVDs, coronary artery disease (CAD) is a group of abnormalities in blood vessels supplying the heart muscle [1]. CAD is caused by a surplus of calcium in the coronary artery trees. Excessive calcium can narrow the arteries and increase the risk of heart attack [2,3]. Therefore, an early assessment and diagnosis can significantly reduce the life-threatening nature of this CVD and improve quality of life for the afflicted patients [2,3]. In modern medical imaging modalities, computed tomography (CT), magnetic resonance imaging (MRI), and ultrasound are used to assist in identifying abnormal findings in the human body for early assessment and diagnosis [4,5,6,7]. Recently, the non-gated and non-invasive chest CT has been used to provide potential support for investigative imaging tests to interpret cardiac function states [8,9,10]. More detailed characteristics of chest CT images are described in Appendix A. Cardiac segmentation in chest CT images has played a key role in partitioning the whole chest CT image into a number of anatomically meaningful regions, known as regions of interest (ROIs). Typically, the anatomical ROIs for cardiac segmentation focus on the four chambers of the heart—the left ventricle (LV), right ventricle (RV), left atrium (LA) and right atrium (RA)—or the entire heart anatomy, including the four chambers, coronary arteries and descending thoracic aorta (DA) [8,9,10]. The manual process of cardiac segmentation can be time-consuming and labor intensive for radiologists. To overcome this burden, fully automatic methods have been proposed by applying computer-aided technologies [1,7,8,9,10]. These techniques have been built based on earlier approaches [8] such as graph-based segmenting [11,12], mean-thresholding [13,14,15,16,17] and fuzzy clustering methods [18,19,20,21]. Later, the deep learning approach showed promising successful performance [7,9,10]. The deep learning approach has two learning manners, including supervised learning [22,23,24,25,26], which requires a ground truth to align the loss function, and unsupervised learning [27,28,29], which learns the features without ground truth labeling. Labeling the ground truth for cardiac segmentation requires the extensive manual annotation of images by the radiologist, which can be time-consuming and labor intensive [7,8,9,10,22,23,24,25,26]. Due to the difficulty in obtaining annotated data and the required expertise as an annotator, an unsupervised approach has been considered in this study. The goal of cardiac segmentation is to partition the whole chest CT image into cardiac anatomical ROIs, with respect to not only the dissimilarity of each pixel’s value, but also the meaningful structure (i.e., geometrical position) of cardiac anatomical ROIs. For example, the Hounsfield units (the quantitative scale of chest CT images) of heart muscles and other muscles in the body are in the same range (almost identical). Detailed information of the Hounsfield unit (HU) of chest CT images is described in Appendix B. Additionally, the cardiac anatomical substructures (i.e., the four chambers, DA and coronary artery) are usually formed in one shape that is mostly arranged in the center of a chest CT image [3,30]. Inspired by the performance of the mean-thresholding method [13,14,15,16,17], we adapted it for this current study by utilizing the K-Means clustering method as a threshold criterion. The K-Means clustering method [31,32] is a simple unsupervised method, which exploits Euclidean distances to compute the mean of all given pixels and assigns pixels into k different clusters based on the nearest mean. However, the anatomical structure of cardiac tissues and the quantitative scale (i.e., HU) of chest CT images are very complicated for cardiac segmentation when using the K-Means clustering method directly. Therefore, we exploited a mathematical morphology method [33,34] to enhance the shifting of the mean threshold. In this paper, semantic whole-heart segmentation combining K-Means clustering as a threshold criterion of mean-thresholding and the mathematical morphology method as a threshold shifting enhancer is proposed. In the proposed approach, the K-Means method is utilized to automatically cluster pixels, while the mathematical morphology method is used to remove noise, fill holes and convex hull; also, prior knowledge of chest anatomical structures [3,30] is required to assist in the awareness of geometrical positions. The silhouette scoring method [32,35] is applied to evaluate the performance of K-Means clustering, while overall accuracy (OA) and mean intersection over union (mIoU) are calculated to evaluate the overall performance of cardiac segmentation. The remaining parts of the paper are organized as follows. Section 2 describes related works of earlier approaches, such as graph-based, mean-threshold and fuzzy clustering, in addition to deep learning approaches such as supervised and unsupervised learning methods. Section 3 describes a step-by-step hierarchical flow of the proposed methodology. Experimental results and comparison discussions are analyzed in Section 4 and Section 5, respectively. Section 6 concludes our study. Finally, supplemental literature reviews of CT, HU and algorithm tables are presented in Appendix A, Appendix B and Appendix C, respectively.

2. Related Works

Shi et al. [11] conducted natural image segmentation by proposing a normalized cut criterion. The method addressed images as a graph partition problem. The normalized cut was used as a measure of dissimilarity between subgroups of a graph. The eigenvalue was utilized to minimize the criteria of the normalized cut. Pedro et al. [12] conducted natural image segmentation through a pairwise region comparison method. The method also addressed images as a graph partition problem. The greedy decision was used as a measurement criterion. Dorin et al. [13] conducted natural image segmentation by proposing a recursive mean shift procedure to generate M-estimators. The M-estimators were utilized in detecting the modes of the density. The method produced significant results with a low dimension of the space. Larrey-Ruiz et al. [14] conducted cardiac segmentation in chest CT images using multiple threshold values. The thresholding was calculated by the mean value of statistical local parameters (i.e., pixels are in one slice) and global parameters (i.e., pixels are across all slices of a full volume). Huo et al. [15] conducted weakly unsupervised cardiac segmentation as an initial step for coronary calcium detection. The method contained many unrelated anatomical ROIs, such as the spine, ribs, and muscles. The method was adapted from Liao et al. [16], utilizing a convex hull of lungs as the base parameters. Rim et al. [17] conducted whole-heart segmentation by adapting the cardiac segmentation method of Huo et al. [15] and Liao et al. [16]. The method used an alternative threshold of a convex hull of lungs and a convex hull of ribs as a base parameter. As the alternative threshold was manually defined by the mathematical geometry of the lungs, the method is an ad hoc set of each image. Radha et al. [18] conducted brain image segmentation using an intelligent fuzzy level set. Quantum particle swarm optimization was proposed to improve the steadiness and meticulousness in order to reduce opening sensitivity. The results for this method showed that it could optimize up to 15% more than the original fuzzy level set method. However, the method experienced challenges with neoplastic or degenerative ailment images. Versaci et al. [19] conducted image edge detection by proposing a new approach of fuzzy divergence and fuzzy entropy. The proposed fuzzy entropy used fuzzy divergence as the distances between fuzzified images, which were computed by means of fuzzy divergence. Chanda et al. [20] conducted cardiac MRI image segmentation using a fuzzy-based approach. The method was based on morphological, threshold-based segmentation and fuzzy-based edge detection. The method achieved more than 90% accuracy. Kong et al. [21] conducted an image segmentation method using an intuitionistic fuzzy C-Means clustering algorithm. The method defined a modified non-membership function, an initial clustering center based on grayscale features, an improved nonlinear kernel function and a new measurement of fuzzy entropy. The method outperformed existing algorithms, but it took a lot of computational time. Ronneberger et al. [22] conducted an electron microscopy (EM) image segmentation using supervised deep learning by defining U-net architecture. The network consisted of 23 convolutional layers. The experiment was trained with 30 images of size 512 × 512 and received an average IOU of 92%. The U-net model has become the gold standard for biomedical image segmentation. Payer et al. [23] conducted whole-heart segmentation using supervised deep learning by defining two-step CNN architecture. The experiment was trained on CT images with 20 volumes of size 300 × 300 × 188 and achieved an average dice similarity coefficient of 90.8%. Ahmed et al. [24] conducted whole-heart segmentation using supervised deep learning by defining a CNN network and using stacked denoising auto-encoders. The experiment was trained on CT images of eight subjects and achieved an average accuracy of 93.77%. Liao et al. [25] conducted 3D whole heart segmentation using supervised deep learning by defining a multi-modality (i.e., CT and MRI) transfer learning network with adversarial training. The network introduced the attention mechanism into the U-net network. The experiment was trained on 60 CT volumes with a dice value of 0.914. Max et al. [26] conducted multi-modality (i.e., CT and MRI) 3D whole-heart segmentation using supervised deep learning by defining a shape encoder–decoder network. The experiment was trained on 15 CT volumes with a dice value of 0.653. Xia et al. [27] conducted natural image segmentation using unsupervised deep learning by defining a W-net architecture. The network was a concatenation of two U-net networks. The experiment was trained on 11,530 images and achieved a probabilistic rand index (PRI) of 0.86. Joyce et al. [28] conducted multi-class medical image segmentation on Myocardium (MYO), left ventricle (LV) and right ventricle (RV) regions using unsupervised deep learning by defining a generative adversarial network (GAN) model. The experiment was trained on 20 CT volumes with a dice value of 0.51. Perone et al. [29] conducted medical image segmentation using unsupervised deep learning by defining a self-ensembling architecture. The experiment was trained on MRI images from 100 subjects and achieved the best dice value of 0.847.

3. Methodology

3.1. Overall Workflow of the Proposed Method

The segmentation process is intended to generate a binary mask of whole-heart anatomical ROIs, including the four chambers, coronary arteries, and DA, as shown in Figure 1, Figure 2 and Figure 3 and listed in Appendix C. As the cardiac anatomical ROIs are formed in one shape, arranged mostly in the center of a chest CT image [3,30], a step-by-step hierarchical process to enhance the mean-threshold method from the corner towards the center region of the chest CT image is proposed. Firstly, the air substance at the corner of a chest CT image is filtered by a convex hull of foreground mask. Then, other human body substances such as fat, muscles and lungs are filtered by a convex hull of lung mask. Finally, the spine substance is filtered by a convex hull of spine mask. More details of each step are explained in the following sub-sections.
Figure 1

Overall workflow of the step-by-step hierarchical cardiac segmentation process using K-Means clustering and mathematical morphology method. Abbreviations: CT, computed tomography; HU, hounsfield unit.

Figure 2

Examples of the step-by-step hierarchical process: (a) whole chest CT image; (b) foreground mask; (c) convex hull of foreground mask; (d) lung mask; (e) convex hull of lung mask; (f) intermediate heart mask; (g) spine mask; (h) convex hull of spine mask; (i) heart mask; and (j) segmented heart image.

Figure 3

Examples of intermediate heart segmentation: (a) intermediate heart image; and (b) corresponded frequency histogram.

3.2. Grayscale Conversion

Firstly, the raw CT image is read in HU pixels, which is the standard radio-density scale. One CT image has a resolution of (i.e., width × height × channels), with the range around [−1000, +1000]. In the current study, 56 images per volume are trimmed. Thus, one CT volume is in size. The HU pixels are converted into grayscale pixels using the standard range normalization, as shown in Equation (1). where is a set of HU pixels, , and is a set of grayscale pixels, and . and are the minimum and maximum value of HU pixels, respectively. An example of a chest CT image in grayscale pixels is shown in Figure 2a.

3.3. Convex Hull of Foreground Mask

The K-Means method (known as the unsupervised technique in clustering literature [31,32]) is used to cluster the whole chest CT image into foreground (i.e., anatomical ROIs such as fat, muscle, lungs, heart and other) and background (i.e., air) clusters. Given clusters, and the set of grayscale pixels , the centroid clusters are calculated by minimizing the function , as shown in Equation (2). where the centroid clusters . The foreground binary mask , is the result of the thresholding condition, as shown in Equation (3). where the foreground threshold is calculated by the mean of centroid clusters . Lines randomly situated at the bottom of the chest in the CT images were observed, as shown in Figure 2a, which are considered as noise. To remove this noise, the morphological binary opening operation [33,34] is conducted with the default structure element from [34]. The operation is iterated twice: . Then, the morphological convex hull operation [33,34] is adapted to cover all human body substances as foreground, as shown in Equation (4). where is the foreground mask and is the default structure elements with i = 1, 2, 3, 4. Given that is a foreground mask (, the (where r = 1, 2, 3, …) is iteratively applied by a hit-or-miss transform with ; and when it converges (i.e., is equal to ), it is united with , which is referred to as . Then, the convex hull is a union of . The examples and generation processes of the foreground mask and the convex hull of foreground mask are shown in Figure 2b,c and Table A2, respectively.
Table A2

Generation process of a convex hull of foreground mask.

Algorithm 1: Generate a convex hull of foreground mask
Input: Whole chest CT image in grayscale format (512 × 512)
Output: Convex hull of foreground mask in binary format (512 × 512)
 1: Begin
 2: centroid_clusters = KMean(input_image, n_cluster = 2)
 3: fg_threshold = mean(centroid_clusters)
 4: if input_image > fg_threshold:
 5:  fg_mask = 1
 6: else
 7:  fg_mask = 0
 8: fg_mask = morphology.opening(fg_mask, n_iteration = 2)
 9: ch_fg_mask = morphology.convex_hull(fg_mask)
10: End

3.4. Convex Hull of Lung Mask

The process of generating a convex hull of lung mask is intended to remove fat, muscle and rib substances. Firstly, the lung mask is computed. The foreground threshold is used to compute the lung mask. The enhanced grayscale pixels , are computed to assist the thresholding. Then, the lung mask , is computed as shown in Equation (5). There are blood vessels within the lungs, which result in many small holes in lung mask . To fill those small holes, the morphological binary closing operation [33,34] is conducted with the default structure element from [34]. The operation is iterated twice: . Then, the convex hull of lung mask , is computed by the morphological convex hull operation [33,34] adapted from Equation (4), as shown in Equation (6). The intermediate heart mask , is a bitwise AND operation between the convex hull of lung mask and an inversion of lung mask , which is computed as shown in Equation (7). where is a bitwise NOT operator: . The examples and generation processes of the lung mask, convex hull of lung mask and intermediate heart mask are shown in Figure 2d–f and Table A3, respectively.
Table A3

Generation process of a lung mask, a convex hull of lung mask and an intermediate heart mask.

Algorithm 2: Generate a lung mask, a convex hull of lung mask and an intermediate heart mask
Input: Whole chest CT image in grayscale format (512 × 512);Convex hull of foreground mask in binary format (512 × 512)
Output: Intermediate heart mask in binary format (512 × 512)
 1: Begin
 2: foreach pixel in input_image:
 3:  if ch_fg_mask == 0:
 4:   fg_pixel = 255
 5:  else
 6:   fg_pixel = pixel
 7: if fg_pixel < fg_threshold:
 8:  lung_mask = 1
 9: else
10:  lung_mask = 0
11: lung_mask = morphology.closing(lung_mask, n_iteration = 2)
12: ch_lung_mask = morphology.convex_hull(lung_mask)
13: int_heart_mask = ch_lung_mask AND invert(lung_mask)
14: End

3.5. Convex Hull of Spine Mask

The process of generating a convex hull of spine mask is intended to remove spine pixels and other substances under the spine and DA. Firstly, the spine mask is computed. The enhanced intermediate heart grayscale pixels , are computed to assist the thresholding, as shown in Figure 3a. Then, the centroid clusters of the K-Means method adapted from Equation (2) are computed to divide the enhanced grayscale pixels into background, heart and spine clusters with a value of . The spine pixels are brighter compared to the background and heart pixels, as shown in Figure 3a. We can assume that spine pixels are in the last right cluster in Figure 3b. Therefore, the spine threshold is the maximum centroid of clusters . Then, the spine binary mask , , is a result of a thresholding condition, which is defined as shown in Equation (8). The morphological binary closing and opening operation are applied to fill holes and remove islands, respectively. Each operation is iterated twice: and then, . The spine mask does not cover pixels of substances around the spine and under the DA, as shown in Figure 2g. To remove those pixels, the convex hull of spine mask is computed. and which are denoted as the top and center coordinates of the white convex polygon in the spine mask , are computed by the region properties function and from [34]: and , respectively. Then, the convex hull of spine mask , is defined, as shown in Equation (9). where and are a set of coordinates of spine mask and is a set of row axes of convex hull spine mask . Then, the heart mask , is a bitwise AND operation between the intermediate heart mask and an inversion of spine mask convex hull , which is computed as shown in Equation (10). where is a bitwise NOT operator: . The examples and generation processes of the spine mask, convex hull of spine mask and heart mask are shown in Figure 2g–i and Table A4, respectively.
Table A4

Generation process of a spine mask, a convex hull of spine mask and a heart mask.

Algorithm 3: Generate a spine mask, a convex hull of spine mask and a heart mask
Input: Whole chest CT image in grayscale format (512 × 512);Intermediate heart mask in binary format (512 × 512)
Output: Heart mask in binary format (512 × 512)
 1: Begin
 2: foreach pixel in input_image:
 3:  if int_heart_mask == 0:
 4:   int_heart_pixel = 0
 5:  else
 6:   int_heart_pixel = pixel
 7: centroid_clusters = KMean(int_heart_pixel, n_cluster = 3)
 8: spine_threshold = max(centroid_clusters)
 9: if int_heart_pixel > spine_threshold:
10:  spine_mask = 1
11: else
12:  spine_mask = 0
13: spine_mask = morphology.closing(spine_mask, n_iteration = 2)14: spine_mask = morphology.opening(spine_mask, n_iteration = 2)15: (x0,y0) = measure.regionprob(spine_mask)16: (xc,yc) = measure.regionprob(spine_mask)
17: ch_spine_mask = spine_mask[y0 : , : xc] is set to 1
18: ch_spine_mask = ch_spine_mask[yc : , : ] is set to 1
19: heart_mask = int_heart_mask AND invert(ch_spine_mask)
20: End

3.6. Heart Pixel Segmentation

This section explains how to filter the heart image from the heart mask. Given whole chest HU pixels , and the heart mask , , the segmented heart pixels , are computed as shown in Equation (11). The examples and generation processes of segmented heart images are shown in Figure 2j and Table A5, respectively.
Table A5

Generation process of segmented heart pixels.

Algorithm 4: Segment a heart image
Input: Whole chest CT image in HU format (512 × 512);Heart mask in binary format (512 × 512)
Output: Segmented heart pixel in HU format (512 × 512)
 1: Begin
 2: foreach pixel in input_image:
 3:  if heart_mask == 0:
 4:   heart_pixel = −1000
 5:  else
 6:   heart_pixel = pixel
 7: End

3.7. Convex Hull of Lung Mask Refinement

The proposed segmentation method depends on the convex hull of lung mask, similar to Huo et al. [14] and Rim et al. [15]. Empirically, when the lungs are not well surrounding the heart region, the convex hull of lung mask fails to compute intermediate heart mask using Equation (7). In this case, the convex hull of lung mask is refined, as shown in Figure 4. The lungs were observed to vanish little by little from the CT image when the liver appears. Thus, within 56 slices per CT volume, the last lungs-well-surrounded slice is recorded and used as the global parameters for the next slice. Then, the local and global parameters are combined to compute the convex hull of lung mask.
Figure 4

Examples of refining a convex hull of lung mask: (a) lung mask of the (n−i)th slice; (b) convex hull of the (n−i)th slice; (c) lung mask of the nth slice; and (d) refined convex hull of lung mask of the nth slice.

Given is the number of slices in one volume (i.e., ) and is the top row axis of the white convex polygon in the lung mask and , , the convex hull of lung mask , is computed by a bitwise OR operation, as shown in Equation (12). where and () are the current and the last lungs-well-surrounded slice in , respectively. The examples and generation processes of the refined convex hull of lung mask are shown in Figure 4 and Table A6, respectively.
Table A6

Generation process of a refined convex hull of lung mask.

Algorithm 5: Refine a convex hull of lung mask
Input: Lung mask in binary format (512 × 512 × 56)
Output: Refined convex hull of lung mask in binary format (512 × 512 × 56)
 1: Begin
 2: for n in range(56): 3: if y0 of lung_mask [n] > y0 of lung_mask [n-i]:
 4:  ch_lung_mask [n] = ch_lung_mask [n] OR ch_lung_mask [n-i]
 5: End

4. Experimental Results

4.1. Experimental Setup

The current experiment was conducted on a Windows 10 computer with an Intel Core™ i7-9700 CPU @ 3.00 GHZ, 32.0 GB RAM, and an NVIDIA GeForce RTX 2070 graphics card. The code was written in Python language with the Scikit image library [34] and the K-Mean method of the Scikit learn library [32]. The method was applied on a dataset from Soonchunhyang University Cheonan Hospital [36]. The dataset was acquired randomly from 500 subjects who were scanned using a CT scanner (Phillips iCT 256) during 2019. The chest CT slices were captured in diverse ranges containing between 56 to 84 slices. The first 56 slices were selected for the current study. The resolution of each slice was the same at 512 × 512 pixels. The FOV was 250 × 250 mm and the slice thickness was 2.5 mm. All data were stored in dicom 3.0 format [37].

4.2. Silhouette Score

The silhouette method [32,35] was applied to score how well the K-Mean method [31,32] separated the clusters. The formula of the silhouette method is shown in Equation (13). where is the average distance from the th pixel to all pixels in the same cluster and is the average distance from th pixel to all pixels in the closest cluster. The has a range of [−1, +1]. If , it means that the clusters are not well separated. If , it means that the clusters are well separated. If , it means that the clusters are overlapping. The evaluation was analyzed on three cases where denotes the number of images: (1) the first 30 slices from the 1st to 30th slice: ; (2) the last 36 slices from the 31st to 56th slice: ; and (3) a full volume per subject from the 1st to 56th slice: , as shown in Table 1.
Table 1

Silhouette scores for 500 subjects. The shilhouette scores of three cases are described in all filtering (foreground, spine, and overall foreground and spine filtering).

MinMaxMean±SD
Foreground filtering
1st to 30th slice0.01500.43410.28910.0879
31st to 56th slice−0.10020.57120.29040.1050
1st to 56th slice−0.10020.57120.29030.1040
Spine filtering
1st to 30th slice0.41760.63320.53670.0455
31st to 56th slice0.37560.673590.53560.0571
1st to 56th slice0.37560.67350.53560.0565
Overall Foreground and Spine filtering
1st to 30th slice0.41760.53370.41290.0605
31st to 56th slice0.14670.60530.41300.0762
1st to 56th slice0.14670.60530.41300.0753
The K-Means method was applied twice for foreground filtering and spine filtering in Equation (2) of Section 3.3 and Equation (8) of Section 3.5, respectively. For foreground filtering, the K-Means method of the three cases achieved mean scores of 0.2891, 0.2904 and 0.2903, respectively. For spine filtering, the K-Means method of the three cases achieved mean scores of 0.5367, 0.5356 and 0.5356, respectively. For overall foreground and spine filtering, the K-Means method for the three cases achieved mean scores of 0.4129, 0.4130 and 0.4130, respectively. We noticed that the K-Means method performed the spine clustering with higher scores than the foreground clustering. Additionally, the K-Means method achieved almost the same mean scores for all three cases in overall foreground and spine filtering.

4.3. Human Visual Inspection Evaluation

An intersection over union of each subject, an overall accuracy and a mean intersection over union were calculated for overall segmentation, as shown in Equation (14). where is the number of subjects (i.e., 500 subjects) and is the th subject in . is the number of images. , and represent number of images of true positives, false positives and false negatives of the segmentation result, respectively. and evaluate the overall quality of the segmentation, and the of each subject evaluates the quality of all images per subject. Since our segmentation method does not have a ground truth for validating, we conducted a visual inspection manually in which the human error rate was assumed to be around 5%. If the segmented image consists of four chambers, the coronary arteries and the DA, it is considered as a whole-heart segmentation, as shown in Figure 5a,b,e,f,i,j. If the segmented image consists of four chambers and the coronary arteries but misses DA, it is considered as a four-chamber segmentation, as shown in Figure 5c,d,g,h,k,l.
Figure 5

Examples of segmented heart results. (a,b) TP of whole-heart segmentation; (c,d) TP of four-chamber segmentation; (e,f) FP of whole-heart segmentation; (g,h) FP of four-chamber segmentation; (i,j) FN of whole-heart segmentation; (k,l) FN of four-chamber segmentation. Abbreviations: IoU, intersection over union; TP, true positive; FP, false positive; FN, false negative.

For the first 30 slices, the segmentation achieved high performance on both whole-heart and four-chamber segmentations with and values of 55.10%, 71.46%, 82.62% and 82.62% respectively, as shown in Table 2. Among the 500 subjects, there were 316 and 341 subjects whose was higher than the for whole-heart and four-chamber segmentations, respectively, as shown in Figure 6a,b. Additionally, the minimum and maximum were 0% and 100%, respectively. Among the 500 subjects, there were 58 and 14 subjects whose was 0%, while there were 234 and 247 subjects whose was 100% for whole-heart and four-chamber segmentations, respectively, as shown in Figure 6c,d.
Table 2

Evaluation results of 500 subjects (%). The OA, mIoU, IoU (min) and IoU (max) of three cases are described in both segmentations (whole-heart and four-chamber segmentation). Abbreviations: OA, overall accuracy; mIoU, mean intersection over union; IoU (min), minimum IoU; IoU (max), maximum IoU.

OA mIoU IoU min IoU max
Whole-heart
1st to 30th slice55.1071.460.00100.00
31st to 56th slice8.3712.600.0069.23
1st to 56th slice34.9041.260.0085.45
Four-chamber
1st to 30th slice82.6282.620.00100.00
31st to 56th slice10.3515.420.0069.23
1st to 56th slice50.9154.100.0089.13
Figure 6

Number of subjects: (a) whose was lower than the (blue) and higher than the (orange) for whole-heart segmentation; (b) whose was lower than the (blue) and higher than the (orange) of four-chamber segmentation; (c) whose was minimum (blue) and maximum (orange) for whole-heart segmentation; and (d) whose was minimum (blue) and maximum (orange) for four-chamber segmentation.

For the last 36 slices, the segmentation achieved low performance on both whole-heart and four-chamber segmentations with and values of 8.37%, 12.60%, 10.35% and 15.42%, respectively. Among the 500 subjects, there were 198 and 215 subjects whose was higher than the for whole-heart and four-chamber segmentation, respectively. Additionally, the minimum and maximum are 0% and 69.23%, respectively. Among the 500 subjects, there were 222 and 180 subjects whose was 0%, while there were 1 and 1 subjects whose was 69.23% for whole-heart and four-chamber segmentation, respectively. For the full volume, the segmentation achieved good performance on both whole-heart and four-chamber segmentations with and values of 34.90%, 41.26%, 50.91% and 54.10%, respectively. Among the 500 subjects, there were 283 and 315 subjects whose was higher than the for whole-heart and four-chamber segmentation, respectively. Additionally, the minimum and maximum were 0%, 0%, 85.45% and 89.13%, respectively. Among the 500 subjects, there were 51 and 11 subjects whose was 0%, while there were 1 and 1 subjects whose was 85.45% and 89.13% for whole-heart and four-chamber segmentation, respectively.

5. Discussion

Among the mean-thresholding methods proposed by [13,14,15,16,17], there is one paper by Larrey-Ruiz et al. [14] in which cardiac segmentation on 32 chest CT images was conducted by defining multiple threshold values. The thresholding was calculated by the mean value of statistical local parameters (i.e., pixels were in one slice) and global parameters (i.e., pixels were across all slices of a whole volume). Table 3 shows comparison results for the top 32 subjects of our proposed method with Larrey-Ruiz et al. [14] for whole-heart segmentation. For the first 30 slices, our results outperform Larrey-Ruiz et al. [14] in both and (maximum accuracy) with values of 100% and 100%, and 94.42% and 99.81%, respectively. For the full volume, Larrey-Ruiz et al. [14] outperforms our method in with values of 87.64% and 73.66%, respectively. However, our method outperforms Larrey-Ruiz et al. [14] in with values of 67.85% and 46.95%, respectively.
Table 3

Comparison results of 32 subjects (%).

Whole-HeartOur MethodLarrey-Ruiz et al. [14]
1st to 30th slice—OA 100.00 94.42
1st to 30th slice—A max 100.00 99.81
31st to 56th slice—OA34.28-
1st to 56th slice—OA73.66 87.64
1st to 56th slice—A min 67.85 46.95
Among the unsupervised deep learning approaches proposed by [27,28,29], there is one paper by Joyce et al. [28] in which multi-class medical image segmentation was conducted on MYO, LV and RV regions on 20 CT volumes with a dice value () of 0.51. For the top 20 subjects, our result of a full volume for whole-heart segmentation outperforms Joyce et al. [28] with an of 0.7833, as shown in Table 4.
Table 4

Comparison results of 20 subjects.

Whole-HeartOur MethodJoyce et al. [28]
1st to 56th slice0.7833 (mIoU)0.51 (mDice)

6. Conclusions

This paper presented semantic whole-heart segmentation combining K-Means clustering as a threshold criterion of the mean-thresholding method and the mathematical morphology method as a threshold shifting enhancer. The experiment was conducted on 500 subjects in two cases: (1) 56 slices per volume containing full heart scans, and (2) 30 slices per volume containing about half of the top of heart scans before the liver appears. In both cases, the results showed an average silhouette score of the K-Means method, with a value of 0.4130. Additionally, the experiment on 56 slices per volume achieved an OA and mIoU of 34.90% and 41.26%, respectively; while the performance result on the first 30 slices per volume achieved an OA and mIoU of 55.10% and 71.46%, respectively. High performance was achieved when the heart was well surrounded by lungs. Otherwise, low performance was achieved. The low performance was likely caused by the lack of filtering of the liver, as both the HU pixels and geometrics of the organ could not be used as a criterion for thresholding. Additionally, the goal of the proposed research was to segment the whole heart. However, the results showed that the four-chamber segmentation outperformed the whole-heart segmentation. The outperformance was due to a failure in generating the convex hull of spine mask. There are limitations in this study, such as the failure of DA segmentation and liver removal, but the main contributions of our proposed method can be summed up as the following: (1) we proposed fully unsupervised semantic whole-heart segmentation from chest CT images; (2) we proposed the K-Means method as a thresholding criterion and the mathematical morphology method as a threshold-shifting enhancer; and (3) we demonstrated good performance for the first 30 slices, which will be able to be used as an initial step for other cardiac applications in other CADs. Finally, the future direction of our research is to conduct an unsupervised deep learning approach to overcome the abovementioned limitations.
Table A1

Hounsfield unit substances in chest CT images.

SubstanceHounsfield Unit
Air−1000
Lung−500
Fat−100~−50
Water0
Muscle+10~+40
Liver+40~+60
Bone+700~+1000
  10 in total

Review 1.  Cardiac findings on non-gated chest computed tomography: A clinical and pictorial review.

Authors:  Rene Epunza Kanza; Christian Allard; Michel Berube
Journal:  Eur J Radiol       Date:  2015-12-03       Impact factor: 3.528

2.  Coronary Calcium Detection using 3D Attention Identical Dual Deep Network Based on Weakly Supervised Learning.

Authors:  Yuankai Huo; James G Terry; Jiachen Wang; Vishwesh Nath; Camilo Bermudez; Shunxing Bao; Prasanna Parvathaneni; J Jeffery Carr; Bennett A Landman
Journal:  Proc SPIE Int Soc Opt Eng       Date:  2019-03-15

3.  scikit-image: image processing in Python.

Authors:  Stéfan van der Walt; Johannes L Schönberger; Juan Nunez-Iglesias; François Boulogne; Joshua D Warner; Neil Yager; Emmanuelle Gouillart; Tony Yu
Journal:  PeerJ       Date:  2014-06-19       Impact factor: 2.984

4.  Unsupervised domain adaptation for medical imaging segmentation with self-ensembling.

Authors:  Christian S Perone; Pedro Ballester; Rodrigo C Barros; Julien Cohen-Adad
Journal:  Neuroimage       Date:  2019-03-19       Impact factor: 6.556

5.  MMTLNet: Multi-Modality Transfer Learning Network with adversarial training for 3D whole heart segmentation.

Authors:  Xiangyun Liao; Yinling Qian; Yilong Chen; Xueying Xiong; Qiong Wang; Pheng-Ann Heng
Journal:  Comput Med Imaging Graph       Date:  2020-08-22       Impact factor: 4.790

6.  Multimodal 3D medical image registration guided by shape encoder-decoder networks.

Authors:  Max Blendowski; Nassim Bouteldja; Mattias P Heinrich
Journal:  Int J Comput Assist Radiol Surg       Date:  2019-11-18       Impact factor: 2.924

7.  Evaluate the Malignancy of Pulmonary Nodules Using the 3-D Deep Leaky Noisy-OR Network.

Authors:  Fangzhou Liao; Ming Liang; Zhe Li; Xiaolin Hu; Sen Song
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-02-14       Impact factor: 10.451

Review 8.  Overview of the Whole Heart and Heart Chamber Segmentation Methods.

Authors:  Marija Habijan; Danilo Babin; Irena Galić; Hrvoje Leventić; Krešimir Romić; Lazar Velicki; Aleksandra Pižurica
Journal:  Cardiovasc Eng Technol       Date:  2020-11-02       Impact factor: 2.495

9.  Deep Learning Techniques for Medical Image Segmentation: Achievements and Challenges.

Authors:  Mohammad Hesam Hesamian; Wenjing Jia; Xiangjian He; Paul Kennedy
Journal:  J Digit Imaging       Date:  2019-08       Impact factor: 4.056

Review 10.  Deep Learning for Cardiac Image Segmentation: A Review.

Authors:  Chen Chen; Chen Qin; Huaqi Qiu; Giacomo Tarroni; Jinming Duan; Wenjia Bai; Daniel Rueckert
Journal:  Front Cardiovasc Med       Date:  2020-03-05
  10 in total
  1 in total

1.  Multi-constraints based deep learning model for automated segmentation and diagnosis of coronary artery disease in X-ray angiographic images.

Authors:  Mona Algarni; Abdulkader Al-Rezqi; Faisal Saeed; Abdullah Alsaeedi; Fahad Ghabban
Journal:  PeerJ Comput Sci       Date:  2022-06-03
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.