| Literature DB >> 34072714 |
Bingjiang Qiu1,2,3, Jiapan Guo2,3, Joep Kraeima1,4, Haye Hendrik Glas1,4, Weichuan Zhang5,6, Ronald J H Borra7, Max Johannes Hendrikus Witjes1,4, Peter M A van Ooijen2,3.
Abstract
PURPOSE: Classic encoder-decoder-based convolutional neural network (EDCNN) approaches cannot accurately segment detailed anatomical structures of the mandible in computed tomography (CT), for instance, condyles and coronoids of the mandible, which are often affected by noise and metal artifacts. The main reason is that EDCNN approaches ignore the anatomical connectivity of the organs. In this paper, we propose a novel CNN-based 3D mandible segmentation approach that has the ability to accurately segment detailed anatomical structures.Entities:
Keywords: 3D virtual surgical planning (3D VSP); accurate mandible segmentation; convolutional neural network; oral and maxillofacial surgery
Year: 2021 PMID: 34072714 PMCID: PMC8229770 DOI: 10.3390/jpm11060492
Source DB: PubMed Journal: J Pers Med ISSN: 2075-4426
Figure 1Illustration of three prevalent strategies for the feeding of input data to the classic encoder–decoder-based convolutional neural networks. (a) Use of 2D EDCNN network for 2D slice-based object segmentation. (b) Use of 2D EDCNN for the image segmentation based on several adjacent slices from the volumetric data. (c) Use of 3D EDCNN based on 3D patches cropped from the complete volumetric data.
Figure 2The overall graphic scheme of the proposed methods. The architecture of RCNNSeg with two components: (a) the RCNNSeg and its loss drawn with recurrent connections; (b) the same seen as a time-unfolded computational graph, where each node is now associated with one particular time instance.
Quantitative comparison of segmentation performance in the UMCG dataset between the proposed RCNNSeg and the classic EDCNNs. The values in the square brackets indicate the standard deviation of the corresponding measurements. We mark in bold the best performance in each metric.
| 2D U-Net | 95.95 [±2.24] | 0.3615 [±0.3366] | 4.0145 [±4.6487] |
| 2.5D U-Net | 96.34 [±1.99] | 0.4053 [±0.7565] |
|
| 3D U-Net | 77.73 [±6.74] | 17.2808 [±6.0045] | 133.7464 [±22.8779] |
| RUnetSeg |
|
| 2.3975 [±4.6051] |
| 2D SegU-Net | 96.30 [±2.06] | 0.2794 [±0.2447] | 3.7958 [±4.3662] |
| 2.5D SegU-Net | 96.69 [±2.12] | 0.4210 [±0.6111] | 4.9574 [±7.5637] |
| 3D SegU-Net | 81.88 [±7.14] | 19.0109 [±6.7765] | 137.2283 [±17.1781] |
| RSegUnetSeg |
|
|
|
| 2D Att U-Net | 94.21 [±3.34] | 0.6929 [±0.8370] | 5.1368 [±3.2194] |
| 2.5D Att U-Net | 93.87 [±2.89] | 0.5188 [±0.3327] | 4.9223 [±4.6204] |
| 3D Att U-Net | 83.92 [±5.43] | 16.2428 [±3.9300] | 124.1773 [±14.2461] |
| RAttUnetSeg |
|
|
|
Figure 33D view of a case from the UMCG head and neck dataset. (a–m) Ground truth, 2D U-Net, 2.5D U-Net, 3D U-Net, RUnetSeg, 2D SegU-Net, 2.5D SegU-Net, 3D U-Net, RSegUnetSeg, 2D Att U-Net, 2.5D Att U-Net, 3D Att U-Net, and RAttUnetSeg. We use cyan to indicate the correctly segmented mandible compared to the ground truth. Pink represents the regions that were missed by the algorithms, while yellow indicates the segmented regions that are not the mandible. The red circles indicate the coronoids of the mandibles that are often missed by the traditional EDCNNs, and the yellow circles indicate parts of the mandible body.
Figure 4Examples of the automatic segmentation of mandibles in the UMCG dataset. (a) Ground truth segmentation on the three examples. (b–m) Segmentation results obtained on the example slices from 2D U-Net, 2.5D U-Net, 3D U-Net, RUnetSeg, 2D SegU-Net, 2.5D SegU-Net, 3D SegU-Net, RSegUnetSeg, 2D Att U-Net, 2.5D Att U-Net, 3D Att U-Net, and RAttUnetSeg. Cyan indicates the correctly segmented mandible compared to the ground truth. Pink represents the regions that were missed by the algorithms, while yellow indicates the segmented non-mandible regions.
Quantitative comparison of the segmentation performance between the proposed RCNNSeg-based approaches and the EDCNN-based methods on the PDDCA dataset. The values in the square brackets indicate the standard deviation of the corresponding measurements. We mark in bold the best performance in each metric.
| 2D U-Net | 94.15 [±1.31] | 0.1827 [±0.0915] | 2.0547 [±1.4431] |
| 2.5D U-Net | 94.19 [±1.25] | 0.1915 [±0.0669] | 1.7512 [±0.6539] |
| 3D U-Net | 91.85 [±5.32] | 3.7577 [±6.1869] | 36.9138 [±66.9059] |
| RUnetSeg |
|
|
|
| 2D SegU-Net | 94.69 [±1.33] | 0.1765 [±0.0671] | 1.5067 [±0.6938] |
| 2.5D SegU-Net | 94.76 [±1.20] | 0.1532 [±0.0622] | 1.6856 [±0.6426] |
| 3D SegU-Net | 93.08 [±2.80] | 2.4289 [±5.8637] | 24.1133 [±62.1808] |
| RSegUnetSeg |
|
|
|
| 2D Att U-Net | 92.99 [±1.25] | 0.2924 [±0.2523] | 3.1848 [±4.0571] |
| 2.5D Att U-Net | 92.75 [±1.34] | 0.2502 [±0.0887] | 2.1815 [±1.0656] |
| 3D Att U-Net | 90.14 [±8.50] | 6.3894 [±11.7528] | 54.2182 [±72.3141] |
| RAttUnetSeg |
|
|
|
Comparison of segmentation performance between the state-of-the-art methods and our proposed RCNNSeg approach; bold font indicates the best three performers for each measurement.
| Multi-atlas [ | 91.7 [±2.34] | - | 2.4887 [±0.7610] |
| AAM [ | 92.67 [±1] | - | 1.9767 [±0.5945] |
| ASM [ | 88.13 [±5.55] | - | 2.832 [±1.1772] |
| CNN [ | 89.5 [±3.6] | - | - |
| NLGM [ | 93.08 [±2.36] | - | - |
| AnatomyNet [ | 92.51 [±2] | - | 6.28 [±2.21] |
| FCNN [ | 92.07 [±1.15] | 0.51 [±0.12] | 2.01 [±0.83] |
| FCNN+SRM [ | 93.6 [±1.21] | 0.371 [±0.11] | 1.5 [±0.32] |
| CNN+BD [ |
| 0.29 [±0.03] | - |
| HVR [ | 94.4 [± 1.3] | 0.43 [± 0.12] | - |
| Cascade 3D U-Net [ | 93 [±1.9] | - |
|
| Multi-view [ | 94.1 [±0.7] | 0.28 [±0.14] | - |
| RUnetSeg |
|
|
|
| RSegUnetSeg |
|
|
|
| RAttUnetSeg | 93.87 [±1.29] |
| 1.6397 [±0.6219] |