| Literature DB >> 34336808 |
Houcheng Su1, Bin Lin1, Xiaoshuang Huang1, Jiao Li1, Kailin Jiang2, Xuliang Duan1.
Abstract
Colonoscopy is currently one of the main methods for the detection of rectal polyps, rectal cancer, and other diseases. With the rapid development of computer vision, deep learning-based semantic segmentation methods can be applied to the detection of medical lesions. However, it is challenging for current methods to detect polyps with high accuracy and real-time performance. To solve this problem, we propose a multi-branch feature fusion network (MBFFNet), which is an accurate real-time segmentation method for detecting colonoscopy. First, we use UNet as the basis of our model architecture and adopt stepwise sampling with channel multiplication to integrate features, which decreases the number of flops caused by stacking channels in UNet. Second, to improve model accuracy, we extract features from multiple layers and resize feature maps to the same size in different ways, such as up-sampling and pooling, to supplement information lost in multiplication-based up-sampling. Based on mIOU and Dice loss with cross entropy (CE), we conduct experiments in both CPU and GPU environments to verify the effectiveness of our model. The experimental results show that our proposed MBFFNet is superior to the selected baselines in terms of accuracy, model size, and flops. mIOU, F score, and Dice loss with CE reached 0.8952, 0.9450, and 0.1602, respectively, which were better than those of UNet, UNet++, and other networks. Compared with UNet, the flop count decreased by 73.2%, and the number of participants also decreased. The actual segmentation effect of MBFFNet is only lower than that of PraNet, the number of parameters is 78.27% of that of PraNet, and the flop count is 0.23% that of PraNet. In addition, experiments on other types of medical tasks show that MBFFNet has good potential for general application in medical image segmentation.Entities:
Keywords: MBFFNet; colonoscopy; fusion network; medical image segmentation; multi-branch feature
Year: 2021 PMID: 34336808 PMCID: PMC8317500 DOI: 10.3389/fbioe.2021.696251
Source DB: PubMed Journal: Front Bioeng Biotechnol ISSN: 2296-4185
FIGURE 1Pspnet structure.
FIGURE 2Unet structure.
FIGURE 3Multi-branch feature fusion network. The backbone initially extracts the features of images, and images with different subsampling multiples are superimposed in an hourglass image pyramid (subsample images have a size larger than 128, and up-sample images have a size smaller than 128, which is equal to 1 × 1 standard convolution for images with size larger than 128). To maximize the use of cross-channel and cross-resolution image branches, each branch is up-sampled and multiplied by the previous layer. Finally, the predicted image of the original image size is obtained.
Image enhancement setting parameters.
| Brightness | −0.2 to 0.2 |
| Zoom | −0.75 to 2 |
| Horizontal flip | 0.5 |
| Shift | 0.5 |
| Rotation | −0.5 to 0.5 |
| Channel transformation | 10 |
FIGURE 4Image enhancement renderings.
FIGURE 5Comparison of model effect. The red represents True Positive (TP), indicating that the predicted polyp area is actually a polyp area. Blue represents False Positive (FP), indicating that the predicted polyp area is actually a non-nuclear area. The green represents FN (False Negative), which means that the predicted polyp area is actually a polyp area.
Evaluation index of polyp segmentation mIOU, F-score, and Dice loss with CE.
| UNet ( | 0.8883 | 0.9354 | 0.1719 |
| LinkNet ( | 0.8711 | 0.9238 | 0.1911 |
| U2Net ( | 0.8950 | 0.9398 | 0.1528 |
| UNet++ ( | 0.8895 | 0.9364 | 0.1642 |
| UNet+++ ( | 0.8831 | 0.9312 | 0.1827 |
| PraNet ( | 0.9347 | 0.9612 | 0.1012 |
| PspNet ( | 0.8612 | 0.8972 | 0.2453 |
| Deeplabv3+ ( | 0.8452 | 0.8872 | 0.3214 |
| FCN8 ( | 0.8563 | 0.8945 | 0.2752 |
| DnlNet ( | 0.8657 | 0.9143 | 0.2064 |
| OcrNet ( | 0.8801 | 0.9210 | 0.1953 |
| PointRend ( | 0.8585 | 0.9074 | 0.2153 |
| MBFFNet | 0.8952 | 0.9450 | 0.1602 |
Analysis of the number of parameters and the number of calculation.
| UNet ( | 12 | 24.89 | 56.33 |
| LinkNet ( | 3 | 11.53 | 1.23 |
| U2Net ( | 18 | 96.25 | 40.24 |
| UNet++ ( | 20.5 | 36.16 | 135.24 |
| UNet+++ ( | 16 | 18.27 | 211.09 |
| PraNet ( | 13 | 16.16 | 20.37 |
| PspNet ( | 11.5 | 15.11 | 25.57 |
| Deeplabv3+ ( | 16.5 | 134.27 | 27.78 |
| FCN8 ( | 78.5 | 30.34 | 6390 |
| DnlNet ( | 15.5 | 50.13 | 50110 |
| OcrNet ( | 5 | 70.35 | 40530 |
| PointRend ( | 37.5 | 47.69 | 14640 |
| MBFFNet | 5.5 | 23.74 | 15.09 |
FIGURE 6Comparison between the accuracy of different models and flop count.
256 × 256 polyp image segmentation FPS.
| Unet ( | 4 | 3 | 45 | 21 |
| LinkNet ( | 19 | 16 | 115 | 88 |
| U2Net ( | 2 | 2 | 23 | 14 |
| UNet++ ( | 2 | 2 | 22 | 10 |
| UNet+++ ( | 2 | 1 | 16 | 8 |
| MBFFNet | 8 | 7 | 55 | 28 |
FPS segmentation of 64 × 64 polyp images.
| Unet ( | 20 | 19 | 152 | 90 |
| LinkNet ( | 84 | 68 | 138 | 141 |
| U2Net ( | 13 | 14 | 31 | 23 |
| UNet++ ( | 10 | 11 | 98 | 55 |
| UNet+++ ( | 9 | 9 | 90 | 68 |
| MBFFNet | 33 | 31 | 163 | 112 |
FIGURE 7Segmentation effect of liver lesions.
mIOU evaluation index of multi-class medical image segmentation.
| UNet ( | 0.9078 | 0.9691 | 0.8106 | 0.8020 | 0.9854 | 0.9641 |
| LinkNet ( | 0.8983 | 0.9166 | 0.7457 | 0.7711 | 0.9803 | 0.9279 |
| U2Net ( | 0.9126 | 0.9732 | 0.8070 | 0.8031 | 0.9854 | 0.9609 |
| UNet++ ( | 0.9108 | 0.9697 | 0.8083 | 0.8023 | 0.9849 | 0.9733 |
| UNet+++ ( | 0.9134 | 0.9715 | 0.8077 | 0.7995 | 0.9858 | 0.9707 |
| PraNet ( | 0.9453 | 0.9897 | 0.8762 | 0.8862 | 0.9903 | 0.9801 |
| PspNet ( | 0.7892 | 0.9568 | 0.5464 | 0.4891 | 0.9667 | 0.9551 |
| Deeplabv3+ ( | 0.7871 | 0.9661 | 0.5449 | 0.4890 | 0.9721 | 0.9623 |
| FCN8 ( | 0.9041 | 0.9815 | 0.6687 | 0.7172 | 0.9853 | 0.9645 |
| MBFFNet | 0.9132 | 0.9704 | 0.8127 | 0.8061 | 0.9884 | 0.9709 |
Dice loss with CE evaluation index for multi-class medical image segmentation.
| UNet ( | 0.1264 | 0.0548 | 0.2222 | 0.3310 | 0.0146 | 0.0377 |
| LinkNet ( | 0.1423 | 0.0714 | 0.3258 | 0.3822 | 0.0215 | 0.0779 |
| U2Net ( | 0.1191 | 0.0471 | 0.2344 | 0.3327 | 0.0148 | 0.0411 |
| UNet++ ( | 0.1213 | 0.0547 | 0.2248 | 0.3202 | 0.0151 | 0.0276 |
| UNet+++ ( | 0.1199 | 0.0514 | 0.2351 | 0.3331 | 0.0144 | 0.0307 |
| PraNet ( | 0.0921 | 0.0321 | 0.1453 | 0.2145 | 0.0101 | 0.0219 |
| PspNet ( | 0.3046 | 0.0743 | 0.6231 | 0.8119 | 0.0351 | 0.0801 |
| Deeplabv3+ ( | 0.3068 | 0.0565 | 0.6285 | 0.8169 | 0.0290 | 0.0792 |
| FCN8 ( | 0.1318 | 0.0350 | 0.4485 | 0.4874 | 0.0145 | 0.0407 |
| MBFFNet | 0.1303 | 0.0638 | 0.2545 | 0.3254 | 0.0130 | 0.0303 |
FIGURE 8Comparison between the accuracy of different models and Flop count.
Multi-class medical image segmentation F-score evaluation index.
| UNet ( | 0.9502 | 0.9842 | 0.8872 | 0.8864 | 0.9926 | 0.9815 |
| LinkNet ( | 0.9446 | 0.9803 | 0.8379 | 0.8657 | 0.9900 | 0.9616 |
| U2Net ( | 0.9527 | 0.9864 | 0.8846 | 0.8873 | 0.9926 | 0.9798 |
| UNet++ ( | 0.9519 | 0.9845 | 0.8855 | 0.8865 | 0.9923 | 0.9863 |
| UNet+++ ( | 0.9532 | 0.9855 | 0.8851 | 0.8846 | 0.9928 | 0.9850 |
| PraNet ( | 0.9732 | 0.9912 | 0.9213 | 0.9274 | 0.9912 | 0.9883 |
| PspNet ( | 0.8729 | 0.9778 | 0.6350 | 0.6223 | 0.9828 | 0.9712 |
| Deeplabv3+ ( | 0.8714 | 0.9827 | 0.6337 | 0.6170 | 0.9857 | 0.9653 |
| FCN8 ( | 0.9478 | 0.9906 | 0.7722 | 0.8278 | 0.9925 | 0.9671 |
| MBFFNet | 0.9604 | 0.9839 | 0.8895 | 0.8928 | 0.9926 | 0.9851 |