Literature DB >> 35310182

Deep Neural Networks for Medical Image Segmentation.

Priyanka Malhotra1, Sheifali Gupta1, Deepika Koundal2, Atef Zaguia3, Wegayehu Enbeyle4.   

Abstract

Image segmentation is a branch of digital image processing which has numerous applications in the field of analysis of images, augmented reality, machine vision, and many more. The field of medical image analysis is growing and the segmentation of the organs, diseases, or abnormalities in medical images has become demanding. The segmentation of medical images helps in checking the growth of disease like tumour, controlling the dosage of medicine, and dosage of exposure to radiations. Medical image segmentation is really a challenging task due to the various artefacts present in the images. Recently, deep neural models have shown application in various image segmentation tasks. This significant growth is due to the achievements and high performance of the deep learning strategies. This work presents a review of the literature in the field of medical image segmentation employing deep convolutional neural networks. The paper examines the various widely used medical image datasets, the different metrics used for evaluating the segmentation tasks, and performances of different CNN based networks. In comparison to the existing review and survey papers, the present work also discusses the various challenges in the field of segmentation of medical images and different state-of-the-art solutions available in the literature.
Copyright © 2022 Priyanka Malhotra et al.

Entities:  

Mesh:

Year:  2022        PMID: 35310182      PMCID: PMC8930223          DOI: 10.1155/2022/9580991

Source DB:  PubMed          Journal:  J Healthc Eng        ISSN: 2040-2295            Impact factor:   2.682


1. Introduction

Image segmentation involves partitioning an input image into different segments with strong correlation with the region of interest (RoI) in the given image [1, 2]. The aim of medical image segmentation [3] is to represent a given input image in a meaningful form to study the anatomy, identify the region of interest (RoI), measure the volume of tissue to measure the size of tumor, and help in the deciding the dose of medicine, planning of treatment prior to applying radiation therapy, or calculating the radiation dose. Image segmentation helps in analysis of medical images by highlighting the region of interest. Segmentation techniques can be utilized for brain tumor boundary extraction in MRI images, cancer detection in biopsy images, mass segmentation in mammography, detection of borders in coronary angiograms, segmentation of pneumonia affected area in chest X-rays, etc. A number of medical image segmentation algorithms have been developed and are in demand as there is a shortage of expert manpower [4]. The earlier image segmentation models were based on traditional image processing approaches [3, 5] which include thresholding and edge-based and region-based techniques. In thresholding technique, pixels were allocated to different categories in accordance with the range of values where a particular pixel lies. In edge-based technique, a filter was applied to an image; it classifies the pixels as edged or nonedged in accordance with the filter output. In region-based segmentation methods, neighbouring pixels having similar values and the groups of pixels having dissimilar values were split. Medical image segmentation is difficult task due to various restrictions inflict by the medical image procurement procedure, the type of pathology, and different biological variations [6]. The analysis of medical images can be done by experts and there is a shortage of medical imaging experts [7]. In the last few years, deep learning networks had contributed to the development of newer image segmentation models with improvement in performance. The deep neural networks had achieved high accuracy rates on different popular datasets. The image segmentation techniques can be broadly classified as semantic segmentation and instance segmentation. Semantic segmentation can be considered as a problem of classifying pixels. In this segmentation technique, each pixel in the image is labelled to a certain class. Instance segmentation detects and delineates each object of interest present in the input image. The present work covers the recent literature in medical image segmentation. The work provides a review on different deep learning-based image segmentation models and explains their architecture. Many authors have worked on the review of medical image segmentation task. Table 1 gives the description of few review papers utilizing deep CNN in the field of medical image segmentation.
Table 1

Description of few review papers in medical image segmentation.

Ref.YearModels discussedPerformance metricsDatasetChallengesRemarks
[8]2017CNNNo coverageNo coverageChallenges with CNN coveredImage classification, object detection, segmentation, and registration mechanisms discussed
[9]2017Stacked autoencoder, deep belief network, and deep Boltzmann machineNo coverageNo coverageNo coverage
[10]2018CNN, R-CNNImage classification metrics discussed but segmentation metrics not coveredMedical image modalities coveredNo coverageAll areas of medical image analysis discussed
[11]2019CNN. FCN, U-Net, VNet, CRN, and RNNNo coverageCoveredChallenges and possible solutions discussed
[12]2020Supervised, weakly supervised models (RNN, U-Net)No coverageCoveredChallenges and possible solutions discussed----
[13]2021CNN, FCN, DeepLab, SegNet, U-Net, and VNetCoveredCoveredChallenges discussed but the solutions not discussed
OursCNN,FCN,R-CNN, fast R-CNN, faster R-CNN, mask R-CNN, U-Net, VNet, and DeepLabCoveredCoveredChallenges and possible state-of-the-art solutions discussedPaper provides extended coverage to the different deep neural networks for image segmentation
All the aforementioned survey literatures discuss the various deep neural networks. This survey paper does not only focus on summarizing the different deep learning approaches but also provides an insight into the different medical image datasets used for training deep neural networks and also explains the metrics used for evaluating the performance of a model. The present work also discusses the various challenges faced by DL based image segmentation models and their state-of-the-art solutions. The paper has several contributions which are as follows: Firstly, the present study provides an overview of the current state of the deep neural network structures utilized for medical image segmentation with their strengths and weaknesses Secondly, the paper describes the publicly available medical image segmentation datasets Thirdly, it presents the various performance metrics employed for evaluating the deep learning segmentation models Finally, the paper also gives an insight into the major challenges faced in the field of image segmentation and their state-of-the-art solutions The organization of the rest of the paper is given in Table 2 [14].
Table 2

Structure of the paper.

S. no.Main sectionSubsection
1IntroductionIntroduction and motivation literature review major contributions
2Deep neural network structuresArtificial neural network convolutional neural network encoder-decoder models regional convolutional network deepLab model comparison, limitations, and advantages/Table 3
3Application of deep neural network to medical image segmentationDeep learning-based system literature review on DNN based image segmentation models for different organs summary on deep learning-based medical image segmentation methods (Table 4)
4Medical image segmentation datasetsTypes and format of dataset different types of modalities summary of medical image segmentation datasets (Table 5)
5Evaluation metricsImportance of metrics popular image segmentation algorithm performance metrics
6Major challenges and state-of-the-art solutionsDataset challenges with DL model possible solution to the problems related to dataset and DL model
7Future directionMotivation for further study and future research
8ConclusionConcluding remarks

2. Deep Neural Network Structures

Deep learning is the most essential approach to artificial intelligence. Deep learning algorithm uses various layers to construct an artificial neural network. An artificial neural network (ANN) consists of [52] input layer, hidden layer(s), and output layer. The input layer of the network receives the signal, an output layer makes decision regarding the input, and between the input and output layers there are hidden layers which perform computations (shown in Figure 1). A deep neural network consists of many hidden layers between input and output layers.
Figure 1

Artificial neural network (ANN) model.

This section provides a review of different deep learning neural networks employed for image segmentation task. The different deep neural network structures generally employed for image segmentation can be grouped as shown in Figure 2.
Figure 2

Different types of deep neural network architectures for image segmentation.

2.1. Convolutional Neural Network

A convolutional neural network or CNN (see Figure 3) consists of a stack of three main neural layers: convolutional layer, pooling layer, and fully connected layer [52, 53]. Each layer has its own role. The convolution layer detects distinct features like edges or other visual elements in an image. Convolution layer performs mathematical operation of multiplication of local neighbours of an image pixel with kernels. CNN uses different kernels for convolving the given image for generating its feature maps. Pooling layer reduces the spatial (width, height) dimensions of the input data for the next layers of neural network. It does not change the depth of the data. This operation is called as subsampling. This size reduction decreases the computational requirements for upcoming layers. The fully connected layers perform high-level reasoning in NN. These layers integrate the various feature responses from the given input image so as to provide the final results.
Figure 3

Convolutional neural network architecture.

Different CNN models have been reported in the literature, including AlexNet [54], GoogleNet [55], VGG [56], Inception[57], SequeezeNet [58], and DenseNet [59]. Here, each network uses different number of convolutions and pooling layers with important process blocks inbetween them. The CNN models have been employed mostly for classification task. In [60], SqueezeNet and GoogleNet have been employed to classify brain MRI images into three different categories. The CNN segmentation models performance is limited by the following: The fully connected layers in CNN cannot manage different input sizes A convolutional neural network with a fully connected layer cannot be employed for object segmentation task, as the presence of number of objects of interest in the image segmentation task is not fixed, so the length of the output layer cannot be constant

2.1.1. Fully Convolutional Network

In fully convolutional network (FCN), only convolutional layers exist. The different existing in CNN architectures can be modified into FCN by converting the last fully connected layer of CNN into a fully convolutional layer. The model designed by [61] can output spatial segmentation map and can have dense pixel-wise prediction from the input image of full size instead of performing patch-wise predictions. The model uses skip connections which perform upsampling on feature maps from final layer and fuses it with the feature map of previous layers. The model thus produces a detailed segmentation in just one go. The conventional FCN model however has the following limitations [62]: It is not fast for real time inference and it does not consider the global context information efficiently. In FCN, the resolution of the feature maps generated at the output is downsampled due to propagation through alternate convolution and pooling layers. This results in low resolution predictions in FCN with fuzziness in object boundaries. An advanced FCN called ParseNet [63] has been also reported; it utilises global average pooling to attain global context. The approaches incorporating models such as conditional random fields and Markov random field into DL architecture have been also reported.

2.2. Encoder-Decoder Models

Encoder-decoder based models employ two-stage model to map data points from the input domain to the output domain. The encoder stage compresses the given input, x to latent space representation, while the decoder predicts the output from this representation. The different types of encoder-decoders based models generally employed for medical image segmentation are discussed as follows:

2.2.1. U-Net

U-Net model [64] has a downsampling and upsampling part. The downsampling section with FCN like architecture extracts features using 3 × 3 convolutions to capture context. The upsampling part performs deconvolution to decrease the number of computed feature maps. The feature maps generated by downsampling or contracting part are fed as input to upsampling part so as to avoid any loss of information. The symmetric upsampling part provides precise localization. The model generates a segmentation map which categorizes each pixel present in the image. The U-Net model offers the following advantages: U-Net model can perform efficient segmentation of images using limited number of labelled training images U-Net architecture combines the location information obtained from the downsampling path and the contextual information obtained from upsampling path to predict a fair segmentation map U-Net models also have few limitations, stated as follows: Input image size is limited to 572 × 572 In the middle layers of deeper UNET models, the learning generally slows down which causes the network to ignore the layers with abstract features The skip connections of the model impose a restrictive fusion scheme which causes accumulation of the same scale feature maps of the encoder and decoder networks To overcome these limitations, the different variants of U-Net architecture have been proposed in the literature: U-Net++ [65], Attention U-Net [66], and SD-UNet [67].

2.2.2. VNet

It is also an FCN-based model employed for medical image segmentation [68]. VNet architecture has two parts, compression and decompression network. The compression network comprises convolution layers at each stage with residual function. These convolution layers utilized volumetric kernels. The decompression network extracts feature and expands the spatial representation of low resolution feature maps. It gives two-channel probabilistic segmentation for both foreground and background regions.

2.3. Regional Convolutional Network

Regional convolutional network has been utilized for object detection and segmentation task. The R-CNN architecture presented in [69] generates region proposal network for bounding boxes using selective search process. These region proposals are then warped to standard squares and are forwarded to a CNN so as to generate feature vector map as output. The output dense layer consists of features extracted from the image and these features are then fed to classification algorithm so as to classify the objects lying within the region proposal network. The algorithm also predicts the offset values for increasing the precision level of the region proposal or bounding box. The processes performed in R-CNN architecture are shown in Figure 4. The use of basic RCN model is restricted due to the following:
Figure 4

R-CNN architecture.

It cannot be implemented in real time as it takes around 47 seconds to train the network for classification task of 2000 region proposals in a test image. The selective search algorithm is a predetermined algorithm. Therefore, learning does not take place at that stage. This could lead to the generation of unfavourable candidate region proposals. To overcome these drawbacks, different variants of R-CNN, fast R-CNN, faster R-CNN, and mask R-CNN have been proposed in the literature.

2.3.1. Fast R-CNN

In R-CNN, the proposed regions of image overlap and same CNN computations are carried again and again. The fast R-CNN reported by [70] is fed with an input image and a set of object proposals. The CNN then generates convolutional feature maps. After that, the ROI pooling layer reshapes each object proposal into a feature vector of fixed size. The feature vectors are sent to the last fully connected layers of the model. At the end, the computed ROI feature vector is fed to Softmax layer for predicting the class and offset values of the proposed region [71]. The fast R-CNN is slower due to the use of selective search algorithm.

2.3.2. Faster R-CNN

In R-CNN and fast R-CNN, the proposed regions were created using a process of selective search and were a slow process. So, in faster R-CNN architecture given by [72], a single convolutional network was deployed to carry out both region proposals and classification task. The model employs a region proposal network (RPN), passing the sliding window on the top of the entire CNN feature map. For each window, it outputs K different potential boundary boxes with their respective scores representing position of object. These bounding boxes fed to fast R-CNN generate the precise classification boxes.

2.3.3. Mask R-CNN

He et al. in [73] extended faster R-CNN to present Mask R-CNN for instance segmentation. The model can detect objects in a given image and generates a high-quality segmentation mask for each object in an image. It uses RoI-Align layer to conserve the exact spatial locations of the given image. The region proposal network (RPN) generated multiple RoIs using a CNN. The RoI-Align network generates multiple bounding boxes which are warped into fixed dimensions. The warped features computed in the previous step are fed to fully connected layer so as to create classification using softmax layer. The model has three output branches with one branch computing bounding box coordinates, second branch determining associated classes, and the last branch evaluating the binary mask for each RoI. The model trains all the branches jointly. The bounded boxes are improved by employing regression model. The mask classifier outputs a binary mask for each RoI.

2.4. DeepLab Model

DeepLab model employs pretrained CNN model ResNet-101/VGG-16 with atrous convolution to extract the features from an image [74]. The use of atrous convolutions gives the following benefits: It controls the resolution of feature responses in CNNs It converts image classification network into a dense feature extractor without the requirement of learning of any more parameters employs conditional random field (CRF) to produce fine segmented output The various variants of DeepLab have been proposed in the literature including DeepLabv1, DeepLabv2, DeepLabv3, and DeepLabv3+. In DeepLabv1 [75], the input image is passed through deep CNN layer with one or two atrous convolution layers (see Figure 5). This generates a coarse feature map. The feature map is then upsampled to the size of original image by using bilinear interpolation process. The interpolated data is applied to fully connect conditional random field to obtain the final segmented image.
Figure 5

DeepLab architecture.

In DeepLabv2 model, multiple atrous convolutions are applied to input feature map at different dilation rates. The outputs are fused together. Atrous spatial pyramid pooling (ASPP) segments the objects at different scales. The ResNet model used the atrous convolution with different rates of dilation. By using atrous convolution, information from large effective field can be captured with reduced number of parameters and computational complexity. DeepLabv3 [20] is an extension of DeepLabv2 with added image level features to the atrous spatial pyramid pooling (ASPP) module. It also utilises batch normalization so as to easily train the network. DeepLabv3+ model combines the ASPP module of DeepLabv3 with encoder and decoder structure. The model uses Xception model for feature extraction. The model also employed atrous and depth-wise separable convolution to compute faster. The decoder section merges the low- and the high-level features which correspond to the structural details and semantic information. DeepLabv3+ [76] consists of an encoding and a decoding module. The encoding path extracts the required information from the input image using atrous convolution and backbone network like MobileNetv2, PNASNet, ResNet, and Xception. The decoding path rebuilds the output with relevant dimensions using the information from the encoder path.

2.5. Comparison of Different Deep Learning-Based Segmentation Methods

The different deep neural networks discussed in the above sections are employed for different applications. Each model has its own advantages and limitations. Table 3 gives a brief comparison between different deep learning-based image segmentation algorithms.
Table 3

Comparison between different image segmentation algorithms.

Deep learning algorithmAlgorithm descriptionAdvantagesLimitations
CNNIt consists of three main neural layers, which are convolutional layers, pooling layers, and fully connected layers(a) It is simple(b) It involves feeding segments of an image as input to the network, which labels the pixels(a) It cannot manage different input sizes(b) Fixed size of output layer causes difficulty in segmentation task
FCNAll fully connected layers of CNN are replaced with the fully convolutional layersThe model outputs a spatial segmentation map instead of classification scoresIt is hard to train a FCN model to get good performance
U-NetIt combines the location information obtained from the downsampling path and the contextual information obtained from upsampling path to predict segmentation mapIt can perform efficient segmentation of images using limited number of labelled training images(a) Input image size is limited to 572 × 572. (b) The skip connections of the model impose a restrictive fusion scheme causing accumulation of the same scale feature maps of the network
VNetIt performs convolutions on each stage using volumetric kernels of size 5 × 5 × 5It can be applied to 3D data for segmentation
R-CNNIt uses selective search algorithm to extract 2000 regions from the image called region proposals(a) It predicts the presence of an object within the region proposals(b) It also predicts four offset values to increase the precision of the bounding box(a) Huge amount of time is needed to train network to classify 2000 region proposals per image (b) It cannot be implemented in real time(c) Selective search algorithm is a fixed algorithm
Fast R-CNNIt uses selective search algorithm which takes the whole image and region proposals as input in its CNN architecture in one forward propagationIt improves mean average precision (mAP) as compared to R-CNNThere is high computation time due to selective search region proposal generation algorithm
Faster R-CNNIt uses region proposal networkIt generates the bounding boxes of different shapes and sizesThere is lower computation time
Mask R-CNNIt gives three outputs for each object in the image: its class, bounding box coordinates, and object maska) It is simple and flexible approach b) It is current state-of-the-art technique for image segmentation taskThere is high training time
DeepLabv1a) It uses atrous convolution to extract the features from an imageb) It also uses conditional random field (CRF) to capture fine detailsa) There is high speed due to atrous convolution b) Localization of object boundaries improved by combining DCNNs and probabilistic graphical modelsUse of CRFs makes algorithm slow
DeepLabv2It usesatrous spatial pyramid pooling (ASPP) and applies multiple atrous convolutions with different sampling rates to the input feature map and fuses them togetherAtrous spatial pyramid pooling (ASPP) robustly segments objects at multiple scalesThere are challenges in capturing fine object boundaries
DeepLabv3It uses atrous separable convolution to capture sharper object boundariesIt can segment sharper targetsIt still needs more refinement for object boundaries
DeepLabv3+It extends DeepLabv3 by adding a decoder module to refine the segmentation results along the object boundariesThere is better segmentation performance as compared to deepLabv3It is a large model with number of parameters to train. So, while training on higher resolution images and batch sizes, it needs large GPU memory.

3. Applications of Deep Neural Networks in Medical Image Segmentation

Deep learning networks had contributed to various applications like image recognition and classification, object detection, image segmentation, and computer vision. A block diagram representing deep learning-based system is given in Figure 5. The first step in deep learning system consists of collecting data [77]. The collected data is then analyzed and preprocessed to be available in the format acceptable to the next block. The preprocessed data is further divided into training, validation, and testing dataset. A deep neural network-based model is selected and trained. The trained model is tested and evaluated. At the end, the analysis of the complete designed system is carried out. This basic layout of deep learning models (shown in Figure 6) is employed in various medical applications [78] including image segmentation. In image segmentation, the objects in image are subdivided. The aim of medical image segmentation is to identify region of interest (RoI) like tumor and lesion. The automatic segmentation of the medical images is really a difficult task because medical images are usually complex in nature due to presence of different artifacts, inhomogeneity in intensity, etc. Different deep learning models have been proposed in the literature. The choice of a particular deep learning model depends on various factors like body part to be segmented, imaging modality employed, and type of disease as different body parts and ailments have different requirements.
Figure 6

Basic layout of typical deep learning-based system.

A 2D and 3D CNN based fully automated framework have been presented by [15] to segment cardiac MR images into left and right ventricular cavities and myocardium. The authors in [18] designed a deep CNN with layers performing convolution, pooling, normalization, and others to segment brain tissues in MR images. Christ et al. in [30] presented a design in which two cascaded FCN were employed to segment liver and further the lesions within ROI were segmented. The final segmentation was produced by dense 3D conditional random field. Hamidian et al. in [25] converted 3D CNN with fixed field of view into a 3D FCN and generated the score map for the complete volume of CT images in one go. The authors employed the designed network for segmentation of pulmonary nodules in chest CT images. The authors concluded that by employing FCN speed of the network increases and there is fast generation of output scores. In [32], authors employed FCN for liver segmentation in CT images. In [27], authors proposed a fully convolution spatial and channel squeeze ad excitation module for segmentation of pneumothorax in chest X-ray images. Gordienko et al. [26] reported a U-Net based CNN for segmentation of lungs and bone shadow exclusion techniques on 2D CXRs images. Zhang et al. in [19] designed SDRes U-Net model, which embedded the dilated and separable convolution into residual U-Net architecture. The network was employed for segmenting brain tumor present in MR images. In [33], the authors proposed the use of Multi-ResUNet architecture for segmentation. The authors concluded that the use of Multi-ResUNet model generates better results in lesser number of training epochs as compared to the standard U-Net model. In [29], the authors segmented pneumothorax on CT images. The authors compared the performance of U-Net model with PSPNet. Ferreira [17] employed U-Net model to automatically segment heart in the short-axis DT-CMR images. The authors in [68] further designed a FCN network for segmenting 3D MRI volumes and employed a VNet based network to segment prostate in MRI images. Poudel et al. in [16] developed a recurrent fully convolutional network (RFCN) to detect and segment body organ. The given design ensures fully automatic segmentation of heart in cardiac MR images. The authors concluded that the RFCN architecture reduces the computational time, simplifies segmentation pipeline, and also enables real time application. Mulay et al. in [31] presented a nested edge detection and Mask R-CNN network for segmentation of liver in CT and MR images. The input images were firstly preprocessed by applying image enhancement so as to produce the sketch of the abdomen area. The network enhances input images for edge map. At last, the authors employed Mask R-CNN for segmenting liver from the edge maps. In [28], authors designed a CheXLocNet based on Mask R-CNN to segment area of pneumothorax from chest radiographs. In [22], authors suggested a recurrent neural network utilizing multidimensional LSTM. The authors arranged the computations in pyramidal fashion. The authors had shown that the PyraMiD-LSTM design can parallelize for 3D data and utilized the design for pixel-wise segmentation of MR images of brain. Table 4 summarizes the different DL based models employed for segmentation in medical images.
Table 4

Summary on deep learning-based medical image segmentation methods.

OrganSegmented areaModel utilizedDatasetModalityRemarks
CardiacCardiac, left, and right ventricular cavities and myocardium [15]2D/3d CNNADC2017Cardiac MR images
Heart [16]RFCNMICCAI2 2009 challenge datasetCardiac MR imagesRFCN reduces computational time, simplifies segmentation, and enables real time applications
Heart [17]U-NetDT-CMR imagesU-Net automated the DT-CMR postprocessing, supporting real time results
BrainBrain tissues [18]2D CNNMultimodal MR imagesModel performance increases by employing multiple modalities
Brain tumor [19]SDResU-NetMR imagesU-Net has generalization capability
Brain [20]Voxel-wise residual networkMRBrainSMRI
Brain [21]DNNISBI 2012 EMTEM
Pixel-wise brain segmentation [22]MD-LSTMMRBrainS13Brain MR imagesIt can parallelize for 3D data
Brain tumor core [23]FCN, U-NetMR imagesBounding box technique used
Brain tumor [24]DeepLabCT imagesDeepLab with conditional random fields produces high accuracy
LungsPulmonary nodules [25]3D FCNLIDC datasetChest CT imagesIncreased speed of screening
Lung segmentation [26]JU-Net based CNNJSRTCXR____
Pneumothorax segmentation [27]FC-DenseNet with SCSE modulePACSChest X-ray imagesSpatial weighted cross-entropy loss function improves precision at boundaries
Pneumothorax segmentation [28]Mask R-CNNSIIM-ACRChest X-ray imagesBounding box regression helps in improving classification
Pneumothorax segmentation [29]U-Net and PSPNetRoutine chest CT datasetChest CT images
LiverLiver and tumor segmentation [30]Cascaded FCNDIRCAD datasetCT and MRI imagesSeparate set of filters applied at each stage improves segmentation
Liver segmentation [31]HED-mask R-CNNCHAOS challengeCT and MR imagesHigh segmentation accuracy obtained
Liver segmentation [32]FCNMICCAI SLiver07 datasetCT images
Reproductive systemProstate [33]VNet3D MRI
Digestive systemPancreas [34]Recurrent NN (LSTM)NIH-CT-82, ufl-mri-79Abdominal CT and MRI imagesRNN performs better than HNN and UNET
BreastBreast masses [35]DBN + CRF/SSVMDDSM-BCRP, INbreast databasesMammogramsCRF model is faster than SSVM
EyesRetinal blood vessels [36]U-Net with modificationsDRIVE/STARERetinal imagesModification allows precise and faster segmentation of blood vessels
Retinal blood vessels [37]U-Net, LadderNetDRIVE/STARE/CHASERetinal images

ADC: Alzheimer Disease Center. MICCAI: Medical Image Computing and Computer Assisted Intervention. MRBrainS: MR brain segmentation. ISBI: IEEE International Symposium on Biomedical Imaging. LIDC: Lung Image Database Consortium. JSRT: Japanese society of radiological technology. PACS: Picture Archiving and Communication System. SIIM-ACR: Society for Imaging Informatics in Medicine-American College of Radiology. DIRCAD: 3D image reconstruction for comparison of algorithm database. CHAOS: combined (CT-MR) healthy abdominal organ segmentation challenge. DDSM: digital database for screening mammography. DRIVE: digital retinal images for vessel extraction. STARE: Structural Analysis of Retinal Dataset. CHASE: Combined Healthy Abdominal Organ Segmentation Challenge.

4. Medical Image Segmentation Datasets

Data is important in deep learning models. Deep learning models require large amount of data. The data plays an important role. It is difficult to collect the medical image data as there are data privacy rules governing collection and labelling of data and also it requires time-consuming explanation to be performed by experts [79]. The medical image datasets can be categorized into three different categories: 2D images, 2.5D images, and 3D images [2]. In 2D medical images, each information element in image is called pixels. In 3D medical images, each element is called voxel. 2.5D refers to RGB images. The 3D images are also sometimes represented as a sequential series of 2D slices. CT, MR, PET, and ultrasound pixels represent 3D voxels. The images may exist in JPEG, PNG, or DICOM format. The medical imaging is performed in different types of modalities [2], such as CT scan, ultrasound, MRI, mammograms, positron emission tomography (PET), and X-ray of different body parts. MR imaging allows achieving variable contrast image by employing different pulse sequences. MR imaging gives the internal structure of chest, liver, brain, pelvis, abdomen, etc. CT imaging uses X-rays to obtain the information about the structure and function of the body parts. CT imaging is used for diagnosis of disease in brain, abdomen, liver, pelvis, chest, spine, and CT based angiography. Figure 7 shows MRI and CT image of brain. Mammography is a technique that uses X-rays to capture the images of the internal structure of the breast. Chest X-rays (CXR) imaging is a photographic image depicting internal composition of chest which is produced by passing X-rays through the chest and these rays are being absorbed by different amounts of different components in the chest [31]. The important publicly available medical image datasets are summarized in Table 5.
Figure 7

(a) MR image of brain. (b) CT scan of brain [30].

Table 5

Summary of medical image segmentation datasets.

Organ examinedImaging modalityDataset nameDataset sizeDimensionsImage formatSegmented areaUsed in reference
BrainMRIBraTS1 20182853D (240 × 240 × 155)NIFTIGliomas tumor[38]
KneeMRISK110603D (0.39 × 0.39 × 1.0)NIFTIBones and cartilage[39]
OA1ZIB5073D (0.36 × 0.36 × 0.7)NIFTIBones and cartilage
EyesRetinal imagesDRIVE402D (768 × 584)JPEGRetinal vessels[40]
Retinal images retinal imagesPALM2 STARE1200 20-- 700 × 605JPEG JPEGLesions in pathological myopia blood vessels[41, 42]
Abdominal areaCTCHAOS340512 × 512DICOMLiver and vessels[43]
MRICHAOS1202D (256 × 256)DICOM
ChestChest X-raySIIM-ACR42D (1024 × 1024)DICOMPneumothorax[44]
Chest X-ray CTSCR5 SegTHOR247 602D (2048 × 2048) -----JPEG-----Lungs, heart, and clavicles segmentation of heart, aorta, trachea, and esophagus[45, 46]
KidneyCTKiTS6 19300-----NIFTIKidney tumor[47]
LiverWSI CTPAIP ----50 2013D 3DLiver cancer tumor[48]
CardiacMRI303DLeft atrium[49]
LungCT CTLuna7 16 DSB8888 13972D 2DMetaImageNodules nucleus segmentation[50, 51]

ACR: Society for Imaging Informatics in Medicine-American College of Radiology. BraTS: Brain Tumor Segmentation. CHAOS: Combined Healthy Abdominal Organ Segmentation Challenge. DSB: Data Science Bowl. KiTS: kidney tumor segmentation challenge. Luna: Lung Nodule Analysis. PALM: Pathologic Myopia Challenge. SCR: Segmentation in Chest Radiographs.

5. 5. Evaluation Metrics

A metric helps in evaluating the performance of any designed model. The metrics provide the accuracy of the designed model. The popular metrics employed for assessing effectiveness of any designed segmentation algorithm are represented in terms of the following [80]: True positive (TP) represents that both the actual data class and the class of predicted data are true. True negative (TN) represents that both the actual data class and the class of predicted data are false. False positive (FP) represents that the actual data class is false while the class of predicted data is true. False negative (FN) represents that the actual data class is true while the class of predicted data is false.

5.1. Precision

Precision is an evaluation metric that tells us about the proportion of input data cases that are reported to be true and represented in [81].

5.2. Recall

Recall represented in (2) gives the percentage of the total relevant results which had been correctly classified by the model [81].

5.3. F1 Score

F1 score tells about models accuracy as represented in the following equation. It is defined as the harmonic average of the precision and recall values [81]:

5.4. Pixel Accuracy

It gives the percentage of pixels in a given input image which are correctly classified by the model [82]:

5.5. Intersection over Union

Intersection over union (IoU) or Jaccard index [82] is a metric commonly used for checking the performance of image segmentation algorithm. It is the amount of intersecting area between the predicted image segment and the ground truth mask, divided by the total area of union between the predicted segment mask and the ground truth mask: where A represents ground truth. B represents predicted segmentation. Mean IoU is employed for evaluating modern segmentation algorithm. Mean IoU is the average of IoU for each class.

5.6. Dice Coefficient

It is defined in the following equation and termed as twice the amount of intersection area between the segment predicted and the ground truth divided by the total number of pixels in both the predicted segment and ground truth image [83]:

6. Major Challenges and State-of-the-Art Solutions

The medical image segmentation field has gained advantage from deep learning, but still it is a challenging task to employ deep neural networks due to the following.

6.1. Challenges with Dataset

The different challenges related to the dataset include the following: Limited Annotated Dataset. Deep learning network models require large amount of data. The data required for training is well annotated. The dataset plays an important role in various DL based medical procedures [84]. In medical image processing, the collection of large amounts of annotated medical images is tough [85]. Also, performing annotation on fresh medical images is tedious and expensive and requires expertise. Several large-scale datasets are publicly available. A list of few such datasets is provided in Table 2. There is still a need of more challenging datasets which can enable better training of DL models and are capable of handling dense objects. Typically, the existing 3D datasets [86] are not so large and few of them are synthetic, so more challenging datasets are required. The size of the existing medical image datasets can be increased by (a) application of image augmentation transformations like rotating image by different angles, flipping image vertically or horizontally, cropping, and shearing image. These augmentation techniques can boost the system performance. (b) The application of transfer learning from efficient models can provide solution to the problem of limited data [87]. (c) Finally comes synthesizing data collected from various sources [87]. Class Imbalance in Datasets. Class imbalance is intrinsic in various publicly available medical image datasets. A highly imbalanced data poses great difficulty in training DL model and makes model accuracy misleading, for example, in a patient data, where the disease is relatively rare and occurs only in 10% of patients screened. The overall designed model accuracy would be high as most of the patients do not have the disease and will reach local minima [88, 89]. The problem of class imbalance can be solved by (a) oversampling the data; the amount of oversampling depends on the extent of imbalance in the dataset. (b) Second, by changing the evaluation or performance metric, the problem of dataset imbalance can be handled. (c) Data augmentation techniques can be applied to create new data samples. (d) By combining minority classes, dataset class imbalance problem can also be handled. Sparse Annotations. Providing full annotation for 3D images is a time-consuming task and is not always possible. So, partial labelling of information slices in 3D images is done. It is really challenging to train DL model based on these sparsely annotated 3D images [85]. In case of sparsely annotated dataset, weighted loss function can be applied to the dataset. The weights for the unlabeled data in the available dataset are all set to zero, so as to learn only from the pixels which are labelled. Intensity Inhomogeneities. In pathology images, colour and intensity inhomogeneities [90] are common. Intensity inhomogeneities cause shading over the image. It is more specific in the segmentation of MR images. Also, the TEM images have brightness variations due to presence of nonuniform support films. The segmentation process becomes tedious due to these variations. For correcting intensity inhomogeneities [90], different algorithms are employed and many nonparametric techniques are proposed in the literature. Prefiltering operation can be employed before segmentation to remove inhomogeneities. Also, intensity inhomogeneities are taken care of by improvement in scanning devices. Complexities in Image Texture. In medical images, there may be different artifacts present during manipulation of images. The different sensors and electronic components used for capturing images create noise in the image [11, 91]. In the captured image, gray levels can be very close to each other and there may be weak image boundaries. There may be overlap in tissues and presence of irregularities like skin lines and hair in dermoscopic images. All these complexities cause difficulty in identification of region of interest in medical images. To remove different artefacts and noises from the image, different image enhancement techniques are used before segmentation. The image enhancement technique suppresses the noise in the image and preserves the integrity of the edges of the image.

6.2. Challenges with DL Models

The important challenging issues related to the training of DNN for robust segmentation of the medical images are as follows: Overfitting the Model. Overfitting of the model refers to the instance when the model learn the details and regularities in training dataset with high accuracy compared with the unprocessed data instance. It mainly occurs while training the model with a small size training data [9]. Overfitting can be handled [88] by (a) increasing the size of dataset by applying augmentation techniques. (b) Dropout techniques [92] also help in handling overfitting by discarding the output of some of the random set of network neurons during each iteration. Memory Efficient Models. Medical image segmentation models require large amount of memory [93]. In order to make these models compatible with certain devices like mobile phones, the models are required to be simplified. Simpler models and model compression techniques can reduce memory requirements for a DL model. Training Time. The training of deep neural network architecture needs time. In image segmentation, fast convergence of training time for deep NN is required. The solution to this problem is (a) application of batch normalization [93]. It refers to locating the pixel values around 0 by subtracting the pixel values from the mean value of the image. It is effective in providing fast convergence. (b) Also, adding pooling layers to reduce dimension of parameters can also provide faster convergence. Vanishing Gradient. Deep neural network faces the problem of vanishing gradient [94]. It occurs as the final gradient loss is not able to be backpropagated to earlier layers. The vanishing gradient problem is more pronounced in 3D models. There are several solutions to the problem of gradient vanishing. (a) By upscaling the intermediate hidden layer output using deconvolution and softmax [91], the auxiliary losses and the original loss of hidden layers are combined to strengthen the gradient value. (b) Also, by carefully initializing weights [95], for the network, we can combat the problem of vanishing gradient. Computational Complexity. Deep learning algorithm performing feature analysis needs to operate at a high level of computational efficiency. These algorithms need high performance computing devices and GPU [96]. Some of the top algorithms may require supercomputers for training the model, which may not be available. To combat these issues, the researcher has to consider the specific number of parameters to attain a limited level of accuracy.

7. Future Direction

The image segmentation techniques have come far away from manual image segmentation to automated segmentation using machine learning and deep learning approaches. The ML/DL based approaches can generate segmentation on large set of images. It helps in identification of meaningful objects and diagnosis of diseases in the images. The image segmentation techniques discussed in the paper can be explored by future researchers for application to various datasets. The future work may include a comparative study of the different existing deep learning models discussed in the paper on the publicly available datasets. Also, different combination of layers and classifiers can be explored to improve the accuracy of image segmentation model. There is still a requirement of an efficient solution to improve performance of image segmentation model. So, the various new deep learning model designs can be explored by future researchers.

8. Conclusion

Deep learning-based automated diagnosis of diseases from medical images had become the latest area of research. In the present work, we had summarized the most popular DL based models employed for segmentation of medical images with their underlined advantages and disadvantages. An overview of the different medical image dataset employed for segmentation of diseases and the various performance metrics utilized for evaluating the performance of image segmentation algorithm is also provided. The paper also investigates the different challenges faced in segmentation of medical images using the deep networks and discusses the different state-of-the-art solutions to overcome these challenges. With advances in technology, deep learning plays a very important role in segmentation of images. The different studies reviewed in Section 3 confirm that applications of deep neural networks in medical image segmentation task outperform the traditional image segmentation techniques. The present work will help the researchers in designing neural network architectures in the medical field for diagnosis of disease. Also, the researchers will become aware with the possible challenges in the field of deep learning-based medical image segmentation and the state-of-the-art solutions. This review paper provides the reference material and the valuable research in the area of medical image segmentation [97].
  28 in total

1.  Interaction in the segmentation of medical images: a survey.

Authors:  S D Olabarriaga; A W Smeulders
Journal:  Med Image Anal       Date:  2001-06       Impact factor: 8.545

2.  National survey to identify subspecialties at risk for physician shortages in Canadian academic radiology departments.

Authors:  Kai-Ling Ng; Jo Yazer; Mohammed Abdolell; Peter Brown
Journal:  Can Assoc Radiol J       Date:  2010-04-10       Impact factor: 2.248

3.  Automating in vivo cardiac diffusion tensor postprocessing with deep learning-based segmentation.

Authors:  Pedro F Ferreira; Raquel R Martin; Andrew D Scott; Zohya Khalique; Guang Yang; Sonia Nielles-Vallespin; Dudley J Pennell; David N Firmin
Journal:  Magn Reson Med       Date:  2020-04-23       Impact factor: 4.668

4.  Fully Convolutional Networks for Semantic Segmentation.

Authors:  Evan Shelhamer; Jonathan Long; Trevor Darrell
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-05-24       Impact factor: 6.226

5.  MultiResUNet : Rethinking the U-Net architecture for multimodal biomedical image segmentation.

Authors:  Nabil Ibtehaz; M Sohel Rahman
Journal:  Neural Netw       Date:  2019-09-04

6.  Automated medical image segmentation techniques.

Authors:  Neeraj Sharma; Lalit M Aggarwal
Journal:  J Med Phys       Date:  2010-01

7.  Interactive Medical Image Segmentation Using Deep Learning With Image-Specific Fine Tuning.

Authors:  Guotai Wang; Wenqi Li; Maria A Zuluaga; Rosalind Pratt; Premal A Patel; Michael Aertsen; Tom Doel; Anna L David; Jan Deprest; Sebastien Ourselin; Tom Vercauteren
Journal:  IEEE Trans Med Imaging       Date:  2018-07       Impact factor: 10.048

8.  Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation.

Authors:  Konstantinos Kamnitsas; Christian Ledig; Virginia F J Newcombe; Joanna P Simpson; Andrew D Kane; David K Menon; Daniel Rueckert; Ben Glocker
Journal:  Med Image Anal       Date:  2016-10-29       Impact factor: 8.545

9.  Automated segmentation and diagnosis of pneumothorax on chest X-rays with fully convolutional multi-scale ScSE-DenseNet: a retrospective study.

Authors:  Qingfeng Wang; Qiyu Liu; Guoting Luo; Zhiqin Liu; Jun Huang; Yuwei Zhou; Ying Zhou; Weiyun Xu; Jie-Zhi Cheng
Journal:  BMC Med Inform Decis Mak       Date:  2020-12-15       Impact factor: 2.796

10.  RA-UNet: A Hybrid Deep Attention-Aware Network to Extract Liver and Tumor in CT Scans.

Authors:  Qiangguo Jin; Zhaopeng Meng; Changming Sun; Hui Cui; Ran Su
Journal:  Front Bioeng Biotechnol       Date:  2020-12-23
View more
  2 in total

1.  Marker-controlled watershed with deep edge emphasis and optimized H-minima transform for automatic segmentation of densely cultivated 3D cell nuclei.

Authors:  Tuomas Kaseva; Bahareh Omidali; Eero Hippeläinen; Teemu Mäkelä; Ulla Wilppu; Alexey Sofiev; Arto Merivaara; Marjo Yliperttula; Sauli Savolainen; Eero Salli
Journal:  BMC Bioinformatics       Date:  2022-07-21       Impact factor: 3.307

2.  Computed Tomography Angiography and B-Mode Ultrasonography under Artificial Intelligence Plaque Segmentation Algorithm in the Perforator Localization for Preparation of Free Anterolateral Femoral Flap.

Authors:  Dan Shen; Xuehui Huang; Yinwei Huang; Dandan Zhou; Shasha Ye
Journal:  Contrast Media Mol Imaging       Date:  2022-09-28       Impact factor: 3.009

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.