| Literature DB >> 35602289 |
Abstract
Medical image interpretation is an essential task for the correct diagnosis of many diseases. Pathologists, radiologists, physicians, and researchers rely heavily on medical images to perform diagnoses and develop new treatments. However, manual medical image analysis is tedious and time consuming, making it necessary to identify accurate automated methods. Deep learning-especially supervised deep learning-shows impressive performance in the classification, detection, and segmentation of medical images and has proven comparable in ability to humans. This survey aims to help researchers and practitioners of medical image analysis understand the key concepts and algorithms of supervised learning techniques. Specifically, this survey explains the performance metrics of supervised learning methods; summarizes the available medical datasets; studies the state-of-the-art supervised learning architectures for medical imaging processing, including convolutional neural networks (CNNs) and their corresponding algorithms, region-based CNNs and their variants, fully convolutional networks (FCN) and U-Net architecture; and discusses the trends and challenges in the application of supervised learning methods to medical image analysis. Supervised learning requires large labeled datasets to learn and achieve good performance, and data augmentation, transfer learning, and dropout techniques have widely been employed in medical image processing to overcome the lack of such datasets.Entities:
Keywords: Convolutional neural network (CNN); Deep learning; FCN; Fast R-CNN; Faster R-CNN; Mask R-CNN; Medical image processing; Supervised learning; U-Net
Year: 2022 PMID: 35602289 PMCID: PMC9112642 DOI: 10.1007/s42979-022-01166-1
Source DB: PubMed Journal: SN Comput Sci ISSN: 2661-8907
Fig. 1Use of deep learning techniques for essential medical tasks: a classification, b detection, and c segmentation [16]
Fig. 2ROC curves of different deep learning algorithms’ predictions [25]
Performance metrics for various image processing tasks, including classification, detection, and segmentation
| Classification | Detection | Segmentation |
|---|---|---|
AP113
Where AP114
Where mAP5
|
1TP = True Positive, TN = True Negative, FP = False Positive, and FN = False Negative
2IoU = Intersection over Union, B = Prediction bounding box, and B = Ground truth bounding box
3AP11 = 11-point interpolation, Pinterp(r) = precision replaced with the maximum precision whose recall value is ≥ r where P(r˜) is the precision at each recall level r
4All point interpolation where n is the number of all recall levels, and Pinterp(r) is the precision replaced with the maximum precision whose recall value is ≥ r
5Mean Average Precision where N is the number of classes and (AP) is AP of i class
6|G ∩ S| is the area of intersection between a set of pixels for ground truth (G) and a set of pixel for prediction segmentation (S), and |G ∪ S| is the area of union between (G) and (S)
7GtoS is the average Hausdorff distance from ground truth to segmentation where StoG is the average Hausdorff distance from segmentation to ground truth
Fig. 3CNN architecture
Fig. 4Convolution layer extracts features from an input image by calculating the dot product of a 3 × 3 filter and an 8 × 8 input image to produce a 6 × 6 feature map. The filter represents a vertical line feature
Fig. 5Batch Normalization’s algorithm [40]
Fig. 6Architectures of VGG-16 and VGG-19 [51]
Fig. 7GoogleNet inception module [37]. a Naïve inception model. b Inception module with dimension reduction
Fig. 8Residual block architecture [44]. a Residual block. b Residual block with dimension reduction
Fig. 9The difference between: a traditional learning and b transfer learning [53]
Fig. 10R-CNN architecture [55]
Fig. 11Fast R-CNN architecture [55]
Fig. 12Faster R-CNN architecture [55]
Fig. 13Mask R-CNN architecture [55]
Fig. 14FCN architecture [58]
Fig. 15U-Net architecture [60]
Medical image datasets for various diseases
| Datasets | Type of image | Number of cases | Number of images | Download links |
|---|---|---|---|---|
| IXI dataset [ | MRI | – | 600 | |
| OASIS-3 [ | MRI | 1000 + | 2000 + | |
| Multimodal Brain Tumor Image Segmentation Benchmark [ | MRI | 2000 | 8000 | Multimodal Brain Tumor Segmentation Challenge 2020: Data | CBICA | Perelman School of Medicine at the University of Pennsylvania (upenn.edu) |
| BancoWeb LAPIMO database [ | Mammography | 320 | 1473 | |
| Breast Cancer Histopathological Database (BreakHis) [ | Microscopic | 82 | 9109 | |
| Breast Cancer Wisconsin Dataset [ | Mammography | – | 569 | |
| INbreast dataset [ | Mammography | 115 | 410 | |
| Mammographic Image Analysis Society (MIAS) [ | Mammography | 161 | 322 | |
| Digital Database for Screening Mammography (DDSM) [ | Mammography | 2620 | 10,480 | |
| Intel and MobileODT on Kaggle [ | Cervical cancer screening | – | 1480 | |
| ISBI Challenge Database [ | Cytology image | – | 16 real EDF image and 945 synthetic images | |
| PAP Smear Benchmark Database [ | Microscopic | – | 917 | |
| SIPaKMeD Database [ | Pap smear slides | – | 4049 | |
| DiaretDB0 [ | Fundus | – | 130 | |
| DiaretDB1 [ | Fundus | – | 89 | |
| E-Ophtha [ | Fundus | – | 381 | |
| Kaggle DR Challenge [ | Fundus | – | 88,702 | |
| Messidor [ | Fundus | – | 1200 | |
| Messidor-2 [ | Fundus | – | 1784 | |
| Lung Image Database Consortium and Image Database (LIDC/IDRI) [ | CT | 1018 | 7371 | |
| National Lung Screening Trial (NLST) [ | CT | 53,454 | 75,000 | |
| DermNet NZ [ | Clinical | – | 20,000 + | |
| Dermofit Image Library [ | Dermoscopic | – | 1300 | |
| ISIC 2019 [ | Dermoscopic | – | 25,331 | |
| PH2 Dataset [ | Dermoscopic | – | 200 | |
| Interactive Atlas of Dermoscopy (EDRA) | Dermoscopic | 1000 + | 2000 + | |
| International Skin Imaging Collaboration (ISIC 2020) [ | Dermoscopic | 2000 | 33,126 | |
Medical image processing task challenges
| Medical Image processing task challenges | Medical dataset |
|---|---|
| Brain tumors segmentation in multimodal magnetic resonance imaging (MRI) scans [ | Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) |
| Women’s cervix type classification based on images [ | Intel and MobileODT on Kaggle |
| Detection of Diabetic Retinopathy (DR) by Kaggle [ | Kaggle DR Challenge |
| Lung nodule detection challenges [ | Lung Image Database Consortium and Image Database (LIDC/IDRI) |
| International Skin Imaging Collaboration (ISIC) challenges for Skin lesion detection [ | ISIC 2016, 2017, 2018, 2019, 2020 dataset |
| The Medical Segmentation Decathlon (MSD) [ | MSD dataset |
| The COVID-19–20 Lung CT Lesion Segmentation Challenge [ | CT images in COVID-19 and COVID-19-AR datasets |
Fig. 16Distribution of supervised learning techniques for various medical image tasks
Fig. 17Distribution of supervised learning techniques for various diseases
Fig. 18Accuracies for transferred features from multiple and individual networks
Fig. 19Lowest and highest accuracy of supervised learning networks
Fig. 20Distributions of papers discussing techniques to overcome the overfitting problem over the years