Literature DB >> 32617333

A review of the application of deep learning in medical image classification and segmentation.

Lei Cai1, Jingyang Gao1, Di Zhao2.   

Abstract

Big medical data mainly include electronic health record data, medical image data, gene information data, etc. Among them, medical image data account for the vast majority of medical data at this stage. How to apply big medical data to clinical practice? This is an issue of great concern to medical and computer researchers, and intelligent imaging and deep learning provide a good answer. This review introduces the application of intelligent imaging and deep learning in the field of big data analysis and early diagnosis of diseases, combining the latest research progress of big data analysis of medical images and the work of our team in the field of big data analysis of medical imagec, especially the classification and segmentation of medical images. 2020 Annals of Translational Medicine. All rights reserved.

Keywords:  Big medical data; classification; deep learning; object detection; segmentation

Year:  2020        PMID: 32617333      PMCID: PMC7327346          DOI: 10.21037/atm.2020.02.44

Source DB:  PubMed          Journal:  Ann Transl Med        ISSN: 2305-5839


Introduction

Since 2006, deep learning has emerged as a branch of the machine learning field in people’s field of vision. It is a method of data processing using multiple layers of complex structures or multiple processing layers composed of multiple nonlinear transformations (1). In recent years, deep learning has made breakthroughs in the fields of computer vision, speech recognition, natural language processing, audio recognition and bioinformatics (2). Deep learning has been praised as one of the top ten technological breakthroughs since 2013 due to its considerable application prospects in data analysis. The deep learning method simulates the human neural network. By combining multiple nonlinear processing layers, the original data is abstracted layer by layer, and different levels of abstract features are obtained from the data and used for target detection, classification or segmentation. The advantage of deep learning is to replace the manual acquisition feature with unsupervised or semi-supervised feature learning and hierarchical feature extraction efficient algorithms (3). Medical care is about the health of people. At present, the amount of medical data is huge, but it is crucial to make good use of this huge medical data to contribute to the medical industry. Although the amount of medical data is huge, there are still many problems: medical data is diverse, including maps, texts, videos, magnets, etc.; due to different equipment used, the quality of data varies greatly; data presents fluctuating characteristics, over time and specific events change; due to differences in individuals, the law of the disease has no universal applicability (4). There are many factors that cannot be dealt with in the existence of these problems. Medical imaging is a very important part of medical data. This paper first introduces the application of deep learning algorithms in medical image analysis, expounds the techniques of deep learning classification and segmentation, and introduces the more classic and current mainstream network models. Then we detailed the application of deep learning in the classification and segmentation of medical images, including fundus, CT/MRI tomography, ultrasound and digital pathology based on different imaging techniques. Finally, it discusses the possible problems and predicts the development prospects of deep learning medical imaging analysis.

Deep learning architectures

Deep learning algorithms

Deep learning has developed into a hot research field, and there are dozens of algorithms, each with its own advantages and disadvantages. These algorithms cover almost all aspects of our image processing, which mainly focus on classification, segmentation. is an overview of some typical network structures in these areas.
Figure 1

The typical network structures of the deep learning.

The typical network structures of the deep learning.

Classification

Using deep learning for image classification is earliest rise and it also a subject of prosperity. Among them, convolutional neural network (CNN) is the most widely structure. Since Krizhevsky et al. proposed AlexNet based on deep learning model CNN in 2012 (5), which won the championship in the ImageNet image classification of that year, deep learning began to explode. In 2013, Lin et al. proposed the network in network (NIN) structure, which uses global average pooling to reduce the risk of overfitting (6). In 2014, GoogLeNet and VGGNet both improved the accuracy on the ImageNet dataset (7,8). GoogLeNet has further developed the v2, v3 and v4 versions to improve performance (9-11). For the shortcomings of CNN on the input size fixed requirements, He et al. proposed spatial pyramid pooling (SPP) model to enhance the robustness of the input data (12). With the deepening of the deep learning model, He et al. proposed the residual network ResNet for the problem of model degradation that may occur, and continue to advance the deep learning technology (13). Take AlexNet as an example. In 2012, the AlexNet adopted an 8-layer network structure consisting of five convolutional layers and three fully connected layers. After each convolution in five convolutional layers, a maximum pooling performed to reduce the amount of data. AlexNet accepts 227×227 pixels’ input data. After five rounds of convolution and pooling operations, the 6×6×256 feature matrix finally sent to the fully connected layer. The sixth layer of the fully connected layer sets up 4,096 convolution kernels, and the linear feature value of 4,096 size obtained by the dropout operation. After the last two layers, we get 1,000 float types output data, which is the final prediction result. AlexNet’s error rate in ImageNet was 15.3%, which was much higher than the 26.2% in second place. At the same time, its activation function is not sigmoid but adopted ReLU, and proved that the ReLU function is more effective. VGG16 first proposed by VGG Group of Oxford University. Compared with AlexNet, it uses several consecutive 3×3 kernels instead of the larger convolution kernel in AlexNet like 11×11 and 5×5. For a given receptive field range, the effect of using several small convolution kernels is better than using a larger convolution kernel, because the multi-layer nonlinear layer can increase the network depth to ensure more complex patterns are learned, and the computational cost is also more small. GoogLeNet, which launched in the same year as VGGNet, also achieved good results. Compared to VGGNet, GoogLeNet designed a module called inception. It’s a dense structure with a small number of convolution kernels of each size, and use 1×1 convolutional layer to reduce the amount of computation.

Segmentation

Semantic segmentation is an important research field of deep learning. With the rapid development of deep learning technology, excellent semantic segmentation neural networks emerge in large numbers and continuously become state-of-the-art in various segmentation competitions. Since CNN’s success in the classification field, people started to try CNN for image segmentation. Although CNN can accept images of any size as input, CNN will lose some details while pooling for extracting features, and it will lose the space information of input image due to the fully connected layers at the end of the network. So it’s difficult for CNN to pinpoint which category certain pixels belong to. With the development of deep learning technology, some segmentation networks based on convolution structure are derived. The fully convolutional network (FCN) (14) proposed by Long et al. is the originator of the semantic segmentation networks. It replaces the fully connected layers of the classification network VGG16 with convolutional layers and retains the spatial information of the feature map and achieves pixel-level classification. Finally, FCN uses the deconvolution and fusing feature maps to restore the image, and provides the segmentation result of each pixel by softmax. Since the fully connected layer with dense connections is replaced by the convolutional layer which is locally connecting and weights sharing, the FCN greatly reduces the parameters that need to be trained. The performance of the FCN on the Pascal VOC 2012 datasets (15) has increased by 20% compared to the previous method, reaching 62.2% of the mIOU. U-Net (16) was proposed by Olaf based on FCN, and has been widely used in medical imaging. Based on the idea of FCN deconvolution to restore image size and feature, U-Net constructs the encoder-decoder structure in the field of semantic segmentation. The encoder gradually reduces the spatial dimension by continuously merging the layers to extract feature information, and the decoder portion gradually restores the target detail and the spatial dimension according to the feature information. Among them, the step of the encoder gradually reducing the image size is called downsampling, and the step of the decoder gradually reducing the image details and size is called upsampling. Different from the fusion operation of the direct addition feature when the FCN is upsampled, the U-Net upsampling process first uses the concatenate operation to splicing the feature maps before the up-sampling of the encoder and the downsampling of the decoder. After concatenation the feature map is deconvolved. Different from the conventional convolution, pooling, and other operations, this strategy of directly utilizing shallow features is called skip connection. U-Net adopts the skip connection strategy of splicing to make full use of the features of the downsampling part of the encoder to be used for upsampling. To achieve a more refined reduction, this strategy is applied to shallow feature information of all scales to achieve a better reduction effect. SegNet (17) is a depth semantic segmentation network designed by Cambridge to solve autonomous driving or intelligent robots, which is also based on the encoder-decoder structure. SegNet’s encoder and decoder each have 13 convolution layers. The convolutional layer of the encoder corresponds to the first 13 convolutional layers of VGG16. The upsampling part of the decoder uses UnPooling. SegNet records the element position information of the maximum pooling operation when the encoder is downsampled, and restores the image according to the position information when sampling on the decoder. SegNet with this strategy does not require learning when upsampling, and SegNet training is more accurate and faster than FCN. In order to fuse the context information under multi-scale at the same level, PSPNet (18) proposes a pooled pyramid structure, which realizes image segmentation in which the target environment can be understood, and solves the problem that FCN cannot effectively deal with, the relationship problem between global information and scenes. Its pooled pyramid structure can aggregate context information of different regions, thereby improving the ability to obtain global information.

Deep learning development framework

While the deep learning technology is developing in theory, the software development framework based on deep learning theory is also booming.

Convolutional architecture for fast feature embedding (Caffe)

Caffe was born in Berkeley, California and now hosted by BVLC. Caffe features high-performance, seamless switching between CPU and GPU modes, and cross-platform support for Windows, Linux and Mac. Caffe has three basic atomic structures of Blobs, Layers and Nets, and its programming framework is implemented under these three atoms. It highly abstracts the structure of the deep neural network in terms of the “Layer”, and significantly optimizes the execution efficiency through some elaborate design, and it has flexibility based on maintaining efficient implementation.

Tensorflow

TensorFlow is an open source software library that uses data flow diagrams for numerical calculations. Google officially opened the computing framework TensorFlow on November 9, 2015, and officially released Google TensorFlow version 1.0 in 2017, marking its official use in the production environment. The TensorFlow calculation framework can well support various algorithms for deep learning such as CNN, RNN and LSTM, but its application is not limited to deep learning, but also supports the construction of general machine learning. TensorFlow’s components are excellent, and it provides powerful visualization capabilities through TensorBoard, which can generate very powerful visual representations of real-world network topologies and performance. At the same time, it supports heterogeneous distributed computing, which can run on multiple GPUs at the same time, and can automatically run the model on different platforms. Because TensorFlow developed in C++, it has high-performance.

PyTorch

Pytorch is the python version of torch, a neural network framework that is open sourced by Facebook and specifically targeted at GPU-accelerated deep neural network programming. Unlike Tensorflow’s static calculation graph, Pytorch’s calculation graph is dynamic, and the calculation graph can be changed in real-time according to the calculation needs. In January 2017, the Facebook Artificial Intelligence Institute (FAIR) team opened up PyTorch on GitHub and quickly occupied the top of the GitHub hotlist. PyTorch immediately attracted widespread attention as soon as it was launched, and quickly became popular in research.

High-performance computing based on GPU

The key factors of image processing in medical imaging field are imaging speed, image size and resolution. Due to the limitation of the hardware, the processing of medical images calculated according to sequence. It is also due to the lack of computing resources that the processing of these images wastes a lot of valuable time of doctors and patients. In recent years, GPU has made great progress and moved towards the direction of general computing. Its data processing capacity far exceeds that of CPU, which makes it possible to realize high-performance computing on ordinary computers. The full name of the GPU is the Graphics Processing Unit, a microprocessor that performs image computing on PCs, workstations, game consoles and some mobile devices. In August 1999, NVIDIA released a GeForce 256 graphics chip codenamed NV10. Its architecture is very different from that of the CPU. At the beginning of its birth, it was mainly oriented to the rendering of graphic images. Like the CPU, the GPU is a processor in the graphics card that designed to perform complex mathematical and geometric calculations that are required for graphics rendering. With the GPU, CPU does not need to perform graphics processing work, and can perform other system tasks, which can greatly improve the overall performance of the computer.

Deep learning for medical imaging analysis

With the development of deep learning, computer vision uses a lot of deep learning to deal with various image problems. Medical image as a special visual image has attracted the attention of many researchers. In recent years, various types of medical image processing and recognition have adopted deep learning methods, including fundus images, endoscopic images, CT/MRI images, ultrasound images, pathological images, etc. At present, deep learning technology is mainly used in classification and segmentation in medical images. shows the main medical application scenarios of deep learning.
Figure 2

Deep learning application in medical image analysis. (A) Fundus detection; (B,C) hippocampus segmentation; (D) left ventricular segmentation; (E) pulmonary nodule classification; (F,G,H,I) gastric cancer pathology segmentation. The staining method is H&E, and the magnification is ×40.

Deep learning application in medical image analysis. (A) Fundus detection; (B,C) hippocampus segmentation; (D) left ventricular segmentation; (E) pulmonary nodule classification; (F,G,H,I) gastric cancer pathology segmentation. The staining method is H&E, and the magnification is ×40.

The classification of medical image

Diabetic retinopathy detection

In the field of deep learning, image classification and its application have made great progress this year. On the one hand, the academic circles have made great efforts to design a variety of efficient CNN models, which have achieved high accuracy and even exceeded the human recognition ability. On the other hand, the application of CNN model in medical image analysis has become one of the most attractive directions of deep learning. In particular, the retinal fundus image obtained from fundus camera has become one of the key research objects of deep learning in the field of image classification. The main method for studying related fundus diseases using deep learning techniques is to classify and detect fundus images, such as diabetic retinopathy detection and glaucoma detection. The following lists the deep learning methods applied and fundus image analysis in the past 3 years. These methods mainly use the large scale dataset to train deep CNN model and perform disease classification detection on fundus images. The deep CNN used to update iterations with the development of deep learning techniques, from the earliest shallow CNN model to the deep CNN model or some combination models, and the use of migration learning, data augmentation and other new methods and techniques.
Table 1

The application of deep network in the fundus detection

LiteratureTarget taskNetwork structureMethod introductionResult
Liskowski et al. 2016 (19)Vessel segmentationDeep CNN + complex data preparationProposing a supervised segmentation technique that uses a deep neural network. Using a relatively recent machine learning formalism of structured prediction to produce segmentation resultsROC >0.99; accuracy of classification >0.97
Fu et al. 2016 (20)Vessel segmentationFully CNNs + conditional random fieldsFormulating the vessel segmentation to a boundary detection problem. Using FCN and CRF to generate a vessel probability map and give a binary classification resultState-of-the-art vessel segmentation performance on the DRIVE and STARE datasets
Dasgupta et al. 2017 (21)Vessel segmentationFCNFormulating the segmentation task as a multi-label inference task and utilize the implicit advantages of the combination of CNNs and structured predictionOn DRIVE dataset; accuracy 95.33%; AUC: 0.974
Zhu et al. 2017 (22)Vessel segmentationExtreme learning machineExtracting feature vectors from pixels. Constructing matrix for pixels based on feature vectors and the manual labels. give ELM the matrix and get the binary retinal vascular segmentationAverage accuracy 0.9607; sensitivity 0.7140; specificity 0.9868
Hu et al. 2018 (23)Vessel segmentationMultiscale CNN + CRF + improved cross-entropy lossCombining with CNN and fully CRFs. Developing a multiscale CNN with an improved cross-entropy loss functionCompetitive in the sensitivity while ensuring accuracy
Fu et al. 2018 (24)Optic disc and cup segmentation for glaucoma detectionM-NetConstructing image pyramid to achieve multiple level receptive field sizes. The U-shape CNN learn the rich hierarchical representation. The side-output layer provides early classification result. Using multi-label loss function to generate final segmentation resultState-of-the-art OD and OC segmentation result on ORIGA data set

CNN, convolutional neural network; FCN, fully convolutional network.

CNN, convolutional neural network; FCN, fully convolutional network. We mainly work of detecting fundus diseases in transfer learning. It’s very difficult to obtain large-scale medical annotation set, and transfer learning is an effective method to solve the problem of small data. In order to find the potential factors between the accuracy and the type of primary models, we process transfer learning using the pre-trained model, for example CaffeNet, GoogleNet, VGG19. The experimental results show that the transfer learning based on the pre-trained CNN model is introduced to solve the problems in medical image analysis, and some effective results are produced. The shows our classification model of fundus.
Figure 3

Fundus retina detection using deep learning network.

Fundus retina detection using deep learning network.

Ultrasound detection of breast nodules

The current deep learning technology has achieved research results in the field of ultrasound imaging such as breast cancer, cardiovascular and carotid arteries. Compared with traditional machine learning, deep learning can automatically filter features to improve recognition performance based on multi-layer models. Deep learning has become an important tool for ultrasonic image recognition with its high efficiency and accuracy, which can effectively improve the diagnostic accuracy. In the field of ultrasound imaging of breast nodules, Chen et al. combined the detection methods of texture features and morphology (25,26), using AlexNet model to find the nodules in the image and predicting the benign and malignant, the AUC value reached 0.9325 (27). Shalev-Shwartz et al. applied deep learning to the segmentation of the left ventricular ultrasound image of the heart, and achieved a better segmentation effect. In the study of automatic classification of fetal facial ultrasound images, Yu integrated CNN and random two-coordinate descent optimization algorithm (28), achieving 96.98% accuracy. At the same time, Yu uses the ResNet model in the automatic recognition of melanoma in dermoscopic images based on deep aggregation features (29,30). The AUC has reached more than 80%. In the classification of breast nodules, we mainly propose an enhancement method of data preprocessing, called adaptive contrast enhancement (ACE) method. Directly sending the original data to the neural network training often has a poor effect. On this basis, we combine the ACE algorithm to enhance the ultrasound image. The ACE algorithm can calculate the relative pixel value of the target point and surrounding pixels by difference to correct the final pixel value, which has a good enhancement effect on the image. shows the enhanced image of breast.
Figure 4

Image processed by the ACE algorithm. (A) represents the picture before processing, and (B) represents the processed picture; (C,D) is also similar. ACE, adaptive contrast enhancement.

Image processed by the ACE algorithm. (A) represents the picture before processing, and (B) represents the processed picture; (C,D) is also similar. ACE, adaptive contrast enhancement.

Pulmonary nodule screening

Pulmonary nodule disease is a common lung disease. shows the whole picture of pulmonary nodule in CT image. The accuracy rate of common X-ray chest film in the diagnosis of pulmonary nodules is less than 50%, and even people with normal chest film can be detected to infer sarcoidosis. With CT becoming the main detection method of pulmonary nodule diagnosis, more and more physical examinations begin to include lung cancer screening detection. According to relevant data statistics, the detection rate of pulmonary nodule has increased 5 times in recent years. With the development of deep learning technology, a series of deep learning methods are emerging to detect pulmonary nodules.
Figure 5

The whole picture of pulmonary nodule. (A) shows the lung image; (B) represents the position of pulmonary nodule in the lung image. (C,D) is also similar.

The whole picture of pulmonary nodule. (A) shows the lung image; (B) represents the position of pulmonary nodule in the lung image. (C,D) is also similar. We compared a variety of existing lung nodule detection methods and found that the deep learning algorithm greatly improved the detection rate of nodules compared with the traditional machine learning algorithm. On the other hand, the deep learning algorithm also has its bottleneck in the detection of pulmonary nodules. First, the high detection rate of lung nodules depends on a large number of data sets, and at the same time depends on the accuracy of the annotation data. Secondly, as the number of network layers increases, the accuracy rate is not very large, which indicates that the deep learning algorithm itself has its limitations.

The segmentation of medical image analysis

Early detection of Alzheimer’s disease (AD)

Brain MRI analysis is mainly for the segmentation of different brain regions and the diagnosis of brain diseases, such as brain tumor segmentation (31), schizophrenia diagnosis, early diagnosis of Parkinson’s syndrome (32) and early diagnosis of AD. Among them, the broadest field of deep learning applications is the early diagnosis of AD. AD diagnosis based on deep learning is mainly based on segmentation of hippocampus, cortical thickness and brain volume in brain MRI images. Sarraf et al. (33) trained AD samples for sMRI and fMRI using the well-known LeNet-5 framework in CNNs, respectively, yielding 98.84% and 96.85% accuracy, respectively. This is the first time based on deep learning to analyzing fMRI data. Payan et al. (34) proposed a 3D CNN for AD diagnosis based on SAE pre-training. The 3D MRI scan is randomly selected to pre-train the SAE, and the trained SAE weights are used for 3D-CNN convolution filtering pre-training. Finally, the 3D-CNN fully connected layer is fine-tuned, but fine-tuning needs to be at the expense of a large amount of computational complexity during the training phase. Hosseini-Asl et al. (35) and other analysis based on sMRI feature extraction technology limits the accuracy of AD classification, they proposed a new depth-supervised adaptive 3D-CNN network, in which 3D-CAE learning and automatic extraction Identify AD features and capture changes caused by AD. The 3D-CAE pre-trained convolution filter is further applied to another set of data fields, such as the CAD Dementia pre-trained AD neuroimaging (ADNI) data set. In the early diagnosis of AD, our research group also carried out related experiments and proposed a deep learning method based on enhanced AlexNet. According to the characteristics of AD, we focus on the how to precise the hippocampus. We compared the segmentation of hippocampus with mainstream deep learning networks, including FCN, Unet, SegNet, Unet-3Dand Mask-RCNN. shows the segmented result of these networks.
Figure 6

The hippocampus segmentation result. (A) FCN segmentation; (B) Unet segmentation; (C) SegNet segmentation; (D) Unet-3d segmentation; (E) Mask-RCNN segmentation. FCN, fully convolutional network.

The hippocampus segmentation result. (A) FCN segmentation; (B) Unet segmentation; (C) SegNet segmentation; (D) Unet-3d segmentation; (E) Mask-RCNN segmentation. FCN, fully convolutional network. Furthermore, we use the dice coefficient as a measure. From the , we can see that Mask-RCNN achieves high precision in hippocampus segmentation. At the same time, we noticed that the effect of Unet is far less than the effect of Unet-3D. The reason may be that MRI imaging itself is a three-dimensional form, so the use of three-dimensional convolution can better interpret the segmented object.
Figure 7

It is the training accuracy (A) and loss (B) diagram of five kinds of segmentation networks. FCN, fully convolutional network.

It is the training accuracy (A) and loss (B) diagram of five kinds of segmentation networks. FCN, fully convolutional network.

Left ventricular segmentation

Cardiac MRI analysis diagnoses heart disease by dividing the left ventricle to measure left ventricular volume, ejection fraction, and wall thickness. Among them, deep learning is widely used in left ventricular segmentation. In recent years, deep learning algorithms for left ventricular segmentation on MRI image have emerged in an endless stream. Poudel et al. (36) proposed a recurrent full convolutional network (RFCN) that learns image representation from the entire 2D slice stack and utilizes inter-slice spatial dependence through internal memory cells, which can detect and segment anatomical Combined into a single architecture for end-to-end training, it significantly reduces computation time, simplifies the segmentation process, and enables real-time applications. Isensee et al. (37) proposed a method for classifying into a fully automated processing pipeline by integrating segmentation and disease, which uses a set of U-Net structures to segment the cardiac structure at each point in the cardiac cycle. Liao et al. (38) designed a detector combined with a neural network classifier to detect the ROI containing LV. The LV in the ROI is then segmented using a deep full CNN called hypercolumns. The 2D segmentation results are integrated into different images to estimate the volume. This model uses end-to-end training and uses the real volume directly as a label. Lieman-Sifry et al. (39) developed the FastVentricle architecture based on the ENet architecture. The FastVentricle architecture is a FCN architecture for ventricular partitioning that runs four times faster than the best ventricular partitioning structure and six times less memory. Maintain good clinical accuracy. In this field, our group proposed a precise left cardiac contour segmentation model based on group normalization and nearest neighbor interpolation, which is called GNNI U-Net. We constructed a convolution module based on group normalization method for fast and accurately feature extraction, and an up-sampling module for feature restoring based on nearest neighbor interpolation method. Our model got Dice coefficient of 0.937 on the Sunnybrook dataset and 0.957 and on the LVSC dataset, which achieves high precision on left cardiac contour segmentation. shows our left ventricular segmentation effect.
Figure 8

Segmentation results of GNNI U-net on Sunnybrook dataset (A) and LVSC dataset (B). The figure shows the segmentation results of the left ventricular contour image of GNNI U-net in six different scales of the two data sets. The green area is the correctly divided contour area, the blue area is the mis-segmented area, and the red area is the leak-divided area.

Segmentation results of GNNI U-net on Sunnybrook dataset (A) and LVSC dataset (B). The figure shows the segmentation results of the left ventricular contour image of GNNI U-net in six different scales of the two data sets. The green area is the correctly divided contour area, the blue area is the mis-segmented area, and the red area is the leak-divided area.

Gastric cancer pathological

Pathological diagnosis is the “gold standard” of various cancer diagnosis methods, which plays an important role in the medical field (40). The contradiction between a large number of market demands and a shortage of talents in the field of pathology gives a great opportunity for deep learning in this field, and digital pathology technology also makes computer-assisted diagnosis possible (41). Nowadays, deep learning technology has been applied to the pathological diagnosis of lung cancer, breast cancer and gastric cancer. Its application mainly includes early tumor screening and benign and malignant diagnosis of tumor. In the field of lung cancer pathology, Zhang et al. established the “early computer diagnosis system for lung cancer” to detect lung cancer pathological sections, so that several major types of lung cancer can be detected (42). In the field of breast cancer pathology, Qaiser et al. established a CNN-based intelligent image diagnosis system with a judgment result of pathological sections (cancer and non-cancer tissue) as high as 0.833, four cancer classifications (normal tissue, benign tumor, original The accuracy of the results of both cancer and invasive cancer is as high as 0.778, which can reach the level of pathologists (43). Hanna and others based on the deep learning artificial intelligence automatic human epidermal growth factor receptor scoring system to score human epidermal growth factor receptor, the results are better than pathologists (44). An intelligent image diagnosis system constructed by Ehteshami Bejnordi et al. The diagnosis of gastric cancer pathological images can reach 69.9% accuracy compared with pathologists (45), Yoshida et al. reached 0.556 in three classifications of gastric cancer pathology (normal tissues, adenomas, cancer cells) (46). In the field of gastric cancer pathology our research team has established a benign and malignant diagnostic system based on gastric cancer pathology based on deep learning, with a sensitivity of over 97% (47). However, the above methods include that the initial work of our group rarely involves the field of image segmentation. In fact, the segmentation of pathological images is indeed a problem. We further designed a multi-input model called MIFNet to segment the lesions in the pathological image, and increase the dice coefficient to 81.87% in the segmentation of gastric cancer case images, much higher than some existing segmentation models. Such as U-Net (67.73%), SegNet (63.89%) and PSPNet (60.51%). shows some sample experimental data. What we need to do is get the label of the style shown in , which marks the tissue area of the lesion on the image, as shown in .
Figure 9

Overview of gastric cancer pathology. (A) Pathological image with a cancerous tissue region; (B) label region corresponding to the cancerous tissue in (A); (C) pathological image of cancer-free tissue; (D) the effect of magnifying observation of the cancerous tissue in (A); (E) the effect of extracting the original pathological image and the label and extracting it into the diseased tissue; (F) the effect of magnifying observation for a cancer-free image. The staining method is H&E, and the magnification is ×40.

Overview of gastric cancer pathology. (A) Pathological image with a cancerous tissue region; (B) label region corresponding to the cancerous tissue in (A); (C) pathological image of cancer-free tissue; (D) the effect of magnifying observation of the cancerous tissue in (A); (E) the effect of extracting the original pathological image and the label and extracting it into the diseased tissue; (F) the effect of magnifying observation for a cancer-free image. The staining method is H&E, and the magnification is ×40.

Conclusions and future research

Deep learning is one of the powerful tools for medical image analysis. It has been successfully applied in target detection, segmentation, classification and registration. The development of deep learning in the medical field depends on the accumulation of medical big data, while the medical data itself has multi-modal characteristics, which provides a large amount of rich data for deep learning. In terms of disease treatment, deep learning can not only find the lesion area, but also discriminate and classify specific lesions. When the lesion is positive, many detection networks can also segment the lesion area. While deep learning reflects its own advantages, it still has certain shortcomings. The deep learning model relies heavily on data sets. Each deep learning network requires massive data for training, which makes the data set acquisition more demanding. The root cause is that the pixel features from the original input image are too complex, so it is a future development trend to focus on designing a network with a smaller data size. Deep learning widely applies all aspects of medical image analysis, including ophthalmology, neuroimaging, ultrasound, etc. With the development of deep learning, more and more medical fields will apply deep learning technology, and future deep learning will not only focus on the single aspect of neuroimaging but also other aspects of genomics and bioinformatics. The reason why deep learning can develop so rapidly in the medical field is inseparable from a large number of clinical practices. How to better apply deep learning to all stages of medical treatment becomes a more challenging task. It depends on two aspects: one is the constantly updated iteration of technology, and the other is the continuous accumulation of medical experience. At present, a number of excellent algorithms have emerged in the fields of driverlessness, natural language processing, computer vision, etc. These algorithms have attracted great attention in their respective fields, and how to use these advanced deep learning algorithms is an aspect worthy of our researchers’ constant thinking and innovation. Deep learning on the medical imaging applications is not limited to the detection of big data routine diseases, but also effective solutions for rare diseases. At present, we mainly use the method of migration learning for small sample data to converge and achieve prediction results. However, there are certain limitations in migration learning. Not all rare diseases can be predicted in this way, which brings new challenges and opportunities for the diagnosis of intractable diseases. With the advent of the 5G era, it provides a new and broad space for medical deep learning. Our traditional machine learning algorithms are concentrated at the software level, and there is very little technology involved in deep learning at the hardware level. The combination of 5G technology and deep learning technology enables the machine to achieve true intelligence. At the same time, the continuous development of intelligent medical devices and medical robots promotes the realization of deep learning at the hardware level, and greatly facilitates patient treatment. The article’s supplementary files as
  16 in total

1.  Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.

Authors:  Kaiming He; Xiangyu Zhang; Shaoqing Ren; Jian Sun
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2015-09       Impact factor: 6.226

2.  Estimation of the Volume of the Left Ventricle From MRI Images Using Deep Neural Networks.

Authors:  Fangzhou Liao; Xi Chen; Xiaolin Hu; Sen Song
Journal:  IEEE Trans Cybern       Date:  2017-12-20       Impact factor: 11.448

Review 3.  Histopathological image analysis: a review.

Authors:  Metin N Gurcan; Laura E Boucheron; Ali Can; Anant Madabhushi; Nasir M Rajpoot; B Yener
Journal:  IEEE Rev Biomed Eng       Date:  2009-10-30

4.  Computer-aided diagnosis of breast DCE-MRI using pharmacokinetic model and 3-D morphology analysis.

Authors:  Teh-Chen Wang; Yan-Hao Huang; Chiun-Sheng Huang; Jeon-Hor Chen; Guei-Yu Huang; Yeun-Chung Chang; Ruey-Feng Chang
Journal:  Magn Reson Imaging       Date:  2013-12-17       Impact factor: 2.546

5.  Segmenting Retinal Blood Vessels With Deep Neural Networks.

Authors:  Pawel Liskowski; Krzysztof Krawiec
Journal:  IEEE Trans Med Imaging       Date:  2016-03-24       Impact factor: 10.048

6.  Retinal vessel segmentation in colour fundus images using Extreme Learning Machine.

Authors:  Chengzhang Zhu; Beiji Zou; Rongchang Zhao; Jinkai Cui; Xuanchu Duan; Zailiang Chen; Yixiong Liang
Journal:  Comput Med Imaging Graph       Date:  2016-05-30       Impact factor: 4.790

7.  High concordance between immunohistochemistry and fluorescence in situ hybridization testing for HER2 status in breast cancer requires a normalized IHC scoring system.

Authors:  Allen M Gown; Lynn C Goldstein; Todd S Barry; Steven J Kussick; Patricia L Kandalaft; Patricia M Kim; Christopher C Tse
Journal:  Mod Pathol       Date:  2008-05-16       Impact factor: 7.842

8.  Alzheimer's disease diagnostics by a 3D deeply supervised adaptable convolutional network.

Authors:  Ehsan Hosseini-Asl; Mohammed Ghazal; Ali Mahmoud; Ali Aslantas; Ahmed M Shalaby; Manual F Casanova; Gregory N Barnes; Georgy Gimel'farb; Robert Keynton; Ayman El-Baz
Journal:  Front Biosci (Landmark Ed)       Date:  2018-01-01

9.  Automatic detection and classification of breast tumors in ultrasonic images using texture and morphological features.

Authors:  Yanni Su; Yuanyuan Wang; Jing Jiao; Yi Guo
Journal:  Open Med Inform J       Date:  2011-07-27

10.  HER2 challenge contest: a detailed assessment of automated HER2 scoring algorithms in whole slide images of breast cancer tissues.

Authors:  Talha Qaiser; Abhik Mukherjee; Chaitanya Reddy Pb; Sai D Munugoti; Vamsi Tallam; Tomi Pitkäaho; Taina Lehtimäki; Thomas Naughton; Matt Berseth; Aníbal Pedraza; Ramakrishnan Mukundan; Matthew Smith; Abhir Bhalerao; Erik Rodner; Marcel Simon; Joachim Denzler; Chao-Hui Huang; Gloria Bueno; David Snead; Ian O Ellis; Mohammad Ilyas; Nasir Rajpoot
Journal:  Histopathology       Date:  2017-10-27       Impact factor: 5.087

View more
  20 in total

1.  Improving mammography lesion classification by optimal fusion of handcrafted and deep transfer learning features.

Authors:  Meredith A Jones; Rowzat Faiz; Yuchen Qiu; Bin Zheng
Journal:  Phys Med Biol       Date:  2022-02-21       Impact factor: 3.609

2.  Machine-assisted interpolation algorithm for semi-automated segmentation of highly deformable organs.

Authors:  Dishane C Luximon; Yasin Abdulkadir; Phillip E Chow; Eric D Morris; James M Lamb
Journal:  Med Phys       Date:  2021-11-27       Impact factor: 4.071

3.  Augmented reality navigation with ultrasound-assisted point cloud registration for percutaneous ablation of liver tumors.

Authors:  Longfei Ma; Hanying Liang; Boxuan Han; Shizhong Yang; Xinran Zhang; Hongen Liao
Journal:  Int J Comput Assist Radiol Surg       Date:  2022-06-15       Impact factor: 3.421

4.  Artificial Intelligence in Computer Vision: Cardiac MRI and Multimodality Imaging Segmentation.

Authors:  Alan C Kwan; Gerran Salto; Susan Cheng; David Ouyang
Journal:  Curr Cardiovasc Risk Rep       Date:  2021-08-04

5.  Synthetic 18F-FDG PET Image Generation Using a Combination of Biomathematical Modeling and Machine Learning.

Authors:  Mohammad Amin Abazari; Madjid Soltani; Farshad Moradi Kashkooli; Kaamran Raahemifar
Journal:  Cancers (Basel)       Date:  2022-06-03       Impact factor: 6.575

6.  Fast Multi-Focus Fusion Based on Deep Learning for Early-Stage Embryo Image Enhancement.

Authors:  Vidas Raudonis; Agne Paulauskaite-Taraseviciene; Kristina Sutiene
Journal:  Sensors (Basel)       Date:  2021-01-28       Impact factor: 3.576

7.  TMD-Unet: Triple-Unet with Multi-Scale Input Features and Dense Skip Connection for Medical Image Segmentation.

Authors:  Song-Toan Tran; Ching-Hwa Cheng; Thanh-Tuan Nguyen; Minh-Hai Le; Don-Gey Liu
Journal:  Healthcare (Basel)       Date:  2021-01-06

8.  Deep learning based neural network application for automatic ultrasonic computed tomographic bone image segmentation.

Authors:  Fradi Marwa; El-Hadi Zahzah; Kais Bouallegue; Mohsen Machhout
Journal:  Multimed Tools Appl       Date:  2022-02-16       Impact factor: 2.577

9.  Cyclist Effort Features: A Novel Technique for Image Texture Characterization Applied to Larynx Cancer Classification in Contact Endoscopy-Narrow Band Imaging.

Authors:  Nazila Esmaeili; Axel Boese; Nikolaos Davaris; Christoph Arens; Nassir Navab; Michael Friebe; Alfredo Illanes
Journal:  Diagnostics (Basel)       Date:  2021-03-03

10.  Metal Artifact Reduction in Spectral X-ray CT Using Spectral Deep Learning.

Authors:  Matteo Busi; Christian Kehl; Jeppe R Frisvad; Ulrik L Olsen
Journal:  J Imaging       Date:  2022-03-17
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.