Literature DB >> 33530242

Automatic quality assessment for 2D fetal sonographic standard plane based on multitask learning.

Bo Zhang1, Han Liu2, Hong Luo1, Kejun Li3.   

Abstract

ABSTRACT: The quality control of fetal sonographic (FS) images is essential for the correct biometric measurements and fetal anomaly diagnosis. However, quality control requires professional sonographers to perform and is often labor-intensive. To solve this problem, we propose an automatic image quality assessment scheme based on multitask learning to assist in FS image quality control. An essential criterion for FS image quality control is that all the essential anatomical structures in the section should appear full and remarkable with a clear boundary. Therefore, our scheme aims to identify those essential anatomical structures to judge whether an FS image is the standard image, which is achieved by 3 convolutional neural networks. The Feature Extraction Network aims to extract deep level features of FS images. Based on the extracted features, the Class Prediction Network determines whether the structure meets the standard and Region Proposal Network identifies its position. The scheme has been applied to 3 types of fetal sections, which are the head, abdominal, and heart. The experimental results show that our method can make a quality assessment of an FS image within less a second. Also, our method achieves competitive performance in both the segmentation and diagnosis compared with state-of-the-art methods.
Copyright © 2021 the Author(s). Published by Wolters Kluwer Health, Inc.

Entities:  

Mesh:

Year:  2021        PMID: 33530242      PMCID: PMC7850658          DOI: 10.1097/MD.0000000000024427

Source DB:  PubMed          Journal:  Medicine (Baltimore)        ISSN: 0025-7974            Impact factor:   1.817


Introduction

Background

Fetal sonographic (FS) examinations are widely applied in clinical settings due to its noninvasive nature, reduced cost, and real-time acquisition.[ FS examinations are consisted of first, second, and third trimester examinations, and limited examination,[ which covers a range of critical inspections such as evaluation of a suspected ectopic pregnancy,[ and confirmation of the presence of an intrauterine pregnancy.[ The screening and evaluation of fetal anatomy are critical during the second and third trimester examinations. The screening is usually assessed by ultrasound after approximately 18 weeks’ gestational (menstrual) age. According to a survey,[ neonatal mortality in the United States in 2016 was 5.9 deaths per 1000 live births, and birth defects are the leading cause of infant deaths, accounting for 20% of all infant deaths. Besides, congenital disabilities occur in 1 in every 33 babies (about 3% of all babies) born in the United States each year. In this case, the screening and evaluation of fetal anomaly will provide crucial information to families prior to the anticipated birth of their child on diagnosis, underlying etiology, and potential treatment options, which can greatly improve the survival rate of the fetus. However, the physiological evaluation of fetal anomaly requires well-trained and experienced sonographers to obtain standard planes. Although a detailed quality control guideline was developed for the evaluation of standard plan,[ the accuracy of the measurements is highly dependent on the operator's training skill and experience. According to a study,[ intraobserver and interobserver variability exist in routine practice, and inconsistent image quality can lead to variances in specific anatomic structures captured by different operators. Furthermore, in areas where medical conditions are lagging, there is a lack of well-trained doctors, which makes FS examinations impossible to perform. To this end, automatic approaches for FS image quality assessment are needed to ensure that the image is captured as required by guidelines and provide accurate and reproducible fetal biometric measurements.[ To obtain standard planes and assess the quality of FS images, it is necessary that all the essential anatomical structures in the imaging should appear full and remarkable with clear boundary.[ For each medical section, there are different essential structures. In our research, we consider 3 medical sections: the heart section, the head section, and the abdominal section. The essential structures corresponding to these sections are given in Table 1. The list of essential anatomical structures used to evaluate the image quality is defined by the guideline[ and further refined by 2 senior radiologists with more than 10 years of experience of FS examination at the West China Second Hospital Sichuan University, Chengdu, China. A comparison of standard and nonstandard planes can be illustrated in Figure 1.
Table 1

Essential anatomical structures for different sections.

Section nameEssential anatomical structure
HeadCavum septi pellucidi
Thalamus
Third ventricle
Brain midline
Lateral sulcus
Choroid plexus
AbdominalSpine
Umbilical vein
Aorta
Stomach
HeartUmbilical vein
Umbilical vein
Right ventricle
Right atrium
Descending aorta
Figure 1

Comparison of the standard plane (upper row) and nonstandard plane (lower row) in 3 sections. A, The lower abdominal FS image does not show the umbilical vein. B, The lower head FS image does not show the brain midline and the choroid plexus. C, The lower heart FS image does not show the left ventricle, left atrium, and descending aorta.

Essential anatomical structures for different sections. Comparison of the standard plane (upper row) and nonstandard plane (lower row) in 3 sections. A, The lower abdominal FS image does not show the umbilical vein. B, The lower head FS image does not show the brain midline and the choroid plexus. C, The lower heart FS image does not show the left ventricle, left atrium, and descending aorta. There are various types of challenges concerning the automatic quality control of FS images. As illustrated in Figure 2, the main challenges can be divided into 3 types: the first type is that the image usually suffers from the influence of noise and shadowing effect, the second type is that similar anatomical structures could be confused due to the low resolution of the images, and the third type is that the fetal location during the scanning is unstable which will cause the rotation of some anatomical structure. The first type of challenges can only be solved by using more advanced scanning machines, but we can tackle the rest 2 challenges by a more scientific approach. Specifically, the purpose of our research can be summarized as follows:
Figure 2

Illustration of different types of challenges. A, The white arrows show the substantial interference of noise and shadowing effect. B, In the upper graph, the blue box represents the real brain midline, and the orange box is the confusing anatomical structure with a similar shape. In the lower figure, the orange box represents the actual gallbladder, and the blue box represents the real umbilical vein. These 2 structures have a very similar shape. C, FS images with different fetal positions, which will cause significant variations for the appearance of the images. FS = fetal sonographic.

Propose an automatic fetal sonographic image quality control framework for the segmentation and classification of the 2-dimensional fetal heart standard plane, which is highly robust against the interference of image rotation and similar structures, and the segmentation speed is quite fast to meet the clinical requirements fully. Improve the accuracy of the detection and classification further compared with state-of-the-art methods by using many recent advanced object detection technologies. Generalize the framework so that it can be well applied to other standard planes. Illustration of different types of challenges. A, The white arrows show the substantial interference of noise and shadowing effect. B, In the upper graph, the blue box represents the real brain midline, and the orange box is the confusing anatomical structure with a similar shape. In the lower figure, the orange box represents the actual gallbladder, and the blue box represents the real umbilical vein. These 2 structures have a very similar shape. C, FS images with different fetal positions, which will cause significant variations for the appearance of the images. FS = fetal sonographic.

Related work

In recent years, deep learning techniques have been widely applied in many medical imaging fields due to the technique's stability and efficiency, such as anatomical object detection and segmentation [ and brain abnormalities segmentation.[ Accordingly, many intelligent automatic diagnostic techniques for FS images have been proposed. For example, Deepika et al[ proposed a novel framework to diagnose the fetal anomaly by using ultrasound images. In the framework, it adopts U-Net architecture with Hough transformation to segment the abdominal region, and then a multistage convolutional neural network (CNN) is designed to extract the hidden features of FS images. The experiment shows it outperforms other CNN-based approach.[ Lin et al[ proposed a multitask CNN framework to address the problem of standard plane detection and quality assessment of fetal head ultrasound images. Under the framework, they introduced prior clinical and statistical knowledge to reduce the false detection rate further. The detection speed of this method is quite fast, and the result achieves promising performance compared with state-of-the-art methods.[ Xu et al[ proposed an integrated learning framework based on deep learning to perform view diagnosis and landmark detection of the structures in the fetal abdominal ultrasound image simultaneously. The automatic framework achieved a higher diagnosis accuracy better than clinical experts, and it also reduced landmark-based measurement errors.[ Wu et al proposed a computerized FS image quality assessment scheme to assist the quality control in the clinical obstetric examination of the fetal abdominal region. This method utilizes the local phase features along with the original fetal abdominal ultrasound images as input to the neural network. The proposed scheme achieved competitive performance in both view diagnosis and region localization.[ Chang et al[ proposed an automatic mid-sagittal plane (MSP) assessment method for categorizing the 3D fetal ultrasound images. This scheme also analyzes corresponding relationships between resulting MSP assessments and several factors, including image qualities and fetus conditions. It achieves a correct high rate for the results of MSP detection. Kumar and Sriram et al proposed an automatic method for fetal abdomen scan-plane identification based on 3 critical anatomical landmarks: the spine, stomach, and vein. In their approach, a Biometry Suitability Index (BSI) is proposed to judge whether the scan plane can be used for biometry based on detected anatomical landmarks. The results of the proposed method over video sequences were closely similar to the clinical expert's assessment of scan-plane quality for biometry.[ Baumgartner et al[ proposed a novel framework based on convolutional neural networks to automatically detect 13 standard fetal views in freehand 2D ultrasound data and provide localization of the anatomical structures through a bounding box. A notable innovation is that the network learns to localize the target anatomy using weak supervision based on image-level labels only.[ Namburete et al[ proposed a multitask, fully convolutional neural network framework to address the problem of 3D fetal brain localization, alignment to a referential coordinate system, and structural segmentation. This method optimizes the network by learning features shared within the input data belonging to the correlated tasks, and it achieves a high brain overlap rate and low eye localization error.[ However, there are no existing automatic quality control methods for fetal heart planes, and the detection accuracy of existing methods on other planes is relatively low due to the use of the outdated design of neural networks. Therefore, it is desirable to propose a more efficient framework that can not only provide accurate clinical assessment in fetal heart plane but can also increase the segmentation accuracy in other planes.

Methods

The framework of our methods can be illustrated in Figure 3. First, the original image is smoothed by the Gaussian filter, and input to feature extraction network (FEN). Second, FEN will extract a deep level feature of image by convolutional neural network and input to Region Proposal Network (RPN) and Class Prediction Network (CPN) respectively. Then CPN will judge whether the organs meet the standard as well as predict the class, and RPN will locate the position of essential organs with the help of feature pyramid network. Lastly, the 2 networks will combine information together and output the final result. In this section, we will briefly introduce the network structure and then elaborate the feature extraction, the region of interest (ROI) localization, and the organ diagnosis in detail. Our study is approved by Ethics Committee of West China Second Hospital Sichuan University.
Figure 3

The framework of our method. We train the network end-to-end to ensure the best performance. The framework contains 3 sections: Feature Extraction Network (FEN), Region Proposal Network (RPN), and Class Prediction Network (CPN). The will help to extract the deep-level features of the image with the help of the relation module and Spatial Pyramid Pooling (SPP) layer, which is the input to RPN and CPN. The RPN will locate the position of essential structures based on the anchors generated by Feature Pyramid Network (FPN), and the CPN will help to judge and classify the structures. The final output will be a quality assessment of each essential structure and its location.

The framework of our method. We train the network end-to-end to ensure the best performance. The framework contains 3 sections: Feature Extraction Network (FEN), Region Proposal Network (RPN), and Class Prediction Network (CPN). The will help to extract the deep-level features of the image with the help of the relation module and Spatial Pyramid Pooling (SPP) layer, which is the input to RPN and CPN. The RPN will locate the position of essential structures based on the anchors generated by Feature Pyramid Network (FPN), and the CPN will help to judge and classify the structures. The final output will be a quality assessment of each essential structure and its location.

Feature extraction network

In the feature extraction network, we have made many improvements compared with the traditional CNN-based approaches: the convolutional neural network is used as a thematic framework, and many state-of-the-art deep learning techniques such as relation module, spatial pyramid pooling (SPP) layer, are integrated into the framework to further increase the feature extraction efficiency. The CNN has unique advantages in speech recognition and image processing with its special structure of local weight sharing, which can greatly reduce the number of parameters and improve the accuracy of recognition.[ CNN typically consists of pairs of convolutional layers and average pooling layers and fully connected (FC) layers. In convolutional layer, several output feature maps can be obtained by the convolutional calculation between input layer and kernel. Specifically, suppose denotes the mth output feature map in layer n, denotes the kth feature map in n – 1 layer, denotes the kernel generating that feature map, then we can get: where is the bias term in the nth layer, relu denotes rectified linear unit, and is defined as: relu(x)=max(x,0). It is also worth mentioning that we use global average pooling (GAP) instead of local pooling for pooling layers. The aim is to use GAP to replace FC layer, which can regularize the structure of the entire network to prevent overfitting.[ The setting of convolution layer is shown in Table 2.
Table 2

The setting of convolutional layer.

LayerKernel sizeChannel depthStride
C131282
C232562
C335122
C4310242
C5320482
The setting of convolutional layer. To fully utilize relevant features between objects and further improve segmentation accuracy, we introduce the relation module presented by Hu.[ Specifically, first the geometry weight is defined as: where and are geometric features, is a dimensional lifting transformation by using concatenation. After that, the appearance weight is defined as: where and are the pixel weights from the previous network. Then the relation weight indicating the impact from other objects is computed as: Lastly, the relation feature of the whole object set with respect to the object is defined as This module achieves a great performance in the instance recognition and duplicate removal, which increases the segmentation accuracy significantly. The SPP layer we use here denotes the SPP layer presented by He et al.[ Specifically, the response map after FC layer is divided into 1 × 1 (pyramid base), 2 × 2 (lower middle of the pyramid), 4 × 4 (higher middle of the pyramid), 16 × 16 (pyramid top) 4 submaps and do max pooling separately. A problem with the traditional CNN network for feature extraction is that there is a strict limit on the size of the input image, this is because there is a need for the FC layer to complete the final classification and regression tasks, and since the number of neurons of the FC layer is fixed, the input image to the network must also have fixed size. Generally, there are 2 ways of fixing input image size: cropping and wrapping, but these 2 operations either cause the intercepted area not to cover the entire target or bring image distortion, thus applying SPP is necessary. The SPP network also contributes to multisize extraction features and is highly tolerant to target deformation. The design of Bottle Net borrows the idea of Residual Networks.[ A common problem with deep networks is that gradient depth and gradient explosions are prone to occur as depth deepens. The main reason of this phenomenon is the over-fitting problem caused by the loss of information. Since each convolutional layer or pooling layer will downsample the image, a lossy compression effect could be produced. With network going deeper, these images will appear some strange phenomena, which is that obviously different categories of images produce similarly stimulating effect on the network. This reduction in the gap will make the final classification effect less than ideal. To let our network extract deeper features more efficiently, we add the residual network structure to our model. The basic implementation is given in Figure 2. By introducing the data output of the previous layers directly into the input part of the latter data layer, it is realized that original vector data and the subsequently down sampled data are used together as the data input of the latter layer, which introduced a richer dimension. In this way, the network can learn more features of the image.

ROI localization with RPN

The RPN is designed to localize the ROI that encloses the essential organs given in Table 1. To achieve this goal, we first use a feature pyramid network (FPN)[ to generate candidate anchors instead of traditional RPN network used in Faster-RCNN.[ FPN could connect the high-level features of low-resolution and high-semantic information with the low-level features of high-resolution and low-semantic information from top to bottom, so that features at all scales have rich semantic information. Specifically, the setting of FPN is shown in Table 3.
Table 3

The setting of FPN.

Pyramid levelStrideSize
3832
41664
532128
664256
7128512

FPN = feature pyramid network.

The setting of FPN. FPN = feature pyramid network. In the training process, we define the metrics of intersection over union (IoU) to evaluate the goodness of ROI localization: where A is a computerized ROI and B is a manually labeled ROI (Ground Truth). In the training process, we set the samples with IoU higher than 0.5 as positive samples, and IoU lower than 0.5 as negative samples.

Judging and predicting class with CPN

For different sections, we use CPN to classify essential organs. For thalamus section, there are cavum septi pellucidi and thalamus to be classified. For abdominal section, there are stomach bubble, spine, and umbilical vein to be classified. For heart section, there are left ventricle, left atrium, right ventricle, and right atrium to be classified. To improve classification accuracy, we choose focal loss[ as the loss function. In the training process of neural network, the internal parameters can be adjusted with the minimization of the loss function of all training samples. The proposed focal loss enables highly accurate dense object detection in the presence of vast number of background examples, which is suitable in our model. The loss function can be defined as: where γ is the focusing parameter, and γ ≥ 0. is defined as: y represents the truth label of a sample, and p represents the probability that the neural network predicts this class.

Results

In this section, we will start with a brief explanation of the process of obtaining and making data sets for training and testing. Then a systematic evaluation scheme will be proposed to test the efficacy of our method in FS examinations. The evaluation is carried out in 4 parts. First, we investigate the performance of ROI localization; we will use Mean Average Precision (mAP) and box-plot to evaluate it. Second, we quantitatively analyze the performance of classification with common indicators: accuracy (ACC), specificity (Spec), sensitivity (Sen), precision (Pre), F1-score (F1), and area under the receiver of operation curve (AUC). Third, we demonstrate the accuracy of our scheme when compared with experienced sonographers. Fourth, we do the running time analysis and sensitivity analysis of our method.

Data preparation

All the FS images used for training and testing our model were acquired from the West China Second Hospital Sichuan University from April 2018 to January 2019. The FS images were recorded with a conventional hand-held 2-D FS probe on pregnant women in the supine position, by following the standard obstetric examination procedure. The fetal gestational ages of all subjects ranged from 20 to 34 weeks. All FS images were acquired with a GE Voluson E8 and Philips EPIQ 7 scanner. There are, in total, 1325 FS images of the head section, 1321 FS images of the abdominal section, and 1455 FS images of the heart section involved for the training and testing of our model. The training set, validation set, and test set of each section are all divided by a ratio of 3:1:1. The ROI labeling of essential structures in each section is achieved by 2 senior radiologists with more than 10 years of experience in the FS examination by marking the smallest circumscribed rectangle of the positive sample. The negative ROI samples are randomly collected from the background of the images.

Evaluation metrics

For testing the performance of ROI localization, first, we define the metrics of IoU between prediction and ground truth and use box-plots to evaluate ROI localization intuitively. As illustrated before, IoU is defined as: where A is computerized ROI, and B is ground truth (manually labeled) ROI. Second, we use average precision (AP) to quantitatively evaluate the segmentation results of each essential anatomical structure and mAP to illustrate the overall quality of ROI localization. To test the performance of classification results, we use several popular evaluation metrics. Suppose TP represents the number of true positives of a certain class, FP is the number of false positives, FN is the number of false negatives, and TN is the number of true negatives, then the definitions of ACC, specificity (Spec), sensitivity (Sen), precision (Pre), and F1-score (F1) are as follows: The area under the AUC is defined as the area under the receiver operating characteristic (ROC) curve, which is equivalent to the probability that a randomly chosen positive example is ranked higher than a randomly chosen negative example.[ The confusion matrix is also a common indicator to visualize the performance of diagnosis in supervised machine learning, we use it to illustrate the performance of our method in each anatomical structure.[ To show the effectiveness of advanced techniques we add to the framework, and 2 different structures are also tested, where NRM means the removal of the relation module, and NSPP means the removal of the SPP layer in the feature extraction network. By comparing the difference in classification and segmentation results, it is clear to see their impact on overall network performance. To analyze the time complexity of our method, we use floating point operations (FLOPs),[ which is a common method in describing the complexity of CNN. Specifically, in the convolutional layer, FLOPs is defined as: where is the input channel size, is the output channel size, the convolutional kernel size is , and the output feature map size is: . In the fully connected layer, FLOPs is defined as: In the global average pooling layer, FLOPs is defined as: where denotes the input feature map height, and denotes the input feature map width.

Results of ROI localization

To demonstrate the efficacy of our method in localizing the position of essential anatomical structures in FS images, we carry out the experimental evaluation in 2 parts. First, we use box-plots to evaluate ROI localization intuitively. Second, we use AP and mAP to illustrate the quality of ROI localization quantitatively. For the head standard plane, there is already a state-of-the-art method proposed for the quality assessment[ (denoted as Lin), so we have compared its results with our method. Also, to show the effectiveness of advanced object segmentation techniques we add to the network, our methods have also been compared with other popular object detection frameworks, including SSD,[ YOLO,[ Faster R-CNN.[ The test of the effectiveness of the relation module we add to the network is also carried out, with Non-NM denoting the framework without the relation module. As shown in Figure 4  , our method has achieved a high IoU in all 3 sections. Specifically, for the head section, the median of IoU values in all the anatomical structures is above 0.955. Also, for the heart section and the abdominal section, the median is above 0.945 and 0.938, respectively. Also, the minimum of IoU values for all 3 sections is above 0.93. As a comparison, the state-of-the-art framework for the quality assessment of the fetal abdominal images proposed by Wu et al[ has only achieved a median of below 0.9. It proves the effectiveness of our method in localizing ROI. As shown in Table 4, we observe that our method has the highest mAP compared with the method proposed by Lin et al and other popular object detection frameworks. Also, we have improved the segmentation accuracy significantly in TV and CSP and overcome the limitation in Lin's method. This is because our method could detect flat and smaller anatomical structures more precisely. It is worth mentioning that after adding the relation module to our network, the segmentation accuracy has been significantly improved in all the anatomical structures, which proves the effectiveness of this module. As shown in Table 5, since it is our first attempt to evaluate the image quality in the heart section, so we have only compared our method with state-of-the-art object segmentation frameworks. We observe that our approach has the highest average precision in all the anatomical structures. Also, as shown in Table 6, we have achieved quite promising segmentation accuracy. It proves that our framework is generalized and can be well applied to the quality assessment of other standard planes.
Figure 4

Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union.

Table 4

Comparisons about detection results between our method and other methods in head section.

MethodTVBMTCPCSPLSmAP
SSD40.5682.7572.6154.2163.4375.4164.66
YOLOv235.4379.3138.5662.7083.6783.3163.83
Faster R-CNN VGG1673.5694.6593.4180.5987.3594.7887.39
Faster R-CNN Resnet5072.4895.4092.7885.4784.7195.3187.69
Lin82.5098.9593.8995.8289.9298.4693.26
Non-NM71.4294.3889.9282.4586.7892.4586.23
Our method86.1298.8794.2193.7695.5797.9294.41

BM = brain midline, CP = choroid plexus, CSP = cavum septi pellucidi, LS = lateral sulcus, mAP = mean average precision, T = thalamus, TV = third ventricle.

Table 5

Comparisons about detection results between our method and other methods in heart section.

MethodRVRALVLADAOmAP
SSD60.2762.4367.6154.2174.6763.84
YOLOv271.3169.3971.5662.7090.7473.14
Faster R-CNN VGG1685.5281.1587.1180.5995.5785.99
Faster R-CNN Resnet5089.4483.5990.7885.4794.0188.66
Non-NM84.3585.4391.2382.4595.8687.86
Our method93.0395.7195.6593.7697.3495.10

DAO = descending aorta, LA = left atrium, LV = left ventricle, mAP = mean average precision, RA = right atrium, RV = right ventricle.

Table 6

Comparisons about detection results between our method and other methods in abdominal section.

MethodSTUVSPAOmAP
SSD80.7483.4376.2362.1375.63
YOLOv288.2185.4778.6167.7180.00
Faster R-CNN VGG1690.2592.1588.7282.5488.42
Faster R-CNN Resnet5091.2993.5990.8581.3489.27
Non-NM93.4191.7694.3890.3492.47
Our method97.3397.7796.2594.1696.38

AO = aorta, mAP = mean average precision, SP = spine, ST = stomach, UV = umbilical vein.

Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union. Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union. Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union. Comparisons about detection results between our method and other methods in head section. BM = brain midline, CP = choroid plexus, CSP = cavum septi pellucidi, LS = lateral sulcus, mAP = mean average precision, T = thalamus, TV = third ventricle. Comparisons about detection results between our method and other methods in heart section. DAO = descending aorta, LA = left atrium, LV = left ventricle, mAP = mean average precision, RA = right atrium, RV = right ventricle. Comparisons about detection results between our method and other methods in abdominal section. AO = aorta, mAP = mean average precision, SP = spine, ST = stomach, UV = umbilical vein.

Results of diagnosis accuracy

To illustrate the performance of our model in classifying the essential anatomical structures, we first use area of ROC and confusion matrix to characterize the performance of the classifier visually, then we use several authoritative indicators to measure it quantitatively: ACC, Spec, Sen, Pre, and F1. Also, to show the effectiveness of our proposed network in diagnosis, we have compared our method with other popular classification networks, including AlexNet,[ VGG16, VGG19,[ and ResNet50.[ The comparison with Lin's method is also carried out. As shown in Figure 5, it is observed that the classifier achieves quite promising performance in all the 3 sections with the true positive rate reaching 100% while the false positive rate is less than 10%. Also, the ROC achieves at 0.96, 0.95, and 0.98 for the head section, abdominal section, and heart section, respectively.
Figure 4 (Continued)

Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union.

ROC curves of classification results in 3 sections. ROC = receiver operating characteristic. From Figure 6, it is clear that our method achieves a quite superior performance in every anatomical structure of different sections with the true positive rate reaching nearly 100%.
Figure 4 (Continued)

Box-plots of IoU values for 3 sections. The 3 lines on each box represent the 4 quartiles of the IoU values. IoU = intersection over union.

Confusion matrix of diagnosis methods. (A) represents abdominal section; (B) represents heart section; and (C) represents heart section. From Table 7, we can observe that the classification results of our method are superior to other state-of-the-art methods. Specially, we achieve the best results with a precision of 94.63%, a specificity of 96.39%, and an AUC of 98.26%, which are better than Lin's method. The relative inferior results in sensitivity, accuracy, and F1-score can be further improved if we add prior clinical knowledge into our framework.[Tables 8 and 9 illustrate the classification results in abdominal and heart section. We can observe that our method has achieved quite promising results in most indicators compared with existing methods. It demonstrates the effectiveness of our proposed method in classifying anatomical structures of all the sections.
Table 7

Comparisons about classification results between our method and other methods in head section.

IndicatorPrecSenACCF1SpecAUC
AlexNet90.2191.2892.2291.7792.4695.29
VGG1692.3492.4891.7990.6893.7196.37
VGG1994.3193.7292.9191.0194.3397.14
Resnet5094.3693.0193.9694.2192.9197.59
Lin93.5793.5794.3793.5795.0098.18
Our method94.6392.4194.3193.1796.3998.26

ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity.

Table 8

Comparisons about classification results between our method and other methods in abdominal section.

IndicatorPrecSenACCF1SpecAUC
AlexNet91.3491.2492.1190.4391.5393.12
VGG1693.2193.5991.3390.9192.2392.76
VGG1993.4294.2792.1293.9192.9193.21
Resnet5094.5294.1093.6693.7695.9296.32
Our method95.6793.5696.3193.1797.9298.54

ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity.

Table 9

Comparisons about classification results between our method and other methods in heart section.

IndicatorPrecSenACCF1SpecAUC
AlexNet90.3491.4492.3290.7891.2893.87
VGG1692.7693.7192.1393.1292.3393.43
VGG1993.6894.3193.3792.8893.5494.41
Resnet5095.9194.3193.5293.4594.4293.56
Our method96.7194.7393.3295.9194.4995.67

ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity.

Comparisons about classification results between our method and other methods in head section. ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity. Comparisons about classification results between our method and other methods in abdominal section. ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity. Comparisons about classification results between our method and other methods in heart section. ACC = accuracy, AUC = the area under the receiver of operation curve, F1 = F1 score, Prec = precision, Sen = sensitivity, Spec = specificity.

Running time analysis and sensitivity analysis

We test the running time of detecting a single FS image for different single-task and multitask networks in a workstation equipped with 3.60 GHz Intel Xeon E5-1620 CPU and a GP106-100 GPU. The results are given in Table 10. It is observed that detecting a single frame could only cost 0.871s, which is fast enough to meet clinical needs. Also, it is observed that although the network parameters of our method and FLOPs are much more than Faster R-CNN + VGG16, there is not much difference in segmentation time, this is because our network shared many low-level features, which could achieve a more efficient segmentation with using only a few parameters.
Table 10

The detection speed and parameters of different single-task and multitask methods.

MethodFLOPs (B)Speed (s)Parameters (M)
Single-task
 AlexNet0.710.01261.10
 VGG1615.650.023138.37
 VGG1919.820.025143.68
 ResNet504.120.05222.56
 YoLo v259.280.561231.26
Multitask
 Faster R-CNN VGG16169.230.721432.31
 Our method186.40.871502.92

B = billion, FLOPs = floating point operations, M = million.

The detection speed and parameters of different single-task and multitask methods. B = billion, FLOPs = floating point operations, M = million. In the CNN model, we usually need to try a series of parameters to get the best performance for the model. There are many parameters that could affect the results of a CNN model such as learning rate, epoch, regularization loss, etc, but generally, the learning rate and weight decay have a great impact on it. Therefore, we change the learning rate and weight decay to illustrate the sensitivity of our method. As shown in Table 11, by altering the learning rate, the change of mAP does not exceed 1.7%; by changing the weight decay, the change of mAP does not exceed 2.2%, demonstrating the robustness of our method.
Table 11

Comparisons about classification results between our method and other methods in abdominal section.

Learning rateEpochRegularization lossWeight decaymAP
0.000000110050.000195.1
0.00000110050.000193.4
0.0000110050.000194.8
0.000000110050.000296.1
0.00000110050.000295.3
0.0000110050.000296.3
0.000000110050.000394.7
0.00000110050.000393.3
0.0000110050.000394.1

mAP = mean average precision.

Comparisons about classification results between our method and other methods in abdominal section. mAP = mean average precision. Figures 7–9 depict the comparison of our results with the manually labeled images by experts in the head section, abdominal section, and heart section, respectively. Our method displays the classification and segmentation results simultaneously to assist in sonographers’ observation. More comparison between our results and ground truth is given in Figure 10. It can be seen that our method is perfectly aligned with professional sonographers.
Figure 5

ROC curves of classification results in 3 sections. ROC = receiver operating characteristic.

Figure 7

Demonstration that our results perfectly match with the annotations of ground truth in the heart section. The classification results in the left white box are the ground truth labeled by professional radiologists, and the results in the right white box are the detection results of our method. “1” means the anatomical structure meets the quality requirement, and “0” means the structure does not meet the requirement.

Figure 8

Demonstration that our results perfectly match with the annotations of ground truth in the abdominal section. The classification results in the left white box are the ground truth labeled by professional radiologists, and the results in the right white box are the detection results of our method. “1” means the anatomical structure meets the quality requirement, and “0” means the structure does not meet the requirement.

Demonstration that our results perfectly match with the annotations of ground truth in the heart section. The classification results in the left white box are the ground truth labeled by professional radiologists, and the results in the right white box are the detection results of our method. “1” means the anatomical structure meets the quality requirement, and “0” means the structure does not meet the requirement. Demonstration that our results perfectly match with the annotations of ground truth in the abdominal section. The classification results in the left white box are the ground truth labeled by professional radiologists, and the results in the right white box are the detection results of our method. “1” means the anatomical structure meets the quality requirement, and “0” means the structure does not meet the requirement. Demonstration that our results perfectly match with the annotations of ground truth in the head section. The classification results in the left white box are the ground truth labeled by professional radiologists, and the results in the right white box are the detection results of our method. “1” means the anatomical structure meets the quality requirement, and “0” means the structure does not meet the requirement. More comparisons between our results with ground truth. The first 2 rows show the results in the abdominal section, the middle 2 rows show the results in the heart section, and the last 2 rows show the results in the head section. For every section, the upper row represents the ground truth, and the lower row represents our results.

Discussion

In this paper, an autonomous image quality assessment approach for FS images was investigated. The experimental results show that our proposed scheme achieves a highly precise ROI localization and a considerable degree of accuracy for all the essential anatomical structures in the 3 standard planes. Also, the conformance test shows that our results are highly consistent with those of professional sonographers, and running time tests show that the image segmentation speed per frame is much higher than sonographers, which means this scheme can effectively replace the work of sonographers. In our proposed network, to further improve segmentation and classification accuracy, we also modify the recently published advanced object segmentation technologies and adapt them to our model. The experiment shows these modules are highly useful, and the overall performance is better than the state-of-the-art methods such as the FS image assessment framework proposed by Lin et al.[ After the Feature Extraction Network, we also divide the network into Region Proposal Network and Class Prediction Network. Accordingly, the features in the segmentation network can avoid interfering with the features in the classification network, so the segmentation accuracy is further increased. Also, the segmentation speed can be significantly improved, as the classification and localization are performed simultaneously. Although our method achieves quite promising results, there are still some limitations. First, for the training sets, we regard the manually labeled FS images by 2 professional sonographers as the ground truth, but the results of manual labeling will have some accidental deviation even though they all have more than 10 years of experience. In future studies, we will invite more professional clinical expects to label the FS images and collect more representative datasets. Second, there still remain some segmentation and classification errors in our results. This is because our evaluation criteria are rigorous, and the midsection of a single anatomical structure could lead to a negative score on the image. Third, all the FS images are collected from GE Voluson E8 and Philips EPIQ 7 scanner; however, different types of ultrasonic instruments will produce different ultrasound images, which may cause our method not to be applied well to the FS images produced by other machines. Our proposed method further boosts the accuracy in the assessment of two-dimensional FS standard plane. Although t3-dimensional and 4-dimensional ultrasound testing are popular recently, they are mainly utilized to meet the needs of pregnant women and their families to view baby pictures instead of serving the diagnosing purpose visually. Two-dimensional ultrasound images are still the most authoritative basis for judging fetal development.[ As illustrated before, there are still many challenges for the automatic assessment of 2D ultrasound images, such as shadowing effects, similar anatomical structures, different fetal positions, etc. To overcome these challenges and further promote the accuracy and robustness of segmentation and classification, it may be useful to add some prior clinical knowledge[ and more advanced attention modules to the network. In the future, we will also investigate the automatic selection technology for finding the standard scanning plane, which will find a standard plane containing all the essential anatomical structures without sonographers’intervention.

Acknowledgments

The authors acknowledge Sichuan University West China Second Hospital for providing the fetal ultrasound image datasets.

Author contributions

Conceptualization: Bo Zhang, Hong Luo. Data curation: Bo Zhang, Hong Luo. Formal analysis: Bo Zhang, Hong Luo. Funding acquisition: Hong Luo. Investigation: Bo Zhang, Kejun Li. Methodology: Bo Zhang, Han Liu, Kejun Li. Project administration: Hong Luo. Resources: Hong Luo. Software: Bo Zhang, Han Liu. Supervision: Hong Luo. Validation: Han Liu. Visualization: Han Liu. Writing – original draft: Han Liu. Writing – review & editing: Bo Zhang, Han Liu, Kejun Li.
  18 in total

1.  The evidence base for miscarriage diagnosis: better late than never.

Authors:  B Thilaganathan
Journal:  Ultrasound Obstet Gynecol       Date:  2011-11       Impact factor: 7.299

2.  Evaluation and comparison of current fetal ultrasound image segmentation methods for biometric measurements: a grand challenge.

Authors:  Sylvia Rueda; Sana Fathima; Caroline L Knight; Mohammad Yaqub; Aris T Papageorghiou; Bahbibi Rahmatullah; Alessandro Foi; Matteo Maggioni; Antonietta Pepe; Jussi Tohka; Richard V Stebbing; John E McManigle; Anca Ciurte; Xavier Bresson; Meritxell Bach Cuadra; Changming Sun; Gennady V Ponomarev; Mikhail S Gelfand; Marat D Kazanov; Ching-Wei Wang; Hsiang-Chou Chen; Chun-Wei Peng; Chu-Mei Hung; J Alison Noble
Journal:  IEEE Trans Med Imaging       Date:  2013-08-06       Impact factor: 10.048

3.  Multi-Scale Deep Reinforcement Learning for Real-Time 3D-Landmark Detection in CT Scans.

Authors:  Florin-Cristian Ghesu; Bogdan Georgescu; Yefeng Zheng; Sasa Grbic; Andreas Maier; Joachim Hornegger; Dorin Comaniciu
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2017-12-12       Impact factor: 6.226

4.  Focal Loss for Dense Object Detection.

Authors:  Tsung-Yi Lin; Priya Goyal; Ross Girshick; Kaiming He; Piotr Dollar
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2018-07-23       Impact factor: 6.226

5.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.

Authors:  Shaoqing Ren; Kaiming He; Ross Girshick; Jian Sun
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-06-06       Impact factor: 6.226

6.  SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound.

Authors:  Christian F Baumgartner; Konstantinos Kamnitsas; Jacqueline Matthew; Tara P Fletcher; Sandra Smith; Lisa M Koch; Bernhard Kainz; Daniel Rueckert
Journal:  IEEE Trans Med Imaging       Date:  2017-07-11       Impact factor: 10.048

7.  Multi-task learning for quality assessment of fetal head ultrasound images.

Authors:  Zehui Lin; Shengli Li; Dong Ni; Yimei Liao; Huaxuan Wen; Jie Du; Siping Chen; Tianfu Wang; Baiying Lei
Journal:  Med Image Anal       Date:  2019-09-06       Impact factor: 8.545

8.  Ultrasound evaluation of ectopic pregnancy including correlation with human chorionic gonadotrophin levels.

Authors:  S E Chambers; B B Muir; N G Haddad
Journal:  Br J Radiol       Date:  1990-04       Impact factor: 3.039

9.  Pregnancy of unknown location: a consensus statement of nomenclature, definitions, and outcome.

Authors:  Kurt Barnhart; Norah M van Mello; Tom Bourne; Emma Kirk; Ben Van Calster; Cecilia Bottomley; Karine Chung; George Condous; Steven Goldstein; Petra J Hajenius; Ben Willem Mol; Thomas Molinaro; Katherine L O'Flynn O'Brien; Richard Husicka; Mary Sammel; Dirk Timmerman
Journal:  Fertil Steril       Date:  2010-10-14       Impact factor: 7.329

10.  Detecting Anatomical Landmarks From Limited Medical Imaging Data Using Two-Stage Task-Oriented Deep Neural Networks.

Authors:  Jun Zhang; Mingxia Liu; Dinggang Shen
Journal:  IEEE Trans Image Process       Date:  2017-06-28       Impact factor: 10.856

View more
  2 in total

1.  Medical Professional Enhancement Using Explainable Artificial Intelligence in Fetal Cardiac Ultrasound Screening.

Authors:  Akira Sakai; Masaaki Komatsu; Reina Komatsu; Ryu Matsuoka; Suguru Yasutomi; Ai Dozen; Kanto Shozu; Tatsuya Arakaki; Hidenori Machino; Ken Asada; Syuzo Kaneko; Akihiko Sekizawa; Ryuji Hamamoto
Journal:  Biomedicines       Date:  2022-02-25

2.  Deep Learning-Based Computer-Aided Fetal Echocardiography: Application to Heart Standard View Segmentation for Congenital Heart Defects Detection.

Authors:  Siti Nurmaini; Muhammad Naufal Rachmatullah; Ade Iriani Sapitri; Annisa Darmawahyuni; Bambang Tutuko; Firdaus Firdaus; Radiyati Umi Partan; Nuswil Bernolian
Journal:  Sensors (Basel)       Date:  2021-11-30       Impact factor: 3.576

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.