Literature DB >> 35602200

Deep learning-based detection and segmentation of diffusion abnormalities in acute ischemic stroke.

Chin-Fu Liu1,2, Johnny Hsu3, Xin Xu3, Sandhya Ramachandran1,2, Victor Wang1,2, Michael I Miller1,2,4, Argye E Hillis5,6, Andreia V Faria3.   

Abstract

Background: Accessible tools to efficiently detect and segment diffusion abnormalities in acute strokes are highly anticipated by the clinical and research communities.
Methods: We developed a tool with deep learning networks trained and tested on a large dataset of 2,348 clinical diffusion weighted MRIs of patients with acute and sub-acute ischemic strokes, and further tested for generalization on 280 MRIs of an external dataset (STIR).
Results: Our proposed model outperforms generic networks and DeepMedic, particularly in small lesions, with lower false positive rate, balanced precision and sensitivity, and robustness to data perturbs (e.g., artefacts, low resolution, technical heterogeneity). The agreement with human delineation rivals the inter-evaluator agreement; the automated lesion quantification of volume and contrast has virtually total agreement with human quantification.
Conclusion: Our tool is fast, public, accessible to non-experts, with minimal computational requirements, to detect and segment lesions via a single command line. Therefore, it fulfills the conditions to perform large scale, reliable and reproducible clinical and translational research.
© The Author(s) 2021.

Entities:  

Keywords:  Brain imaging; Stroke

Year:  2021        PMID: 35602200      PMCID: PMC9053217          DOI: 10.1038/s43856-021-00062-8

Source DB:  PubMed          Journal:  Commun Med (Lond)        ISSN: 2730-664X


Introduction

Stroke is a major cause of death and long-term disability in US, with increasing mortality rates among middle-aged adults[1]. Defining the stroke core in diffusion-weighted images (DWIs), a magnetic resonance imaging (MRI) sequence highly sensitive to the acute lesion, is a major benchmark for acute treatment. The importance of fast locating and objectively quantifying the acute damage has been reinforced by many stroke trials[2-5]. Clinical research also relays on objective quantification of lesions to access brain function (as lesion-symptoms studies[6]), to stratify populations in clinically relevant groups, and to model prognosis[7-10]. Due to the great variability of stroke population and lesions in both biological and technical aspects, stroke research depends on large datasets and, consequently, on automated tools to process them with accuracy and high levels of reproducibility. Therefore, accessible tools that can fast and efficiently detect and segment diffusion abnormalities in acute strokes are well anticipated by the clinical and research communities. Traditional machine learning algorithms use texture features[11], asymmetric features[12], abnormal voxel intensity on histogram[13] or probabilistic maps of populations[14,15], or advanced statistical models such as support vector machine[16] and random forest[17], to detect and segment diffusion abnormalities in acute strokes. Although these methods represent resourceful statistical approaches and can work reasonably in high-resolution images and homogeneous protocols, the low-level features these methods utilize such as intensity, spatial, and edge information reduce their capability to capture the large variation in lesion pattern, especially in typical clinical, low-resolution, and noisy data. As the prosperous progress in deep learning (DL) in the past two decades, several DL neural network models[18-20], such as the convolutional neural networks (CNNs), performed better brain lesion detection and segmentation than traditional methods. A landmark was the introduction of UNet[21], to re-utilize coarse semantical features via skip connections from encoders to decoders in order to predict better segmentation. Further developments of UNet variants, such as Mnet, DenseUnet, Unet++, and Unet3+[22-24] optimized the features utilization. The emergence of attention-gate techniques[25-27] conditioned networks to focus on local semantical features. Recent studies applied “attention UNets”, for example, to predict final ischemic lesions from baseline MRIs[28,29]. Nevertheless, segmenting medical images with 3D networks still remains a challenge. The 3D networks suffer from issues like gradient vanishing and lack of generalization, due to their complex architectures. Training a large number of parameters requires a large number of training samples, which is not typically available when utilizing real clinical images. As a result, most prior DL methods[20,30,31] were developed in 2D architectures, ignoring the 3D contextual information. In addition, they were mostly trained and/or tested on 2D lesioned brain slices; therefore, they are not generalizable in clinical applications where there is no prior information about the lesion location. More recently, “2.5D” networks[32], 3D dual-path local patch-wise CNNs, DeepMedic[31], and other CNNs proposed in the Ischemic Stroke Lesion Segmentation (ISLES) challenge[18], aimed to improve local segmentation. However, they still do not fully utilize the whole-brain 3D contextual information which might lead to more false positives, especially in “lesion-like” artifacts commonly seen in clinical DWIs. Furthermore, previous networks were mostly evaluated on segmenting late subacute and chronic ischemic abnormalities in high-resolution MRIs, with research-suited homogenous protocols[32-34] which do not reflect the challenges of real clinical data. To our best knowledge, Zhang et al.[35] were the first to reveal the potential of the 3D whole-brain dense networks on low-resolution clinical acute lesion segmentation. However, the relatively modest sample size utilized (242 clinical low-resolution images for training, validating and testing the models, plus 36 downsampled high-resolution images to test for generalization), compared to the large amounts of trainable parameters and high variations of stroke biological features, raises questions about the generalization and robustness of the derived models. The requirement of large amounts of data for training and testing DL models is a common challenge for achieving good generalization and efficiency in practical applications. Artificial data augmentation does not necessarily mirror the biological heterogeneity of strokes, and imperfectly reflects the noise, low resolution, and technical variability of clinical data. Models developed in modest and homogeneous samples, or in artificial augmented data, might be less efficient in heterogeneous and noisy clinical samples[36]. Lastly, the lack of accessibility reduces the potential generalization and translation of previous methods: currently, there is no platform that allows regular users (not imaging experts) to readily and rapidly detect and segment diffusion abnormalities in acute stroke clinical images. In this study, we developed DL network ensembles for the detection and segmentation of lesions from acute and early subacute ischemic strokes in brain MRIs. Our network ensembles were trained and validated in 1390 clinical 3D images (DWIs) and tested in 459 images. The lesion core was manually delineated in all the images, serving as “ground truth”. Furthermore, we evaluated false positives on extra 499 DWIs of brains with “not visible” lesions. To our best knowledge, this is the largest clinical set ever used. We also tested our trained models on an external population of 140 individuals scanned at the hyperacute stroke phase and 24 h later (STIR dataset[37]). Our results show that our model ensembles are in general comparable to 3D pathwise CNNs (as “DeepMedic”) for lesion segmentation, performing superiorly for detection and segmentation of small lesions, with much lower false-positive rate. Our model is generalizable and robust on the external dataset, over the acute phases. Finally, our model is readily and publicly available as a tool with minimal implementation requirements, enabling non-expert users to detect and segment lesions in about one minute in their local computers, with a single command line.

Methods

Cohort

This study included MRIs of patients admitted to the Comprehensive Stroke Center at Johns Hopkins Hospital with the clinical diagnosis of ischemic stroke, between 2009 and 2019 (flowchart for data inclusion in Fig. 1). It utilizes data from an anonymized dataset (IRB00228775), created under the waiver of informed consent because the image is anonymized. We have complied with all relevant ethical regulations and the guidelines of the Johns Hopkins Institutional Review Board, that approved this study (IRB00290649).
Fig. 1

Flowchart of data and analysis.

The traced box shows the independent testing samples.

Flowchart of data and analysis.

The traced box shows the independent testing samples. From the 2348 DWIs quality-controlled for clinical analysis, 499 images did not show visible lesion. These mostly include transitory ischemic strokes (TIA) or strokes with volume under the image resolution. They are called images with “not visible” lesions and were used to calculate false positives in this study. The other 1849 DWIs showed lesions classified by a neuroradiologist as a result of acute or early subacute ischemic stroke, with no evidence of hemorrhage. The lesion core was defined in DWI, in combination with the apparent diffusion coefficient maps (ADC) by two experienced evaluators and was revised by a neuroradiologist until reaching a final decision by consensus. This manual definition was saved as a binary mask (stroke = 1, brain and background = 0) in the original image space of each subject. The DWIs with ischemic lesions were randomly split into a training dataset (n = 1390), which was used to train and validate all subsequent models, and testing dataset (n = 459), which was used exclusively for testing models. A second external dataset, STIR, was used to test the generalization of our models in a completely unrelated population (flowchart in Fig. 1). We have complied with all relevant STIR regulations for data usage. From the STIR dataset, we included 140 subjects scanned twice: at the hyperacute stroke event (here called “STIR 1”) and up to 24 h later (“STIR 2”). The longitudinal aspect of STIR enables to evaluate the performance of our models according to the stroke’s onset. The demographic, lesion, and scanner profiles for all the datasets are summarized in Table 1. The distribution according to arterial territories (MCA > PCA > VEB > ACA) and the demographic characteristics reflect the general population of stroke patients. MRIs were obtained on seven scanners from four different manufacturers, in different magnetic fields (1.5T (60%) and 3T), with more than a hundred different protocols. The DWIs had high in-plane (axial) resolution (1 × 1 mm, or less), and typical clinical high slice thickness (ranging from 3 to 7 mm). Although a challenge for imaging processing, the technical heterogeneity promotes the potential generalization of the resulting developed tools. The demographics (age, sex, race, time from symptoms onset, and NIHSS) and lesion characteristics (e.g., volume, intensity, location) were balanced between training and testing ischemic sets (Supplementary Table 1).
Table 1

Population, lesion, and scanner profiles, per dataset.

DatasetInternal datasetExternal dataset
Training—ischemicTesting—ischemicTesting—not visibleSTIR 1STIR 2
Number of subjects1390459499140
Age in years (median [IQR])62.5 [52, 73]62.0 [54, 72]61.0 [52, 71]73.0 [60.81]
Sex
 Male739 (53.2%)256 (55.8%)251 (50.3%)64 (45.7%)
 Female651 (46.8%)203 (44.2%)248 (49.7%)76 (54.3%)
Race/ethnicity
 African American591 (42.5%)220 (47.9%)236 (47.3%)38 (27.1%)
 Caucasian384 (27.6%)138 (30.1%)190 (38.1%)98 (70.0%)
 Asian30 (2.2%)12 (2.6%)9 (1.8%)4 (3.9%)
 Missing data385 (27.7%)89 (19.4%)64 (12.8%)0 (0.0 %)
NIHSS (median [IQR]; missing)4.0 [1.0, 8.0]; 6813.0 [1.0, 8.25]; 1151.0 [0.0, 3.0]; 25310.0 [5.5, 16.5]; 1
Symptoms onset to MRI in hours
 <265 (4.7%)37 (8.1%)27 (5.4%)
 2–6169 (12.2%)65 (14.2%)63 (12.6%)
 6–12122 (8.8%)106 (23.1%)60 (12.0%)<3 h~24 h
 12–24349 (25.1%)143 (31.2%)177 (35.5%)
 >24108 (7.8%)12 (2.6%)43 (8.6%)
 Missing data577 (41.5%)96 (20.9%)129 (25.9%)
Lesioned hemisphere
 Left627 (45.1%)196 (42.7%)80 (57.1%)
 Right540 (38.9%)210 (45.8%)N.A.57 (40.7%)
 Bilateral223 (16.0%)53 (11.5%)3 (2.2%)
Vascular territories
 MCA806 (58.0%)276 (60.1%)111 (79.3%)
 PCA258 (18.6%)79 (17.2%)N.A.12 (8.6%)
 VB235 (16.9%)67 (14.6%)15 (10.7%)
 ACA91 (6.5%)37 (8.1%)2 (1.4%)
Lesion volume in ml (median [IQR])
 Any vascular territory4.39 [1.05, 21.78]4.62 [1.09, 27.84]7.15 [1.44.28.36]17.48 [3.21.64.43]
 MCA7.49 [1.78, 28.32]6.58 [1.57, 44.23]
 PCA2.53 [0.61, 14.57]2.92 [0.64, 15.16]N.A.
 VB1.35 [0.49, 8.94]1.74 [0.39, 10.75]
 ACA3.39 [0.78, 8.72]2.85 [0.89, 7.39]
Lesion contrast in DWI (median [IQR])
 Any vascular territory3.31 [2.28, 4.51]3.26 [2.23, 4.41]1.67 [1.24, 2.48]3.13 [2.31, 3.99]
 MCA3.28 [2.32, 4.50]3.39 [2.47, 4.48]N.A.
 PCA3.20 [2.12, 4.39]2.87 [2.13, 4.20]
 VB3.72 [2.49, 5.33]3.39 [2.28, 5.16]
 ACA2.76 [2.05, 3.89]2.24 [1.55, 3.31]
MRI manufacturer*
 Manufacturer 1 (Siemens)1207 (86.8%)433 (94.3%)457 (91.5%)0 (0%)
 Manufacturer 2 (Phillips)13 (0.9%)2 (0.4%)5 (1%)39 (27.9%)
 Manufacturer 3 (GE)142 (10.2%)22 (4.8%)29 (5.8%)101 (72.1%)
 Manufacturer 4 (other)28 (2%)2 (0.4%)8 (1.6%)0 (0%)
MRI magnetic field*
 1.5 T930 (66.9%)269 (58.6%)315 (63.1%)104 (74.3%)
 3.0 T460 (33.1%)190 (41.4%)184 (36.9%)36 (25.7%)
Voxel size in mm3 (median [IQR])
 Height/width1.20 [0.90, 1.29]1.20 [0.60, 1.23]1.20 [0.90, 1.30]0.88 [0.86, 0.94]0.88 [0.86, 0.94]
 Thickness5.00 [4.0, 5.0]5.00 [4.0, 5.0]5.00 [4.0, 5.0]7.00 [4.0, 7.0]7.00 [4.0, 7.0]

ACA, PCA, MCA stand for anterior, posterior, and middle cerebral artery territories, VB stands for vertebro-basilar territory. IQR stands for interquartile range. Statistical significant differences in distributions between testing and training datasets are marked with “*”; P values are in Supplementary Table 3.

Population, lesion, and scanner profiles, per dataset. ACA, PCA, MCA stand for anterior, posterior, and middle cerebral artery territories, VB stands for vertebro-basilar territory. IQR stands for interquartile range. Statistical significant differences in distributions between testing and training datasets are marked with “*”; P values are in Supplementary Table 3.

Lesion delineation

The lesion delineations were performed using ROIEditor. The evaluators looked for hyperintensities in DWI and/or hypo intensities (<30% average brain intensity) in ADC. Additional contrasts were used to rule out chronic lesions or microvascular white matter disease. A “seed growing” tool in ROIEditor was often used to achieve a broad segmentation, followed by manual adjustments. The segmentation was performed by two individuals highly experienced in lesion tracing (more than 10 years of experience). In addition, they were trained by detailed instructions and illustrative files, in a subset of 130 cases (randomly selected 10% of the dataset). These cases were then revised by a neuroradiologist, discussed with the evaluators, and (blinded) retraced, revised, and re-analyzed after 2 weeks. The interevaluator index of agreement, Dice score, in this set was 0.76 ± 0.14; the intra-evaluator Dice was 0.79 ± 0.12. Although this is a satisfactory level of agreement, it demonstrates that human segmentation is not totally reproducible, even when performed by experienced, trained evaluators. After consistent consensus agreement was achieved in the initial set, the evaluators started working on the whole dataset. The neuroradiologist revised all the segmentation and identified the sub-optimum cases that were subsequently retraced. The segmentations were revised as many times as necessary until reaching a final decision by consensus among all evaluators.

Image preprocessing

To convert the images to a common space, where the next preprocessing steps will take place, we: (1) resampled DWI, B0 (the image in the absence of diffusion gradients), and ADC into 1 × 1 × 1 mm3, (2) skull-stripped with an in-house “UNet BrainMask Network”, (3) used “in-plane” (IP) linear transformations to map the images to the standard MNI (Montreal Neurological Institute) template, (4) normalized the DWI intensity to reduce the variability and increase the comparability among subjects, (5) “down-sampled” (DS) images to reduce the memory resource requirement in the next steps (e.g., DL networks). Details of these procedures follow.

Automated skull stripping—BrainMask network

To decrease computational complexity and time required for skull stripping, we built an “UNet BrainMask Network” using DWI, B0 images, and gold-standard brain masks. The gold standards brain masks were generated as follows: first, all DWI and B0 images were resampled into 1 × 1 × 1 mm3 and skull striped by a level-set algorithm (available under ROIStudio, in MRIStudio[38]), with W5 = 1.2 and 4, respectively (see explanation about the choice of parameters in MRIstudio website[38]). The brain masks were the union of masking on DWI and B0. Then, these brain masks were further manually corrected by our annotators, serving as ground true for the “UNet BrainMask Network”. To train our “UNet BrainMask Network”, all images are mapped to MNI and downsampled to 4 × 4 × 4 mm3. The final brain mask inferenced by the network was then post-processed by the closing and the “binary_fill_holes” function from Python scipy module, upsampled to 1 × 1 × 1 mm3, and dilated by one voxel with image smoothing. The Dice agreement between the “gold-standard” brain masks and those obtained with our network was above 99.9%, in an independent test dataset of 713 brains. The average processing time was about 19 s (against 4.3 min taken by the level-set algorithm[38]), making it suitable for large-scale, fast processing.

IP-MNI and IP-MNI DS space

We used sequential linear transformations (3D rigid translation followed by scaling/rotation along x and y axis) to map B0, less affected by the acute stroke, into JHU_MNI_B0[38], a template in MNI space. No rotation was performed along the interslice coordinate (z axis), aiming to preserve the image contrast and the continuity of manual annotations, and to ameliorate issues related to the low resolution and voxel anisotropy. We called this standardized space “in-plane MNI"(IP-MNI). The IP-MNI images were then padded into 192 × 224 × 192 voxels, to facilitates the next step of “down-sampling” (DS) by (2, 2, 4) in (x, y, z) axes and max-pooling operations, which reduces the memory resource requirement in DL networks. The voxel dimension of the downsampled images are 96 × 112 × 48 voxels, which is called IP-MNI DS space.

DWI intensity normalization

Intensity normalization increases the comparability between subjects and, as normalizing images to a standardized space, is crucial for diverse image analytical processes. Although the lesion might affect intensity distribution, we assume that the majority of brain voxels are from healthy tissue and can be a good reference for intra- and interindividual comparison. We used bimodal Gaussian function in Eq. (1) to fit the intensity histogram of DWI and cluster two groups of voxels: the “brain tissue” (the highest peak) and “non-brain tissue” (the lowest peak at lowest intensities, composed mostly by cerebrospinal fluid).where a, b, c are the coefficients of the scale, mean, and standard deviation of Gaussian distribution. a, b, c are calculated by least-square fitting the bimodal Gaussian function to the intensity histogram of individual DWI. DWI intensities are normalized to make the “brain tissue” intensity with zero mean and one standard deviation. Supplementary Fig. 1A shows that the DWI intensity distribution of voxels in a brain with ischemic lesions (blue) and in a brain with “not visible” lesion (orange) prior to (left column), and post-to (right column) intensity normalization. We note that the preservation of the minor peak at high intensities in the brain with ischemic lesion indicates the preservation of the lesion contrast after normalization. Supplementary Fig. 1B–D shows the distribution of DWI intensities in groups of images, prior to and post-to intensity normalization. We note that the distributions are much more homogeneous, and the individual variations are smaller after intensity normalization. More importantly, intensity differences between different magnetic fields and scan manufacturers are ameliorated after intensity normalization. Our choice for this intensity normalization approach is, therefore, based on its success to capture the contrast between lesioned voxels and normal tissue voxels, while minimizing variations in intensity range across subjects. The influence of other intensity normalization approaches, from simple z score normalization up to different self-supervised methods[39,40], is subject to further evaluation.

Lesion segmentation with unsupervised (“classical”) methods

We tried two “classical” methods for lesion segmentation: (1) “t-scores” maps of DWI and ADC for individual brains[14], via comparing them to the set of brains with “not visible” lesions, and (2) ischemic probabilistic maps of abnormal voxels, here called “IS”, via the modified c-fuzzy algorithms[15].

T-score method[14]

After mapping all the images to a common template (JHU-EVE MNI[38]), we computed the t-score maps of DWI and ADC, t, t, for each individual s compared to all controls, as in Nazari-Farsani et al.[14]. The images were smoothed by a Gaussian filter with different full width at half-maximum (FWHM), W. The lesion voxels were detected by thresholding the t-scores maps (i.e., t > σ, t > σ) and thresholding DWI intensities individually (i.e., t > σ). These three thresholds and FWHM were optimized by cross-validation in the training dataset. We searched all parameters’ configurations as follows: W ∈ {2, 3, 4, 5}, σ ∈ {1, 1.5, 2, 2.5, 3, 3.5}, σ ∈ {1, 1.5, 2, 2.5, 3, 3.5}, σ ∈ {1.5, 2, 2.5, 3, 3.5}. σ, σ, and σ are in z score scale, rather than percentage. W is in the unit of the pixel. The best model was W = 4, σ = 2, σ = 1, σ = 3.5. The Dice and Net Overlap (defined as[14]) of the top four configurations are summarized in Supplementary Table 2.

Modified c-fuzzy method[15]

We first created intensity mean and standard deviation “templates” in IP-MNI space, I(x, y, z) and I(x, y, z) for s = dwi and s = adc, separately, using the normalized DWI and ADC of cases with “not visible” lesions. We then modified and computed the dissimilarity equation for individual DWI and ADC, I(x, y, z) for s = dwi and s = adc, compared to I(x, y, z) and I(x, y, z) similar as in Guo et al.[15].where α is the parameter used for controlling sensitivity. The probability map of abnormal low-intensity (H1) and high-intensity (H2) voxels were computed as: To generate the probabilistic maps of ischemic voxels, P, we looked for these voxels to have abnormal high intensity (H2) in DWI and abnormal low intensity (H1) in ADC compared to the templates, and also relatively high intensity in its own individual DWI, is defined as follows:where Q(⋅) is the standard normal distribution Q function, and t(x, y, z) is the t-score of DWI voxel intensity, compared to the mean and standard deviation of the whole-brain voxels, μ and σ, in each individual DWI. Like what was described for the t-score method, the DWI and ADC images were smoothed by a Gaussian filter with different FWHM. The six following parameters’ combinations were optimized by cross-validation in training dataset: W ∈ {2, 3, 4, 5}, α ∈ {0.5, 1, 1.5, 2, 2.5}, λ ∈ {2, 3, 4}, α ∈ {0.5, 1, 1.5, 2, 2.5}, λ ∈ {2, 3, 4}, σ ∈ {2, 2.5, 3, 3.5}. σ is in z score scale, rather than percentage. W is in the unit of the pixel. The best parameters’ configuration was W = 2, α = 1.5, λ = 4, α = 0.5, λ = 2, σ = 2. The performance in all datasets is summarized in Supplementary Table 3. As shown in Supplementary Fig. 2, the modified c-fuzzy method performed better than the t-score method for lesion segmentation. In addition, the t-score method performed worse on our testing dataset (Dice about 0.37) than what is described by the developers[14] (Dice about 0.5) Therefore, the classical t-score method was considered insufficiently efficient to segment lesions in our large and heterogeneous clinical dataset.

Lesion segmentation with Deep Learning: DAGMNet implementation details

3D DAGMNet architecture

The architecture of our proposed 3D DAGMNet, depicted in Fig. 2, is equipped with intraskip connections as UNet3+[24], fused multiscale contextual information block, deep supervision, L1 regularization on final predicts, dual attention gate (DAG)[25-27], self-normalized activation (SeLU)[41], and batch normalization. The details of the important components and training techniques/parameters are outlined in the following subsections.
Fig. 2

Our proposed model for lesion detection and segmentation.

a The architecture of the DAGMNet. In DAG, sAG stands for spatial attention gate and cAG stands for channel attention gate. “N" denotes the number features (N = 32 for our final deployed DAGMNet). b Flowchart of the input image’s dimension for training networks and flowchart of the lesion predict’s dimension for inferencing networks in IP-MNI (high-resolution) space.

Our proposed model for lesion detection and segmentation.

a The architecture of the DAGMNet. In DAG, sAG stands for spatial attention gate and cAG stands for channel attention gate. “N" denotes the number features (N = 32 for our final deployed DAGMNet). b Flowchart of the input image’s dimension for training networks and flowchart of the lesion predict’s dimension for inferencing networks in IP-MNI (high-resolution) space. DAGMNet was intuitively designed to capture semantical features directly from input images at different receptive scale levels, as MNet, and to segment lesions of various volumes with consistent efficiency, via deep supervision at each level. This aims to conquer the drawback that although big lesions can be easily detected at different receptive scale levels in the original generic UNet, small lesions features could be dropped after downsampling pathway/max-pooling layer. Furthermore, The introduction of the interskip connections between layers in UNet structure and intraskip connections as UNet3+ help model share and re-utilize features between different receptive scale levels with lower computational complexity than DenseUNet and UNet++. The final fuse block combines all-scale semantic features (from small lesion volume to large ones) to generate the final predict output. At the end of the fusion block, L1-regularization on predicts prevents the networks from false-positive claiming. To overcome the high variability in lesion volume, shape, and location, DAG was utilized to condition the networks to emphasize the most informative components (both spatial and channel-wise) from encoders’ semantic features at each level prior to decoders, increasing the sensitivity to small lesions or lesions with subtle contrast. In DAG, spatial attention gate (sAG) was used to spatially excite the receptive for the most abnormal voxels, like the hyperintensity in DWI or predicts from the third channel ("IS"). On the other hand, the channel attention gate (cAG) was included to excite the most semantical/morphological features associated to ischemic lesions from artifacts. The inclusion of information from classical methods (“IS”) as the third channel aimed to help the networks to focus on abnormal voxels, even if in small clusters (small lesions). In addition, the batch-normalization technique and SeLU activation function aim to self-normalize the networks, avoiding the gradient vanish problem usually faced in complex 3D networks.

Encoder and decoder

As shown in Fig. 2, each encoder or decoder convolution block is composed with two consecutive convolution layers with batch normalization and SeLU activation. The number of features at the first level is denoted as “N”. For our proposed model, N was 32. At the encoder part, input images were downsampled by (2, 2, 2) to each receptive scale level and encoded by an encoder convolution block with N = 32 for each level. At the second, third, and fourth levels, the encoded features from the previous level were concatenated with the features at the present level and furthered encoded by an encoder block with feature number 2 × N, 4 × N, and 8 × N, respectively. At the decoder part, each level has an decoder convolution block with deep supervision. The feature numbers of decoders at the first, second, third, and fourth levels are N, 2 × N, 4 × N, and 8 × N. The input for decoders at the first, second, and third-level are features from dual attention gate concatenated with the upsampled decoded features from the second, third, and fourth level, respectively. The decoder at the fourth level only takes the features from dual attention gate as input at the same level. The final predict is fused by a fuse convolution block at the end. The fuse convolution block takes concatenated features decoded from decoders at each level as inputs. The intraskip connection at each level to fuse convolution block was implemented by transposing up-sampling layers.

Dual attention gate (DAG)

DAG has two parts: channel attention gate (cAG) and spatial attention gate (sAG). Both have two parts, squeezing and exciting parts as depicted in Fig. 2a. cAG block could be considered as a self-attention function intrinsically exciting the most important feature channels. cAG squeeze all spatial features into a channel-wise descriptor by a global average pooling layer and a global max-pooling layer. The most activated local spatial receptive fields would lead to the highest global average/max for its corresponding channel. Then the channel-wise dependencies are calibrated by a fully connected dense layer with sigmoid activation to generate a set of weights of the aggregated spatial information. Each channel will be excited by its corresponding weight post-to the cAG block. sAG block could be considered as a self-attention function intrinsically exciting the most important receptive fields. sAG squeezes channel-wise features into spatial statistics with Global Maximum Pooling (GM pooling), Global Average Pooling (GA pooling), and 1 × 1 × 1 convolution layer. Any local spatial semantical features will be excited by a 5 × 5 × 5 convolution layer with SeLU activation to condition network’s attention locally. The sum of both excited spatial local features from both inputs is further calibrated by a 1 × 1 × 1 convolution layer with sigmoid activation. sAG and cAG both take two inputs: “input 1” from low-level encoded features directly from images at the same receptive scale, “input 2” is all accumulated encoded features from all previous levels, including the current scale level.

SeLU: self-normalization activation function

Compared to rectified linear unit (ReLU), SeLU activation function was reported to have more power to reduce gradient vanish problem among more complicated network structures. Besides, its self-normalization makes the network learn more from morphological features and be less dependent on image contrast. Hence, we used SeLU activation function in our proposed networks.

Deep supervision

Deep supervision is adopted in our model to learn hierarchical representations at each level. Each decoder and the fuse block are fed into a 3-by-3-by-3 convolution layer with sigmoid activation to generate side output, which is supervised by our ground true lesion annotation at its corresponding level.

Loss function

To ameliorate the imbalanced voxel classes issue (between the number of lesion and non-lesion voxels) and regularize the false-positive rate predicted by networks, a hybrid loss function described in Eq. (6) was utilized to train our proposed models.where L is the loss function supervised at the final output of the fusion block X and L is the loss function supervised at the side output of the decoders at each level X as follows:where w = 1, w = 1, w = 1e − 5 due to hyperparameter tuning, L is the generalized dice loss function defined as[42], L is the balanced binary cross-entropy defined as[21], and L1(p) is the L1 regularization on all predicted voxels defined aswhere p( is the predicts from networks at (x, y, z) coordinates. The loss function is optimized during training step by ADAM optimizer with learning rate = 3E − 4. Learning rate will be factor by 0.5 when loss function is on a plateau over 5 epochs with minimum learning rate = 1E − 5.

The dimension of the inputs and predicts during training and inferencing

The networks were trained and inferenced in IP-MNI DS space, which is 96 × 112 × 48 voxels. All images (DWI, ADC, IS) in IP-MNI space were downsampled (DS) along x, y, and z axis with stride of (2, 2, 4) into 16 smaller volumes in 96 × 112 × 48 × 3 voxels for 3-channel models, as shown in Fig. 2b. The input shape of networks with 3 channels is 96 × 112 × 48 × 3. During training step, as shown in Fig. 2b, one of these 16 downsampled volumes are randomly selected to be the inputs of 3-channel networks for each subject in a selected batch (batch size = 4). This re-sampling strategy aims to increase the robustness of our networks to the image’s spatial shifting and inhomogeneity. To make efficient backpropagation for training networks, the downsampled volumes were standard normalized to zero mean and unit-variant within the brain mask region for both DWI and ADC channels. During inferencing step, the 16 lesion predicts from networks in IP-MNI DS space were stacked according to the way their inputs volumes were downsampled in the original coordinates, to construct the final predict in IP-MNI space (Fig. 2c). Then, the lesion mask was “closed” with connectivity 1 voxel. The predicted lesion binary mask (predicted value > 0.5) in IP-MNI space was then mapped back to individual original space. We refined the final prediction by removing clusters with <5 pixels in each slice, which is the smallest size of lesions defined by human evaluators.

Hyperparameter

For searching hyperparameters, such as weights for loss functions and networks structures, 20% of subjects from the training dataset were randomly selected as validation dataset, with the same random states for all experiment models. In each experiment, once the loss functions converged in validation dataset along the training epochs (200 epochs at top, early stops at 100 epochs if training and validation loss function converged early), we selected the best model from the snapshot models at every 10 epochs in the validation dataset. We chose maximum training epoch as 200 because most experiments of benchmark models converged after 80 to 120 epochs. For each experiment, we trained the same-type networks independently, with different training set and different resampled validation set, at least twice to check if similar performance would be achieved and avoid overfitting. Once the networks parameters (including weights for loss or regularization, different network layers, depth...etc.) were finalized according to their best dice performance in the validation sets, we used the whole training dataset, including the validation dataset, to train the final deployed models and to make the loss function and dice scores converge to the similar level as the previous experiment. This allowed us to fully use all training dataset and capture the population variation. All Testing datasets and the STIR datasets were totally unexploited till this step was done.

System implementation

All the evaluated methods and models were built with TensorFlow[43] (tensorflow-gpu version is 2.0.0) and Keras[44] (2.3.1) framework on Python 3.6. Imaging processing and analysis were built with Nibabel[45], Scipy[46], Dipy[47], and Scikit-image[48,49]. The experiments run on a machine with an Intel Core (Intel(R) Xeon(R) CPU E5-2620 v4 @ 2.10GHz) with 2 NVIDIA TITAN XP GPUs (with CUDA 10.1). The average inference time over all testing dataset and memory requirements are listed in Table 2.
Table 2

Performance summary.

MetricsDatasetDeepMedicDAGMNet_CH3DAGMNet_CH2UNet_CH3UNet_CH2FCN_CH3FCN_CH2
Dice scoreTesting (n = 459)0.74 (0.17); 0.790.76 (0.16); 0.810.75 (0.17); 0.800.75 (0.18); 0.810.74 (0.20); 0.800.68 (0.20); 0.720.66 (0.20); 0.71
STIR 2 (n = 140)0.76 (0.18); 0.820.75 (0.21); 0.820.75 (0.21); 0.810.73 (0.24); 0.820.73 (0.24); 0.820.70 (0.22); 0.750.68 (0.24); 0.75
STIR 1 (n = 140)0.55 (0.27); 0.600.51 (0.30); 0.590.48 (0.32); 0.580.49 (0.31); 0.590.48 (0.32); 0.580.49 (0.28); 0.550.44 (0.30); 0.46
Testing L (n = 163)0.85 (0.09); 0.870.83 (0.10); 0.860.84 (0.09); 0.860.85 (0.09); 0.880.84 (0.10); 0.870.81 (0.10); 0.840.80 (0.11); 0.83
STIR 2 L (n = 76)0.84 (0.13); 0.880.81 (0.18); 0.880.82 (0.16); 0.890.81 (0.20); 0.890.81 (0.21); 0.880.79 (0.18); 0.840.77 (0.21); 0.86
STIR 1 L (n = 50)0.67 (0.25); 0.780.64 (0.28); 0.770.64 (0.29); 0.790.59 (0.30); 0.720.62 (0.30); 0.760.61 (0.28); 0.740.59 (0.30); 0.73
Testing M (n = 144)0.74 (0.13); 0.760.75 (0.14); 0.800.74 (0.14); 0.770.76 (0.14); 0.790.74 (0.16); 0.790.67 (0.15); 0.710.66 (0.15); 0.68
STIR 2 M (n = 43)0.73 (0.13); 0.770.75 (0.13); 0.770.75 (0.15); 0.780.72 (0.20); 0.780.71 (0.22); 0.770.66 (0.16); 0.700.66 (0.16); 0.70
STIR 1 M (n = 51)0.53 (0.24); 0.590.49 (0.28); 0.590.43 (0.30); 0.500.45 (0.31); 0.570.42 (0.30); 0.370.47 (0.24); 0.530.39 (0.26); 0.42
Testing S (n = 152)0.63 (0.18); 0.670.68 (0.19); 0.73*0.66 (0.22); 0.720.65 (0.22); 0.720.62 (0.25); 0.690.54 (0.22); 0.580.51 (0.22); 0.56
STIR 2 S (n = 21)0.52 (0.21); 0.560.51 (0.25); 0.550.48 (0.27); 0.510.49 (0.28); 0.550.53 (0.26); 0.620.45 (0.24); 0.480.40 (0.22); 0.42
STIR 1 S (n = 39)0.43 (0.25); 0.520.37 (0.29); 0.480.34 (0.31); 0.290.41 (0.31); 0.480.38 (0.32); 0.470.37 (0.25); 0.420.32 (0.27); 0.38
PrecisionTesting (n = 459)0.76 (0.21); 0.820.83 (0.17); 0.88*0.81 (0.18); 0.870.80 (0.18); 0.860.81 (0.19); 0.870.70 (0.22); 0.750.68 (0.22); 0.73
STIR 2 (n = 140)0.75 (0.19); 0.790.80 (0.20); 0.87*0.78 (0.20); 0.850.78 (0.21); 0.840.80 (0.19); 0.850.72 (0.20); 0.780.73 (0.20); 0.78
STIR 1 (n = 140)0.62 (0.26); 0.670.62 (0.31); 0.720.55 (0.33); 0.640.65 (0.31); 0.770.66 (0.32); 0.780.57 (0.28); 0.650.57 (0.33); 0.69
SensitivityTesting (n = 459)0.78 (0.17); 0.83*0.73 (0.19); 0.770.74 (0.21); 0.790.76 (0.21);0.830.71 (0.23); 0.780.71 (0.21); 0.770.69 (0.23); 0.76
STIR 2 (n = 140)0.82 (0.21); 0.91*0.76 (0.24); 0.850.78 (0.25); 0.870.76 (0.28); 0.900.75 (0.28); 0.880.74 (0.26); 0.850.72 (0.27); 0.82
STIR 1 (n = 140)0.59 (0.32); 0.650.52 (0.33); 0.620.53 (0.37); 0.650.48 (0.35); 0.530.46 (0.35); 0.530.52 (0.32); 0.610.43 (0.33); 0.41
Subject detection rateTesting (n = 459)1.00 (0.05);0.99 (0.08);0.98 (0.12);0.99 (0.11);0.98 (0.15);0.98 (0.13);0.97 (0.17);
[0.99, 1.00][0.99, 1.00][0.97, 1.00][0.98, 1.00][0.96, 0.99][0.97, 0.99][0.96, 0.99]
STIR 2 (n = 140)0.99 (0.08);0.98 (0.14);0.99 (0.12);0.98 (0.14);0.97 (0.17);0.98 (0.14);0.99 (0.12);
[0.98,1.01][0.95, 1.00][0.97, 1.01][0.95, 1.00][0.94, 1.00][0.95, 1.00][0.97, 1.01]
STIR 1 (n = 140)0.96 (0.20);0.90 (0.30);0.84 (0.36);0.87 (0.33);0.85 (0.36);0.91 (0.29);0.85 (0.36);
[0.92, 0.99]0.85, 0.95[0.78, 0.90][0.82, 0.93][0.79, 0.91][0.86, 0.96][0.79, 0.91]
Spearman correlation of dice and lesion volume sizeTesting (n = 459)0.62 [0.57, 0.68]0.44 [0.37, 0.51]0.48 [0.41, 0.55]0.53 [0.46, 0.59]0.53 [0.46, 0.59]0.63 [0.57, 0.68]0.65 [0.59, 0.70]
STIR 2 (n = 140)0.68 [0.58, 0.76]0.49 [0.36, 0.61]0.54 [0.41, 0.65]0.55 [0.42, 0.65]0.51 [0.37, 0.62]0.60 [0.48, 0.69]0.59 [0.48, 0.69]
STIR 1 (n = 140)0.42 [0.28, 0.55]0.42 [0.27, 0.55]0.42 [0.27, 0.55]0.30 [0.14, 0.44]0.36 [0.21, 0.50]0.44 [0.29, 0.56]0.42 [0.28, 0.55]
Spearman correlation of dice and lesion DWI contrastTesting (n = 459)0.60 [0.54, 0.66]0.65 [0.59, 0.70]0.61 [0.55, 0.66]0.62 [0.56, 0.68]0.64 [0.59, 0.69]0.64 [0.59, 0.69]0.65 [0.59, 0.70]
STIR 2 (n = 140)0.45 [0.31, 0.57]0.57 [0.44, 0.67]0.54 [0.41, 0.65]0.54 [0.41, 0.65]0.55 [0.42, 0.65]0.52 [0.38, 0.63]0.56 [0.43, 0.66]
STIR 1 (n = 140)0.52 [0.38, 0.63]0.56 [0.43, 0.66]0.41 [0.26, 0.54]0.51 [0.37, 0.62]0.40 [0.25, 0.53]0.45 [0.30, 0.57]0.42 [0.28, 0.55]
Spearman correlation of dice and lesion ADC contrastTesting (n = 459)−0.33 [−0.41, −0.24]−0.48 [−0.55, −0.41]−0.47 [−0.53, −0.39]−0.44 [−0.51, −0.36]−0.46 [−0.53, −0.38]−0.41 [−0.48, −0.33]−0.40 [−0.48, −0.32]
STIR 2 (n = 140)−0.31 [−0.45, −0.15]−0.37 [−0.51, −0.22]−0.36 [−0.50, −0.21]−0.40 [−0.53, −0.25]−0.42 [−0.55, −0.28]−0.38 [−0.51, −0.23]−0.41 [−0.54, −0.26]
STIR 1 (n = 140)−0.24 [−0.39, −0.08]+−0.30 [−0.44, −0.14]−0.27 [−0.42, −0.11]+−0.29 [−0.44, −0.13]−0.30 [−0.44, −0.14]−0.13 [−0.29, 0.03]+−0.20 [−0.35, −0.03]+
Spearman correlation of lesion and predict volume sizeTesting (n = 459)0.97 [0.96, 0.97]0.97 [0.97, 0.98]0.97 [0.96, 0.97]0.97 [0.96, 0.98]0.97 [0.96, 0.97]0.97 [0.96, 0.97]0.97 [0.96, 0.97]
STIR 2 (n = 140)0.97 [0.96, 0.98]0.96 [0.94, 0.97]0.96 [0.94, 0.97]0.93 [0.90, 0.95]0.89 [0.86, 0.92]0.95 [0.93, 0.96]0.94 [0.91, 0.96]
STIR 1 (n = 140)0.87 [0.83, 0.91]0.84 [0.79, 0.89]0.80 [0.73, 0.85]0.81 [0.74, 0.86]0.79 [0.72, 0.85]0.84 [0.78, 0.88]0.79 [0.72, 0.85]
Spearman correlation of lesion and predict DWI contrastTesting (n = 459)0.87 [0.85, 0.89]0.89 [0.86, 0.90]0.88 [0.86, 0.90]0.87 [0.84, 0.89]0.88 [0.86, 0.90]0.86 [0.83, 0.88]0.85 [0.82, 0.87]
STIR 2 (n = 140)0.83 [0.77, 0.88]0.81 [0.74, 0.86]0.85 [0.80, 0.89]0.87 [0.82, 0.90]0.88 [0.84, 0.91]0.83 [0.77, 0.87]0.82 [0.76, 0.87]
STIR 1 (n = 140)0.61 [0.50, 0.71]0.70 [0.61, 0.78]0.59 [0.47, 0.69]0.69 [0.58, 0.77]0.74 [0.65, 0.81]0.59 [0.48, 0.69]0.50 [0.36, 0.62]
Spearman correlation of lesion and predict ADC contrastTesting (n = 459)0.77 [0.74, 0.81]0.84 [0.81, 0.86]0.83 [0.80, 0.86]0.82 [0.79, 0.85]0.83 [0.80, 0.86]0.80 [0.77, 0.83]0.80 [0.76, 0.83]
STIR 2 (n = 140)0.85 [0.80, 0.89]0.86 [0.81, 0.90]0.91 [0.87, 0.93]0.87 [0.82, 0.90]0.93 [0.90,0.95]0.90 [0.86, 0.93]0.84 [0.79, 0.88]
STIR 1 (n = 140)0.51 [0.38, 0.63]0.58 [0.46, 0.68]0.52 [0.38, 0.63]0.55 [0.42, 0.66]0.57 [0.44, 0.67]0.53 [0.39, 0.64]0.47 [0.32, 0.59]
Median of false positivesNot visible (n = 499)14000000
Number of subjects whose FP > 10 voxelsNot visible (n = 499)27513278553618288
False positive subject detection rateNot visible (n = 499)0.55 (0.50);0.26 (0.44);0.16 (0.36);0.11 (0.31);0.07 (0.26);0.36 (0.48);0.18 (0.38);
[0.51, 0.59]*[0.23, 0.30][0.12, 0.19][0.08, 0.14][0.05, 0.09][0.32, 0.41][0.14, 0.21]
False positive subject detection rate (retrospect evaluation)Not visible (n = 499)0.53 (0.50);0.24 (0.43);0.14 (0.35);0.10 (0.30);0.06 (0.24);0.34 (0.47);0.15 (0.36);
[0.48, 0.57]*[0.21, 0.28][0.11, 0.17][0.07, 0.12][0.04, 0.08][0.30, 0.38][0.12, 0.18]
Number of trainable parametersAll24.5 M10.7 M10.7 M10.0 M10.0 M10.1 M10.1M
CPU inference time in secondsTesting (n = 459)85.6830.10 (0.52)29.09 (0.46)19.30 (0.34)18.71 (0.37)7.15 (0.52)6.80 (0.55)
GPU inference time in secondsTesting (n = 459)14.975.91 (0.44)4.82 (0.30)3.78 (0.18)3.59 (0.18)2.40 (0.18)2.26(0.18)

Metrics (dice, precision, sensitivity) are represented as “mean (standard deviation); median”; subject detection rate is represented as “mean (standard deviation); [95% CI]”. The correlations are shown as “correlation coefficient; [95% CI]”. “+” indicates no significant correlations (P value>1E − 3); all the other correlations were significant with P value≤1E − 3. In dataset column, L = large; M = moderate; S = small lesion groups. The statistical significant difference between DAGMNet_CH3 and DeepMedic is labeled by “*”.

Performance summary. Metrics (dice, precision, sensitivity) are represented as “mean (standard deviation); median”; subject detection rate is represented as “mean (standard deviation); [95% CI]”. The correlations are shown as “correlation coefficient; [95% CI]”. “+” indicates no significant correlations (P value>1E − 3); all the other correlations were significant with P value≤1E − 3. In dataset column, L = large; M = moderate; S = small lesion groups. The statistical significant difference between DAGMNet_CH3 and DeepMedic is labeled by “*”.

Evaluation of efficiency and comparison among methods

The performances metrics, Dice score, precision, sensitivity, and subject detection rate (SDR), are defined as follows.TP, FP, and FN, respectively, denote the number of true positives, false positives, and false negatives in the unit of the voxel. The SDR represents the percentage of the subjects whose lesions are detected. Therefore, high SDR indicates good performance in the datasets of subjects with visible lesions (our testing dataset and STIR) but indicates low performance in the dataset of subjects with “not visible” lesions, because the predicted lesion voxels would be false positives. Hence, in the dataset of subjects with “not visible” lesions, it is false-positive SDR. All models were compared to the manual delineation of the stroke (used here as “gold-standard”) on original raw space. All models were trained, validated, and optimized in the same training dataset, and tested in the independent testing dataset and in the external dataset, STIR. For fair comparison, all models were designed to utilize similar number of trainable parameters, except for DeepMedic, whose default network uses more trainable parameters than all other networks (Table 2). Besides comparing global performance in the whole brain, we compared models’ performance according to lesion volume and location (by vascular territories) and analyzed the Spearman correlation coefficient of performance with the lesion volume and contrast in DWI and ADC. The “scipy.stats” module is used to calculate the Spearman correlation coefficient. The 95%CI for SDR is calculated with z score = 1.96 and the 95% CI for the Spearman correlation coefficient is calculated by Fishers’ transformation with z score = 1.96. The lesion contrast in DWI, κ, and in ADC, κ, are defined as the ratio between the average intensity of the lesion voxels and the average intensity of the “normal brain tissue” voxels in DWI and ADC, respectively. High κ and low κ are expected in acute and subacute lesions which tend to be “bright” in DWI and “dark” in ADC. Furthermore, false-positive analysis was conducted on voxel- and subject-basis among the 499 images with “not visible” lesions.

Statistical significance testing

The statistical significance testing was performed by ANOVA test in module “bioinfokit” for continuous data, and by Chi-square test in “scipy.stats.chi2_contingency” module for categorical data. The statistical significance on the demographic, lesion, and image profiles distribution between training and testing datasets is shown in Supplementary Table 1. For race/ethnicity and MRI manufacturer, we compared the major groups. For symptoms onset to MRI time, we compared groups with onset time < 6 and ≥6 h. The statistical comparison among models’ performance in the Testing and STIR dataset is shown in Supplementary Fig. 3. The statistical comparison of models’ performance regarding to demographic, lesion, and image profiles is illustrated in Supplementary Fig. 4.
  30 in total

1.  Lesion load of the corticospinal tract predicts motor impairment in chronic stroke.

Authors:  Lin L Zhu; Robert Lindenberg; Michael P Alexander; Gottfried Schlaug
Journal:  Stroke       Date:  2010-04-08       Impact factor: 7.914

Review 2.  Mapping human brain lesions and their functional consequences.

Authors:  Hans-Otto Karnath; Christoph Sperber; Christopher Rorden
Journal:  Neuroimage       Date:  2017-10-16       Impact factor: 6.556

3.  Automatic Segmentation of Acute Ischemic Stroke From DWI Using 3-D Fully Convolutional DenseNets.

Authors:  Rongzhao Zhang; Lei Zhao; Wutao Lou; Jill M Abrigo; Vincent C T Mok; Winnie C W Chu; Defeng Wang; Lin Shi
Journal:  IEEE Trans Med Imaging       Date:  2018-03-30       Impact factor: 10.048

4.  Tissue at Risk and Ischemic Core Estimation Using Deep Learning in Acute Stroke.

Authors:  Y Yu; Y Xie; T Thamm; E Gong; J Ouyang; S Christensen; M P Marks; M G Lansberg; G W Albers; G Zaharchuk
Journal:  AJNR Am J Neuroradiol       Date:  2021-03-25       Impact factor: 4.966

5.  Fully automatic acute ischemic lesion segmentation in DWI using convolutional neural networks.

Authors:  Liang Chen; Paul Bentley; Daniel Rueckert
Journal:  Neuroimage Clin       Date:  2017-06-13       Impact factor: 4.881

6.  Fully Automatic Segmentation of Acute Ischemic Lesions on Diffusion-Weighted Imaging Using Convolutional Neural Networks: Comparison with Conventional Algorithms.

Authors:  Ilsang Woo; Areum Lee; Seung Chai Jung; Hyunna Lee; Namkug Kim; Se Jin Cho; Donghyun Kim; Jungbin Lee; Leonard Sunwoo; Dong Wha Kang
Journal:  Korean J Radiol       Date:  2019-08       Impact factor: 3.500

7.  A comparison of automated lesion segmentation approaches for chronic stroke T1-weighted MRI data.

Authors:  Kaori L Ito; Hosung Kim; Sook-Lei Liew
Journal:  Hum Brain Mapp       Date:  2019-07-26       Impact factor: 5.038

8.  Use of Deep Learning to Predict Final Ischemic Stroke Lesions From Initial Magnetic Resonance Imaging.

Authors:  Yannan Yu; Yuan Xie; Thoralf Thamm; Enhao Gong; Jiahong Ouyang; Charles Huang; Soren Christensen; Michael P Marks; Maarten G Lansberg; Gregory W Albers; Greg Zaharchuk
Journal:  JAMA Netw Open       Date:  2020-03-02

9.  Does stroke location predict walk speed response to gait rehabilitation?

Authors:  P Simon Jones; Valerie M Pomeroy; Jasmine Wang; Gottfried Schlaug; S Tulasi Marrapu; Sharon Geva; Philip J Rowe; Elizabeth Chandler; Andrew Kerr; Jean-Claude Baron
Journal:  Hum Brain Mapp       Date:  2015-11-19       Impact factor: 5.038

10.  Attention gated networks: Learning to leverage salient regions in medical images.

Authors:  Jo Schlemper; Ozan Oktay; Michiel Schaap; Mattias Heinrich; Bernhard Kainz; Ben Glocker; Daniel Rueckert
Journal:  Med Image Anal       Date:  2019-02-05       Impact factor: 8.545

View more
  2 in total

1.  Development of deep learning-assisted overscan decision algorithm in low-dose chest CT: Application to lung cancer screening in Korean National CT accreditation program.

Authors:  Sihwan Kim; Woo Kyoung Jeong; Jin Hwa Choi; Jong Hyo Kim; Minsoo Chun
Journal:  PLoS One       Date:  2022-09-29       Impact factor: 3.752

2.  Left Hemisphere Bias of NIH Stroke Scale Is Most Severe for Middle Cerebral Artery Strokes.

Authors:  Emilia Vitti; Ganghyun Kim; Melissa D Stockbridge; Argye E Hillis; Andreia V Faria
Journal:  Front Neurol       Date:  2022-06-14       Impact factor: 4.086

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.