| Literature DB >> 30801055 |
Ghalib A Bello1, Timothy J W Dawes1,2, Jinming Duan1,3, Carlo Biffi1,3, Antonio de Marvao1, Luke S G E Howard4, J Simon R Gibbs2,4, Martin R Wilkins5, Stuart A Cook1,2,6, Daniel Rueckert3, Declan P O'Regan1.
Abstract
Motion analysis is used in computer vision to understand the behaviour of moving objects in sequences of images. Optimising the interpretation of dynamic biological systems requires accurate and precise motion tracking as well as efficient representations of high-dimensional motion trajectories so that these can be used for prediction tasks. Here we use image sequences of the heart, acquired using cardiac magnetic resonance imaging, to create time-resolved three-dimensional segmentations using a fully convolutional network trained on anatomical shape priors. This dense motion model formed the input to a supervised denoising autoencoder (4Dsurvival), which is a hybrid network consisting of an autoencoder that learns a task-specific latent code representation trained on observed outcome data, yielding a latent representation optimised for survival prediction. To handle right-censored survival outcomes, our network used a Cox partial likelihood loss function. In a study of 302 patients the predictive accuracy (quantified by Harrell's C-index) was significantly higher (p = .0012) for our model C=0.75 (95% CI: 0.70 - 0.79) than the human benchmark of C=0.59 (95% CI: 0.53 - 0.65). This work demonstrates how a complex computer vision task using high-dimensional medical image data can efficiently predict human survival.Entities:
Keywords: Heart Failure; Hypertension, Pulmonary; Machine Learning; Magnetic Resonance Imaging, Cine; Motion; Survival
Year: 2019 PMID: 30801055 PMCID: PMC6382062 DOI: 10.1038/s42256-019-0019-2
Source DB: PubMed Journal: Nat Mach Intell ISSN: 2522-5839
Patient characteristics at baseline (date of MRI scan). WHO, World Health Organization; BP, Blood pressure; LV, left ventricle; RV, right ventricle.
| Characteristic | n | % or Mean ±SD |
|---|---|---|
| Age (years) | 62.9 ±14.5 | |
| Body surface area ( | 1.92 ±0.25 | |
| Male | 169 | 56 |
| Race | ||
| Caucasian | 215 | 71.2 |
| Asian | 7 | 2.3 |
| Black | 13 | 4.3 |
| Other | 28 | 9.3 |
| Unknown | 39 | 12.9 |
| WHO functional class | ||
| I | 1 | 0 |
| II | 45 | 15 |
| III | 214 | 71 |
| IV | 42 | 14 |
| Haemodynamics | ||
| Systolic BP (mmHg) | 131.5 ±25.2 | |
| Diastolic BP (mmHg) | 75 ±13 | |
| Heart rate (beats/min) | 69.8 ± 22.5 | |
| Mean right atrial pressure (mmHg) | 9.9 ±5.8 | |
| Mean pulmonary artery pressure (mmHg) | 44.1 ±12.6 | |
| Pulmonary vascular resistance (Wood units) | 8.9 ±5.0 | |
| Cardiac output (l/min) | 4.3 ±1.5 | |
| LV Volumetry | ||
| LV ejection fraction (%) | 61 ± 11.1 | |
| LV end diastolic volume index (ml/m) | 110 ± 37.4 | |
| LV end systolic volume index (ml/m) | 44 ± 22.9 | |
| RV Volumetry | ||
| RV ejection fraction (%) | 38 ± 13.7 | |
| RV end diastolic volume (ml/m) | 194 ± 62 | |
| RV end systolic volume (ml/m) | 125 ± 59.3 | |
| RV Strain | ||
| Longitudinal (%) | -16.8 ± 4.7 | |
| Radial (%) | +18.0 ± 4.4 | |
| Circumferential (%) | -9.6 ± 7.0 |
Figure 1A) An example of an automatic cardiac image segmentation of each short-axis cine image from apex (slice 1) to base (slice 9) across 20 temporal phases. Data were aligned to a common reference space to build a population model of cardiac motion. B) Trajectory of right ventricular contraction and relaxation averaged across the study population plotted as looped pathlines for a sub-sample of 100 points on the heart (magnification factor of x4). Colour represents relative myocardial velocity at each phase of the cardiac cycle. A surface-shaded model of the heart is shown at end-systole. These dense myocardial motion fields for each patient were used as an input to the prediction network. LV, left ventricular; RV, right ventricular.
Figure 2Kaplan-Meier plots for A) a conventional parameter model using a composite of manually-derived volumetric measures, and B) a deep learning prediction model (4Dsurvival) whose input was time-resolved three dimensional models of cardiac motion. For both models, patients were divided into low- and high-risk groups by median risk score. Survival function estimates for each group (with 95% confidence intervals) are shown. For each plot, the Logrank test was performed to compare survival curves between risk groups (conventional parameter model: χ2 = 5.5, p = .019; 4Dsurvival: χ2 = 15.6, p < .0001)
Hyperparameter search ranges for DL network (first column) and optimum hyperparameter values in final model (second column)
| Hyperparameter | Search Range | Optimized Value |
|---|---|---|
| Dropout | [0.1, 0.9] | 0.71 |
| # of nodes in hidden layers | [75, 250] | 78 |
| Latent code dimensionality ( | [5, 20] | 13 |
| Reconstruction loss penalty ( | [0.3, 0.7] | 0.6 |
| Learning Rate | [10−6, 10−4.5] | 10−4.86 |
| | [10−7, 10−4] | 10−5.65 |
Figure 3A) A 2-dimensional projection of latent representations of cardiac motion in the 4Dsurvival network labelled by survival time. A visualisation of RV motion is shown for two patients with contrasting risks. B) Saliency map showing regional contributions to survival prediction by right ventricular motion. Absolute regression coefficients are expressed on a log-scale.
Figure 5The architecture of the segmentation algorithm. A fully convolutional network takes each stack of cine images as an input, applies a branch of convolutions, learns image features from fine to coarse levels, concatenates multi-scale features and finally predicts the segmentation and landmark location probability maps simultaneously. These maps, together with the ground truth landmark locations and label maps, are then used in the loss function (see Equation 1) which is minimised via stochastic gradient descent.
Figure 6The prediction network (4Dsurvival) is a denoising autoencoder that takes time-resolved cardiac motion meshes as its input (right ventricle shown in solid white, left ventricle in red). For the sake of simplicity two hidden layers, one immediately preceding and the other immediately following the central layer (latent code layer), have been excluded from the diagram. The autoencoder learns a task-specific latent code representation trained on observed outcome data, yielding a latent representation optimised for survival prediction that is robust to noise. The actual number of latent factors is treated as an optimisable parameter.