| Literature DB >> 35664012 |
Yang Nan1, Javier Del Ser2,3, Simon Walsh1, Carola Schönlieb4, Michael Roberts4,5, Ian Selby6, Kit Howard7, John Owen7, Jon Neville7, Julien Guiot8,9, Benoit Ernst8,9, Ana Pastor10, Angel Alberich-Bayarri10, Marion I Menzel11,12, Sean Walsh13, Wim Vos13, Nina Flerin13, Jean-Paul Charbonnier14, Eva van Rikxoort14, Avishek Chatterjee15, Henry Woodruff15, Philippe Lambin15, Leonor Cerdá-Alberich16, Luis Martí-Bonmatí16, Francisco Herrera17,18, Guang Yang1,19,20.
Abstract
Removing the bias and variance of multicentre data has always been a challenge in large scale digital healthcare studies, which requires the ability to integrate clinical features extracted from data acquired by different scanners and protocols to improve stability and robustness. Previous studies have described various computational approaches to fuse single modality multicentre datasets. However, these surveys rarely focused on evaluation metrics and lacked a checklist for computational data harmonisation studies. In this systematic review, we summarise the computational data harmonisation approaches for multi-modality data in the digital healthcare field, including harmonisation strategies and evaluation metrics based on different theories. In addition, a comprehensive checklist that summarises common practices for data harmonisation studies is proposed to guide researchers to report their research findings more effectively. Last but not least, flowcharts presenting possible ways for methodology and metric selection are proposed and the limitations of different methods have been surveyed for future research.Entities:
Keywords: Information fusion; data harmonisation; data standardisation; domain adaptation; reproducibility
Year: 2022 PMID: 35664012 PMCID: PMC8878813 DOI: 10.1016/j.inffus.2022.01.001
Source DB: PubMed Journal: Inf Fusion ISSN: 1566-2535 Impact factor: 17.564
Comparison of existing data harmonisation review studies.
| Survey | |||||
| Period | ∼2020 | ∼2019 | ∼2020 | ∼2021 | ∼2021 |
| # of reviewed studies | N/A | 23 | 49 | 42 | 96 |
| Domain | Radiomics | Radiomics | Radiomics | Radiomics | Radiomics, Gene, Pathology |
| Metric | × | × | × | × | √ |
| Checklist | × | × | × | × | √ |
| Guidance | × | √ | × | × | √ |
| Meta-analysis | × | √ | × | × | √ |
“# of reviewed studies” indicates the number of included papers in the survey.
Terminology of computational data harmonisation.
| Terminology | Definitions |
| Cohort | A group of data acquired by the same acquisition protocol and devices |
| Subjects | Patients (objects) involved in the study |
| Category | The classes that were involved in the study, e.g., cancer vs. normal |
| Cases | Samples (a subject can produce multiple samples with different acquisition protocols) involved in the study |
| Cohort bias | The non-biological related variances caused by acquisition protocols (also named as “batch effect” in gene expression studies) |
| Source cohorts | The cohort that needs to be harmonised from |
| Reference cohort | The cohort that needs to be harmonised to |
Fig. 1Visualised differences in (a) radiomics and (b) pathology images. (a) a lung tumour captured on the same CT scanner with 6 different acquisition protocols (From [13]). (b) H&E stained tissue images from different sites [14].
Summary of the reproducibility/repeatability studies.
| Reference | Intra-repro | Inter-repro | Repeatability | Condition | Variables | Object | Modality |
| Jha et al. | 30.7% | 14.3% | 82.2% (888/1080) | ICC >0.90 | Slice Sickness | Phantoms | CT |
| Emaminejad et al. | 8.0% | / | / | CCC>0.90 | Reconstruction | Patients | CT |
| 7.5% | / | / | CCC>0.90 | Radiation Dose | Patients | CT | |
| Kim et al. | 11.0% | / | / | CCC>0.85 | Acceleration Factors | Patients | MRI |
| Ymashita et al. | / | 5.6% | / | CCC>0.90 | Different Scanners | Patients | CECT |
| Fiset et al. | / | 22.6% | / | ICC >0.90 | Different Scanners | Patients | MRI |
| Saeedi et al. | 20.5% | / | / | CoV< 5% | Tube Voltage | Phantoms | CT |
| 30% | / | / | CoV< 5% | Tube Current | Phantoms | CT | |
| Meyer et al. | 20.8% | / | / | Radiation Dose | Patients | CT | |
| 52.8% | / | / | Reconstruction | Patients | CT | ||
| 39.6% | / | / | Reconstruction | Patients | CT | ||
| 12.3% | / | / | Slice Sickness | Patients | CT | ||
| Perrin et al. | 24.8% | / | / | CCC>0.90 | Injection Rates | Patients | CECT |
| 13.4% | / | / | CCC>0.90 | Resolution | Patients | CECT | |
| Midya et al. | 11.7% | / | / | CCC>0.90 | Tube Current | Phantoms | CT |
| 19.8% | / | / | CCC>0.90 | Noise | Phantoms | CT | |
| 63.3% | / | / | CCC>0.90 | Reconstruction | Patients | CT | |
| Altazi et al. | 21.5% | / | / | Mean difference <25% | Reconstruction | Patients | PET |
| Zhao et al. | 11.2% | / | / | CCC>0.90 | Reconstruction | Patients | CT |
| / | / | 69.7% | CCC>0.90 | / | Patients | CT | |
| Hu et al. | / | / | 64.0% | ICC>0.80 | / | Patients | CT |
| Choe et al. | 15.2% | / | / | CCC>0.85 | Reconstruction | Patients | CT |
CCC: concordance correlation coefficient; ICC: intraclass correlation coefficient; CoV: coefficient of variation; : R-squared; CT: computed tomography; MRI: magnetic resonance imaging; CECT: consecutive contrast-enhanced computed tomography; PET: positron emission tomography.
Fig. 2Workflow of developing a computational data harmonisation method.
Fig. 3Literature selection procedure.
Fig. 4Taxonomy of computational data harmonisation strategies.
Fig. 5Illustration of adversarial learning methods.
Fig. 6Taxonomy of harmonisation metrics. The visualization and cohort classification assessment are not presented due to their limited subcategory.
Fig. 8Number of publications and years in terms of data properties and modalities. The public data is the open source data that can be acquired, the in-house data is not available from the internet. The percentage in the top left subfigure is the ratio of studies that were conducted on the public dataset.
Fig. 9Harmonisation strategies in terms of different modalities. ‘IFL’ indicates invariant feature learning approaches, “Img Pro” refers to image processing approaches. The percentage of sub-methods is annotated with the abbreviations of sub-methods in each pie chart.
Fig. 10Evaluation metrics in terms of different modalities.
Data scale (image size) in previous studies.
| Small | Middle | Large | N/A | |
| Radiomics | 9.1% (6) | 18.2% (12) | 3.0% (2) | 69.7% (46) |
| Pathology | 5.0% (1) | 15.0% (3) | 60.0% (12) | 20% (4) |
* The small, middle, large image sizes are defined as , , ε , respectively, N/A indicates there is no report of image size.
Checklist for Computational Data Harmonisation in Digital Healthcare (CHECDHA) criteria.
| Category | Item | Explanation | Example | |
| Motivation | Background | The application field of the dataset(s) | Information fusion of DW-MRI data from different scanners | |
| Importance | Why this study is conducted, how important it is | Dramatically increase the statistical power and sensitivity of clinical studies | ||
| Data | Common | Dataset | What the dataset(s) is (are), how it is (they are) collected (details of acquisition protocols, entry and exit criteria) | |
| Property | Whether the dataset(s) is (are) in-house or public, provide the access link if appropriate | Public/In-house | ||
| Pre-processing | How the dataset is pre-processed | Z-score normalisation | ||
| Ground truth | What the ground truth is and how it is generated | Cohort | ||
| Partition | For machine learning, how the dataset is partitioned into training, validation, and testing subsets in terms of the number of samples, patients | 7:2:1 for training, validation and test | ||
| Augmentation | For machine learning, how the dataset is augmented | Randomized flip, rotation | ||
| Specific | MRI sequence | What the MRI sequence is | Diffusion-weighted | |
| Region | Which region(s) of the body or the subject in the dataset is (are) covered | Brain | ||
| Slice size | What the sizes of each slice are | 512 | ||
| Pixel/Voxel size | What the physical length of a pixel/voxel is | 0.25 mm/ 1 | ||
| WSI size | What the sizes of the whole slide images are | 12,000 | ||
| Patch size | What the extracted image patches are | 256 | ||
| mmp | What the microns per pixel in the level-0 scan are | – | ||
| Model | Workflow | What the procedures of train and inference are, illustrated by the flow chart(s) if appropriate. | – | |
| Learning approaches | What the learning method is. e.g., supervised learning, un/semi-supervised learning | Semi-supervised learning | ||
| Architecture | What the structure of the proposed neural network is, if appropriate | nnUNet | ||
| Task | The description of main tasks conducted on harmonised datasets, e.g., lesion segmentation/classification. | Tumour Segmentation | ||
| Input domain | What the input modality of the proposed method is | 3-D images / 2D feature vectors | ||
| Input size | The input sizes of the model | |||
| Loss | What the optimisation functions are during the training. | Dice and cross-entropy loss | ||
| Open-source | Whether the source code is available or not, provide the link if appropriate. | Open-source code www.github.com... | ||
| Platform | The learning library used to build the model | TensorFlow 2.5.0 | ||
| Evaluation | Statistical Analysis | What the evaluation methods of statistical analysis are | ANOVA-test | |
| Metric | What indicators are used to evaluate harmonisation performance, e.g., the ratio of the reproducible features, coefficient of variation, Pearson correlation coefficient. | Intra-class correlation coefficient (>0.9 is considered reproducible) | ||
| Comparison | What existing approaches are used to compare the performance of the proposed method | stVAE | ||
| Visualisation | What approaches are used to visualise the data distribution before and after harmonisation strategies | t-SNE/UMAP/PCA | ||
| Result | Result | What the quantitative values of evaluation metrics are. | – | |
| Time-consuming | The computational time of the proposed method and the comparisons. | 30 s per case | ||
| Discussion | Novelty | What the innovation of the proposed method is. | – | |
| Strength | The importance/significance of the issue addressed by the proposed method. | – | ||
| Limitation | What remained and unsolved issues are. | – | ||
| Future works | Whether there will be potential studies in the future. | – | ||
Fig. 12Workflow of conducting data harmonisation studies guided by the checklist.
Fig. 13Flowchart of how to select data harmonisation strategies.
Fig. 14Flowchart of how to select harmonisation metrics.
Fig. 7Taxonomy of applications that involved computational data harmonisation strategies.
Fig. 11Scales of cohorts in gene expression and radiomics studies.