Literature DB >> 33847779

Distributed learning: a reliable privacy-preserving strategy to change multicenter collaborations using AI.

Margarita Kirienko1,2, Martina Sollini3,4, Gaia Ninatti2, Daniele Loiacono5, Edoardo Giacomello5, Noemi Gozzi6, Francesco Amigoni5, Luca Mainardi5, Pier Luca Lanzi5, Arturo Chiti2,6.   

Abstract

PURPOSE: The present scoping review aims to assess the non-inferiority of distributed learning over centrally and locally trained machine learning (ML) models in medical applications.
METHODS: We performed a literature search using the term "distributed learning" OR "federated learning" in the PubMed/MEDLINE and EMBASE databases. No start date limit was used, and the search was extended until July 21, 2020. We excluded articles outside the field of interest; guidelines or expert opinion, review articles and meta-analyses, editorials, letters or commentaries, and conference abstracts; articles not in the English language; and studies not using medical data. Selected studies were classified and analysed according to their aim(s).
RESULTS: We included 26 papers aimed at predicting one or more outcomes: namely risk, diagnosis, prognosis, and treatment side effect/adverse drug reaction. Distributed learning was compared to centralized or localized training in 21/26 and 14/26 selected papers, respectively. Regardless of the aim, the type of input, the method, and the classifier, distributed learning performed close to centralized training, but two experiments focused on diagnosis. In all but 2 cases, distributed learning outperformed locally trained models.
CONCLUSION: Distributed learning resulted in a reliable strategy for model development; indeed, it performed equally to models trained on centralized datasets. Sensitive data can get preserved since they are not shared for model development. Distributed learning constitutes a promising solution for ML-based research and practice since large, diverse datasets are crucial for success.
© 2021. The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature.

Entities:  

Keywords:  Clinical trial; Distributed learning; Ethics; Federated learning; Machine learning; Privacy

Mesh:

Year:  2021        PMID: 33847779      PMCID: PMC8041944          DOI: 10.1007/s00259-021-05339-7

Source DB:  PubMed          Journal:  Eur J Nucl Med Mol Imaging        ISSN: 1619-7070            Impact factor:   9.236


Introduction

Artificial intelligence has gained significant attention because of the achievements of machine learning (ML) and deep learning algorithms that rapidly accelerate research and transform practices in multiple fields, including medicine. However, data-driven learning necessitates “big data”-sets not to suffer from overfitting and underspecification. Indeed, studies on a small sample size using ML are affected by an inherent methodological bias that might undermine their validity. The adequate sample size is thus crucial for ML as for classical statistics [1-3]. An adequate sample is a hurdle in rare diseases (e.g. thymic malignancies, sarcomas) or low-prevalence conditions (e.g. refractory lymphoma, iodine-negative thyroid cancer) [4]. The small sample size is acknowledged as a limitation in almost every research study on image mining [1]. Other commonly recognized weaknesses in image mining studies are the monocentric retrospective design (i.e. localized learning) and the lack of independent validation. Collectively, all these aspects negatively affect the reproducibility and the generalizability of the results [1, 5]. These limitations might be overcome by multicenter or benchmarking trials. However, benchmarking trials require considerable infrastructural efforts to develop data repository platforms, while traditional multicenter studies (i.e. centralized learning) are affected by many logistical difficulties mainly related to sharing of clinical and imaging data. Data transfer is indeed burdened by legal, ethical, and privacy issues [5, 6]. Given these constraints, distributed learning has emerged as a strategy for effective collaboration between centres while preserving governance and regulatory aspects [7]. Distributed learning aims to train one or more machine learning models within a network of nodes, each one owning a local dataset. Individual institutions do not share patients’ data externally. Just post-processed data in the form of model updates (e.g. coefficients and weight parameters) are shared among centres to build the final model [8, 9] (Fig. 1). Distributed learning methods may be distinguished according to computational principles (e.g. data parallelism and communication topology [10]) (Fig. 2) [11-13]. However, some general principles, (i) how the model parameters are displaced over the network of nodes, (ii) how nodes interact and which type of data they exchange, (iii) limitations on the kind of disclosed data, (iv) technical and technological constraints related to the task, are relevant to design the best-distributed learning model. Accordingly, distributed learning methods include different approaches, namely ensemble [14, 15], split [16], and federated learning [17, 18] described in details in Fig. 3.
Fig. 1

Distributed learning framework

Fig. 2

Schematic overview of the most popular distributed learning methods. Differences among methods are summarized according to two general design principles: (i) how the model parameters are displaced over the network of nodes, local, local + shared, and shared (y-axis), and (ii) how nodes interact and which type of data they exchange, no, partial, and complete exchange (x-axis)

Fig. 3

Distributed learning methods

Distributed learning framework Schematic overview of the most popular distributed learning methods. Differences among methods are summarized according to two general design principles: (i) how the model parameters are displaced over the network of nodes, local, local + shared, and shared (y-axis), and (ii) how nodes interact and which type of data they exchange, no, partial, and complete exchange (x-axis) Distributed learning methods Therefore, distributed learning dealing with a network of nodes, each one owning a local dataset, has been proposed as a method to cross the hurdles related to patient data sharing [7, 9]. However, there are some uncertainties related to distributed learning performance compared to centrally trained models. Therefore, the present scoping review, providing an overview of distributed learning in medical research, aimed at assessing the non-inferiority of distributed learning over centrally and locally trained models. The non-inferiority of distributed learning over centrally and locally trained models is an essential requirement to appoint distributed learning as a suitable approach for data sharing within multicenter collaborations.

Materials and methods

Literature search and selection

We performed an extensive literature search using the term “distributed learning” OR “federated learning” in the PubMed/MEDLINE and EMBASE databases. No start date limit was used, and the search was extended until July 21, 2020. Two authors (MS and GN) independently searched the literature and performed an initial screening of identified titles and abstracts. The following exclusion criteria were applied: (a) articles outside the field of interest (i.e. distributed learning); (b) guidelines or expert opinion, review articles and meta-analyses, editorials, letters or commentaries, and short abstracts presented at conferences and scientific meetings; (c) articles not in the English language; and (d) studies not testing distributed approaches using medical data (i.e. electronic health records, genomic data, signals, or images). Selected papers were retrieved in full text, and reference lists were screened in order to identify additional records. Screened papers were included in the scoping review when considered eligible by both reviewers. In case of discrepancies, papers were reviewed by a third researcher, blinded to previous assessments. The majority vote was used to include/exclude a paper finally.

Analysis

Each selected article was tabulated in an Excel ®2017 (Microsoft®, Redmond, WA) file. The following information was collected: (i) clinical setting; (ii) type of data (clinical data, images, genomic data); (iii) source of data (open-access or local dataset); (iv) type of distributed network (simulated versus real); (v) machine learning algorithms’ architecture; (vi) distributed learning approach (ensembling, split, or federated learning); (vii) performance metrics; (viii) comparative analysis (distributed versus centralized, distributed versus local); (ix) analysis on data distribution among nodes; (x) other results. Later on, selected studies were classified according to their aim, namely (a) risk prediction; (b) diagnosis; (c) prognosis; and (d) treatment side effect/adverse drug reaction prediction. Papers having more than one aim were analysed as separate works.

Results

Study selection

Overall, the search criteria resulted in 387 articles. After removing duplicates and initial screening of titles and abstracts, 347 papers were excluded, and the remaining 40 were retrieved in full text. Subsequent analysis of full-text articles excluded 22 papers (two using non-medical data, 13 outside the field of interest, five review articles). Eight additional research studies of interest were identified as screening reference lists. A total of 26 studies were eventually selected. Figure 4 summarizes the process of study selection.
Fig. 4

Study selection

Study selection

Overall results

Distributed learning was based on federated learning with [12, 19–21] or without [22-43] other methods in the great majority of the selected studies (4/26 and 22/26, respectively); in few articles (3/26), the ensembling approach was preferred [38, 40, 42]. Distributed training was compared to centralized and/or localized training in 21/26 [12, 19, 20, 22–24, 26, 28–31, 34–43] and 14/26 [12, 21, 22, 25, 27, 28, 30, 32, 33, 36, 37, 39, 40, 42] selected papers, respectively. Area under the curve (AUC) [19, 22–25, 27–29, 36, 43] and accuracy [12, 21, 27, 34, 35, 37, 39–41] were the most commonly used metrics to assess models’ performance in the selected papers (10/26 and 7/26). Other metrics included mean squared error [26, 31, 39, 41], dice score [20, 32, 33], F1 score [27, 42], precision [27, 37], normalized mutual information and signal-to-noise ratio [39], recall [37], sensitivity and positive predictive value [27], hazard ratio [38], relative bias, error ratio, and odd ratio [30]. Figure 5 summarizes the studies by the aim.
Fig. 5

Summary of the topics covered by selected studies. (ADR, adverse drug reaction)

Summary of the topics covered by selected studies. (ADR, adverse drug reaction)

Risk prediction

Three out of the 26 included papers developed a distributed learning framework for risk prediction (Table 1). Brisimi et al. [29] and Wang et al. [38] used clinical data to predict hospitalization risks for cardiac events and re-hospitalization for heart failure, respectively. Duan et al. [30] used distributed clinical data to build a risk model for foetal loss in pregnant women. Wu et al. [35] instead performed a federated genome-wide association studies analysis to predict the risk of developing ankylosing spondylitis in healthy individuals. In the four studies, distributed and centralized learning model performance was compared. In all cases, the two approaches showed substantial equivalence.
Table 1

Studies using distributed learning models for risk prediction

ReferenceData, typeAimDistributed networkMachine learning modelDistributed learning methodDistributed vs centralized learningDistributed vs localized learning
[29]Clinical dataPrediction of risk of hospitalisations for cardiac eventsSimulated distribution among 5 and 10 nodesSVMFederated learningFederated learning performs close to centralized learningNA
[30]Clinical dataPrediction of foetal loss riskSimulated distribution among 10 nodesLogistic regressionFederated learningFederated learning performs close to centralized learningFederated learning outperforms localized learning models
[35]Genomic dataPrediction of risk of developing ankylosing spondylitisReal distribution among 3 nodesPCAFederated learningFederated learning performs close to centralized learningNA
[38]Clinical dataPrediction of readmission for heart failure within 30 daysSimulated distribution among 50 and 200 nodesCox proportional hazards modelEnsemblingEnsembling performs close to centralized learningNA

NA, not assessed; SVM, support vector machine

Studies using distributed learning models for risk prediction NA, not assessed; SVM, support vector machine

Diagnosis

Out of the 26 papers on distributed learning, 15 focused on diagnosis (Table 2). Among these, three works [12, 34, 42] used different datasets as input for their experiments. Specifically, Chang et al. [12] tested distributed learning to analyse retinal fundus images and mammograms, Balachandar et al. [34] to analyse retinal fundus images and chest x-ray, and Tuladhar et al. [42] to analyse digital fine-needle aspiration (FNA) images and clinical data.
Table 2

Studies using distributed learning models for diagnosis

ReferenceData, typeAimDistributed networkMachine learning modelDistributed learning methodDistributed vs. centralized learningDistributed vs. localized learning
[42]Images (digital FNA images)Detection of breast cancerSimulated distribution among many different numbers of nodesANN, SVM, RFEnsemblingEnsembling performs close to centralized learningEnsembling outperforms localized learning models
[42]Clinical dataDetection of MCIReal distribution among 43 nodesANN, SVM, RFEnsemblingCentralized learning outperforms ensemblingEnsembling outperforms localized learning models
[42]Clinical dataDetection of diabetesSimulated distribution among many different numbers of nodesANN, SVM, RFEnsemblingEnsembling performs close to centralized learningEnsembling outperforms localized learning models
[42]Clinical dataDetection of heart diseaseSimulated distribution among many different numbers of nodesANN, SVM, RFEnsemblingEnsembling performs close to centralized learningEnsembling outperforms localized learning models
[43]Image-derived data (radiomic features from head and neck CT scans)Prediction of HPV status in head and neck cancerReal distribution among 6 nodesLogistic regressionFederated learningFederated learning performs close to Centralized LearningNA
[12](Retinal fundus images)Diabetic retinopathy detectionSimulated distribution among 4 nodesCNNFederated learning + ensemblingFederated learning performs close to centralized LearningNA
[12]Images (mammograms)Breast cancer detectionSimulated distribution among 4 nodesCNNFederated learning + ensemblingFederated learning performs close to centralized learningNA
[27]Clinical data (ICU data)Disease detectionDistribution among 2 nodes based on the data provisioning systemAutoencoderFederated learningNAFederated learning outperforms localized learning models
[28]Clinical data (ICU data)Disease detectionSimulated distribution among 2 and 3 nodesHashingFederated learningFederated learning performs close to centralized learningFederated learning outperforms localized learning models
[20]Images (brain MRI)Brain tumour segmentationSimulated distribution among 4, 8,16, and 32 nodesCNNFederated learning, CIIL, IILFederated learning performs close to centralized learningNA
[31]Images (brain MRI)Classification of neurodegenerative diseasesReal distribution among 4 nodesPCAFederated learningFederated learning performs close to centralized learningNA
[32]Images (head CT scans)Haemorrhage segmentationReal distribution among 4 nodes (only 2 nodes used for training)CNNFederated learningNAFederated learning outperforms localized learning models
[33]Images (head CT scans)Haemorrhage segmentationReal distribution among 2 nodesCNNFederated learningNAFederated learning outperforms localized learning models
[34]Images (retinal fundus images)Diabetic retinopathy detectionSimulated distribution among 4 nodesCNNFederated learningFederated learning performs close to centralized learningNA
[34]Images (chest X-rays)Thoracic disease classificationSimulated distribution among 4 nodesCNNFederated learningFederated learning performs close to centralized learningNA
[21]Images (brain fMRI)Autism spectrum disorder detectionReal distribution among 4 nodesANNFederated learning + ensemblingNAFederated learning outperforms localized learning models
[36]Images (chest CT scans)Pneumonia classificationReal distribution among 4 nodesCNNFederated learningFederated learning performs close to centralized learning (except on viral pneumonia other than COVID-19)Federated learning outperforms localized learning models
[39]Images (digital FNA images)Breast lesion classificationSimulated distribution among 5 nodesSemi-supervised ANN + ELMFederated learningCentralized learning outperforms federated learningFederated learning outperforms localized learning models
[40]Images (brain MRI)Schizophrenia detectionReal distribution among 4 nodesSVMEnsemblingEnsembling performs close to centralized learningEnsembling outperforms localized learning models
[41]Images (digital FNA images)Breast lesion classificationSimulated distribution among 10 nodesANNFederated learningFederated learning performs close to centralized learningNA

ANN, artificial neural network; CIIL, cyclic institutional incremental learning; CNN, convolutional neural network; CT, computed tomography; ELM, extreme learning machine; FNA, fine-needle aspiration; HPV, human papilloma virus; ICU, intensive care unit; IIL, institutional incremental learning; MCI, mild cognitive impairment; MRI, magnetic resonance imaging; NA, not assessed; PCA, principal component analysis; RF, random forest; SVM, support vector machine

Studies using distributed learning models for diagnosis ANN, artificial neural network; CIIL, cyclic institutional incremental learning; CNN, convolutional neural network; CT, computed tomography; ELM, extreme learning machine; FNA, fine-needle aspiration; HPV, human papilloma virus; ICU, intensive care unit; IIL, institutional incremental learning; MCI, mild cognitive impairment; MRI, magnetic resonance imaging; NA, not assessed; PCA, principal component analysis; RF, random forest; SVM, support vector machine In the group of 26 papers, seven articles aimed at using distributed clinical [27, 28, 42, 43] or imaging [12, 21, 34, 40, 42] data to detect several diseases. Included diseases were diabetic retinopathy, breast cancer, cardiac diseases, diabetes, and neuropsychiatric disorders. Five studies instead tested distributed learning using diagnostic images for various classification tasks, namely differentiation of neurodegenerative diseases [31], thoracic pathologies [34], types of pneumonia [36], benign vs malignant breast lesions [39, 41], and HPV+ vs HPV- head and neck cancer [43]. Finally, three studies [20, 31, 32] utilized distributed head computed tomography (CT) and brain magnetic resonance imaging (MRI) data to perform segmentation of intraparenchymal haemorrhages and brain tumours. In 13/15 papers focusing on the diagnosis, the authors compared distributed and centralized approaches for model training. The performance achieved using a model trained on distributed data was always comparable to the one obtained using a centralized approach, except for a few experiments [36, 42]. Specifically, in the experiment by Tuladhar et al. [42] on real data distribution, the skewness of the training dataset towards patient examples affected the ensembling approach’s performance in predicting mild cognitive impairment (MCI), but not those of the centralized learning. However, this experiment was characterized not only by imbalanced classes (189 MCI patients versus 94 healthy subjects) but also by a limited number of observations (average of 4 patients and two healthy cases) for each node [42]. Finally, centralized learning outperformed distributed learning in classifying non-COVID-19 viral pneumonia [36]. Xu et al. [34] developed a distributed CNN model [36](Xu et al. 2020) [26–28, 34]to classify CT scan as normal or with pneumonia, distinguishing among COVID-19 or other viral pneumonia and bacterial pneumonia. The federated model performed close to the centralized one in all cases, except non-COVID-19 viral pneumonia, which was scarcely represented in the whole dataset (76 cases) [36].

Prognosis

Five out of the 26 analysed articles on distributed learning were aimed at prognostication (Table 3).
Table 3

Studies using distributed learning models for prognosis

ReferenceData, typeAimDistributed networkMachine learning modelDistributed learning methodDistributed vs. centralized learningDistributed vs. localized learning
[24]Clinical dataPost-treatment 2-year survival predictionReal distribution among 3 nodesSVM with ADMMFederated learningFederated learning performs close to centralized learningNA
[25]Clinical dataPost-treatment 2-year survival predictionReal distribution among 8 nodesBayesian networkFederated learningNALocalized learning models sometimes outperform federated learning
[43]Image-derived data (radiomic features from head and neck CT scans)Post-treatment 2-year survival predictionReal distribution among 6 nodesLogistic regressionFederated learningFederated learning performs close to centralized learningNA
[26]Clinical dataPrediction of hospital LoSSimulated distribution among 3, 6, and 12 nodesLinear regressionFederated learningFederated learning performs close to centralized learningNA
[19]Clinical dataPrediction of mortalityReal distribution among 50 nodesAutoencoder, K-means, and nearest neighbourFederated learning + ensemblingFederated learning performs close to centralized learningNA
[19]Clinical dataPrediction of ICU LoSReal distribution among 50 nodesAutoencoder, K-means, and nearest neighbourFederated learning + ensemblingFederated learning performs close to centralized learningNA

ADMM, alternative direction method of multipliers; CT, computed tomography; ICU, intensive care unit; LoS, length of stay; NA, not assessed; SVM, support vector machine

Studies using distributed learning models for prognosis ADMM, alternative direction method of multipliers; CT, computed tomography; ICU, intensive care unit; LoS, length of stay; NA, not assessed; SVM, support vector machine Jochems et al. [24] set up a network of systematic clinical data collection and sharing through distributed learning to predict post-treatment 2-year survival in 894 non-small cell lung cancer (NSCLC) patients. The same group [25] confirmed the previous results on a larger cohort, including more than 20,000 patients in eight centres across different countries. Bogowicz et al. [43] used a distributed approach to train a radiomic model to predict 2-year survival in a cohort of 1174 patients with head and neck cancer. Dankar et al. [26] and Huang et al. [19] used a distributed algorithm to estimate the patient length of stay in hospital and ICU, respectively. Moreover, Huang et al. [19] used the same approach to predict inpatient mortality. In four out of these five prognostic studies, predictive models trained using distributed and centralized data were compared (Table 3). In all cases, models did not show significant differences in terms of performance.

Adverse drug reaction/side effect prediction

Three out of the 26 studies used a distributed approach to predict either adverse drug reaction (ADR) or treatment side effect (Table 4).
Table 4

Studies using distributed learning models for adverse drug reaction/side effect prediction

ReferenceData, typeAimDistributed networkMachine learning modelDistributed learning methodDistributed vs. centralized learningDistributed vs. localized learning
[22]Clinical dataPrediction of post-radiotherapy dyspnoeaReal distribution among 5 nodesBayesian networkFederated learningFederated learning performs close to centralized learningFederated learning performs close to localized learning
[23]Clinical dataPrediction of post-radiotherapy dyspnoeaReal distribution among 5 nodesSVM with ADMMFederated learningFederated learning performs close to centralized learningNA
[37]Clinical dataPrediction of opioid chronic useSimulated distribution among 10 nodesSVM, single-layer perceptron, logistic regression (all of them trained with SGD)Federated learningFederated learning performs close to centralized learningFederated learning outperforms localized learning models
[37]Clinical dataPrediction of antipsychotics side effectsSimulated distribution among 10 nodesSVM, single-layer perceptron, logistic regression (all of them trained with SGD)Federated learningFederated learning performs close to centralized learningFederated learning outperforms localized learning models

ADMM, alternative direction method of multipliers; NA, not assessed; SGD, stochastic gradient descent; SVM, support vector machine

Studies using distributed learning models for adverse drug reaction/side effect prediction ADMM, alternative direction method of multipliers; NA, not assessed; SGD, stochastic gradient descent; SVM, support vector machine Among these, Choudhury et al. [37] performed two distributed learning experiments and built models to predict opioid chronic usage and antipsychotic side effects using clinical data. Jochems et al. [22] tested the distributed approach using clinical parameters and built a Bayesian network model to predict dyspnoea in NSCLC patients treated with radiotherapy. The same group [23] confirmed the distributed approach’s potential in a second study using a different machine learning algorithm to predict the same outcome. In the three studies, authors compared the distributed and centralized algorithms’ performance and showed substantial equivalence between the two approaches.

Discussion

The present scoping review aimed to gather and assess the research evidence to answer whether multi-institutional collaboration in machine learning research and practice can rely on distributed learning as on a conventional centralized approach. Our analysis confirmed that distributed learning is an effective strategy for multi-institutional collaboration with the potential advantage of being privacy-preserving. Although distributed learning has been proposed to share data guaranteeing privacy-preserving issues, it does not per se guarantee security and privacy by design. Indeed, it might be possible to retrieve estimations of the original data through a reverse engineering approach from the shared weights. Nonetheless, distributed learning should be considered as the prerequisite infrastructure to address governance and regulatory compliance. Indeed, a distributed network may be easily empowered by specific privacy preservation methods (e.g. differential privacy and cryptographic techniques) [7, 9]. To propose an effective privacy-preserving methodology in multicenter collaboration, the assessment of distributed learning performance compared to a centralized approach is the first-step requirement. Evaluation of the effectiveness of privacy-preserving methods was out of the present work’s scope and should be evaluated in future investigations. Regardless of the aim (risk prediction, diagnosis, prognosis, or treatment side effect/adverse drug reaction prediction), the type of input (clinical data, images, or genetic data), the method (federated or ensembling), and the classifier (e.g. artificial neural networks, support vector machine, random forest), distributed training performed close to centralized training in almost all the experiments [12, 19, 20, 22–24, 26, 28–31, 34–43] (Tables 1-4). Centralized learning showed better performances than the distributed approach in two experiments among those reported in the studies by Xu et al. [36] and Tuladhar et al. [42]. However, both these experiments were burdened by a limited number of observations, probably responsible for the overfitting-related failure. Particularly, Xu et al. [36] built a model using a dataset consisting mainly of patients with SARS-CoV-2 or bacterial pneumonia (34% and 27%, respectively) and healthy subjects (33%). Other viral non-COVID pneumonia, which per se included several entities characterized by specific features (e.g. respiratory syncytial virus, cytomegalovirus, influenza A), represented only 6% of the entire dataset [36]. Similarly, the ensembling approach tested by Tuladhar et al. [42] dealt with unbalanced groups (67% MCI patients versus 33% healthy subjects), but the experiment also suffered from a limited number of observations for each node (average of 6 cases in the 43 nodes). Notably, the other experiments performed by Tuladhar et al. [42] dealing with simulated distributed data demonstrated that the dataset’s skewness did not affect the final model’s performance. This finding indicated that the imbalance of groups determined the model’s poor generalizability when the number of nodes and observations in each node is not sufficiently heterogenous to be representative for the target population. Indeed, when a sufficient number of “nodes” and of “observations for each node” is included in the model, distributed learning is not prone to the inherent limitations of locally trained models (e.g. imbalanced groups) and the overfitting related to the peculiarity of each site resulting in a competitive generalisable model. Distributed models outperformed locally trained models [21, 27, 28, 30, 32, 33, 36, 37, 39, 40, 42] with rare exceptions [22, 25]. Specifically, the distributed model developed by Jochems et al. [22] was as efficient as localized learning models in predicting post-radiotherapy dyspnoea. Diversely, Deist et al. [25] found that localized learning models sometimes outperform federated learning in 2-year survival prediction, suggesting that unobserved confounding factors or diverse outcome collection standards may affect the model’s performance. Preprocessing data harmonization – if possible – (e.g. image resampling, a clear methodology and uniform criteria for data collection) could positively impact data integration, simplifying multi-institutional collaboration for large-scale analytics [25, 44]. This scoping review aimed to produce evidence on the efficiency of the distributed learning regardless of the type of input data. We included papers dealing with clinical data (n = 12), images (hand-extracted features = 1, images = 11), both clinical data and images (n = 1), and genetic data (n = 1). This variability demonstrates that distributed learning may be implemented within any multicenter collaboration. Indeed, in precision medicine, an increasing number and kind of variables (lifestyle habits, constitutional factors, clinical, pathological, imaging, and “omics” data including treatment-related information) need to be considered to build predictive models. Medicine is evolving from a “disease-based” vision towards a “patient-based” approach based on multidimensional data that describe physiological and pathological conditions [6, 45, 46]. Distributed learning may represent a reliable framework for future multi-institutional and multidisciplinary research and clinical practice. On the other hand, the “digital twin” approach has been proposed to respond to the challenge of the integration and analysing of large amounts of data within a dynamic framework. This technology is based on the construction of a digital model of an individual patient (depicting the molecular profile, the physiological status, and the life style habits) to virtually test a multitude of treatments to choose the optimal one [47, 48]. Therefore, both the medical and technological conceptual views that aim to provide a detailed and extensive description of an individual’s multidimensionality may benefit from a distributed learning framework. Differences in terms of aims, input data, and methods (i.e. algorithms, data distribution among nodes, number of nodes, and metrics to assess model performance) made selected papers hardly comparable. Consequently, a preferred distributed learning method tailored for each specific task cannot be recommended. Nonetheless, ensembling emerged to be incredibly convenient when input data are heavily heterogeneous; it can be applied to any machine learning algorithm and uses different learning models for each node [14, 15]. Split learning, owing to its layered architecture, is fit for deep neural networks [16]. Finally, federated learning, which parallelly trains local models and aggregates their updates into a “central” node, can be efficiently utilized with many different machine learning algorithms [17]. Therefore, when setting up a distributed learning network, artificial data and simulated networks may be developed according to the model’s particular setting and objective to get preliminary performance results and compare distributed vs non-distributed approaches. Technical constraints and artificial intelligence (AI) acceptance may be the main barriers to widespread distributed learning in healthcare. Technical challenges consist of computational burden and the communication overload (i.e. the amount of data shared among nodes) that need to meet the infrastructure constraints. In this regard, the number of nodes and kinds of data distributed among nodes is crucial hyperparameters that need to be tuned. Nonetheless, as a proof of concept, some commercial and open source solutions (e.g. DistriM from Oncoradiomics [49], Varian Learning Portal from Varian [50], Clara platform from Nvidia [51], and GRIN from Genomics Research and Innovation Network [52]) have recently become available, supporting the feasibility of implementing a distributed learning infrastructure. Moreover, the lack of confidence and critical appraisal of the ML-based tools by healthcare personnel may limit technology implementation. Consequently, educational material and programs [6, 17, 53] involving clinicians, researchers, and regulatory officers are progressively getting available to promote the awareness on opportunities of multi-institutional trials and practices based on distributed learning towards a responsible AI. Additionally, the trustworthiness of AI-based methods is challenged by the barrier of explainability. The so-called eXplainable AI (XAI) field is growing, intending to develop responsible AI and encourage experts and professionals to embrace the new technology’s benefits to overcome the limitation related to explainability [54]. Furthermore, data have to be structured or preprocessed before their use to train a model. Distributed learning strategies were first introduced to analyse clinical data from electronic health records (EHRs). The rigid structure of EHRs has been influenced by research data collection systems and technology advances applied in clinical trials. Indeed, in clinical research, study design, data collection, analysis, and sharing have evolved over the last 20 years [55]. At present, data within clinical trials are recorded according to demanding rules aimed at making them standardized and structured or semi-structured. Overall, these approaches lead to high-quality information, adequate assessment of outcomes, endpoints, events, timely data analysis, reliable models, and efficient and comprehensive results’ publication and dissemination. The trend towards structured data is getting translated to routine medical EHRs. Generally, EHRs share common terminology, codes, and sections that could be easily collected from different centres or countries and subsequently analysed, while other domains, including medical imaging, are less structured. In the past years, several initiatives and registries containing structured data have been developed mainly as public health and descriptive epidemiological tools [55, 56]. The CancerLinQ has been the first and the foremost initiative sponsored by the American Society of Clinical Oncology and its Institute for Quality to collect “protected” EHRs data with the final goal of assessing, monitoring, and improving delivered care in cancer patients [57]. However, all these initiatives were challenged by privacy and security concerns on collecting, using, and patient information disclosure. Additionally, concerns related to ethical and reliability aspects of AI-based algorithms may influence distributed learning technology spread. A potential drawback of AI prediction is its dependence on the data being used to train the algorithm. Training data have to represent the diseases and patient populations under evaluation and be balanced to perform when exposed to diverse patient data. Additionally, underspecification in ML pipelines may contribute to the failure of ML algorithms in a real-world deployment. Underspecification occurs when many distinct solutions can solve the problem equivalently [58]. Modelling pipelines need to explicitly account for this problem, also in the medical imaging domain. The great potential of distributed learning has been already recognized by companies as proofed by the EHR4CR project. Within that project, pharmaceutical companies created a precompetitive data environment that relied on distributed learning methods that allowed them to safely perform clinical trials, sharing commercially confidential information [59]. This scoping review has some limitations. The high heterogeneity of included studies in terms of aims, input data, methods, and performance metrics prevented a comparison between studies, leaving no space for systematic considerations. Therefore, the present review aimed at assessing if multisite collaboration could efficiently rely on distributed learning. Therefore, our interest focused on assessing the non-inferiority of distributed learning over centrally and locally trained models instead of producing a systematic review. Nonetheless, the review gathered evidence that distributed learning is feasible, safe, and non-inferior for localized and centralized learning. Given this premise, we expect that distributed learning’s usefulness and applicability will be objects of many more investigations in different medical settings shortly. In conclusion, distributed learning-based models showed to be reliable; indeed, they performed equally to models trained on centralized datasets. Sensitive data can get preserved by distributed learning since they are not shared for model development. Distributed learning constitutes a promising solution, especially for AI research, since large, diverse datasets are crucial for success. We foresee distributed learning being the bridge to large-scale, multi-institutional collaboration in research and medical practice.
  2 in total

1.  Multi-Institutional Deep Learning Modeling Without Sharing Patient Data: A Feasibility Study on Brain Tumor Segmentation.

Authors:  Micah J Sheller; G Anthony Reina; Brandon Edwards; Jason Martin; Spyridon Bakas
Journal:  Brainlesion       Date:  2019-01-26

2.  Predicting Adverse Drug Reactions on Distributed Health Data using Federated Learning.

Authors:  Olivia Choudhury; Yoonyoung Park; Theodoros Salonidis; Aris Gkoulalas-Divanis; Issa Sylla; Amar K Das
Journal:  AMIA Annu Symp Proc       Date:  2020-03-04
  2 in total
  2 in total

Review 1.  Diagnostic accuracy and potential covariates of artificial intelligence for diagnosing orthopedic fractures: a systematic literature review and meta-analysis.

Authors:  Xiang Zhang; Yi Yang; Yi-Wei Shen; Ke-Rui Zhang; Ze-Kun Jiang; Li-Tai Ma; Chen Ding; Bei-Yu Wang; Yang Meng; Hao Liu
Journal:  Eur Radiol       Date:  2022-06-27       Impact factor: 7.034

2.  Ensemble Learning Using Individual Neonatal Data for Seizure Detection.

Authors:  Ana Borovac; Steinn Gudmundsson; Gardar Thorvardsson; Saeed M Moghadam; Paivi Nevalainen; Nathan Stevenson; Sampsa Vanhatalo; Thomas P Runarsson
Journal:  IEEE J Transl Eng Health Med       Date:  2022-08-23
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.