| Literature DB >> 33015372 |
Nicola Rieke1,2, Jonny Hancox3, Wenqi Li4, Fausto Milletarì1, Holger R Roth5, Shadi Albarqouni2,6, Spyridon Bakas7, Mathieu N Galtier8, Bennett A Landman9, Klaus Maier-Hein10,11, Sébastien Ourselin12, Micah Sheller13, Ronald M Summers14, Andrew Trask15,16,17, Daguang Xu5, Maximilian Baust1, M Jorge Cardoso12.
Abstract
Data-driven machine learning (ML) has emerged as a promising approach for building accurate and robust statistical models from medical data, which is collected in huge volumes by modern healthcare systems. Existing medical data is not fully exploited by ML primarily because it sits in data silos and privacy concerns restrict access to this data. However, without access to sufficient data, ML will be prevented from reaching its full potential and, ultimately, from making the transition from research to clinical practice. This paper considers key factors contributing to this issue, explores how federated learning (FL) may provide a solution for the future of digital health and highlights the challenges and considerations that need to be addressed.Entities:
Keywords: Medical imaging; Medical research
Year: 2020 PMID: 33015372 PMCID: PMC7490367 DOI: 10.1038/s41746-020-00323-1
Source DB: PubMed Journal: NPJ Digit Med ISSN: 2398-6352
Fig. 1Example federated learning (FL) workflows and difference to learning on a Centralised Data Lake.
a FL aggregation server—the typical FL workflow in which a federation of training nodes receive the global model, resubmit their partially trained models to a central server intermittently for aggregation and then continue training on the consensus model that the server returns. b FL peer to peer—alternative formulation of FL in which each training node exchanges its partially trained models with some or all of its peers and each does its own aggregation. c Centralised training—the general non-FL training workflow in which data acquiring sites donate their data to a central Data Lake from which they and others are able to extract data for local, independent training.
Fig. 2Overview of different FL design choices.
FL topologies—communication architecture of a federation. a Centralised: the aggregation server coordinates the training iterations and collects, aggregates and distributes the models to and from the Training Nodes (Hub & Spoke). b Decentralised: each training node is connected to one or more peers and aggregation occurs on each node in parallel. c Hierarchical: federated networks can be composed from several sub-federations, which can be built from a mix of Peer to Peer and Aggregation Server federations (d)). FL compute plans—trajectory of a model across several partners. e Sequential training/cyclic transfer learning. f Aggregation server, g Peer to Peer.
| 1: |
| 2: Initialise global model: |
| 3: |
| 4: |
| 5: Send |
| 6: Receive model updates and number of local training iterations |
| 7: |
| 8: |
| 9: |
| 10: |
| 11: |