| Literature DB >> 33411624 |
Suraj Rajendran1,2, Jihad S Obeid3, Hamidullah Binol4, Ralph D Agostino5, Kristie Foley6, Wei Zhang1, Philip Austin7, Joey Brakefield8, Metin N Gurcan4, Umit Topaloglu1,4,5.
Abstract
PURPOSE: Building well-performing machine learning (ML) models in health care has always been exigent because of the data-sharing concerns, yet ML approaches often require larger training samples than is afforded by one institution. This paper explores several federated learning implementations by applying them in both a simulated environment and an actual implementation using electronic health record data from two academic medical centers on a Microsoft Azure Cloud Databricks platform.Entities:
Mesh:
Year: 2021 PMID: 33411624 PMCID: PMC8140794 DOI: 10.1200/CCI.20.00060
Source DB: PubMed Journal: JCO Clin Cancer Inform ISSN: 2473-4276
FIG A1.Single weight training mechanism.
FIG A2.Cyclical weight training mechanism.
FIG A3.Distributions of each dataset across classes.
A Comprehensive Distribution of the Two Institutions' Datasets Across Each Feature
FIG A4.The workflow of the federated learning environment in databricks.
ANN Model Performances on Institution 1 Test Data: Base 1 (Model Trained on Institution 1's Data), Base 2 (Model Trained on Institution 2's Data), Single Weight Model A (Institution 1 Trains the Model First), and Single Weight Model B (Institution 2 Trains the Model First); ANN Model Performances on Institution 2 Test Data: Base 1 (Model Trained on Institution 1's Data), Base 2 (Model Trained on Institution 2's Data), Single Weight Model A (Institution 1 Trains the Model First), and Single Weight Model B (Institution 2 Trains the Model First)
LR Model Performances on WF's Test Data: Base 1 (Model Trained on WF's Data), Base 2 (Model Trained on MUSC's Data), Single Weight Model A (WF Trains the Model First), and Single Weight Model B (MUSC Trains the Model First); LR Model Performances on MUSC's Test Data: Base 1 (Model Trained on WF's Data), Base 2 (Model Trained on MUSC's Data), Single Weight Model A (WF Trains the Model First), and Single Weight Model B (MUSC Trains the Model First)
LR Model Performances on Institution 1 Test Data: Base 1 (Model Trained on Institution 1's Data), Base 2 (Model Trained on Institution 2's Data), Single Weight Model A (Institution 1 Trains the Model First), and Single Weight Model B (Institution 2 Trains the Model First); LR Model Performances on Institution 2 Test Data: Base 1 (Model Trained on Institution 1's Data), Base 2 (Model Trained on Institution 2's Data), Single Weight Model A (Institution 1 Trains the Model First), and Single Weight Model B (Institution 2 Trains the Model First)
FIG 1.ROC curves corresponding to performance metrics in tables. (A) ROC curve based on ANN models' performances against institution 1 test data. (B) ROC curve based on ANN models' performances against institution 2 test data. (C) ROC curve based on LR models' performances against institution 1 test data. (D) ROC curve based on LR models' performances against institution 2 test data. ANN, artificial neural network; LR, logistic regression; ROC, receiver operating characteristic.
ANN Model Performances on WF's Test Data: Base 1 (Model Trained on WF's Data), Base 2 (Model Trained on MUSC's Data), Single Weight Model A (WF Trains the Model First), and Single Weight Model B (MUSC Trains the Model First); ANN Model Performances on MUSC's Test Data: Base 1 (Model Trained on WF's Data), Base 2 (Model Trained on MUSC's Data), Single Weight Model A (WF Trains the Model First), and Single Weight Model B (MUSC Trains the Model First)
FIG 2.ROC curves corresponding to performance metrics in tables. (A) ROC curve based on ANN models' performances against WF's test data. (B) ROC curve based on ANN models' performances against the MUSC's test data. (C) ROC curve based on LR models' performances against WF's test data. (D) ROC curve based on LR models' performances against the MUSC's test data. ANN, artificial neural network; LR, logistic regression; MUSC, Medical University of South Carolina; ROC, receiver operating characteristic; WF, Wake Forest.