Literature DB >> 30892656

Fair compute loads enabled by blockchain: sharing models by alternating client and server roles.

Tsung-Ting Kuo1, Rodney A Gabriel1,2, Lucila Ohno-Machado1,3.   

Abstract

OBJECTIVE: Decentralized privacy-preserving predictive modeling enables multiple institutions to learn a more generalizable model on healthcare or genomic data by sharing the partially trained models instead of patient-level data, while avoiding risks such as single point of control. State-of-the-art blockchain-based methods remove the "server" role but can be less accurate than models that rely on a server. Therefore, we aim at developing a general model sharing framework to preserve predictive correctness, mitigate the risks of a centralized architecture, and compute the models in a fair way.
MATERIALS AND METHODS: We propose a framework that includes both server and "client" roles to preserve correctness. We adopt a blockchain network to obtain the benefits of decentralization, by alternating the roles for each site to ensure computational fairness. Also, we developed GloreChain (Grid Binary LOgistic REgression on Permissioned BlockChain) as a concrete example, and compared it to a centralized algorithm on 3 healthcare or genomic datasets to evaluate predictive correctness, number of learning iterations and execution time.
RESULTS: GloreChain performs exactly the same as the centralized method in terms of correctness and number of iterations. It inherits the advantages of blockchain, at the cost of increased time to reach a consensus model. DISCUSSION: Our framework is general or flexible and can also address intrinsic challenges of blockchain networks. Further investigations will focus on higher-dimensional datasets, additional use cases, privacy-preserving quality concerns, and ethical, legal, and social implications.
CONCLUSIONS: Our framework provides a promising potential for institutions to learn a predictive model based on healthcare or genomic data in a privacy-preserving and decentralized way.
© The Author(s) 2019. Published by Oxford University Press on behalf of the American Medical Informatics Association.

Entities:  

Keywords:  batch machine learning; blockchain distributed ledger technology; clinical information systems; decision support systems; privacy-preserving predictive modeling

Mesh:

Year:  2019        PMID: 30892656      PMCID: PMC7787356          DOI: 10.1093/jamia/ocy180

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


INTRODUCTION

Healthcare and genomics are some of the most important types of data for cross-institution predictive modeling that estimates patient outcomes by analyzing observed data and generating scientific evidence using data from multiple institutions. Specifically, as the volume of the shared healthcare or genomic data increases, the “learned” model (ie, the parameters identified by the learning algorithms) becomes more generalizable, thus improves the predictive correctness for each of the participating institutions. Initiatives such as ClinVar, aim at the same goal and allow researchers to perform case-based predictions. To avoid potential risks of improper protected health information (PHI) data disclosure in direct data sharing, several privacy-preserving algorithms, such as GLORE and EXPLORER, have been proposed to exchange only the model (ie, aggregated parameters) but not the observation-level patient data. As shown in Figure 1A , such methods are mostly based on a centralized architecture (ie, they require a central server as an intermediary) that includes 2 roles: a central server (the “server” role) and multiple sites (the “client” role). The server controls the learning process and combines the models generated by the clients. An analog is the shared computation with people’s personal computers, such as the Fight AIDS @ Home project. Although these approaches preserve privacy during the learning process, such a centralized architecture may be vulnerable to attacks (Figure 1B) due to its single point of control, mutable data or records, change provenance trail, and partial visibility.
Figure 1.

Comparison of privacy-preserving learning methods. (A) Basic idea of learning on a centralized architecture (eg, GLORE). Each site computes its local model and submits the model to the central server. The central server combines the models into a global one, and submits the models back to each site. This process continues until the global model converges. (B) Types of attack related to a centralized architecture. The central server is taken over by an attacker during the model learning process. The potential risks are as follows: (1) single point of control: because the attacker now controls the learning process, a site (s in this example) may be assigned more computational work than the others; (2) mutable data/records: if the attacker submits a falsified or fabricated global model to a site (eg, s), no other sites can detect this misconduct, because the transparency about the models is limited; (3) change provenance: the attacker may change the ownership of the local model submitted by a site (eg, s) to another site (eg, s); and (4) partial visibility: a site (s) can only access its own local models and the global models, and therefore cannot see the models from the other sites (s, s, and s). (C) Basic idea of learning on a decentralized architecture (eg, GloreChain). There is no central server. Instead, each site exchanges its models to improve the predictive correctness on a peer-to-peer blockchain network. (D) Reducing risks by using a decentralized architecture. Site s is taken over by an attacker. The risks described in panel B can be mitigated as follows: (1) no single point of control: although the attacker can change or even stop the computational work on site s, the loading of the other sites cannot be easily raised because the learning process is not controlled by s; the attacker cannot overload any other site as would be the case in panel B; (2) immutable data or records: if the attacker who took over s tries to submit a falsified or fabricated model to another site (eg, s), all other sites (eg, s and s) will receive the tampered model on the blockchain network, and thus the probability of detecting such misconduct increases; (3) data provenance: the attacker cannot claim that a model generated by s came from s, because on a blockchain network the source of each model is recorded and is verifiable; and (4) complete visibility: a site (eg, s) can easily see all models from all sites, because on a blockchain network every site has a full copy of all models.

Comparison of privacy-preserving learning methods. (A) Basic idea of learning on a centralized architecture (eg, GLORE). Each site computes its local model and submits the model to the central server. The central server combines the models into a global one, and submits the models back to each site. This process continues until the global model converges. (B) Types of attack related to a centralized architecture. The central server is taken over by an attacker during the model learning process. The potential risks are as follows: (1) single point of control: because the attacker now controls the learning process, a site (s in this example) may be assigned more computational work than the others; (2) mutable data/records: if the attacker submits a falsified or fabricated global model to a site (eg, s), no other sites can detect this misconduct, because the transparency about the models is limited; (3) change provenance: the attacker may change the ownership of the local model submitted by a site (eg, s) to another site (eg, s); and (4) partial visibility: a site (s) can only access its own local models and the global models, and therefore cannot see the models from the other sites (s, s, and s). (C) Basic idea of learning on a decentralized architecture (eg, GloreChain). There is no central server. Instead, each site exchanges its models to improve the predictive correctness on a peer-to-peer blockchain network. (D) Reducing risks by using a decentralized architecture. Site s is taken over by an attacker. The risks described in panel B can be mitigated as follows: (1) no single point of control: although the attacker can change or even stop the computational work on site s, the loading of the other sites cannot be easily raised because the learning process is not controlled by s; the attacker cannot overload any other site as would be the case in panel B; (2) immutable data or records: if the attacker who took over s tries to submit a falsified or fabricated model to another site (eg, s), all other sites (eg, s and s) will receive the tampered model on the blockchain network, and thus the probability of detecting such misconduct increases; (3) data provenance: the attacker cannot claim that a model generated by s came from s, because on a blockchain network the source of each model is recorded and is verifiable; and (4) complete visibility: a site (eg, s) can easily see all models from all sites, because on a blockchain network every site has a full copy of all models. To mitigate these risks, one plausible idea is to adopt a decentralized architecture, as shown in Figure 1C. Without a central server (ie, no server role), the benefits of the decentralized solution include: no single point of control, immutable data or records, ascertained data provenance, and complete visibility (Figure 1D). Based on this idea, recent studies such as ModelChain and ExplorerChain proposed to combine privacy-preserving learning with the blockchain technology,, which is a distributed chain of transaction blocks and has been proposed by many researchers for various healthcare or genomic applications., ModelChain/ExplorerChain execute learning algorithms on the peer-to-peer permissioned blockchain network (ie, participants are preauthorized) and utilize the metadata of the transaction to disseminate the partially trained models. In the design of ModelChain and ExplorerChain, the server role is removed to achieve decentralization (Figure 2A ). Without a server role to oversee and combine the local models, ModelChain/ExplorerChain are actually approximate methods to integrate the learning algorithm with blockchain, and thus have reduced correctness when compared with centralized algorithms such as EXPLORER.
Figure 2.

Comparison of the methods to integrate predictive modeling in a decentralized architecture. (A) Inclusion of only the “client” role to approximate model learning. One of the sites (eg, s) serves as a client at each learning iteration. (B) Inclusion of both client and “server” roles to perform exact model computation. Every site serves as a “client,” and 1 of the sites (eg, s) serves as the “server,” at each learning iteration.

Comparison of the methods to integrate predictive modeling in a decentralized architecture. (A) Inclusion of only the “client” role to approximate model learning. One of the sites (eg, s) serves as a client at each learning iteration. (B) Inclusion of both client and “server” roles to perform exact model computation. Every site serves as a “client,” and 1 of the sites (eg, s) serves as the “server,” at each learning iteration. Owing to the importance of the predictive correctness for cross-institutional healthcare or genomic modeling, the inclusion of both server and client roles can ensure the learning algorithm performs as well as the centralized methods. On the other hand, the fairness of the compute loads sharing should also be considered. That is, every site should share the loads of acting as the server in a fair way. Such fairness is important because the server site is responsible for the combination of the partial models from all sites, and the combination process involves non-negligible computational costs and thereby high energy consumption.

Objective

We sought to develop a general privacy-preserving predictive model sharing framework that achieves 3 goals: (1) preserves the predictive correctness, (2) mitigates the risks of a centralized architecture, and (3) computes the models in a fair way, facilitating cross-institution healthcare or genomic studies and quality improvement initiatives.

MATERIALS AND METHODS

In our proposed framework, to achieve the first goal of preserving predictive correctness, every site serves as both a server and client while computing the model (Figure 2B). With this approach, the high level of correctness can be preserved because the learning process is exactly the same as that of centralized learning. To achieve the second goal of mitigating the risks of centralized architecture, we adopt the peer-to-peer blockchain technology, and can thus inherit the benefits of blockchain (eg, no single point of control). To achieve the third goal of computing the models in a fair way, we propose to adopt the round-robin approach in which every site alternatively serves as server for each learning iteration, as shown in Figure 3 . This approach can avoid computational unfairness, such as what happens when one or few sites acts significantly more times as the server than the other sites. Note that permissioned blockchain platforms such as MultiChain, utilize the round-robin approach as a fast and low computational cost consensus protocol to mitigate the slow and high energy consumption of blockchain.
Figure 3.

An illustration of the alternating client-server method. At each learning iteration, 1 site (eg, s, s and s) serves as the “server” in an alternating manner.

An illustration of the alternating client-server method. At each learning iteration, 1 site (eg, s, s and s) serves as the “server” in an alternating manner. To examine our proposed blockchain-enabled fair compute loads framework, we developed GloreChain (Grid Binary LOgistic REgression on Permissioned BlockChain) as a concrete example. The 3 main components of GloreChain—batch model learning, blockchain data/network, and consensus learning algorithm—are described sequentially in the following 3 subsections, followed by the details of our implementation.

The GLORE batch model learning

There are 2 major types of privacy-preserving predictive modeling algorithms. The first type includes online methods (eg, EXPLORER) that update the model using partial data sequentially and focus on the efficiency of the retraining process (ie, when the data are updated, the model does not need to be completely retrained). The second type includes batch methods (eg, GLORE) that update the model using all data at the same time and focus on learning a model that is exactly the same as the one trained by “traditional” logistic regression (ie, disseminate all data to a single server first and then perform learning). Although our framework is general and can adopt both online and batch methods while still being decentralized, for GloreChain, we adopt the batch learning algorithm GLORE. This is because we focus on obtaining a high-level of predictive correctness. GLORE applies the Newton-Raphson method, and decomposes the derivatives of the log-likelihood function to estimate model coefficients. We denote the client and the server parts of the learning algorithms in GLORE as GLORE-Client and GLORE-Server, respectively.

The blockchain data and network

GloreChain utilizes the metadata of the transaction to disseminate partially trained predictive models (ie, a set of aggregated numerical coefficients or parameters) and relevant meta information. The design of GloreChain is shown in Figure 4 . As shown in Figure 4A, each transaction represents an action of a model, and stores the partial model with meta information in the metadata of the transaction. Table 1 describes the details of each data field stored in the metadata of the transaction on the blockchain. Only partially trained models are disseminated on-chain, and the observation-level patient data are stored off-chain, to improve privacy protection. Also, there is no transaction amount (ie, the coin value transferred within every transaction is 0), and thus GloreChain takes advantage of only the distributed data ledger but not of the cryptocurrency aspect of blockchain.
Figure 4.

The design of GloreChain. (A) The simplified blocks and transactions. The details of the data fields stored in the metadata of the transaction are described in Table 1. (B) The implementation architecture of GloreChain, based on an 8-site configuration. The Blockchain-Connector is a software component we programmed as an interface of the GloreChain and the underlying MultiChain blockchain platform. PHI: protected health information; VM: virtual machine.

Table 1.

The on-chain data of GloreChain demonstrating an example of the transfer action of a model.

FieldDescriptionPossible ValuesExample
Model MeanThe mean vector of the GLORE partial model13A numerical vector with its length equals m + 1[−1.046593, −63.844112, 4.276981]
Model CovarianceThe variance-covariance matrix of the GLORE partial model13A numerical (m + 1) x (m + 1) square symmetric matrix[[1.413845, 43.833862, 22.948767], [43.833862, 2985.832248, 720.785133], [22.948767, 720.785133, 487.095031]]
FlagThe type of action a site has taken to the modelUNKNOWN, INITIALIZE, UPDATE, TRANSFER, CONSENSUS, TEST, CLEARTRANSFER
ResultThe value of the evaluation metric when the learning process is completeA numerical value between 0 and 10.921875
From SiteThe site that has submitted the modelA unique name or identifier representing the siteSan Diego Hospital
To SiteThe site that will receive the modelA unique name or identifier representing the siteSan Diego Hospital
TimeThe time that the site submitted the modelA timestamp2018–10–16 11: 29: 41
IterationThe current iteration of the learning processA non-negative integer7

In this example, m = 2 is the number of the covariates in the dataset. The partial model of GLORE contains both Model Mean (eg, the m + 1 = 3-dimensional vector) and Model Covariance (eg, the (m + 1) x (m + 1) = 3 x 3 matrix), while the final model is the consensus mean vector. The Flag is TRANSFER, representing the submission of a global model to the blockchain (in contrast to UPDATE, representing the submission of a local model). The Result indicates the value of the evaluation metric for correctness (eg, the full area under the receiver operating characteristic curve). In this round, San Diego Hospital acts as the “server,” and thus it is both the From Site and To Site. The Time field stands for the timestamp of the transaction, and the Iteration is the number of learning iterations

The on-chain data of GloreChain demonstrating an example of the transfer action of a model. In this example, m = 2 is the number of the covariates in the dataset. The partial model of GLORE contains both Model Mean (eg, the m + 1 = 3-dimensional vector) and Model Covariance (eg, the (m + 1) x (m + 1) = 3 x 3 matrix), while the final model is the consensus mean vector. The Flag is TRANSFER, representing the submission of a global model to the blockchain (in contrast to UPDATE, representing the submission of a local model). The Result indicates the value of the evaluation metric for correctness (eg, the full area under the receiver operating characteristic curve). In this round, San Diego Hospital acts as the “server,” and thus it is both the From Site and To Site. The Time field stands for the timestamp of the transaction, and the Iteration is the number of learning iterations The design of GloreChain. (A) The simplified blocks and transactions. The details of the data fields stored in the metadata of the transaction are described in Table 1. (B) The implementation architecture of GloreChain, based on an 8-site configuration. The Blockchain-Connector is a software component we programmed as an interface of the GloreChain and the underlying MultiChain blockchain platform. PHI: protected health information; VM: virtual machine. Regarding the type of the underlying blockchain network, a permissioned one (ie, only permitted participants can join the network) is suitable for GloreChain. We only store the aggregated partial models on-chain, but a permissioned blockchain network provides an additional layer of privacy protection. It should also be noted that GloreChain provides nonfinancial incentives (ie, improved predictive power) instead of financial ones (ie, coins) for each participating site to contribute to the computation (eg, learning the models, creating the blocks, and verifying the transactions). In a complex network, however, not all sites participate in a study, so it is important that all institutions (ie, nodes in the network) know how to build a model and are thus not dependent on a single institution to act as a “server.”

The Proof-of-Equity consensus learning algorithm

We designed a new algorithm, Proof of Equity (PoE), to determine a fair order for each site to serve as the server in a round-robin way. The PoE algorithm determines the order alphabetically using the unique name or identifier that represents a site. Each site first submits to the blockchain its unique name or identifier (eg, “San Diego Hospital” and “Davis Hospital”), and then every site retrieves the names or identifiers from blockchain to determine the order, alphabetically (eg, “Davis Hospital” → “San Diego Hospital”). Thereafter, the learning process starts, and each site follows the predetermined order to serve as the server for each iteration, in a round-robin way (eg, “Davis Hospital” → “San Diego Hospital” → “Davis Hospital” → “San Diego Hospital” →…). This process continues until the model converges or the maximum number of learning iterations is reached. We then regard the final predictive model as the consensus model. A running example of the PoE algorithm is shown in Figure 5 , while the details are described in Algorithm 1 (the high-level PoE algorithm in GloreChain). There are 4 hyperparameters in the algorithm: the polling time period Δ, the waiting time period Θ, the maximum iteration Ω, and the total number of participating sites N. The size of the model (including mean and variance) is O(m2), where m is the total number of covariates.
Figure 5.

An example of the Proof-of-Equity (PoE) algorithm (ie, model consensus protocol). The notations are as follows: I denotes the initial transaction on site s for determining the order, Li is the local model at iteration i on site s, and Gi is the global model in iteration i. (A) Order determination. Each site s first starts the ordering process by submitting an initial transaction I, which includes the unique name/identifier of the site (Los Angeles Hospital, Irvine Hospital, Davis Hospital, and San Diego Hospital in our example), to the blockchain. Then, each site collects the initial transactions from all sites, and then starts the checking process by ordering the unique name or identifier in an alphabetical order (eg,  → → → , or s → s → s → s). In this round-robin way, the same order is determined on each site (eg, s → s → s → s → s → s → s → s →…). (B) Initialization of the predictive model. Next, each site starts to compute the initial local model (in local iteration 0), and submits its local model (ie, L0 on site s) to the blockchain. Then, based on the predetermined order, the first “server” site (s) computes the initial global model (in global iteration 0), and submits the global model (G0) to the blockchain. Note that, every site serves as a “client,” while only s acts as the “server,” in this initial learning process. Therefore, sites s, s, and s do not perform global model computation to avoid duplicated work and unnecessary energy consumption. (C) The first iteration. At iteration 1, each site first computes its local model (ie, L1 on site s) based on the initial global model (G0), and submits L1 to the blockchain. Then, the next server site (s) computes the global model (G1) and submits G1 to the blockchain. Other sites (s, s, and s) skip the global model computation and serve as clients only. (D) The final or consensus iteration. The learning process repeats until a converged model is identified or the maximum number of iterations is reached. For the example shown in panel D, at iteration 7, site s acts as the server, and after the global model computation it determines that the global model (G7) converged. That is, G7 is very close to G6, the global model in the previous iteration, based on the convergence criterion (10−6 precision in our experiments). In this case, s submits this consensus predictive model G7 to the blockchain, and the PoE algorithm is complete.

An example of the Proof-of-Equity (PoE) algorithm (ie, model consensus protocol). The notations are as follows: I denotes the initial transaction on site s for determining the order, Li is the local model at iteration i on site s, and Gi is the global model in iteration i. (A) Order determination. Each site s first starts the ordering process by submitting an initial transaction I, which includes the unique name/identifier of the site (Los Angeles Hospital, Irvine Hospital, Davis Hospital, and San Diego Hospital in our example), to the blockchain. Then, each site collects the initial transactions from all sites, and then starts the checking process by ordering the unique name or identifier in an alphabetical order (eg,  → → → , or s → s → s → s). In this round-robin way, the same order is determined on each site (eg, s → s → s → s → s → s → s → s →…). (B) Initialization of the predictive model. Next, each site starts to compute the initial local model (in local iteration 0), and submits its local model (ie, L0 on site s) to the blockchain. Then, based on the predetermined order, the first “server” site (s) computes the initial global model (in global iteration 0), and submits the global model (G0) to the blockchain. Note that, every site serves as a “client,” while only s acts as the “server,” in this initial learning process. Therefore, sites s, s, and s do not perform global model computation to avoid duplicated work and unnecessary energy consumption. (C) The first iteration. At iteration 1, each site first computes its local model (ie, L1 on site s) based on the initial global model (G0), and submits L1 to the blockchain. Then, the next server site (s) computes the global model (G1) and submits G1 to the blockchain. Other sites (s, s, and s) skip the global model computation and serve as clients only. (D) The final or consensus iteration. The learning process repeats until a converged model is identified or the maximum number of iterations is reached. For the example shown in panel D, at iteration 7, site s acts as the server, and after the global model computation it determines that the global model (G7) converged. That is, G7 is very close to G6, the global model in the previous iteration, based on the convergence criterion (10−6 precision in our experiments). In this case, s submits this consensus predictive model G7 to the blockchain, and the PoE algorithm is complete.

Algorithm 1

: The local data D, the polling time period Δ, the waiting time period Θ, the maximum iteration Ω, and the total number of participating sites N. : The consensus batch predictive model M. Step 1. Submit an initial transaction to the blockchain. Step 2. Check the blockchain every time Δ until the initial transactions from all N sites are received, and determine the learning order alphabetically using the unique name or identifier of each site. Step 3. Wait for time Θ to let every site determine the learning order. Step 4. Initialize the global model G by setting all coefficients to zeroes. Step 5. Wait for time Θ to let every site initialize its global model. Step 6. Loop until G converges or the maximum iteration Ω is reached. Step 6.1. Wait for time Θ, update the local model L using G and local data D through the GLORE-Client learning algorithm, and submit L to the blockchain. Step 6.2. If this site is the next server site, check the blockchain every time Δ until all N local models are received, update model G using all local models received through the GLORE-Server learning algorithm, and then submit G to the blockchain. Step 6.3. If this site is not the next server site, check the blockchain every time Δ until the next G is found. Step 7. The consensus model is M = G.

The implementation of GloreChain

The architecture of GloreChain is demonstrated in Figure 4B. GloreChain, as well as the blockchain-connector and the two GLORE components (ie, GLORE-Client and GLORE-Server, as introduced in Section 3.1) were written in Java. The code of GLORE was refactored to the application programming interface that can be called by GloreChain, while the original modeling capabilities remained the same. Note that PHI was only used to compute the local model (ie, GLORE-Client) and was not disseminated to the blockchain network. Based on a recent survey of the blockchain platforms, we adopted MultiChain, in GloreChain, because (1) MultiChain is based on the popular Bitcoin Blockchain, and (2) MultiChain is developed to serve as a general-purpose permissioned blockchain network. We used its default parameters and consensus protocol (ie, Mining Diversity), in our implementation. The computation environment for GloreChain is the iDASH (integrating Data for Analysis, Anonymization, and SHaring) cloud network,, a private cloud network compliant with requirements from the Health Insurance Portability and Accountability Act. We simulated the multisite scenarios (2, 4, and 8 sites in our experiment) on iDASH virtual machines (VMs). Each VM included 2 of Intel Xeon 2.30 GHz CPUs, 8 GB of RAM, and 100GB of storage.

Datasets

We evaluated GloreChain on 3 healthcare or genomic test datasets, and each of them had a binary outcome. First, the Edinburg (Edin) dataset was used to predict whether myocardial infarction was present, based on phenotyping features. Next, the cancer biomarkers (CA) dataset was used to predict whether the cancer was present based on biomarkers (ie, CA-19 and CA-125). Finally, the total hip arthroplasty (THA) dataset, was used to predict whether a patient’s actual hospital length of stay was greater than the expected length of stay (3 days) for the unilateral primary THA surgery at University of California, San Diego (UCSD), based on various covariates including demographics, osteoarthritis grade, surgical approach, and comorbidities. Table 2 summarizes the statistics of the 3 datasets.
Table 2.

Statistics of the test datasets evaluated in our experiments, including the percentage of the positive or negative classes

DatasetEdinCATHA
DescriptionMyocardial infarctionCancerLength of hospitalization >3 days
Covariates9234
Observations1253141960
Class Distribution (Positive/Negative)0.219/0.7810.638/0.3620.278/0.722
OutcomePresence of diseasePresence of cancerHospital length of stay for total hip arthroplasty is >3 days

The values for the myocardial infarction (Edin) and cancer biomarker (CA) datasets are reproduced from Wang et al and the values for the length of hospitalization (total hip arthroplasty [THA]) dataset are reproduced from Kuo et al.

Statistics of the test datasets evaluated in our experiments, including the percentage of the positive or negative classes The values for the myocardial infarction (Edin) and cancer biomarker (CA) datasets are reproduced from Wang et al and the values for the length of hospitalization (total hip arthroplasty [THA]) dataset are reproduced from Kuo et al. These test datasets were used to evaluate privacy-preserving predictive models in the previous studies.,, For the THA dataset, the Institutional Review Board at UCSD approved this research with Project Number 171344X on February 9, 2018. Also, the Human Research Protections Program at UCSD exempted the informed consent requirement, because the THA dataset was defined as Health Insurance Portability and Accountability Act de-identified and contained no sensitive patient health information.

Experiment settings

The goal of our experiments was to evaluate whether GloreChain, containing the alternating server and client roles, could in practice provide the exact same predictive power as centralized methods, while obtaining the benefits of the underlying decentralized blockchain network. Therefore, we compared GloreChain with GLORE, the state-of-the-art batch learning method, in our experiments. The convergence criterion was set to 10−6 precision, the same value described in Wu et al, for both methods. The hyperparameters for GloreChain were configured as the following: the polling time period Δ = 1 (second), the waiting time period Θ = 30 (seconds), the maximum iteration Ω = 20, and the total number of participating sites N = 2, 4, or 8. We selected the time period hyperparameters Δ and Θ according to the learning speed and the network latency based on a previous study. The maximum iteration Ω was selected based on the average iteration numbers for convergence (ie, 7 and 12 iterations for the Edin and the CA datasets, respectively) Also, we checked the latest N transactions with the size of the transaction metadata >20 in the PoE algorithm. Note that a waiting time was added in each iteration for GLORE for the purpose of synchronization. We chose this per-iteration waiting time in GLORE to be the same as the value of Θ (ie, 30 seconds) in GloreChain, to retain the fairness of the comparison. We evenly and randomly split each dataset for the 2-, 4-, and 8-site scenarios. For each site, the dataset was randomly divided into 80% and 20% training and the testing records, respectively. Note that each training or testing dataset preserved a similar class distribution as described in Table 2, and contained at least 1 positive and 1 negative record. The full area under the receiver-operating characteristic curve (AUC) on the test datasets was adopted as our evaluation measure. We used the averaged test AUC among all N sites as the result for the predictive correctness. We repeated the previously mentioned process (ie, data splitting, model learning, and averaged test AUC computation) over 30 trials. We also compared consensus iterations and the execution time of GloreChain and GLORE. To collect the results, a pausing time of 240 seconds was added between each trial for both GloreChain and GLORE, and this per-trial waiting time was deducted in the execution time computation. It should be noted that, for the N site configuration, GLORE actually requires N + 1 VMs (ie, N VMs for the clients and 1 VM for the server), while GloreChain requires exactly N VMs.

RESULTS

The comparison results for predictive correctness and number of learning iterations are shown in Table 3. For both correctness and iterations, the results of GloreChain are exactly the same as the ones from GLORE. In general, the mean AUC, the standard deviation of AUC, and the mean of the iterations remained stable in different scenarios (ie, 2, 4, and 8 sites) for the same dataset.
Table 3.

The predictive correctness and number of learning iterations for the myocardial infarction (Edin), CA, and length of hospitalization (THA) datasets for the 2-, 4-, and 8-site scenarios

DatasetnCorrectness (AUC)
Number of Iterations
GLORE
GloreChain
GLORE
GloreChain
MeanSDMeanSDMeanSDMeanSD
Edin20.9650.0130.9650.0136.9670.1836.9670.183
40.9620.0100.9620.0107.0000.0007.0000.000
80.9590.0130.9590.0137.4332.3737.4332.373
CA20.8930.0540.8930.05411.6330.4911.6330.490
40.8660.0710.8660.07111.8330.37911.8330.379
80.9000.0580.9000.05811.8000.40711.8000.407
THA20.7360.0340.7360.03417.5005.68617.5005.686
40.7410.0460.7410.04617.5005.68617.5005.686
80.7220.0400.7220.04017.0006.10317.0006.103

The evaluation metric for correctness is the averaged full AUC among N sites for 30 trials.

AUC: area under the receiver-operating characteristic curve; CA: cancer biomarkers; Edin: Edinburg; THA: total hip arthroplasty.

The predictive correctness and number of learning iterations for the myocardial infarction (Edin), CA, and length of hospitalization (THA) datasets for the 2-, 4-, and 8-site scenarios The evaluation metric for correctness is the averaged full AUC among N sites for 30 trials. AUC: area under the receiver-operating characteristic curve; CA: cancer biomarkers; Edin: Edinburg; THA: total hip arthroplasty. On the other hand, the standard deviations of an iteration in some combinations of datasets and scenarios (ie, Edin for 8 sites, and THA for 2, 4, and 8 sites) were relatively high. Therefore, we further investigated the iterations per trial for each dataset and for each scenario. As shown in Figure 6 , 1 of the 30 trials for Edin and the 8-site configuration reaches the maximum iteration Ω (ie, 20). Also, for the THA dataset, most of the trials reach the maximum iteration, while some of the trials converged in 5 iterations. Although increasing the maximum iteration Ω can potentially improve the predictive power, our goal is to evaluate whether GloreChain performs exactly the same as GLORE in terms of correctness, instead of improving the predictive power for each dataset. Also, current correctness results are already comparable to ones shown in the previous studies.,, Therefore, we kept the hyperparameter Ω as 20 in our experiments.
Figure 6.

The pretrial iterations on the 3 scenarios (2, 4, and 8 sites) for each of the 30 trials, with the maximum iteration Ω set to 20. (A) Edinburg (Edin) dataset, (B) cancer biomarkers (CA) dataset, (C) total hip arthroplasty (THA) dataset.

The pretrial iterations on the 3 scenarios (2, 4, and 8 sites) for each of the 30 trials, with the maximum iteration Ω set to 20. (A) Edinburg (Edin) dataset, (B) cancer biomarkers (CA) dataset, (C) total hip arthroplasty (THA) dataset. Table 4   illustrates the total and the per-iteration execution time results. GloreChain has higher total execution time when compared with GLORE because of the additional time requirement of the PoE algorithm to determine the learning order of the institutions at the beginning of the modeling process, as described in Section 3.3. Also, GloreChain had higher running time (about 3–8 times slower than GLORE in terms of the per-iteration running time), because GloreChain adopts a blockchain network and therefore incurs additional time to create and synchronize the transactions and blocks.
Table 4.

The execution time (in seconds) for N sites in 30 trials.

DatasetnOverall
Per Iteration
GLORE
GloreChain
GLORE
GloreChain
TotalRunningTotalRunningTotalRunningTotalRunning
Edin2239.1960.168342.7570.63834.3340.02449.2000.092
4240.1580.219350.9640.64434.3080.03150.1380.092
8253.1390.228367.3160.76034.0550.03149.4150.102
CA2379.0940.178488.3260.71732.5870.01541.9760.062
4385.1450.182505.6840.77432.5470.01542.7340.065
8384.1210.192509.4400.85632.5530.01643.1730.073
THA2555.3220.275687.9531.70131.7330.01639.3120.097
4555.3010.302702.5451.90831.7310.01740.1450.109
8540.3130.357691.7712.88331.7830.02140.6920.170

All measurements are in seconds and are averaged over N sites. The total time (ie, the “real” time in the Linux system) includes both running time (ie, the “user + sys” time in the Linux system) and synchronization time (ie, total minus running time). We divided the total time by the mean of iteration (as shown in Table 3) to compute the per-iteration time. The additional per-trial pausing time (240 seconds) was deducted in the computation.

CA: cancer biomarkers; Edin: Edinburg; THA: total hip arthroplasty.

The execution time (in seconds) for N sites in 30 trials. All measurements are in seconds and are averaged over N sites. The total time (ie, the “real” time in the Linux system) includes both running time (ie, the “user + sys” time in the Linux system) and synchronization time (ie, total minus running time). We divided the total time by the mean of iteration (as shown in Table 3) to compute the per-iteration time. The additional per-trial pausing time (240 seconds) was deducted in the computation. CA: cancer biomarkers; Edin: Edinburg; THA: total hip arthroplasty. GLORE used 3, 5, and 9 VMs for the 2-, 4-, and 8-site scenarios, respectively, while GloreChain used exactly 2, 4 and 8 VMs for the corresponding scenarios. Therefore, the actual difference of the aggregated running time on all VMs is smaller. For example, the per-iteration running time for the Edin dataset and the 2-site setting of GLORE and GloreChain are 0.024 and 0.092, respectively (GloreChain takes 4 times longer than GLORE). However, considering the aggregated running time by multiplying the number of VMs (3 for GLORE and 2 for GloreChain) in the same dataset and scenario, GloreChain takes only 2.5 times longer than GLORE.

DISCUSSION

Based on the results, GloreChain, with its alternating server and client roles, had exactly the same predictive power as well as number of learning iterations when compared with GLORE. Additionally, GloreChain possessed the advantages of adopting the blockchain technology, such as no single point of control. The cost of inheriting these additional benefits from blockchain technology was a slight increase in execution time (including both running and synchronization time) when compared with an equivalent approach that did not use blockchain. This slight increase was minimal when compared with what would have been the energy consumption if we had decided to utilize the typical Proof-of-Work protocol, that made Bitcoin blockchain famous worldwide for promoting the use of computer “farms” by block miners., We used instead a low energy-consumption protocol (ie, Mining Diversity) that was sufficiently secure to run GloreChain on a permissioned blockchain network (eg, MultiChain)., In fact, we used fewer machines to run GloreChain than we would have used had we followed other approaches (eg, GLORE) because the central server and its backups were not needed due to the dis-intermediation and redundancy features provided by blockchain technology. Although we adopted GLORE in GloreChain, the core privacy-preserving learning can be replaced by any other centralized algorithm. That is, our framework, containing both server and client roles, provides a general and flexible framework that supports any client-server–based algorithms. In related methods, such as those used in ModelChain/ExplorerChain, the server is replaced by a consensus algorithm to determine the learning order using model errors, stored on-chain as a field in the metadata of the transaction., Therefore, ModelChain/ExplorerChain can only integrate online learning algorithms with blockchain. GloreChain does not require model errors to determine the order during the learning process, and instead includes a field “Result” to store the value of AUC for computing the overall predictive correctness results (Figure 4A and Table 1). The 3 intrinsic challenges of the blockchain networks: transparency or confidentiality, speed or scalability, and the threat of a 51% attack (ie, the blockchain network is taken over by the majority of malicious nodes), are not critical for GloreChain because: (1) GloreChain only disseminates partial models and not PHI on the blockchain, and thus minimizes the risk of confidentiality breaches; (2) the execution time of GloreChain is large (40–50 seconds per iteration) when compared with the transaction time of the underlying blockchain platform (eg, 1000 transactions per second at a maximum for MultiChain), and the difference may be much larger for the use cases on big data; and (3) GloreChain is based on a permissioned blockchain network and the participating nodes are preauthorized, therefore mitigating the risk of a 51% attack. In the PoE algorithm, we used the alphabetical order of the unique site name or identifier to determine the next server site, and more sophisticated methods may be applied so that the institutions at the top of the order do not work more than others. For example, a different order may be assigned every N iterations (eg, “s → s → s → s,” followed by “s → s → s → s”). Although our method is relatively simple, in the long run, the computational cost and energy consumption for each site can still be as “fair” as more complicated methods. Note that our framework is based on a “semitrust” assumption, where every site would like to improve the predictive correctness by sharing the aggregated partial model, but might be “curious” about other sites. Moreover, we assume syntactic and semantic interoperability for each site (ie, they adopted the same data format, meaning, and standards). Regarding the hyperparameters, the polling time period Δ decides the balance of system responsiveness and network burden, the waiting time period Θ determines execution speed while considering potential network latency, and the maximum iteration Ω represents the trade-off between predictive correctness and learning efficiency. These hyperparameters can be adjusted based on the actual use case scenarios. Also, the size of the metadata of the transaction is about 5KB for THA (ie, the largest dataset in our experiment), and is way below the default size limit of MultiChain (ie, 2 MB). It should also be noted that, although we implemented GloreChain based on MultiChain, the GloreChain framework itself is platform independent, and can adopt other blockchain platforms such as Ethereum or Hyperledger by changing the Blockchain-Connector (as shown in Figure 4B). Finally, the deployment on the iDASH private cloud compute environment also improved the security level of the permissioned blockchain network.

Limitations

The limitations for this study are as follows: (1) our framework was not evaluated on very high-dimensional datasets, which can create large predictive models and therefore have impacts on the size of the metadata of the transactions, the speed to disseminate partial models over the blockchain network, and the additional process to adjust the hyperparameters; (2) several use case situations were not tested, such as nonrepresentative samples, very different data distribution among all sites, and low-quality models due to poor data; (3) privacy-preserving concerns, such as those often raised by differential privacy advocates (ie, the potential privacy breach because of the very small data size in some sites), were not fully studied; and (4) more investigations about the potential ethical, legal, and social implications arising from decentralized computer system access are yet to be conducted.

CONCLUSION

By including both server and client roles in each learning iteration, adopting blockchain as the underlying peer-to-peer network, and alternating the roles for each site in a round-robin way, our proposed framework preserves the prediction correctness, obtains the benefits of decentralization, and ensures the computational fairness for predictive model building. GloreChain, an example of our framework, demonstrates the capability to reach exactly the same predictive power and number of learning iterations as the state-of-the-art centralized method. Considering the critical and sensitive nature of healthcare or genomic data, the exchange of additional execution time for benefits such as no single point of control may be considered attractive by some institutions. Although more evaluations and refinements are warranted, our framework provides a promising potential for multiple institutions to collaboratively learn a healthcare or genomic predictive model in a privacy-preserving and decentralized way, using blockchain platform that are maintained by a large community of software developers worldwide, as opposed to custom software.

FUNDING

The study was supported by National Institutes of Health grants K99HG009680 and R00HG009680 (T-TK) and R01GM118609 (LO-M); and the National Human Genome Research Institute of the National Institutes of Health and R00HG009680. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

AUTHOR CONTRIBUTORS

T-TK contributed in conceptualization, data curation, formal analysis, funding acquisition, investigation, methodology, project administration, resources, software, validation, visualization, and writing (original draft). RAG contributed in data curation and writing (review & editing). LO-M contributed in conceptualization, funding acquisition, project administration, resources, supervision, and writing (review and editing).
  28 in total

1.  Privacy issues in personalized medicine.

Authors:  Laszlo T Vaszar; Mildred K Cho; Thomas A Raffin
Journal:  Pharmacogenomics       Date:  2003-03       Impact factor: 2.533

2.  A Proposed Solution and Future Direction for Blockchain-Based Heterogeneous Medicare Data in Cloud Environment.

Authors:  Harleen Kaur; M Afshar Alam; Roshan Jameel; Ashish Kumar Mourya; Victor Chang
Journal:  J Med Syst       Date:  2018-07-10       Impact factor: 4.460

3.  To share or not to share: that is not the question.

Authors:  Lucila Ohno-Machado
Journal:  Sci Transl Med       Date:  2012-12-19       Impact factor: 17.956

4.  Blockchain Technology: Applications in Health Care.

Authors:  Suveen Angraal; Harlan M Krumholz; Wade L Schulz
Journal:  Circ Cardiovasc Qual Outcomes       Date:  2017-09

5.  A framework for secure and decentralized sharing of medical imaging data via blockchain consensus.

Authors:  Vishal Patel
Journal:  Health Informatics J       Date:  2018-04-25       Impact factor: 2.681

6.  Healthcare Blockchain System Using Smart Contracts for Secure Automated Remote Patient Monitoring.

Authors:  Kristen N Griggs; Olya Ossipova; Christopher P Kohlios; Alessandro N Baccarini; Emily A Howson; Thaier Hayajneh
Journal:  J Med Syst       Date:  2018-06-06       Impact factor: 4.460

7.  A secure protocol for protecting the identity of providers when disclosing data for disease surveillance.

Authors:  Khaled El Emam; Jun Hu; Jay Mercer; Liam Peyton; Murat Kantarcioglu; Bradley Malin; David Buckeridge; Saeed Samet; Craig Earle
Journal:  J Am Med Inform Assoc       Date:  2011-05-01       Impact factor: 4.497

8.  Comparison of blockchain platforms: a systematic review and healthcare examples.

Authors:  Tsung-Ting Kuo; Hugo Zavaleta Rojas; Lucila Ohno-Machado
Journal:  J Am Med Inform Assoc       Date:  2019-05-01       Impact factor: 4.497

9.  ClinVar: public archive of interpretations of clinically relevant variants.

Authors:  Melissa J Landrum; Jennifer M Lee; Mark Benson; Garth Brown; Chen Chao; Shanmuga Chitipiralla; Baoshan Gu; Jennifer Hart; Douglas Hoffman; Jeffrey Hoover; Wonhee Jang; Kenneth Katz; Michael Ovetsky; George Riley; Amanjeev Sethi; Ray Tully; Ricardo Villamarin-Salomon; Wendy Rubinstein; Donna R Maglott
Journal:  Nucleic Acids Res       Date:  2015-11-17       Impact factor: 16.971

Review 10.  Blockchain distributed ledger technologies for biomedical and health care applications.

Authors:  Tsung-Ting Kuo; Hyeon-Eui Kim; Lucila Ohno-Machado
Journal:  J Am Med Inform Assoc       Date:  2017-11-01       Impact factor: 4.497

View more
  13 in total

1.  Comparison of Smart Contract Blockchains for Healthcare Applications.

Authors:  Hongru Yu; Haiyang Sun; Danyi Wu; Tsung-Ting Kuo
Journal:  AMIA Annu Symp Proc       Date:  2020-03-04

Review 2.  Functional genomics data: privacy risk assessment and technological mitigation.

Authors:  Gamze Gürsoy; Tianxiao Li; Susanna Liu; Eric Ni; Charlotte M Brannon; Mark B Gerstein
Journal:  Nat Rev Genet       Date:  2021-11-10       Impact factor: 53.242

3.  Potent Blockchain-Enabled Socket RPC Internet of Healthcare Things (IoHT) Framework for Medical Enterprises.

Authors:  Abdullah Lakhan; Tor Morten Groenli; Arnab Majumdar; Pattaraporn Khuwuthyakorn; Fida Hussain Khoso; Orawit Thinnukool
Journal:  Sensors (Basel)       Date:  2022-06-08       Impact factor: 3.847

4.  The Current State of Research, Challenges, and Future Research Directions of Blockchain Technology in Patient Care: Systematic Review.

Authors:  Polina Durneva; Karlene Cousins; Min Chen
Journal:  J Med Internet Res       Date:  2020-07-20       Impact factor: 5.428

5.  EXpectation Propagation LOgistic REgRession on permissioned blockCHAIN (ExplorerChain): decentralized online healthcare/genomics predictive model learning.

Authors:  Tsung-Ting Kuo; Rodney A Gabriel; Krishna R Cidambi; Lucila Ohno-Machado
Journal:  J Am Med Inform Assoc       Date:  2020-05-01       Impact factor: 4.497

6.  The anatomy of a distributed predictive modeling framework: online learning, blockchain network, and consensus algorithm.

Authors:  Tsung-Ting Kuo
Journal:  JAMIA Open       Date:  2020-07-06

7.  iDASH secure genome analysis competition 2018: blockchain genomic data access logging, homomorphic encryption on GWAS, and DNA segment searching.

Authors:  Tsung-Ting Kuo; Xiaoqian Jiang; Haixu Tang; XiaoFeng Wang; Tyler Bath; Diyue Bu; Lei Wang; Arif Harmanci; Shaojie Zhang; Degui Zhi; Heidi J Sofia; Lucila Ohno-Machado
Journal:  BMC Med Genomics       Date:  2020-07-21       Impact factor: 3.063

8.  Privacy-preserving model learning on a blockchain network-of-networks.

Authors:  Tsung-Ting Kuo; Jihoon Kim; Rodney A Gabriel
Journal:  J Am Med Inform Assoc       Date:  2020-03-01       Impact factor: 4.497

9.  Blockchain in Healthcare: Insights on COVID-19.

Authors:  Antonio Fusco; Grazia Dicuonzo; Vittorio Dell'Atti; Marco Tatullo
Journal:  Int J Environ Res Public Health       Date:  2020-09-30       Impact factor: 3.390

Review 10.  Blockchain in Health Care Innovation: Literature Review and Case Study From a Business Ecosystem Perspective.

Authors:  Shuchih Ernest Chang; YiChian Chen
Journal:  J Med Internet Res       Date:  2020-08-31       Impact factor: 5.428

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.