1,243
Views
0
CrossRef citations to date
0
Altmetric
Research Article

A data sharing method for remote medical system based on federated distillation learning and consortium blockchain

ORCID Icon, , , , &
Article: 2186315 | Received 04 Oct 2022, Accepted 27 Feb 2023, Published online: 13 Mar 2023

Abstract

With the development of Medical Internet of Things (MIoT) technology and the global COVID-19 pandemic, hospitals gain access to patients’ health data from remote wearable medical equipment. Federated learning (FL) addresses the difficulty of sharing data in remote medical systems. However, some key issues and challenges persist, such as heterogeneous health data stored in hospitals, which leads to high communication cost and low model accuracy. There are many approaches of federated distillation (FD) methods used to solve these problems, but FD is very vulnerable to poisoning attacks and requires a centralised server for aggregation, which is prone to single-node failure. To tackle this issue, we combine FD and blockchain to solve data sharing in remote medical system called FedRMD. FedRMD use reputation incentive to defend against poisoning attacks and store reputation values and soft labels of FD in Hyperledger Fabric. Experimenting on COVID-19 radiography and COVID-Chestxray datasets shows our method can reduce communication cost, and the performance is higher than FedAvg, FedDF, and FedGen. In addition, the reputation incentive can reduce the impact of poisoning attacks.

1. Introduction

With the ongoing COVID-19 pandemic transmission, people's health is being seriously threatened (Rahman et al., Citation2020). Traditional medical technology requires patients to go to the hospital in person. With the development of the remote medical system, it can reduce patients’ commuting time to the hospital and improve doctors’ diagnostic efficiency (Bashshur et al., Citation2020). While benefiting from remote medical system infuses new vitality into clinical scientific research, epidemic prevention and public health management, it also brings privacy leakage. According to the OCR HHS, the number of victims of medical data leakage increased over 1.5 times compared to 2020 (Maria, Citation2022). Due to policy and legal restrictions and the awareness of patients to protect their privacy and security, it is difficult to share data from the hospital's local server with other institutions. Lack of sufficient data samples in hospital servers for machine learning (Zheng et al., Citation2022), which results in isolated data islands. Therefore, there is an urgent need for a secure method of data sharing in remote medical systems.

Fortunately, Federated learning (FL) is a distributed computing technique (McMahan et al., Citation2017), that solves isolated data islands. Traditional machine learning transmits collected medical data to a cloud server for training. FL reduces the risk of data leakage by deploying the model on a local device and transferring the model training parameters. Instead of patients’ medical data directly to a cloud server with high computing power. This way guarantees the privacy of the user's data and solves the data isolations problem.

However, the latest research shows that there are still some critical issues for FL. For example, first, Melis et al. (Citation2019) proof gradient updates may reveal important information about customer training. Attackers can recover data from local servers uploaded gradients (Hitaj et al., Citation2017). In addition, the FL training process contains millions of parameters because of the large CNN model. Transmitting these massive model parameters will produce a vast communication cost and increase the risk of malicious attacks during the model training process. Medical data is often heterogeneous in the remote medical system. Second, Yang and Sun (Citation2022) believe that only interacting common information from training results on different devices can improve local personalised training results more effectively.

In order to solve the problem of medical data heterogeneity in the remote medical system and the high communication cost of FL. Knowledge distillation combined with FL provides a new solution (Hard et al., Citation2018). Wang et al. (Citation2020) proposed a knowledge distillation method for object recognition. They achieved this by deploying the model on the cloud and fog nodes and the online bidirectional model heterogeneous training on the distributed data set. Oh et al. (Citation2020) proposed a distributed machine learning framework with high communication efficiency and privacy protection to solve the problem of asymmetric uplink and downlink capacity. The output of the local model using the FD algorithm is uploaded to the server in the uplink, the global model parameters of FL are downloaded to the downlink. Sattler et al. (Citation2021) used auxiliary data for unsupervised pre-training and deterministic weighted integrated distillation. The differential privacy method is used to make a weighted ensemble prediction for the auxiliary data according to the determinism of each client model. Kumar et al. (Citation2021) proposed a secure framework for FL in a hostile environment. Knowledge distillation and model inversion methods to ensure security and privacy features and reduce communication cost.

Although many excellent solutions to solve the problem of heterogeneous FL models and communication overhead, we have noticed that they can't control the quality of data transmitted by participants in the process of FL training. The attacker will upload low-quality data and affect the performance of the global model. Thus it is affected by poisoning attacks. And it relies on a centralised server aggregation model, but the central server is prone to failure (Fung et al., Citation2018). Blockchain technology allows model aggregation and task training to be securely distributed. Blockchain is defined as a peer-to-peer distributed ledger for recording transactions, stored in a chain of connected blocks (Crosby et al., Citation2016). The development of blockchain technology provides FL with a decentralised global model convergence environment in which all model updates are verified. It is stored in the blockchain in a distributed manner through the blockchain consensus mechanism. Smart contract calls are used to select nodes in FL, thus controlling the uploaded data to be of high quality. Hyperledger Fabric as a representative of the consortium blockchain. It is an open-source, enterprise-grade implementation platform maintained by IBM and the Linux Foundation (Androulaki et al., Citation2018). Unlike Bitcoin and Ethereum, Hyperledger Fabric does not have any cryptocurrency.

In this paper, we propose a data sharing method in remote medical system called FedRMD based on FD and consortium blockchain, respectively. To summarise, our major contributions are as follows.

  • We propose an FD approach called FedRMD to solve the problem of patient data heterogeneity in remote medical systems. While traditional FL transmits model parameters, our approach transmits the model output logit, thus greatly reducing the communication cost.

  • We propose a reputation incentive based consortium blockchain to improve the quality of data in FL. Blockchain provides a decentralised global model convergence environment, and store the global model of the task publishers and the reputation values of the data providers.

  • We experiment with FedRMD on COVID-19 Radiography and COVID-Chestxray datasets. Extensive experiments show that our algorithm achieves higher accuracy compared with the FedAvg, FedDF, FedGen. And our method can resist the influence of poisoning attack in FL model training.

The rest of the paper as follows: Section 2 details the current state of research federated distillation learning and blockchain-based data sharing. Section 3 introduces problem definition. Section 4 elaborates our proposed FedRMD algorithm, and the reputation incentive of blockchain. Section 5 presents the experimental setting of federated distillation learning and blockchain. Section 6 analysis the results of our experiments. Section 7 summarises our research and future work.

2. Related work

2.1. Knowledge distillation in federated learning

Knowledge distillation in FL has recently emerged as an effective method for addressing user heterogeneity. We have conducted extensive research on the current area of combining FL with distillation learning. Li and Wang (Citation2019) proposed FedMD using transfer learning and knowledge distillation. This framework allows clients to design different network structures based on their computational power. It protects the privacy security of the dataset and the model under the condition of jointly training a model. However, it requires less computation than centralised FL. Chang et al. (Citation2019) in Cronus, each client uses the local and soft-label public datasets jointly for local training. However, the model he used is large, and the high-dimensional parameter vectors are vulnerable to privacy and security attacks. Li et al. (Citation2019) designed a personalised distillation environment for each client rather than solving a global model. Lin et al. (Citation2020) proposed FedDF training models. This knowledge distillation method combined with FL reduces privacy risks and cost and allows flexible aggregation of heterogeneous client models without many restrictions on the model's university and data structure. Zhu et al. (Citation2021) proposed a data-free knowledge distillation method to address heterogeneous FL, called FedGen, in which a lightweight generator was learned to integrate user information in a data-free manner. The server uses local training to coordinate and broadcast to users. Cheng et al. (Citation2021) proposed a framework that can prevent negative and malicious knowledge transfer. This framework further improves the robustness of FL against poisoning attacks. It reduces the cost of communication between the server and the client, but this framework has some limitations, such as the dependency on marking as public datasets. Gong et al. (Citation2021) proposed a single-issue FL framework, called FedAD. They theoretically preserve the efficiency of using only untagged and domain name public data and efficiently available network bandwidth resources, demonstrating the applicability of FedAD to real-world cross-institutional learning through medical image data. Nguyen et al. (Citation2022) proposed a distributed FL approach that uses knowledge distillation to ensure data privacy and protection, with each node operating independently and without access to external data. Xing et al. (Citation2022) proposed a multitask time series classification system for multitasking using the FD. The scheme matches each user in the system with a partner. In order to their weights are the closest of all uploaded weights. However, the framework is still composed of a central server and multiple mobile users, which is vulnerable to the failure of the central server and can cause the system to crash. Han et al. (Citation2022) proposed an unsupervised FL framework that does not require clients to share data features. He used knowledge distillation methods to learn bi-directionally from dispersed and heterogeneous data. They also proved that the framework could enhance the current unsupervised FL algorithm. He et al. (Citation2022) proposed a selective self-distillation method for FL. It imposes adaptive constraints on local updates by self-distilling knowledge of the global model and selectively weighting it by evaluating class and sample level confidence.

2.2. Blockchain-based data sharing

Nguyen et al. (Citation2019) proposed a medical data-sharing solution using mobile cloud computing and blockchain. The data is collected by interacting with the medical sharing system through mobile devices, storing the patient's data address in the blockchain and the data set in the local server. They also implement an access control mechanism based on medical personnel's public key and data address to achieve decentralised data storage and sharing. Lu et al. (Citation2019) employ blockchain to form a data-sharing platform that benefits data privacy and security. On this basis, Lu et al. (Citation2020) utilise a shard-based blockchain protocol for preserving FL's convergence when large-scale nodes are involved. Zhou et al. (Citation2020) proposed a decentralised multi-community training framework, which utilises blockchain to maintain a global model within each community. Communication between the community and other newer models follows the full reduction protocol. El Rifai et al. (Citation2020) integrated FL and blockchain in a healthcare environment to use diabetes data sets to predict decision support tools. They proposed a smart contract to achieve transparency and immutability in information sharing. Tan et al. (Citation2021) proposed a blockchain to control patients’ COVID-19 medical records. When the user meets the attribute access policy of the blockchain, the key can be obtained through the cloud to decrypt the information and control malicious users through tracking lists. Kumar et al. (Citation2021) proposed a FL model based on blockchain by invoking smart contracts to upload the weights of local training models to the blockchain network and train a deep learning model on the decentralised network to obtain the latest information of COVID-19 patients. Egala et al. (Citation2021) proposed a blockchain-based electronic health record through smart contracts to provide convenient services. However, the data is stored in an on-chain database, which makes the system unstable and less scalable. Zheng et al. (Citation2022) proposed a mutual authentication and key agreement scheme based on the Internet of Vehicles and blockchain. Storing the vehicle's authentication information on the blockchain reduces the time for cross-domain authentication of the vehicle and uses lightweight computing to protect the user's private information. Yao et al. (Citation2022) proposed a multi-dimension traceable privacy-preserving health code scheme based on blockchain. It prevents health code information from being tampered with and supports the traceability of the virus transmission chain. Qu et al. (Citation2022) designed a quantum electronic medical record system and proposed a blockchain network based on quantum cryptography. The use of quantum authentication systems and the automatic formation of time stamps by linking quantum blocks to controlled activities reduce the amount of storage space required.

Finally, Table  gives a comparison of related work with FedRMD. The results show that our method is designed for COVID-19 and efficient communication and defends against poisoning attacks and single node failures.

Table 1. Comparison of related work with FedRMD.

3. Problem definition

Figure  shows the framework of FL in remote medical system. First, remote medical equipment transmit the encryption medical data to the local hospital server. We assume a set of N hospitals, which stores a local dataset denoted as {Dn}n{1,,N}. The local hospitals n experience loss over Dn is defined as: (1) Ln(θ)=1{Dn}{xn,yn}DnL(θ,xn,yn)(1) where L(θ,xn,yn) is the loss function of the training medical data sample xn and yn be the label. θ be the model parameter in the global server and local hospitals. FL can be assumed to be the centralised data Dc distributed to N hospitals, expressed as {Dc}={Dn}n=1N. Generally speaking, FL enables multiple hospitals to train a global model G without sharing their data, and through sharing of their training parameters θ. The aim of FL can be formulated as the following problem: (2) minθ1Nn=1NLn(θ),Ln(θ)=1{Dn}n=1{Dn}L(θ,xn,yn)(2)

However, in the remote medical system, the distribution of medical data stored in various hospital institutions differs. After executing SGD on each hospital server, the collected parameters are aggregated in the global server. However, the aggregation method cannot solve the problem of patient data heterogeneity. In addition, the quality of parameters uploaded by participants can't be guaranteed in the process of FL training, and it requires a centralised server for global aggregation. Hence, it is easy to crash the whole system due to a single node failure. Therefore, we propose our approach FedRMD combining FD and blockchain to solve the above problems.

Figure 1. Framework of FL in remote medical system.

Figure 1. Framework of FL in remote medical system.

4. Method

4.1. Overview

Our remote medical system comprises block nodes, task publishers for FL, and medical data providers. Task publishers such as hospital doctors and medical institutions release FL tasks according to their needs to find data providers and select appropriate data providers to participate in FL training by invoking smart contracts. Data providers, mainly medical institutions and hospitals, use remote medical equipment to supervise patient health and collect patient medical data. After that, they use local servers to train the FL model and transmit soft labels to the blockchain nodes. The blockchain of peer nodes is mainly responsible for accepting the FL tasks of the task publisher and the soft targets uploaded by the data provider's server. It also provides a robust aggregation environment for aggregating FL models and updating reputation values. Our proposed training process in one communication round is shown in Figure .

Figure 2. Overview of proposed method.

Figure 2. Overview of proposed method.

We provide a reliable data option scheme for FL tasks. Blockchain ensures that the data uploaded by the hospital can be instantly traceable and has greater transparency. We can use smart contracts to control the quality of the uploaded data. It consists of six steps. As shown in Figure , a clear training flow can be represented by the following steps:

Step 1: Publish FL task. Task publishers set reputation thresholds based on their requirements and invoke smart contracts to publish the FL task after consensus certification to produce the genesis block, which includes the initial global model, communication rounds, and reputation threshold.

Step 2: Select data providers. The reputation value of each hospital is stored in the blockchain. If the hospital meets the task publisher's reputation threshold, it will be selected to participate in the FL exercise as a data provider. The selected participants download the global model from the blockchain.

Step 3: Upload soft labels. The selected data provider downloads the global model from the distributed ledger to the local server and uses the local health data to train the model. After local training, the hospital server generates the extracted feature local knowledge, and the local server of the hospital uploads the soft labels to the peer node. New blocks are generated through consensus authentication.

Step 4: Update reputation values. The peer node updates the hospital's reputation value based on the result of each round. The number of FL tasks determines the data provider's reputation weight. After the training, the peer node combines the recent reputation value and reputation weight to form a final reputation value, which will be used as the reputation value for the selected data provider for the next task. After consensus certification, the peer node generates a new block, and the reputation values are added to the blockchain. Therefore, all task publishers can train the model by selecting hospitals with high-quality data through reputation values.

Step 5: Update the global model: Unlike traditional FL, the server for global model aggregation is replaced by blockchain nodes randomly selected by the blockchain system. The soft targets stored in the blockchain are used for global aggregation. After averaging the uploaded soft targets, the new aggregated global model is added to the blockchain through consensus authentication. If communication rounds are not completed, continue to the next round of training.

Step 6: Get optimised global model. The task publisher finally downloads the optimised global model from the consortium blockchain. Our approach ensures that hospitals can safely aggregate in a decentralised environment without contributing local data sets.

4.2. Main construction

To solve the problem of heterogeneous remote medical data and against poison attack, algorithm 1 summarises the proposed FedRMD algorithm.

4.2.1. Knowledge distillation

During each distillation step t, we select a subset Ht based on the reputation values of hospitals in the blockchain. This subset downloaded the global model from the blockchain and training with its own local dataset. Where ξ is the learning rate. Such a training process often takes a huge number of update steps to converge. Parameters of the model on the local hospital after t steps of stochastic gradient descent (SGD) iterations are denoted as follows: (3) Mtn={Mtnξ~Lt(Mt1n)t|τ01ni=1n[θt1iξ~Li(θt1i)]t|τ=0(3) Knowledge distillation is the process of distilling knowledge from a large and well-trained teacher model to a small student model. We use the Kullback-Leibler (Van Erven & Harremos, Citation2014) divergence to ensemble all hospitals’ soft labels. Neural networks typically produce class probabilities by using a softmax function (Wang et al., Citation2018) as follows: (4) Zt(θ)=softmax(θ)=exp(γ(xi)nT)i=1nexp(γ(xi)T)forxDlog(4) Let xi represent as input the logit of i-th health sample. Where γ is the softmax function, which will output the prediction score of local model Zt. To enable the peak probability congruence among hospitals, we require the refined peak probabilities of all hospitals to be a constant value τ. After local knowledge distillation, hospitals send the soft label Zt of the updated local models to the peer node for global aggregation.

4.2.2. Global aggregation

The global model loss function in the peer node. The global knowledge distillation loss encourage to bridge the gap between the task publisher’s soft target Zt and the soft label ZS received from the data providers. The global loss function is expressed as The global distillation loss of the task publisher’s dataset is {Dt} the cross-entropy loss (Le et al., Citation2021) which is: (5) mingθg=i=1Ng(Zt(Dt),ZS(Dt)i)N(5) where gθ represents the parameter of the global model and Zt(Dt) is the soft output of the global model G. ZS(Dt)i is the soft target obtained from hospital i. On the peer node, it trains the global model G with the reference dataset Dt, and the outcomes from local hospital servers extract the generalised knowledge. Afterwards, the peer node broadcasts its knowledge, which as the soft target of the updated global model to data providers’ servers. During the FedRMD training process, the soft label instead of the model parameters can significantly reduce the communication cost of the training process.

4.2.3. Reputation incentive

First, we set that each hospital had a peer node exchanging information, which means that there are N peer nodes in the blockchain system. We assume that i is a new node added to the blockchain system. And the initial reputation value is determined by the reputation values of other nodes, which is (6) Rvalueiinit=nNRvaluenN(6) where n is one of the hospitals selected to participate in FL. We use Beta reputation (Josang & Ismail, Citation2002) for a hospital i is calculated as the recent reputation value, which is as follows: (7) Rvaluentsrect=αα+β(7) where α=positive(n)+1 and β=negative(n)+1 is the accumulated positive and negative reputation values at one point in time. Where nts is the hospital n participates in one of FL tasks published by the system. We also design reputation weight to encourage hospitals to actively participate in FL, which is: (8) Rweightn=tsTs(8) Let ts represent as the number of times n accepts FL task. Where Ts is the sum of FL tasks in the blockchain. So, the hospital n final reputation value is (9) Rvaluenfinal=λRvalueninit+Rvaluenrect×Rweightn(9) where λ is the weight of the node’s initial reputation. It gradually decreases with the progress of with FL training. To reduce the impact of low data quality submitted due to high initial reputation values.

5. Experimental settings

5.1. Datasets and learning model

We validate the efficacy of the FedRMD algorithm with the COVID-19 radiography (Chowdhury et al., Citation2020; Rahman et al., Citation2021) and COVID-Chestxray (Cohen et al., Citation2020) dataset. COVID-19 radiography dataset has 3616 COVID-19 positive cases along with 10,192 normal, 6012 Lung Opacity (Non-COVID lung infection), and 1345 viral pneumonia images. COVID-Chestxray dataset is created by assembling medical images from websites and publications and currently contains 123 frontal view X-rays. To build an open public dataset of chest X-ray and CT images of patients which are positive or suspected of COVID-19 or other viral and bacterial pneumonias. In all experiments, the neural network model is ResNet-18 (He et al., Citation2016).

5.2. Baseline methods

We select the classic FL algorithm and the most advanced FD algorithm for comparison. To demonstrate the high accuracy of our algorithm in the heterogeneous medical data.

  • FedAvg (McMahan et al., Citation2017): This method is a baseline of FL, after each round of the training process, all trained parameters in each participated client will be sent to the central server for aggregation.

  • FedDF (Lin et al., Citation2020): This method consists of two chief components: ensemble and knowledge distillation via out-of-domain data. It uses unlabelled datasets from other domains to achieve model fusion in order to protect the privacy of FL.

  • FedGen (Zhu et al., Citation2021): This method directly regulates local model updating using the extracted knowledge. It is a data-free FD method with flexible parameter sharing.

5.3. Evaluation environment

The hardware configuration is shown in Table . The software configuration is shown in Table .

Table 2. Hardware configuration.

Table 3. Software configuration.

5.3.1. Federated distillation learning details

We assume each computer to be a hospital server that stores the remote medical data. We also have a computer emulated as a global server. During the implementation, we use Dirichlet distribution Dir(α) to simulate the heterogeneous data among patients, where a smaller α shows a higher data heterogeneity.

We set α = 0.2, 0.5 and 1. The local hospital datasets are unbalanced and have a few training samples. Communication rounds T = 200. In each communication round, we randomly select hospitals to take part in the learning process. The local update and distillation, the batch size B = 32. The local update steps and the distillation steps as 15 epochs in each communication round. The step sizes of the local update and distillation are set to 0.02 and 0.002, respectively. The number of global epochs R = 4. And Table  shows the hyperparameter that is used in our experiments.

Table 4. Hyperparameter list.

5.3.2. Consortium blockchain details

The experiment established a reliable transaction service by building blockchain in Hyperledger Fabric 2. 3.1. We use the PBFT as consensus algorithm, running on 10 peer nodes. And the SHA-256 encryption algorithm is used as the secure hashing algorithm. The data size for reputational values ranges from 100KB to 200KB.

5.4. Evaluation metrics

We use accuracy (ACC), Recall, and Precision as the evaluation indicators. Furthermore, we also use the Area Under Curve (AUC) to compare the classification results comprehensively. AUC is an average of true positive rates over all possible values of the false positive rate. To do this, we first obtain True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN). ACC can be calculated as follows: (10) ACC=TP+TNTP+TN+FP+FN(10) Recall can be calculated as follows: (11) Recall=TP TP + FN(11)

Precision can be calculated as follows: (12) Precision=TPTP + FP(12)

We use transactions per second (TPS) as the indicator for performance of systems that handle routine transactions and record-keeping. TPS can be calculated as follows: (13) TPS=Number of transactionsNumber of seconds(13)

At present, there is no international standard for blockchain performance indicators, according to existing Chinese Blockchain Application Security Technology (DB43/T 1842-2020), the test standards are shown in Table .

Table 5. Blockchain evaluation standards.

6. Results and analysis

6.1. Performance evaluation

To simulate realistic medical scenarios, the communication resources of various hospital servers differ. In data fusion with global services, only servers from some local hospitals are involved. Therefore, we set the number of local hospital services is N for each round of local servers taking part in the FedRMD process in each communication round. We use the COVID-Chestxray dataset as patients’ data stored in local hospitals, where the data heterogeneity α = 1.

We use the same colour to represent the number of hospitals participating in the training. It showed the experimental results in Figure . For all values of all N, the accuracy gradually increases as the number of communication rounds increases. When the training process converges, the accuracy of the training varies slightly. Furthermore, the reason is that with the increase in the number of hospitals participating in the training of the model, the more variety of knowledge the task requester's global model acquires, the better the performance of the model and the more accurate the results. However, in the case of low data quality will have a negative impact on the model training of FL. Therefore, we use the reputation mechanism to control the data quality and improve the quality of FL training.

Figure 3. Accuracy of different number of hospitals.

Figure 3. Accuracy of different number of hospitals.

Although test accuracy is similar, the average time cost per round varies widely, as shown in Figure . For example, when 80% of hospitals are involved, the value of communication cost is 16.18s. However, 40% of hospitals are involved, and the communication cost is 8.25s. The more hospitals involved in the training, the greater the medical data that needs to be trained and the more extended time cost. While the amount of time spent training locally in FL does not increase with the number of other participants, the combination with blockchain requires consistent data stored in the distributed ledger. As the amount of data uploaded by the hospital to the blockchain increases, the time taken through the consensus mechanism increases. We hope to reduce the time of consensus in the future.

Figure 4. Time cost of different number of hospitals.

Figure 4. Time cost of different number of hospitals.

We set each hospital server with a different reputation value in the range of [0,1]. In the experiment, we set the reputation threshold to 0.2 and 0.5 to simulate the process of the blockchain dynamically screening hospital servers to participate in training. We used our proposed FedRMD algorithm for 200 rounds to study the effect of local reputation values on model convergence. Figure  shows that the training loss is larger when the reputation incentive is not used. The training loss decreases when the reputation value set to 0.2. The training loss is smaller when the reputation value set to 0.5. Therefore, using reputation inventive can improve the efficiency of FL, providing high-quality data sharing for FL training.

Figure 5. Training loss during poisoning attacks.

Figure 5. Training loss during poisoning attacks.

Caliper can easily interface with multiple blockchain platforms and shield the underlying details, and users only need to be responsible for designing specific test processes to obtain the visual performance test report. We use Caliper 0.5.0 to test the performance of Hyperledger Fabric, and we show the HTML report in Figure . According to the test report, we got the success rate, TPS, average latency, and so on. Compared with the test standard of blockchain, our system has passed the stress test, and the success rate has reached more than 95%. The system runs at a low load without breakdown. And all the test results meet the performance evaluation standards listed in Table .

Figure 6. Performance report of Hyperledger Fabric.

Figure 6. Performance report of Hyperledger Fabric.

As shown in Figure , we test the average latency of 12,000 blocks. Respectively, and the average latency time also increased as the number of blocks increased, reaching an average latency of 519.6 ms when the number of blocks reached 12,000. Figure  shows that the system's TPS reaches a maximum of 281.9 when the number of blocks reaches 5000. But with the number of blocks increases, the system's TPS decreases slightly. The performance degradation is mainly due to system overwork. Key technologies such as sort execution verification and asynchronous block generation reduce the overhead of deterministic execution and block generation. Therefore, we can improve the TPS of the system through two improvements. On the one hand, it improves the performance of the consensus layer. For example, reduce the number of nodes, increase the consensus speed, modify the data consistency method, and broadcast rotation. On the other hand, the Sharding technique can be used, where the main chain only synchronises the block header and validation information, and the content method is completed with many sharding. Many high-frequency trades are placed off-chain, with only the blockchain recording the final transaction information. But the TPS of Hyperledger fabric system can still reach 250. Therefore, it is feasible to storage the reputation values and the results of model training.

Figure 7. Average latency of blocks.

Figure 7. Average latency of blocks.

Figure 8. Transactions per second of blocks.

Figure 8. Transactions per second of blocks.

6.2. Performance comparison

Table  shows that we use the COVID-Chestxray dataset in the case of heterogeneous α = 1 for experiments. Figure  shows that 80% of the local hospital servers are involved in the training process. When the accuracy of the model does not improve significantly, we think the model reaches convergence. From the figure, we can see that our algorithm model converges faster. Other methods can only randomly select data providers to train the model. Ours uses a reputation incentive mechanism to control the quality of data through blockchain, so our method is more accurate. Figure  shows that AUC of our approach and the three baseline approaches. Compared to the traditional FL, the communication cost generated by FD depends only on the model output dimension. In our method, the communication cost of uploading from a local hospital server to block nodes is determined by the size of the soft labels.

Figure 9. Accuracy trends on the COVID-Chestxray dataset.

Figure 9. Accuracy trends on the COVID-Chestxray dataset.

Figure 10. AUC scores on the COVID-Chestxray dataset.

Figure 10. AUC scores on the COVID-Chestxray dataset.

Table 6. Different methods on COVID-Chestxray.

Table  shows that we use COVID-19 radiography dataset in the case of heterogeneous α = 0.2 for experiments. From tables clearly shows that our proposed FedRMD is better than the other three algorithms in medical data is heterogeneous. Obviously, our method is 12% higher than FedAvg. FedGen also gets a good training result. FedGen has 6% higher training accuracy than FedDF. In FedRMD, the local hospital model needs to classify the resulting images, but also needs to participate in other FL tasks hospitals, which is more accurate than others that rely entirely on generators to transmit information. Figure  shows that 40% of the local hospital servers are involved in the training process. Although our accuracy is similar to FedGen sometimes, our final training results were 3% higher than FedGen. Figure  shows that AUC of our approach and the three baseline approaches.

Figure 11. Accuracy trends on the COVID-19 radiography dataset.

Figure 11. Accuracy trends on the COVID-19 radiography dataset.

Figure 12. AUC scores on the COVID-19 radiography dataset.

Figure 12. AUC scores on the COVID-19 radiography dataset.

Table 7. Different methods on COVID-19 radiography.

6.3. Security analysis

In traditional FL, cryptographic encryption technology encrypts the local model parameters and gradient updates so that attackers cannot obtain any information. But, it cannot control the quality of the data source data. So, we propose a reputation incentive based on consortium blockchain to reduce the attack of malicious nodes, and increase the weight of involved in the FL tasks encourage the hospital to actively participate. The transaction is only passed into the block after the consortium blockchain has verified it. So blockchain guarantees that the task publisher will have access to high-quality data. Even if malicious attackers can participate in interactive training at the beginning. After many iterations, their reputation values become low. Their reputation values are below the reputation threshold. Reputation incentive successfully defended against the poisoning attacks.

7. Conclusion and future work

In this paper, we propose a new FL algorithm called FedRMD for data security sharing in the remote medical system. Combining knowledge distillation and blockchain technology solves two key problems in FL. First, the model accuracy decreases due to data heterogeneity, and the communication cost is high. Second, the quality of data uploaded by participants cannot be controlled during FL training, which is affected by poisoning attacks. To comprehensively solve the above problems, we use Hyperledger Fabric to design a reputation-based consensus to help FL task publishers get excellent global models. We use the knowledge distillation method to transfer soft targets rather than model parameters during the model training process, significantly reducing communication overhead. The task publisher uses the reputation mechanism to select the data provider, and constantly adjusts the reputation value of the participants in each round of training, so as to control the uploading of toxic data by malicious nodes and successfully resist the influence of poisoning attacks on the model training. In addition, the combination of FL and blockchain solves the system paralysis caused by the failure of the central server and improves the robustness of the remote medical system. Extensive experiments under the heterogeneous COVID-19 data show that our method performs well compared to the FedAvg, FedDF, and FedGen.

In future work, we will study how to improve the consensus mechanism of blockchain to improve the sharing efficiency, how to dynamically adjust the size of the reputation threshold, improve the flexibility of the reputation mechanism, and enable more medical institutions to share data while improving the accuracy of the model. In addition, at present, we only provide remote medical data sharing in the Hyperledger Fabric, so we will consider how to solve the problem of cross-chain data sharing and provide a more convenient data-sharing environment for the remote medical system.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported in part by the Natural Science Foundation of Hainan Province [621RC508], State Key Laboratory of Information Security [2022-MS-04], the Science Project of Hainan University [KYQD(ZR)-21075].

References

  • Androulaki, E., Barger, A., Bortnikov, V., Cachin, C., Christidis, K., De Caro, A., Enyeart, D., Ferris, C., Laventman, G., Manevich, Y., Muralidharan, S., Murthy, C., Nguyen, B., Sethi, M., Singh, G., Smith, K., Sorniotti, A., Stathakopoulou, C., Vukolić, M., … Yellick, J. (2018, April). Hyperledger fabric: A distributed operating system for permissioned blockchains. In Proceedings of the Thirteenth EuroSys Conference (pp. 1–15). https://doi.org/10.1145/3190508.3190538
  • Bashshur, R. L., Doarn, C. R., Frenk, J. M., Kvedar, J. C., Shannon, G. W., & Woolliscroft, J. O. (2020). Beyond the COVID pandemic, telemedicine, and health care. Telemedicine and e-Health, 26(11), 1310–1313. https://doi.org/10.1089/tmj.2020.0328
  • Chang, H., Shejwalkar, V., Shokri, R., & Houmansadr, A. (2019). Cronus: Robust and heterogeneous collaborative learning with black-box knowledge transfer. arXiv preprint arXiv:1912.11279. https://doi.org/10.48550/arXiv.1912.11279
  • Cheng, S., Wu, J., Xiao, Y., & Liu, Y. (2021). Fedgems: Federated learning of larger server models via selective knowledge fusion. arXiv preprint arXiv:2110.11027. https://doi.org/10.48550/arXiv.2110.11027
  • Chowdhury, M. E., Rahman, T., Khandakar, A., Mazhar, R., Kadir, M. A., Mahbub, Z. B., & Islam, M. T. (2020). Can AI help in screening viral and COVID-19 pneumonia? IEEE Access, 8, 132665–132676. https://doi.org/10.1109/ACCESS.2020.3010287
  • Cohen, J. P., Morrison, P., & Dao, L. (2020). COVID-19 image data collection. arXiv preprint arXiv:2003.11597. https://doi.org/10.48550/arXiv.2003.11597
  • Crosby, M., Pattanayak, P., Verma, S., & Kalyanaraman, V. (2016). Blockchain technology: Beyond bitcoin. Applied Innovation, 2(6–10), 71. https://scet.berkeley.edu/reports/blockchain.
  • Egala, B. S., Pradhan, A. K., Badarla, V., & Mohanty, S. P. (2021). Fortified-chain: A blockchain-based framework for security and privacy-assured internet of medical things with effective access control. IEEE Internet of Things Journal, 8(14), 11717–11731. https://doi.org/10.1109/JIOT.2021.3058946
  • El Rifai, O., Biotteau, M., de Boissezon, X., Megdiche, I., Ravat, F., & Teste, O. (2020, August). Blockchain-based federated learning in medicine. In International Conference on Artificial Intelligence in Medicine (pp. 214–224). Springer. https://doi.org/10.1007/978-3-030-59137-3_20
  • Fung, C., Yoon, C. J., & Beschastnikh, I. (2018). Mitigating sybils in federated learning poisoning. arXiv preprint arXiv:1808.04866. https://doi.org/10.48550/arXiv.1808.04866
  • Gong, X., Sharma, A., Karanam, S., Wu, Z., Chen, T., Doermann, D., & Innanje, A. (2021). Ensemble attention distillation for privacy-preserving federated learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 15076–15086). https://doi.org/10.1109/ICCV48922.2021.01480
  • Han, S., Park, S., Wu, F., Kim, S., Wu, C., Xie, X., & Cha, M. (2022). FedX: Unsupervised federated learning with cross knowledge distillation. In European Conference on Computer Vision (pp. 691–707). Springer. https://doi.org/10.1007/978-3-031-20056-4_40
  • Hard, A., Rao, K., Mathews, R., Ramaswamy, S., Beaufays, F., Augenstein, S., & Ramage, D. (2018). Federated learning for mobile keyboard prediction. arXiv preprint arXiv:1811.03604. https://doi.org/10.48550/arXiv.1811.03604
  • He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 770–778). https://doi.org/10.1109/CVPR.2016.90
  • He, Y., Chen, Y., Yang, X., Yu, H., Huang, Y. H., & Gu, Y. (2022). Learning critically: Selective self-distillation in federated learning on non-iid data. IEEE Transactions on Big Data. https://doi.org/10.1109/TBDATA.2022.3189703
  • Hitaj, B., Ateniese, G., & Perez-Cruz, F. (2017, October). Deep models under the GAN: Information leakage from collaborative deep learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security (pp. 603–618). https://doi.org/10.1145/3133956.3134012
  • Josang, A., & Ismail, R. (2002, June). The beta reputation system. In Proceedings of the 15th Bled Electronic Commerce Conference (Vol. 5, pp. 2502–2511).
  • Kumar, A., Purohit, V., Bharti, V., Singh, R., & Singh, S. K. (2021). Medisecfed: Private and secure medical image classification in the presence of malicious clients. IEEE Transactions on Industrial Informatics, 18(8), 5648–5657.
  • Kumar, R., Khan, A. A., Kumar, J., Golilarz, N. A., Zhang, S., Ting, Y., & Wang, W. (2021). Blockchain-federated-learning and deep learning models for COVID-19 detection using CT imaging. IEEE Sensors Journal, 21(14), 16301–16314. https://doi.org/10.1109/JSEN.2021.3076767
  • Le, H. Q., Shin, J. H., Nguyen, M. N., & Hong, C. S. (2021, September). Distilling knowledge in federated learning. In 2021 22nd Asia-Pacific Network Operations and Management Symposium (APNOMS) (pp. 196–201). IEEE. https://doi.org/10.23919/APNOMS52696.2021.9562670
  • Li, D., & Wang, J. (2019). FedMD: Heterogenous federated learning via model distillation. arXiv preprint arXiv:1910.03581. https://doi.org/10.48550/arXiv.1910.03581
  • Li, T., Sahu, A. K., Zaheer, M., Sanjabi, M., Talwalkar, A., & Smithy, V. (2019, November). Feddane: A federated newton-type method. In 2019 53rd Asilomar Conference on Signals, Systems, and Computers (pp. 1227–1231). IEEE. https://doi.org/10.1109/IEEECONF44664.2019.9049023
  • Lin, T., Kong, L., Stich, S. U., & Jaggi, M. (2020). Ensemble distillation for robust model fusion in federated learning. Advances in Neural Information Processing Systems, 33, 2351–2363. https://doi.org/10.48550/arXiv.2006.07242
  • Lu, Y., Huang, X., Dai, Y., Maharjan, S., & Zhang, Y. (2019). Blockchain and federated learning for privacy-preserved data sharing in industrial IoT. IEEE Transactions on Industrial Informatics, 16(6), 4177–4186. https://doi.org/10.1109/TII.2019.2942190
  • Lu, Y., Huang, X., Zhang, K., Maharjan, S., & Zhang, Y. (2020). Blockchain empowered asynchronous federated learning for secure data sharing in internet of vehicles. IEEE Transactions on Vehicular Technology, 69(4), 4298–4311. https://doi.org/10.1109/TVT.2020.2973651
  • Maria, N. (2022, February). Telehealth: A new frontier in medicine – and security. Kaspersky. https://securelist.com/telehealth-report-2020-2021/105642/
  • McMahan, B., Moore, E., Ramage, D., Hampson, S., & Arcas, B. A. (2017, April). Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics (pp. 1273–1282). https://doi.org/10.48550/arXiv.1602.05629
  • Melis, L., Song, C., De Cristofaro, E., & Shmatikov, V. (2019, May). Exploiting unintended feature leakage in collaborative learning. In 2019 IEEE Symposium on Security and Privacy (SP) (pp. 691–706). IEEE. https://doi.org/10.48550/arXiv.1805.04049
  • Nguyen, D. C., Pathirana, P. N., Ding, M., & Seneviratne, A. (2019). Blockchain for secure EHRs sharing of mobile cloud based e-health systems. IEEE Access, 7, 66792–66806. https://doi.org/10.1109/ACCESS.2019.2917555
  • Nguyen, T. V., Dakka, M. A., Diakiw, S. M., VerMilyea, M. D., Perugini, M., Hall, J. M. M., & Perugini, D. (2022). A novel decentralized federated learning approach to train on globally distributed, poor quality, and protected private medical data. Scientific Reports, 12(1), 1–12. https://doi.org/10.1038/s41598-022-12833-x
  • Oh, S., Park, J., Jeong, E., Kim, H., Bennis, M., & Kim, S. L. (2020). Mix2FLD: Downlink federated learning after uplink federated distillation with two-way mixup. IEEE Communications Letters, 24(10), 2211–2215. https://doi.org/10.1109/LCOMM.2020.3003693
  • Qu, Z., Zhang, Z., & Zheng, M. (2022). A quantum blockchain-enabled framework for secure private electronic medical records in internet of medical things. Information Sciences, 612, 942–958. https://doi.org/10.1016/j.ins.2022.09.028
  • Rahman, A., Hossain, M. S., Alrajeh, N. A., & Alsolami, F. (2020). Adversarial examples – security threats to COVID-19 deep learning systems in medical IoT devices. IEEE Internet of Things Journal, 8(12), 9603–9610. https://doi.org/10.1109/JIOT.2020.3013710
  • Rahman, T., Khandakar, A., Qiblawey, Y., Tahir, A., Kiranyaz, S., Kashem, S. B. A., & Chowdhury, M. E. (2021). Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Computers in Biology and Medicine, 132, Article 104319. https://doi.org/10.1016/j.compbiomed.2021.104319
  • Sattler, F., Korjakow, T., Rischke, R., & Samek, W. (2021). FedAUX: Leveraging unlabeled auxiliary data in federated learning. IEEE Transactions on Neural Networks and Learning Systems. https://doi.org/10.1109/TNNLS.2021.3129371
  • Tan, L., Yu, K., Shi, N., Yang, C., Wei, W., & Lu, H. (2021). Towards secure and privacy-preserving data sharing for COVID-19 medical records: A blockchain-empowered approach. IEEE Transactions on Network Science and Engineering, 9(1), 271–281. https://doi.org/10.1109/TNSE.2021.3101842
  • Van Erven, T., & Harremos, P. (2014). Rényi divergence and Kullback-Leibler divergence. IEEE Transactions on Information Theory, 60(7), 3797–3820. https://doi.org/10.1109/TIT.2014.2320500
  • Wang, C., Yang, G., Papanastasiou, G., Zhang, H., Rodrigues, J. J., & de Albuquerque, V. H. C. (2020). Industrial cyber-physical systems-based cloud IoT edge for federated heterogeneous distillation. IEEE Transactions on Industrial Informatics, 17(8), 5511–5521. https://doi.org/10.1109/TII.2020.3007407
  • Wang, H., Wang, Y., Zhou, Z., Ji, X., Gong, D., Zhou, J., & Liu, W. (2018). Cosface: Large margin cosine loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 5265–5274). https://doi.org/10.1109/CVPR.2018.00552
  • Xing, H., Xiao, Z., Qu, R., Zhu, Z., & Zhao, B. (2022). An efficient federated distillation learning system for multitask time series classification. IEEE Transactions on Instrumentation and Measurement, 71, 1–12. https://doi.org/10.1109/TIM.2022.3201203
  • Yang, Z., & Sun, Q. (2022). A dynamic global backbone updating for communication-efficient personalised federated learning. Connection Science, 34(1), 2240–2264. https://doi.org/10.1080/09540091.2022.2114428
  • Yao, S., Jing, P., Li, P., & Chen, J. (2022). A multi-dimension traceable privacy-preserving prevention and control scheme of the COVID-19 epidemic based on blockchain. Connection Science, 34(1), 1654–1677. https://doi.org/10.1080/09540091.2022.2077912
  • Zheng, J., Wang, X., Yang, Q., Xiao, W., Sun, Y., & Liang, W. (2022). A blockchain-based lightweight authentication and key agreement scheme for internet of vehicles. Connection Science, 1–24. https://doi.org/10.1080/09540091.2022.2032602
  • Zheng, Z., Zhou, Y., Sun, Y., Wang, Z., Liu, B., & Li, K. (2022). Applications of federated learning in smart cities: Recent advances, taxonomy, and open challenges. Connection Science, 34(1), 1–28. https://doi.org/10.1080/09540091.2021.1936455
  • Zhou, S., Huang, H., Chen, W., Zhou, P., Zheng, Z., & Guo, S. (2020). Pirate: A blockchain-based secure framework of distributed machine learning in 5G networks. IEEE Network, 34(6), 84–91. https://doi.org/10.1109/MNET.001.1900658
  • Zhu, Z., Hong, J., & Zhou, J. (2021, July). Data-free knowledge distillation for heterogeneous federated learning. In International Conference on Machine Learning (pp. 12878–12889). https://doi.org/10.48550/arXiv.2105.10056