Next Article in Journal
Designing for the Metaverse: A Multidisciplinary Laboratory in the Industrial Design Program
Previous Article in Journal
A Simple Model of Knowledge Scaffolding Applied to Wikipedia Growth
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RingFFL: A Ring-Architecture-Based Fair Federated Learning Framework

1
School of Computer Science (National Pilot Software Engineering School), University of Posts and Telecommunication, Beijing 100876, China
2
Zhongguancun Laboratory, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Future Internet 2023, 15(2), 68; https://doi.org/10.3390/fi15020068
Submission received: 10 January 2023 / Revised: 1 February 2023 / Accepted: 6 February 2023 / Published: 9 February 2023

Abstract

:
In the ring-architecture-based federated learning framework, security and fairness are severely compromised when dishonest clients abort the training process after obtaining useful information. To solve the problem, we propose a Ring- architecture-based Fair Federated Learning framework called RingFFL, in which we design a penalty mechanism for FL. Before the training starts in each round, all clients that will participate in the training pay deposits in a set order and record the transactions on the blockchain to ensure that they are not tampered with. Subsequently, the clients perform the FL training process, and the correctness of the models transmitted by the clients is guaranteed by the HASH algorithm during the training process. When all clients perform honestly, each client can obtain the final model, and the number of digital currencies in each client’s wallet is kept constant; otherwise, the deposits of clients who leave halfway will be compensated to the clients who perform honestly during the training process. In this way, through the penalty mechanism, all clients either obtain the final model or are compensated, thus ensuring the fairness of federated learning. The security analysis and experimental results show that RingFFL not only guarantees the accuracy and security of the federated learning model but also guarantees the fairness.

1. Introduction

With the development of IoT and the widespread adoption of mobile devices, artificial intelligence (AI) technology is widely used in many aspects of daily life [1,2]. Considering security issues, federated learning (FL) comes into being, which enables collaborative training by exchanging local model parameters instead of raw data among devices [3,4]. The training data are kept locally during the training process, thus effectively protecting users’ privacy [5,6]. This advantage in turn motivates more users to participate in FL, which leads to more accurate global models [7,8].
Traditional FL typically uses a central server to coordinate the collaboration of clients, which is called star architecture [9,10]. Specifically, in each training round, the server updates the global model by aggregating the local model parameters generated by the clients from their respective local models. The server broadcasts the updated global model to the clients. As a result, most of the clients participating in the training will make simultaneous transmissions in a short period. However, the explosive growth in traffic can have a severe impact on the network, thus limiting the number of clients that can participate in FL [11,12]. At the same time, the trustworthiness of the clients and vulnerability to single-point attacks pose hidden risks to the security and reliability of FL [13,14].
Some scholars start by changing the traditional star architecture of FL to solve the above problems and propose distributed FL architectures [15,16]. At present, distributed FL architectures are mainly divided into two categories: fully connected architectures [17,18,19] and ring architectures [20,21,22,23]. Under the fully connected architecture, clients form a P2P network, in which each client transmits local model parameters to neighboring clients, receives local model parameters transmitted from neighbors to complete model aggregation, and continues to transmit until the models finally converge. However, this approach brings a large amount of redundant communication, and it is difficult to guarantee the convergence of the model because each client performs aggregation locally [24]. In the ring architecture, clients are connected in a ring communication network in a certain order, and during the training process, each client only needs to transmit the local model parameters to the next client. The next client will either continue training based on the model received or transmit to the next client after aggregating it with its local model. Through multiple rounds of loop transmission, the converged global model is finally obtained. The ring structure achieves better results due to its more efficient communication efficiency and better convergence.
In the above scenario, it is usually assumed that all clients are honest, that is, all clients participate in the training process in its entirety [25]. In fact, a client may maliciously abandon the subsequent training after obtaining a result in their favor or leave midway unintentionally for technical reasons such as unstable network conditions, and the impact of such dishonest behavior (called escape attack) on the ring structure is catastrophic [26]. Firstly, the model training may be unsustainable due to a break in the ring communication at a certain point, which will lead to low accuracy of FL. Secondly, the intermediate result obtained by the dishonest client contains data contributions from honest clients, but honest clients did not obtain any benefit, which is unfair to them. Therefore, the escape attack needs to be avoided as much as possible to ensure the fairness of all clients participating in the model training.
Motivated by the above analysis, we propose a Ring-architecture-based Fair Federated Learning framework (RingFFL), in which we design a penalty mechanism to guarantee the fairness of FL, that is, the client either obtains the model that is transmitted correctly or is compensated with a certain amount of digital currency. The HASH comparison guarantees the correctness of the models transmitted during the training process. Meanwhile, blockchain is introduced in the training process to guarantee the immutability of the clients’ deposits and compensation operations. Finally, the security analysis and experimental evaluation are carried out. The main contributions of this paper are as follows:
  • To ensure the fairness of FL training, we propose a penalty mechanism where clients need to pay deposits in a given order before participating in the training. If all clients perform honestly, they will all obtain the final global model, otherwise, deposits of dishonest clients will be compensated to the honest clients. In other words, clients either obtain the final global model or are compensated, thus ensuring fairness.
  • To enhance the security of FL training, we propose the HASH comparison mechanism and introduce the blockchain. Through HASH comparison, the correctness of the models transmitted by the clients during the training process is guaranteed. With the immutability of the blockchain, the security of the client’s deposit payment and compensation operations is guaranteed.
  • To evaluate the performance of our mechanism, we perform a security analysis of the proposed mechanism and conduct experiments using MNIST and CIFAR10 data sets. The security analysis and experimental results show that our mechanism ensures security and fairness while maintaining high accuracy.
The rest of the paper is organized as follows. In Section 2, the related work is reviewed. In Section 3, the scenario, threat and adversary model, as well as the work flow of RingFFL are introduced. In Section 4, the proposed penalty mechanism is specifically described. In Section 5, the security analysis and numerical results are presented. Section 6 concludes the paper with a summary and points out the future directions.

2. Related Works

In this section, we briefly describe the challenges of FL first. For the challenges of poor scalability and vulnerability to single-point attacks of star architectures in FL, we provide a detailed description of the distributed FL framework based on fully connected architecture and ring architecture.

2.1. Challenges of FL

Federated learning, a pioneering technique to protect data privacy, which enables training machine learning models on distributed data sets, has been widely used in various fields such as IoT, IoV, and Data Access [27]. However, FL still faces many challenges, such as expensive communication costs, systems heterogeneity, statistical heterogeneity, and privacy concerns [28]. To address the problem of expensive communication costs, Paragliola et al. proposed a new learning strategy that reduced the total number of parameters shared in the FL process and evaluated the trade-off between the requirement to reduce communication costs and the need to guarantee the highest classification performance [29]. To address the problem of system heterogeneity, Lin et al. proposed an ensemble refinement scheme for model aggregation, where a central classifier was trained from unlabeled data of the clients’ model output, a technique that allowed flexible aggregation of heterogeneous client models [30]. To address the statistical heterogeneity problem, Liu et al. proposed a client–edge–cloud hierarchical FL system supported by the HierFAVG algorithm for better communication–computation tradeoffs [31]. To address the problem of data heterogeneity, Paragliola et al. investigated and evaluated the performance and behavior of a federated model in the presence of catastrophic forgetting events in the context of nonstationary data [32]. Huang et al. illustrated how data skewing could affect the performance of FL algorithms and then proposed a new algorithm, FedMix, which adapted existing FL algorithms and achieved better performance [33]. For security issues, Li et al. proposed q-Fair Federated Learning (q-FFL) for fairer accuracy allocation among devices, which outperformed the existing baseline in terms of fairness, flexibility, and efficiency [34]. Fang et al. proposed a multiparty privacy-preserving machine learning framework (PFMLP) based on partial homomorphic encryption and FL to achieve protection of data and model parameters during training [35].

2.2. Distributed FL Framework

The distributed FL framework based on fully connected architecture completes model training through P2P model transmission between clients, and the implementation of the architecture often requires the support of blockchain, Device-to-Device (D2D), and so on [17,18,19]. Samarakoon et al. proposed a novel method based on distributed FL to estimate the tail distribution of vehicle formation lengths, approaching the accuracy of the centralized solution while reducing the amount of data exchanged by 79% [17]. Li et al. proposed a decentralized blockchain-based FL framework called the Blockchain-based Federated Learning framework with Committee consensus (BFLC), which avoided attacks by malicious central servers on global models and users’ private data, and designed an innovative committee consensus mechanism for BFLC that could effectively reduce consensus computation and malicious attacks [18]. Qu et al. proposed a novel blockchain-based FL scheme called FL-Block that balanced the privacy protection of fog computing with the attendant inefficiencies, excelling in privacy protection, efficiency, and resistance to poisoning attacks [36]. Xing et al. proposed distributed stochastic gradient descent (DSGD) to achieve large-scale deployment of FL in wireless communication scenarios [37]. Zhang et al. proposed an FL scheme using D2D communication (D2D-FedAvg), where D2D grouping, primary UE selection, and D2D exit were implemented in the learning process, resulting in a complete D2D-assisted FL averaging algorithm [19]. Although the problem of poor scalability and vulnerability to single-point attacks of star architectures is solved to some extent by using fully connected architecture, it brings a lot of redundant communication consumption, and it is difficult to guarantee the convergence of the model due to the lack of global aggregation.
Some scholars have designed a ring architecture to complete the training through the ring transmission of models while ensuring lower communication redundancy and higher convergence [20,21,22,23]. Xu et al. proposed an optimization method to minimize FL uplink transmission time in wireless networks and designed a ring all-reduce architecture based on D2D, which significantly reduced the transmission time [20]. Wang et al. proposed a ring-topology-based distributed FL (RDFL) scheme for deep generative modeling (DGM), which provided communication efficiency and maintained training performance to boost DGMs in target tasks compared with existing FL works [21]. Lee et al. designed a cyclic learning scheme based on the ring topology to reduce the number of FL training iterations and improve the performance without adding any additional computational cost [22]. Lee et al. proposed a new algorithm called TornadoAggregate to improve FL accuracy and scalability by promoting a ring architecture. Experimental results show that TornadoAggregate improved test accuracy by 26.7% and achieved near-linear scalability [23].
The above schemes usually assume that the clients are honest. However, the assumption is difficult to be satisfied in realistic situations. A malicious client sending escape attacks during the training process can be fatal to FL based on the ring architecture. Therefore, we propose a fair FL framework to solve the problem. With the proposed penalties mechanism, clients are effectively prevented from escaping from the training after obtaining valid information.

3. Framework Overview

In this section, we focus on the scenario, threat, and adversary model and workflow of the proposed RingFFL.

3.1. The Scenario of RingFFL

We consider the scenario where a ring architecture is formed among different clients to perform FL and no server exists. Without loss of generality, we assume that there are N clients. All clients involved in FL first agree on the order of the ring architecture. As shown in Figure 1, the black dashed line indicates the communication link between clients. The whole training process is divided into two phases, in which one is the deposit payment phase (red dashed line) and the other is the model transmission phase (blue solid line). In the deposit payment phase, all clients will prepay the deposits according to the proposed mechanism. In the model transmission phase, the local model parameters of the previous client are sent to the next client. After all clients obtain the global model for one round of FL, they will continue training according to the same rules until a predefined number of rounds or a loss function convergence threshold is reached.

3.2. Threat and Adversary Model

Threat Model: In our mechanism, there are N clients cooperating to train the FL model. We assume that all clients train local models honestly, and a subset of them are malicious or are potentially corrupted by a malicious adversary when transmitting local model parameters.
Adversary Goal: The adversary has two goals. One is to manipulate clients to send false local model parameters to mislead the global model of FL. The other is to abort during training to obtain useful intermediate information without effort, which can lead to some clients’ contributing but no benefit, and some clients cannot obtain the FL model. Both threats will not only reduce the accuracy of FL but also destroy the fairness.

3.3. The Work Flow of RingFFL

The work flow of RingFFL is shown in Figure 2. Suppose there are N distributed clients participating in the FL. In order to ensure the fairness, RingFFL is designed based on the blockchain. It is divided into four steps.
Step 1 
Initialization
Each client generates a random number to form a private key, computes the corresponding public key, and produces a digital currency wallet address. Apart from this, they perform data preprocessing on their own data sets.
Step 2 
Local  Training
Each client trains a local model based on its private data set. Since all clients do not want to reveal their sensitive information, we use FL to protect the clients’ raw data. The local model parameters sent by the clients during the FL process can be protected using many methods, such as differential privacy [13]. It is not the focus of our approach and is not described in detail.
Step 3 
Transactions on the blockchain
Transactions on the blockchain phase are divided into three parts: roof deposits, ladder deposits, and acknowledgment. We introduce them below.
(3.1) 
Roof deposits
According to the topology of the ring structure, all clients agree on a sequence. Except for the last client, all the others make the same amount of deposits to it and broadcast all transactions to the network.
(3.2) 
Ladder deposits
Each client pays some deposits according to the predetermined rule to its previous client. All transactions will be broadcasted to the network.
(3.3) 
Acknowledgement
All clients will provide their local model parameters to acknowledge digital currencies deposited in (3.1) and (3.2). Successful transactions will be recorded on the blockchain.
Step 4 
Output
All local model parameters will be aggregated to generate the global model in a round and shared among all clients. The process will be repeated until a predetermined number of training rounds or a convergence threshold is reached. If all clients behave honestly, each client’s digital currencies will neither increase nor decrease in the end. Otherwise, dishonest clients’ deposits will be used to compensate for every honest client that does not obtain the final model of the FL.

4. Penalty Mechanism in RingFFL

In FL, each distributed client trains a local model according to its data set locally. Then, all the local model parameters can be safely aggregated into a global model to be used for more accurate prediction. However, each client may be corrupted and does not share their local model with other clients. In this case, they can synthesize a more accurate model to improve their prediction accuracy, while other clients cannot, although they share their local models.
To deal with this, we put forward RingFFL to ensure that if a dishonest client refuses to send their local model to honest clients after receiving others’ local models, the honest ones who do not obtain the results will obtain a certain amount of digital currency compensation provided by the dishonest ones. The proposed mechanism is applicable to FL with ring architecture, and clients that can form the ring topology can participate in FL, such as mobile devices, IoT terminals, and self-driving vehicles. In the paper, we deal only with the static adversaries and assume that all clients will honestly train their local models. The main notations used are shown in Table 1.
To gain a better understanding of the mechanism, motivated by [38,39], we introduce the process of digital currency changing on the blockchain. Firstly, sender S of the deposits broadcasts a transaction to the network. The transaction involves the number of deposits c o i n s b , a boolean circuit C · , and the time limit t. Secondly, the recipient R of the deposits provides evidence e in time t. Finally, if e satisfies C e = 1 , the transaction will be recorded on the blockchain, and c o i n s b will be sent to R’s wallet. If R cannot provide e or C e = 0 in time t, then the transaction will not be recorded, and c o i n s b will be returned to S’s wallet. The penalty mechanism is detailed below.
Step 1 
Initialization
At first, each client uses the SHA256 algorithm [40] to generate a 256-bit random number as the private key, which is used to generate the corresponding public key and used for digital signatures in the transactions. Next, they calculate the corresponding public key according to the Elliptic Curve Cryptography algorithm [41]. Once more, based on the RIPEMD160 [42] and SHA256 algorithms [40], each client generates a digital currency wallet address. Each client must use their private key for signature authentication to transmit the digital currencies in their wallet. At last, each client performs data preprocessing on their own data set, which mainly includes data cleaning and data integration.
Step 2 
Local training
The data sets of the clients P i are T i i = 1 , 2 , , N , respectively. In the local training phase, all P i i = 1 , 2 , , N construct local model parameters M i i = 1 , 2 , , N according to T i . Then, each client calculates
H i = H a s h M i
and outputs H i i N to P i i = 1 , 2 , , N .
Step 3 
Transactions on the blockchain
In our mechanism, paying deposits is a necessary condition for clients to participate in the FL, even for honest clients. It is believed that the more digital currency clients hold, the more reliable they are. Clients who leave frequently midway will hold less and less digital currency and cannot even meet the deposit amount required to participate in the FL, and thus cannot participate in the training. Transactions on the blockchain phase consist of three parts, which are introduced in detail below.
(3.1) 
Roof deposits
We introduce the roof deposits phase in two cases: normal and abnormal situations. The normal situation is the case that all clients perform honestly as required, and the abnormal situation is the case that there are clients who do not perform the corresponding process as required during the training. The situations are presented in detail below.
(a)
Normal situation.
Each client P i i = 1 , , N 1 simultaneously deposits c o i n s b to P N under the condition C N , and the deadline is t N . The detail is that each P i i = 1 , 2 , , N 1 broadcasts a transaction d e p o s i t , T i d , T i d , P i , P N , C N , t N , b to the network, in which T i d is the session identifier and T i d is the transaction identifier. Ignore other transactions with the same T i d sent by P i i = 1 , 2 , , N 1 to P N before the end of RingFFL. The process is denoted as
A c k i , N : P i b , t N C N P N
(b)
Abnormal situation.
If the transaction d e p o s i t , T i d , T i d , P i , P N , C N , t N , b is not recorded on the blockchain for any one of P i i = 1 , , N 1 , all P i i = 1 , , N 1 wait until time t N + 1 to receive A c k i , N ’s reimbursement message r e i m b u r s e , T i d , T i d , P i , P N , C N , t N , c o i n s b to P i ’s wallet, and the RingFFL stops, in which i = 1 , , N 1 .
(3.2) 
Ladder deposits
Like roof deposits, we introduce the ladder deposits phase in two cases: normal and abnormal situations.
(a)
Normal situation.
After the clients, except P N , complete the roof deposits, all clients perform the process of ladder deposits in turn. First, P N broadcasts a transaction ( d e p o s i t , T i d , T i d , P N , P N 1 , C N 1 , t N 1 , N 1 b ) to the network as a deposit to P N 1 under condition C N 1 with deadline t N 1 . Similarly, P N 1 deposits c o i n s N 2 b to P N 2 under condition C N 2 . It continues until P 2 deposits c o i n s b to P 1 under condition C 1 . For i = N 1 down to i = 1 , the process is denoted as
A c k i + 1 , i : P i + 1 i · b , t i C i P i
(b)
Abnormal situation.
If there is no transaction within the network that client P i + 1 sends deposits to P i i = 1 , , N 1 as described above, all P j j i do not perform the ladder deposits and wait for A c k i , N ’s reimbursement according to Equation (2). All P j j > i need to wait until the end of the RingFFL to judge whether their ladder deposits are acknowledged.
(3.3) 
Acknowledgment
Like roof deposits, we introduce the acknowledgment phase in two cases: normal and abnormal situations.
(a)
Normal situation.
In the acknowledgment phase, we define E i + 1 and C i as follows:
E i + 1 = Δ E x p a n d i + 1 , E i ; M i + 1 = E i M i + 1
C i M 1 M i ; H 1 , H 2 , , H i = H a s h M 1 ? ̲ ̲ H 1 H a s h M i ? ̲ ̲ H i
in which H i is associated with Equation (1), M i is the model parameters received from neighboring client, and A B denotes a connection between A and B.
P 1 first provides evidence E 1 = E x p a n d 1 , ; M 1 according to Equation (4) to acknowledge A c k 2 , 1 from P 2 . When P 1 provides the acknowledgment message ( a c k n o w l e d g e , T i d , T i d , P 2 , P 1 , C 1 , t 1 , b , E 1 ) , miners will verify whether the transaction d e p o s i t , T i d , T i d , P 2 , P 1 , C 1 , t 1 , b exists and that C 1 E 1 ; H 1 = 1 is true according to Equation (5). If the verification passes, the message a c k n o w l e d g e , T i d , T i d , P 2 , P 1 , C 1 , t 1 , b , E 1 is made public to the network, and the transaction d e p o s i t , T i d , T i d , P 2 , P 1 , C 1 , t 1 , b is recorded on the blockchain later. The c o i n s b in a c k n o w l e d g e , T i d , T i d , P 2 , P 1 , C 1 , t 1 , c o i n s b is sent to P 1 ’s wallet. Afterwards, P 2 acknowledges A c k 3 , 2 from P 3 , similarly. It continues in turn until P N 1 acknowledges A c k N , N 1 from P N using evidence E N 1 according to Equations (4) and (5). At last, P N uses E N to acknowledge A c k i , N (Equation (2)) from all P i i = 1 , 2 , , N 1 .
(b)
Abnormal situation.
For i = 1 , 2 , , N 1 , if P i does not acknowledge A c k i + 1 , i according to Equation (3), then P i + 1 will end the RingFFL and wait for reimbursement from A c k i + 1 , i . Furthermore, when i + 1 N , P i + 1 waits for reimbursements from A c k i + 1 , N according to Equation (2).
Step 4 
Output
Like roof deposits, we introduce the output phase in two cases: normal and abnormal situations.
(a)
Normal situation.
In the output phase, if P N acknowledges A c k i , N according to Equation (2)), then E N (Equation (4)) is made public, which is the evidence that meets C N (Equation (5)). Each P i i = 1 , , N outputs M 1 M N , uses the FedAvg algorithm [43] for model aggregation to obtain the current round of global model M R o u n d g l o b = 1 N N M i , and continues to the next round of training.
In the RingFFL, for i = 1 , , N 1 , A c k i , N (Equation (2)) locks c o i n s b of P i . If P i 1 successfully executes A c k i , i 1 ’s acknowledgment (Equation (3)), then P i will give P i 1 c o i n s i 1 b and the condition C i 1 (Equation (5)) provided by P i 1 will be made public. According to the model parameters M i of P i and C i 1 , P i can successfully execute A c k i + 1 , i ’s acknowledgment (Equation (3)) and obtain c o i n s i b from P i + 1 . That is, if all clients execute the RingFFL honestly, the number of each client’s digital currencies neither increase nor decrease.
(b)
Abnormal situation.
If P N does not acknowledge any A c k i , N (Equation (2)), then each client outputs ⊥.
For i = 1 , , N , suppose P j j i 1 are honest clients. If P i terminates the RingFFL after obtaining M 1 M 2 M i 1 , then P i 1 can successfully acknowledge c o i n s i 1 b from P i ’s ladder deposits, and P k ’s c o i n s k 1 b in the ladder deposits will be reimbursed, in which k = i + 1 , i + 2 , , N . Since P N cannot provide C N (Equation (5)), each client’s roof deposits will be reimbursed. As a result, the total number of P k ’s digital currencies is unchanged, in which k = i + 1 , i + 2 , , N . The number of each P j ’s digital currencies increases by b, and the number of P i ’s digital currencies decreases by i 1 b . Thus, the RingFFL guarantees the fairness of FL. The overall process of RingFFL is illustrated in Algorithm 1.
To better understand the whole process, we take 5 clients as an example to demonstrate the penalty mechanism in RingFFL, which is shown in Figure 3.
Algorithm 1 The proposed RingFFL
Input: 
T 1 , T 2 , , T N , M i n t , M A X .
Output: 
Federated Learning Model M.
1:
Set R o u n d = 1
2:
for   i = 1 : N   do
3:
    P i trains local models M i based on initial model M i n t , calculates H i , and broadcasts H i i N to P i .
4:
end for
5:
for   i = 1 : N 1   do
6:
    P i deposits c o i n s b to P N under the condition C N , which is denoted as A c k i , N .
7:
   if P i is honest then
8:
     Process A c k i , N is executed.
9:
   else
10:
     Transaction d e p o s i t , T i d , T i d , P i , P N , C N , t N , b is not recorded on the blockchain.
11:
     for i = 1 : N 1 do
12:
         P i waits to receive A c k i , N ’s reimbursement message ( r e i m b u r s e , T i d , T i d , P i , P N , C N , t N , c o i n s b ) .
13:
        End RingFFL
14:
     end for
15:
   end if
16:
end for
17:
for   i = 1 : N 1   do
18:
    P i + 1 deposits c o i n s i · b to P i under the condition C i , which is denoted as A c k i + 1 , i .
19:
   if P i is honest then
20:
     Process A c k i + 1 , i is executed.
21:
   else
22:
     for  j = 1 : i 1 do
23:
         P j does not perform the ladder deposits and wait for A c k i , N ’s reimbursement.
24:
     end for
25:
     for j = i + 1 : N 1 do
26:
         P j wait for the end of RingFFL to judge whether its ladder deposit is acknowledged.
27:
     end for
28:
   end if
29:
end for
30:
for   i = 1 : N   do
31:
   if P i is honest then
32:
      P i outputs M R o u n d g l o b = 1 N N M i
33:
   else
34:
      P i outputs ⊥
35:
     End RingFFL
36:
   end if
37:
end for
38:
if Round<MAX then
39:
    R o u n d = R o u n d + 1
40:
   Return 2
41:
else
42:
   for i = 1 : N do
43:
      P i outputs M = M R o u n d g l o b
44:
     End RingFFL
45:
   end for
46:
end if

5. Security Analysis and Numerical Results

In the section, we focus on the security analysis and numerical results of the proposed RingFFL.

5.1. Security Analysis

In this paper, we use blockchain in the ring-architecture-based FL to address the aforementioned security threats.
(1)
Guaranteeing the data security of the clients: We use FL with ring architecture to train the model, and the local model parameters instead of the raw data are transmitted during the training process. The raw data are stored locally from start to finish, thus ensuring the security of the clients’ data.
(2)
Guaranteeing the security of the clients’ deposits: In the process of deposit payment and deposit refund, we use blockchain to guarantee the security of the transactions. The security mechanism based on consensus can avoid security attacks such as double spending, thus ensuring the security of clients’ deposits during the training process.
(3)
Guaranteeing the correctness of the transmitted local model parameters: To prevent the dishonest clients from transmitting false local model parameters, the acknowledgment process needs to validate the hash values of the model parameters, and only the correct ones will pass the validation. The client will then receive the corresponding amount of digital currencies.
(4)
Guaranteeing the fairness of the FL: To avoid dishonest clients from aborting during the FL process, the penalty mechanism can deduct the digital currency deposits of the aborting clients to compensate clients with additional information loss. In this way, the fairness of FL is guaranteed.

5.2. Numerical Results

We perform experiments using two well-known data sets MNIST [44] and CIFAR10 [45] to verify the effect of RingFFL, which are widely used in FL. MNIST is a handwritten digit data set, containing 60,000 training samples and 10,000 test samples, in which each image is a single digit from 0 to 9. The CIFAR10 data set consists of 60,000 color images of 10 classes, which contains 50,000 training images and 10,000 test images and each image is 32 × 32 in size. According to the distribution of data among clients, we conducted experiments for the cases where the data sets of clients are independent and identically distributed (IID) [46] and nonindependent and identically distributed (non-IID) [47], respectively. Whether clients behave honestly or dishonestly in FL is independent of the distribution of their data sets. When simulating the case of IID, we uniformly sample the data with labels 0–9 as the data set for each client, and the distribution of clients’ training data obeys a uniform distribution. In the case of simulating the situation of non-IID, we first preprocess the data by dividing the original data into 20 parts, in which each containing half of all the data in a particular category. Then, we put every 2 parts together as a client’s data set according to i , j categories, where i j , which is a common method to simulate non-IID in FL.
To demonstrate the accuracy of RingFFL for MNIST and CIFAR10 data sets in IID and non-IID cases, respectively, Figure 4 and Figure 5 are given.
Figure 4 and Figure 5 compare the accuracy with the communication round for the centralized algorithm (Central Data in the figures) and our approach in the case of IID and non-IID experimented with MNIST and CIFAR10 data set, respectively. It can be seen that the accuracy of RingFFL gradually increases as the number of communication rounds increase. After reaching a certain number of communication rounds, the accuracy tends to be stable and close to the accuracy of the centralized algorithm. Since CIFAR10’s images are derived from the real world, the scale, features, and colors of objects are different in addition to having a lot of noise, which greatly increases the difficulty of recognition. Nevertheless, the accuracy of the experiments conducted by CIFAR10 is close to the accuracy of the centralized algorithm both in the IID and non-IID cases, which meets the requirements of FL.
To demonstrate the effect of escape attacks on the FL, Figure 6 and Figure 7 are given. In our experiments, we assume that there are clients left in the 5th communication round of FL.
Figure 6 and Figure 7 compare the accuracy of the experiments using the MNIST and CIFAR10 data sets in normal and abnormal (with clients leaving) situations in IID and non-IID cases, respectively. It can be seen that the accuracy of the experiments using the MNIST and CIFAR10 data sets decreases substantially when there are clients leaving during the FL process, both in the IID and non-IID cases. For example, in the IID case of the experiment using the MNIST data set, the accuracy of FL in the normal situation is 96.29%, while the accuracy of FL in the abnormal situation with clients leaving is 81.31%. Similarly, in the non-IID case of the experiment using the CIFAR10 data set, the accuracy of FL in the normal situation is 85.13%, while the accuracy of FL in the abnormal situation with clients leaving is 45.56%.
To demonstrate the fairness of the RingFFL, we performed simulation experiments. Suppose there are 5 clients participating in the FL. The situation where all clients are honest is shown in Figure 8a. The situations where a client leaves during the roof deposit and ladder deposit are shown in Figure 8b,c, respectively. The situation where a client leaves after obtaining useful information is shown in Figure 8d.
As shown in Figure 8a, when all clients behave honestly, the number of digital currencies in each client’s wallet neither increases nor decreases at the end of FL. Client 1’s reimbursement comes from client 2’s ladder deposits of b, and similarly, client 2, client 3, and client 4’s reimbursement come from client 3, client 4, and client 5’s ladder deposits of 2 b , 3 b and 4 b , respectively. Client 5’s reimbursement comes from client 1–client 4’s roof deposits of b × 4 = 4 b .
The situation where a client aborts during the roof deposit is shown in Figure 8b. Suppose that client 3 leaves. After client 3 aborts during the roof deposit, RingFFL stops and does not continue the following process. Since no local model parameters are transmitted, client 1, client 2, and client 4’s roof deposits of b will be fully refunded.
The situation where a client aborts during the ladder deposit is shown in Figure 8c. Suppose that client 3 leaves. As can be seen from Figure 8c, after client 3 aborts during the ladder deposit, client 1 and client 2 will also not start the ladder deposit phase. Because no local model parameters are transmitted, ladder deposits for client 4 of 3 b and client 5 of 4 b and roof deposits for client 1–client 4 of b will be refunded.
The situation where a client aborts during the ladder deposit is shown in Figure 8d. Suppose that client 3 leaves after obtaining useful information. It can be seen from Figure 8d that, in the case where client 3 aborts after obtaining useful information, since client 3 does not transmit the model parameters to client 4, they do not obtain the number of ladder deposits of 3 b from client 4, which will be returned to client 4. Similarly, the number of ladder deposits of 4 b from client 5 to client 4 will also be returned to client 5. Client 1–client 4’s roof deposits to client 5 will also be returned in full. However, client 1 correctly transmitted the model parameters to client 2, so client 1 received both client 2’s ladder deposits of b and client 5’s returned roof deposits of b. That is, the number of digital currencies of client 1 eventually increases by b. Similarly, client 2 correctly transmits the model parameters to client 3, so client 2 obtains both the ladder deposits of 2 b from client 3 and the roof deposits of b refunded from client 5. After subtracting the number of ladder deposits of b given by client 2 to client 1, the number of digital currencies in client 2’s wallet also increases by b in the end. In addition, as can be seen in Figure 8d, client 3’s digital currencies are reduced by 2 b , while the digital currencies of client 1 and client 2, who trained honestly and transmitted the local model parameters to the next client, are both increased by b. This corresponds to client 3’s deposits being used to compensate for the losses of client 1 and client 3. Client 4 and client 5 receive neither the model parameters nor the additional digital currencies. Thus, the fairness of FL is achieved.
To illustrate the universality of the proposed mechanism to the number of clients, we increased the number of clients and reconduct the experiments. As is shown in Figure 9, Figure 10 and Figure 11, similar results are obtained with 10 clients, 15 clients, and 20 clients participating in FL training, respectively, in which we assume that client 8 leaves halfway.
Figure 8, Figure 9, Figure 10 and Figure 11 show the case where one client leaves halfway. When more than one client leaves, the penalty mechanism still works. We conducted a simulation experiment with three clients leaving midway (client 8, client 14, and client 18), and the experimental results are shown in Figure 12.
It can be seen from Figure 12a that when three clients do not execute roof deposits, all roof deposits of other clients will be refunded. As can be seen from Figure 12b, since client 18 should make ladder deposits first among the three leaving clients, when client 18 discontinues ladder deposits, all the remaining clients will no longer perform ladder deposits and wait for the refund of their respective roof deposits. From Figure 12c, it can be seen that since client 8 performs the acknowledgment step first among the three leaving clients, after client 8 leaves with useful information, their ladder deposits will be used to compensate for the loss of client 1–client 7. Although client 14 and client 18 also leave midway, they do not obtain additional useful information, so all their deposits are returned. The fairness of FL is still guaranteed when multiple clients leave in the middle of the process.
To demonstrate the effect of including blockchain on training time, we give Figure 13 and Figure 14.
It can be seen from Figure 13 and Figure 14 that including blockchain will take a little longer than excluding blockchain experiments with the MNIST and CIFAR10 data sets in both the IID and non-IID cases. If blockchain is not included, the time consumption mainly consists of the models’ local training time, model transmission, model validation, and model aggregation time. In the case of including blockchain, the transaction confirmation time for paying deposits and deposit acknowledgment is increased, in which the confirmation operation of paying deposits and the models’ local training can be executed in parallel. So, the increased time consumption mainly lies in the extra time for the confirmation operation of paying deposits over local training and the confirmation time of deposit acknowledgment. In the experiment, we use Ethereum 2.0, which has a transaction confirmation time of 12–14 s. If a blockchain with a shorter consensus time such as Fabric is used, it can reduce the time consumption even more.

6. Conclusions

In this paper, we propose a ring-architecture-based fair federated learning framework RingFFL. The penalty mechanism is designed to ensure the fairness of FL in the case that the clients launch an escape attack during the FL process. That is, either all clients obtain the FL model, or clients who have suffered losses will be compensated with the digital currencies from the clients that launched escape attacks. Specifically, each client trains a local model based on its data and then transmits the local model parameters to other clients according to the rules designed in this paper. When dishonest clients obtain other clients’ model parameters and abort, the honest clients who do not obtain the models will obtain a certain amount of digital currencies provided by dishonest clients using blockchain as a compensation, thus achieving the fairness of FL.
In the future, we plan to study the fairness of FL in the presence of more attacks. For example, when considering a poisoning attack in which the corrupted client manipulates the distribution of training data by inserting carefully crafted samples in the training set in order to change the model behavior and degrade the model performance; how to achieve fairness while guaranteeing the performance of FL is a question worth studying. In addition, taking asynchronous FL into account and investigating the case where there are clients who can join the training at any time may also be an interesting future direction.

Author Contributions

Methodology, L.H., X.H. and D.L.; supervision, Y.Z.; writing—original draft, L.H.; writing—review and editing, X.H., D.L. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “National Key Research and Development Program of China under Grant (No. 2020YFE0200500)”.

Data Availability Statement

Data that support the findings of this study is available from the corresponding author upon reasonable request.

Acknowledgments

This work is supported by the National Key Research and Development Program of China under Grant (No.2020YFE0200500). The authors gratefully appreciate the anonymous reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Firouzi, F.; Farahani, B.; Marinšek, A. The convergence and interplay of edge, fog, and cloud in the AI-driven Internet of Things (IoT). Inf. Syst. 2022, 107, 101840. [Google Scholar] [CrossRef]
  2. Ignatov, A.; Malivenko, G.; Plowman, D.; Shukla, S.; Timofte, R. Fast and accurate single-image depth estimation on mobile devices, mobile ai 2021 challenge: Report. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2545–2557. [Google Scholar]
  3. Hard, A.; Rao, K.; Mathews, R.; Ramaswamy, S.; Beaufays, F.; Augenstein, S.; Eichner, H.; Kiddon, C.; Ramage, D. Federated learning for mobile keyboard prediction. arXiv 2018, arXiv:1811.03604. [Google Scholar]
  4. Alam, T.; Gupta, R. Federated Learning and Its Role in the Privacy Preservation of IoT Devices. Future Internet 2022, 14, 246. [Google Scholar] [CrossRef]
  5. Li, L.; Fan, Y.; Tse, M.; Lin, K.Y. A review of applications in federated learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
  6. Al-Saedi, A.A.; Boeva, V.; Casalicchio, E. FedCO: Communication-Efficient Federated Learning via Clustering Optimization. Future Internet 2022, 14, 377. [Google Scholar] [CrossRef]
  7. Banabilah, S.; Aloqaily, M.; Alsayed, E.; Malik, N.; Jararweh, Y. Federated learning review: Fundamentals, enabling technologies, and future applications. Inf. Process. Manag. 2022, 59, 103061. [Google Scholar] [CrossRef]
  8. Asad, M.; Aslam, M.; Jilani, S.F.; Shaukat, S.; Tsukada, M. SHFL: K-Anonymity-Based Secure Hierarchical Federated Learning Framework for Smart Healthcare Systems. Future Internet 2022, 14, 338. [Google Scholar] [CrossRef]
  9. Yang, Q.; Liu, Y.; Cheng, Y.; Kang, Y.; Chen, T.; Yu, H. Federated learning. Synth. Lect. Artif. Intell. Mach. Learn. 2019, 13, 1–207. [Google Scholar] [CrossRef]
  10. Chen, Z.; Liao, W.; Tian, P.; Wang, Q.; Yu, W. A Fairness-Aware Peer-to-Peer Decentralized Learning Framework with Heterogeneous Devices. Future Internet 2022, 14, 138. [Google Scholar] [CrossRef]
  11. Niknam, S.; Dhillon, H.S.; Reed, J.H. Federated learning for wireless communications: Motivation, opportunities, and challenges. IEEE Commun. Mag. 2020, 58, 46–51. [Google Scholar] [CrossRef]
  12. Hamer, J.; Mohri, M.; Suresh, A.T. FedBoost: A Communication-Efficient Algorithm for Federated Learning. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020; Volume 119, pp. 3973–3983. [Google Scholar]
  13. Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [Google Scholar] [CrossRef]
  14. Kang, J.; Xiong, Z.; Niyato, D.; Xie, S.; Zhang, J. Incentive mechanism for reliable federated learning: A joint optimization approach to combining reputation and contract theory. IEEE Internet Things J. 2019, 6, 10700–10714. [Google Scholar] [CrossRef]
  15. AbdulRahman, S.; Tout, H.; Ould-Slimane, H.; Mourad, A.; Talhi, C.; Guizani, M. A survey on federated learning: The journey from centralized to distributed on-site learning and beyond. IEEE Internet Things J. 2020, 8, 5476–5497. [Google Scholar] [CrossRef]
  16. Wang, T.; Liu, Y.; Zheng, X.; Dai, H.N.; Jia, W.; Xie, M. Edge-based communication optimization for distributed federated learning. IEEE Trans. Netw. Sci. Eng. 2021, 9, 2015–2024. [Google Scholar] [CrossRef]
  17. Samarakoon, S.; Bennis, M.; Saad, W.; Debbah, M. Distributed federated learning for ultra-reliable low-latency vehicular communications. IEEE Trans. Commun. 2019, 68, 1146–1159. [Google Scholar] [CrossRef]
  18. Li, Y.; Chen, C.; Liu, N.; Huang, H.; Zheng, Z.; Yan, Q. A blockchain-based decentralized federated learning framework with committee consensus. IEEE Netw. 2020, 35, 234–241. [Google Scholar] [CrossRef]
  19. Zhang, X.; Liu, Y.; Liu, J.; Argyriou, A.; Han, Y. D2D-assisted federated learning in mobile edge computing networks. In Proceedings of the 2021 IEEE Wireless Communications and Networking Conference (WCNC), Nanjing, China, 29 March–1 April 2021; pp. 1–7. [Google Scholar] [CrossRef]
  20. Xu, Z.; Tian, W.; Liu, Y.; Ning, W.; Wu, J. A Ring Topology-based Optimization Approach for Federated Learning in D2D Wireless Networks. arXiv 2022, arXiv:2212.02830. [Google Scholar]
  21. Wang, Z.; Hu, Y.; Yan, S.; Wang, Z.; Hou, R.; Wu, C. Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems. Electronics 2022, 11, 1548. [Google Scholar] [CrossRef]
  22. Lee, Y.; Park, S.; Kang, J. Fast-Convergent Federated Learning via Cyclic Aggregation. arXiv 2022, arXiv:2210.16520. [Google Scholar]
  23. Lee, J.w.; Oh, J.; Lim, S.; Yun, S.Y.; Lee, J.G. Tornadoaggregate: Accurate and scalable federated learning via the ring-based architecture. arXiv 2020, arXiv:2012.03214. [Google Scholar]
  24. Qu, Y.; Uddin, M.P.; Gan, C.; Xiang, Y.; Gao, L.; Yearwood, J. Blockchain-enabled federated learning: A survey. ACM Comput. Surv. 2022, 55, 1–35. [Google Scholar] [CrossRef]
  25. Boenisch, F.; Dziedzic, A.; Schuster, R.; Shamsabadi, A.S.; Shumailov, I.; Papernot, N. When the curious abandon honesty: Federated learning is not private. arXiv 2021, arXiv:2112.02918. [Google Scholar]
  26. Lyu, L.; Yu, H.; Yang, Q. Threats to federated learning: A survey. arXiv 2020, arXiv:2003.02133. [Google Scholar]
  27. Śmietanka, M.; Pithadia, H.; Treleaven, P. Federated Learning for Privacy-Preserving Data Access. 2020. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3696609 (accessed on 15 September 2020).
  28. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
  29. Paragliola, G. Evaluation of the trade-off between performance and communication costs in federated learning scenario. Future Gener. Comput. Syst. 2022, 136, 282–293. [Google Scholar] [CrossRef]
  30. Lin, T.; Kong, L.; Stich, S.U.; Jaggi, M. Ensemble distillation for robust model fusion in federated learning. Adv. Neural Inf. Process. Syst. 2020, 33, 2351–2363. [Google Scholar]
  31. Liu, L.; Zhang, J.; Song, S.; Letaief, K.B. Client-Edge-Cloud Hierarchical Federated Learning. In Proceedings of the ICC 2020-2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  32. Paragliola, G. A federated learning-based approach to recognize subjects at a high risk of hypertension in a non-stationary scenario. Inf. Sci. 2023, 622, 16–33. [Google Scholar] [CrossRef]
  33. Huang, Y.; Hu, C. Toward Data Heterogeneity of Federated Learning. arXiv 2022, arXiv:2212.08944. [Google Scholar]
  34. Li, T.; Sanjabi, M.; Beirami, A.; Smith, V. Fair resource allocation in federated learning. arXiv 2019, arXiv:1905.10497. [Google Scholar]
  35. Fang, H.; Qian, Q. Privacy preserving machine learning with homomorphic encryption and federated learning. Future Internet 2021, 13, 94. [Google Scholar] [CrossRef]
  36. Qu, Y.; Gao, L.; Luan, T.H.; Xiang, Y.; Yu, S.; Li, B.; Zheng, G. Decentralized privacy using blockchain-enabled federated learning in fog computing. IEEE Internet Things J. 2020, 7, 5171–5183. [Google Scholar] [CrossRef]
  37. Xing, H.; Simeone, O.; Bi, S. Decentralized federated learning via SGD over wireless D2D networks. In Proceedings of the 2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Atlanta, GA, USA, 26–29 May 2020; pp. 1–5. [Google Scholar] [CrossRef]
  38. Kumaresan, R.; Vaikuntanathan, V.; Vasudevan, P.N. Improvements to Secure Computation with Penalties. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; ACM: New York, NY, USA, 2016; pp. 406–417. [Google Scholar] [CrossRef]
  39. Desai, H.B.; Ozdayi, M.S.; Kantarcioglu, M. Blockfla: Accountable federated learning via hybrid blockchain architecture. In Proceedings of the Eleventh ACM Conference on Data and Application Security and Privacy, Virtual Event, 26–28 April 2021; pp. 101–112. [Google Scholar] [CrossRef]
  40. Courtois, N.T.; Grajek, M.; Naik, R. Optimizing sha256 in bitcoin mining. In Proceedings of the International Conference on Cryptography and Security Systems, Lublin, Poland, 29 August 2014; pp. 131–144. [Google Scholar] [CrossRef]
  41. Hankerson, D.; Menezes, A. Elliptic Curve Cryptography; Springer: Berlin, Germany, 2011. [Google Scholar]
  42. Suhaili, S.; Watanabeb, T. Design of Optimized Pipelined RIPEMD-160 with High Frequency and Throughput. J. Adv. Res. Comput. Appl. 2016, 3, 17–27. [Google Scholar]
  43. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.Y. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; Volume 54, pp. 1273–1282. [Google Scholar]
  44. Deng, L. The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web]. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  45. Ho-Phuoc, T. CIFAR10 to Compare Visual Recognition Performance between Deep Neural Networks and Humans. arXiv 2018, arXiv:1811.07270. [Google Scholar]
  46. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Kiddon, C.; Konečný, J.; Mazzocchi, S.; McMahan, B.; et al. Towards Federated Learning at Scale: System Design. In Proceedings of the Machine Learning and Systems, Stanford, CA, USA, 31 March–2 April 2019; Volume 1, pp. 374–388. [Google Scholar]
  47. Hsieh, K.; Phanishayee, A.; Mutlu, O.; Gibbons, P.B. The Non-IID Data Quagmire of Decentralized Machine Learning. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
Figure 1. The scenario of RingFFL.
Figure 1. The scenario of RingFFL.
Futureinternet 15 00068 g001
Figure 2. The work flow of RingFFL.
Figure 2. The work flow of RingFFL.
Futureinternet 15 00068 g002
Figure 3. An example of the penalty mechanism in RingFFL.
Figure 3. An example of the penalty mechanism in RingFFL.
Futureinternet 15 00068 g003
Figure 4. Accuracy comparison of FL for MNIST data set in IID and non-IID cases.
Figure 4. Accuracy comparison of FL for MNIST data set in IID and non-IID cases.
Futureinternet 15 00068 g004
Figure 5. Accuracy comparison of FL for CIFAR10 data set in IID and non-IID cases.
Figure 5. Accuracy comparison of FL for CIFAR10 data set in IID and non-IID cases.
Futureinternet 15 00068 g005
Figure 6. Accuracy comparison of FL for MNIST data set in IID and non-IID cases with clients leaving.
Figure 6. Accuracy comparison of FL for MNIST data set in IID and non-IID cases with clients leaving.
Futureinternet 15 00068 g006
Figure 7. Accuracy comparison of FL for CIFAR10 data set in IID and non-IID cases with clients leaving.
Figure 7. Accuracy comparison of FL for CIFAR10 data set in IID and non-IID cases with clients leaving.
Futureinternet 15 00068 g007
Figure 8. The circulation of digital currencies when there are 5 clients participating in FL.
Figure 8. The circulation of digital currencies when there are 5 clients participating in FL.
Futureinternet 15 00068 g008
Figure 9. The circulation of digital currencies when there are 10 clients participating in FL.
Figure 9. The circulation of digital currencies when there are 10 clients participating in FL.
Futureinternet 15 00068 g009
Figure 10. The circulation of digital currencies when there are 15 clients participating in FL.
Figure 10. The circulation of digital currencies when there are 15 clients participating in FL.
Futureinternet 15 00068 g010
Figure 11. The circulation of digital currencies when there are 20 clients participating in FL.
Figure 11. The circulation of digital currencies when there are 20 clients participating in FL.
Futureinternet 15 00068 g011
Figure 12. The circulation of digital currencies when there are 20 clients participating in FL and three leave halfway.
Figure 12. The circulation of digital currencies when there are 20 clients participating in FL and three leave halfway.
Futureinternet 15 00068 g012
Figure 13. Comparisons of training time including and excluding blockchain of FL for MNIST data set in IID and non-IID cases.
Figure 13. Comparisons of training time including and excluding blockchain of FL for MNIST data set in IID and non-IID cases.
Futureinternet 15 00068 g013
Figure 14. Comparisons of training time including and excluding blockchain of FL for CIFAR10 data set in IID and non-IID cases.
Figure 14. Comparisons of training time including and excluding blockchain of FL for CIFAR10 data set in IID and non-IID cases.
Futureinternet 15 00068 g014
Table 1. Summary of main notations.
Table 1. Summary of main notations.
ParametersDescription
P i The i-th client
NThe number of all clients
T i Client i’s private data set
M i Client i’s local model parameters
CConditions for obtaining digital currencies
EEvidence needed to obtain digital currencies
M i n t Initial model of FL
M A X Maximum communication rounds of FL
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, L.; Huang, X.; Li, D.; Zhang, Y. RingFFL: A Ring-Architecture-Based Fair Federated Learning Framework. Future Internet 2023, 15, 68. https://doi.org/10.3390/fi15020068

AMA Style

Han L, Huang X, Li D, Zhang Y. RingFFL: A Ring-Architecture-Based Fair Federated Learning Framework. Future Internet. 2023; 15(2):68. https://doi.org/10.3390/fi15020068

Chicago/Turabian Style

Han, Lu, Xiaohong Huang, Dandan Li, and Yong Zhang. 2023. "RingFFL: A Ring-Architecture-Based Fair Federated Learning Framework" Future Internet 15, no. 2: 68. https://doi.org/10.3390/fi15020068

APA Style

Han, L., Huang, X., Li, D., & Zhang, Y. (2023). RingFFL: A Ring-Architecture-Based Fair Federated Learning Framework. Future Internet, 15(2), 68. https://doi.org/10.3390/fi15020068

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop