Next Article in Journal
A Novel Combined Method for Measuring the Three-Dimensional Rotational Angle of a Spherical Joint
Next Article in Special Issue
Modeling Structured Dependency Tree with Graph Convolutional Networks for Aspect-Level Sentiment Classification
Previous Article in Journal
Dynamic Tensor Modeling for Missing Data Completion in Electronic Toll Collection Gantry Systems
Previous Article in Special Issue
DaylilyNet: A Multi-Task Learning Method for Daylily Leaf Disease Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Personalized Fair Split Learning for Resource-Constrained Internet of Things

1
College of Science, North China University of Science and Technology, Tangshan 063210, China
2
Hebei Key Laboratory of Data Science and Application, Tangshan 063210, China
3
Tangshan Key Laboratory of Data Science, Tangshan 063210, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(1), 88; https://doi.org/10.3390/s24010088
Submission received: 3 November 2023 / Revised: 3 December 2023 / Accepted: 20 December 2023 / Published: 23 December 2023
(This article belongs to the Special Issue Emerging Machine Learning Techniques in Industrial Internet of Things)

Abstract

:
With the flourishing development of the Internet of Things (IoT), federated learning has garnered significant attention as a distributed learning method aimed at preserving the privacy of participant data. However, certain IoT devices, such as sensors, face challenges in effectively employing conventional federated learning approaches due to limited computational and storage resources, which hinder their ability to train complex local models. Additionally, in IoT environments, devices often face problems of data heterogeneity and uneven benefit distribution between them. To address these challenges, a personalized and fair split learning framework is proposed for resource-constrained clients. This framework first adopts a U-shaped structure, dividing the model to enable resource-constrained clients to offload subsets of the foundational model to a central server while retaining personalized model subsets locally to meet the specific personalized requirements of different clients. Furthermore, to ensure fair benefit distribution, a model-aggregation method with optimized aggregation weights is used. This method reasonably allocates model-aggregation weights based on the contributions of clients, thereby achieving collaborative fairness. Experimental results demonstrate that, in three distinct data heterogeneity scenarios, employing personalized training through this framework exhibits higher accuracy compared to existing baseline methods. Simultaneously, the framework ensures collaborative fairness, fostering a more balanced and sustainable cooperation among IoT devices.

1. Introduction

In recent years, with the rapid advancements in the Internet of Things (IoT) and distributed computing driven by big data, IoT devices have been extensively deployed. Particularly in the Industrial Internet of Things (IIoT) domain, a multitude of sensor devices are dispersed across various locations for the purpose of network data collection. However, aggregating network big data to centralized servers may potentially infringe upon the data privacy of individual enterprises or clients, potentially leading to the formation of multiple data silos [1]. To address the challenges stemming from the development of IoT and the application-related “data silos” phenomenon, federated learning (FL) [2] has gained increasing attention as a distributed learning method. FL allows multiple participants to collaboratively train a shared global model without sharing raw data, thus ensuring the protection of participant data privacy. However, in IoT environments, some resource-constrained IoT devices cannot easily train complex local models using traditional FL methods due to limitations in computational and storage resources.
Split learning (SL), as an evolved form of FL, proves to be more suitable for resource-constrained IoT devices [3]. SL allows one to split complex client-side local models, offloading most of the computational tasks of the model to a central server while retaining only subsets of the model at the client end. This method effectively reduces the computational and storage burden on IoT devices, making them more adaptable to resource-limited environments. In addition, in complex IoT environments, there are also issues of data heterogeneity and uneven distribution of benefits among devices. On the one hand, data heterogeneity manifests as each device being influenced by its environment and individual preferences, resulting in a unique data distribution and significant differences in data volume. For instance, in the realm of sensors, distinct sensor devices exhibit unique data distributions and noticeable disparities in data volume due to differences in sensing capabilities and deployment environments. Due to data heterogeneity, models trained through SL may have difficulty in effectively capturing common features of different data types across devices, which may lead to the degradation of the model’s performance on a specific device or even prevent affected devices from participating in SL training [4]. On the other hand, the uneven distribution of benefits emerges due to imbalances in data quality and scale among devices. Some smaller-scale IoT entities, despite having lesser data volume, possess more representative or higher-quality data, playing a crucial role in model training. However, despite their substantial contributions, these entities often receive less benefit compared to larger-scale entities [5]. Such an imbalance might hinder these smaller entities from achieving expected outcomes during the training process, prompting them to opt out of cooperation, leading to the termination of SL training. Hence, in complex IoT environments, a mechanism is needed to address the issues encountered in SL training, catering to device-specific requirements while ensuring fairness.
To address the aforementioned issues, a personalized and fair split learning framework, termed Split Learning with Personalized Fairness (SplitLPF), has been proposed. The framework begins by splitting the model into a three-segment U-shaped structure during model training. In this structure, the central server acts as the offloading entity, undertaking the burden of intensive computational tasks to facilitate the training of intricate models on resource-constrained client devices. Simultaneously, client devices retain personalized model subsets locally to cater to the specific individualized needs of different clients, enhancing their ability to address the challenges posed by data heterogeneity. Furthermore, during the model-aggregation phase, the central server adopts an optimized aggregation weight methodology. This method ensures fair benefit distribution among various clients in federated learning by reasonably allocating model-aggregation weights based on client contributions, utilizing estimations derived from differences in dataset sizes and gradient directions. This approach incentivizes the active participation of contributors in the model-training process, thereby achieving collaborative fairness.
The main contributions of this paper are as follows:
(1)
This paper proposes a personalized U-shaped split architecture, designed to fulfil specific individualized requirements of different clients in complex IoT environments while effectively training sophisticated models on resource-constrained client devices;
(2)
This paper introduces a model-aggregation method based on contribution estimation. By estimating the contribution size of each client, the aggregation weights of the model are reasonably assigned to ensure a fairer distribution of benefits among different clients;
(3)
The experimental results demonstrate that our proposed framework not only exhibits high accuracy in its personalized training approach under different data heterogeneity scenarios, it also ensures fairness in collaboration and promotes more balanced and sustainable cooperation among clients.

2. Background and Related Work

2.1. Distributed Learning in the IoT

In the realm of the IoT, distributed learning has emerged as a pivotal technology, offering effective solutions for handling large-scale dispersed sensor and device data. In this context, distributed learning methods such as federated learning and split learning and their variants play significant roles in IoT [6]. These methods underscore privacy preservation [7], decentralization [8], and resource efficiency [9], holding promise to propel the development of IoT, fostering intelligent decision-making and enabling the realization of data-driven applications.

2.1.1. Federated Learning

In an FL system, two primary entities can be divided: there are the data owners collaborating in model training, referred to as clients, and there is the model owner who is responsible for coordinating the training process and model aggregation, termed the central server [10]. Let N = { 1 , 2 , . . . , K } represent a set containing K clients, each of which possesses its local dataset D k ( k K ) , constituting the entire dataset D = k = 1 K D k . The FL system is shown in Figure 1a. The FL system in the initial phase of model training, and the central server first initializes the model and broadcasts the initial global model to the selected clients. Subsequently, each selected client employs its local dataset to train its local model, uploading the trained model parameters (gradients) to the central server. The central server aggregates all received client gradients, producing a new global gradient, ultimately using this global gradient to update all clients’ models. This local training and global aggregation process iterate in a loop until the global model converges, thereby completing the entire model-training process. However, some resource-constrained IoT devices face challenges in effectively utilizing traditional federated learning methods to train complex local models.

2.1.2. Split Learning and Its Variants

To address the challenge of resource constraints, SL proposes to use servers to collaboratively train local models without the need to scale down the model and obtain raw data from the client [11]. SL involves dividing the complete model ω into client-side ω C and server-side models ω S . The parallel split learning (PSL) system is shown in Figure 1b, where each client performs forward propagation steps in parallel during PSL training until a specific clipping layer is reached. The output of this cut layer is then transmitted to the server, where the remaining part of the model undergoes forward propagation to generate prediction results. The training loss is computed on the server using labels and predicted values. Subsequently, a backward propagation step is performed on the server until it reaches the cut layer [12]. Afterwards, the gradients are sent back to the respective clients, enabling parallel execution of backward propagation on the first segment of the model (the client-side model). This process iterates until the final model is obtained. In this configuration, there is no need for data or weight aggregation to accomplish the training of complex models. However, in certain configurations, appropriate variants of SL are applied, some of which involve operations such as aggregating model weights and aggregating intermediate data [13,14,15]. In weight aggregation-based SL, clients transmit their local model weights to a trusted third party for the aggregation of these weights. For instance, in SplitFed learning [13] as illustrated in Figure 1c, the model is divided into two segments and trained in a manner similar to PSL. After each client completes a training round, its local model weights are transferred to the server, which employs the FedAvg aggregation method to integrate these local models. Joshi et al. [14] proposed the SplitFed learning framework SFPL with positive labels to improve the deep learning model training for resource-constrained IoT clients. In intermediate data aggregation-based SL, intermediate data like activations or gradient information from different clients are aggregated. For instance, in SGLR [15] shown in Figure 1d, the server aggregates local gradients from clients instead of model parameters. Additionally, SGLR divides the learning rate into server-side and client-side learning rates, adjusting them separately to support parallelism among multiple clients. SGLR effectively reduces the volume of shared information among clients and mitigates client decoupling issues [16].
The model partitioning employed in the aforementioned SL and its variants has somewhat increased communication costs due to the transmission of intermediate data during the model-training process, which needs to be exchanged in each iteration, resulting in higher communication overhead. To mitigate the communication cost in SL, Chen et al. [17] proposed an asynchronous training scheme based on loss, where gradients are transmitted at specified intervals and quantized using 8-bit floating-point before transmission, thereby reducing the communication cost of SL. Oh et al. [18] introduced a communication-efficient SL framework called SplitFC, utilizing adaptively compressed intermediate features and gradient vectors based on varying dispersion levels in vectors. They combined adaptive feature dropping and adaptive feature quantization as compression strategies to minimize communication overhead. Lin et al. [19] presented an efficient PSL framework, EPSL, which aggregates the last layer’s activation gradients during backpropagation. They conducted joint optimization on EPSL’s subchannel allocation, power control, and layer pruning to reduce communication costs. While most methods for reducing communication costs adopt techniques like value quantization and feature compression, these may significantly impact the accuracy of the global model. However, in [20], the number of communications is reduced and the model quality is improved by transmitting more information and smaller synthetic data. This paper is inspired by that, and lightweight synthetic data transfer is chosen to alleviate some communication pressure.

2.2. Distributed Learning Challenges and Strategies in IoT

2.2.1. Challenges to Distributed Learning in the IoT

In the complex IoT environment, SL model splitting proves effective at addressing resource-constrained challenges. However, the complexity and specificity of the IoT environment poses additional challenges, notably data heterogeneity and uneven benefit distribution. These challenges hold significant importance in both academic research and practical applications.
  • Data Heterogeneity
In IoT environments, particularly in IIoT environments, the diversity among different sensor devices, arising from variations in sensing capabilities, environmental conditions, and data characteristics, results in significant differences in collected data. This diversity poses challenges for traditional distributed learning methods, which commonly operate under the assumption of data being identically and independently distributed (IID). Consequently, there is a pressing need to explore novel methodologies to address the issue of data heterogeneity based on the assumption of non-independent and identically distributed (Non-IID) data.
  • Uneven Benefit Distribution
The imbalance in data quality and quantity among different devices results in varied benefits obtained from their contributions. However, fairness becomes particularly crucial when engaging in data sharing and collaboration among multiple sensor devices. Given that these devices are typically distributed across different locations and collect diverse types of data, an effective collaborative approach is required to achieve integrated sensing and analysis. Therefore, there is a need for fair model aggregation so that the contribution of each device is justly recognized, thus effectively improving the problem of unequal distribution of benefits and promoting a more balanced cooperative relationship.

2.2.2. Strategies to Distributed Learning in the IoT

In this subsection, we aim to review the representative techniques (strategies) proposed to address the typical challenges of distributed learning in the IoT context. Given the focal point of this paper on personalized fair SL, and the fact that SL-related work has been discussed in detail earlier, this subsection will provide the necessary background information on personalized processing and fairness design.
  • Personalized Processing
To address the challenges posed by the non-IID data, an effective strategy involves personalized handling at both the data and model levels to mitigate data heterogeneity and deliver high-quality personalized models for each client. Personalized FL encompasses various approaches, including those based on data, models, and architecture [21]. Data-based methods aim to reduce the heterogeneity in client data distributions to alleviate client drift issues. This approach can be realized using techniques such as data enhancement [22] and client selection mechanisms [23]. Model-based approaches aim to learn a robust global FL model while catering to personalized services for each individual client. This can be achieved by learning global and local features of each client [24], leveraging historical memory to learn global representations [25], and improving local models through transfer learning [26]. Architecture-based methods focus on providing tailored and personalized model architectures for individual clients. The approach can be implemented using techniques such as parameter decoupling [27] and knowledge distillation [28], where parameter decoupling mainly provides a personalization layer for each client.
In the realm of personalized SL, relatively fewer researchers have delved into this area. Among them, Wadhwa et al. [29] proposed that PFSL, which employs transfer learning for pre-training. Subsequently, by freezing certain shared layer weights and updating the unfrozen ones, each client can train their model for varying numbers of epochs, thereby enabling personalized operations. Han et al. [30] proposed SplitGP, leveraging multi-output neural networks to capture both generalized and personalized learning requirements. This approach empowers client models to optimize for their respective primary tasks during the training process, fostering robust personalized capabilities. However, the aforementioned personalized SL methods predominantly focus on model-based approaches. In this paper, we aim to adopt an architecture-based approach. We intend to meet customized personalized requirements through the design of foundational and personalized model architectures.
  • Fairness Design
Fairness design can effectively address the challenges posed by uneven benefits distribution. In this context, both model fairness and collaborative fairness have garnered significant attention as distinct perspectives in fairness design. Model fairness concerns whether the global model constructed in federated learning maintains fairness in its treatment of different participants [31], aiming to enhance uniformity in performance across all involved parties. On the other hand, collaborative fairness places greater emphasis on the cooperative interaction among participants in federated learning. This entails ensuring equal opportunities for engagement and contribution among all participants during training, alongside fair treatment of each participant’s contributions [32].
For model fairness, Hu et al [33] proposed the FedMGDA+ method, which performs multi-objective optimization by optimizing the loss function of each FL client individually and simultaneously to avoid sacrificing the model performance of any client. Cui et al. [34] proposed a constrained multi-objective optimization framework, learning a model that satisfies the fairness constraints of all clients with consistent performance by optimizing the agent’s maximal function involving all objectives. Li et al. [35] proposed the Ditto method, employing inter-client fine-tuning to minimize individual losses significantly. They augmented this by adding regularization terms, bringing personalized models closer to the optimal global model to achieve consistent model performance. In contrast to model fairness, collaborative fairness garners higher attention due to its direct involvement with collaboration and contributions among clients. Lyu et al. [36] proposed a collaborative fairness framework using a reputation mechanism to gauge participant contributions, converging toward different models to ensure fairness. Xu et al. [37] designed a gradient-reward mechanism, in which fairness is ensured by sparsification to treat the aggregated gradient downloaded by the server as an expected marginal contribution to each client. However, in most collaborative fairness designs, more or less shapely values [38] were used to participate in the contribution assessment, resulting in high computational cost. Therefore, in this paper an approximation is used to measure the contribution size to alleviate the computational pressure.

3. SplitLPF Frame

The paper introduces a personalized and fair SL framework termed SplitLPF, as illustrated in Figure 2. In this framework, clients split the model and offload intricate computational tasks to a central server; this flexible splitting enables tasks to be assigned based on the characteristics of the device. During the model-training process, clients and the central server collaboratively execute the foundational model training. Subsequently, clients engage in more flexible personalized model training. This personalized training helps to improve the accuracy and performance of devices (such as sensors), making them better adapted to specific application scenarios and environments. During the model-aggregation process, model weights are optimized based on client contribution estimations, enabling the aggregation of split models. This ensures fairness among participants, providing an incentive for active participation in model training. To minimize communication overhead, the framework adopts a lightweight synthetic data transmission method. The SplitLPF framework amalgamates the strengths of both the central server and clients, effectively leveraging resources to address challenges faced by resource-constrained devices in complex IoT environments. Simultaneously, it ensures fairness in model training.
The design of the SplitLPF framework is divided into three components: model partition, personalized training, and fairness mechanism guarantee. Subsequently, detailed introduction will be provided for each of these components.

3.1. Model Split

In SL model training, it is common to split the model into two segments: the client-side model ω C and the server-side model ω S . The bulk of computational tasks is offloaded to the server-side model ω S to better accommodate the model training requirements of resource-constrained devices. This approach simultaneously reduces the need for transmitting sensitive data, consequently lowering the risk of client data exposure.
Consider a system with a single central server and N clients; each client has its own private dataset (local dataset) { D 1 , D 2 , . . . , D N } , and D k denotes the dataset owned by the k-th participant. Similar to FL model training, the goal of SL training is to collaboratively determine an optimal model ω * through the cooperation of clients and the central server, fulfilling the minimization of the loss function as shown in Equation (1):
min ω F ( ω ) = min ω k = 1 N p k F k ( ω )
where p k denotes the weight of participant k local model when aggregated at the central server, satisfying p k 0 and k = 1 N p k = 1 . It is commonly set as p k = | D k | k = 1 N | D k | , where F k ( ω ) denotes the local loss function of participant k. This function is further defined as the sum of the loss functions between the client-side model ω C and the server-side model ω S , as illustrated by Equation (2):
F k ( ω ) = 1 | D k | x D k ( C ( x ; ω C ) + S ( A k ; ω S ) )
where C ( ) and S ( ) denote the loss functions of the client-side models and server-side models, respectively. A k denotes the intermediate result (such as activation), D k denotes the client-side dataset, and x denotes the local data. In SL, clients train the model up to its splitting layer using raw data and then transmit the intermediate results (activation data A k ) from this splitting layer to the server. The server employs the received intermediate results from clients to train the remaining layers of the model, completing the forward propagation of the model. Subsequently, the server conducts backpropagation up to the splitting layer and forwards the intermediate results (gradient g) to the clients. Utilizing the gradient, each client performs its backpropagation on the split model, iteratively updating it to accomplish the entire model training. Each client k updates the client-side model based on the local gradient information as shown in Equation (3):
ω C , k t + 1 = ω C , k t η t F C , k ( ω C , k t )
where ω C , k t denotes the local model of client k in round t, η t denotes the learning rate in round t, and F C , k ( ) denotes the gradient of the local model of client k. The server updates the server-side model according to Equation (4):
ω S , k t + 1 = ω S , k t η t F S , k ( ω S , k t )
where ω S , k t denotes the server model updated for round t based on the intermediate results of client k, and F S , k ( ) denotes the gradient of the server model. In resource-constrained environments, some devices may have difficulty in transmitting large volumes of data (such as activations) for long periods of time due to unstable network connections, limited bandwidth, and low device performance, and thus they may not be able to effectively participate in model training. However, in this paper we address the communication pressure by substituting bulk data transmission with lightweight synthetic data. In distributed learning, model compression primarily occurs at two main places: the local model update transmission from the client to the server in the upstream, and the global model update transmission from the server to the client in the downstream [39]. Similar to compression work, we replace model updates with synthetic data in both upstream and downstream data transmission. In the conventional SL model training, A k is upstream transmission data, F S , k ( ) is downstream transmission data, and clients k and servers attempt to find synthetic data D C k s y n and D S k s y n as replacements for transmission data. The most intuitive approach is to minimize the distance metric between the model generated using synthetic data and the model generated from real data [40]. Therefore, we optimize the synthetic data using the objective function of the 2 distance between simulated model weights and real model weights, employing stochastic gradient descent (SGD) to generate appropriate lightweight synthetic data D s y n , as shown in Equation (5):
D s y n S G D ( D s y n ; f ( ω r a w ; ω s y n ) )
where ω s y n = M ( D s y n ; ω ) denotes the model generated using synthetic data, M ( ) denotes the model update function, and ω r a w denotes the model generated using real data. f ( ) denotes the loss function based on the 2 distance. Subsequently, the server involved in upstream communication and the client involved in downstream communication can use this synthetic data to recover local activation data or gradient information. The server and client use the same process as the generation of D s y n to recover the original data D r a w . This communication scheme has been widely applied in the literature [40,41,42,43], and specific details are not further elaborated on in this paper. In comparison to large batches of real data, lightweight synthetic data, while having the same input dimensions, achieves similar feature representations to large batches of data in smaller batches and can be recovered as real data with less computation. Based on the experimental observations in Section 4.4, we find that the use of lightweight synthetic data transmission reduces the communication cost by almost approximately 24% compared to the existing personalized fair SL.

3.2. Personalized Training

In complex IoT environments, model partitioning effectively alleviates the computational burden on resource-constrained clients and enhances prediction accuracy to some extent. However, in IoT environments, particularly in IIoT domains, various types of sensor devices may possess different sensing capabilities, disparate data distributions, and varying data volumes, exacerbating the problem of data heterogeneity. This data heterogeneity issue can potentially impact model performance. In extreme scenarios, data types, data distributions, or even data contents among different sensor devices might be entirely dissimilar. Under such circumstances, it is difficult to use SL model training methods to handle this heterogeneity effectively. In this context, personalized training becomes paramount.
In the model training phase, this paper further subdivided the traditional SL to form a three-stage U-shaped structure. The client-side retains the input and output layers locally and performs the computation of the loss function, while the central server performs the complex hidden layer computation. The specific process of training is as follows: during each global iteration, individual clients compute the first part (front part) of the model and upload the intermediate data to the central server in parallel. Leveraging more robust computational capabilities, the central server swiftly conducts corresponding forward propagation, completing the second part (middle part) of the model training. During this period, the clients and server collaborate to achieve the generalization phase of model training. Subsequently, the central server sends the results back to the client, which executes the SGD optimization algorithm to compute the prediction results, loss values, and gradients based on the respective local learning rates to complete the personalization phase training (back part) locally. Then, clients utilize the gradients to update their local parameters and transmit them back to the central server. The central server uses the gradient results to update the intermediate model parameters and sends the results back to the clients. Finally, the clients update the parameters of the first part of the model. In this process, clients and servers first complete the base model training, after which the clients independently execute personalized model training aimed at optimizing the model to better adapt to the characteristics of diverse devices, thereby enhancing the model’s performance and adaptability. Throughout the personalized training, clients retain label data locally, preventing other clients and servers from inferring model specifics or local data through label data. This ensures the enhancement of model training efficiency while safeguarding data privacy.
The client-side model ω C training can be further broken down into two components: base model training ω C B and personalized model training ω C P . Each round of update of the base model and personalized model is represented as Equations (6) and (7):
ω C B , k t + 1 = ω C B , k t η t F C B , k ( ω C B , k t )
ω C P , k t + 1 = ω C P , k t η k t F C P , k ( ω C P , k t )
where F C B , k ( ) and F C P , k ( ) denote the model gradient for client k base model and personalized model training, respectively. η k t is the local learning rate of client k in round t, and η t denotes the global learning rate in round t. During personalized training, the pseudo-code for the client to perform the model update locally is shown in Algorithm 1.

3.3. Fairness Mechanism

Personalized training exhibits significant advantages in addressing data heterogeneity issues in complex IoT environments. However, in the typical model-training process, the aggregation of model parameters typically relies on the dataset’s size to allocate weights for achieving a weighted average model aggregation. This approach may lead to the uneven distribution of benefits, making it challenging for smaller-scale devices possessing high-quality data to achieve expected outcomes during model training. This leads to smaller devices opting out of the collaboration and terminating model training. This poses a serious threat to the process of collaborative learn as the contribution of each device is not negligible. To address this challenge, the introduction of mechanisms for collaborative fairness is necessary.
Collaborative fairness aims to ensure the fair recognition of each device’s contribution. However, the size of the dataset by itself is insufficient to provide enough information to accurately reflect the contribution of each local client. Therefore, relying solely on dataset size-based weighted aggregation might not be the optimal aggregation strategy.
Algorithm 1 ClientUpdate
Input: 
D k : local dataset; N: number of clients; η t : global learning rate; η k t : local learning rate; Y k : the true labels;
Output: 
ω C P , k t : personalized model; ω C B , k t : base model;
1:
Procedure ClientForwordProp( ω C B , k t ):
2:
if epoch t = 0 then
3:
    Set the local activations A C B , k t = ;
4:
else
5:
     D S k s y n ServerUpdate()// Receive data transmitted by the server;
6:
    Recover the global ω ^ C B t based on D S k s y n ;
7:
end if
8:
Forward propagate the local data D k to the ω C B , k t cutting layer and obtain the local activations A C B , k t ;
9:
Generate synthetic data D C k s y n based on A C B , k t , | D k | , and ω C B , k t ;
10:
Send the synthetic data D C k s y n to the central server;
11:
Wait for ClientBackProp( ω C P , k t ) to complete;
12:
end procedure
13:
Procedure ClientBackProp( ω C P , k t ):
14:
D S k s y n ServerUpdate() // Receive data transmitted by the server;
15:
Recover A S , k t and d A S , k t based on D S k s y n ;
16:
if state = Personalized training then
17:
    Forward propagation with A S , k t on ω C P , k t ;
18:
    Calculate Y ^ k and loss calculation with Y k and Y ^ k ;
19:
    Back-propagation calculate C ( ω C P , k t ) ;
20:
    Generate D C k s y n using d A C P , k t : = C ( A S , k t ; ω C P , k t ) and send it to the server;
21:
    Update the personalized model ω C P , k t + 1 ω C P , k t η k t C ( ω C P , k t ) ;
22:
else
23:
    Calculate C ( ω C B , k t ) using d A S , k t ;
24:
    Update the base model ω C B , k t + 1 ω C B , k t η t C ( ω C B , k t ) ;
25:
end if
26:
end procedure
In this case, a viable aggregation strategy involves fairly distributing benefits by considering the contribution levels of the clients comprehensively. It might be beneficial to incorporate the discrepancy between local clients and the global model into the consideration for benefit distribution. Intuitively, certain clients with high-quality data provide more trustworthy and informative gradient information, enabling the model to update parameters more accurately and converge faster towards the global optimum, which results in a model whose gradient direction is closer to that of the global model. Consequently, clients whose gradient direction is more similar to that of the global model might contribute more significantly to the global model. This implies that the contribution level of clients might be related to the similarity of their gradient direction with the global model’s gradient direction. Hence, gradient direction disparity could serve as a useful supplementary metric. In other words, considering both dataset size and gradient direction disparity for assessing contribution levels seems more comprehensive. Section 3.5 provides theoretical analysis about this; it promotes fairness in benefit distribution through contribution estimation. The measurement of fairness could be defined as follows:
Definition 1. 
If the test performance distribution of model ω 1 is more uniform than that of model ω 2 , which is expressed as: s t d ( F k ( ω 1 ) ) < s t d ( F k ( ω 2 ) ) , it is said that model ω 1 is fairer than model ω 2 , where F k ( ) represents the test loss on client k, and s t d ( ) denotes the standard deviation.
The specific implementation steps of the fairness mechanism based on contribution estimation are as follows.
First, in order to reflect the gradient direction discrepancy, cosine similarity is used for measurement. In the t-th round, the discrepancy between client k and the global model is as shown in Equation (8):
d k = c o s ( F k ( ω k t ) , F ( ω t 1 ) )
where F k ( ω k ) denotes the local gradient of client k, and F ( ω ) denotes the global gradient. The global gradient can be expressed as: F ( ω ) = k = 1 N p k F k ( ω k ) , where p k denotes the aggregation weights of the model.
Then, the discrepancy values are further normalized as shown in Equation (9):
d ¯ k = | d k | k = 1 N | d k |
Finally, the aggregation weights of the model are optimized according to the size of the client’s contribution, and the optimized aggregation weights are as shown in Equation (10):
p k = α n k + ( 1 α ) d ¯ k
where α is a hyperparameter and n k = | D k | k = 1 N | D k | denotes the dataset size. In Equation (10), there is a free-rider attack by some clients that are small and in the opposite direction of the global model training, making the aggregation weights negative, and this can be circumvented using the ReLU function. Therefore, it can be normalized as shown in Equation (11):
p k = R e L U ( p k ) k = 1 N R e L U ( p k )
The central server can aggregate the global model based on the optimized model weights, denoted as: ω ^ = k = 1 N p k ω k . The pseudo-code for the server-side model update process is shown in Algorithm 2.
Algorithm 2 ServerUpdate
Input: 
C t : client set; η t : global learning rate;
Output: 
p k t : optimized model-aggregation weights;
1:
Initialize ω S 0 ;
2:
for each client k C t in parallel do
3:
     D C k s y n ClientUpdate() //Receive data transmitted by the client;
4:
    Recover A C B , k t , d A C P , k t and ω C B , k t using D C k s y n ;
5:
    Forward propagation with A C B , k t on ω S , k t and obtain the activations A S , k t ;
6:
    Back-propagation calculate S ( ω S , k t ) using d A C P , k t ;
7:
    Calculate d k t using Equation (7) //where F k ( ω k t ) C ( ω C P , k t ) ;
8:
    Normalize d k t using Equation (8) and collect the client dataset sizes n k ;
9:
    Calculate the optimized model-aggregation weights p k t using Equation (9);
10:
    Model aggregation ω ^ C B t = k = 1 N p k t ω C B , k t ;
11:
    Generate D S k s y n based on A S , k t , d A S , k t , and ω ^ C B t , and send it to the clients;
12:
    Update the server model ω S , k t + 1 ω S , k t η t S ( ω S , k t ) ;
13:
end for

3.4. SplitLPF Algorithm Analysis

The pseudo-code for the training process of SplitLPF as shown in Algorithm 3.
Algorithm 3 SplitLPF
Input: 
E: global iteration times; N: number of clients;
Output: 
ω C P , k t : personalized model; p k t : optimized model-aggregation weights;
1:
Split the model into ω C B , ω S , ω C P ;
2:
for  t 1 to E do
3:
    for  k 1 to N in parallel do
4:
        ClientForwordProp( ω C B , k t ) // Client forward propagation;
5:
        ServerUpdate( ω S , k t ) // Updates to the offloading model for the central server;
6:
         ω C P , k t ClientBackProp( ω C P , k t 1 ) // Client-side personalized training and back propagation;
7:
         p k t ServerUpdate( d A C P , k t ) //Server-side back propagation with aggregation of client-side base models and update of server-side models;
8:
        ClientBackProp( d A S , k t ) //The client performs back propagation and updates the local model;
9:
    end for
10:
end for

3.4.1. Complexity Analysis

In the training process of SplitLPF, each client trains the local model and sends it to the server (line 4), with a time complexity that equals o ( 1 ) . The central server forward propagates to update the intermediate model (line 5), with a time complexity that equals o ( N ) . Each client performs personalized training and back propagates it (line 6), with a time complexity that equals o ( 1 ) . The central server back propagation aggregates client-side base models and updating server-side models (line 7), with a time complexity that equals o ( N ) . The client back propagates and updates the local model (line 8), with a time complexity that also equals o ( 1 ) . Thus, the total time complexity of one model update for client k is o ( 2 N ) . The total time complexity of SplitLPF is o ( N 2 E ) through the cycle iteration. In terms of space complexity, SplitLPF additionally stores w t , D s y n and η t , and they are both fixed-size parameters. Therefore, SplitLPF also has the same space complexity, o ( N ) , with SL as well.

3.4.2. Total Cost Analysis

Assume that N is the number of clients, P is the total data size, C is the size of the activations that pass through the model trimming layer when considering one input sample from a single client, S is the total synthetic data size, R is the rate of communication between the client and the server, T is the time taken for one forward and backward propagation on the full model with dataset of size P, T 1 is the time to generate synthetic data based on real data, T a g g is the time required to perform the model aggregation, | W | is the the size of the full model, and β is the ratio of the full model size available to the client in SL/SplitFed, i.e., | W C | = β | W | , since clients download and upload client model updates before and after training respectively. Therefore, the size of each client communication becomes 2 β | W | . Based on the above assumptions, we compare the per-client communication cost, total communication cost, and total model training time in different approaches, as shown in Table 1.

3.5. Fairness Theory Analysis

This subsection will delve into exploring the convergence bounds of model training. Subsequently, based on this exploration, the effects of aggregation weights p k , dataset size n k , and local and global gradient discrepancy d k on convergence are obtained, and it is further concluded that dataset size n k and gradient direction discrepancy d k should be taken into account when determining the aggregation weights p k .
Given the personalized model partitioning employed in this paper, for the sake of simplicity in analysis, we will only consider the influence of client base models during model aggregation, i.e., ω k ω C B . Assumptions such as local objective function smoothness and bounded variance of the gradient used in the literature [44,45,46,47] are first employed in order to analyze the convergence of the model.
Theorem 1. 
Assuming the objective function is L-smooth and the gradients have bounded variance, when we set the global learning rate η < 1 L , the upper bound of convergence of the optimization is shown in Equation (12):
min t E | | F ( w t ) | | 2 2 1 4 ( H + 2 K ) [ 1 η T ( F ( w 0 ) F inf ) + 2 H B + 2 A H N k = 1 N d k + K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) ]
where H = N k = 1 N ( n k p k ) 2 , K = L 2 γ τ = 0 γ 2 η 2 τ ( 1 4 η 2 L 2 τ ) , A and B are both positive constants, and F inf and σ 2 correspond to the bounded scalar and variance of gradients, respectively.
A detailed proof is provided in Appendix A.
Considering stricter boundary conditions typically corresponds to superior optimization outcomes. Hence, this paper delves into exploring the impact of the parameter aggregation weight p k on the upper-bound conditions. From Theorem 1, it is apparent that the weight p k is primarily correlated with H, which corresponds to the three components in the convergence bounds. To shrink the bounds, consider optimizing p k to minimize the upper bound. As a result, a concise expression for the relationship between the aggregation weight p k , the dataset size n k , and the discrepancy d k between the local and global gradient can be derived.
Lemma 1. 
Assuming η < 1 / L , for more stringent upper bounds, the aggregation weight p k can be characterized as shown in Equation (13):
p k n k + J d k
where J is a positive constant.
Detailed proofs can be found in Appendix B.
In stricter upper limits, the aggregation weight p k correlates with the dataset size n k and the discrepancy d k between local and global gradients. For resource-constrained devices, the dataset size is typically relatively smaller. Therefore, further adjustments in the contributions of n k and d k can lead to the optimal model weight. Optimizing Lemma 1 appropriately results in the expression for the aggregation weight in this paper, i.e., p k = α n k + ( 1 α ) d k , as shown in Equation (10). Thus, fine-tuning the hyperparameters α provides an optimal selection for model performance.

4. Experiment and Result Analysis

The performance of SplitLPF in IoT environments was evaluated through experiments. Consider a typical IoT application scenario as shown in Figure 3. In general, IoT scenarios usually contain both strong devices (resource-rich clients) and weak devices (resource-constrained clients). Weak devices typically possess limited computational and storage resources, such as sensors and RFID tag cards. In contrast, strong devices have more computational and storage capabilities, capable of handling more complex tasks and algorithms, such as NVIDIA Jetson Nano and cloud servers. The experiments simulated a multi-machine deployment to adapt to the distributed environment requirements of split learning. To better manage resources and reduce unnecessary waste, a configuration comprising 10 commonly used IoT devices (including four Raspberry Pi, four STM32 boards, and two NVIDIA Jetson Nano) and one central server was chosen. In this configuration, the scaled-down Raspberry Pi and STM32 boards emulated weak devices, while the NVIDIA Jetson Nano emulated a strong device. In the experiment, the hardware environment of IoT devices encompassed CPU Cortex-A53, Cortex-M0, and Cortex-A57 ARM, with RAM capacities of 1G and 2G. Operating systems included Raspbian GNU/Linux10 (buster), FreeRTOS, and Ubuntu 18, with Python 3.7 as the development language. The central server (a laptop) featured an Intel i5-13500HQ CPU, 32GB RAM, 1TB ROM, and GPU RTX4060, and it ran on Windows 11. Software environment comprised Pycharm as the development environment and Python 3.9 as the programming language. In order to validate the effectiveness of SplitLPF, this paper conducts experiments on Fashion-MNIST (FMNIST), Extended MNIST (EMNIST), and CIFAR-10 datasets, all of which use the PyTorch framework for training deep learning models. These experiments aim to demonstrate the performance of SplitLPF under different datasets.

4.1. Datasets

The experiment evaluated the proposed framework using three popular datasets. FMNIST and CIFAR-10 datasets contain images from 10 categories each, while EMNIST, an extension of MNIST, includes 62 categories. All three datasets were divided into a training set and a test set using a standard data splitting approach. To simulate data heterogeneity seen in real-world scenarios, we used three forms of non-IID data segmentation for each dataset, as shown in Figure 4. Unbalanced Data Distribution (UDD): This scenario simulated uneven data distribution by altering the number of label categories within client datasets. In this case, a different number of labeling categories are distributed among different clients, i.e., weak clients usually have fewer labeling categories than strong clients but not less than one. For instance, among 10 clients, different subsets of label categories were allocated to each client, as illustrated in Figure 4a,d,g. Unbalanced Data Size (UDS): Data size was allocated to clients based on a power-law rule, creating a scenario where some clients had fewer data sizes while others had more. All clients possessed all label categories, but weak clients typically had less data than strong clients. For example, for the CIFAR-10 dataset with 10 clients, 500, 1000, 2000, and 4000 data are assigned to 4, 3, 2, and 1 clients, as shown in Figure 4b,e,h. Unbalanced Data Distribution and Size (UDDS): We use the Dirichlet distribution [48] D i r N ( β ) for data partitioning, where N is the number of clients and β determines the degree of non-independent homogeneous distribution, i.e., the smaller the value of β , the more unbalanced the data distribution. In this way, the scenario of uneven data distribution and data size between clients is simulated. For example, the D i r 10 ( 0.5 ) distribution for 10 clients is shown in Figure 4c,f,i. Across these simulated heterogeneous data scenarios, the weak client is significantly weaker than the strong client in terms of both data distribution and data size, and these simulated scenarios mimic, as much as possible, the different data heterogeneity scenarios that may occur in real IoT environments.

4.2. Baseline and Experimental Setup

The paper compares SplitLPF with state-of-the-art methods focused on personalization and fairness, including FedPer [49], which employs an architecture-based personalized federated learning method combining base and personalized layers; FedPAC [50], which achieves personalized federated learning through feature alignment and collaborative classifiers; FedCI [51], which employs client evaluation and uses evaluation outcomes as fairness-aware aggregation weights in federated learning; Ditto [35], an expandable federated multi-task learning framework that achieves fair federated learning through personalization; SplitGP [30], a personalized splitting learning method capable of capturing both generalization and personalization functionalities; and PFSL [29], a personalized fair splitting learning approach that combines transfer learning with lightweight personalization and cooperative fairness. Additionally, the comparison includes the benchmark federated learning method, FedAvg, and the federated splitting learning method, SplitFed [13]. For the model settings, a convolutional neural network (CNN) consisting of three convolutional layers and three fully connected layers was employed to process the FMNIST and EMNIST datasets. For the CIFAR-10 dataset, a ResNet18 neural network with a depth of 18 was used. In the split learning scenario, we offload most of the computational tasks of the model to the central server, and the number of front, middle, and back layers in the model structure is allocated based on the computational latency and computational complexity. Among them, the back part of the model retains the last layer of full connectivity with dropout operation to avoid model inference. SGD was utilized for model optimization during training, with an initial global learning rate set to η = 0.05 and a momentum change rate of 0.9. In terms of parameter settings, the training batch size was set to B = 128, the local iterations for all methods were set to e = 2, and the global iterations were set to E = 100. The default weight coefficient was set to α = 0.5, and the non-identical distribution degree β was set to 0.5. In the personalization settings, the local learning rate η c was dynamically adjusted based on the model accuracy achieved by each client during training.

4.3. Evaluation Index

To compare the performance of SplitLPF with other methods in terms of personalization and fairness, this paper employed a variety of evaluation metrics, including Test Accuracy and Personalized Accuracy: this metric is used to assess the model’s performance by comparing the test accuracy achieved during model training; Scaled Pearson Correlation (PC) [52]: this metric is used to compare the test accuracy φ obtained when clients train independently with the test accuracy θ obtained when they collaborate through fairness mechanisms in FL, and it is calculated as P C = C o v ( φ , θ ) σ φ σ θ , where C o v represents covariance and σ represents standard deviation; and Jain’s Fairness Index (JFI) [53]: this metric is used to compare the accuracy distributions of N clients on a local dataset and combine their accuracy scores into a vector s, which is denoted as J F I = i = 1 N s i 2 N i = 1 N s i 2 . In the literature [52,53], PC and JFI are used to assess method fairness. In addition, this paper calculates the standard deviation of test performance between clients according to Definition 1 to further measure fairness.

4.4. Analysis of Experimental Results

To validate the performance of SplitLPF, a comparison was made against other state-of-the-art methods across three datasets. The average test accuracy of model training under three distinct data heterogeneity scenarios is shown in Table 2.
From the Table 2, it is easy to see that SplitLPF presents satisfactory performance in terms of test accuracy for different datasets in all three cases of data heterogeneity. Notably, in all three datasets, adopting the UDDS scenario with Dirichlet distribution for data partitioning showcased higher test accuracy compared to focusing solely on uneven data distribution in UDD scenarios or differing data sizes in UDS scenarios. The reason for this is that the data distribution and the number of data in the UDDS scenario are appropriately compensated for each client, so that each client has access to a certain number of data, allowing the model to learn and adapt to different data features in a more comprehensive way, thus improving the accuracy of the model on the test set. In addition, in Table 2, the average test accuracy of SplitLPF exceeds that of the benchmark federated learning method FedAvg, and for the benchmark method SplitFed with the same splitting strategy, the test accuracy of SplitLPF is significantly higher than that of this benchmark as well. This indicates the superior performance of SplitLPF in the experimental evaluation. The enhanced performance of SplitLPF might stem from its utilization of optimized methods for aggregating model weights, which effectively integrate pertinent information from client devices, resulting in improved global model accuracy. In addition, consider that in the model design, SplitFed will make the accuracy rate somewhat lower because it needs noise to the uploaded gradients and labels to protect the privacy leakage during the model-training process.
Compared to other state-of-the-art methods, such as the FedPer, FedPAC, and SplitGP methods for personalized settings, it can be seen from Table 2 that SplitLPF all performs slightly higher than other personalized schemes. For instance, in the UDS scenario of the FMNIST dataset, SplitLPF demonstrates average test accuracy improvements of 0.98%, 0.33%, and 0.82% for FedPer, FedPAC, and SplitGP, respectively. Therefore, compared with existing methods, SplitLPF shows good performance under different datasets and data heterogeneity conditions. A horizontal comparison among existing personalized methods from Table 2 reveals that both FedPAC and SplitGP achieve higher average test accuracy than FedPer. This is primarily attributed to FedPAC employing effective feature alignment strategies and a robust classifier collaboration mechanism, which leads to a more consistent feature space among different clients, facilitating better collaboration and integration of information from various client sources. Meanwhile, SplitGP leverages controllable parameters to adjust the boundaries between personalization and generalization, effectively promoting collaborative training among clients. In contrast, while FedPer’s personalized layer offers a certain degree of client customization, its integration capability might be slightly lacking, resulting in a slightly lower average test accuracy compared to FedPAC and SplitGP methods. PFSL, a split learning approach that considers both personalization and fairness, demonstrates an average test accuracy close to that of SplitLPF and surpasses other baseline methods. This can be attributed to PFSL’s utilization of transfer learning strategies and freezing weights to achieve client personalization effectively, enabling it to adapt efficiently to various data heterogeneity scenarios.
In order to better demonstrate the effectiveness of personalization in SplitLPF, all the personalization methods as well as the benchmark methods were simulated and compared with SplitLPF. The experimental performance was evaluated on FMNIST and CIFAR-10 datasets, and the experimental results are shown in Figure 5. The experimental results are depicted in Figure 5. It is evident from Figure 5 that SplitLPF outperforms other methods in terms of personalized accuracy, while benchmark methods show relatively lower accuracy in comparison to personalized methods. Among the personalized methods considered, two split learning methods, PFSL and SplitGP, were assessed. Both adopted a model partitioning strategy that aimed to adapt better to resource-constrained devices. It can be observed from Figure 5 that both the methods perform well in terms of performance. PFSL leverages transfer learning, enabling the model to utilize abundant source domain data to assist in personalized training on resource-constrained weaker clients, thereby enhancing the model’s generalization ability. Simultaneously, model partitioning allows personalized training to focus more on specific tasks and data features at the client’s local level, effectively boosting accuracy. Meanwhile, SplitGP, through parameter control, adjusts the degree of personalization, enabling it to adapt to local data features on clients while maintaining generalization over global data. This balance allows the model to better cater to the requirements of different clients, consequently enhancing accuracy.
In order to validate the fairness achieved by SplitLPF through optimized model-aggregation weights, Table 3 presents the comparative results between SplitLPF and other baseline methods in terms of fairness metrics. Scaled Pearson Correlation (PC), Jain’s Fairness Index (JFI), and standard deviation (Std) are employed in Table 3 as indicators measuring fairness. Higher PC and JFI values alongside lower Std typically indicate a more balanced and fair distribution of benefits within the evaluation. Observing Table 3, it can be noted that SplitLPF achieves higher levels of fairness compared to other methods. Specifically, on the FMNIST and EMNIST datasets, SplitLPF achieves a PC (%) of over 90, indicating strong correlation. Moreover, the JFI (%) on all datasets exceeds 98. In addition, the standard deviation of SplitLPF on the EMNIST and CIFAR-10 datasets is kept at a low take. This means that the performance gap between different clients is small, and both achieve good performance.
In fairness-driven distributed learning approaches such as FedCI, Ditto, and PFSL, it is apparent from Table 3 that they show better performance than methods that focus on personalization. The reason behind this phenomenon likely lies in the different strategies adopted by these methods. In FedCI, fairness is achieved by fostering collaboration and information sharing, ensuring that each participant obtains reasonable gains in federated learning, thus enhancing overall fairness. Ditto achieves a higher level of fairness through a personalization mechanism and a reasonable distribution of proceeds. In contrast, PFSL maintains not only high levels of personalized accuracy but also superior fairness. This illustrates that in method design, personalization and fairness can be compatible. Taking inspiration from this, the present study optimizes both personalized models and fairness methods, aiming to better satisfy the dual requirements of personalization and fairness.
In the fairness mechanism, the contribution weights of dataset size and gradient direction disparity to model performance are adjusted by tuning the hyperparameter α . Different values of the hyperparameter α will be attempted to assess the impact of the contribution ratio of the dataset size and gradient direction disparity on SplitLPF’s performance. The experimental results are presented in Figure 6 and Figure 7. We can see that Figure 6 evaluates the impact of hyperparameters α on the experimental results on the FMNIST dataset, and Figure 7 evaluates the impact of hyperparameters on the experimental results on the CIFAR-10 dataset. We controlled the value of the hyperparameter a between [0–1], specifically taking values of 0.1, 0.2, 0.4, 0.5, 0.6, 0.8, and 0.9. For each of these values, 100 global iterations and two local iterations are conducted.
From the results in Figure 6 and Figure 7, it can be observed that with different data heterogeneity and different datasets, SplitLPF’s performance is significantly better when α is set to 0.4, 0.5, and 0.6 compared to when α is set to 0.1, 0.2, 0.8, and 0.9. Additionally, the performance is slightly lower when α is set to 0.1 and 0.9 compared to other values. This indicates that relying solely on either local dataset size or the disparity between gradient direction and the global gradient is not the optimal choice. Only by properly allocating the weights for both factors and ensuring an appropriate contribution ratio between them, it is more likely to obtain the best model weights and provide the optimal selection for model performance. Based on this, we optimize the model performance by adjusting the value of the hyperparameter α in order to better optimize the model performance.
In order to evaluate the resource consumption of SplitLPF, we compared several split learning methods that introduce certain communication overheads due to model partitioning. These include SplitFed, SplitGP, and PSFL, as well as the benchmark federated learning method FedAvg. Figure 8 illustrates the comparison of these methods in terms of time and communication overhead. From Table 1, it is evident that the communication in the federated learning approach mainly comes from downloading and uploading models, while the data communication in the split learning approach consists of transmitting intermediate results and synchronizing client models. Therefore, the communication cost incurred by FedAvg is significantly lower than that of SplitFed, SplitGP, PSFL, and SplitLPF.
In Figure 8b, the disparity in communication overhead resulting from training different models can be observed. The vertical coordinates on the left side of the figure represent the criteria for the communication overhead of different methods when training on the CNN model, and the vertical coordinates on the right side represent the criteria for the communication overhead of different methods when training on the ResNet18 model. Although SplitLPF incurs higher communication costs due to model partitioning, it employs a lightweight synthetic data transfer approach. This method utilizes smaller batch data to learn representative information from large-scale data effectively and can recover data information efficiently, thereby reducing some communication load. Observing Figure 8b, it is evident that compared to the existing personalized fair SL method PSFL, SplitLPF’s communication overhead is reduced by approximately 24%.
In Figure 8a, the comparison of time expenditures for training over 20 rounds across different methods is displayed. The power consumption of IoT devices intensifies as the training time increases, and some resource-constrained devices may power down midway due to load. Therefore, the selected training rounds were optimized to be the appropriate choice. PFSL and SplitLPF exhibit reduced time expenditures compared to other split learning methods. This is largely attributed to the fact that they employ the parallel training of clients and that the servers effectively support the parallel processing of client data. Moreover, SplitLPF’s time expenditure is slightly lower than PFSL, possibly due to the increased efficiency resulting from reduced communication data volume. In summary, compared to other split learning methodologies, SplitLPF demonstrates higher efficiency and lower communication costs in terms of resource consumption.

4.5. Extended Discussion

Privacy Analysis

In the proposed SplitLPF framework, each client undergoes specific personalized design to better adapt to the characteristics of different devices, enhancing model performance and adaptability. The central server (referred to as the third party) only serves the role of executing offloading computational tasks to participate in model training, without delving into the detailed client-side operations or accessing information about their data labels. Furthermore, when updating the client’s base model, a separate global learning rate is used for model updates. In contrast, the update of the client personalized model is controlled by the local learning rate. Clients locally store data labels and dynamic local learning rates. Thus, unlike traditional two-phase split learning, in this framework, the server lacks access to client data labels, significantly increasing the difficulty of model inversion attacks. In addition, unlike the SplitFed learning approach, SplitLPF does not require data to be noised during data transmission. Instead, lightweight synthetic data is used for transmission, effectively reducing the risk of data leakage. In specific scenarios, a public-private key encryption mechanism can be employed to further enhance data confidentiality, although this might lead to a slight increase in communication overhead.

5. Conclusions

In this paper, we propose SplitLPF, a personalized and fair split learning framework for resource-constrained IoT devices, which aims to address the difficulty of resource-constrained clients in traditional federated learning to efficiently train complex local models, as well as to deal with the challenges posed by the heterogeneity of data and the uneven distribution of benefits in IoT environments. In the IIoT domain, data heterogeneity arises from the diversity of sensing capabilities, environmental conditions, and data characteristics of different devices. Traditional SL training methods may lead to performance degradation in such a heterogeneous environment. On the other hand, benefit imbalance manifests as some smaller devices contributing significantly but not receiving proportional benefits during model training. This could lead to their withdrawal from cooperation, endangering the entire learning process. The SplitLPF framework achieves improved adaptation to specific application scenarios for resource-constrained IoT devices through model splitting, flexible task allocation, and support for personalized training. In terms of model aggregation, an optimized model weight aggregation method is employed to ensure fair benefit distribution by reasonably allocating the contribution proportion, thereby providing the best model performance selection, and to achieve collaborative fairness. This promotes a more balanced and sustainable collaboration. The experimental results demonstrate that the SplitLPF framework achieves higher accuracy through personalized training methods, even in scenarios with diverse data heterogeneity. Furthermore, the framework ensures collaboration fairness, providing a more stable foundation for cooperation among IoT devices. This research offers a novel approach to address distributed learning issues in the IoT, with the potential for significant breakthroughs in IoT applications.
However, the study also has limitations. For instance, SplitLPF still incurs certain communication overheads, particularly during the model segmentation and aggregation phases, thereby increasing the communication burden. Additionally, the SplitLPF method restricts the amount of shared information among different clients. This limitation might hinder the model from fully utilizing the data features of all clients, potentially reducing the model’s generalization ability. Therefore, the next step involves further addressing the communication overhead issues posed by SplitLPF. We plan to consider integrating more efficient communication technologies, such as the FedD3 framework [54], which distills instances through the dataset, requires only one-time communication, and greatly reduces communication costs. By introducing such efficient communication schemes, we aim to effectively tackle the communication cost issue while enhancing the efficiency and performance of distributed learning models, making SplitLPF adaptable to more different scenarios. In future work, we also plan to enhance the generalization ability of the model by improving the information exchange strategy. This improvement will be achieved by augmenting non-private portions of the model parameters, feature summaries, or gradient information. By doing so, the model can share partial information more extensively, fostering increased communication and information sharing among the involved parties.

Author Contributions

Conceptualization, X.C.; methodology, H.C.; software, H.C.; validation, H.C.; formal analysis, X.C.; investigation, L.P.; data curation, L.P. and Y.B.; writing—original draft preparation, H.C., X.C. and L.P.; writing—review and editing, H.C. and X.C.; and visualization, H.C. and Y.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number U20A20179.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The fashion-mnist dataset can be downloaded at https://github.com/zalandoresearch/fashion-mnist/ (accessed on 15 July 2023), the cifar10 dataset can be downloaded at https://www.cs.toronto.edu/kriz/cifar.html (accessed on 1 June 2023), and the emnist dataset can be downloaded at https://www.itl.nist.gov/iaui/vip/cslinks/EMNIST/ (accessed on 15 July 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoTInternet of Things
IIoTIndustrial Internet of Things
FLFederated learning
SLSplit learning
SplitLPFSplit Learning with Personalized Fairness
PSLParallel split learning
IIDIdentically and independently distributed
Non-IIDNon-independent and identically distributed
SGDStochastic gradient descent
FMNISTFashion-MNIST
EMNISTExtended MNIST
UDDUnbalanced Data Distributed
UDSUnbalanced Data Size
UDDSUnbalanced Data Distributed and Size
PCPearson Correlation
JFIJain’s Fairness Index
StdStandard deviation
CNNConvolutional neural network

Appendix A

Considering the simplified model training objective in this paper, the convergence proof is provided below. First, the global objective function before optimization is defined as F ( ω ) = k = 1 N n k F k ( ω ) , where k = 1 N n k = 1 . Let G k be the local accumulated stochastic gradients, represented as G k = 1 γ τ = 1 γ F k ( ω k ( t , τ ) ) . For each round, the global model update can be described as: w t + 1 w t = η k = 1 N p k G k , where k = 1 N p k = 1 . This analysis relies on assumptions related to local objective function smoothness and bounded gradient variance, as found in the literature [44,45,46,47].
Proposition A1. 
(Smoothness). For each local objective function that is L-Lipschitz smooth, it has for all k [ 1 , N ] : | | F ( ω k + 1 ) F ( ω k ) | | L | | ω k + 1 ω k | | , its equivalence condition is:
F ( ω k + 1 ) F ( ω k ) F ( ω k ) , ω k + 1 ω k + L 2 | | ω k + 1 ω k | | 2
Proposition A2. 
(Bounded gradient and bounded variance). For each client i, the local stochastic gradient is unbiased, i.e., E [ G k ( x ; ω k ) ] = F k ( ω ) . Moreover, it possesses a bounded variance, i.e., E [ G k ( x ; ω k ) F k ( ω ) ] σ 2 .
Proposition A3. 
(Bounded discrepancy and bounded scalar). For each local objective function F k ( ω ) , constants A and B greater than zero exist, such that: | F k ( ω ) F ( ω ) | 2 A d k + B , where d k represents the local-global discrepancy. The global objective function is bounded by F inf .
Proposition A4. 
(Conditional expectation). For a sequence of random matrices { M t } t T satisfying the zero conditional expectation E [ M t | M t 1 , M t 2 , . . . , M 1 ] = 0 and the independence assumption P [ M t | M t 1 , M t 2 , . . . , M 1 ] = P [ M t ] , there is E [ | | t = 1 T M t | | 2 ] = t = 1 T E [ | | M t | | 2 ] .
Proof of Theorem A1. 
Next, we will present the derivation of the convergence process. Given that the global gradient is smooth, based on proposition (A1), we can obtain:
F ( w t + 1 ) F ( w t ) F ( w t ) , ( w t + 1 w t ) + L 2 | | w t + 1 w t | | 2 = η F ( w t ) , k = 1 N p k G k C 1 + L η 2 2 | | k = 1 N p k G k | | 2
According to | | a | | 2 + | | b | | 2 | | a b | | 2 2 a , b = 0 , C 1 is further expressed as:
C 1 = F ( w t ) , k = 1 N p k G k = 1 2 | | F ( w t ) | | 2 + 1 2 | | k = 1 N p k G k | | 2 1 2 | | F ( w t ) k = 1 N p k G k | | 2 C 2
C 2 can be expressed as:
C 2 = | | F ( w t ) k = 1 N p k G k | | 2 = | | k = 1 N ( n k p k ) F k ( w t ) + k = 1 N p k ( F k ( w t ) G k ) | | 2
According to | | a + b | | 2 2 | | a | | 2 + 2 | | b | | 2 and Cauchy’s inequality, C 2 is further expressed as:
C 2 2 | | k = 1 N ( n k p k ) F k ( w t ) | | 2 + 2 | | k = 1 N p k ( F k ( w t ) G k ) | | 2 2 k = 1 N | | n k p k | | 2 k = 1 N | | F k ( w t ) | | 2 C 3 + 2 k = 1 N p k | | F k ( w t ) G k | | 2 C 4
According to | | a + b | | 2 2 | | a | | 2 + 2 | | b | | 2 , C 3 can be expressed as:
C 3 = k = 1 N | | F k ( w t ) | | 2 = k = 1 N | | ( F k ( w t ) F ( w t ) ) + F ( w t ) | | 2 2 k = 1 N | | F k ( w t ) F ( w t ) | | 2 + 2 k = 1 N | | F ( w t ) | | 2
According to proposition (A3), C 3 is further expressed as:
C 3 2 k = 1 N ( A d k + B ) + 2 k = 1 N | | F ( w t ) | | 2 2 N ( | | F ( w t ) | | 2 + B ) + 2 A k = 1 N d k
In C 4 , according to proposition (A4), C 4 can be expressed as:
C 4 = | | F k ( w t ) G k | | 2 = | | F k ( w t ) 1 γ τ = 0 γ F k ( w ( t , τ ) ) | | 2 = | | 1 γ τ = 0 γ ( F k ( w t ) F k ( w ( t , τ ) ) | | 2 = 1 γ τ = 0 γ | | F k ( w t ) F k ( w ( t , τ ) ) | | 2
According to proposition (A1), C 4 is further expressed as:
C 4 = L 2 γ τ = 0 γ | | w t w ( t , τ ) | | 2 = L 2 γ τ = 0 γ η 2 | | λ = 0 τ G k ( ω k ( t , λ ) ) | | 2 = L 2 η 2 γ τ = 0 γ | | λ = 0 τ ( G k ( ω k ( t , λ ) ) F k ( ω ( t , λ ) ) ) + F k ( ω ( t , λ ) ) | | 2 C 5
From | | a + b | | 2 2 | | a | | 2 + 2 | | b | | 2 and proposition (A1) and (A2), C 5 can be expressed as:
C 5 2 λ = 0 τ | | G k ( ω k ( t , λ ) ) F k ( ω ( t , λ ) ) | | 2 + 2 λ = 0 τ | | F k ( ω ( t , λ ) ) | | 2 2 τ σ 2 + 2 λ = 0 τ | | ( F k ( ω ( t , λ ) ) F k ( ω t ) ) + F k ( ω t ) | | 2 2 τ σ 2 + 4 λ = 0 τ | | F k ( ω ( t , λ ) ) F k ( ω t ) | | 2 + 4 τ | | F k ( ω t ) | | 2 2 τ σ 2 + 4 τ L 2 | | ω ( t , λ ) ω t | | + 4 τ | | F k ( ω t ) | | 2
Integrating | | ω ( t , λ ) ω t | | 2 from C 5 into C 4 , we can obtain:
| | w t w ( t , τ ) | | 2 = 2 η 2 τ σ 2 + 4 η 2 τ L 2 | | ω ( t , λ ) ω t | | 2 + 4 η 2 τ | | F k ( ω t ) | | 2
Transform it into:
| | w t w ( t , τ ) | | 2 = 1 ( 1 4 η 2 L 2 τ ) ( 2 η 2 τ σ 2 + 4 η 2 τ | | F k ( ω t ) | | 2 )
| | F k ( ω t ) | | 2 can be further derived from C 3 to obtain C 4 :
C 4 L 2 γ τ = 0 γ 1 ( 1 4 η 2 L 2 τ ) ( 2 η 2 τ σ 2 + 4 η 2 τ | | F k ( ω t ) | | 2 ) L 2 γ τ = 0 γ 1 ( 1 4 η 2 L 2 τ ) ( 2 η 2 τ σ 2 + 8 η 2 τ ( | | F ( w t ) | | 2 + A k = 1 N d k + B ) )
Let H = N k = 1 N | | n k p k | | 2 , K = L 2 γ τ = 0 γ 2 η 2 τ ( 1 4 η 2 L 2 τ ) ; then, C 2 can be expressed as:
C 2 4 H ( | | F ( w t ) | | 2 + B ) + 4 A H N k = 1 N d k + 2 k = 1 N p k K ( σ 2 + 4 ( | | F ( w t ) | | 2 + A k = 1 N d k + B ) ) 4 H ( | | F ( w t ) | | 2 + B ) + 4 A H N k = 1 N d k + + 2 K ( σ 2 + 4 ( | | F ( w t ) | | 2 + A k = 1 N p k d k + B ) )
Substituting C 2 into C 1 gives:
C 1 = 1 4 ( H + 2 K ) 2 | | F ( w t ) | | 2 + 1 2 | | k = 1 N p k G k | | 2 2 H B 2 A H N k = 1 N d k K ( σ 2 + 4 ( A k = 1 N p k d k + B ) )
Integrate into:
F ( w t + 1 ) F ( w t ) η [ 1 4 ( H + 2 K ) 2 | | F ( w t ) | | 2 2 H B 2 A H N k = 1 N d k K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) ] + η ( L η 1 ) 2 | | k = 1 N p k G k | | 2
when L η 1 < 0 , i.e., η < 1 / L , we can obtain:
F ( w t + 1 ) F ( w t ) η [ 1 4 ( H + 2 K ) 2 | | F ( w t ) | | 2 2 H B 2 A H N k = 1 N d k K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) ]
Translocation and according to proposition (A2) and (A3):
min t E | | F ( w t ) | | 2 = 1 T t = 0 T E | | F ( w t ) | | 2 2 1 4 ( H + 2 K ) [ 1 η T ( F ( w 0 ) F inf ) + 2 H B + 2 A H N k = 1 N d k + K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) ]
which completes the proof. □

Appendix B

Proof of Lemma A1. 
To simplify the analysis, the convergent bounds in Theorem 1 are represented as five terms, i.e, N 1 = 2 1 4 ( H + 2 K ) , N 2 = 1 η T ( F ( w 0 ) F inf ) , N 3 = 2 H B , N 4 = 2 A H N k = 1 N d k , N 5 = K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) , where H = N k = 1 N | | n k p k | | 2 , and it is evident that it consists of four parts related to the aggregation weights p k . When there is a greater disparity between the dataset sizes n k and the aggregation weights p k , the value of H increases, leading to larger values for N 3 and N 4 , which can often amplify the bounds. From N 4 and N 5 , we can observe that having different d k values might help in reducing the bounds. To minimize the bounds, we consider transforming the convergence bound expression, N 2 + N 3 + N 4 + N 5 / N 1 , into N 2 + N 3 + N 4 + N 5 ε N 1 , and then we tackle the following optimization problem to achieve minimization:
Y ( p k ) = 1 η T ( F ( w 0 ) F inf ) + 2 H B + 2 A H N k = 1 N d k + K ( σ 2 + 4 ( A k = 1 N p k d k + B ) ) 1 4 ( H + 2 K ) 2 ε
where k = 1 N p k = 1 and p k > 0 . For further optimization, we can set the derivative of Y ( p k ) to zero to achieve the minimization, as follows:
Y ( p k ) = 4 B N ( n k p k ) + 4 A ( n k p k ) k = 1 N d k + 4 A d k + 4 ε N ( n k p k ) = 0
Consolidating, we obtain:
p k = n k + A d k B N + A k = 1 N d k + ε N
This leads to:
p k n k + J d k
J is a constant greater than zero.
which completes the proof. □

References

  1. Tang, J.; Ding, X.; Hu, D.; Guo, B.; Shen, Y.; Ma, P.; Jiang, Y. FedRAD: Heterogeneous Federated Learning via Relational Adaptive Distillation. Sensors 2023, 23, 6518. [Google Scholar] [CrossRef] [PubMed]
  2. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Kiddon, C.; Konečnỳ, J.; Mazzocchi, S.; McMahan, B.; et al. Towards federated learning at scale: System design. Proc. Mach. Learn. Syst. 2019, 1, 374–388. [Google Scholar]
  3. Poirot, M.; Vepakomma, P.; Chang, K.; Kalpathy-Cramer, J.; Gupta, R.; Raskar, R. Split Learning for collaborative deep learning in healthcare. arXiv 2019, arXiv:1912.12115. [Google Scholar]
  4. Mendieta, M.; Yang, T.; Wang, P.; Lee, M.; Ding, Z.; Chen, C. Local learning matters: Rethinking data heterogeneity in federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8397–8406. [Google Scholar]
  5. Xu, X.; Lyu, L. A Reputation Mechanism Is All You Need: Collaborative Fairness and Adversarial Robustness in Federated Learning. arXiv 2020, arXiv:2011.10464. [Google Scholar]
  6. Gao, Y.; Kim, M.; Abuadbba, S.; Kim, Y.; Thapa, C.; Kim, K.; Camtep, S.A.; Kim, H.; Nepal, S. End-to-End Evaluation of Federated Learning and Split Learning for Internet of Things. In Proceedings of the 2020 International Symposium on Reliable Distributed Systems (SRDS), Shanghai, China, 21–24 September 2020. [Google Scholar]
  7. Yan, S.; Zhang, P.; Huang, S.; Wang, J.; Sun, H.; Zhang, Y.; Tolba, A. Node Selection Algorithm for Federated Learning Based on Deep Reinforcement Learning for Edge Computing in IoT. Electronics 2023, 12, 2478. [Google Scholar] [CrossRef]
  8. Sun, Y.; Ochiai, H.; Esaki, H. Decentralized deep learning for multi-access edge computing: A survey on communication efficiency and trustworthiness. IEEE Trans. Artif. Intell. 2021, 3, 963–972. [Google Scholar] [CrossRef]
  9. Xia, T.; Deng, Y.; Yue, S.; He, J.; Ren, J.; Zhang, Y. HSFL: An Efficient Split Federated Learning Framework via Hierarchical Organization. In Proceedings of the 2022 18th International Conference on Network and Service Management (CNSM), Thessaloniki, Greece, 31 October–4 November 2022; pp. 1–9. [Google Scholar]
  10. Duan, Q.; Hu, S.; Deng, R.; Lu, Z. Combined Federated and Split Learning in Edge Computing for Ubiquitous Intelligence in Internet of Things: State-of-the-Art and Future Directions. Sensors 2022, 22, 5983. [Google Scholar] [CrossRef] [PubMed]
  11. Shin, J.; Ahn, J.; Kang, H.; Kang, J. FedSplitX: Federated Split Learning for Computationally-Constrained Heterogeneous Clients. arXiv 2023, arXiv:2310.14579. [Google Scholar]
  12. Gawali, M.; Arvind, C.; Suryavanshi, S.; Madaan, H.; Gaikwad, A.; Bhanu Prakash, K.; Kulkarni, V.; Pant, A. Comparison of privacy-preserving distributed deep learning methods in healthcare. In Proceedings of the 25th Annual Conference on Medical Image Understanding and Analysis (MIUA 2021), Oxford, UK, 12–14 July 2021; pp. 457–471. [Google Scholar]
  13. Thapa, C.; Arachchige, P.C.M.; Camtepe, S.; Sun, L. Splitfed: When federated learning meets split learning. Proc. Aaai Conf. Artif. Intell. 2022, 36, 8485–8493. [Google Scholar] [CrossRef]
  14. Joshi, P.; Thapa, C.; Hasanuzzaman, M.; Scully, T.; Afli, H. Federated Split Learning with Only Positive Labels for resource-constrained IoT environment. arXiv 2023, arXiv:2307.13266. [Google Scholar]
  15. Pal, S.; Uniyal, M.; Park, J.; Vepakomma, P.; Raskar, R.; Bennis, M.; Jeon, M.; Choi, J. Server-side local gradient averaging and learning rate acceleration for scalable split learning. arXiv 2021, arXiv:2112.05929. [Google Scholar]
  16. Zhou, T.; Hu, Z.; Wu, B.; Chen, C. SLPerf: A Unified Framework for Benchmarking Split Learning. arXiv 2023, arXiv:2304.01502. [Google Scholar]
  17. Chen, X.; Li, J.; Chakrabarti, C. Communication and Computation Reduction for Split Learning using Asynchronous Training. In Proceedings of the 2021 IEEE Workshop on Signal Processing Systems (SiPS), Coimbra, Portugal, 19–21 October 2021. [Google Scholar]
  18. Oh, Y.N.; Lee, J.; Brinton, C.G.; Jeon, Y.S. Communication-Efficient Split Learning via Adaptive Feature-Wise Compression. arXiv 2023, arXiv:2307.10805. [Google Scholar]
  19. Lin, Z.; Zhu, G.; Deng, Y.; Chen, X.; Gao, Y.; Huang, K.; Fang, Y. Efficient Parallel Split Learning over Resource-constrained Wireless Edge Networks. arXiv 2023, arXiv:2303.15991. [Google Scholar]
  20. Xiong, Y.; Wang, R.; Cheng, M.; Yu, F.; Hsieh, C.J. Feddm: Iterative distribution matching for communication-efficient federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16323–16332. [Google Scholar]
  21. Tan, A.Z.; Yu, H.; Cui, L.; Yang, Q. Towards Personalized Federated Learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 9587–9603. [Google Scholar] [CrossRef]
  22. Duan, M.; Liu, D.; Chen, X.; Liu, R.; Tan, Y.; Liang, L. Self-Balancing Federated Learning With Global Imbalanced Data in Mobile Systems. IEEE Trans. Parallel Distrib. Syst. 2021, 32, 59–71. [Google Scholar] [CrossRef]
  23. Yang, M.; Wang, X.; Zhu, H.; Wang, H.; Qian, H. Federated learning with class imbalance reduction. In Proceedings of the 2021 29th European Signal Processing Conference (EUSIPCO), Dublin, Ireland, 23–27 August 2021; pp. 2174–2178. [Google Scholar]
  24. Zhang, J.; Hua, Y.; Wang, H.; Song, T.; Xue, Z.; Ma, R.; Cao, J.; Guan, H. GPFL: Simultaneously Learning Global and Personalized Feature Information for Personalized Federated Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  25. Marfoq, O.; Neglia, G.; Kameni, L.; Vidal, R. Personalized Federated Learning through Local Memorization. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  26. Liu, Y.; Kang, Y.; Xing, C.; Chen, T.; Yang, Q. A secure federated transfer learning framework. IEEE Intell. Syst. 2020, 35, 70–82. [Google Scholar] [CrossRef]
  27. Su, R.; Pang, X.; Wang, H. A novel parameter decoupling approach of personalised federated learning for image analysis. IET Comput. Vis. 2023, 17, 913–924. [Google Scholar] [CrossRef]
  28. Zhu, Z.; Hong, J.; Zhou, J. Data-free knowledge distillation for heterogeneous federated learning. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 12878–12889. [Google Scholar]
  29. Wadhwa, M.; Gupta, G.R.; Sahu, A.; Saini, R.; Mittal, V. PFSL: Personalized & Fair Split Learning with Data & Label Privacy for thin clients. arXiv 2023, arXiv:2303.10624. [Google Scholar]
  30. Han, D.J.; Kim, D.Y.; Choi, M.; Brinton, C.G.; Moon, J. Splitgp: Achieving both generalization and personalization in federated learning. In Proceedings of the IEEE INFOCOM 2023—IEEE Conference on Computer Communications, New York City, NY, USA, 17–20 May 2023; pp. 1–10. [Google Scholar]
  31. Chu, L.; Wang, L.; Dong, Y.; Pei, J.; Zhou, Z.; Zhang, Y. Fedfair: Training fair models in cross-silo federated learning. arXiv 2021, arXiv:2109.05662. [Google Scholar]
  32. Lin, X.; Xu, X.; Ng, S.K.; Foo, C.S.; Low, B. Fair yet Asymptotically Equal Collaborative Learning. In Proceedings of the 40th International Conference on Machine Learning (ICML 2023), Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  33. Hu, Z.; Shaloudegi, K.; Zhang, G.; Yu, Y. Federated Learning Meets Multi-Objective Optimization. IEEE Trans. Netw. Sci. Eng. 2022, 9, 2039–2051. [Google Scholar] [CrossRef]
  34. Cui, S.; Pan, W.; Liang, J.; Zhang, C.; Wang, F. Addressing algorithmic disparity and performance inconsistency in federated learning. In Proceedings of the Thirty-Fifth Annual Conference on Neural Information Processing Systems (NeurIPS 2021), Virtual, 6–14 December 2021; Volume 34, pp. 26091–26102. [Google Scholar]
  35. Li, T.; Hu, S.; Beirami, A.; Smith, V. Ditto: Fair and robust federated learning through personalization. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 6357–6368. [Google Scholar]
  36. Lyu, L.; Xu, X.; Wang, Q.; Yu, H. Collaborative fairness in federated learning. In Federated Learning: Privacy and Incentive; Springer: Berlin/Heidelberg, Germany, 2020; pp. 189–204. [Google Scholar]
  37. Xu, X.; Lyu, L.; Ma, X.; Miao, C.; Foo, C.S.; Low, B.K.H. Gradient driven rewards to guarantee fairness in collaborative machine learning. In Proceedings of the Thirty-Fifth Annual Conference on Neural Information Processing Systems (NeurIPS 2021), Virtual, 6–14 December 2021; Volume 34, pp. 16104–16117. [Google Scholar]
  38. Fan, Z.; Fang, H.; Zhou, Z.; Pei, J.; Friedlander, M.P.; Zhang, Y. Fair and efficient contribution valuation for vertical federated learning. arXiv 2022, arXiv:2201.02658. [Google Scholar]
  39. Sattler, F.; Wiedemann, S.; Muller, K.R.; Samek, W. Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3400–3413. [Google Scholar] [CrossRef] [PubMed]
  40. Hu, S.; Goetz, J.; Malik, K.; Zhan, H.; Liu, Z.; Liu, Y. Fedsynth: Gradient compression via synthetic data in federated learning. arXiv 2022, arXiv:2204.01273. [Google Scholar]
  41. Goetz, J.; Tewari, A. Federated learning via synthetic data. arXiv 2020, arXiv:2008.04489. [Google Scholar]
  42. Zhou, Y.; Shi, M.; Li, Y.; Sun, Y.; Ye, Q.; Lv, J. Communication-efficient Federated Learning with Single-Step Synthetic Features Compressor for Faster Convergence. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 17–24 June 2023; pp. 5031–5040. [Google Scholar]
  43. Dankar, F.K.; Madathil, N. Using Synthetic Data to Reduce Model Convergence Time in Federated Learning. In Proceedings of the 2022 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), Istanbul, Turkey, 10–13 November 2022; pp. 293–297. [Google Scholar]
  44. Avdiukhin, D.; Kasiviswanathan, S. Federated learning under arbitrary communication patterns. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 425–435. [Google Scholar]
  45. Yu, H.; Jin, R.; Yang, S. On the linear speedup analysis of communication efficient momentum SGD for distributed non-convex optimization. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7184–7193. [Google Scholar]
  46. Murata, T.; Suzuki, T. Bias-variance reduced local sgd for less heterogeneous federated learning. arXiv 2021, arXiv:2102.03198. [Google Scholar]
  47. Ye, R.; Xu, M.; Wang, J.; Xu, C.; Chen, S.; Wang, Y. FedDisco: Federated Learning with Discrepancy-Aware Collaboration. arXiv 2023, arXiv:2305.19229. [Google Scholar]
  48. Zhang, J.; Li, Z.; Li, B.; Xu, J.; Wu, S.; Ding, S.; Wu, C. Federated learning with label distribution skew via logits calibration. In Proceedings of the International Conference on Machine Learning, Virtual, 25–22 July 2022; pp. 26311–26329. [Google Scholar]
  49. Arivazhagan, M.G.; Aggarwal, V.; Singh, A.K.; Choudhary, S. Federated learning with personalization layers. arXiv 2019, arXiv:1912.00818. [Google Scholar]
  50. Xu, J.; Tong, X.Y.; Huang, S.L. Personalized Federated Learning with Feature Alignment and Classifier Collaboration. arXiv 2023, arXiv:2306.11867. [Google Scholar]
  51. Song, T.; Tong, Y.; Wei, S. Profit allocation for federated learning. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 2577–2586. [Google Scholar]
  52. Shi, Y.; Yu, H.; Leung, C. Towards fairness-aware federated learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–17. [Google Scholar] [CrossRef]
  53. Jiang, M.; Roth, H.R.; Li, W.; Yang, D.; Zhao, C.; Nath, V.; Xu, D.; Dou, Q.; Xu, Z. Fair Federated Medical Image Segmentation via Client Contribution Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16302–16311. [Google Scholar]
  54. Song, R.; Liu, D.; Chen, D.Z.; Festag, A.; Trinitis, C.; Schulz, M.; Knoll, A. Federated learning via decentralized dataset distillation in resource-constrained edge environments. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023; pp. 1–10. [Google Scholar]
Figure 1. Comparison of distributed learning. (a) Federated Learning; (b) Parallel Split Learning; (c) SplitFed learning; (d) SGLR.
Figure 1. Comparison of distributed learning. (a) Federated Learning; (b) Parallel Split Learning; (c) SplitFed learning; (d) SGLR.
Sensors 24 00088 g001
Figure 2. SplitLPF framework.
Figure 2. SplitLPF framework.
Sensors 24 00088 g002
Figure 3. Experimental scenario setup in IoT simulated environment.
Figure 3. Experimental scenario setup in IoT simulated environment.
Sensors 24 00088 g003
Figure 4. Three non-IID forms of FMNIST, CIFAR-10, and EMNIST; (a) FMNIST UDD; (b) FMNIST UDS; (c) FMNIST UDDS; (d) CIFAR-10 UDD; (e) CIFAR-10 UDS; (f) CIFAR-10 UDDS; (g) EMNST UDD; (h) EMNST UDS; (i) EMNST UDDS.
Figure 4. Three non-IID forms of FMNIST, CIFAR-10, and EMNIST; (a) FMNIST UDD; (b) FMNIST UDS; (c) FMNIST UDDS; (d) CIFAR-10 UDD; (e) CIFAR-10 UDS; (f) CIFAR-10 UDDS; (g) EMNST UDD; (h) EMNST UDS; (i) EMNST UDDS.
Sensors 24 00088 g004
Figure 5. Personalized accuracy of FMNIST and CIFAR10 under different non-IID data; (a) FMNIST UDD; (b) FMNIST UDS; (c) FMNIST UDDS; (d) CIFAR-10 UDD; (e) CIFAR-10 UDS; (f) CIFAR-10 UDDS.
Figure 5. Personalized accuracy of FMNIST and CIFAR10 under different non-IID data; (a) FMNIST UDD; (b) FMNIST UDS; (c) FMNIST UDDS; (d) CIFAR-10 UDD; (e) CIFAR-10 UDS; (f) CIFAR-10 UDDS.
Sensors 24 00088 g005
Figure 6. Model accuracy and loss of SplitLPF under different α conditions (based on FMNIST dataset); (a) Performance of FMNIST under UDD; (b) Performance of FMNIST under UDS.
Figure 6. Model accuracy and loss of SplitLPF under different α conditions (based on FMNIST dataset); (a) Performance of FMNIST under UDD; (b) Performance of FMNIST under UDS.
Sensors 24 00088 g006
Figure 7. Model accuracy and loss of SplitLPF under different α conditions (based on CIFAR-10 dataset); (a) Performance of CIFAR-10 under UDD; (b) Performance of CIFAR-10 under UDS.
Figure 7. Model accuracy and loss of SplitLPF under different α conditions (based on CIFAR-10 dataset); (a) Performance of CIFAR-10 under UDD; (b) Performance of CIFAR-10 under UDS.
Sensors 24 00088 g007
Figure 8. Comparison of time overhead and communication overhead between different methods; (a) Time overhead; (b) Communication overhead.
Figure 8. Comparison of time overhead and communication overhead between different methods; (a) Time overhead; (b) Communication overhead.
Sensors 24 00088 g008
Table 1. Total cost analysis of the different approaches for one global epoch.
Table 1. Total cost analysis of the different approaches for one global epoch.
MethodsCommunication Cost Per ClientTotal Communication CostTotal Model Training Time
FL 2 | W | 2 N | W | T + 2 | W | R + T a g g
SL 2 P C N + 2 β | W | 2 P C + 2 β N | W | T + 2 β N | W | R + 2 P C R
SplitFed 2 P C N + 2 β | W | 2 P C + 2 β N | W | T + 2 β | W | R + 2 P C N R + T a g g
SplitLPF 4 S N 4 S T + 2 T 1 + 4 S N R + T a g g
Table 2. Average test accuracy (%) of each method with different non-IID data.
Table 2. Average test accuracy (%) of each method with different non-IID data.
DatasetsFMNIST EMNISTCIFAR-10
MethodUDD UDS UDDS UDD UDS UDDS UDD UDS UDDS
FedAvg86.1480.6991.7175.4369.5083.4761.4040.8271.56
FedPer86.1981.9891.7276.1270.2482.9261.9740.9471.72
FedPAC86.7882.3792.1476.3870.5885.6462.7343.7772.16
FedCI86.9682.2292.0976.3970.7985.6862.0543.1172.12
Ditto86.7081.2891.6075.8270.2385.4861.7642.9272.11
SplitFed85.6778.7788.4467.6366.2381.5259.6140.0270.84
SplitGP86.3082.1391.9376.2370.7285.0961.6642.5971.67
PFSL87.4082.4392.3276.4370.9886.1462.2246.1472.48
SplitLPF87.2882.9592.8576.8871.8687.1562.9646.9772.88
Table 3. The fairness index of each method is compared under different datasets.
Table 3. The fairness index of each method is compared under different datasets.
DatasetsFMNISTEMNISTCIFAR-10
MethodPCJFIStdPCJFIStdPCJFIStd
FedAvg85.1593.9217.4787.6997.547.9173.8793.896.47
FedPer85.9494.4717.0290.5397.737.8873.5794.936.30
FedPAC85.3194.0516.8589.8197.206.7473.4393.376.19
FedCI88.3595.2016.1592.0797.725.2475.2095.184.66
Ditto87.9895.4916.4691.5697.956.2074.5496.884.98
SplitFed72.7290.9620.3655.4495.0033.1969.6489.798.37
SplitGP78.7491.5517.5577.4696.1515.1572.5992.967.33
PFSL89.4396.1615.0394.5197.963.0676.4595.453.27
SplitLPF91.6998.7614.7391.2198.322.6478.5798.393.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, H.; Chen, X.; Peng, L.; Bai, Y. Personalized Fair Split Learning for Resource-Constrained Internet of Things. Sensors 2024, 24, 88. https://doi.org/10.3390/s24010088

AMA Style

Chen H, Chen X, Peng L, Bai Y. Personalized Fair Split Learning for Resource-Constrained Internet of Things. Sensors. 2024; 24(1):88. https://doi.org/10.3390/s24010088

Chicago/Turabian Style

Chen, Haitian, Xuebin Chen, Lulu Peng, and Yuntian Bai. 2024. "Personalized Fair Split Learning for Resource-Constrained Internet of Things" Sensors 24, no. 1: 88. https://doi.org/10.3390/s24010088

APA Style

Chen, H., Chen, X., Peng, L., & Bai, Y. (2024). Personalized Fair Split Learning for Resource-Constrained Internet of Things. Sensors, 24(1), 88. https://doi.org/10.3390/s24010088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop