Next Article in Journal
Health Risk Assessment Using Machine Learning: Systematic Review
Previous Article in Journal
Precise Orientation Estimation for Rotated Object Detection Based on a Unit Vector Coding Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Rate-Splitting-Based Distributed Edge Computing via Multi-Group Information Recycling

The Electronic Information School, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(22), 4403; https://doi.org/10.3390/electronics13224403
Submission received: 28 September 2024 / Revised: 2 November 2024 / Accepted: 5 November 2024 / Published: 11 November 2024
(This article belongs to the Special Issue Distributed Intelligence Technologies for Smart Cities of the Future)

Abstract

:
To address the straggling effect in distributed edge computing, existing methods often introduce extra computation or communication costs. Recently, information recycling has emerged as an efficient solution that avoids such extra overhead. Nonetheless, the performance of the existing information recycling-assisted rate-splitting-based distributed edge computing scheme significantly hinges on the single common stream, whose rate is limited by the edge node (EN) with the worst channel quality. To this end, a multi-group information recycling scheme for rate-splitting-based distributed edge computing is proposed, where the ENs are clustered into multiple groups, each with its own common stream. In this way, the proposed scheme alleviates the communication limitation of a single common stream, thereby boosting the information recycling mechanism for promoted computation collaboration. A K-medoids-based grouping algorithm is designed for the general multi-antenna case, and a more efficient continuous partitioning-based grouping algorithm is proposed for the special case of single-antenna. Besides, a convex–concave procedure-based algorithm is developed to solve the corresponding latency optimization problem. Simulations demonstrate that the proposed scheme offers significant and robust advantages across various communication and computation conditions. In the considered scenarios, the proposed scheme can substantially reduce the processing latency by up to 51.0% compared to the conventional information recycling scheme.

1. Introduction

In recent years, edge computing has played an increasingly vital role in enhancing the quality of computational services [1,2]. However, the mounting scale of edge intelligence applications, (e.g., big data, augmented reality, and artificial intelligence) is pushing the limits of edge computing. These large-scale and computationally intensive tasks have rendered traditional single-node edge computing insufficient, entailing distributed edge computing, which has emerged as a promising computation paradigm [3,4,5]. This approach partitions and distributes the original large-scale and intensive tasks among multiple edge nodes (ENs), exploiting their computational resources for faster processing.
Nonetheless, the straggling effect presents a notable challenge in distributed edge computing. Specifically, some ENs, known as stragglers, may experience substantial delays or even fail to complete their tasks due to various factors, including hardware heterogeneity, imbalanced workload allocation, I/O contentions, and hardware failures [6,7,8]. This significantly degrades the overall performance, as the system needs to wait for feedback from all the ENs in distributed edge computing [9]. To alleviate the straggling issues, several strategies have been considered in the literature. One typical approach is task replication [10,11,12], where redundant copies of the original task are distributed across multiple ENs, and the execution delay is determined by the fastest EN despite some stragglers. Besides task replication, coded computing [8] has recently been advocated, where the partitioned subtasks are encoded into a set of coded subtasks via coding theory [13]. Thanks to the well-designed coding redundancy, the mobile user can reconstruct the original task result using the feedback from only a subset of ENs, thus avoiding the need to wait for stragglers. In addition, workload reallocation [14,15] is another commonly adopted method, which allows slower ENs to re-transmit part of their remaining workloads to other faster ENs to minimize the overall processing latency. Nevertheless, despite the effectiveness of these pioneering works in mitigating the straggling effect, they often introduce redundant computation or require additional data transmission, which may still limit the overall performance.
To handle straggling without incurring extra computation and communication costs, a physical-layer mechanism known as information recycling has been recently proposed [16], in view of the widespread employment of the spectrum-efficient non-orthogonal transmission techniques (e.g., rate-splitting multiple access (RSMA) [17,18] and power-domain non-orthogonal multiple access (NOMA) [19]) in distributed edge computing [20,21,22,23,24]. The key insight of information recycling is that, under many non-orthogonal transmission techniques involving successive interference cancellation (SIC), each EN has to decode part of the information intended for other ENs [25,26]. Instead of directly discarding this part of the information after interference cancellation, the information recycling mechanism proactively recycles this information (for free), by exploiting the SIC procedure. With the recycled information, faster ENs that complete their own task earlier can help execute the corresponding part of the task for the straggling ENs, instead of waiting for the straggling ENs, thereby mitigating the straggling effect without extra overhead. A concrete example of the information recycling mechanism has been proposed for rate-splitting-based distributed edge computing systems [16]. The performance of this conventional scheme heavily hinges on the single common stream in rate-splitting. Specifically, a larger common stream can increase the portion of tasks available for information recycling and thus promote computation collaboration. However, the rate of the common stream is fundamentally limited by the EN with the worst channel quality. This constraint undermines the benefits of information recycling, since enforcing a large common stream may aggravate the impact of the rate limitation on transmission, leading to sub-optimal communication performance. To the best of our knowledge, developing a new distributed edge computing scheme that can overcome the communication limitation of the single common stream remains an open challenge.
To this end, a novel scheme, coined as multi-group information recycling for rate-splitting-based distributed edge computing, is proposed in this work. Particularly, it will cluster the ENs into multiple groups and offer one common stream for each group, where collaborative computation can still be enabled through information recycling. Since each common stream is decoded by a smaller number of ENs, the communication bottleneck of the single common stream can be alleviated, thereby facilitating the common stream transmission. As a result, the information recycling mechanism can be further boosted, promoting computation collaboration among the ENs for processing latency reduction. The major contributions of this work are summarized as follows:
  • A multi-group information recycling scheme for rate-splitting-based distributed edge computing is proposed in this work for processing latency reduction;
  • To obtain a good EN grouping strategy, a K-medoids-based grouping algorithm is designed by jointly considering the channel alignment and the channel strength for the general multi-antenna case, and a more efficient continuous partitioning-based grouping algorithm is proposed for the special case of single-antenna;
  • By exploiting the inherent structure of the corresponding latency optimization problem and bounding via difference of convex (DC) programming, a convex–concave procedure (CCCP)-based algorithm is developed to find a reasonably good solution.
The remainder of this paper is organized as follows. Related works are briefly reviewed in Section 2, and the system model and the conventional information recycling mechanism are presented in Section 3. In Section 4, the proposed scheme is presented, and the corresponding latency optimization problem is formulated. The latency optimization algorithms are developed in Section 5. Simulation results are presented in Section 6. Finally, conclusions and future works are discussed in Section 7.
Notations: · H denotes the Hermitian transpose. C a × b represents a complex matrix with dimension a × b . CN a , b represents the distribution of a circularly symmetric complex Gaussian random vector with mean vector a and covariance matrix b . E · represents expectation. R · obtains the real part of a complex number. · is the absolute value. tr · is the trace.

2. Related Work

Several methods have been considered in the literature to mitigate the straggling effect in distributed edge computing, including task replication [10,11,12], coded computing [27,28,29], as well as workload reallocation [14,15]. For example, in [10], the original computation task is duplicated and assigned to multiple ENs so that the execution latency is determined by the fastest EN rather than the stragglers. The tradeoff between expected completion time and task replication ratio is explored in [11] to identify the optimal replication strategy for various computation straggling models. In [12], a balanced grouping replication strategy is further proposed, focusing on optimizing the redundancy level to minimize average job compute time while preserving task completion time predictability. In addition to task replication, coded computing has gained traction for the ability to mitigate straggling via coding theory [13]. Specifically, the maximum distance separable (MDS) codes are used in [27] for distributed task offloading. It is shown that MDS-coded offloading can effectively mitigate the straggling effect as the mobile user can reconstruct the original task result using feedback from only a subset of fast ENs. Besides, entangled polynomial codes are proposed in [28] to achieve the optimal tradeoff between the recovery threshold (i.e., the minimum number of ENs required to recover the original task) and the per-EN computation workload through different matrix partitioning. In [29], the impact of the coding parameter selection under different task execution models is investigated to alleviate the straggling effect and minimize the processing latency. Moreover, workload reallocation is another widely employed method for straggler mitigation. For instance, in [14], when a faster EN completes its task and becomes idle, part of the straggling ENs’ remaining workloads can be re-transmitted to this faster EN, thereby rebalancing workload and addressing straggling. In [15], by detecting straggling ENs continuously in the computation process, the mobile user can promptly reassign tasks to faster ENs, thereby minimizing idle time and improving overall system performance. Nevertheless, while these methods effectively mitigate the straggling effect, their performance remains constrained. Specifically, task replication and coded computing incur increased communication and computational costs, because replicated or coded subtasks are often larger than the subtasks in traditional distributed computing due to the introduced task redundancy. Besides, workload reallocation necessitates task data retransmission, which brings extra communication costs and limits the overall performance.
On the other hand, significant research has been conducted on strengthening communication in distributed edge computing using non-orthogonal transmission techniques. For instance, in [22], the computation subtasks are offloaded to the ENs via NOMA, and the processing latency is minimized by jointly optimizing transmission duration and workload allocation. In [23], the benefits of NOMA are investigated for both uplink and downlink transmissions in mobile edge computing. In [24], the total multi-user processing latency of an RSMA-based distributed edge computing system is optimized and the superiority of RSMA in serving distributed edge computing is manifested as compared to NOMA. However, most of these pioneering works mainly consider how non-orthogonal transmission strengthens the communication in distributed edge computing, while the straggling issue is largely ignored.
To handle straggling without incurring extra computation and communication costs, the information recycling mechanism is proposed in [16], which can be integrated into the non-orthogonal transmission techniques. In particular, by recycling the task information intended for other ENs in the decoded common stream, the faster ENs can help execute the corresponding part of the task for the slower ENs, thereby mitigating the straggling effect. However, the effectiveness of the existing information recycling-assisted distributed edge computing scheme highly hinges on the single common stream in rate-splitting and the rate of the common stream is inherently limited by the EN with the poorest channel quality. This constraint diminishes the advantages of information recycling, as dedicating resources to a single common stream can promote collaborative computation but may also degrade overall communication performance. In contrast to these pioneering works, by clustering ENs into multiple groups and transmitting several common streams, the proposed multi-group information recycling scheme can alleviate the single common stream’s communication limitation to facilitate the transmission of the common streams, thus promoting the information recycling mechanism for enhanced computation collaboration among the ENs.

3. System Model

In this section, the system model of rate-splitting-based distributed edge computing with stragglers will be presented, together with the conventional information recycling mechanism.

3.1. System Model

Consider a distributed edge computing system consisting of a mobile user and K ENs indexed by K = { 1 , 2 , , K } . The mobile user wirelessly offloads its computation-intensive task W to these K ENs for faster processing. Specifically, the computation task W can be divided into K independent subtasks { W 1 ,   W 2 ,   ,   W K } (Take (large-scale) matrix multiplication, a fundamental operation in many machine learning algorithms [30], as an example of the practical application. The original matrix multiplication task Fb will be partitioned into K sub-tasks { F 1 b ,   F 2 b ,   ,   F K b } with F T = F 1 T ,   F 2 T ,   ,   F K T ). Each sub-task W k will be offloaded to EN k for processing (As in the existing literature [1,23,31,32], it is assumed that the size of the computing results is much smaller compared to the offloaded subtasks, making the downlink transmission latency negligible).

3.1.1. Communication Phase

For better spectrum efficiency in communications, it is assumed that the mobile user employs RSMA to offload the subtasks to the K ENs during the communication phase. Specifically, each subtask W k is split into a common part W k c and a private part W k p (In the matrix multiplication, this corresponds to partition F k along the rows into two parts, F k c and F k p , where F k T = F k c , T , F k c , T ). In the wireless physical layer, all the common parts { W 1 c ,   W 2 c ,   ,   W K c } are encoded into a single common stream s c , while each private part W k p is encoded into a private stream s k p . By employing superposition coding, the transmitted signal x from the mobile user is given by
x = w c s c + k K w k p s k p ,
where w c C N t × 1 and w k p C N t × 1 are the precoding vectors for the common stream s c and the private stream s k p , respectively, with N t denoting the number of antennas of the mobile user. The received wireless signal y k at EN k can be expressed as
y k = h k H x + n k ,
where h k C N t × 1 represents the gain of the wireless channel between the mobile user and EN k, and n k CN ( 0 , σ 2 ) is the corresponding additive white Gaussian noise. According to the decoding rules of rate-splitting [18], each EN k first decodes the common stream s c by treating the interference from all private streams as noise and then subtracts it from the received signal using SIC. Thereafter, EN k decodes its private stream s k p by treating the remaining interference from the other private streams as noise. Hence, by plugging (1) into (2), the signal-to-interference-plus-noise ratios (SINRs) of the common stream s c and the private stream s k p to EN k are, respectively, given by
γ k c = h k H w c 2 j K h k H w j p 2 + σ 2 ,   k K ,
γ k p = h k H w k p 2 j K k h k H w j p 2 + σ 2 ,   k K .
With the above SINRs, the rates R k p ’s of the private steams should admit
R k p B log 2 ( 1 + γ k p ) , k K ,
where B is the system bandwidth. To ensure that the common stream s c can be successfully decoded by all ENs, the rate R c of the common stream should admit
R c min k K B log 2 ( 1 + γ k c ) .
As the rate R c of the common stream is shared among all ENs to offload { W 1 c ,   W 2 c ,   ,   W K c } , it has k = 1 K R k c = R c , where R k c is the uplink transmission rate of the common part W k c of subtask W k . Hence, the overall uplink transmission rate R k to EN k is given by
R k = R k c + R k p .

3.1.2. Computation Phase

In the computation phase, each EN k processes its received offloaded subtask W k . To account for the straggling effect [8,9], the computation rate C k of EN k can be modeled as follows [29]
C k = C ¯ Δ k .
Here, C ¯ = f e β denotes the maximum computation rate of the ENs, where f e is the maximum processor frequency, and β represents the number of CPU cycles required to process one bit of data. Δ k is a non-negative random variable representing the degradation in computation capability due to the straggling effect. The variables Δ k k = 1 K are assumed to be independent and identically distributed (i.i.d.).

3.2. Conventional Information Recycling

The idea of conventional information recycling for rate-splitting-based distributed edge computing [16] is introduced in the sequel. According to the rate-splitting transmission principle [18], each EN k will decode the task messages W 1 c , , W k 1 c , W k + 1 c , , W K c intended for other ENs when decoding the common stream s c to obtain its own message W k . In conventional distributed computing, after decoding, these messages will be discarded directly. As a result, each EN can only compute its own subtask W k , and the faster EN that completes its own task earlier can do nothing but wait for the straggling ENs. As shown in Figure 1a, both EN 1 and EN 3 can do nothing but wait for the straggling EN 3 and remain idle, which undoubtedly wastes the computation capability of the system. In contrast, the information recycling mechanism enables the ENs to proactively recycle the messages W 1 c , , W k 1 c , W k + 1 c , , W K c , as shown in Figure 1b. In this way, each EN will have full knowledge of the common parts { W 1 c , W 2 c , , W K c } of all the subtasks W k k = 1 K . When a certain EN i is straggling, another faster EN j can help compute (the common part of) its subtask based on the recycled information W i c .
The specific process of collaborative computing for the common parts is introduced as follows and summarized in Algorithm 1. Each common part task W k c is divided into A k mini-tasks { W k , u c } u = 1 A k , where A k = D k c D m . In the matrix multiplication, each common part F k c is further partitioned along the rows into A k smaller parts F k , 1 c , F k , 2 c , , F k , A k c to form mini-tasks F k , u c b u = 1 A k , where F k c , T = F k , 1 c , T , F k , 2 c , T , , F k , A k c , T . Here, D k c and D m are the sizes of the task W k c and each mini-task, respectively, (note that D m is a parameter determined during the engineering implementation process; therefore, its specifics are beyond the scope of this paper). The system also maintains an ordered task queue for each EN k, which is initialized as { W k , u c } u = 1 A k . Each EN sequentially processes the mini-tasks according to their queuing order. Upon completing the mini-task W k , u c , EN j sends a finishing signal S k , u fin to inform the mobile user. Besides, whenever a certain EN j completes all its assigned mini-tasks and the overall computation task has not yet been finished, it can assist other straggling ENs in computing their mini-tasks by information recycling. Specifically, if straggling EN i has the longest unfinished task queue at the current time, and task W k , v c at the end of its task queue, the mobile user will send a recycling signal S k , v rec to EN j and a canceling signal S k , v can to EN i. The signals S k , v rec and S k , v can direct EN j to begin computing W k , v c using information recycling and EN i to remove W k , v c from its task queue, respectively. This collaboration continues until the overall computation task is completed, allowing fast ENs and slow ENs to collaboratively compute the common parts { W 1 c , W 2 c , , W K c } . An illustrative example is depicted in Figure 2. With the information recycling mechanism enabling computation collaboration, theoretically, the computation capabilities of all ENs can be captured for the computation of W 1 c , W 2 c , , W K c and the overall computation rate C rec can be made to
C rec = k = 1 K C k .
Algorithm 1 The collaborative computing process of information recycling
1:
Initialize: For each EN k, divide the common part W k c into mini-tasks and initialize the ordered task queue { W k , u c } u = 1 A k ;
2:
Each EN starts to sequentially process the mini-tasks according to their queuing order;
3:
while The overall computation task is not completed do
4:
  for  j = 1 K  do
5:
   if EN j completes a mini-task W k , u c  then
6:
     EN j sends a finishing signal S k , u fin to the mobile user;
7:
   end if
8:
   if EN j receives a recycling signal S k , v rec  then
9:
     EN j adds W k , v c to its task queue and begins computing W k , v c via information recycling;
10:
   end if
11:
   if EN j receives a canceling signal S k , v can  then
12:
     EN j removes W k , v c from its task queue;
13:
   end if
14:
  end for
15:
  if The mobile user receives a finishing signal S k , u fin from EN j and determines that EN j has completed all assigned mini-tasks then
16:
   Finding the EN i with the longest unfinished task queue, and task W k , v c at the end of its task queue, the mobile user sends a recycling signal S k , v rec to EN j and a canceling signal S k , v can to EN i;
17:
  end if
18:
end while

4. Proposed Multi-Group Information Recycling

4.1. Multi-Group Information Recycling Scheme

The performance of the conventional information recycling assisted distributed edge computing highly depends on the single common stream [16]. Specifically, when more resources are allocated to the common stream, the common parts W 1 c , W 2 c , , W K c will occupy a larger portion of the task, thus facilitating the information recycling mechanism for straggler mitigation. Nonetheless, enforcing a too-strong common stream may not be necessarily good for communication, as the rate of the common stream s c is restricted by the EN with the worst channel quality. This may offset the performance gain brought by the information recycling mechanism. For this reason, a new information recycling scheme that can overcome the communication limitation of the single common stream is required. To this end, a multi-group information recycling scheme is proposed for rate-splitting-based distributed edge computing below.
Assume that the K ENs are divided into G mutually-exclusive groups and each group g contains K g ENs with g G K g = K . For the communication phase, as shown in Figure 3a, the common parts W k c | k K g of the ENs in each group g are combined as W K g c and encoded into the common stream s K g c in the wireless layer. Meanwhile, the private part W k p is still encoded into s k p . By employing superposition coding, the transmitted signal x from the mobile user is given by
x = g G w K g c s K g c + k K w k p s k p ,
where w K g c C N t × 1 is the precoding vector for the common stream s K g c . By plugging (10) into (2), the received signal y k at EN k is given by
y k = g G h k H w K g c s K g c + j K h k H w k p s j p + n k .
The decoding procedure of EN k in group g (i.e., k K g ) is performed as follows: EN k in group g decodes its common stream s K g c by treating other common streams and all private streams as interference and removes it from the received signal using SIC. Thereafter, EN k decodes its private stream s k p by treating the other common streams and private streams as interference. Accordingly, the signal-to-interference-plus-noise ratios (SINRs) of the common stream s K g c and the private stream s k p to EN k in group g are given as follows
γ K g , k c = h k H w K g c 2 g G , g g h k H w K g c 2 + j K h k H w j p 2 + σ 2 , g G , k K g ,
γ k p = h k H w k p 2 g G , g g h k H w K g c 2 + j K , j k h k H w j p 2 + σ 2 , k K .
With the above SINRs, the achievable rate R k p of the private stream s k p admits
R k p B log 2 1 + γ k p , k K .
To ensure that each common stream s K g c can be successfully decoded by the ENs in group g, the achievable rate R K g c of the common stream s K g c should admit
R K g c min k K g B log 2 1 + γ K g , k c , g G .
For the computation phase, as shown in Figure 3b, the ENs in each group g will first collaboratively compute the common parts W k c | k K g with information recycling. Next, each EN k will compute its private part W k p independently.
Remark 1. 
In the proposed scheme, each common stream s K g c is decoded by a smaller number of ENs instead of all ENs. As a result, the communication bottleneck of the single common stream can be alleviated, which allows more resources to be allocated to form stronger common streams. Consequently, the size of the common parts W 1 c , W 2 c , , W K c can be increased, thereby boosting the advantage of information recycling.

4.2. Problem Formulation

As presented before, the original task W with size D will finally be converted into G combined common parts W K 1 c , W K 2 c , , W K G c and K private parts W 1 p , W 2 p , , W K p for transmission. The uplink transmission time T u , K g c of the combined common part W K g c is given by
T u , K g c = D K g c R K g c ,
where D K g c is the size of W K g c . The uplink transmission time T u , k p of the private part W k p of the EN k is given by
T u , k p = D k p R k p ,
where D k p is the size of W k p . As the common streams and the private streams are transmitted concurrently, the overall uplink transmission time T u should admit
T u max max g G T u , K g c , max k K T u , k p = max max g G D K g c R K g c , max k K D k p R k p .
In the computation phase, the ENs in each group g will first collaboratively compute the combined common part W K g c = W k c | k K g . By (9), the corresponding computation latency T e , K g c in group g is given by
T e , K g c = D K g c k K g C k + η ,
where η is the cost for the signaling transmission (c.f., Section 3.2). Next, the ENs in group g start to compute their own private part, and the computation latency T e , k p of the private part W k p is given by
T e , k p = D k p C k .
Accordingly, the computation latency T e , K g p for the private parts in group g is given by
T e , K g p = max k K g D k p C k .
Hence, the overall latency T e for the total computation phase is given by
T e = max g G T e , K g c + T e , K g p = max g G D K g c k K g C k + η + max k K g D k p C k .
Based on the above, the expected latency minimization problem can be formulated as follows
P 1 : min N , w , D , R T u + E T e s . t . ( 14 ) ,   ( 15 ) ,   ( 18 ) ,   ( 22 ) ,
g G D K g c + k K D k p = D ,
tr w w H P t ,
g G K g = K ,
K i K j = , i , j G ,
where N = K 1 , K 2 , , K G , w = w K 1 c , w K 2 c , , w K G c , w 1 p , w 2 p , , w K p , D = D K 1 c , D K 2 c , , D K G c , D 1 p , D 2 p , , D K p , and R = R K 1 c , R K 2 c , , R K G c , R 1 p , R 2 p , , R K p are compact representations of the corresponding variables. The expectation in P 1 is over the randomness of the computation rates { C k } k = 1 K of the ENs (c.f. (8)). Constraint (24) restricts the total transmit power of the mobile user to P t .

5. Optimization Algorithm

Since the original problem P 1 is a non-convex combinatorial problem involving random variables, it turns out to be quite challenging to solve it directly. To this end, the optimization of the problem P 1 is suboptimally decomposed into two steps: deriving an effective grouping strategy, and optimizing the rest of the variables under the given grouping strategy.

5.1. EN Grouping

The optimal grouping strategy can be identified by exhaustively searching through all potential grouping strategies N ’s. However, this brute-force approach is computationally infeasible due to its prohibitively high complexity, which scales approximately as O G K . To this end, alternative methods can be employed to derive a sub-optimal grouping strategy with significantly lower complexity. Intuitively, if EN i and EN j exhibit sufficiently different channel conditions, clustering them together and transmitting a shared common stream will likely harm the communication rate due to the mismatch in their channel characteristics. Hence, it is essential to design an algorithm that can cluster ENs based on channel similarity to obtain a good grouping strategy. With this consideration, a K-medoids [33] based grouping algorithm is designed by jointly considering the channel alignment and the channel strength for the general multi-antenna case (It is worth noting that other heuristic clustering algorithms, such as K-means [34], can also be used for EN grouping in the proposed scheme). Besides, a more efficient continuous partitioning-based grouping algorithm is proposed for the special case of single-antenna.
Remark 2. 
As collaborated by the simulation results in Section 6, both the K-medoids-based grouping algorithm and the continuous partitioning-based grouping algorithm can lead to a reasonably good grouping strategy.
For the K-medoids-based grouping algorithm, both channel strength and channel alignment are jointly considered to assess the similarity between the channel characteristics of different ENs. Accordingly, the distance of the EN i and j is defined as
d i , j = l h i 2 h j 2 max { h i 2 , h j 2 } + ( 1 l ) 1 h i H h j 2 h i 2 h j 2 ,
where l 0 , 1 is a weighting parameter; the first term and the second term account for the channel similarity in magnitude and direction, respectively.
Remark 3. 
The intuition of (22) is the following: as shown in (12) and (15), decreasing the variation in channel strengths within each group positively impacts the transmission of common streams. Similarly, reducing differences in channel alignment yields comparable benefits.
With this distance metric, the standard K-medoids clustering algorithm can be applied, yielding a grouping strategy with complexity O K 2 G [33]. The K-medoids-based grouping algorithm has significantly lower complexity compared to the naive brute-force search while achieving a reasonable performance for problem P 1 . The specific process is summarized in Algorithm 2.
Algorithm 2 K-medoids-based grouping algorithm
Input: 
Channel state information { h 1 , h 2 , , h K } , number of groups G, distance metric d ( i , j ) ;
Output: 
Groups K 1 , K 2 , , K G ;
1:
Randomly select G ENs M = { M 1 , M 2 , , M G } as the initial medoids and set M = M ;
2:
repeat
3:
  Set M = M and K g = , g G ;
4:
  for  i = 1 , , K  do
5:
   Compute j = arg min j G d ( i , M j ) and Set K j = K j { i } ;
6:
  end for
7:
  for  g = 1 , , G  do
8:
   Select a new medoid M g that minimizes the total intra-group distance: M g = arg min j K g i K g d ( i , j ) ;
9:
  end for
10:
until  M = M
In the single-antenna case, besides the K-medoids-based grouping algorithm, a more efficient continuous partitioning-based grouping algorithm is further proposed. Specifically, in this case, the channel alignment term in the distance metric becomes irrelevant, focusing solely on channel strength. Without loss of generality, assume that the channel strengths of all ENs satisfy h 1 h 2 h K . The continuous partition-based grouping algorithm defines each group g as K g = c g 1 + 1 , c g 1 + 2 , , c g , where c 0 = 0 < c 1 < < c g 1 < c g = K . By selecting different values for c 1 , c 2 , c g 1 from the set 1 , 2 , , K 1 , this algorithm can generate K 1 G 1 candidate strategies, denoted as C . Besides, through searching through all grouping strategies in C , the continuous partitioning-based grouping algorithm can also achieve a good performance with much lower complexity than the brute-force search approach. In particular, this algorithm can achieve the same performance as the naive brute-force search approach in a special case of problem P 1 , where D k p = D 1 p , k K . The latency optimization problem of this special case can be formulated as follows
P 2 : min N , w , D , R T u + E T e s . t . ( 14 ) ,   ( 15 ) ,   ( 18 ) ,   ( 22 ) ,   ( 23 ) ,   ( 24 ) ,   ( 25 ) ,   ( 26 ) ,
D k p = D 1 p , k K .
Proposition 1. 
For problem P 2 , the optimal grouping strategy belongs to C .
Proof. 
Please see Appendix A.    □

5.2. Optimization Algorithm with Given Grouping

In this subsection, the optimization algorithm with a given grouping strategy N will be introduced. Specifically, consider the following problem
P 3 : min w , D , R T u + E T e s . t . ( 14 ) ,   ( 15 ) ,   ( 18 ) ,   ( 22 ) ,   ( 23 ) ,   ( 24 ) .
In the objective function of P 3 , due to the randomness of the computation capabilities C k ’s, it turns out to be difficult to find E T e in a closed form. To this end, by (22), a lower bound of the objective function can be obtained as follows
T u + E T e T u + max g G E D K g c k K g C k + η + max k K g D k p C k T u + max g G E D K g c k K g C k + η + max k K g E D k p C k .
Accordingly, the following optimization problem will be considered in the sequel
P 4 : min w , D , R T u + max g G E D K g c k K g C k + η + max k K g E D k p C k s . t . ( 14 ) ,   ( 15 ) ,   ( 18 ) ,   ( 23 ) ,   ( 24 ) .
Remark 4. 
As corroborated by the simulation results, using the solution obtained by optimizing this lower bound, a quite good expected processing latency can be achieved for P 1 .
By introducing the auxiliary variables t = t e , t e , K 1 , , t e , K G , problem P 4 can be readily transformed into the following equivalent problem
P 5 : min w , D , R , t T u + t e s . t . ( 14 ) ,   ( 15 ) ,   ( 18 ) ,   ( 23 ) ,   ( 24 ) ,
t e t e , K g , g G ,
t e , K g E D K g c k K g C k + η + E D k p C k , k K g .
In the sequel, the non-convex problem P 5 will be first transformed into a DC programming problem. In particular, by introducing the auxiliary variables ν = ν K g , k c , ν k p | g G , k K g to P 5 , constraints (14) and (15) can be equivalently written as
(34) R K g c B log 2 ν K g , k c , g G , k K g , (35) R k p B log 2 ν k p , k K , (36) ν K g , k c 1 + γ K g , k c , g G , k K g , (37) ν k p 1 + γ k p , k K .
By invoking (12) and (13), (36) and (37) can be further transformed into the following DC form
g G , g g h k H w K g c 2 + j K h k H w j p 2 + σ 2 g G h k H w K g c 2 + j K h k H w j p 2 + σ 2 ν K g , k c 0 ,
g G , k K g ,
g G , g g h k H w K g c 2 + j K , j k h k H w j p 2 + σ 2 g G , g g h k H w K g c 2 + j K h k H w j p 2 + σ 2 ν k p 0 ,
g G , k K g .
Besides, it can be obtained that (18) can be rewritten into the following DC form
D K g c + Ω T u , R K g c 0 , g G ,
D k p + Ω T u , R k p 0 , k K ,
where Ω a , b 1 / 4 · a b 2 1 / 4 · a + b 2 . Besides, it is not difficult to verified that constraint (32) is linear about t e and t e , K g , and constraint (33) is linear about t e , K g , D K g c and D k p . Consequently, P 5 can be equivalently transformed into the following DC programming problem
P 6 : min w , D , R , t , ν T u + t e
s . t . ( 23 ) ,   ( 24 ) ,   ( 32 ) ,   ( 33 ) ,   ( 34 ) ,   ( 35 ) ,   ( 38 ) ,   ( 39 ) ,   ( 40 ) ,   ( 41 ) .
The above DC programming problem P 6 can be efficiently solved by the CCCP technique [35], which is briefly introduced in the following. Consider DC programming problems which have the form
min x f 0 ( x ) g 0 ( x ) s . t . f a ( x ) g a ( x ) 0 , a = 1 , , m ,
where x is the optimization variable and f a and g a are convex for a = 0 , 1 , , m . The fundamental concept of CCCP is to solve a sequence of approximate convex problems iteratively. At the i + 1 -th iteration, the approximated problem is given by the following:
min x f 0 ( x ) g ^ 0 ( x ; x i ) s . t . f a ( x ) g ^ a ( x ; x i ) 0 , a = 1 , , m ,
where g ^ a ( x ; x i ) = g a ( x i ) + g a ( x i ) ( x x i ) and x i is the optimal solution of problem (44) at the i-th iteration. As the number of iterations increases, the optimal solution x i will converge to a Karush–Kuhn–Tucker (KKT) point of the original problem [35].
For the considered problem P 6 , at the i + 1 -th iteration of the CCCP procedure, the concave parts in both the objective function and the constraints will be linearized at the point w i , D i , R i , t i , ν i , where w i = w K 1 c , i , w K 2 c , i , , w K G c , i , w 1 p , i , w 2 p , i , , w K p , i , D i = D K 1 c , i , D K 2 c , i , , D K G c , i , D 1 p , i , D 2 p , i , , D K p , i , R i = R K 1 c , i , R K 2 c , i , , R K G c , i , R 1 p , i , R 2 p , i , , R K p , i , t i = t e i , t e , K 1 i , , t e , K G i and ν i = ν K g , k c , i , ν k p , i | k K , g G are compact representations of the corresponding optimization variables obtained in the i-th iteration. Specifically, as the objective function (42) and the constraints (24), (32), (33), (34) and (35) are convex, they do not require the treatment of linearization. For the constraints (38) and (39), by linearizing their concave parts, their approximations for the i + 1 -th iteration are, respectively, given by
g G , g g h k H w K g c 2 + j K h k H w j p 2 + σ 2 + g G h k H w K g c 2 + j K h k H w j p 2 + σ 2 × ν K g , k c / ν K g , k c , i 2 2 R g G w K g c , i H h k h k H w K g c + j K w j p , i H h k h k H w j p + σ 2 / ν K g , k c , i 0 ,
g G , k K g ,
g G , g g h k H w K g c 2 + j K , j k h k H w j p 2 + σ 2 + g G , g g h k H w K g c 2 + j K h k H w j p 2 + σ 2 × ν k p / ν k p , i 2 2 R g G , g g w K g c , i H h k h k H w K g c + j K w j p , i H h k h k H w j p + σ 2 / ν k p , i 0 ,
g G , k K g .
Similarly, for the constraints (40) and (41), their approximations for the i + 1 -th iteration are, respectively, given by
D K g c + Φ T u , R K g c ; T u i , R K g c , i 0 , g G ,
D k p + Φ T u , R k p ; T u i , R k p , i 0 , k K ,
where Φ i a , b ; a i , b i 1 4 ( a b ) 2 + 1 4 ( a i + b i ) 2 1 2 ( a i + b i ) ( a + b ) and T u i can be computed by using (18) and D i , R i .
Based on the above, given the optimal solution w i , D i , R i , t i , ν i at the i-th iteration, the approximate convex problem of P 6 for the i + 1 -th iteration is given by
P 7 : min w , D , R , t , ν T u + t e s . t . ( 23 ) ,   ( 24 ) ,   ( 32 ) ,   ( 33 ) ,   ( 34 ) ,   ( 35 ) ,   ( 45 ) ,   ( 46 ) ,   ( 47 ) ,   ( 48 ) .
The convex problem P 7 can be solved by standard methods [36]. Specifically, with the interior point method, the computational complexity for solving problem P 7 is approximately O K + G N t 3.5 [36]. The above procedure will be repeated until convergence as summarized in Algorithm 3. Additionally, since the problem size of P 7 does not scale with the number of iterations, the computation complexity of Algorithm 3 is O K + G N t 3.5 I , where I is the total number of iterations required for the convergence.
Algorithm 3 CCCP-based optimization for problem P 3
Input: 
Tolerance ϵ , power constraint P t , distributions of the computation rates of all ENs { C k } i = 1 K , task size D;
Output: 
Precoder w * , task partitioning vector D * , rate vector R * ;
1:
Set iteration index i = 0 ;
2:
Find an initial feasible point w 0 , D 0 , R 0 , t 0 , ν 0 of P 5 ;
3:
repeat
4:
  Using w i , D i , R i , t i , ν i obtained from the last iteration, solve problem P 7 and obtain an optimal solution w i + 1 , D i + 1 , R i + 1 , t i + 1 , ν i + 1 ;
5:
  Compute the corresponding objective function O E i + 1 ;
6:
  Update iteration index i = i + 1 ;
7:
until  O E i O E i 1 ϵ ;
8:
return  w * = w i , D * = D i , R * = R i .

6. Simulation Results

In this section, simulation results are presented to collaborate on the effectiveness of the proposed scheme. In the simulations, the number K of the ENs is set to 8. The size of the original computation task is set to 1 × 10 6 bits. For communication, the one-ring model [37] is adopted to characterize the channel between the mobile user and each EN, assuming that the mobile user’s antennas have a uniform linear array structure [25,38,39]. Besides, the location angle of EN k with respect to the mobile user’s antennas is set to 2 π k K and the angular spread of the scattering associated with EN k follows uniform distribution on π 36 , π 18 . The average channel power gain ρ is set to 85 dB. The bandwidth B is set to 1 (MHz); the noise power σ 2 is set to 1 × 10 10 (W); the maximum transmit power P t of the mobile user is set to 1 (W). The number of antennas of the mobile user N t is set to 6. For computation, processing each bit of the computation task is assumed to take β = 400 CPU cycles. The maximum processor frequency f e is set to 2.5 (GHz). The disturbed part Δ k of the computation rate is assumed to follow Bernoulli distribution [29,40,41] with parameter p k = 0.3; that is, P { Δ k = S } = p k and P { Δ k = 0 } = 1 p k . Here, S is the maximum value of the disturbed part and is set to 19 20 C ¯ . The cost η for the signaling transmission of the collaboration is set to 1 × 10 2 second. The weighting parameter l in (27) is set to 0.5. To demonstrate the advantage of the proposed scheme, three baseline schemes are considered. The first baseline is the naive (uncoded) distributed edge computing with neither information recycling nor coded computing. The second baseline is conventional coded computing with the optimal coding parameter. The third baseline is the conventional information recycling for distributed edge computing, (i.e., the proposed scheme with G = 1 and K G 1 = K ). Besides, the K-medoids algorithm and continuous partitioning algorithm are both demonstrated in the single-antenna case for the proposed scheme.
The task processing latency of the proposed scheme and those of the three baselines under different B with N t = 1 and N t = 6 are compared in Figure 4 and Figure 5, respectively. It can be seen that the performance gain brought by the proposed scheme is significant. For instance, when N t = 1 and the bandwidth is 1 (MHz), the proposed scheme with continuous partitioning algorithm can substantially reduce the processing latency to about 0.29 (s) and achieves about 59.8 % and 51.0 % performance gains as compared to Baseline 2 and Baseline 3. Intuitively, the reason is that the proposed scheme does not introduce extra task coding redundancy as compared to Baseline 2 and alleviates the communication limitation by transmitting multiple common streams as compared to Baseline 3. Besides, it can be observed from Figure 4 that the proposed scheme with the continuous partitioning algorithm achieves a better performance than that with the heuristic K-medoids algorithm. To better demonstrate the advantage of the proposed scheme over Baseline 3, the common streams’ power ratio g = 1 G w K g c 2 P t and task ratio g = 1 G D K g D of the proposed scheme and Baseline 3 are compared in Figure 6. It can be observed that the proposed scheme has a higher power ratio g = 1 G w K g c 2 P t and task ratio g = 1 G D K g D as compared to Baseline 3. For example, when the bandwidth is 0.5 (MHz), the power ratio and task ratio of the proposed scheme are 0.91 and 0.75, respectively, while the power ratio and task ratio of Baseline 3 are 0.62 and 0.23, respectively. The reason is that, by alleviating the rate limitation of the single common stream through the transmission of multiple common streams, the proposed scheme can allocate more power to enlarge the common streams, thereby increasing the task ratio of the common part(s) and promoting computation collaboration for processing latency reduction.
Besides, the task processing latency of the proposed scheme and those of the three baselines under different ρ with N t = 1 and N t = 6 are compared in Figure 7 and Figure 8, respectively. It can be seen that the advantages of the proposed scheme are still robust. For instance, when the average power gain ρ is 90 (dB), the proposed scheme can substantially reduce the processing latency to about 0.39 (s) and 0.30 (s), respectively, which further demonstrates the effectiveness of the proposed scheme.
Moreover, the communication phase latency and the computation phase latency under different f e with N t = 6 are compared in Figure 9 and Figure 10, respectively. It can be seen that the proposed scheme achieves both lower communication latency and computation latency as compared to Baseline 3, further demonstrating its superiority in overcoming the communication limitation of the single common stream to facilitate communication and promote information recycling for enhanced computation collaboration. Meanwhile, it can be seen from Figure 9 that, the proposed scheme has much lower communication latency than Baseline 2. The reason is that, due to the introduction of task redundancy for straggler mitigation, Baseline 2 (i.e., coded computing) incurs additional communication overhead, resulting in relatively high communication latency.
Furthermore, the task processing latency of the proposed scheme and those of the three baselines with K = 12 under different B and ρ are compared in Figure 11 and Figure 12, respectively.
It can be observed from Figure 11 and Figure 12 that, the performance gain brought by the proposed scheme scheme is still significant. For example, when the bandwidth is 0.5 (MHz), the proposed scheme can substantially reduce the processing latency to about 0.33 (s) and achieve about 28.5 % and 31.3% performance gains as compared to Baseline 2 and Baseline 3.

7. Conclusions and Further Works

In this work, a novel multi-group information recycling scheme is proposed for rate-splitting-based distributed edge computing, which can better utilize the information recycling mechanism for processing latency reduction. In particular, by clustering ENs into different groups and transmitting multiple common streams, the proposed scheme can overcome the communication limitation of the single common stream and allocate more resources to the common streams to foster information recycling for straggler mitigation. To achieve an effective EN grouping strategy, a K-medoids-based algorithm is developed by jointly considering the channel alignment and channel strength for the general multi-antenna case, and a more efficient continuous partitioning-based algorithm is proposed for the specific case of single-antenna. Besides, through bounding the corresponding latency optimization problem via DC programming, a CCCP-based algorithm is developed to find a good solution. Simulation results show that the proposed scheme can substantially outperform the baselines and greatly enhance the system’s performance.
Extension to group number optimization, advanced grouping strategies exploration, practical implementation and compatibility with existing infrastructures, imperfect channel state information problem, security and vulnerability issues related to the common streams transmission, energy consumption optimization and distributed network congestion, as well as integrating information recycling into other non-orthogonal transmission techniques for distributed edge computing are all worthwhile future works.

Author Contributions

Conceptualization, W.L. and X.H.; methodology, W.L. and X.H.; software, W.L.; validation, W.L.; formal analysis, W.L.; writing—original draft preparation, W.L. and X.H.; writing—review and editing, W.L. and X.H.; funding acquisition, X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of Proposition 1

Proof. 
Consider a grouping strategy N = { K g } g = 1 G C , which admits
a , b G , inf K a , sup K a inf K b , sup K b .
Without loss of generality, assume that inf K b < inf K a and the optimal point of P 2 is N , w , D , R . By contradiction, it can be verified that K a inf K b , sup K b . In the following, it will be proved that the performance of N , w , D , R can be achieved by another point N , w , D , R constructed as follows
K g = K g , g G , g a , b ,
K a = K a o sup K b ,
K b = K b sup K b o ,
w K g c = w K g c , , g G ,
w k p = w k p , , k K , k o , sup K b ,
w o p = w o p , g G , g b w K g c , 2 + j K w j p , 2 + σ 2 / h o 2 g G , g a w K g c , 2 + j K w j p , 2 + σ 2 / h o 2 1 / 2 ,
w sup K b p = w sup K b p , g G , g a w K g c , 2 + j K w j p , 2 + σ 2 / h sup K b 2 g G , g b w K g c , 2 + j K w j p , 2 + σ 2 / h sup K b 2 1 / 2 ,
D = D ,
R = R ,
where o K a inf K b , sup K b . The feasibility of the point N , w , D , R is presented in the following. First, it is shown that the optimal point N , w , D , R satisfies min k K R k p , min k K log ( 1 + γ k p , ) . Assume that min k K R k p , > min k K log ( 1 + γ k p , ) and let j = arg min k K log ( 1 + γ k p , ) . As R j p , log ( 1 + γ j p , ) and min R k p , k = 1 K R j p , , a contradiction arises, following that min R k p , k = 1 K min log ( 1 + γ k p , ) k = 1 K . Next, it is shown that the optimal point N , w , D , R admits
γ i p , = γ j p , , i , j K .
Assume that k K , γ k p , > min k K γ k p , , where γ k p , admits (13). Multiplying w k p , by a factor α , where 0 < α < 1 , a new precoder w = w K 1 c , , , w K G c , , w 1 p , , , α w k p , , , w K p , can be constructed, where w admits that
γ k p , = min k K γ k p , , γ k p , > γ k p , , k K k , γ K g , k c , > γ K g , k c , .
By (14) and (15), a feasible point N , w , D , R can be further constructed, satisfying
R k p , = min k K log 1 + γ k p , = min k K log 1 + γ k p , min k K R k p , , k K , R K g c , = R K g c , , g G .
By (18) and (22), it can be readily obtained that the point N , w , D , R can achieve the same performance as the point N , w , D , R . Further using the spare power ( 1 α ) w k p , 2 , a point N , P t tr w w , H ( 1 α ) w k p 2 w , D , R ^ with better performance, where R ^ R , can be constructed, thus contradicting the assumption. Therefore, it follows that the point N , w , D , R admits k K , γ k p , = min k K γ k p , (i.e., γ i p , = γ j p , , i , j K ), which yields that
h o H w o p , 2 g G , g a h o H w K g c , 2 + j K , j o h o H w j p , 2 + σ 2 = h sup K b H w sup K b p , 2 g G , g b h sup K b H w K g c , 2 + j K , j sup K b h sup K b H w j p , 2 + σ 2 h o H w o p , 2 g G , g a h o H w K g c , 2 + j K h o H w j p , 2 + σ 2 = h sup K b H w sup K b p , 2 g G , g b h sup K b H w K g c , 2 + j K h sup K b H w j p , 2 + σ 2 .
Accordingly, in the special case of single-antenna, it can be computed that
w o p 2 + w sup K b p 2 w o p , 2 w sup K b p , 2 = w K a c , 2 w K b c , 2 w o p , 2 g G , g a w K g c , 2 + j K w j p , 2 + σ 2 / h o 2 w sup K b p , 2 g G , g b w K g c , 2 + j K w j p , 2 + σ 2 / h sup K b 2 = 0 .
By (A5), (A6), and (A15), it can be obtained that w is feasible for P 2 . Moreover, by (13), (A7) and (A15), it can be computed that
γ o p = w o p 2 g G , g b w K g c 2 + j K w j p 2 w o p 2 + σ 2 / h o 2 = 1 g G , g b w K g c , 2 + j K w j p , 2 + σ 2 / h o 2 w o p 2 1 = h o H w o p , 2 g G , g a h o H w K g c , 2 + j K , j o h o H w j p , 2 + σ 2 = γ o p , .
Similarly, it can be derived that γ sup K b p = γ sup K b p , . By (12), (13), (A5), (A6), and (A15), it can be readily obtained that
min k K g γ K g c = min k K g γ K g c , , g G , γ k p = γ k p , , k K , k o , sup K b .
Further, by (14) and (15), it can be obtained that the point N , w , D , R is feasible for P 2 and achieves the same performance as the optimal point of the previous grouping strategy N . Note that N and N admits the following relationship
K a inf K b , sup K b = K a inf K b , sup K b { o } .
Hence, by recursively adjusting the grouping strategy as (A2)–(A4), a grouping strategy N * = { K g * } g = 1 G can achieve the same performance as N , where
K g * = K g , g G , g a , b , K a * inf K b * , sup K b * = .
Further adjusting the grouping strategy N * , it can be obtained that a grouping strategy N * * = { K g * * } g = 1 G can achieve the same performance as N , where
K i * * inf K j * * , sup K j * * = , i , j G .
It can be readily verified that N * * belongs to C , which yields the fact that N C is not the optimal grouping strategy. Thus, the proposition is proved. □

References

  1. Mao, Y.; You, C.; Zhang, J.; Huang, K.; Letaief, K.B. A Survey on Mobile Edge Computing: The Communication Perspective. IEEE Commun. Surv. Tuts. 2017, 19, 2322–2358. [Google Scholar] [CrossRef]
  2. Wang, X.; Han, Y.; Leung, V.C.M.; Niyato, D.; Yan, X.; Chen, X. Convergence of Edge Computing and Deep Learning: A Comprehensive Survey. IEEE Commun. Surv. Tuts. 2020, 22, 869–904. [Google Scholar] [CrossRef]
  3. Gong, X. Delay-Optimal Distributed Edge Computing in Wireless Edge Networks. In Proceedings of the IEEE INFOCOM, Toronto, ON, Canada, 6–9 July 2020. [Google Scholar]
  4. Chen, M.; Gunduz, D.; Huang, K.; Saad, W.; Bennis, M.; Feljan, A.V.; Poor, H.V. Distributed Learning in Wireless Networks: Recent Progress and Future Challenges. IEEE J. Sel. Areas Commun. 2021, 39, 3579–3605. [Google Scholar] [CrossRef]
  5. Nguyen, C.T.; Nguyen, D.N.; Hoang, D.T.; Phan, K.T.; Niyato, D.; Pham, H.A.; Dutkiewicz, E. Elastic Resource Allocation for Coded Distributed Computing over Heterogeneous Wireless Edge Networks. IEEE Trans. Wirel. Commun. 2023, 22, 2636–2649. [Google Scholar] [CrossRef]
  6. Dutta, S.; Jeong, H.; Yang, Y.; Cadambe, V.; Low, T.M.; Grover, P. Addressing unreliability in emerging devices and non-von Neumann architectures using coded computing. Proc. IEEE 2020, 108, 1219–1234. [Google Scholar] [CrossRef]
  7. Zhu, J.; Li, S. Generalized Lagrange Coded Computing: A Flexible Computation-Communication Tradeoff. In Proceedings of the IEEE ISIT, Espoo, Finland, 26 June–1 July 2022. [Google Scholar]
  8. Ng, J.S.; Lim, W.Y.B.; Luong, N.C.; Xiong, Z.; Asheralieva, A.; Niyato, D.; Leung, C.; Miao, C. A Comprehensive Survey on Coded Distributed Computing: Fundamentals, Challenges, and Networking Applications. IEEE Commun. Surv. Tuts. 2021, 23, 1800–1837. [Google Scholar] [CrossRef]
  9. Lee, K.; Lam, M.; Pedarsani, R.; Papailiopoulos, D.; Ramchandran, K. Speeding Up Distributed Machine Learning Using Codes. IEEE Trans. Inf. Theory 2018, 64, 1514–1529. [Google Scholar] [CrossRef]
  10. Shah, N.B.; Lee, K.; Ramchandran, K. When Do Redundant Requests Reduce Latency? IEEE Trans. Commun. 2016, 64, 715–722. [Google Scholar] [CrossRef]
  11. Wang, D.; Joshi, G.; Wornell, G.W. Efficient Straggler Replication in Large-Scale Parallel Computing. ACM Trans. Model. Perform. Eval. Comput. Syst. 2019, 4, 1–23. [Google Scholar] [CrossRef]
  12. Behrouzi-Far, A.; Soljanin, E. Efficient Replication for Fast and Predictable Performance in Distributed Computing. IEEE/ACM Trans. Netw. 2021, 29, 1467–1476. [Google Scholar] [CrossRef]
  13. Ramamoorthy, A.; Das, A.B.; Tang, L. Straggler-Resistant Distributed Matrix Computation via Coding Theory: Removing a Bottleneck in Large-Scale Data Processing. IEEE Signal Process. Mag. 2020, 37, 136–145. [Google Scholar] [CrossRef]
  14. Blumofe, R.D.; Leiserson, C.E. Scheduling multithreaded computations by work stealing. J. ACM 1999, 46, 720–748. [Google Scholar] [CrossRef]
  15. Harlap, A.; Cui, H.; Dai, W.; Wei, J.; Ganger, G.R.; Gibbons, P.B.; Gibson, G.A.; Xing, E.P. Addressing the straggler problem for iterative convergent parallel ML. In Proceedings of the ACM Symposium on Cloud Computing, Santa Clara, CA, USA, 5–7 October 2016. [Google Scholar]
  16. Liang, W.; Li, T.; He, X. Information Recycling Assisted Collaborative Edge Computing for Distributed Learning. In Proceedings of the IEEE INFOCOM WKSHPS, Hoboken, NJ, USA, 17–20 May 2023. [Google Scholar]
  17. Clerckx, B.; Mao, Y.; Jorswieck, E.A.; Yuan, J.; Love, D.J.; Erkip, E.; Niyato, D. A Primer on Rate-Splitting Multiple Access: Tutorial, Myths, and Frequently Asked Questions. IEEE J. Sel. Areas Commun. 2023, 41, 1265–1308. [Google Scholar] [CrossRef]
  18. Mao, Y.; Dizdar, O.; Clerckx, B.; Schober, R.; Popovski, P.; Poor, H.V. Rate-Splitting Multiple Access: Fundamentals, Survey, and Future Research Trends. IEEE Commun. Surv. Tuts. 2022, 24, 2073–2126. [Google Scholar] [CrossRef]
  19. Dai, L.; Wang, B.; Ding, Z.; Wang, Z.; Chen, S.; Hanzo, L. A Survey of Non-Orthogonal Multiple Access for 5G. IEEE Commun. Surv. Tuts. 2018, 20, 2294–2323. [Google Scholar] [CrossRef]
  20. Park, S.H.; Lee, H. Completion Time Minimization of Fog-RAN-Assisted Federated Learning with Rate-Splitting Transmission. IEEE Trans. Veh. Technol. 2022, 71, 10209–10214. [Google Scholar] [CrossRef]
  21. Nguyen, T.H.; Park, H.; Kim, M.; Park, L. DRL-Enabled RSMA-Assisted Task Offloading in Multi-Server Edge Computing. In Proceedings of the ICOIN, Ho Chi Minh City, Vietnam, 17–19 January 2024. [Google Scholar]
  22. Zhu, B.; Chi, K.; Liu, J.; Yu, K.; Mumtaz, S. Efficient Offloading for Minimizing Task Computation Delay of NOMA-Based Multiaccess Edge Computing. IEEE Trans. Commun. 2022, 70, 3186–3203. [Google Scholar] [CrossRef]
  23. Ding, Z.; Fan, P.; Poor, H.V. Impact of Non-Orthogonal Multiple Access on the Offloading of Mobile Edge Computing. IEEE Trans. Wirel. Commun. 2019, 67, 375–390. [Google Scholar] [CrossRef]
  24. Diamanti, M.; Pelekis, C.; Tsiropoulou, E.E.; Papavassiliou, S. Delay Minimization for Rate-Splitting Multiple Access-Based Multi-Server MEC Offloading. IEEE/ACM Trans. Netw. 2024, 32, 1035–1047. [Google Scholar] [CrossRef]
  25. Dai, M.; Clerckx, B.; Gesbert, D.; Caire, G. A Rate Splitting Strategy for Massive MIMO with Imperfect CSIT. IEEE Trans. Wirel. Commun. 2016, 15, 4611–4624. [Google Scholar] [CrossRef]
  26. Wang, Y.; Wong, V.W.S.; Wang, J. Flexible Rate-Splitting Multiple Access with Finite Blocklength. IEEE J. Sel. Areas Commun. 2023, 41, 1398–1412. [Google Scholar] [CrossRef]
  27. Ko, D.; Chae, S.H.; Choi, W. MDS Coded Task Offloading in Stochastic Wireless Edge Computing Networks. IEEE Trans. Wirel. Commun. 2022, 21, 2107–2121. [Google Scholar] [CrossRef]
  28. Yu, Q.; Maddah-Ali, M.A.; Avestimehr, A.S. Straggler Mitigation in Distributed Matrix Multiplication: Fundamental Limits and Optimal Coding. IEEE Trans. Inf. Theory 2020, 66, 1920–1933. [Google Scholar] [CrossRef]
  29. Peng, P.; Soljanin, E.; Whiting, P. Diversity/Parallelism Trade-Off in Distributed Systems with Redundancy. IEEE Trans. Inf. Theory 2022, 68, 1279–1295. [Google Scholar] [CrossRef]
  30. Bekkerman, R.; Bilenko, M.; Langford, J. Scaling up Machine Learning: Parallel and Distributed Approaches; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  31. Ding, Z.; Xu, J.; Dobre, O.A.; Poor, H.V. Joint Power and Time Allocation for NOMA–MEC Offloading. IEEE Trans. Veh. Technol. 2019, 68, 6207–6211. [Google Scholar] [CrossRef]
  32. Kiani, A.; Ansari, N. Edge Computing Aware NOMA for 5G Networks. IEEE Internet Things J. 2018, 5, 1299–1306. [Google Scholar] [CrossRef]
  33. Park, H.S.; Jun, C.H. A simple and fast algorithm for K-medoids clustering. Expert Syst. Appl. 2009, 36, 3336–3341. [Google Scholar] [CrossRef]
  34. Ahmed, M.; Seraj, R.; Islam, S.M.S. The k-means algorithm: A comprehensive survey and performance evaluation. Electronics 2020, 9, 1295. [Google Scholar] [CrossRef]
  35. Sun, Y.; Babu, P.; Palomar, D.P. Majorization-Minimization Algorithms in Signal Processing, Communications, and Machine Learning. IEEE Trans. Signal Process. 2017, 65, 794–816. [Google Scholar] [CrossRef]
  36. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  37. Goldsmith, A. Wireless Communications; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  38. Li, Z.; Ye, C.; Cui, Y.; Yang, S.; Shamai, S. Rate Splitting for Multi-Antenna Downlink: Precoder Design and Practical Implementation. IEEE J. Sel. Areas Commun. 2020, 38, 1910–1924. [Google Scholar] [CrossRef]
  39. Sadeghabadi, E.; Blostein, S. Low Complexity Rate Splitting Using Hierarchical User Grouping. In Proceedings of the IEEE ICC WKSHPS, Montreal, QC, Canada, 14–23 June 2021. [Google Scholar]
  40. Behrouzi-Far, A.; Soljanin, E. Redundancy Scheduling in Systems with Bi-Modal Job Service Time Distributions. In Proceedings of the Allerton, Monticello, IL, USA, 24–27 September 2019. [Google Scholar]
  41. Tandon, R.; Lei, Q.; Dimakis, A.G.; Karampatziakis, N. Gradient Coding: Avoiding Stragglers in Distributed Learning. In Proceedings of the ICML, Sydney, NSW, Australia, 6–11 August 2017. [Google Scholar]
Figure 1. Comparison between the conventional distributed computing scheme and the conventional information recycling scheme.
Figure 1. Comparison between the conventional distributed computing scheme and the conventional information recycling scheme.
Electronics 13 04403 g001
Figure 2. An illustrative example of the collaborative computing among ENs under information recycling. Suppose that the task queues of the three ENs are initialized as { W 1 , k c } k = 1 4 , { W 2 , k c } k = 1 8 , and { W 3 , k c } k = 1 6 , respectively. (a) After it finished the mini-task W 1 , 2 c , EN 1 sends the finishing signal S 1 , 2 fin to the mobile user to indicate the completion. Meanwhile, the mobile user has already received three finishing signals { S 1 , 1 fin , S 2 , 1 fin , S 3 , 1 fin } from the three ENs. (b) EN 1 finishes its last assigned task W 1 , 4 c and sends a finishing signal S 1 , 4 fin to the mobile user. (c) Upon receiving the finishing signal S 1 , 4 fin , the mobile user is notified that EN 1 has completed all its assigned mini-tasks. As EN 2 has the a recycling signal S 2 , 8 rec and a canceling signal S 2 , 8 can to EN 1 and EN 2, respectively. (d) EN 1 receives the recycling signal S 2 , 8 rec and starts the computation of W 2 , 8 c . EN 2 receives the canceling signal S 2 , 8 can and removes the mini-task W 2 , 8 c from its task queue.
Figure 2. An illustrative example of the collaborative computing among ENs under information recycling. Suppose that the task queues of the three ENs are initialized as { W 1 , k c } k = 1 4 , { W 2 , k c } k = 1 8 , and { W 3 , k c } k = 1 6 , respectively. (a) After it finished the mini-task W 1 , 2 c , EN 1 sends the finishing signal S 1 , 2 fin to the mobile user to indicate the completion. Meanwhile, the mobile user has already received three finishing signals { S 1 , 1 fin , S 2 , 1 fin , S 3 , 1 fin } from the three ENs. (b) EN 1 finishes its last assigned task W 1 , 4 c and sends a finishing signal S 1 , 4 fin to the mobile user. (c) Upon receiving the finishing signal S 1 , 4 fin , the mobile user is notified that EN 1 has completed all its assigned mini-tasks. As EN 2 has the a recycling signal S 2 , 8 rec and a canceling signal S 2 , 8 can to EN 1 and EN 2, respectively. (d) EN 1 receives the recycling signal S 2 , 8 rec and starts the computation of W 2 , 8 c . EN 2 receives the canceling signal S 2 , 8 can and removes the mini-task W 2 , 8 c from its task queue.
Electronics 13 04403 g002
Figure 3. An illustration of the proposed multi-group information recycling scheme.
Figure 3. An illustration of the proposed multi-group information recycling scheme.
Electronics 13 04403 g003
Figure 4. Comparison of processing latency under different B ( N t = 1 ).
Figure 4. Comparison of processing latency under different B ( N t = 1 ).
Electronics 13 04403 g004
Figure 5. Comparison of the processing latency under different B ( N t = 6 ).
Figure 5. Comparison of the processing latency under different B ( N t = 6 ).
Electronics 13 04403 g005
Figure 6. Comparison of the power ratio g = 1 G w K g c 2 P t and task ratio g = 1 G D K g D under different B ( N t = 6).
Figure 6. Comparison of the power ratio g = 1 G w K g c 2 P t and task ratio g = 1 G D K g D under different B ( N t = 6).
Electronics 13 04403 g006
Figure 7. Comparison of processing latency under different ρ ( N t = 1 ).
Figure 7. Comparison of processing latency under different ρ ( N t = 1 ).
Electronics 13 04403 g007
Figure 8. Comparison of the processing latency under different ρ ( N t = 6 ).
Figure 8. Comparison of the processing latency under different ρ ( N t = 6 ).
Electronics 13 04403 g008
Figure 9. Comparison of communication phase latency under different f e ( N t = 6 ).
Figure 9. Comparison of communication phase latency under different f e ( N t = 6 ).
Electronics 13 04403 g009
Figure 10. Comparison of computation phase latency under different f e ( N t = 6 ).
Figure 10. Comparison of computation phase latency under different f e ( N t = 6 ).
Electronics 13 04403 g010
Figure 11. Comparison of processing latency under different B ( K = 12 ).
Figure 11. Comparison of processing latency under different B ( K = 12 ).
Electronics 13 04403 g011
Figure 12. Comparison of processing latency under different ρ ( K = 12 ).
Figure 12. Comparison of processing latency under different ρ ( K = 12 ).
Electronics 13 04403 g012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, W.; He, X. Enhancing Rate-Splitting-Based Distributed Edge Computing via Multi-Group Information Recycling. Electronics 2024, 13, 4403. https://doi.org/10.3390/electronics13224403

AMA Style

Liang W, He X. Enhancing Rate-Splitting-Based Distributed Edge Computing via Multi-Group Information Recycling. Electronics. 2024; 13(22):4403. https://doi.org/10.3390/electronics13224403

Chicago/Turabian Style

Liang, Wanlin, and Xiaofan He. 2024. "Enhancing Rate-Splitting-Based Distributed Edge Computing via Multi-Group Information Recycling" Electronics 13, no. 22: 4403. https://doi.org/10.3390/electronics13224403

APA Style

Liang, W., & He, X. (2024). Enhancing Rate-Splitting-Based Distributed Edge Computing via Multi-Group Information Recycling. Electronics, 13(22), 4403. https://doi.org/10.3390/electronics13224403

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop