Dynamic Resource Allocation for Network Slicing with Multi-Tenants in 5G Two-Tier Networks
Abstract
:1. Introduction
- Enhanced Mobile Broadband (eMBB) [8]: provides enhanced bandwidth transmission speeds of up to 10 Gbps to improve existing communication services and offer users a seamless transmission experience. The slice is currently utilized for applications such as streaming and broadcasting, live video, and Augmented Reality/Virtual Reality (AR/VR).
- Ultra-Reliable and Low Latency Communications (URLLC) [9]: intended for communication applications with high reliability (i.e., error rates less than ) and low time latency (i.e., less than 1 ms) requirements. The slice is currently allocated mainly to critical applications such as remote surgery and autonomous driving.
- Massive Machine Type Communications (mMTC) [10]: provides the means to satisfy the communication needs of up to 1 million connected devices per square kilometer. The slice is currently used mainly to support the communication requirements of Internet of Things (IoT) devices.
- We proposed a method for dynamic resource allocation and task offloading optimization for 5G two-tier multi-tenant network slicing that considered the real-time resource usage of nodes in the network model. This enabled the determination of optimal routing and resource allocation for new service requests, with the goal of satisfying QoS requirements and maximizing resource utilization. Therefore, our approach effectively avoided problems such as premature resource depletion and blocking due to excessive resource allocation to nodes.
- Developed Minimum Cost Resource Allocation (MCRA) and Fast Latency Decrease Resource Allocation (FLDRA) dynamic resource allocation algorithms that worked with Cloud-only (CO), Edge-only (EO), and Parallel Cloud-Edge Hybrid (PH) network structures. Simulation results demonstrated that the MCRA algorithm calculated weights based on the current resource usage of each node to achieve optimal resource availability. Secondly, the FLDRA algorithm prioritized minimizing the node with the highest E2E delay while satisfying multi-tenant service quality requirements. Therefore, the effectiveness is also superior to Upper-tier First with Latency-bounded Overprovisioning Prevention (UFLOP). On the other hand, among our proposed three network models, the CO model experienced the earliest blocking, followed by the EO model, and finally the PH model. The reason is that the PH model supports offloading to the cloud. When comparing the CO and EO models, due to the difference in the number of nodes, under the same latency constraints, the EO model consumes fewer resources per request. In other words, the PH model can accommodate a higher arrival rate based on the same blocking rate. Overall, these algorithms and models significantly reduced service request blocking rates and improved resource utilization.
- Our proposed resource allocation approach considered the varying QoS requirements of emerging 5G applications for multi-tenant and multi-network slicing. It ensured that each tenant’s service quality assurance requirements were satisfied by optimizing resource allocation for each tenant’s network slice, thereby enhancing the overall efficiency and effectiveness of the multi-tenant 5G network and improving the quality of service provided to tenants.
2. Related Works
2.1. One-Tier Architecture
2.2. Two-Tier Architecture
2.3. Graph-Based Deep Learning and Reinforcement Learning
3. System Architecture
3.1. System Model
- Chien et al. [20] did not consider the core network. Thus, the present study defines a two-tier architecture that takes explicit account of this network.
- In [20], the Transport Network (TN) was considered a component of the overall architecture and had its own queue. However, in reality, the TN can be modeled as a simple communication queue compared to the core network, since it is a transmission environment that consists of multiple network devices. Therefore, the TN is redundant, and in our proposed architecture, we have removed it.
- In [20], the RAN and edge server are separated into different locations within the architecture. However, in real 5G network environments, the edge servers coexist with the RAN at the base station. As a result, in the architecture proposed in this study, the communication resources and queueing model between the RAN and edge server are ignored.
- Additionally, in [20], the resource allocation tasks for different service requests are handled separately. In other words, the allocated resource for each service request is based simply on that assigned to the first request. That is, no provision is made for adjusting the allocated resources dynamically across different service requests based on changes in the available resources at the nodes. In our proposed algorithm, we aim to dynamically allocate the given resources to each request. This involves adjusting the allocation of resources in response to changes in available resources at the nodes. By performing dynamic allocation, our algorithm can more effectively and efficiently Satisfy the resource requirements of each request, ultimately improving the overall performance of the system.
- In implementing the service diversion strategy in [20], the diversion ratio on different nodes is simply increased or decreased in increments of 1% until the resource limit is reached. However, this approach does not consider the relationship between the cost and the E2E delay, or the resource allocation ratios under each slice. Thus, the present study proposes a resource allocation strategy which optimizes the tradeoff between the cost weight and the minimum delay in each slice when adjusting the diversion ratio at the nodes.
3.2. Problem Definition
- Inputs: The two-tier architecture (C, E, CN, RN, UE); the number of network slices supported (L); the tenants and their SLAs (T, S, , D, ); the service requests from the tenants (, , , r, ); and the j th service request from the i th tenant in the l th slice ().
- Output: The resources required by each node to satisfy the current request (, , , , , ).
- Constraints:
- The specific resources of each tier (C, E, CN, RN, UE) required by each tenant must not exceed the corresponding SLA () resource allocation limit.
- The calculated E2E delay must satisfy the corresponding delay constraint for each tenant slice ().
- Objective: Maximize the system resource availability .
4. Proposed Methodology
4.1. Resource Allocation Work Flow
4.2. Latency Modeling
5. Algorithm Designs
5.1. Determine Resource Allocation for Each Request in Cloud-Only and Edge-Only Models
Algorithm 1 Resource allocation determination for each request in Cloud-only (CO) and Edge-only (EO) Model |
Input: Output:
|
5.2. Determine Resource Allocation for Each Request in Parallel Cloud-Edge Hybrid Model
Algorithm 2 Resource allocation determination for each request in Parallel Cloud-Edge Hybrid (PH) Model |
Input: Output:
|
5.3. Minimum Cost Resource Allocation Algorithm
Algorithm 3 Minimum Cost Resource Allocation (MCRA) algorithm |
Input: Output:
|
5.4. Fast Latency Decrease Resource Allocation Algorithm
Algorithm 4 Fast Latency Decrease Resource Allocation (FLDRA) algorithm |
Input: Output:
|
6. Simulation
6.1. Simulation Environment
- Tenant 1:
- eMBB: High data rates and traffic density performance criteria as Urban macro
- URLLC: Process automation–remote control
- mMTC: Low data rate and maintaining battery life
- Tenant 2:
- eMBB: 4K 360-degree VR panoramic video (VR IMAX and VR concert)
- URLLC: Intelligent transport systems–infrastructure backhaul
- mMTC: Medical monitoring
6.2. Simulation Results
6.2.1. Performance of MCRA and FLDRA Algorithms in Cloud-Only Model
- Average DelayFigure 6 shows the average delays of the MCRA and FLDRA algorithms in the different slices of the two tenants in the CO model. Note that the horizontal axis shows the request arrival rate of the entire system, while the vertical axis shows the average E2E delay. The results show that, through an appropriate incremental adjustment of the resources (), both algorithms enable the E2E delay constraints of the three slices to be satisfied for both tenants.Although both algorithms satisfy the SLA of each slice, the average delay of the MCRA algorithm is less than that of the FLDRA algorithm. This finding is reasonable since the FLDRA algorithm selects the node which results in the largest delay drop (as computed by Table 3, Equations (3a) and (4a), and line 10 of the Algorithm 4). Consequently, the delay gradually decreases with each adjustment. When the average delay satisfies the SLA, the system will stop adjusting, resulting in the node being very close to the SLA. By contrast, the MCRA algorithm calculates weights based on each node’s current resource usage and allocates resources to nodes with more remaining resources (as shown in Equation (6)). Therefore, when the algorithm selects the node with the lowest cost, it will inevitably cause the average delay to be lower than the SLA, and even significantly lower than the SLA, resulting in a greater reduction of the E2E delay.
- Resource AvailabilityFigure 7a,d shows the available resources at each node in the eMBB slices of the two tenants when applying the MCRA and FLDRA algorithms, respectively, and using the CO model. The horizontal axis represents the request arrival rate of the entire system, while the vertical axis represents the remaining resources at each node. A detailed analysis shows that the standard deviation of the available resources at each node under different arrival rates is around 0.01–0.05 for the MCRA algorithm and 0.2–0.35 for the FLDRA algorithm. In other words, the MCRA algorithm results in a more uniform utilization of the node resources in the eMBB slice than the FLDRA algorithm for each tenant.In the eMBB slice, the required resources and latency constraints of the T1 requests are smaller than those of T2. However, the T1 requests have a longer survival time (T1 = 5000 ms vs. T2 = 1000 ms, Table 3), and consequently, the resource utilization of T1 is greater than that of T2. From the perspective of the number of resource allocation adjustments, FLDRA selects nodes which yield the greatest reduction in the E2E delay. Hence, each node has been adjusted approximately the same number of times. However, due to the tenfold difference in bandwidth between the core network and RN (i.e., is 10 times that of ), the remaining available capacity at the RN node is relatively small, while that of the core network is relatively large. In addition, for T1 (high data rate and traffic density), the remaining available capacity of the C computing resource is greater than that of any of the other nodes for all values of the traffic arrival rate, while for T2 (360-degree VR panoramic video), the remaining available capacity of the C computing resource is the lowest of all the nodes.Figure 7b,e shows the available resources at each node in the URLLC slices of the two tenants under the two algorithms and the CO model. For both tenants, the standard deviation of the available resources under the MCRA algorithm is around 0.01–0.05, while for the FLDRA algorithm, the standard deviation is around 0.2–0.3. In other words, for both tenants, the MCRA algorithm achieves a more balanced resource utilization across the URLLC slice since it favors nodes with the lowest cost when allocating resources, where this cost is calculated based on the usage of the nodes. In particular, as shown in Algorithm 3. is calculated based on ∂, and the exponential term of ∂ is divided by , which can be interpreted as a proportional adjustment. By contrast, in the FLDRA algorithm, the resource allocation is adjusted based on the delay, and hence the resource usage across the different nodes is relatively unbalanced.In the URLLC slice, T1 has a higher demand for single-request resources than T2, resulting in a relatively higher occupancy of the resources. Although T2 has relatively lower latency constraints, the impact of this on the resource availability is less significant than that of the resource demands of T1. As a result, the availability of the computational resources of T2 is greater than that of T1. In general, due to the strict latency requirements of vehicle-to-vehicle communications, the computational resources allocated to the URLLC slice are slightly higher than those allocated to the other slices, resulting in fewer remaining RN resources and more remaining CN resources. Notably, for T1 (remote control), the remaining available resources at the C computing node are the lowest among all the nodes, while for T2 (infrastructure backhaul), the remaining C computing resources fall between those of the RN and CN, respectively.Figure 7c,f shows the available resources at each node of the mMTC slices when using the two algorithms and the CO model. The standard deviation of the available resources of the different nodes when using the MCRA algorithm is approximately 0.01–0.03. By contrast, that for the FLDRA algorithm is round 0.3–0.4. In other words, as for the eMBB and URLLC slices, the MCRA algorithm results in a more uniform resource utilization than the FLDRA algorithm for both tenants.In the mMTC slice, the latency constraint of T1 is much larger than that of T2 (T1 = 10,000 ms vs. T2 = 50 ms). Thus, it seems intuitive that T1 should require fewer resources to service each request. In Figure 7c, the application of T1 is for maintaining low data rates and battery life. Therefore, in the initial resource design, more resources are provided to the computing nodes and Core Network compared to RN, and requests are assigned to the computing nodes and CN for processing to reduce the frequency of discharging and charging, thereby extending the battery life. As a result, blocking only occurs when the arrival rate reaches 150. In Figure 7f, T2 is used for medical monitoring, and the communication nodes have more initial resources than the computing nodes. However, due to the FLDRA adjustment mechanism, each node experiences a relatively uniform number of adjustments. As a result, the T2 computing nodes with fewer initial resources will experience a reduction in resource availability after several allocations, resulting in a blocking phenomenon occurring at 60 (as shown in Figure 8c).
- Blocking RateFigure 8 shows the blocking rates of the two tenants for the different slices when using the MCRA and FLDRA algorithms in the CO model. The horizontal axis represents the request arrival rate of the entire system, while the vertical axis represents the blocking rate of each slice. In the simulations, the service requests are evenly distributed among the six slices, and hence the blocking rate rises relatively slowly in each case. For all six slices, the blocking rate caused by FLDRA is higher than that caused by MCRA. In the mMTC slice, the low latency constraint of 50 ms of T2 results in blocking even at low arrival rates. Due to its policy of selecting the node which yields the most significant reduction in the E2E delay, the blocking rate of T2 under FLDRA increases significantly as the request arrival rate increases beyond 60. Overall, the results show that, for both algorithms, a stricter latency requirement has a greater impact on the blocking rate, particularly at higher request arrival rates.
6.2.2. Performance of MCRA and FLDRA Algorithms in Edge-Only Model
- Average DelayFigure 9 shows the average delays in the three slices of the two tenants when using the MCRA and FLDRA algorithms under the EO model. The results are similar to those obtained for the CO model. In particular, when the initial E2E delay exceeds 1 s, the resource allocation is adjusted by the MCRA or FLDRA algorithm in such a way as to satisfy the respective latency constraints of the various applications. Both algorithms satisfy the SLA requirements of the two tenants in every slice. However, as for the EO model, the average delay under the MCRA algorithm is lower than that under FLDRA, particularly in the mMTC slice of T1. This is because the T1 mMTC slice has a much greater latency constraint (10,000 ms) compared to the other slices, which results in the significant difference in average latency between the two algorithms.
- Resource AvailabilityFigure 10a,d shows the availability of the resources at each node in the eMBB slices of the two tenants when using the MCRA and FLDRA algorithms and the EO model. For the MCRA algorithm, the standard deviation of the resource availability across the different nodes ranges from 0.01 to 0.05. By contrast, that for the FLDRA algorithm ranges from 0.15 to 0.3. In other words, MCRA results in a more uniform usage of the node resources than FLDRA for both tenants. T1 has lower resource and latency requirements than T2 in the eMBB slice. However, its service requests have a longer survival time (i.e., 5000 ms compared to 1000 ms for T2). Consequently, the resource usage is similar for both tenants in the eMBB slice. However, a detailed inspection reveals that T1 has slightly fewer remaining transmission resources than computational resources, while T2 has slightly fewer remaining computational resources. Finally, the difference in the amount of remaining available resources among the different nodes in the FLDRA algorithm is greater than that under the MCRA algorithm. This can be attributed to the fact that the CN resources are 10 times greater than that of the RN. In FLDRA, the mechanism results in the number of adjustments for each node being almost equal, without giving priority to nodes with more remaining resources for further adjustments. As a result, significant differences in remaining resource availability among nodes occur. By contrast, in MCRA, the system calculates weights based on each node’s current resource usage and allocates resources to nodes with more remaining resources (as shown in Equation (6)). Therefore, the mechanism of MCRA, resulting in different adjustment numbers for each node, promotes a more even distribution of available remaining resources.Figure 10b,e shows the resource availability at each node in the URLLC slices of the two tenants under the MCRA and FLDRA algorithms and EO model. The standard deviation of the resource availability for MCRA is approximately 0.01–0.05, while that for FLDRA it is around 0.15–0.2. In other words, for both tenants, MCRA achieves a more balanced distribution of the resource utilization of the nodes in the slice than FLDR. Although T1 demands more resources than T2 to satisfy its service requests in the URLLC slice, the latency constraints of T2 are more relaxed than those of T1, and hence the resource utilization of T2 is greater than that of T1. Under the MCRA and FLDRA algorithm, the remaining RN transmission resources and computational resources for T1 are more than those for T2. In each slice, the difference in the remaining available resources at the different nodes in the EO model is greater under the FLDRA algorithm than under the MCRA algorithm.Figure 10c,f shows the resource availability at each node of the mMTC slice of the two tenants when using the two resource allocation algorithms and the EO model. The standard deviation of the resource availability under MCRA is approximately 0.03–0.08, while that under FLDRA it is around 0.1–0.25. In other words, MCRA once again shows a relatively more balanced utilization of the node resources than FLDRA for both tenants.In the mMTC slice, T1 requires fewer resources to service its requests, but has a longer survival time. Consequently, the remaining resource availability reduces rapidly as the request arrival rate increases. Conversely, T2 requires more resources to service its requests. As a result, blocking occurs when the resource availability reduces to around 15% (see Figure 10f). In the MCRA algorithm, T1 has more remaining RN transmission resources and computational resources than T2. For both tenants, the disparity in the available resources among the different nodes is greater under the FLDRA algorithm than under the MCRA algorithm.
- Blocking RateFigure 11 shows the blocking rates of the MCRA and FLDRA algorithms for the different slices of the two tenants under the EO model. FLDRA results in a higher blocking rate than MCRA across all six slices, and for the URLLC and mMTC slices in particular. This outcome can be attributed to two main factors. First, the EO model has only three nodes, compared to the five nodes in the CO model, which exacerbates the uneven resource usage tendency of FLDRA and increases the number of necessary resource adjustments, especially for nodes with abnormally high resource usage. Second, computational resources are scarcer on the edge than in the C, which increases the likelihood of resource shortages for requests that require a large amount of computational resources. As a result, the service requests are more likely to be blocked.
6.2.3. Performance of MCRA and FLDRA Algorithms in Parallel Cloud-Edge Hybrid Model
- Average DelayFigure 12 shows the average delays in the three slices of each tenant when using the MCRA and FLDRA algorithms and the PH model. The experimental results are similar to those obtained for the CO and EO models. Both algorithms satisfy the SLAs of the two tenants in all three slices. However, the MCRA algorithm results in a lower average delay than FLDRA.
- Resource AvailabilityFigure 13a,d shows the available resources at each node in the eMBB slice of the two tenants when using the MCRA and FLDRA algorithms and the PH model. The standard deviations of the available resources at each node are around 0.01–0.05 and 0.1–0.15 for the two algorithms, respectively. In other words, MCRA once again results in a more uniform resource utilization than FLDRA for both tenants. Additionally, although T1 requires fewer resources in the eMBB slice than T2 and has higher latency constraints, it has a longer survival time (5000 ms) than T2 (1000 ms) and hence consumes significantly more resources. Moreover, when comparing MCRA and FLDRA, due to FLDRA prioritizes the selection of nodes which achieve the greatest reduction in the E2E delay, and among the EO model and CO model, the former has a higher latency than the latter. Thus, in the PH model, FLDRA leans towards assigning requests to the edge rather than the C, thereby increasing the resource usage at the edge and RN nodes. Furthermore, if the latency constraints cannot be satisfied, the FLDRA allocation mechanism distributes requests equally. However, due to the relatively scarce resources available at the E compared to the C, the E quickly becomes overwhelmed, leading to a limitation of both C and CN usage.Figure 13b,e shows the resource availability at each node in the URLLC slice of the PH model. The standard deviation of the resource availability under the MCRA algorithm is around 0.02, while that under FLDRA is 0.15–0.2. In other words, MCRA results in a more uniform distribution of the resources than FLDRA for both tenants. As with the CO and EO models, T1 requires more resources in the URLLC slice than T2. As a result, even though T2 has lower latency constraints than T1, its impact on the resource usage is not as significant as that for T1. In the comparison of MCRA and FLDRA, FLDRA prioritizes the allocation of requests to E, leading to a significant increased in the resource usage at the edge and RN nodes. Furthermore, when the latency constraints cannot be satisfied, FLDRA adopts a more balanced allocation of the requests, resulting in only a limited usage of the C and CN resources.Figure 13c,f shows the resource availability at each node in the mMTC slices of the two tenants in the PH model when using the two different algorithms. The standard deviations of the resource availability across the different nodes are around 0.015 for the MCRA algorithm and 0.1–0.2 for the FLDRA algorithm. Thus, as with the EO model and CO model, the MCRA algorithm achieves a more uniform resource utilization than the FLDRA algorithm for both tenants. In the mMTC slice, T1 requires only a few resources to process its service requests. However, its long survival time results in a significant reduction in the resource availability as the request arrival rate increases. Conversely, T2 requires a greater amount of resources to satisfy its service requests, and thus blocking occurs when the resource availability reduces to around 25%. Finally, as discussed above, the tendency of the FLDRA algorithm to assign service requests to E processing increases the use of the E and RN resources. Moreover, the adoption of a more balanced allocation when the latency constraints cannot be satisfied results in a limited usage of the C and CN resources.
- Blocking RateFigure 14 shows the blocking rates of the MCRA and FLDRA algorithms in the three slices of the two tenants in the PH model. As with the previous models, the blocking rate of the FLDRA algorithm is higher than that of the MCRA algorithm in all six slices.
6.2.4. Comparison of MCRA and UFLOP Algorithms in Cloud-Only Model
- Average DelayFigure 15 shows the average delays of the MCRA and UFLOP algorithms in the three slices of the two tenants in the CO model. The horizontal axis represents the arrival rate of the requests for the entire system, and the vertical axis represents the average delay. It is seen that both algorithms satisfy the SLA for the E2E delay of each tenant in every slice. However, for all six slices, the average delay of the MCRA algorithm is lower than that of UFLOP.
- Resource AvailabilityFigure 16 shows the resource availability at each node of the six slices under the MCRA and UFLOP algorithms, respectively. For both algorithms, the standard deviation of the resource availability across the different nodes is around 0.03. In other words, both algorithms yield a relatively uniform distribution of the resource usage across the nodes in the CO model. For the UFLOP algorithm, T1 and T2 both suffer blocking in all three slices when the request arrival rate exceeds 30 (see Figure 17c). This observation suggests that the available resources of each node have reached a state of minimum capacity, rendering them unable to accommodate subsequent requests.
- Blocking RateFigure 17 shows the blocking rates at each node of the six slices of the two tenants under the MCRA and UFLOP algorithms and the CO model. The blocking rates of the UFLOP algorithm are higher than those of MCRA in all six slices. In the UFLOP algorithm, all of the slices experience blocking after the arrival of the first few service requests since, because all requests in the same slice do not consider current usage for dynamic adjustment, namely given a fixed resource. By contrast, the MCRA algorithm shares slice resources among all requests from the same tenant and dynamically adjusts resource allocation for each request based on current usage, thus satisfying its E2E constraints.
6.2.5. Comparison of MCRA and UFLOP Algorithms in Edge-Only Model
- Average DelayFigure 18 compares the average delays of the MCRA and UFLOP algorithms in the three slices of each tenant in the EO model. As with the case of the CO model, both algorithms satisfy the SLA requirements of the two tenants in every slice. However, the average delay under the MCRA algorithm is lower than that under the UFLOP algorithm.
- Resource AvailabilityFigure 19 presents the resource availability at each node in three slices of the EO model using MCRA and UFLOP. For both MCRA and UFLOP, the standard deviation of the resource availability at each node across six different slices is approximately 0.02, indicating a relatively uniform utilization of resources. However, when the arrival rate reaches 30, the UFLOP algorithm experiences blocking in both T1 and T2 (see Figure 20c), which leads to the inability to accommodate the next request. This trend is supported by the halt in the decline of remaining resource availability.
- Blocking RateFigure 20 compares the blocking rates of the MCRA and UFLOP algorithms in the three slices of each tenant under the EO model. It is evident that UFLOP has a higher blocking rate than MCRA across all six slices since, as discussed above for the CO model, the UFLOP algorithm can only accommodate a small number of requests before blocking occurs due to the fixed manner in which the resources are allocated within the slice.
6.2.6. Performance of MCRA and UFLOP Algorithms in Parallel Cloud-Edge Hybrid Model
- Average DelayFigure 21 shows the average delays in the three slices of the tenants when using the MCRA and UFLOP algorithms in the PH model. Both algorithms satisfy the SLA requirements of the two tenants in all six slices. However, the average delay under the MCRA algorithm is lower than that under UFLOP.
- Resource AvailabilityFigure 22 shows the resource availability at each node in the three slices of each tenant given the use of the MCRA and UFLOP algorithms and the PH model. For each slice, and both tenants, the standard deviation of the resource availability across the nodes is around 0.02–0.05. Thus, both algorithms result in a relatively balanced resource utilization within each slice. For the UFLOP algorithm, blocking occurs for both tenants when the arrival rate reaches 30 (see Figure 23), which is unable to accommodate the next request.
- Blocking RateFigure 23 shows the blocking rates of the MCRA and UFLOP algorithms in the different slices of the two tenants under the PH model. For all six slices, the blocking rate under UFLOP is higher than that under MCRA. This result is consistent with that obtained under the CO and EO models and confirms that the UFLOP algorithm accommodates significantly fewer requests than MCRA, thus leading to an early blocking of new service requests.
6.2.7. Overall Performance Comparison
- Comparison of Three Algorithms in Same ModelFigure 24 shows the blocking rates of the MCRA, FLDRA, and UFLOP algorithms for the three slices of each tenant in the CO, EO, and PH models. Overall, the results show that, for each model, the MCRA algorithm accommodates the highest request arrival rate, followed by the FLDRA algorithm and then the UFLOP algorithm. As described in Section 6.2.1, Section 6.2.2 and Section 6.2.3, MCRA outperforms FLDRA since it calculates weights based on the current resource usage of each node, and selects nodes with more remaining resources for allocation, resulting in a more evenly distributed amount of remaining resources. On the other hand, FLDRA adjusts the node with the most significant delay reduction, leading to a similar number of adjustments for each node, without prioritizing nodes with more remaining resources. In summary, MCRA can achieve a higher arrival rate without causing blocking, while FLDRA has a lower rate. Furthermore, in Section 6.2.4, Section 6.2.5 and Section 6.2.6, the UFLOP algorithm is prone to rapid blocking of all slices. This is because UFLOP does not dynamically adjust the allocation of resources based on the current node usage. As a result, the resources within the slice are quickly consumed, and new-arriving requests are blocked.
- Comparison of Same Algorithm on Different ModelsFigure 24c,f,i shows the blocking rates in the different slices when using the same algorithm (UFLOP) in the different models. It is observed that the CO model results in the earliest occurrence of blocking, followed by the EO model, and the PH model. The relatively poorer performance of the CO model stems for its use of a large number of nodes, i.e., C, CU, CD, RU and RD. By contrast, the EO model uses just three nodes, i.e., E, RU and RD. Therefore, given the same latency constraint, the EO model consumes fewer resources for each request and therefore experiences later blocking. The PH model has the advantage of cloud-side diversion, which is not available in the EO model. This allows the service requests to be processed either in the C or at the E depending on the current network situation, and hence the blocking rate is reduced.
- Comparison of Three Algorithms Used in Different ModelsFigure 25 shows the blocking rates in each slice of the two tenants when using the three resource allocation algorithms and different network structure models. For each slice, the UFLOP algorithm results in the earliest occurrence of blocking, followed by FLDRA and finally MCRA. Furthermore, for a given resource allocation algorithm, the CO model results in the quickest blocking speed, followed by the EO model, and then the PH model.
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
3GPP | 3rd Generation Partnership Project |
4G | Fourth Generation |
5G | Fifth Generation |
5QI | 5G Quality Indicator |
AN | Access Network |
AR | Augmented Reality |
C | Cloud Servers |
CD | Core Network Downlink |
CN | Core Network |
CO | Cloud-only |
CORD | Central Office Re-architected as a Datacenter |
CPU | Central Processing Unit |
CU | Core Network Uplink |
DEDCA | Dynamic Enhanced Distributed Channel Access |
DRL | Deep Reinforcement Learning |
E | Edge Servers |
E2E | End-to-End |
eMBB | enhanced Mobile Broadband |
EO | Edge-only |
FLDRA | Fastest Latency Decrease Resource Allocation |
GCN | Graph Convolutional Network |
gNB | Next Generation Node B |
GNN | Graph Neural Network |
IoT | Internet of Things |
MCRA | Minimum Cost Resource Allocation |
MEC | Multi-access Edge Computing |
mMTC | Massive Machine Type Communications |
NFV | Network Function Virtualization |
PH | Parallel Cloud-Edge Hybrid |
QoE | Quality of Experience |
QoS | Quality-of-Service |
RN(RAN) | Radio Access Network |
RU | Radio Access Network Uplink |
RD | Radio Access Network Downlink |
SDN | Software Defined Network |
SLA | Service-Level Agreement |
TN | Transport Network |
UE | User Equipment |
UFLOP | Upper-tier First with Latencybounded Overprovisioning Prevention |
URLLC | Ultra-Reliable and Low Latency Communications |
VNF | Virtual Network Function |
VR | Virtual Reality |
References
- 3GPP TR 21.205 V16.0.0, Technical Specification Group Services and System Aspects; Technical Specifications and Technical Reports for a 5G Based 3GPP System, Release 16. 2020. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3441 (accessed on 26 October 2022).
- ITU-T Recommendation E.800SerSup9 (10/21) Guidelines on Regulatory Aspects of Quality of Service. Available online: https://www.itu.int/rec/T-REC-E.800SerSup9/en (accessed on 3 May 2023).
- ITU-T Recommendation P.10/G.100 (11/17) Vocabulary for Performance, Quality of Service and Quality of Experience. Available online: https://www.itu.int/rec/T-REC-P.10 (accessed on 3 May 2023).
- 3GPP TR 21.201 V16.0.0, Technical Specification Group Services and System Aspects; Technical Specifications and Technical Reports for an Evolved Packet System (EPS) Based 3GPP System, Release 16. 2020. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=550 (accessed on 26 October 2022).
- CORD: Re-Inventing Central Offices for Efficiency and Agility. 2016. Available online: https://opennetworking.org/cord/ (accessed on 26 October 2022).
- Shin, M.-K.; Nam, K.-H.; Kim, H.-J. Software-Defined Networking (SDN): A Reference Architecture and Open APIs. In Proceedings of the 2012 International Conference on ICT Convergence (ICTC), Jeju, Republic of Korea, 15–17 October 2012; pp. 360–361. [Google Scholar]
- 3GPP TS 23.501 V16.1.0, Technical Specification Group Services and System Aspects; System Architecture for the 5G System; Stage 2, Release 16. 2019. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3144 (accessed on 3 May 2023).
- ETSI TR 121 915 V15.0.0, Digital Cellular Telecommunications System (Phase 2+) (GSM); Universal Mobile Telecommunications System (UMTS); LTE; 5G; Release Description, Release 15. 2019. Available online: https://www.etsi.org/deliver/etsi_tr/121900_121999/121915/15.00.00_60/tr_121915v150000p.pdf (accessed on 26 October 2022).
- 3GPP TR 23.725 V0.3.0, Technical Specification Group Services and System Aspects; Study on Enhancement of Ultra-Reliable Low-Latency Communication (URLLC) Support in the 5G Core Network (5GC), Release 16. 2018. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3453 (accessed on 26 October 2022).
- ETSI TR 138 913 V16.0.0, 5G; Study on Scenarios and Requirements for Next Generation Access Technologies, Release 16. 2020. Available online: https://www.etsi.org/deliver/etsi_tr/138900_138999/138913/16.00.00_60/tr_138913v160000p.pdf (accessed on 27 October 2022).
- Poryazov, S.A.; Saranova, E.T.; Andonov, V.S. Overall Model Normalization towards Adequate Prediction and Presentation of QoE in Overall Telecommunication Systems. In Proceedings of the 2019 14th International Conference on Advanced Technologies, Systems and Services in Telecommunications (TELSIKS), Nis, Serbia, 23–25 October 2019; pp. 360–363. [Google Scholar]
- Manzanares-Lopez, P.; Malgosa-Sanahuja, J.; Muñoz-Gea, J.P. A Software-Defined Networking Framework to Provide Dynamic QoS Management. IEEE 802.11 Netw. Sens. 2018, 18, 2247. [Google Scholar]
- NGMN 5G P1 Requirements & Architecture Work Stream End-to-End Architecture, Description of Network Slicing Concept. 2016. Available online: https://ngmn.org/wp-content/uploads/160113_NGMN_Network_Slicing_v1_0.pdf (accessed on 26 October 2022).
- Chen, W.-K.; Liu, Y.-F.; Dai, Y.-H.; Luo, Z.-Q. Optimal Qos-Aware Network Slicing for Service-Oriented Networks with Flexible Routing. In Proceedings of the 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 5288–5292. [Google Scholar]
- Xu, X.; Zhang, H.; Dai, X.; Hou, Y.; Tao, X.; Zhang, P. SDN based next generation Mobile Network with Service Slicing and trials. China Commun. 2014, 11, 65–77. [Google Scholar] [CrossRef]
- Rostami, A.; Öhlén, P.; Santos, M.A.S.; Vidal, A. Multi-Domain Orchestration across RAN and Transport for 5G. In Proceedings of the 2016 ACM SIGCOMM Conference, Florianopolis, Brazil, 22–26 August 2016; pp. 613–614. [Google Scholar]
- Katsalis, K.; Nikaein, N.; Schiller, E.; Ksentini, A.; Braun, T. Network Slices toward 5G Communications: Slicing the LTE Network. IEEE Commun. Mag. 2017, 55, 146–154. [Google Scholar] [CrossRef]
- Li, Y.; Xu, L. The Service Computational Resource Management Strategy Based On Edge-Cloud Collaboration. In Proceedings of the 2019 IEEE 10th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 18–20 October 2019; pp. 400–404. [Google Scholar]
- Wang, L.; Jiao, L.; Li, J.; Gedeon, J.; Mühlhäuser, M. MOERA: Mobility-Agnostic Online Resource Allocation for Edge Computing. IEEE Trans. Mob. Comput. 2019, 18, 1843–1856. [Google Scholar] [CrossRef]
- Chien, H.-T.; Lin, Y.-D.; Lai, C.-L.; Wang, C.-T. End-to-End Slicing with Optimized Communication and Computing Resource Allocation in Multi-Tenant 5G Systems. IEEE Trans. Veh. Technol. 2020, 69, 2079–2091. [Google Scholar] [CrossRef]
- Zhang, Q.; Liu, F.; Zeng, C. Adaptive Interference-Aware VNF Placement for Service-Customized 5G Network Slices. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; pp. 2449–2457. [Google Scholar]
- Ren, J.; Yu, G.; He, Y.; Li, G.Y. Collaborative Cloud and Edge Computing for Latency Minimization. IEEE Trans. Veh. Technol. 2019, 68, 5031–5044. [Google Scholar] [CrossRef]
- Wang, P.; Yao, C.; Zheng, Z.; Sun, G.; Song, L. Joint Task Assignment, Transmission, and Computing Resource Allocation in Multilayer Mobile Edge Computing Systems. IEEE Internet Things J. 2019, 6, 2872–2884. [Google Scholar] [CrossRef]
- Guo, M.; Li, L.; Guan, Q. Energy-Efficient and Delay-Guaranteed Workload Allocation in IoT-Edge-Cloud Computing Systems. IEEE Access 2019, 7, 78685–78697. [Google Scholar] [CrossRef]
- Tam, P.; Song, I.; Kang, S.; Ros, S.; Kim, S. Graph Neural Networks for Intelligent Modelling in Network Management and Orchestration: A Survey on Communications. Electronics 2022, 11, 3371. [Google Scholar] [CrossRef]
- Jiang, W. Graph-based Deep Learning for Communication Networks: A Survey. Comput. Commun. 2022, 185, 40–54. [Google Scholar] [CrossRef]
- Yuan, S.; Zhang, Y.; Ma, T.; Cheng, Z.; Guo, D. Graph convolutional reinforcement learning for resource allocation in hybrid overlay–underlay cognitive radio network with network slicing. IET Commun. 2023, 17, 215–227. [Google Scholar] [CrossRef]
- Shao, Y.; Li, R.; Hu, B.; Wu, Y.; Zhao, Z.; Zhang, H. Graph Attention Network-Based Multi-Agent Reinforcement Learning for Slicing Resource Management in Dense Cellular Network. IEEE Trans. Veh. Technol. 2021, 70, 10792–10803. [Google Scholar] [CrossRef]
- Dong, T.; Zhuang, Z.; Qi, Q.; Wang, J.; Sun, H.; Yu, F.R.; Sun, T.; Zhou, C.; Liao, J. Intelligent Joint Network Slicing and Routing via GCN-powered Multi-Task Deep Reinforcement Learning. IEEE Trans. Cogn. Commun. Netw. 2021, 8, 1269–1286. [Google Scholar] [CrossRef]
- Kamath, A.; Palmon, O.; Plotkin, S. Routing and Admission Control in General Topology Networks with Poisson Arrivals. J. Algorithms 1998, 27, 236–258. [Google Scholar] [CrossRef]
- 3GPP TS 22.262 V16.0.0, Technical Specification Group Services and System Aspects; Message Service within the 5G System; Stage 1, Release 16. 2018. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3558 (accessed on 28 February 2023).
- ETSI TR 103 702 V1.1.1, Speech and Multimedia Transmission Quality (STQ); QoS Parameters and Test Scenarios for Assessing Network Capabilities in 5G Performance Measurements. 2020. Available online: https://www.etsi.org/deliver/etsi_tr/103700_103799/103702/01.01.01_60/tr_103702v010101p.pdf (accessed on 28 February 2023).
- 3GPP TS 22.261 V15.8.0, Technical Specification Group Services and System Aspects; Service Requirements for the 5G System; Stage 1, Release 15. 2019. Available online: https://portal.3gpp.org/ChangeRequests.aspx?q=1&versionId=63664&release=190 (accessed on 28 February 2023).
- 3GPP TS 22.261 V18.3.0, Technical Specification Group Services and System Aspects; Service Requirements for the 5G System; Stage 1, Release 18. 2021. Available online: https://portal.3gpp.org/ChangeRequests.aspx?q=1&versionId=73737&release=193 (accessed on 28 February 2023).
- 3GPP TR 22.886 V16.2.0, Technical Specification Group Services and System Aspects; Study on Enhancement of 3GPP Support for 5G V2X Services, Release 16. 2018. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3108 (accessed on 28 February 2023).
- 3GPP TR 21.914 V14.0.0, Technical Specification Group Services and System Aspects; Release 14 Description; Summary of Rel-14 Work Items, Release 14. 2018. Available online: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3179 (accessed on 28 February 2023).
Lit. | Proposed | Network/Server Slicing (RAN/E/CN/C) | Tiers | Dynamic Allocation | Latency Constraint |
---|---|---|---|---|---|
[15] |
| Network -/-/CN/- | 1 | No | No |
[16] |
| Network RAN/-/CN/- | |||
[17] |
| Both -/-/N/C | |||
[18] |
| Both -/E/-/C | |||
[19] |
| Both RAN/E/-/- | Yes | Yes | |
[21] |
| Both -/E/-/C | 2 | No | No |
[22] |
| Both RAN/E/-/C | |||
[23] |
| Both -/E/-/C | |||
[24] |
| Both RAN/E/CN/C | No | Yes | |
[20] |
| Both RAN/E/CN/C | |||
Proposed |
| Both RAN/E/CN/C | Yes | Yes |
Symbol | Description | |
---|---|---|
Entity | C | Cloud Server |
E | Edge server | |
UE | User Equipment | |
CN | Core Network | |
RN | RAN | |
L | Number of service types (slices) supported (e.g., eMBB, URLLC, mMTC) | |
Tenant | T, | is the i-th tenant, T = |
, | The j-th uplink or downlink request of the i-th tenant on different node () | |
The l-th slice of the i-th tenant | ||
SLA of | ||
, | Current RN uplink and downlink resources allocated to | |
, | Current CN uplink and downlink resources allocated to | |
, | Current C and E computing resources allocated to | |
Current C and E computing resources allocated to | ||
, | Availability of RN uplink and downlink resources of . , | |
, | Availability of CN uplink and downlink resources of . , | |
, | Availability of C and E computing resources of . , | |
Slice resource availability of . = | ||
SLA | , | Guaranteed (maximum) RN uplink and downlink resources specified by |
, | Guaranteed (maximum) CN uplink and downlink resources specified by | |
, | Guaranteed (maximum) C and E computing resources specified by | |
The end-to-end (E2E) delay requirement of slice | ||
Service Deployment Request | Request arrival rate of slice | |
Request departure rate of slice | ||
The j- request of slice | ||
The traffic arrival rate of | ||
, | The computing resource demand of | |
, | The network (communication) resource demand of | |
, | RN uplink and downlink resources allocated to | |
, | CN uplink and downlink resources allocated to | |
, | Cloud and edge computing resources allocated to | |
Delay | The estimated computing delay of C of | |
The estimated computing delay of E of | ||
, | The estimated network delay of uplink and downlink of RN of | |
, | The estimated network delay of uplink and downlink of CN of | |
The estimated network delay of different place m of w = , and , where Cloud-only (), Edge-only ()} | ||
Unit Resource Comparison | The cost of allocating one unit of resource at node , where is C, E, RU, RD, CU, CD | |
∂ | The parameter of the resource reservation cost function (default value is 100) | |
The parameter for setting the smallest resource unit | ||
The total E2E cost of using CO/EO in the Parallel Cloud-Edge Hybrid (PH) model | ||
The total E2E delay of using CO/EO in the PH model | ||
The reduction of delay resulting from resource allocation at node X |
Tenant 1 | Tenant 2 | |||||
---|---|---|---|---|---|---|
eMBB | URLLC | mMTC | eMBB | URLLC | mMTC | |
(Uplink resource demand, unit: Mbps) | 2.5 | 20 | 0.14 | 4 | 2 | 0.1 |
(Downlink resource demand, unit: Mbps) | 5 | 1 | 0.125 | 8 | 0.2 | 0.05 |
(Latency constraints, unit: ms) | 500 | 60 | 10,000 | 100 | 30 | 50 |
Survival time (ms) | 5000 | 100 | 10,000 | 1000 | 100 | 500 |
(Departure rate) | 0.2 | 10 | 0.1 | 1 | 10 | 2 |
(Traffic arrival rate, unit: packet/request) | 10 | 1 | 1 | 10 | 5 | 5 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lin, J.-Y.; Chou, P.-H.; Hwang, R.-H. Dynamic Resource Allocation for Network Slicing with Multi-Tenants in 5G Two-Tier Networks. Sensors 2023, 23, 4698. https://doi.org/10.3390/s23104698
Lin J-Y, Chou P-H, Hwang R-H. Dynamic Resource Allocation for Network Slicing with Multi-Tenants in 5G Two-Tier Networks. Sensors. 2023; 23(10):4698. https://doi.org/10.3390/s23104698
Chicago/Turabian StyleLin, Jia-You, Ping-Hung Chou, and Ren-Hung Hwang. 2023. "Dynamic Resource Allocation for Network Slicing with Multi-Tenants in 5G Two-Tier Networks" Sensors 23, no. 10: 4698. https://doi.org/10.3390/s23104698
APA StyleLin, J. -Y., Chou, P. -H., & Hwang, R. -H. (2023). Dynamic Resource Allocation for Network Slicing with Multi-Tenants in 5G Two-Tier Networks. Sensors, 23(10), 4698. https://doi.org/10.3390/s23104698