Next Article in Journal
Multi-Temporal Change Detection Analysis of Vertical Sprawl over Limassol City Centre and Amathus Archaeological Site in Cyprus during 2015–2020 Using the Sentinel-1 Sensor and the Google Earth Engine Platform
Next Article in Special Issue
From Constellation Dithering to NOMA Multiple Access: Security in Wireless Systems
Previous Article in Journal
Q-Meter: Quality Monitoring System for Telecommunication Services Based on Sentiment Analysis Using Deep Learning
Previous Article in Special Issue
A Survey of Rain Attenuation Prediction Models for Terrestrial Links—Current Research Challenges and State-of-the-Art
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization-Based Resource Management Algorithms with Considerations of Client Satisfaction and High Availability in Elastic 5G Network Slices

1
Research Center for Information Technology Innovation, Academia Sinica, Taipei 115, Taiwan
2
Department of Information Management, National Taiwan University, Taipei 10617, Taiwan
3
Graduate Institute of Information Management, National Taipei University, New Taipei City 23799, Taiwan
*
Author to whom correspondence should be addressed.
Current address: 128 Academia Road, Section 2, Nankang, Taipei 115, Taiwan.
These authors contributed equally to this work.
Sensors 2021, 21(5), 1882; https://doi.org/10.3390/s21051882
Submission received: 2 February 2021 / Revised: 18 February 2021 / Accepted: 1 March 2021 / Published: 8 March 2021
(This article belongs to the Special Issue Radio Mobile Communication System)

Abstract

:
A combined edge and core cloud computing environment is a novel solution in 5G network slices. The clients’ high availability requirement is a challenge because it limits the possible admission control in front of the edge cloud. This work proposes an orchestrator with a mathematical programming model in a global viewpoint to solve resource management problems and satisfying the clients’ high availability requirements. The proposed Lagrangian relaxation-based approach is adopted to solve the problems at a near-optimal level for increasing the system revenue. A promising and straightforward resource management approach and several experimental cases are used to evaluate the efficiency and effectiveness. Preliminary results are presented as performance evaluations to verify the proposed approach’s suitability for edge and core cloud computing environments. The proposed orchestrator significantly enables the network slicing services and efficiently enhances the clients’ satisfaction of high availability.

1. Introduction

Novel applications of 5G services are cataloged with three kinds of quality of service (QoS) features: massive machine-type communication (mMTC), enhanced mobile broadband (eMBB), and ultra-reliable low-latency communication (uRLLC) in 5G cellular networks [1]. Two system architectures have been proposed for ensuring the scalability and flexibility of resource allocation and scheduling for the diverse QoS requirements. The first highly acceptable approach is the cloud radio access network (C-RAN). The baseband processing and networking functions are virtually centralized in a resource pool for resource scalability and flexibility [2]. The other one is mobile edge computing (MEC). It supports interactive and real-time applications associated with nearby cloud hosts to obtain the required latency for addressing urgent and distributed requirements [3]. To implement C-RAN and MEC, software-defined networking (SDN) and network function virtualization (NFV) have been integrated to develop network slicing technologies in 5G [4]. Network slices are end-to-end (E2E) mutually separate sets of programmable infrastructure resources with independent control. A slice is a logical network adapted for particular applications regarding the diverse QoS requirements [1].
However, SDN and NFV have paved the way for employing the slicing concept. The network slice-as-a-service (NSaaS) strategy has several problems and challenges [4,5]. Furthermore, the failure of virtualized network functions (VNFs) may impact the QoS for service provisioning of the control plane (e.g., MME) and the data plane (e.g., S-GWsor PDN-GWs), respectively [4]. The related factors would be addressed and collected to design an orchestrator with the resource management efficiently and effectively for the VNFs, such as resource allocation, load balancing, and high availability (HA) for building up a high performance and robust slicing network.
Furthermore, HA is considered a high stringency requirement for providing a sustainable business with emergency response applications to support a reliable performance with zero downtime and costs. Some technologies can be adopted by the redundant array of independent discs, the replication of nodes, or master-slave database redundancy to offer data protection in an HA-equipped system [6,7,8]. From the network operator perspective, 5G network slices’ resources are limited for simultaneous profit maximization. Load balancing is typically adopted by an orchestrator to achieve resource planning and provisioning and enhance performance [4,9].
In this paper, the efficient elastic mechanisms of resource management in virtual machines (VMs) with VNFs are proposed to integrate admission control, load balancing, resource allocation, and HA arrangement in 5G network slices. The system architecture is shown in Figure 1. An integer programming problem is formulated to maximize system utilization, subject to the quality of services, resource requirements, and HA constraints. The resource allocation problem is combined with knapsack and bin-packing problems. Generally, the VNFs or called tasks with various amounts of resource requirements emulated as VMs should be packed into a finite number of servers. A few servers in active status and then stepping forward servers by increasing demands are possible to reduce the operating costs in a bin-packing problem. A set of tasks for applications with diverse weights and values is selected for inclusion to remain within the resource limit and reach the maximum benefit, a well-known knapsack problem. Furthermore, the knapsack problem is the aforementioned combinatorial nondeterministic polynomial time (NP) hard problem. One possible solution is adopted by the Lagrangian relaxation (LR) method efficiently and effectively. Near-optimal solutions provide practical methods to overcome the bottleneck within limited computing resources of slices. The virtualization, parallel computing, and mathematical programming techniques are also developed in this paper to make near-optimal decisions from an NSaaS service provider’s perspective. The proposed resource management algorithms ensure that computing resources are equitably distributed in the slices of cloudlets and core clouds. The computation in the edge and core clouds is performed to achieve maximum resource utilization with HA and minimal cost. Moreover, the admission control scheme with the additional HA requirements is devised to admit the maximum number of jobs. Each job requires various types of resources.
The remainder of this paper is organized as follows: Section 2 reviews related works. Section 3 introduces the mathematical model and problem description. The LR-based solution approach is developed, as shown in Section 4. Section 5 presents computational experiments. Conclusions are drawn in Section 6.

2. Literature Review

A cloud computing environment in 5G network slices supports share-based services in a pay-as-you-go approach. It provides operators with a flexible architecture from the perspective of a network operator. Resource management with complex on-demand traffic in the share-based cloud computing environment is a considerable challenge. Inappropriately designed resource management solutions might lead to the network’s increasing costs. This is related to unsatisfactory quality of services (QoS) and system reliability regarding the penalty of user experience and call-blocking probability due to the virtualized network function failure. Table 1, Table 2 and Table 3 compare the resource management problems investigated with those investigated in related studies in terms of (i) resource allocation, (ii) load balancing, and (iii) admission control.

2.1. Resource Allocation

For the maximization of the benefits of the cloud system and service quality enhancement, Liu et al. [10] developed a joint multi-resource allocation model based on a semi-Markov decision process (SMDP). They solved the optimization problem using linear programming to attain wireless resource allocation and near-optimal decisions in cloud computing. However, cloud service providers fulfill individual requirements to achieve a win-win situation for both parties in terms of computational efficiency, budget balance, and truthfulness. Jin et al. [11] designed an incentive-compatible auction mechanism to appropriately allocate matching among cloud resources and user demands. In this auction model, the buyers are mobile devices, and the sellers are cloudlets. The nearest control center can adopt the role of the auctioneer to reduce transmission costs and latency. Furthermore, by assuming that all cloudlets can provide the same resources with distinct rewards, Liu and Fan [12] presented a two-stage optimization strategy to achieve optimal cloudlet selection in a multi-cloudlet environment and optimal resource allocation in response to the request in the cloudlet. Studies on resource allocation in a cloud computing environment are shown in Table 1. Resource allocation and scheduling algorithms have been proposed in numerous research areas, such as transportation management, industrial management, operational research, computer science, and particularly in real-time operating systems [13,14]. For instance, the earliest deadline first (EDF) is a dynamic scheduling algorithm used in real-time operating systems to allocate computing resources in central processing units (CPUs) using a priority queue. The queue searches for the task with the closest deadline; if a job cannot be completed within its time frame, the operating system must release the job. The main idea is to maximize resource use for several competing interests while balancing the load. A weighted priority for each data flow can be assigned inversely proportional to the respective flow’s anticipated resource consumption [15,16]. Ding et al. [17] proposed a Linux scheduling policy with a priority queue rather than a first-in-first-out (FIFO) queue to improve kernel-based virtual machine performance. Zhao et al. [18] proposed online VM placement algorithms to cost efficiently allocate resources to VMs to increase revenues in a managed server farm. First-fit (FF), FF migration (FFM), least reliable first (LRF), and decreased density greedy (DDG) algorithms are packing strategies relevant to task optimization for achieving desirable performance.
Table 1. Resource allocation comparisons with existing methods. SMDP, semi-Markov decision process; EDF, earliest deadline first; FF, First-fit; FFM, FF migration; LRF, least reliable first; DDG, decreased density greedy.
Table 1. Resource allocation comparisons with existing methods. SMDP, semi-Markov decision process; EDF, earliest deadline first; FF, First-fit; FFM, FF migration; LRF, least reliable first; DDG, decreased density greedy.
ClassificationObjectiveStrategyRelated StudiesProposed Methods
Resource allocationMaximizing revenue and satisfying QoSSMDP; auction; EDF; FIFO; FF; FFM; LRF; DDG[10,11,12,13,14,15,16,17,18]LR and next-fit are adopted
Comparison with related studies
This paper regards offloading tasks as virtual machines with distinct rewards with various levels of demands in the cloudlet and core cloud environment. The Lagrangian relaxation-based approach is proposed to maximize system revenue, subject to constraints such as computing capacity, assignments, and quality of service requirements. The objective is to maximize the total value of tasks by using proposed heuristics to appropriately rearrange resources using the next-fit algorithm to allocate tasks and satisfy application requirements.
Based on the analysis of the previous studies, this paper regards offloading tasks as virtual machines with distinct rewards with various levels of demands in the cloudlets and core clouds. The Lagrangian relaxation-based approach is proposed to maximize system revenue, subject to constraints such as computing capacity, assignments, and quality of service requirements. The approach is adopted by considering finding the optimal solutions other than SMDP within limited states or constraints. The objective is to maximize the total value of tasks using proposed heuristics in polynomial time. It also appropriately rearranges resources using the next-fit algorithm to allocate tasks and satisfy application requirements.

2.2. Load Balancing

Load balancing is a crucial technology that optimizes mobile application performance. Several studies have proposed cloudlet technology solutions [19] to achieve load balancing for mobile devices (Table 2). Jia et al. [19] devised a load-balancing algorithm to balance the workload among multiple cloudlets in wireless metropolitan area networks. A redirection scheme shortened the average offload task response time and improved user experience. Yao et al. [20] studied load balancing for cloudlet nodes. The task allocation problem was formulated as an integer linear problem. A two-step centralized appointment-driven strategy was used to obtain a solution with minimal average response time and a balanced load. From a network operator perspective, time-varying and periodically changing demands are challenges in providing on-demand services. Furthermore, cloud service providers face load-balancing problems in cloud computing environments. The system availability of on-demand access should be considered using an adjustable resource assignment mechanism to satisfy QoS. If resource demands are stochastic, their accurate forecast is severe. Historical data are measured for estimating usage data to determine the load distribution and evaluate the proposed algorithm [21].
Table 2. Load balancing comparisons with existing methods.
Table 2. Load balancing comparisons with existing methods.
ClassificationObjectiveStrategyRelated StudiesProposed Methods
Load balanceMaximum system availabilityBin packing; 0/1 knapsack[19,20,21]Based on the system utilization in global view
Comparison with related studies
In general, as many tasks as possible are admitted to maximize total revenue; however, the supply and demand for resources are unbalanced during peak traffic hours. This study determines which tasks are selected and dispatched to achieve maximum objective values subject to task assignment and limited capacity. The problems are classified as bin-packing and 0/1 knapsack problems. Herein, the assignment decisions are based on resource utilization functions, including a central processing unit, random access memory, hard drive, and bandwidth for each server type. The assignment strategies are designed to fit the QoS requirements, such as those of computing and transmission in variant traffic loads in a global view adopted by the heuristics of bin-packing strategies.
In general, as many tasks as possible are admitted to maximize total revenue; however, the supply and demand for resources are unbalanced during peak traffic hours. This study determines the orchestrator’s decisions for which tasks are selected and dispatched to achieve the maximum objective values subject to task assignment and limited capacity. The problems are classified as the bin-packing and 0/1 knapsack combined problems. Herein, the orchestrator’s decisions are based on resource utilization functions, including a central processing unit, random access memory, hard drive, and bandwidth for each server type. The orchestrator’s objective is also designed for that assignment strategy fitting the QoS requirements, such as those of computing and transmission in variant traffic loads in a global view adopted by the heuristics of bin-packing strategies.

2.3. Admission Control

Preventing server overload and ensuring application performance are the goals of admission control [22]. This mechanism decides whether to admit a particular service request to the server. It is implemented for meeting QoS requirements and achieving service providers’ expected revenue [23]. Hoang et al. [24] developed a semi-Markov decision process (SMDP)-based optimization model for mobile cloud computing that considers constraints such as resources, bandwidth, and QoS to perform admission control tasks. This framework ensures QoS performance and maximizes the reward. Xia et al. [25] aimed to optimize cloud system throughput. An effective admission algorithm based on the proposed resource cost paradigm to model various resource consumptions was devised according to the online request following the current workload. Table 3 presents the comparison of the admission control methods.
Table 3. Admission control comparisons with existing methods.
Table 3. Admission control comparisons with existing methods.
ClassificationObjectiveStrategyRelated StudiesProposed Methods
Admission controlMaximum QoS and throughputSMDP; priority; no-priority; 0/1 knapsack[22,23,24,25]Non-priority and Lagrangian multipliers are adopted
Comparison with related studies
The semi-Markov decision process (SMDP) approach relies on partial data, such as resources, bandwidth, and quality of service (QoS), to decide whether to accept or reject a task. SMDP computing time is inefficient for jobs with time constraints, such as delay and delay tolerance sensitive applications. The priority of user information is unknown for decision making under a computing time constraint. In this paper, call admission control mechanisms within the conditions with non-priority and QoS constraints are jointly considered. The Lagrangian relaxation-based approach is proposed to maximize system revenue combined with the proposed resource allocation methods to admit tasks and satisfy application requirements appropriately.
The SMDP-based approach relies on partial data. The computing time is also inefficient for jobs with time constraints, such as delay and delay tolerance sensitive applications. The priority of user information is unknown for decision making under a computing time constraint. In this paper, call admission control mechanisms within the conditions with non-priority and QoS constraints are jointly considered. The Lagrangian relaxation-based approach is proposed to maximize system revenue combined with the proposed resource allocation methods to appropriately admit tasks and satisfy application requirements.

2.4. Research Scope

This research focuses on resource management in various scenarios in 5G network slices. An optimization-based approach (LR) is used to solve the mathematical programming problem to maximize system revenue. In our proposed cloud computing framework (cloudlets and core clouds) in 5G network slices, two primary research questions are considered. What is the most efficient algorithm for resource allocation within admission control, resource scheduling, and load-balancing policies in a limited-resource cloud computing environment? Is the HA of VMs considerably influenced in the system?
Additionally, the problem is addressed under rapidly increasing data traffic conditions and a limited resource pool to obtain near-optimal policies using a combination of LR approaches. The solutions satisfy the QoS-related constraints in transmission and stand on computation perspectives to fulfill the throughput, delay, and delay jitter requirements. They are compared with other resource management schemes for efficiency and effectiveness.

3. Mathematical Formulation

A mathematical model is proposed to manage an optimization programming problem. It focuses on developing a well-designed algorithm in a parallel computing environment for admission control, resource scheduling, and an HA arrangement to maximize cloud service provider profits in 5G network slices. Based on the combined cloudlet and core cloud network architecture (Figure 1), VMs with VNFs are requested by various applications and require specific resources. They should be allocated to the server k with computation and transmission capacities concerning CPU, RAM, storage, and internal or shared bandwidth, expressed as P k , M k , H k , B k n , and B c , respectively, in cloudlets and core clouds. The proposed model in the orchestrator supports network resource management in front of the cloudlet and core cloud environments when assuming that resource management strategies are followed when facing a batch of requests. Table 4 and Table 5 present lists of the given parameters and decision variables.
The system’s profit is the combination of the rewards of admitting applications, the penalties of rejecting the other applications, and the cost for turning on the servers.
The objective function Z IP is shown in Equation (1), and its goal is to maximize system profits among all VMs requested by the applications.
Z IP = i I R i y i i I P i ( 1 y i ) s S β s z s
To acquire the highest revenue, the optimization problem is shown as:
Objective function : max Z IP = min Z IP subject to : C 1 : ( | W i | + | T i | ) y i j W i s S a i j s + T i s S b i s , i I , C 2 : s S a i j s 1 , i I , j W i , C 3 : s S b i s 1 , i I , T i , C 4 : i I j W i a i j s + i I T i b i s V s , s S , C 5 : i I j W i D i j a i j s + i I T i D i b i s E s , s S , C 6 : i I j W i G i j a i j s + i I T i G i b i s M s , s S , C 7 : i I j W i Q i j a i j s + i I T i Q i b i s H s , s S , C 8 : i I j W i T i X i j f i j s B s n , s S , C 9 : i I j W i T i X i j ( 1 f i j s ) B c , s S , C 10 : a i j s z s , i I , j W i , s S , C 11 : b i s z s , i I , T i , s S , C 12 : a i j s + b i s f i j s + 1 , i I , j W i , T i , s S .
where C1–C3 belong to the admission control and assignment constraints, C4–C9 belong to the capacity constraints, and C10–C12 belong to the HA constraints.
For admission control and assignment constraints, ( | W i | + | T i | ) is the total number of standard or additional HA VMs required by an application, in which the VMs are in disjoint sets, where W i T i = in constraint C1. s S a i j s is the number of standard VMs that must be admitted and allocated into servers. s S b i s is the number of admitted and allocated HA VMs. The total number of allocated HA and standard VMs should be greater than or equal to the requirement, as shown in constraint C1. That is, while application i is completely served, the summation of a i j s and b i s should be greater than and equal to the demand on the right-hand side of constraint C1.
s S a i j s and s S b i s are shown in the constraints C2 and C3, which means when the value of s S a i j s or s S b i s is less than or equal to one, the VMs are inseparably assigned to servers. In other words, a virtual machine is not partially allocated to servers.
For capacity constraints, the resources offered by a server are defined as a set of four factors: the maximum number of VMs in the server s ( V s ), the processing capacity of each CPU core ( E s ), the RAM capacity ( M s ), and the storage capacity ( H s ). The total resources required by VMs for each server cannot exceed its available resources as formulated in the constraints C4–C7.
We also set the internal bandwidth rate ( B s n ) in the server s for internal transmission between VMs. For example, two applications, i and i + 1 , and their index sets of VMs are W i = { 1 i , 2 i } , T i = { 3 i } , W i + 1 = { 1 i + 1 , 2 i + 1 } , and T i + 1 = { 3 i + 1 } , as shown in Figure 2. The link between VM 3 i and VM 2 i + 1 represents the internal bandwidth required by VM 3 i to connect to VM 2 i + 1 in server s. The constraint C8 indicates that the total bandwidth required by all the VMs should not exceed the internal bandwidth B s n in server s. The external bandwidth rate ( B c ) is set for the transmission between servers in the cloud computing environment (cloudlets and core clouds), as illustrated in Figure 2. The link between server s 1 and server s represents the bandwidth requested by VM 1 i to connect to VM 3 i . The constraint C9 indicates that the total bandwidth required by all VMs accepted by servers in a cloud should not exceed the external bandwidth.
In this work, we assume two kinds of VMs, standard VMs and VMs with HA, of the same application, cannot be assigned to the same server, as shown in Figure 2. In the beginning, all servers contain no VMs; then, for example: application i comes in and needs two standard VMs and one VM with HA, as mentioned in the previous paragraph. First, we put two standard VMs, VM 1 i and VM 2 i , in server s 1 , while the two kinds of VMs cannot be assigned to the same server, then VM 3 i has to be assigned to server s. To continue, the next application i + 1 comes in, and we put its standard VMs 1 i + 1 and 2 i + 1 in server s 1 and server s, respectively, while server s 1 and server s have no residual capacity to handle one more VM. Meanwhile, the VM with HA, VM 3 i + 1 , cannot be put in server s 1 or server s due to our exclusive assumption. Thus, VM 3 i + 1 is assigned to server s + 1 .
For HA configuration constraints, the decision variable f i j s for application i is set for the VM assignments. The index sets j and for different kinds of VMs must be collocated to server s and separated into different servers, where j W i and T i . The relationships are expressed as the constraint C12. The constraints C10 and C11 indicate the server power status. If any VM is admitted and assigned to server s, server s must be powered on, and z s should be set to one. The constraint C12 assures that when VMs j and are assigned to the same server s, which means that a i j s and b i s are both set to one, the exclusive setting f i j s must also be one. In other words, the standard VM and the HA VM requested from the same application cannot be allocated to the same server s.

4. Lagrangian Relaxation-Based Solution Processes

The Lagrangian relaxation method is proposed for solving large-scale mathematical programming problems, including optimization problems with linear, integer, and nonlinear programming problems in many practical applications [26]. The key idea is to relax complicated constraints into a primal optimization problem and extend feasible solution regions to simplify the primal problem. Based on the relaxation, the primal problem is transformed into an LR problem associated with Lagrangian multipliers [27,28,29]. Figure 3 illustrates the six procedural steps in the LR method.

4.1. Procedures of Step 1: Relaxation

To separate the feasible region of the primal problem into several subproblems with an independent set of decision variables, the primal problem is transformed into an LR problem. The relaxation method associated with Lagrangian multipliers is applied to the constraints C1 and C4–C13 in Step 1, as presented in Figure 3. Then, the constraints C1 and C4–C12 with the multipliers are added to the primal problem Equation (1), as shown in Equation (2) and denoted as Z LR .
Z LR = i I R i y i + i I P i ( 1 y i ) + s S β s z s + i I μ i 1 [ ( | W i | + | T i | ) y i j W i s S a i j s T i s S b i s ] + s S μ s 2 [ i I j W i a i j s + i I T i b i s V s ] + s S μ s 3 [ i I j W i D i j a i j s + i I T i D i b i s E s ] + s S μ s 4 [ i I j W i G i j a i j s + i I T i G i b i s M s ] + s S μ s 5 [ i I j W i Q i j a i j s + i I T i Q i b i s H s ] + s S μ s 6 [ i I j W i T i X i j f i j s B s n ] + s S μ s 7 [ i I j W i T i X i j ( 1 f i j s ) B c ] + i I j W i s S μ i j s 8 [ a i j s z s ] + i I T i s S μ i s 9 [ b i s z s ] + i I j W i T i s S μ i j s 10 [ a i j s + b i s f i j s 1 ]
Then, the optimization problem can be reformulated as:
Objective function : min Z LR subject to : C 2 , C 3 ,
where y i { 0 , 1 } , a i j s { 0 , 1 } , b i s { 0 , 1 } , f i j s { 0 , 1 } , and z s { 0 , 1 } .

4.2. Procedures of Steps 2 and 3: Decomposition and Solving Subproblems

The LR problem can be decomposed into several independent subproblems, with their related decision variables. The divide-and-conquer approach is used to solve the subproblems correspondingly.

4.2.1. Subproblem 1 (Related to y i )

By extracting items with decision variable y i , the optimization problem of Subproblem 1 can be developed as:
Objective function : min i I ( R i P i + μ i 1 | W i | + μ i 1 | T i | ) y i subject to : y i { 0 , 1 } .
Equation (3) can be divided into | I | independent subproblems. For each application i, where i I , the decision variable y i is set to one when the coefficient ( R i P i + μ i 1 | W i | + μ i 1 | T i | ) is less than zero. Otherwise, y i is set to zero. The run time is O ( | I | ) , and the pseudocode is illustrated in Algorithm 1.
Algorithm 1: Subproblem 1.
  • Input: Given parameters R , P , W , T and Lagrangian
  •     multipliers μ 1 .
  • Output: Decision variable y .
  • Initialize: y i 0 , i I
  • for i = 0 to ( | I | 1 ) do
  •      c R i P i + μ i 1 | W i | + μ i 1 | T i |
  •     if c < 0 then
  •          y i 1
  •     end if
  • end for

4.2.2. Subproblem 2 (Related to a i j s )

By extracting items with decision variable a i j s , the optimization problem of Subproblem 2 can be developed as:
Objective function : min i I j W i s S ( μ i 1 + μ s 2 + μ s 3 D i j + μ s 4 G i j + μ s 5 Q i j + μ i j s 8 + T i μ i j s 10 ) a i j s subject to C 2 : s S a i j s 1 , i I , j W i , a i j s { 0 , 1 } .
Equation (4) can be divided into | I | | W i | | S | cases. The decision variable a i j s is set to one, and the minimum coefficient ( μ i 1 + μ s 2 + μ s 3 D i j + μ s 4 G i j + μ s 5 Q i j + μ i j s 8 + T i μ i j s 10 ) is less than zero and corresponds to alliance subindices i, j, and s. To satisfy the constraint C2, the decision variable should be set to one only if the minimum coefficient with subindex s is a required sorting processes. Otherwise, a i j s is set to zero. The run time is O ( | I | | W i | | S | | T i | ) . The pseudocode is illustrated in Algorithm 2.
Algorithm 2: Subproblem 2.
  • Input: Given parameters D , G , Q and Lagrangian
  •     multipliers μ 1 , μ 2 , μ 3 , μ 4 , μ 5 , μ 8 , μ 10 .
  • Output: Decision variable a .
  • Initialize: a i j s 0 , i I , j W i , s S
  • for i = 0 to ( | I | 1 ) do
  •     for j = 0 to ( | W i | 1 ) do
  •         for s = 0 to ( | S | 1 ) do
  •             c s μ i 1 + μ s 2 + μ s 3 D i j + μ s 4 G i j + μ s 5 Q i j
  •                    + μ i j s 8 + T i μ i j s 10
  •         end for
  •         Find the index m that c m has the minimum value
  •         in c .
  •         if c m < 0 then
  •             a i j m 1
  •         end if
  •     end for
  • end for

4.2.3. Subproblem 3 (Related to b i s )

By extracting items with decision variable b i s , the optimization problem of Subproblem 3 can be developed as:
Objective function : min i I T i s S ( μ i 1 + μ s 2 + μ s 3 D i + μ s 4 G i + μ s 5 Q i + μ i s 9 + j W i μ i j s 10 ) b i s subject to C 3 : s S b i s 1 , i I , T i , b i s { 0 , 1 } .
The solution process of Equation (5) is similar to that of Equation (4) and can be also divided into | I | | T i | | S | subproblems. The decision variable b i s is set to one when the minimum coefficient ( μ i 1 + μ s 2 + μ s 3 D i + μ s 4 G i + μ s 5 Q i + μ i s 9 + j W i μ i j s 10 ) is less than zero and corresponds to alliance subindices i, , and s. Otherwise, b i s is set to zero. The run time of this subproblem is O ( | I | | T i | | S | | W i | ) . The pseudocode is illustrated in Algorithm 3.
Algorithm 3: Subproblem 3.
  • Input: Given parameters D , G , Q and Lagrangian
  •     multipliers μ 1 , μ 2 , μ 3 , μ 4 , μ 5 , μ 9 , μ 10 .
  • Output: Decision variable b .
  • Initialize: b i s 0 , i I , T i , s S
  • for i = 0 to ( | I | 1 ) do
  •     for = 0 to ( | T i | 1 ) do
  •         for s = 0 to ( | S | 1 ) do
  •             c s μ i 1 + μ s 2 + μ s 3 D i + μ s 4 G i + μ s 5 Q i
  •                + μ i s 9 + j W i μ i j s 10
  •         end for
  •         Find the index m that c m has the minimum value
  •         in c .
  •         if c m < 0 then
  •             b i m 1
  •         end if
  •     end for
  • end for

4.2.4. Subproblem 4 (Related to f i j s )

By extracting items with decision variable f i j s , the optimization problem of Subproblem 4 can be developed as:
Objective function : min i I j W i T i s S ( μ s 6 X i j μ s 7 X i j μ i j s 10 ) f i j s subject to : f i j s { 0 , 1 } .
Equation (6) can be divided into | I | | W i | | S | | T i | cases. For the alliance subindices i, j, , and s, the decision variable f i j s is set to one when the coefficient ( μ s 6 X i j μ s 7 X i j μ i j s 10 ) is less than zero. Otherwise, f i j s is set to zero. The run time is O ( | I | | W i | | S | | T i | ) . The pseudocode is illustrated in Algorithm 4.
Algorithm 4:Subproblem 4.
  • Input: Given parameters X and Lagrangian multipliers
  •      μ 6 , μ 7 , μ 10 .
  • Output: Decision variable f .
  • Initialize: f i j s 0 , i I , j W i , T i , s S
  • for i = 0 to ( | I | 1 ) do
  •     for j = 0 to ( | W i | 1 ) do
  •         for = 0 to ( | T i | 1 ) do
  •            for s = 0 to ( | S | 1 ) do
  •                 c μ s 6 X i j μ s 7 X i j μ i j s 10
  •                if c < 0 then
  •                     f i j s 1
  •                end if
  •            end for
  •         end for
  •     end for
  • end for

4.2.5. Subproblem 5 (Related to z s )

By extracting items with decision variable z s , the optimization problem of Subproblem 5 can be developed as:
Objective function : min s S [ β s i I ( j W i μ i j s 8 + T i μ i s 9 ) ] z s subject to : z s { 0 , 1 } .
Equation (7) can be divided into | S | | I | | W i | or | S | | I | | T i | cases. In each case with subindex s, the decision variable z s is set to one when the coefficient [ β s i I ( j W i μ i j s 8 + T i μ i s 9 ) ] , which corresponds to alliance subindex s, is less than zero. Otherwise, z s is set to zero. The run time is O ( | S | | I | | W i | ) or O ( | S | | I | | T i | ) , and the pseudocode is illustrated in Algorithm 5.
Algorithm 5: Subproblem 5.
  • Input: Given parameters β and Lagrangian multipliers
  •      μ 8 , μ 9 .
  • Output: Decision variable z .
  • Initialize: z s 0 , s S
  • for s = 0 to ( | S | 1 ) do
  •      c 0
  •     for i = 0 to ( | I | 1 ) do
  •         for j = 0 to ( | W i | 1 ) do
  •             c c + μ i j s 8
  •         end for
  •         for = 0 to ( | T i | 1 ) do
  •             c c + μ i s 9
  •         end for
  •     end for
  •     if β s c < 0 then
  •          z s 1
  •     end if
  • end for

4.3. Procedure of Step 4: Dual Problem and the Subgradient Method

According to the weak Lagrangian duality theorem [30], the objective values of the LR problem Z LR are the lower bounds (LBs) of the primal problem Z IP with multiples μ i 1 , μ s 2 , μ s 3 , μ s 4 , μ s 5 , μ s 6 , μ s 7 , μ i j s 8 , μ i s 9 , μ i j s 10 0 , i I , j W i , T i , s S . The formulation of the dual problem (D) is constructed to calculate the tightest LB ( max Z D ), where max Z D = min Z LR . Then, the dual problem can be formulated as:
Objective function : max Z D subject to : μ 1 0 , μ 2 0 , μ 3 0 , μ 4 0 , μ 5 0 , μ 6 0 , μ 7 0 , μ 8 0 , μ 9 0 , μ 10 0 .
The subgradient method is commonly used for solving the dual problem by iteratively updating the Lagrangian multipliers [26,31,32].
First, let vector S be a subgradient of Z D . In the q th iteration of the subgradient optimization procedure, the multiplier vector π q = ( μ 1 , q , μ 2 , q , μ 3 , q , μ 4 , q , μ 5 , q , μ 6 , q , μ 7 , q , μ 8 , q , μ 9 , q , μ 10 , q ) is updated by π q + 1 = π q + t q S q . The step size t q is determined by equation t q = λ ( Z IP h Z D ( π q ) ) S q 2 . The denominator S q is the sum of relaxed constraints concerning to the decision variable values based on the q th iteration. Z IP h is the primal objective value in the h th iteration. Z D ( π q ) is the objective value of the dual problem in the q th iteration. λ is a constant, where 0 λ 2 . Accordingly, the optimal objective value of the dual problem is obtained iteratively.

4.4. Procedure of Step 5: Obtaining the Primal Feasible Solutions

Applying the LR and the subgradient methods to solve the LR and dual problems determines a theoretical LB from the primal feasible solution. Crucial information regarding the primal feasible solution can be identified [28]. The feasible region of a mathematical programming problem defined by the solutions must be satisfied by all constraints. A set of primal feasible solutions to Z IP is a subset of the infeasible region solutions to Z LR or Z D . Several alternative methods can be used sophisticatedly to obtain the primal feasible solutions from the observations of the infeasible region. For example, Figure 4 presents an experimental case. The green curve represents the process of obtaining the primal feasible solutions iteratively. The objective is to identify the minimum value of the primal problem ( Z IP h ). Then, the LBs are determined using the subgradient method to iteratively obtain the tightest LB (max Z D ), represented by the purple line. The proposed resource management approach for obtaining primal feasible solutions is called the drop-and-add algorithm. It is a heuristic-based allocation mechanism for optimizing the objective function. It searches for a solution that satisfies not only all the user demands, but also the constraints. The initial solution is adopted using next-fit (NF) or first-fit (FF) as baselines for evaluating solution quality [33]. The proposed algorithm, FF, and NF are simultaneously implemented for result comparison.

4.5. Procedure of Step 6: Drop-and-Add Algorithm

Lagrangian multipliers determined from the dual problem have significant values for evaluating the sensitivity of objective value improvement [26,28,30,32]. Through the subproblems, the integrated weighting factor of applications is represented in (3) as ( R i P i + μ i 1 | W i | + μ i 1 | T i | ) ; therefore, the sum of the values can be used as an index to interpret the application’s significance of i. The corresponding multipliers determine the ordering set of admitting, assigning, and scheduling decision variables. The other decision variables a i j s or b i s and the assignment of indices j, , and s can be regarded as two bin-packing problems. The VM j can be packed into the server s. The VM can be packed into servers except server s to comply with the HA conditions for b i s . For performance evaluation, NF is adopted for sequentially performing the algorithm of assignment for a i j s and then b i s . The flowchart of the drop-and-add algorithm is shown in Figure 5.

5. Computational Experiments

A cloud service provider’s resource parameters and the applications’ demand attributes were simulated in a cloud computing experimental environment and are presented in Table 6. U ( x , y ) means a number uniform distributed between the parameters of x and y.
The algorithms were constructed and implemented to analyze solution quality and improvement ratios (IRs) in several simulation cases. Solution quality is defined as the objective value gap (denoted as GAP) between the proposed algorithm and the LR problem, which is expressed as GAP = | V Drop&Add V LR | | max ( V Drop&Add , V LR ) | × 100 % , where V Drop&Add is the objective value of applying the drop-and-add algorithm and V LR is the objective value of the LR problem. IR NF is expressed as IR NF = V Drop&Add V NextFit | max ( V NextFit , V Drop&Add ) | × 100 % . IR FF is expressed as IR FF = V Drop&Add V FirstFit | max ( V FirstFit , V Drop&Add ) | × 100 % , where V NextFit and V FirstFit are the objective values of employing the NF algorithm or FF algorithm, respectively. The experiments were developed in Python and implemented in a VM on a workstation with a quad-core CPU, 8 GB RAM, and Ubuntu 14.04. The traffic loads of tasks and arrival time intervals were randomly generated. The experimental environment was initialized for a data center including edge and core clouds. VMs were requested by applications to represent computing requirements. Based on resource admission control strategies and scheduling models, the VMs were packed into the corresponding servers in the cloud computing environments.

5.1. Performance Evaluation Case: Traffic Load

This experiment was designed to analyze VMs requested by applications in a time slot, which can be interpreted as a snapshot of the system loading. First, the trend of the objective function with the number of applications as the control variable was examined. The result in Figure 6 indicates that the objective value of (IP)increases with the number of applications. The drop-and-add and LB values were almost the same in some cases (20∼80), indicating that the optimal solution was determined when the GAP was less than 0.80% (the minimized one was 0.60%). Table 7 compares drop-and-add with NF or FF (higher values are preferable) in the primal problem. The maximum IR was 89.48%, with 180 applications in both FF and NF. The penalty of unsatisfied applications significantly increased the objective value for indicating when arriving applications exceeded system capacity with the NF or FF algorithm in the cases of the number of applications being over 100. Otherwise, the drop-and-add algorithm can select valuable applications to pack into servers, which results in a higher objective value than NF or FF in the cases of the number of applications being over 100.

5.2. Performance Evaluation Case: Number of Servers

The cost function is emulated as the capital expenditure (CAPEX) of the cloudlets and core clouds related to network infrastructure deployment of appropriate levels of servers, where the budget is a significant constraint. The cost difference between high- and low-end servers is also typically significant. Service providers generally conduct a comprehensive system analysis to determine the methods of how to manage, control, and operate a system appropriately. The plans, such as turning servers on or off, provide sufficient QoS to applications efficiently and effectively. Furthermore, the servers purchased were deployed at four levels with different capacities in this experiment. Nonhomogeneous servers were deployed under the same limited budget. The rapid increases in data traffic are represented as traffic loads to determine which method delivered superior QoS for applications at affordable costs and preserved revenue. The following are the experimental scenarios tested. Figure 7 and Table 8 present the results of the proposed methods (drop-and-add, LB, NF, and FF). The drop-and-add algorithm attained the most practical objective value (higher is preferable) compared with NF and FF in the case of 40.
Regarding the GAP, some values in the other situations were calculated to determine the difference in rate value between drop-and-add and LB. The GAP values indicate that the minimum value (44.62%) was determined in one of the cases with numerous servers (40). A data center has sufficient resource space in servers with the drop-and-add, NF, and FF algorithms. The maximum improvement ratio is represented by the ratio of improvement of feasible solutions and was 193.73% in numerous servers (i.e., 40). The result reveals that the resource allocation algorithm has a significant impact on system performance with a limited resource constraint.

5.3. Performance Evaluation Case: Effect of HA

As far as crucial business requirements are concerned, cloud service providers should offer high quality, excellent availability, good performance, and reliable services. In this case, the applications are divided into levels by using VM replication for HA requests. The VMs requesting to be in the T i set with HA by the application i asking for VM placement must be mutually and exclusively allocated into different physical servers. The level of exclusivity is called the dual-rate. In the subsequent experiment, the dual-rate was the parameter configured for HA VMs. Moreover, a dual-rate of 0.5 indicates that the HA VMs, | T i | , requested by half of | W i | for application i are allocated to different servers. The dual-rate equals 0.3, which indicates 30% of standard VMs required for application i with HA capability. In Figure 8, it is evident that the dual-rate significantly affected the application satisfaction rate. Thus, the drop-and-add algorithm offers more benefits (higher values are preferable) than NF or FF in the dual-rate cases. The improvement ratios increased significantly when the dual-rate was higher than 0.3, and the maximum IR was 608.39%. The drop-and-add algorithm achieved flexibility and efficiency and could sufficiently obtain the maximum objective value to deal with HA requests. As observed in Table 9, FF and NF performed poorly when the dual-rate was beyond 0.3. The tasks were not assigned to appropriate servers with HA. This resulted in turning on more servers, which corresponded to cost generation.

5.4. Performance Evaluation Case: Time complexity comparison

Table 10 shows the time complexity of the LR-based solution for resource management in a sliced network. We added an existing scheme, brute force, for comparison in Table 10. The time complexity of this scheme was higher than the proposed LR-based algorithm. Furthermore, the corresponding explanations were included to support our statements. The time complexity of the proposed LR-based solutions was O ( N | I | | W i | | T i | | S | ) . Table 10 shows the time complexity of the LR-based solution. The Lagrange dual solution was determined by Sub-problems (4)–(6), which were solved using the minimum algorithm with | S | servers. Each sub-problem required O ( | I | | W i | | T i | ) time minimum coefficient among servers. Since the sub-problems were solved individually by divide-and-conquer algorithms, the time complexity for each of the sub-problems was constant. Thus, the worst case of time complexity among these subproblems was considered significant in each iteration. The Lagrange dual solutions for the sub-problems could be obtained after a maximum number of iterations N, and the time complexity was O ( N | I | | W i | | T i | | S | ) . The number of iterations pre-defined to converge was about 600 based on the experiment results shown in Figure 4. Fortunately, all given parameters and multipliers for the solution did not need the same initial values. The convergence was achieved in a small number of iterations with the previous execution results. Furthermore, the proposed algorithms can be set as the output results at any complete iteration. Thus, the time can be controlled in practice.

5.5. Discussion

To test the proposed algorithm, we designed three experiments involving changes in application demands, the cost of servers, and dual-rates for HA with application requests, as shown in Table 11. The research problem was formulated as a mathematical programming problem to determine both admission control and resource scheduling problems from a service provider perspective. Overall, the drop-and-add algorithm is the most effective for obtaining the optimal solution in the fewest iterations. The optimization-based energy-efficient admission control and resource allocation algorithms have significant benefits in cloud computing systems. The significance of applications was inspired by the coefficient of (3), which sorts applications by a composition of application weights, client satisfaction, and high availability. Therefore, using the LR-based solution approach with multipliers to obtain the primal feasible solutions was suitable for allocating VMs to servers efficiently and effectively. The mathematical formulation was decomposed into five subproblems. It was solved optimally using parallel computation techniques to reduce computation time substantially. Furthermore, the experimental results reveal and confirm that the drop-and-add algorithm was active in a few minutes by the most significant in the sliced network stages shown in Table 7, Table 8, Table 9, Table 10 and Table 11. The following suggested problems and limitations of this paper can be further studied and solved: QoS requirements can be classified into more categories. Resource sharing between multiple service providers is a new research problem that necessitates considering related issues, such as the sharing economy, from multiple network service provider perspectives. Hence, operator pricing policies could be a worthwhile topic.

6. Summary and Conclusions

A mathematical programming model is used to develop resource management by simulating the cloud service provider role in this paper’s cloud computing systems in 5G network slices. The mathematical model is solved using an LR-based approach. The proposed algorithm increases the cloud computing network infrastructure’s flexibility, including cloudlets, and core clouds, to maximize rewards by admitting as many applications as possible. The gaps between upper bounds and lower bounds in the computational experiments demonstrate the drop-and-add heuristic optimal solution qualities. The main contribution is demonstrating that the orchestrator designed the resource management algorithm significantly determined using Lagrangian multipliers to indicate task significance. A promising and straightforward resource allocation approach is proposed to combine client satisfaction and high availability for network planning in a sliced network. The developed optimization-based efficient admission control and resource allocation algorithms are confirmed through various experimental cases. The proposed method has excellent effectiveness and efficiency compared with the LB, FF, and NF solutions based on the experimental results. The resource management mechanisms enables the slicing network as services to efficiently and maximize system revenue in 5G networks.

Author Contributions

Conceptualization, C.-H.H. and F.Y.-S.L.; Formal analysis, C.-H.H. and F.Y.-S.L.; Methodology, C.-H.H., F.Y.-S.L., Y.-C.S., Y.-S.W. and H.-Y.K.; Supervision, F.Y.-S.L., Y.-F.W. and Y.H.; Visualization, C.-H.H. and Y.-F.C.; Writing review and editing, C.-H.H., E.S.-H.F., Y.-F.C. and Y.-F.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by Ministry of Science and Technology (MOST), Taiwan, under Grant Number MOST 109-2221-E-002-144 and MOST 110-2222-E-001-002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ordonez-Lucena, J.; Ameigeiras, P.; Lopez, D.; Ramos-Munoz, J.J.; Lorca, J.; Folgueira, J. Network Slicing for 5G with SDN/NFV: Concepts, Architectures, and Challenges. IEEE Commun. Mag. 2017, 55, 80–87. [Google Scholar] [CrossRef] [Green Version]
  2. Checko, A.; Christiansen, H.L.; Yan, Y.; Scolari, L.; Kardaras, G.; Berger, M.S.; Dittmann, L. Cloud RAN for Mobile Networks—A Technology Overview. IEEE Commun. Surv. Tutor. 2015, 17, 405–426. [Google Scholar] [CrossRef] [Green Version]
  3. Chen, M.; Hao, Y. Task Offloading for Mobile Edge Computing in Software Defined Ultra-Dense Network. IEEE J. Sel. Areas Commun. 2018, 36, 587–597. [Google Scholar] [CrossRef]
  4. Taleb, T.; Ksentini, A.; Sericola, B. On Service Resilience in Cloud-Native 5G Mobile Systems. IEEE J. Sel. Areas Commun. 2016, 34, 483–496. [Google Scholar] [CrossRef] [Green Version]
  5. Li, X.; Samaka, M.; Chan, H.A.; Bhamare, D.; Gupta, L.; Guo, C.; Jain, R. Network Slicing for 5G: Challenges and Opportunities. IEEE Internet Comput. 2017, 21, 20–27. [Google Scholar] [CrossRef]
  6. Al-Fuqaha, A.; Guizani, M.; Mohammadi, M.; Aledhari, M.; Ayyash, M. Internet of Things: A Survey on Enabling Technologies, Protocols, and Applications. IEEE Commun. Surv. Tutor. 2015, 17, 2347–2376. [Google Scholar] [CrossRef]
  7. AWS Marketplace. High Availability. Available online: https://aws.amazon.com/marketplace/solutions/infrastructure-software/high-availability (accessed on 2 February 2021).
  8. Mao, Y.; You, C.; Zhang, J.; Huang, K.; Letaief, K.B. A Survey on Mobile Edge Computing: The Communication Perspective. IEEE Commun. Surv. Tutor. 2017, 19, 2322–2358. [Google Scholar] [CrossRef] [Green Version]
  9. Sotiriadis, S.; Bessis, N.; Amza, C.; Buyya, R. Elastic Load Balancing for Dynamic Virtual Machine Reconfiguration Based on Vertical and Horizontal Scaling. IEEE Trans. Serv. Comput. 2019, 12, 319–334. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, Y.; Lee, M.J.; Zheng, Y. Adaptive Multi-Resource Allocation for Cloudlet-Based Mobile Cloud Computing System. IEEE Trans. Mob. Comput. 2016, 15, 2398–2410. [Google Scholar] [CrossRef]
  11. Jin, A.; Song, W.; Zhuang, W. Auction-Based Resource Allocation for Sharing Cloudlets in Mobile Cloud Computing. IEEE Trans. Emerg. Top. Comput. 2018, 6, 45–57. [Google Scholar] [CrossRef]
  12. Liu, L.; Fan, Q. Resource Allocation Optimization Based on Mixed Integer Linear Programming in the Multi-Cloudlet Environment. IEEE Access 2018, 6, 24533–24542. [Google Scholar] [CrossRef]
  13. Ahmad, A.; Arshad, R.; Mahmud, S.A.; Khan, G.M.; Al-Raweshidy, H.S. Earliest-Deadline-Based Scheduling to Reduce Urban Traffic Congestion. IEEE Trans. Intell. Transp. Syst. 2014, 15, 1510–1526. [Google Scholar] [CrossRef]
  14. Meneguette, R.I.; Boukerche, A.; Pimenta, A.H.M. AVARAC: An Availability-Based Resource Allocation Scheme for Vehicular Cloud. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3688–3699. [Google Scholar] [CrossRef]
  15. Garey, M.R.; Johnson, D.S. Computers and Intractability: A Guide to the Theory of NP-Completeness; W. H. Freeman & Co.: San Francisco, CA, USA, 1990. [Google Scholar]
  16. Hou, I.; Gupta, P. Proportionally Fair Distributed Resource Allocation in Multiband Wireless Systems. IEEE/ACM Trans. Netw. 2014, 22, 1819–1830. [Google Scholar] [CrossRef]
  17. Ding, T.; Hao, Q.; Zhang, B.; Zhang, T.; Huai, L. Scheduling Policy Optimization in Kernel-Based Virtual Machine. In Proceedings of the International Conference on Computational Intelligence and Software Engineering (CiSE), Wuhan, China, 10–12 December 2010; pp. 1–4. [Google Scholar]
  18. Zhao, L.; Lu, L.; Jin, Z.; Yu, C. Online Virtual Machine Placement for Increasing Cloud Provider’s Revenue. IEEE Trans. Serv. Comput. 2017, 10, 273–285. [Google Scholar] [CrossRef]
  19. Jia, M.; Liang, W.; Xu, Z.; Huang, M. Cloudlet Load Balancing in Wireless Metropolitan Area Networks. In Proceedings of the IEEE International Conference on Computer Communications (IEEE INFOCOM), San Francisco, CA, USA, 10–14 April 2016; pp. 1–9. [Google Scholar]
  20. Yao, D.; Gui, L.; Hou, F.; Sun, F.; Mo, D.; Shan, H. Load Balancing Oriented Computation Offloading in Mobile Cloudlet. In Proceedings of the IEEE Vehicular Technology Conference (VTC-Fall), Toronto, ON, Canada, 24–27 September 2017; pp. 1–6. [Google Scholar]
  21. Nguyen, D.D.; Nguyen, H.X.; White, L.B. Reinforcement Learning With Network-Assisted Feedback for Heterogeneous RAT Selection. IEEE Trans. Wirel. Commun. 2017, 16, 6062–6076. [Google Scholar] [CrossRef]
  22. Yuan, H.; Bi, J.; Tan, W.; Li, B.H. CAWSAC: Cost-Aware Workload Scheduling and Admission Control for Distributed Cloud Data Centers. IEEE Trans. Autom. Sci. Eng. 2016, 13, 976–985. [Google Scholar] [CrossRef]
  23. Bashar, A. BN-Based Approach for Predictive Admission Control of Cloud Services. In Proceedings of the IEEE International Advance Computing Conference (IACC), Hyderabad, India, 5–7 January 2017; pp. 59–64. [Google Scholar]
  24. Hoang, D.T.; Niyato, D.; Wang, P. Optimal Admission Control Policy for Mobile Cloud Computing Hotspot with Cloudlet. In Proceedings of the 2012 IEEE Wireless Communications and Networking Conference (WCNC), Paris, France, 1–4 April 2012; pp. 3145–3149. [Google Scholar]
  25. Xia, Q.; Liang, W.; Xu, W. Throughput Maximization for Online Request Admissions in Mobile Cloudlets. In Proceedings of the IEEE Conference on Local Computer Networks (LCN), Sydney, NSW, Australia, 21–24 October 2013; pp. 589–596. [Google Scholar]
  26. Geoffrion, A. Lagrangian Relaxation and its Uses in Integer Programming. Math. Program. 1974, 2, 82–114. [Google Scholar]
  27. Bertsekas, D.P. Multiplier Methods: A Survey. Automatica 1976, 12, 133–145. [Google Scholar] [CrossRef] [Green Version]
  28. Bertsekas, D.P. Constrained Optimization and Lagrange Multiplier Methods; Academic Press: Cambridge, MA, USA, 1982. [Google Scholar]
  29. Hestenes, M.R. Multiplier and Gradient Methods. J. Optim. Theory Appl. 1969, 4, 303–320. [Google Scholar] [CrossRef]
  30. Fisher, M.L. The Lagrangian Relaxation Method for Solving Integer Programming Problems. Manag. Sci. 2004, 50, 1861–1871. [Google Scholar] [CrossRef] [Green Version]
  31. Rockafellar, R.T. A Dual Approach to Solving Nonlinear Programming Problems by Unconstrained Optimization. Math. Program. 1973, 5, 354–373. [Google Scholar] [CrossRef] [Green Version]
  32. Fisher, M.L. An Applications Oriented Guide to Lagrangian Relaxation. Interfaces 1985, 15, 10–21. [Google Scholar] [CrossRef]
  33. Xu, X.; Zhang, J.; Ji, Y.; Li, H.; Gu, R.; Yu, H.; Zhang, J. BBU Aggregation for Maximizing the Resource Utilization in Optical-Enabled Cloud Radio Access Networks. In Proceedings of the International Conference on Optical Communications and Networks (ICOCN), Hangzhou, China, 24–27 September 2016; pp. 1–3. [Google Scholar]
Figure 1. Network topology in slicing networks.
Figure 1. Network topology in slicing networks.
Sensors 21 01882 g001
Figure 2. Representation of internal and external transmissions between VMs in servers.
Figure 2. Representation of internal and external transmissions between VMs in servers.
Sensors 21 01882 g002
Figure 3. Lagrangian relaxation-based solution process flow.
Figure 3. Lagrangian relaxation-based solution process flow.
Sensors 21 01882 g003
Figure 4. Obtaining primal feasible solutions and the tightest lower bound (LB). NF, next-fit.
Figure 4. Obtaining primal feasible solutions and the tightest lower bound (LB). NF, next-fit.
Sensors 21 01882 g004
Figure 5. Flowchart of the drop-and-add algorithm.
Figure 5. Flowchart of the drop-and-add algorithm.
Sensors 21 01882 g005
Figure 6. Objective value by the number of applications.
Figure 6. Objective value by the number of applications.
Sensors 21 01882 g006
Figure 7. Objective value by the number of servers.
Figure 7. Objective value by the number of servers.
Sensors 21 01882 g007
Figure 8. Objective value when scaling the dual-rate.
Figure 8. Objective value when scaling the dual-rate.
Sensors 21 01882 g008
Table 4. Given parameters.
Table 4. Given parameters.
NotationDescription
SIndex set of physical servers in the cloud computing system (cloudlets and core clouds), where  S = { 1 , 2 , , s , , | S | } .
IIndex set of applications on the cloud computing system, where I = { 1 , 2 , , i , , | I | } .
WIndex set of standard VMs, W = i = 1 I W i and W i = { 1 , 2 , , j , , | W i | } , where W i is also an index set of standard VMs required by application i, where i I .
TIndex set of VMs with high availability (HA), T = i = 1 I T i and T i = { 1 , 2 , , , , | T i | } , where T i is also an index set of VMs with HA required by application i, where i I .
NSet of total VMs, N = i = 1 I N i and N i = W i T i , where N i is the total number of standard VMs and VMs with HA required by application i, where i I .
rDual-rate; represents the ratio between the number of standard VMs and the VMs with HA, formulated as T i = W i × r , where i I .
B s n Internal transmission bandwidth of server s, where s S .
B c Shared transmission bandwidth within the cloud computing system (cloudlets and core clouds).
E s Processing capability of each central processing unit (CPU) core in server s, where s S .
M s Total random access memory (RAM) capacity in server s, where s S .
H s Total storage capacity in server s, where s S .
V s Maximum number of VMs allowable on server s, where s S .
β s Cost rate for opening server s, where s S .
R i Reward of admitting application i (application i can be admitted only if the demands on all types of resources are fully satisfied), where i I .
P i Penalty of rejecting application i (application i is rejected only if all of the types of resource requirements are not fully satisfied), where i I .
D i j Total CPU processing capability on VM j required by application i, where i I , j W i .
G i j Total RAM capability required by application i on VM j, where i I , j W i .
Q i j Total storage capability required by application i on VM j, where i I , j W i .
X i j Total transmission channel capacity required by application i between VMs j and , where i I , j W i , T i .
Table 5. Decision variables.
Table 5. Decision variables.
NotationDescription
y i Binary variable, 1 if the application i is completely allocated to and served in the computing system, and 0 otherwise, where i I .
a i j s Binary variable used for task admission control and assignment, 1 if the standard VM j of application i is admitted in the cloud computing networks and allocated to server s, and 0 otherwise, where i I , j W i , s S .
b i s Binary variable used for task admission control and assignment, 1 if the VM with HA of application i is admitted in the cloud computing networks and allocated to server s, and 0 otherwise, where i I , T i , s S .
f i j s Binary variable used for exclusive setting, 1 if VMs j and of application i are allocated to server s, and 0 otherwise, where i I , j W i , T i , s S .
z s Binary variable for server power-on or power-off status, 1 if server s is turned on, and 0 otherwise, where s S .
Table 6. Given parameters for the experiments.
Table 6. Given parameters for the experiments.
Given ParameterValue
Number of servers, | S | 56–84
Number of applications, | I | 24–52
Dual-rate, r0.5
Number of standard VMs for application i, | W i | | W i | U ( 1 , 10 ) , i I
Number of VMs with HA for application i, | T i | T i = W i × r , i I
Total number of VMs for application i, | N i | | N i | = | W i | + | T i | , i I
Host internal bandwidth capacity, B s n (Mbps)120–225
Shared transmission bandwidth within the cloud computing system, B c (Mbps)1000
Host CPU processing capacity, E s (GHz)480–900
Host memory capacity, M s (GB)120–225
Host storage capacity, H s (GB)1200–2250
The maximum number of VMs allowable on server s, V s 8–15
Cost rate for opening server s, β s β s = ( E s + M s + H s + V s ) / 40,000
Reward rate of each application, R i R i j ( D i j 50 ) , i I
Penalty rate of each application, P i P i j ( D i j 100 ) , i I
CPU requests of a task, D i j (GHz) D i j U ( 1 , 120 ) , i I , j W i
Memory requests of a task, G i j (GB) G i j U ( 1 , 30 ) , i I , j W i
Storage requests of a task, Q i j (TB) Q i j U ( 1 , 300 ) , i I , j W i
Total transmission channel capacity, X i j (Mbps) X i j U ( 0 , 30 ) , i I , j W i , T i
Bandwidth requests of a task, C i j s (Mbps) C i j s U ( 1 , 2000 ) , i I , j W i , s s S
Table 7. Comparison of solution qualities under different number of applications.
Table 7. Comparison of solution qualities under different number of applications.
Number of ApplicationsDrop-And-AddLBGap (%)NFFF IR NF (%) IR FF (%)
20123.743124.7600.80123.687123.6870.050.05
40337.910340.0800.64337.910337.9100.000.00
60506.729509.7800.60506.729506.7290.000.00
80623.692627.7600.65623.523623.5230.020.02
100772.445780.6801.06736.595736.5954.874.87
120934.295987.8455.73670.535670.53539.3339.33
140916.9851099.9119.94593.945593.94554.3954.39
160886.9251110.7025.23537.905537.90564.8964.89
180860.7051112.3229.23454.235454.23589.4889.48
In some cases (20∼80), the lower bound (LB) value is extremely close to, but not equal to that of the drop-and-add algorithm. In other cases, the drop-and-add algorithm has a significant improvement rate compared with NF or FF.
Table 8. Comparison of solution qualities under different number of servers.
Table 8. Comparison of solution qualities under different number of servers.
Number of ServersDrop-And-AddLBGap (%)NFFF IR NF (%) IR FF (%)
2034.439137.926300.5−75.601−75.601145.55145.55
40291.198421.11744.6299.13899.138193.73193.73
60291.880498.20170.69108.01108.01170.23170.23
80312.825528.89869.07116.41116.41168.73168.73
100307.056540.08875.89113.894113.894169.6169.6
Table 9. Comparison of solution qualities under different the dual-rate.
Table 9. Comparison of solution qualities under different the dual-rate.
Dual-RateDrop-and-AddLBGap (%)NFFF IR NF (%) IR FF (%)
0.1384.827387.2000.61384.827384.8270.000.00
0.2416.531419.0400.60416.531416.5310.000.00
0.3441.730452.6182.46351.926351.92625.5225.52
0.4438.900480.4939.48202.224202.224117.04117.04
0.5383.708492.96028.47158.026158.026142.81142.81
0.6269.938315.36616.8338.10638.106608.39608.39
Table 10. Time complexity comparison between the proposed LR-based approach and other schemes.
Table 10. Time complexity comparison between the proposed LR-based approach and other schemes.
AlgorithmTime ComplexityAnnotation
Equation (3) O ( | I | ) | I | subproblems determine the value of the decision variable for each application i.
Equation (4) O ( | I | | W i | | S | | T i | ) | I | | W i | subproblems with summation of T i , where i I , determine the binary decision variable a i j s that the coefficient is minimized among | S | .
Equation (5) O ( | I | | T i | | S | | W i | ) | I | | T i | subproblems with summation of j W i , where i I determine the binary decision variable b i s that the coefficient is minimized among | S | .
Equation (6) O ( | I | | W i | | S | | T i | ) | I | | W i | | S | | T i | subproblems determine decision variable f i j s .
Equation (7) O ( | S | | I | | W i | ) or O ( | S | | I | | T i | ) | S | subproblems with summation μ i j s 8 or μ i s 9 determine decision variable z s .
Dual problem (D) O ( N | I | | W i | | T i | | S | ) N times of the maximum complexity by the sub-problems, (4)–(6).
Drop-and-add algorithm O ( N | I | | W i | | T i | | S | ) The algorithm adjusts the values of decision variables y i , a i j s , b i s , f i j s , and z s based on the dual problems. The complexity is the worst case determined by the decision variables of the subproblems.
FF O ( | I | | W i | | T i | | S | ) The resource allocation.
NF O ( | I | | W i | | T i | | S | ) The resource allocation.
Brute force O ( 2 | I | 4 | W i | 2 | T i | 2 | S | 4 ) The total combination of decision variables.
Table 11. Execution time in the experiments.
Table 11. Execution time in the experiments.
Traffic load (number of servers: 80, dual-rate: 0.5)
Number of applications60120180
min (FF, NF)0.0418 s.0.0826 s.0.1067 s.
Proposed method373 s.744 s.1117 s.
Number of servers (number of users: 80, dual-rate: 0.5)
Number of servers4080100
min (FF, NF)0.213 s.0.159 s.0.269 s.
Proposed method236 s.478 s.604 s.
Effect of HA (number of applications: 80, number of servers: 80)
Dual-rate0.20.40.6
min (FF, NF)0.059 s.0.246 s.0.430 s.
Proposed method284 s.437 s.583 s.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hsiao, C.-H.; Lin, F.Y.-S.; Fang, E.S.-H.; Chen, Y.-F.; Wen, Y.-F.; Huang, Y.; Su, Y.-C.; Wu, Y.-S.; Kuo, H.-Y. Optimization-Based Resource Management Algorithms with Considerations of Client Satisfaction and High Availability in Elastic 5G Network Slices. Sensors 2021, 21, 1882. https://doi.org/10.3390/s21051882

AMA Style

Hsiao C-H, Lin FY-S, Fang ES-H, Chen Y-F, Wen Y-F, Huang Y, Su Y-C, Wu Y-S, Kuo H-Y. Optimization-Based Resource Management Algorithms with Considerations of Client Satisfaction and High Availability in Elastic 5G Network Slices. Sensors. 2021; 21(5):1882. https://doi.org/10.3390/s21051882

Chicago/Turabian Style

Hsiao, Chiu-Han, Frank Yeong-Sung Lin, Evana Szu-Han Fang, Yu-Fang Chen, Yean-Fu Wen, Yennun Huang, Yang-Che Su, Ya-Syuan Wu, and Hsin-Yi Kuo. 2021. "Optimization-Based Resource Management Algorithms with Considerations of Client Satisfaction and High Availability in Elastic 5G Network Slices" Sensors 21, no. 5: 1882. https://doi.org/10.3390/s21051882

APA Style

Hsiao, C. -H., Lin, F. Y. -S., Fang, E. S. -H., Chen, Y. -F., Wen, Y. -F., Huang, Y., Su, Y. -C., Wu, Y. -S., & Kuo, H. -Y. (2021). Optimization-Based Resource Management Algorithms with Considerations of Client Satisfaction and High Availability in Elastic 5G Network Slices. Sensors, 21(5), 1882. https://doi.org/10.3390/s21051882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop