Reinforcement-Learning-Assisted Service Function Chain Embedding Algorithm in Edge Computing Networks
Abstract
:1. Introduction
- Distributed-based reinforcement learning algorithm: We designed a RL algorithm that can make dynamic resource allocation decisions based on the real-time state of the system. Through a distributed architecture, the algorithm can handle large-scale smart city networks and has good scalability.
- Embedded algorithm for service function chain: For the embedding problem of the service function chain, we propose an optimized algorithm. This algorithm can make intelligent mapping decisions based on the requirements of service functions and the characteristics of physical resources, thereby improving the overall performance and stability of the system.
2. Related Work
2.1. Research Status of Edge Computing Networks
2.2. Research Status of Network Resource Allocation Algorithms
2.3. Summary of Related Work
3. Network Modeling
3.1. Modeling Edge Computing Networks
3.2. Resource Properties and Constraint Conditions
- Resource Properties
- Constraint Conditions
3.3. Algorithmic Evaluation Metrics
- The long-term acceptance ratio
- The long-term average revenue
- The long-term average revenue–cost ratio ()
3.4. System Models
- State Space: The system state represents the real-time status of the physical network, including real-time computing resources, bandwidth, and other resource information, as well as real-time node latency, network topology, and other network information. Network state can be defined as , where represents the real-time available computing resources, denotes the network bandwidth, denotes the node latency, and describes the network topology.
- Action Space: The action of the agent is the mapping decision taken at a certain moment t, including the embedding scheme and the embedding scheme, defined as , where is a sequence of embedding scheme for , and is for . The embedding scheme is expressed using the following formula:
- Reward: The reward of an intelligent agent is the feedback signal value obtained from the action taken, which is used to guide the optimization direction of the agent. In this work, we consider three factors, including long-term acceptance ratio (), long-term average revenue (R), and long-term average revenue–cost ratio (). The reward is computed using the following formula:
4. Algorithm
4.1. Parameter Settings
- : The computing resources of the physical node determine its carrying capacity, and the node with higher computing resources is better capable of meeting user requirements.
- : Each physical node has at least one link connected to it. denotes the sum of the bandwidth of all the links connected to node . When the of a node is higher, the virtual nodes it hosts can have better link embedding options.
- : The delay performance of a physical node reflects how fast it can process Virtual Network Functions (VNFs). Nodes with lower delay can handle VNFs with stringent delay requirements.
- : Degree denotes the total number of links connected to node . Physical nodes with more adjacent links can have a higher possibility of successful link embedding.
4.2. Environment Perception
4.3. Local Training
Algorithm 1 Local Training |
Input: Output: Probability of SFC being embedded;
|
5. Experimental Analysis
5.1. Experimental Environment and Parameter Settings
5.2. Evaluation Results
5.2.1. Experiment 1: The Long-Term Average Revenue
5.2.2. Experiment 2: The Long-Term SFCR Acceptance Rate
5.2.3. Experiment 3: The Long-Term R/C Ratio
5.2.4. Summary of Evaluation
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Mahmood, O.A.; Abdellah, A.R.; Muthanna, A.; Koucheryavy, A. Distributed edge computing for resource allocation in smart cities based on the IoT. Information 2022, 13, 328. [Google Scholar] [CrossRef]
- Cai, Q.; Zhou, Y.; Liu, L.; Qi, Y.; Pan, Z.; Zhang, H. Collaboration of heterogeneous edge computing paradigms: How to fill the gap between theory and practice. IEEE Wirel. Commun. 2023, 31, 110–117. [Google Scholar] [CrossRef]
- Cao, H.; Wu, S.; Aujla, G.S.; Wang, Q.; Yang, L.; Zhu, H. Dynamic embedding and quality of service-driven adjustment for cloud networks. IEEE Trans. Ind. Inform. 2019, 16, 1406–1416. [Google Scholar] [CrossRef]
- Bechtold, S.; Perrig, A. Accountability in future internet architectures. Commun. ACM 2014, 57, 21–23. [Google Scholar] [CrossRef]
- Fisher, D. A look behind the future internet architectures efforts. ACM SIGCOMM Comput. Commun. Rev. 2014, 44, 45–49. [Google Scholar] [CrossRef]
- Anderson, T.; Peterson, L.; Shenker, S.; Turner, J. Overcoming the Internet impasse through virtualization. Computer 2005, 38, 34–41. [Google Scholar] [CrossRef]
- Houidi, I.; Louati, W.; Ameur, W.B.; Zeghlache, D. Virtual network provisioning across multiple substrate networks. Comput. Netw. 2011, 55, 1011–1023. [Google Scholar] [CrossRef]
- Zhang, P.; Wang, C.; Qin, Z.; Cao, H. A multidomain virtual network embedding algorithm based on multiobjective optimization for Internet of Drones architecture in Industry 4.0. Softw. Pract. Exp. 2022, 52, 710–728. [Google Scholar] [CrossRef]
- Cao, H.; Yang, L.; Zhu, H. Novel node-ranking approach and multiple topology attributes-based embedding algorithm for single-domain virtual network embedding. IEEE Internet Things J. 2017, 5, 108–120. [Google Scholar] [CrossRef]
- Spano, S.; Cardarilli, G.C.; Di Nunzio, L.; Fazzolari, R.; Giardino, D.; Matta, M.; Nannarelli, A.; Re, M. An efficient hardware implementation of reinforcement learning: The q-learning algorithm. IEEE Access 2019, 7, 186340–186351. [Google Scholar] [CrossRef]
- Wang, C.; Liu, L.; Jiang, C.; Wang, S.; Zhang, P.; Shen, S. Incorporating Distributed DRL Into Storage Resource Optimization of Space-Air-Ground Integrated Wireless Communication Network. IEEE J. Sel. Top. Signal Process. 2022, 16, 434–446. [Google Scholar] [CrossRef]
- Abdullaev, I.; Prodanova, N.; Bhaskar, K.A.; Lydia, E.L.; Kadry, S.; Kim, J. Task offloading and resource allocation in iot based mobile edge computing using deep learning. Comput. Mater. Contin. 2023, 76, 1463–1477. [Google Scholar] [CrossRef]
- Chen, Z.; Chen, Z.; Ren, Z.; Liang, L.; Wen, W.; Jia, Y. Joint optimization of task caching, computation offloading and resource allocation for mobile edge computing. China Commun. 2022, 19, 142–159. [Google Scholar] [CrossRef]
- Liu, Q.; Li, J.; Wei, J.; Zhou, R.; Chai, Z.; Liu, S. Efficient multi-user for task offloading and server allocation in mobile edge computing systems. China Commun. 2022, 19, 226–238. [Google Scholar] [CrossRef]
- Gao, J.X.; Hu, B.Y.; Liu, J.L.; Wang, H.C.; Huang, Q.Z.; Zhao, Y. Overbooking-Enabled Task Scheduling and Resource Allocation in Mobile Edge Computing Environments. Intell. Autom. Soft Comput. 2023, 37, 1–16. [Google Scholar] [CrossRef]
- Yang, Y.; Gong, Y.; Wu, Y.C. Energy Sensitive Binary Offloading for reconfigurable-intelligent-surface-assisted wireless powered mobile edge computing. IEEE Internet Things J. 2023, 11, 11593–11605. [Google Scholar] [CrossRef]
- Liu, X.; Yu, J.; Feng, Z.; Gao, Y. Multi-agent reinforcement learning for resource allocation in IoT networks with edge computing. China Commun. 2020, 17, 220–236. [Google Scholar] [CrossRef]
- Wen, J.; Yang, J.; Wang, T.; Li, Y.; Lv, Z. Energy-efficient task allocation for reliable parallel computation of cluster-based wireless sensor network in edge computing. Digit. Commun. Netw. 2023, 9, 473–482. [Google Scholar] [CrossRef]
- Li, Z.; Hu, C.; Wang, W.; Li, Y.; Wei, G. Joint access point selection and resource allocation in MEC-assisted network: A reinforcement learning based approach. China Commun. 2022, 19, 205–218. [Google Scholar] [CrossRef]
- Zhi, Y.; Tian, J.; Deng, X.; Qiao, J.; Lu, D. Deep reinforcement learning-based resource allocation for D2D communications in heterogeneous cellular networks. Digit. Commun. Netw. 2022, 8, 834–842. [Google Scholar] [CrossRef]
- Lin, Q. Dynamic resource allocation strategy in mobile edge cloud computing environment. Mob. Inf. Syst. 2021, 2021, 8381998. [Google Scholar] [CrossRef]
- Li, F.; Fang, C.; Liu, M.; Li, N.; Sun, T. Intelligent Computation Offloading Mechanism with Content Cache in Mobile Edge Computing. Electronics 2023, 12, 1254. [Google Scholar] [CrossRef]
- Xu, J.; Xu, Z.; Shi, B. Deep Reinforcement Learning Based Resource Allocation Strategy in Cloud-Edge Computing System. Front. Bioeng. Biotechnol. 2022, 10, 908056. [Google Scholar]
- Sun, W.; Liu, J.; Yue, Y.; Wang, P. Joint resource allocation and incentive design for blockchain-based mobile edge computing. IEEE Trans. Wirel. Commun. 2020, 19, 6050–6064. [Google Scholar] [CrossRef]
- Wang, C.; Liang, C.; Yu, F.R.; Chen, Q.; Tang, L. Computation offloading and resource allocation in wireless cellular networks with mobile edge computing. IEEE Trans. Wirel. Commun. 2017, 16, 4924–4938. [Google Scholar] [CrossRef]
- Zheng, J.; Pan, Y.; Jiang, S.; Chen, Z.; Yan, F. A Federated Learning and Deep Q-Network based Cooperative Resource Allocation Algorithm for Multi-Level Services in Mobile Edge Computing Networks. IEEE Trans. Cogn. Commun. Netw. 2023, 9, 1734–1745. [Google Scholar] [CrossRef]
- He, Y.; Wang, Y.; Lin, Q.; Li, J. Meta-hierarchical reinforcement learning (MHRL)-based dynamic resource allocation for dynamic vehicular networks. IEEE Trans. Veh. Technol. 2022, 71, 3495–3506. [Google Scholar] [CrossRef]
- Zhang, P.; Li, Y.; Kumar, N.; Chen, N.; Hsu, C.-H.; Barnawi, A. Distributed Deep Reinforcement Learning Assisted Resource Allocation Algorithm for Space-Air-Ground Integrated Networks. IEEE Trans. Netw. Serv. Manag. 2023, 20, 3348–3358. [Google Scholar] [CrossRef]
- Yu, M.; Yi, Y.; Rexford, J.; Chiang, M. Rethinking virtual network embedding: Substrate support for path splitting and migration. ACM SIGCOMM Comput. Commun. Rev. 2008, 38, 17–29. [Google Scholar] [CrossRef]
- Cai, Y.; Wang, Y.; Zhong, X.; Li, W.; Qiu, X.; Guo, S. An approach to deploy service function chains in satellite networks. In Proceedings of the NOMS 2018-2018 IEEE/IFIP Network Operations and Management Symposium, Taipei, Taiwan, 23–27 April 2018; pp. 1–7. [Google Scholar]
Parameter | Values |
---|---|
Physical nodes | 100 |
Physical links | 600 |
CPU capacity | U[50,100] |
Bandwidth capacity | U[50,100] |
Link bandwidth resource | U[1,50] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, W.; Chen, S.; Zhang, P.; Liu, K. Reinforcement-Learning-Assisted Service Function Chain Embedding Algorithm in Edge Computing Networks. Electronics 2024, 13, 3007. https://doi.org/10.3390/electronics13153007
Wang W, Chen S, Zhang P, Liu K. Reinforcement-Learning-Assisted Service Function Chain Embedding Algorithm in Edge Computing Networks. Electronics. 2024; 13(15):3007. https://doi.org/10.3390/electronics13153007
Chicago/Turabian StyleWang, Wei, Shengpeng Chen, Peiying Zhang, and Kai Liu. 2024. "Reinforcement-Learning-Assisted Service Function Chain Embedding Algorithm in Edge Computing Networks" Electronics 13, no. 15: 3007. https://doi.org/10.3390/electronics13153007
APA StyleWang, W., Chen, S., Zhang, P., & Liu, K. (2024). Reinforcement-Learning-Assisted Service Function Chain Embedding Algorithm in Edge Computing Networks. Electronics, 13(15), 3007. https://doi.org/10.3390/electronics13153007