DRL-Based Hybrid Task Offloading and Resource Allocation in Vehicular Networks
Abstract
:1. Introduction
- A multi-user vehicular network (VNET) model based on VEC has been developed for complex and changing VNET scenarios, which considers dynamic factors such as vehicle mobility, task size, channel state, and the distance between vehicles and servers to formulate a joint optimization problem that ensures the system efficiency to reduce the system cost by optimizing the offloading decision and bandwidth allocation;
- In this paper, we propose a novel Hybrid Task Computing Offloading (HTCO) algorithm, which innovatively combines a deep reinforcement learning (DRL) algorithm with a convex optimization algorithm and can perform computation offloading and bandwidth allocation in an arbitrary ratio. Specifically, this paper first transforms the dynamic task offloading problem into a Markov decision problem, uses the (TD3) algorithm to make offloading decisions, and brings the offloading decision variables into the convex optimization algorithm to obtain the bandwidth, which jointly optimizes the system’s latency and energy consumption, resulting in the lowest total cost of the system.
- The effectiveness of the HTCO algorithm is demonstrated by comparison experiments with the TD3 algorithm and other single-scene algorithms. The cost is reduced by 9%, 22%, and 54% compared to other algorithms.
2. Related Work
3. System Model
3.1. System Framework
3.2. Communication Model
3.3. Task Model
3.4. Problem Formulation
4. Hybrid Task Offloading Algorithm (HTCO) Design
4.1. Markov Decision Model for Offloading
- State: In DRL, the state space is used to reflect the environmental conditions of the vehicle. The state of the time slot is shown below:
- 2.
- Action: Based on the purpose of immediate reward maximization, the vehicle takes an action based on the current system status . The action is denoted as
- 3.
- Reward: When the action is performed, the system enters the next state and the vehicle receives a reward from the environment.
4.2. Offloading Decision Based on DRL Algorithm
- Tailoring the dual-Q network:
- 2.
- Delayed update of strategy:
- 3.
- Target strategy smoothing:
4.3. Bandwidth Allocation Based on Convex Optimization
Algorithm 1. HTCO Algorithm |
Input: vehicle edge computing environment parameter Output: action selected by HTCO algorithm and allocated bandwidth 1: Randomly initialize the network parameters and 2: Initialize the target network parameters and 3: Initialize the experience replay buffer 4: for each episode do 5: Random reset environmental parameters and receive the initial observation state s1 6: for each time slot do 7: Observe state s(t) and select action with exploration noise: 8: Use the Lagrange multiplier method to calculate the bandwidth by (27) 9: Execute action and obtain a reward and the next state 10: Store in the experience replay buffer 11: If is not full: store the data in the experience buffer 12: Else: randomly replace other experiences in the buffer 13: Select N samples from the experience replay buffer 14: Compute the value of the target network: 15: Backpropagation to update weights 16: Soft update of the target network 17: 18: 19: end for 20: end for |
5. Experimental Results
5.1. Simulation Environment and Parameter Settings
- All-Local: The task vehicle processes all tasks locally;
- All-MEC: The task vehicle processes all tasks on the MEC server;
- TD3: A deep reinforcement learning offload approach that disregards bandwidth distribution and determines the offload ratio of a task based on task size, channel conditions, distance, and other elements.
5.2. Analysis of Experimental Results
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Storck, C.R.; Duarte-Figueiredo, F. A Survey of 5G Technology Evolution, Standards, and Infrastructure Associated with Vehicle-to-Everything Communications by Internet of Vehicles. IEEE Access 2020, 8, 117593–117614. [Google Scholar] [CrossRef]
- Feng, C.; Han, P.; Zhang, X.; Yang, B.; Liu, Y.; Guo, L. Computation offloading in mobile edge computing networks: A survey. J. Netw. Comput. Appl. 2022, 202, 103366. [Google Scholar] [CrossRef]
- Yan, J.; Bi, S.; Zhang, Y.J.A. Offloading and resource allocation with general task graph in mobile edge computing: A deep reinforcement learning approach. IEEE Trans. Wirel. Commun. 2020, 19, 5404–5419. [Google Scholar] [CrossRef]
- Arena, F.; Pau, G.J. When Edge Computing Meets IoT Systems: Analysis of Case Studies. China Commun. 2020, 17, 14. [Google Scholar] [CrossRef]
- Li, Y.; Ma, H.; Wang, L.; Mao, S.; Wang, G. Optimized Content Caching and User Association for Edge Computing in Densely Deployed Heterogeneous Networks. IEEE Trans. Mob. Comput. 2020, 21, 2130–2142. [Google Scholar] [CrossRef]
- Zhang, J.; Ben Letaief, K. Mobile Edge Intelligence and Computing for the Internet of Vehicles. Proc. IEEE 2019, 108, 246–261. [Google Scholar] [CrossRef]
- Khan, W.U.; Nguyen, T.N.; Jameel, F.; Jamshed, M.A.; Pervaiz, H.; Javed, M.A.; Jäntti, R. Learning-based resource allocation for backscatter-aided vehicular networks. IEEE Trans. Intell. Transp. Syst. 2021, 23, 19676–19690. [Google Scholar] [CrossRef]
- Fatemidokht, H.; Rafsanjani, M.K.; Gupta, B.B.; Hsu, C.H. Efficient and Secure Routing Protocol Based on Artificial Intelligence Algorithms With UAV-Assisted for Vehicular Ad Hoc Networks in Intelligent Transportation Systems. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4757–4769. [Google Scholar] [CrossRef]
- Yue, X.; Liu, Y.; Wang, J.; Song, H.; Cao, H. Software Defined Radio and Wireless Acoustic Networking for Amateur Drone Surveillance. IEEE Commun. Mag. 2018, 56, 90–97. [Google Scholar] [CrossRef]
- Mach, P.; Becvar, Z. Mobile Edge Computing: A Survey on Architecture and Computation Offloading. IEEE Commun. Surv. Tutor. 2017, 19, 1628–1656. [Google Scholar] [CrossRef]
- Zhang, D.; Gong, C.; Zhang, T.; Zhang, J.; Piao, M. A new algorithm of clustering AODV based on edge computing strategy in IOV. Wirel. Netw. 2021, 27, 2891–2908. [Google Scholar] [CrossRef]
- Zhang, D.; Piao, M.; Zhang, T.; Chen, C.; Zhu, H. New algorithm of multi-strategy channel allocation for edge computing. Int. J. Electron. Commun. 2020, 126, 153372. [Google Scholar] [CrossRef]
- Wang, S.; Hu, L.; Cao, L.; Huang, X.; Lian, D.; Liu, W. Attention-based transactional context embedding for next-item recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
- Feng, W.; Zhang, N.; Li, S.; Lin, S.; Ning, R.; Yang, S.; Gao, Y. Latency Minimization of Reverse Offloading in Vehicular Edge Computing. IEEE Trans. Veh. Technol. 2022, 71, 5343–5357. [Google Scholar] [CrossRef]
- Zhang, J.; Guo, H.; Liu, J.; Zhang, Y. Task Offloading in Vehicular Edge Computing Networks: A Load-Balancing Solution. IEEE Trans. Veh. Technol. 2020, 69, 2092–2104. [Google Scholar] [CrossRef]
- Xu, J.; Li, X.; Ding, R.; Liu, X. Energy efficient multi-resource computation offloading strategy in mobile edge computing. Jisuanji Jicheng Zhizao Xitong/Comput. Integr. Manuf. Syst. 2019, 25, 954–961. [Google Scholar]
- Yang, L.; Zhang, H.; Li, M.; Guo, J.; Ji, H. Mobile edge computing empowered energy efficient task offloading in 5G. IEEE Trans. Veh. Technol. 2018, 67, 6398–6409. [Google Scholar] [CrossRef]
- Chen, H.; Todd, T.D.; Zhao, D.; Karakostas, G.J. Wireless and Service Allocation for Mobile Computation Offloading with Task Deadlines. arXiv 2023, arXiv:2301.12088. [Google Scholar] [CrossRef]
- Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. Deep Reinforcement Learning: A Brief Survey. IEEE Signal Process. Mag. 2017, 34, 26–38. [Google Scholar] [CrossRef]
- Liao, Y.; Qiao, X.; Yu, Q.; Liu, Q. Intelligent dynamic service pricing strategy for multi-user vehicle-aided MEC networks. Futur. Gener. Comput. Syst. 2021, 114, 15–22. [Google Scholar] [CrossRef]
- Wu, Y.C.; Dinh, T.Q.; Fu, Y.; Lin, C.; Quek, T.Q.S. A Hybrid DQN and Optimization Approach for Strategy and Resource Allocation in MEC Networks. IEEE Trans. Wirel. Commun. 2021, 20, 4282–4295. [Google Scholar] [CrossRef]
- Chen, C.; Li, H.; Li, H.; Fu, R.; Liu, Y.; Wan, S. Efficiency and Fairness Oriented Dynamic Task Offloading in Internet of Vehicles. IEEE Trans. Green Commun. Netw. 2022, 6, 1481–1493. [Google Scholar] [CrossRef]
- Zhang, L.; Zhou, W.; Xia, J.; Gao, C.; Zhu, F.; Fan, C.; Ou, J. DQN-based mobile edge computing for smart Internet of vehicle. EURASIP J. Adv. Signal Process. 2022, 2022, 45. [Google Scholar] [CrossRef]
- Elgendy, I.A.; Zhang, W.Z.; He, H.; Gupta, B.B.; Abd El-Latif, A.A. Joint computation offloading and task caching for multi-user and multi-task MEC systems: Reinforcement learning-based algorithms. Wirel. Netw. 2021, 27, 2023–2038. [Google Scholar] [CrossRef]
- Hazarika, B.; Singh, K.; Biswas, S.; Li, C.P. DRL-Based Resource Allocation for Computation Offloading in IoV Networks. IEEE Trans. Ind. Inform. 2022, 18, 8027–8038. [Google Scholar] [CrossRef]
- Wu, C.; Huang, Z.; Zou, Y. Delay Constrained Hybrid Task Offloading of Internet of Vehicle: A Deep Reinforcement Learning Method. IEEE Access 2022, 10, 102778–102788. [Google Scholar] [CrossRef]
- Xue, Z.; Liu, C.; Liao, C.; Han, G.; Sheng, Z. Joint Service Caching and Computation Offloading Scheme Based on Deep Reinforcement Learning in Vehicular Edge Computing Systems. IEEE Trans. Veh. Technol. 2023, 72, 6709–6722. [Google Scholar] [CrossRef]
- Hazarika, B.; Singh, K.; Biswas, S.; Mumtaz, S.; Li, C.P. Multi-Agent DRL-Based Task Offloading in Multiple RIS-Aided IoV Networks. IEEE Trans. Veh. Technol. 2023, 1–15. [Google Scholar] [CrossRef]
- Chen, Z.; Wang, X. Decentralized computation offloading for multi-user mobile edge computing: A deep reinforcement learning approach. J. Wireless Commn. Netw. 2020, 2020, 188. [Google Scholar] [CrossRef]
- Van Hasselt, H.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
- Fujimoto, S.; Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 1587–1596. [Google Scholar]
- Hu, X.; Huang, Y. Deep reinforcement learning based offloading decision algorithm for vehicular edge computing. PeerJ Comput. Sci. 2022, 8, e1126. [Google Scholar] [CrossRef]
Parameter | Quantity | Explanation |
---|---|---|
batch size | 64 | Sample size in reply buffer |
10 MHz | Total system bandwidth | |
5 GHz | MEC computing power | |
1.2 GHz | Vehicle computing power | |
10 m/s | Vehicle speed | |
5 | Number of vehicles | |
100 cycles/bit | Number of CPU cycles | |
20–30 dBm | Transmission power | |
50–70 M | Task size | |
0.0001 | Learning rate | |
0.5 | Delay/energy weighting |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, Z.; Jia, Z.; Pang, X. DRL-Based Hybrid Task Offloading and Resource Allocation in Vehicular Networks. Electronics 2023, 12, 4392. https://doi.org/10.3390/electronics12214392
Liu Z, Jia Z, Pang X. DRL-Based Hybrid Task Offloading and Resource Allocation in Vehicular Networks. Electronics. 2023; 12(21):4392. https://doi.org/10.3390/electronics12214392
Chicago/Turabian StyleLiu, Ziang, Zongpu Jia, and Xiaoyan Pang. 2023. "DRL-Based Hybrid Task Offloading and Resource Allocation in Vehicular Networks" Electronics 12, no. 21: 4392. https://doi.org/10.3390/electronics12214392
APA StyleLiu, Z., Jia, Z., & Pang, X. (2023). DRL-Based Hybrid Task Offloading and Resource Allocation in Vehicular Networks. Electronics, 12(21), 4392. https://doi.org/10.3390/electronics12214392