Path Tracking Control for Four-Wheel Independent Steering and Driving Vehicles Based on Improved Deep Reinforcement Learning
Abstract
:1. Introduction
2. Background and Related Work
2.1. DRL-Based Vehicle Control
2.2. Experience Replay Mechanism in DRL
2.3. Representation Learning in DRL
3. Problem Formulation and Analysis
3.1. 4WISD Vehicle Dynamics Model
3.2. Transition Model for DRL
4. Compound Control Framework Based on Improved DRL
4.1. Compound Control Framework
4.2. Group Intelligent Experience Replay
- (1)
- Group : Samples with priorities in the top percentile.
- (2)
- Group : Samples with priorities in the top percentile but not in the top percentile.
- (3)
- Group : Samples with priorities in the bottom percentile.
4.3. Actor-Critic Architecture Based on TIB
Algorithm 1 Proposed GT-TD3 |
|
5. Numerical Simulations
5.1. Convergence and Generalization Analyses
5.2. Performance Analysis of an Improved DRL-Based Path Tracking Controller for 4WISD Vehicles
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A. Stability Analysis of the Compound Control Framework
References
- Zhao, M.; Wang, L.; Ma, L.; Liu, C.; An, N. Development of a four wheel independent-driving and four wheel steering electric testing car. China Mech. Eng. 2009, 20, 319–322. [Google Scholar]
- Dong, Z.R.; Ren, S.; He, P. Kinematics modeling and inverse kinematics simulation of a 4WID/4WIS electric vehicle based on multi-body dynamics. Automot. Eng. 2015, 3, 253–259. [Google Scholar]
- Kumar, P.; Sandhan, T. Path-tracking control of the 4WIS4WID electric vehicle by direct inverse control using artificial neural network. In Proceedings of the 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), Delhi, India, 6–8 July 2023; pp. 1–8. [Google Scholar]
- Zhang, Z.; Zhang, X.; Pan, H.; Salman, W.; Rasim, Y.; Liu, X.; Wang, C.; Yang, Y.; Li, X. A novel steering system for a space-saving 4WS4WD electric vehicle: Design, modeling, and road tests. IEEE Trans. Intell. Transp. Syst. 2016, 18, 114–127. [Google Scholar] [CrossRef]
- Maoqi, L.; Ishak, M.I.; Heerwan, P.M. The effect of parallel steering of a four-wheel drive and four-wheel steer electric vehicle during spinning condition: A numerical simulation. IOP Mater. Sci. Eng. 2019, 469, 012084. [Google Scholar] [CrossRef]
- Li, Y.; Cai, Y.; Sun, X.; Wang, H.; Jia, Y.; He, Y.; Chen, L.; Chao, Y. Trajectory tracking of four-wheel driving and steering autonomous vehicle under extreme obstacle avoidance condition. Veh. Syst. Dyn. 2024, 62, 601–622. [Google Scholar] [CrossRef]
- Hang, P.; Han, Y.; Chen, X.; Zhang, B. Design of an active collision avoidance system for a 4WIS-4WID electric vehicle. IFAC-PapersOnLine 2018, 51, 771–777. [Google Scholar] [CrossRef]
- Hang, P.; Chen, X. Towards autonomous driving: Review and perspectives on configuration and control of four-wheel independent drive/steering electric vehicles. Actuators 2021, 10, 184. [Google Scholar] [CrossRef]
- Wang, Z.; Ding, X.; Zhang, L. Chassis coordinated control for full x-by-wire four-wheel-independent-drive electric vehicles. IEEE Trans. Veh. Technol. 2022, 72, 4394–4410. [Google Scholar] [CrossRef]
- Long, W.; Zhang, Z. Research and design of steering control for four wheel driving mobile robots. Control Eng. China 2017, 24, 2387–2393. [Google Scholar]
- Jin, L.; Gao, L.; Jiang, Y.; Chen, M.; Zheng, Y.; Li, K. Research on the control and coordination of four-wheel independent driving/steering electric vehicle. Adv. Mech. Eng. 2017, 9, 1687814017698877. [Google Scholar] [CrossRef]
- Wang, C.; Heng, B.; Zhao, W. Yaw and lateral stability control for four-wheel-independent steering and four-wheel-independent driving electric vehicle. Proc. Inst. Mech. Eng. Part J. Automob. Eng. 2020, 234, 409–422. [Google Scholar] [CrossRef]
- Yutao, L.; Tianyang, Z.; Xiaotong, X. Time-varying LQR control of four-wheel steer/drive vehicle based on genetic algorithm. J. South China Univ. Technol. (Natural Sci. Ed.) 2021, 49, 9. [Google Scholar]
- Potluri, R.; Singh, A.K. Path-tracking control of an autonomous 4WS4WD electric vehicle using its natural feedback loops. IEEE Trans. Control. Syst. Technol. 2015, 23, 2053–2062. [Google Scholar] [CrossRef]
- Lai, X.; Chen, X.B.; Wu, X.J.; Liang, D. A study on control system for four-wheels independent driving and steering electric vehicle. Appl. Mech. Mater. 2015, 701, 807–811. [Google Scholar] [CrossRef]
- Tan, Q.; Dai, P.; Zhang, Z.; Katupitiya, J. MPC and PSO based control methodology for path tracking of 4WS4WD vehicles. Appl. Sci. 2018, 8, 1000. [Google Scholar] [CrossRef]
- Zhang, X.; Zhu, X. Autonomous path tracking control of intelligent electric vehicles based on lane detection and optimal preview method. Expert Syst. Appl. 2019, 121, 38–48. [Google Scholar] [CrossRef]
- Akermi, K.; Chouraqui, S.; Boudaa, B. Novel SMC control design for path following of autonomous vehicles with uncertainties and mismatched disturbances. Int. J. Dyn. Control. 2020, 8, 254–268. [Google Scholar] [CrossRef]
- Jeong, Y.; Yim, S. Model predictive control-based integrated path tracking and velocity control for autonomous vehicle with four-wheel independent steering and driving. Electronics 2021, 10, 2812. [Google Scholar] [CrossRef]
- Barari, A.; Afshari, S.S.; Liang, X. Coordinated control for path-following of an autonomous four in-wheel motor drive electric vehicle. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2022, 236, 6335–6346. [Google Scholar] [CrossRef]
- Rui, L.; Duan, J. A path tracking algorithm of intelligent vehicle by preview strategy. In Proceedings of the 32nd Chinese Control Conference, Xi’an, China, 26–28 July 2022; pp. 26–28. [Google Scholar]
- Li, Y.; Jiang, Y.; Wang, L.; Cao, J.; Zhang, G. Intelligent PID guidance control for AUV path tracking. J. Cent. South Univ. 2015, 22, 3440–3449. [Google Scholar] [CrossRef]
- Zhang, P.; Zhang, J.; Kan, J. A research on manipulator-path tracking based ondeep reinforcement learning. Appl. Sci. 2023, 13, 7867. [Google Scholar] [CrossRef]
- Li, Z.; Yuan, S.; Yin, X.; Li, X.; Tang, S. Research into autonomous vehicles following and obstacle avoidance based on deep reinforcement learning method under map constraints. Sensors 2023, 23, 844. [Google Scholar] [CrossRef] [PubMed]
- Lu, Y.; Wu, C.; Yao, W.; Sun, G.; Liu, J.; Wu, L. Deep reinforcement learning control of fully-constrained cable-driven parallel robots. IEEE Trans. Ind. Electron. 2022, 70, 7194–7204. [Google Scholar] [CrossRef]
- Chen, H.; Zhang, Y.; Bhatti, U.A.; Huang, M. Safe decision controller for autonomous driving based on deep reinforcement learning in nondeterministic environment. Sensors 2023, 23, 1198. [Google Scholar] [CrossRef]
- Mirmozaffari, M.; Yazdani, M.; Boskabadi, A.; Ahady Dolatsara, H.; Kabirifar, K.; Amiri Golilarz, N. A novel machine learning approach combined with optimization models for eco-efficiency evaluation. Appl. Sci. 2020, 10, 5210. [Google Scholar] [CrossRef]
- Osedo, A.; Wada, D.; Hisada, S. Uniaxial attitude control of uncrewed aerial vehicle with thrust vectoring under model variations by deep reinforcement learning and domain randomization. ROBOMECH J. 2023, 10, 20. [Google Scholar] [CrossRef]
- Huang, S.; Wang, T.; Tang, Y.; Hu, Y.; Xin, G.; Zhou, D. Distributed and scalable cooperative formation of unmanned ground vehicles using deep reinforcement learning. Aerospace 2023, 10, 96. [Google Scholar] [CrossRef]
- Abbas, A.N.; Chasparis, G.C.; Kelleher, J.D. Specialized deep residual policy safe reinforcement learning-based controller for complex and continuous state-action spaces. arXiv 2023, arXiv:2310.14788. [Google Scholar]
- Liu, T.; Yang, Y.; Xiao, W.; Tang, X.; Yin, M. A Comparative Analysis of Deep Reinforcement Learning-Enabled Freeway Decision-Making for Automated Vehicles. IEEE Access 2024, 12, 24090–24103. [Google Scholar] [CrossRef]
- Lin, Y.; McPhee, J.; Azad, N.L. Comparison of deep reinforcement learning and model predictive control for adaptive cruise control. IEEE Trans. Intell. Veh. 2020, 6, 221–231. [Google Scholar] [CrossRef]
- Chen, T.C.; Sung, Y.C.; Hsu, C.W.; Liu, D.R.; Chen, S.J. Path following and obstacle avoidance of tracked vehicle via deep reinforcement learning with model predictive control as reference. In Proceedings of the Multimodal Sensing and Artificial Intelligence: Technologies and Applications III, Munich, Germany, 26–30 June 2023; pp. 91–96. [Google Scholar]
- Selvaraj, D.C.; Hegde, S.; Amati, N.; Deflorio, F.; Chiasserini, C.F. An ML-aided reinforcement learning approach for challenging vehicle maneuvers. IEEE Trans. Intell. Veh. 2022, 8, 1686–1698. [Google Scholar] [CrossRef]
- Li, D.; Zhao, D.; Zhang, Q.; Chen, Y. Reinforcement learning and deep learning based lateral control for autonomous driving [application notes]. IEEE Comput. Intell. Mag. 2019, 14, 83–98. [Google Scholar] [CrossRef]
- Peng, Y.; Tan, G.; Si, H.; Li, J. DRL-GAT-SA: Deep reinforcement learning for autonomous driving planning based on graph attention networks and simplex architecture. J. Syst. Archit. 2022, 126, 102505. [Google Scholar] [CrossRef]
- Li, J.; Wu, X.; Fan, J.; Liu, Y.; Xu, M. Overcoming driving challenges in complex urban traffic: A multi-objective eco-driving strategy via safety model based reinforcement learning. Energy 2023, 284, 128517. [Google Scholar] [CrossRef]
- EL Sallab, A.; Abdou, M.; Perot, E.; Yogamani, S. Deep reinforcement learning framework for autonomous driving. arXiv 2017, arXiv:1704.02532. [Google Scholar] [CrossRef]
- Wei, Q.; Ma, H.; Chen, C.; Dong, D. Deep reinforcement learning with quantum-inspired experience replay. IEEE Trans. Cybern. 2021, 52, 9326–9338. [Google Scholar] [CrossRef]
- Li, Y.; Aghvami, A.H.; Dong, D. Path planning for cellular-connected UAV: A DRL solution with quantum-inspired experience replay. IEEE Trans. Wirel. Commun. 2022, 21, 7897–7912. [Google Scholar] [CrossRef]
- Zhu, P.; Dai, W.; Yao, W.; Ma, J.; Zeng, Z.; Lu, H. Multi-robot flocking control based on deep reinforcement learning. IEEE Access 2020, 8, 150397–150406. [Google Scholar] [CrossRef]
- Na, S.; Niu, H.; Lennox, B.; Arvin, F. Bio-inspired collision avoidance in swarm systems via deep reinforcement learning. IEEE Trans. Veh. Technol. 2022, 71, 2511–2526. [Google Scholar] [CrossRef]
- Ye, X.; Yu, Y.; Fu, L. Deep reinforcement learning based link adaptation technique for LTE/NR systems. IEEE Trans. Veh. Technol. 2023, 72, 7364–7379. [Google Scholar] [CrossRef]
- Ma, J.; Ning, D.; Zhang, C.; Liu, S. Fresher experience plays a more important role in prioritized experience replay. Appl. Sci. 2022, 12, 12489. [Google Scholar] [CrossRef]
- Wang, X.; Luo, Y.; Qin, B.; Guo, L. Power allocation strategy for urban rail HESS based on deep reinforcement learning sequential decision optimization. IEEE Trans. Transp. Electrif. 2022, 9, 2693–2710. [Google Scholar] [CrossRef]
- Osei, R.S.; Lopez, D. Experience replay optimization via ESMM for stable deep reinforcement learning. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 1. [Google Scholar] [CrossRef]
- Liu, Y.; Wang, H.; Wu, T.; Lun, Y.; Fan, J.; Wu, J. Attitude control for hypersonic reentry vehicles: An efficient deep reinforcement learning method. Appl. Soft Comput. 2022, 123, 108865. [Google Scholar] [CrossRef]
- Xiang, G.; Dian, S.; Du, S.; Lv, Z. Variational information bottleneck regularized deep reinforcement learning for efficient robotic skill adaptation. Sensors 2023, 23, 762. [Google Scholar] [CrossRef]
- Zou, Q.; Suzuki, E. Compact goal representation learning via information bottleneck in goal-conditioned reinforcement learning. IEEE Trans. Neural Netw. Learn. Syst. 2024, 1–14. [Google Scholar] [CrossRef]
- Schwarzer, M.; Anand, A.; Goel, R.; Hjelm, R.D.; Courville, A.; Bachman, P. Data-efficient reinforcement learning with self-predictive representations. arXiv 2020, arXiv:2007.05929. [Google Scholar]
- Zhang, A.; McAllister, R.; Calandra, R.; Gal, Y.; Levine, S. Learning invariant representations for reinforcement learning without reconstruction. arXiv 2020, arXiv:2006.10742. [Google Scholar]
- Laskin, M.; Srinivas, A.; Abbeel, P. Curl: Contrastive unsupervised representations for reinforcement learning. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 5639–5650. [Google Scholar]
- Stooke, A.; Lee, K.; Abbeel, P.; Laskin, M. Decoupling representation learning from reinforcement learning. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 9870–9879. [Google Scholar]
- Wei, W.; Fu, L.; Gu, H.; Zhang, Y.; Zou, T.; Wang, C.; Wang, N. GRL-PS: Graph embedding-based DRL approach for adaptive path selection. IEEE Trans. Netw. Serv. Manag. 2023, 20, 2639–2651. [Google Scholar] [CrossRef]
- Qian, Z.; You, M.; Zhou, H.; He, B. Weakly supervised disentangled representation for goal-conditioned reinforcement learning. IEEE Robot. Autom. Lett. 2022, 7, 2202–2209. [Google Scholar] [CrossRef]
- Yarats, D.; Zhang, A.; Kostrikov, I.; Amos, B.; Pineau, J.; Fergus, R. Improving sample efficiency in model-free reinforcement learning from images. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 10674–10681. [Google Scholar]
- Dai, P.; Katupitiya, J. Force control for path following of a 4WS4WD vehicle by the integration of PSO and SMC. Veh. Syst. Dyn. 2018, 56, 1682–1716. [Google Scholar] [CrossRef]
- Besselink, I.J.M.; Schmeitz, A.J.C.; Pacejka, H.B. An improved Magic Formula/Swift tyre model that can handle inflation pressure changes. Veh. Syst. Dyn. 2010, 48, 337–352. [Google Scholar] [CrossRef]
- Wang, X.; Zhang, J.; Hou, D.; Cheng, Y. Autonomous driving based on approximate safe action. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14320–14328. [Google Scholar] [CrossRef]
- Yan, S.; Liu, W.; Li, X.; Yang, P.; Wu, F.; Yan, Z. Comparative study and improvement analysis of sparrow search algorithm. Wirel. Commun. Mob. Comput. 2022, 1, 4882521. [Google Scholar] [CrossRef]
- Tian, Y.; Wang, H.; Zhang, X.; Jin, Y. Effectiveness and efficiency of non-dominated sorting for evolutionary multi-and many-objective optimization. Complex Intell. Syst. 2017, 3, 247–263. [Google Scholar] [CrossRef]
- Wu, A.; Deng, C. TIB: Detecting unknown objects via two-stream information bottleneck. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 611–625. [Google Scholar] [CrossRef]
- Feng, X. Consistent experience replay in high-dimensional continuous control with decayed hindsights. Machines 2022, 10, 856. [Google Scholar] [CrossRef]
- Dankwa, S.; Zheng, W. Twin-delayed ddpg: A deep reinforcement learning technique to model a continuous movement of an intelligent robot agent. In Proceedings of the 3rd International Conference on Vision, Image and Signal Processing, Vancouver, BC, Canada, 26–28 August 2019; pp. 1–5. [Google Scholar]
DRL Algorithm | AER-TD3 | PER-TD3 | GER-TD3 |
---|---|---|---|
Hidden layer dimension | 256 | 256 | 256 |
Batch size | 256 | 256 | 256 |
Discount factor | 0.99 | 0.99 | 0.99 |
Soft update coefficient | 0.05 | 0.05 | 0.05 |
Policy noise | 0.2 | 0.2 | 0.2 |
Noise clipping range | 256 | 256 | 256 |
Policy update frequency | 2 | 2 | 2 |
Priority exponent | × | 0.6 | 0.6 |
Group proportion coefficients | × | × | 0.2, 0.7, 0.1 |
Learning rate | 1 × 10−4 | 1 × 10−4 | 1 × 10−4 |
DRL Training Environment | Parameters | Unit |
---|---|---|
Shift line longitudinal position | [40,180] | m |
Shift line transition length | [25,75] | m |
Longitudinal velocity | [15,20] | m/s |
Longitudinal velocity variation range | [0,20] | m/s |
Smooth disturbance amplitude | [−100,100] | N |
Sudden disturbance amplitude | [−100,100] | N |
Smooth disturbance duration | [5,10] | s |
Vehicle Parameter | Parameters | Unit |
---|---|---|
Vehicle mass | 1477 | kg |
Vehicle yaw inertia | 1536.7 | |
Track width | 1.675 | m |
Distance from CG to front axle | 1.015 | m |
Distance from CG to rear axle | 1.895 | m |
Wheel radius | 0.325 | m |
Wheel mass | 22 | kg |
Wheel moment of inertia | 0.8 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hua, X.; Zhang, T.; Cheng, X.; Ning, X. Path Tracking Control for Four-Wheel Independent Steering and Driving Vehicles Based on Improved Deep Reinforcement Learning. Technologies 2024, 12, 218. https://doi.org/10.3390/technologies12110218
Hua X, Zhang T, Cheng X, Ning X. Path Tracking Control for Four-Wheel Independent Steering and Driving Vehicles Based on Improved Deep Reinforcement Learning. Technologies. 2024; 12(11):218. https://doi.org/10.3390/technologies12110218
Chicago/Turabian StyleHua, Xia, Tengteng Zhang, Xiangle Cheng, and Xiaobin Ning. 2024. "Path Tracking Control for Four-Wheel Independent Steering and Driving Vehicles Based on Improved Deep Reinforcement Learning" Technologies 12, no. 11: 218. https://doi.org/10.3390/technologies12110218
APA StyleHua, X., Zhang, T., Cheng, X., & Ning, X. (2024). Path Tracking Control for Four-Wheel Independent Steering and Driving Vehicles Based on Improved Deep Reinforcement Learning. Technologies, 12(11), 218. https://doi.org/10.3390/technologies12110218