The traditional D*Lite algorithm, the improved D*Lite algorithm, the traditional DWA, the A* and DWA fusion method (A_DWA, which can handle environmental risks and has good dynamic performance, as described in [
21]), and the fusion algorithm proposed in this paper are employed for path planning experiments. Both the traditional D*Lite algorithm and the improved D*Lite algorithm execute path planning on the coarse map, aiming to showcase the improved D*Lite algorithm’s ability to jointly consider risk and distance. The comparison among the traditional DWA, the improved DWA, the A* and DWA fusion algorithm, and the proposed algorithm reflects not only the global optimal ability and dynamic planning ability of the algorithm designed in this paper but also the enhanced algorithm efficiency.
5.1. Static Environment
In order to fully test the performance of the algorithm, we placed a number of obstacles that appeared suddenly on the map during the test; these obstacles were not surrounded by other risk level grids. There are two types of static simulation environments, which are as follows:
(A) The environment with only regional risk and no additional obstacles.
(B) An environment where obstacles are added to the planned path in environment (A), allowing the algorithm to replan the path and observe the performance.
The path planning results of the traditional D*Lite algorithm (
Figure 5) and the improved D*Lite algorithm (
Figure 6) on the coarse map are shown. Additionally, the results for the traditional DWA (
Figure 7), the improved DWA (
Figure 8), the fused algorithm of A* and DWA (
Figure 9), and the proposed fusion algorithm of D*Lite algorithm and DWA (
Figure 10) on the bi-layer map are presented.
In
Figure 5,
Figure 6,
Figure 7,
Figure 8,
Figure 9 and
Figure 10, the start and the goal locations for all cases are set as
sstart = (
xstart,
ystart) = (0, 0), and
sgoal = (
xgoal,
ygoal) = (195–200, 195–200). The circle located in the lower-left corner of each figure represents the starting point, and the blue box in the upper-right corner represents the goal. The light red area indicates the recommended area based on the D*Lite algorithm, and the colored lines are the paths searched by the DWA. The color of the lines indicates the time cost by the algorithms, thereby demonstrating the real-time efficiency of the algorithms. The red circles indicate the key points planned by the A* algorithm in A_DWA. The gray-colored grids represent the driving risk for the agent. Grids with risk levels 0, 1, 2, and 3 are passable, while grids with risk level 4 are obstacles and cannot be traversed by the agent.
To visually compare the performance of the algorithms in various aspects, the following abbreviations are defined for the concepts: the length of the path (LOP), the number of risk grids the planned path passes through (NORG), the accumulation of the risk of the grids the planned path passes through (AROG), the global planning time in the fusion algorithm (GT), and the time cost for the algorithms (Time). The traditional D*Lite algorithm (T_DL), the improved D*Lite algorithm (I_DL), the traditional DWA (T_DWA), the improved DWA (I_DWA), the fused algorithm of A* and DWA(A_DWA), and the proposed algorithm in this paper (D_DWA) are compared based on five specific criteria, as shown in
Table 4,
Table 5,
Table 6 and
Table 7. Among them, since the T_DL and the I_DL algorithm plan the regional paths according to the coarse map (20 × 20), it is not possible to calculate the LOP, and the NORG and AROG of these two algorithms only calculate the number of large grids in coarse maps. In addition, the T_DWA does not consider the road risk level and directly traverses the risk area, so the path length has no reference meaning and is not the object of comparison.
As shown in
Figure 5 and
Figure 6, all scenarios can reach the goal. However, T_DL (
Figure 5) does not consider the risk, and although it can avoid obstacles on the path, it directly traverses the high-risk area. On the other hand, I_DL (
Figure 6) avoids unnecessary high-risk areas and yields a higher-quality path. I_DL provides a global solution in the risk environment, reducing NORG by 57% and AROG by 64% compared to the T_DL in environment (A). When obstacles are introduced, I_DL reduces NORG by 50% and AROG by 64% compared to T_DL. Therefore, I_DL significantly outperforms T_DL. Both algorithms have comparable time costs.
As shown in
Figure 7,
Figure 8,
Figure 9 and
Figure 10, all algorithms can reach the goal. The T_DWA (
Figure 7) directly traverses areas with high risk levels. The I_DWA (
Figure 8) can plan paths through the area with low risk while avoiding obstacles. In environment (A), the NORG of the I_DWA increases by 25% compared with T_DWA. This is because the T_DL directly traverses the high-risk grids, while the I_DWA explores part of the low-risk grids with risk level 1 to maintain the speed and driving angle of the agent. Despite the increase in NORG, the AROG decreases by 11%, indicating higher path safety. In environment (B), using I_DWA results in a 25% reduction in NORG and a 30% decrease in AROG, while maintaining a similar LOP and time cost to that of T_DWA. The path quality of I_DWA is significantly superior to that of T_DWA.
In terms of path quality, D_DWA shows more improvement compared to I_DWA, while A_DWA demonstrates less improvement and exhibits erratic performance. As can be seen from
Table 6 and
Table 7, the LOP of I_DWA, A_DWA, and D_DWA do not differ significantly in either environment (A) or environment (B). In environment (A), A_DWA’s NORG and AROG decrease by 20% and 12% relative to I_DWA, respectively. Meanwhile, D_DWA achieves a remarkable reduction of 44% in NORG and 36% in AROG. In environment (B), A_DWA experiences a 14% increase in NORG and a 6% increase in AROG relative to I_DWA, and the path quality is slightly worse. This is primarily attributed to the global algorithm A*, which strictly constrains the actions of local planning algorithms when obstacles are encountered. On the other hand, D_DWA achieves a 39% reduction in NORG and a 35% reduction in AROG relative to I_DWA. Regarding time cost, A_DWA takes 27.4 times and 34.4 times as long as I_DWA in environment (A) and (B), respectively. In contrast, D_DWA exhibits a significantly lower time cost, with ratios of only 2.0 and 2.1 times. The main reason for this phenomenon lies in the difference in time consumed by the global planning algorithm. As shown in
Table 6 and
Table 7, the GT for D_DWA is much smaller than that for A_DWA. This difference in time is comprised of two main factors: on one hand, the global planner D*Lite consumes less time than the A* algorithm; on the other hand, the bi-layer map significantly reduces the time cost of global planning by decreasing the number of map nodes (from 200 × 200 = 40,000 to 20 × 20 = 400). According to an experiment conducted by the authors, the time expenditure of the global planner D*Lite without bi-layer map processing for D_DWA is 59.28 s and 57.19 s for environment (A) and (B), respectively, which is reduced to 1.2 s and 1.09 s after applying bi-layer map processing.
When comparing the same global–local fusion algorithm, D_DWA, with A_DWA, it becomes evident that D_DWA holds a significant advantage. In environment (A), D_DWA reduces NORG by 29.6%, AROG by 27%, GT by 99.7%, and Time by 92.5% compared to A_DWA, albeit with a 1.3% increase in LOP. In environment (B), although LOP increases by 1.2% when comparing D_DWA to A_DWA, NORG decreases by 46.7%, AROG decreases by 39.3%, GT decreases by 99.7%, and Time decreases by 93.8%.
5.2. Dynamic Environment
In the dynamic environment experiment, two cases are considered.
Figure 11 depicts the simulated environment for Case 1 and Case 2. As before, the circle in the lower-left corner represents the starting point for path planning, the blue box in the upper-right corner represents the goal, and the gray grid indicates the regional risk level. The pink line shows the obstacle’s direction of movement. The varying colors on the color bar to the right represent the time durations consumed by both the algorithms and the obstacles. The other elements in the figure are consistent with
Section 5.1. The definitions of the case are as follows:
Case 1: Uniform obstacle
As shown in
Figure 11a, the environment includes 20 dynamic obstacles that move toward the lower left at a speed of 0.1 m/s.
Case 2: Random obstacles
As shown in
Figure 11b, the environment includes 30 dynamic obstacles. To test dynamic performance, the obstacle locations are randomly set within an x range of 50 to 150 and a y range of 50 to 150. The obstacle velocities are randomly set in a range from 0.0 m/s to 0.25 m/s.
Case 1 was conducted once, and Case 2 was repeated 10 times.
Figure 12,
Figure 13 and
Figure 14 show the trajectory results for I_DWA, A_DWA, and D_DWA in both Case 1 and Case 2.
Figure 12b,
Figure 13b, and
Figure 14b show the trajectories of the agent 10 times only.
Table 8 contains the LOP, NORG, AROG, Time, and GT data for the three algorithms in Case 1.
Table 9 lists the average length of the path (A_LOP), the average number of rough grids (A_NORG), the average accumulated risks of agent (A_AROG), the average Time (A_Time), and the average GT (A_GT) for the three algorithms in Case 2.
Based on
Figure 12,
Figure 13 and
Figure 14, it becomes evident that all three algorithms can successfully choose the optimal path while taking into account both risk and length. However, it is worth noting that the I_DWA, due to its absence of global planning guidance, frequently generates longer paths and traverses risk areas, particularly in complex dynamic obstacle environments. The paths generated by the A_DWA are more distorted due to its reliance on local path planning. The global planning algorithm cannot adjust the recommended path in real time, causing the local planner to initiate obstacle avoidance only upon detection. As a result, the robot requires larger turning angles to circumvent moving obstacles, producing a more tortuous and longer path. In contrast, the global planning algorithm of D_DWA can avoid obstacles in advance, reducing reliance on local planning and yielding smoother paths with lower risk and enhanced quality.
Based on the data presented in
Table 8 and
Table 9, it is clear that the three algorithms perform differently in dynamic obstacle environments, with D_DWA performing best, followed by A_DWA, and then I_DWA. In uniformly dynamic obstacle environments, A_DWA significantly reduces LOP, NORG, and AROG by 17.94%, 34.58%, and 49.19%, respectively, compared to I_DWA. However, A_DWA also increases time cost by 286.68%. In contrast, D_DWA slightly increases LOP by 1.11% compared to A_DWA but reduces NORG, AROG, and Time by 28.57%, 26.60%, and 57.11%, respectively.
In the random dynamic obstacle environment, the path lengths generated by the three algorithms show minimal disparity, while exhibiting significant variations in path risk assessment. A_DWA demonstrates a noticeable improvement in path quality, albeit at the cost of a substantial increase in time consumption. In contrast, D_DWA not only achieves further enhancement in path quality but also effectively reduces time consumption compared to A_DWA. Specifically, when compared to the baseline I_DWA, A_DWA achieves reductions of 12.85% and 14.76% in A_NORG and A_AROG, respectively, while incurring a 257.27% increase in time cost. Conversely, D_DWA achieves reductions of 39.74%, 36.22%, and 63.21% in A_NORG, A_AROG, and A_Time, respectively, compared to A_DWA.
While acknowledging the inherent differences between the global planner algorithms, it is essential to explore the time discrepancies observed between D_DWA and A_DWA. These discrepancies can be attributed to the inherent efficiency differences between the A* and D*Lite algorithms, as well as the application of bi-layer map processing. To investigate how bi-layer map processing enhances time efficiency, this study conducted experiments using the D*Lite algorithm without bi-layer maps in Cases 1 and 2 and recorded the respective time cost. In Case 1, the time cost of D* Lite without bi-layer maps processing was 176.98 s, which is a 40.46% reduction from the 297.26 s recorded with A_DWA’s global planner. After applying bi-layer maps, the time further decreased to 132.64 s, achieving an additional 25.05% reduction. In Case 2, the time cost of D*Lite without bi-layer maps processing was 200.36 s, which was reduced to 116.52 s after application of bi-layer maps processing, representing reductions of 33.13% and 41.84%, respectively. Research findings demonstrate that the application of bi-layer map processing in the D_DWA algorithm effectively enhances its time efficiency.
5.3. Complex Environment
In the dynamic simulation described in
Section 5.2, the risk environment was static, lacking dynamic changes. To more effectively evaluate the proposed algorithm’s adaptability to complex environments, this study introduces a more complex simulation experiment. This new experiment features a dynamic risk environment with moving obstacles, significantly enhancing the realism and complexity of the test scenario. Additionally, to comprehensively compare the algorithm’s performance, we included the A_Q algorithm (as described in [
28]) in our comparison. The A_Q algorithm combines the A* and Q-learning, and its ability to handle complex environments has been verified in real settings. We use risk level to replace the terrain cost metric used in the original paper and evaluate performance under the same environmental conditions. Note that since the goal is too close to the map boundary, Q-learning will regard it as an obstacle and cannot approach it. Therefore, we set the
sgoal = (
xgoal,
ygoal) = (190–200, 190–200). In the A_Q algorithm, we use red lines to represent the global planning results and colored lines to represent the final planned paths. The other elements in the figure are consistent with
Section 5.1.
The dynamic environment includes four distinct risk maps (Risk Maps A to D, as shown in
Figure 15). Given the significant time differences between the three algorithms, transitions between these maps are determined by the greater percentage of the
x or
y coordinates relative to the map’s total length. The specific transition rules are: coordinates less than 15% switch to Risk Map A; those between 15% and 30% switch to Risk Map B; coordinates from 30% to 45% switch to Risk Map C; and coordinates exceeding 45% switch to Risk Map D. Dynamic obstacles are configured exactly as in Case 2. Blue and red pentagrams mark the positions of the agent using the A_DWA and A_Q and the agent using D_DWA, respectively. Furthermore, the starting points, goals, and risk levels are consistent with previous setups.
Figure 16,
Figure 17,
Figure 18,
Figure 19 and
Figure 20 display the trajectories of the A_DWA, A_Q, and D_DWA algorithms at various stages within this complex environment, while
Table 10 details the LOP, NORG, AROG, GT, and Time metrics for the three algorithms.
As illustrated in
Figure 16,
Figure 17,
Figure 18,
Figure 19 and
Figure 20, all three algorithms successfully reached their goals, yet their performances varied significantly in response to environmental changes. D_DWA and A_Q demonstrated active dynamic adjustment capabilities, whereas A_DWA did not. When the environment changes, A_DWA’s target point imposes greater restrictions on the agent, hindering its ability to navigate complex environments. A_Q can switch to local path planning promptly to manage complex environmental changes, but it is limited to making small adjustments due to its inability to anticipate and plan dodges. D_DWA dynamically adjusts its recommended area using its global algorithm, D*Lite, effectively simplifying the challenges faced by local algorithms. For instance, in Map B, the risk environment is less severe than in Map A. D_DWA alters the recommended area to minimize risk exposure. As the risk level increases from Map B to Map C, and further intensifies from Map C to Map D, D*Lite continues to dynamically adjust its recommended area, consistently decreasing the risk associated with the recommended area and ensuring the final algorithm path is of lower risk.
Table 10 shows that D_DWA is significantly better than A_Q, and A_Q is slightly better than A_DWA. Although the path length planned by D_DWA is 7.43% longer than that used by A_DWA, it achieves reductions of 71.95%, 80.34%, 26.98%, and 35.61% in NORG, AROG, GT, and Time, respectively, when compared to the A_DWA algorithm’s path. This indicates that D_DWA significantly lowers risk despite opting for a slightly longer path. In addition, compared with the A_Q algorithm, it achieves reductions of 1.34%, 67.14%, 78.70%, 34.85%, and 37.63% in LOP, NORG, AROG, GT, and Time, respectively. The superior performance of D_DWA can be attributed to its global algorithm, which dynamically adjusts the recommended area, unlike A_DWA’s global algorithm that does not implement real-time changes. As a result, A_DWA’s local planning algorithm reacts only when encountering risk areas, often missing optimal timing to bypass risk areas. Although the A_Q algorithm can cope with dynamic changes to a certain extent, it is not as effective as D_DWA because it cannot avoid risks in advance. Therefore, the path quality of D_DWA is superior to that of the A_DWA algorithm. The simulation results confirm the effectiveness of the proposed method.
It is worth noting that the change in the color of the lines illustrates the distribution of time consumption. Due to the large number of map nodes, the A_DWA and A_Q algorithms’ global time consumption is concentrated in the initial stage. In contrast, the D_DWA algorithm’s global time consumption is spread throughout the entire process. This distribution improves real-time performance and reduces robot waiting time. Therefore, the D_DWA algorithm has lower time costs and higher real-time efficiency compared to the A_DWA algorithm and A_Q algorithm.
To sum up, in complex environments, the D_DWA algorithm outperforms other compared algorithms in terms of planning speed, path quality, and path flexibility. The global–local coupled path planning framework proposed in this study significantly reduces the consumption of computing resources and enhances robustness by optimizing the fusion method and interactions between coupled algorithms.
Compared with existing global–local coupled path planning algorithms, our research introduces a novel perspective for enhancing efficiency. We focus on optimizing how global and local algorithms are integrated to maximize their respective capabilities. The effectiveness of this method in managing complex environments has been confirmed through simulation comparisons. For robots, applying this method can improve the autonomy and safety of tasks, especially for robots that need to operate in unpredictable environments.